Globalization :: Migrating Character Data Using A Full Export And Import

Jul 23, 2013

I have a database in my local machine that doesn't support Turkish characters. My NLS_CHARACTERSET is WE8ISO8859P1, It must be changed to WE8ISO8859P9 , since it supports full Turkish characters. I would like to migrate character data using a full export and import and my strategy is as follows:

1- create a full export to a location in network,

2- create a new database in local machine that it's NLS_CHARACTERSET is WE8ISO8859P9 (I would like to change NLS_LANGUAGE and NLS_TERRITORY by the way)

3- and implement full import to newly created database. I 've implemented first step, but I couldn't implement the second step. I 've created the second step by using toad editor by clicking Create -> New Database but I can not connect the new database. I must connect new database in order to perform full import.

DetailsNLS_LANGUAGE.....................AMERICANNLS_TERRITORY.....................AMERICANLS_CURRENCY..

View 8 Replies


ADVERTISEMENT

Export/Import/SQL Loader :: Full Dump But Data Only Import

Feb 15, 2013

When I do the import the of succeeding dump, I drop the existing schema "SQL> drop user username cascade;" and import dump by " impdp system .... ". I would like to import a dump to an existing instance but only data import and will leave the current packages and other metadata untouched and unchanged on the said existing instance.

1. Do i need to drop user before the import if my requirements are the above?

2. If i need to drop user, what should be script.

3. For the import itself, what parameter should i use?

4. What are the necessaries I need to consider before doing the import.

View 12 Replies View Related

Export/Import/SQL Loader :: Migrating Data From Solid Database To Oracle - Flat Files

Aug 14, 2012

I am migrating data from a Solid Database to Oracle, I am using Flat Files to do that.

1.- I download the data to flat files from Solid
2.- I move the files to Oracle server
3.- I upload the data to Oracle

Now, I have done the 90% of the data base, but I have found some tables that has description columns and in this description the users writes enters, so when I try to upload the data to Oracle SQL loader cannot recognize this characters.

Example:

'25','0.','5.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''
'26','0.','2.','0.','0.','0.','0.','3.','0.','0.','0.','0.','0.','',''
'27','0.','1.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''
'28','0.','1.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''
'29','0.','38.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''
'30','0.','13.','0.','0.','0.','0.','0.','6.','0.','6.','0.','0.','|SE RECHAZA B20CS50SNW ^M
^M
SE RECHAZAN CINCO PZAS ^M
DOS MOD. HSC15I41EH,DOS MOD. HSK15I41EH |Agregó: 06/06/2009 12:22:50
|','DEV. A PROV.'
'31','0.','50.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''
'32','0.','9.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''
'33','0.','2.','0.','0.','0.','0.','0.','0.','0.','0.','0.','0.','',''

How can I solve this ?

View 3 Replies View Related

Export/Import/SQL Loader :: How To Export Full Dump And Metadata Of Particular Table

Apr 9, 2013

let us consider mytest schema is having 6 tables

tname tabtype
myt table
myaxpertlog table
abb table
ccc table
ddd table
xxx table

now from this schema i want full dump and also from myaxpertlog table i required metadata only not records.

c:> export mytest/log file=20130409mytest0904pm.dmp tables=(myaxpertlog) rows=n

if i tried i am get only one table but it does have records.

View 6 Replies View Related

Export / Import - Character-set Is Different?

Apr 18, 2012

I taken a export through exp and i want to imp in some othet database.. But here character-set is different.

Is it import successfully or not..?

View 1 Replies View Related

Export/Import/SQL Loader :: Perform Full Exp / Imp Of Oracle 11g Database?

Jul 30, 2012

i want to perform full export + import of an oracle 11g database as fast as possible. i was thinking to perform the exp+imp on the same command.in exp i can perform something like this :

mknod /oracle/migration/exp_pipe p

exp '/ AS SYSDBA' file= /oracle/migration/exp_pipe full=y | imp system/***@oracle_db file= /oracle/migration/exp_pipe full=y

i know that i can do both action in impdp when using a dblink, but the problem is that some objects in the database cannot be copied via a dblink. the question is if there's a corresponding datapump command to the old exp+imp command i presented.

View 5 Replies View Related

Globalization :: Character Encoding - LATIN-1 Character In UTF-8 DB?

Feb 9, 2013

I am using C++ OCI LIB, to insert some report data from remote OCI client to oracle 11 server. This data is read by another process to create the report.The DB CHARSET is UTF-8. But the report tool expects the data to be ISO08859-1 encoded. So while inserting the data into the database i specify the following LANG and CHARSET for my table colulmn in client:

The TARGET DB CHARSET is UTF-8
NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P1
size_t csid = 871; // UTF-8
OCIAtrSet((void *) bnd1p, (ub4) OCI_HTYPE_BIND,
*(void *)&csid*,
(ub4) 0,
(ub4)OCI_ATTR_CHARSET_ID, errhp);

This solution works for almost every case of ASCII and Extended ASCII Charest but we are facing issues if we have few specific characters to be inserted.f we are trying to insert single beta character [β] through client, the data goes empty to the column.

Beta Character details:
DEC     OCT     HEX     BIN     Symbol           Description
223     337     DF     11011111     ß     Latin small letter sharp s - ess-zed

DB Output after insert single β:
select rawtohex(NAME) from PERSONS where EID=333;

RAWTOHEX(NAME)
---------------------------

But if the string is *"ββ"* everything work fine:
DB Output for "ββ":
select rawtohex(NAME) from PERSONS where EID=333;

RAWTOHEX(NAME)
---------------------------
DFDF

View 6 Replies View Related

Export/Import/SQL Loader :: How To Validate Special Character After Migration

Mar 26, 2013

We have migrated the database from Oracle 10.1.0.3 to 10.2.0.5 by using conventional export and import. (OS - HP.Uinx).,After the migration some characters (like é or à or è) have not been migrated correctly.

how to validate this characters after migration?

Note:

All NLS parameters are same in both source and target database

View 8 Replies View Related

Export/Import/SQL Loader :: Export And Import Of Data Not Table And Data?

Sep 11, 2012

Export and import of data in oracle forms...i have created 02 boutons one for export his trigger like this:

eclare
alrt number;
v_directory varchar2(200) := 'c:ackup'; --- that if the C Drive not the Drive that the windows had installed in it.
path varchar2(100):='back_up'
||to_char(sysdate,'dd_mm_yyyy-hh24_mi_ss');
v_exp varchar2(200) := 'exp hamada/hamada2013@orcl file = '
||v_directory
||''
||path
||'.dmp';
[code]....

this code is correct he expot not only the data but also the creation of the table ....for exemple i do export and everything is good until now and i find the .dmp in the folder backup .. but when i deleted all data from my app and try to import this .dmp iit show me error it tell me thet the table phone is already created...just export the data of phone not the creation of table and data ???? or how can i import just the data from this .dmp ??

View 3 Replies View Related

Globalization :: Database Character Set

Jun 6, 2012

We have production DB 10g with character set US7ASCII. This DB stores Arabic data and English data.Production DB located in HP unix Operating System.

When I query data from DB through SQL developer data is shown as Junk or Unknown characters(Square Boxes).

Client (Workstation from where query is issued from SQL develope- Windows XP OS) Settings: NLS_LANG = AMERICAN_AMERICA.US7ASCII

In Client workstation Oracle 10g client is installed from where I used to query data through SQL developer. The problem is I am unable to see Arabic characters in the sense that it is displayed as Junk character. However English characters and Eneglish numeric values are displayed properly.

I tried below way to make sure that data is not corrupted: Converted "Name" column to hex value (rawtohex) and displayed its HEX value. Executed below query in UTF-8 DB.

select UTL_I18N.RAW_TO_CHAR(hex_value_of-name) from dual;

This displayed Arabic name properly in UTF8 DB.

Character set for this production DB can not be changed at this time. There are many applications which is based on this DB. All these applications are well capable of converting Junk data to Arabic to display in application.

My concern is: What I should required to do to view Arabic data properly through SQL developer? Is there any settings needs to be done at my client workstation?

View 15 Replies View Related

Globalization :: Character Set For Local Language

Dec 2, 2012

I am using oracle 10 g database on windows xp. I have backup of data contains data in local language (Marathi). I want read this data in oracle itself.Which character set need to choose?

View 6 Replies View Related

Globalization :: Bad Character Set In Dump File?

Jul 23, 2013

IMPDP-ing a dump file that someone has handed me over into Oracle XE results in special characters, i.e. Umlauts, being messed up.  

In a hex editor, the dump file shows  a) the token WE8MSWIN1252 near the beginning, but b) Umlauts obviously being encoded in DOS 850, for example "König" is encoded as 4b 94(!) 6e 69 67. Does this prove that the dump file is badly formatted and that I have to resign myself to the complicated approach mentioned at the end of [URL]...

View 4 Replies View Related

Export/Import/SQL Loader :: How To Import Data From Excel File To Table Through Procedure

Jul 2, 2012

How to import data from excel(.xls) file to data base table

I have excel sheet(.xls) data details, I neet to upload details to data base table using procedure

excel sheet is not CSV file, so SQL Loader is not using

any alternative solution for this issue

View 3 Replies View Related

Export/Import/SQL Loader :: How To Filter Some Illegal Rows When Import Data

May 24, 2013

I want to import data in a csv file by SQL Loader.

but , I don't want to import some illegal rows when the column 'name' is null

how can I modify the SQL Loader ctrl file?

View 1 Replies View Related

Export/Import/SQL Loader :: Error During Data Pump Import With Developer

Sep 17, 2012

I try to transfer data from one database to another one through data pump via SQL Developer (data amount is quite important) exporting several tables. Tables export is doing fine, but I encounter the following error when I import the file (I try data only and data + DDL).

"Exception: ORA-39001: argument value invalid dbms_datapump.get_status(64...=
ORA-39001: argument value invalid
ORA-39000: ....
ORA-31619: ...

The file is in the right place, data pump folder of the new database. User is the same on both base, database version are similar.

View 4 Replies View Related

Export/Import/SQL Loader :: Export Data Pump On Remote Location

Oct 11, 2012

I have one prod server (11.1.0.6.0) servers on Windows 2003 R2 64 bit.

Server Name (PRODDB)

I do not have access to that prod server , i want to take one export data pump from my client machine and due to space issue in prod server , i want to keep dump file in my client machine itself. i can take traditional export and keep the dump file in my client machine but i do not know how to achieve the same via data pump ...

How to generate dump file in client machine itself via data pump ?

View 18 Replies View Related

Export/Import/SQL Loader :: Export Data Program (ociuldr) Cannot Run In 64bits Win2008 Environment

Jun 29, 2012

export data program "ociuldr" can not run in 64bits win2008 environment. Where can i download 64 bit version of "ociuldr" program. I have read some article . The article mention that the export data program ociuldr.exe need to recompile into 64 bits version. Finally I also want to ask the import data program sql loader "sqlldr.exe". Can it run in 64 bits environment. Where can i downlaod 64 bits version of "sqlldr" program.

View 1 Replies View Related

Export/Import/SQL Loader :: How To Import Only Data From The Dmp File

Feb 11, 2013

I received dmp file , and i want to import only data from that file ?

How can we achieve that in oracle Oracle 11.2.0.3

View 5 Replies View Related

Globalization :: SQL Loader Loads Special Character Incorrectly

May 31, 2013

We are trying to load some xml files via

sqlldr user/pswd@xe control='C:xmlldr.ctl' data='listoffiles.dat'

where xmlldr.ctl is the following:

load data
CHARACTERSET WE8ISO8859P1
replace
into table LOADER_CTG_PROTOCOL
(

xmlfile lobfile(nct_code) terminated by eof,
nct_code char(15) "substr(:nct_code,1,11)",
created_dt sysdate
)

I was playing with different CHARACTERSET, but some special characters e.g. "greater than or equal" do not get loaded/displayed correctly in the database. Also tried changing NLS_LANG registry key and following some advices in the Oracle doc

[URL].......

View 1 Replies View Related

Globalization :: Convert Chinese Character From US7ASCII To AL32UTF8?

Sep 17, 2012

Source Database: AMERICAN_AMERICA.US7ASCII
Target Database: AMERICAN_AMERICA.AL32UTF8

From the source database, the chinese characters are stored in some schema table. From the csscan result, there are convertiable, truncate, data lossy character. So, I have tried to use exp/imp for the conversion. However, all chinese characters are invalided and cannot be read anymore. How can I convert them from US7ASCCI to UTF8 database?

Also, I have tried build up another database with AMERICAN_AMERICA.ZHT16MSWIN950. The exp/imp is used for conversion again. The chinese characters are readable in AL32UTF8 database.

- source database (US7ASCII)
export NLS_LANG=AMERICAN_AMERICA.US7ASCII
export LANG=AMERICAN_AMERICA.US7ASCII
exp userid='/ as sysdba' file=export.dmp full=y

- target database (AL32UTF8)
export NLS_LANG=AMERICAN_AMERICA.US7ASCII
export LANG=AMERICAN_AMERICA.US7ASCII
imp userid='/ as sysdba' file=export.dmp full=y ignore=y

Result:
from US7ASCII to AL32UTF8:
the chinese characters cannot be read

from US7ASCII to ZHT16MSWIN950:
the chinese characters cannot be read

from ZHT16MSWIN950 to AL32UTF8:
the chinese characters can be read

How can I convert the chinese character from US7ASCCI to UTF8 database?

View 6 Replies View Related

Globalization :: Non-English Character In Oracle 10g Express Edition

Oct 5, 2012

I have a table. It's name is INSTITUTION. It has a NUMBER INS_ID and NVARCHAR2(50) INS_NAME . INS_NAME can contain Turkish characters, such as "ğ,ü,ş,ç,ö". According to business logic, there can not be a repetition on the INS_NAME.User will enter institution name from a textbox in ASP.NET , and I check this name in database from c sharp code, if there is no repetition, we will add this record.

The problem is; when user enter a instition name that contains Turkish character, there is a duplication. If there is a instition name is *"su işleri"* , the both query; SELECT * FROM INSTITUTION WHERE INS_NAME = *'su işleri'*; and SELECT * FROM INSTITUTION WHERE INS_NAME = *'su isleri'*; returns no result, even though there it is.But if instition name is "oracle corporation" (there is no Turkish character) it query successfully. I have the same problem in Toad for Oracle 11.5.1.2. When I query database from toad SELECT * FROM INSTITUTION, the phrase *"su işleri"* has appeared. But when I query SELECT * FROM INSTITUTION WHERE INS_NAME = *'su işleri'*; , there is again no result.When I connect oracle database directly and perform the query SELECT * FROM INSTITUTION , the phrase *"su isleri"* (not *"su işleri"* ) has appeared.

Here are the language settings of the database:

National Language Support
National Language Parameter Value
NLS_CALENDAR______________GREGORIAN
NLS_CHARACTERSET__________WE8MSWIN1252
NLS_COMP__________________BINARY
NLS_CURRENCY______________TL
NLS_DATE_FORMAT__________DD/MM/RRRR
NLS_DATE_LANGUAGE________TURKISH
NLS_DUAL_CURRENCY_________YTL
[code]....

View 8 Replies View Related

Globalization :: Junk Character Insertion In Oracle Table

Feb 6, 2013

How to avoid Junk character insertion in oracle table. I have prepared scripts like this Say

customer - info

After insertion the data is inserted like below in production

Customer ¿ info

We are using command prompt for script execution in production environment. I am using PLSQL developer and SQL developer for development. i cannot see junk data in PLSQL developer and latest SQL developer , but its caught in old version of SQL developer. Also in Application also i can able to figure out junk data.

View 6 Replies View Related

Globalization :: Forms / Reports Chinese Character Support?

Jul 4, 2013

We have an existing db (10.2.0.4.0) and forms (11.1.2.1.0) application, that we're trying to extend to support Chinese characters. We're looking to add some unicode (nvarchar2) columns to existing tables, rather than converting the whole db charset. I've pasted my environment settings below. What I've found so far in trying to create a local (ie. running the form in Builder with local weblogic running) test form, is that I can insert the chars ok (using plsql developer) and the test form can display them correctly, but cannot write them back to the database. They appear as upside down question marks in any records the form has created. 

Env variables:NLS_LANG = .UTF8 Database:NLS_CHARACTERSET = WE8MSWIN1252NLS_NCHAR_CHARACTERSET = AL16UTF16 1)

So, how to get the form to write the characters back into the database correctly? 2) The chinese chars will only be relevant to a few forms inside the app, are there any settings local to the form that will enable unicode support, rather than setting at OS level. ie, an alter session, or equivalent? 3) Oracle Reprts doesn't appear to have an nchar datatype unlike Forms, is there anyway to get Reports (generating PDFs), to include Chinese?

I now have the chars writing back to the db ok. If you do it via an INSERT statement from inside the form, it doesn't work. It appears the value is sent to the db in the normal charset rather than the national charset, and it's written as a question mark. If you pass the value from the form into a back end stored proc though (which does the insert) it works okay.

View 3 Replies View Related

Export Data Using Escape Character

Jul 2, 2012

I have got a requirement where i need to export data from oracle with escape character.

eg. I am using a delimiter 237(í) and if the same character is present in data it should be escaped by escape character eg. /.

Once this file will get created i need to load this file in Netezza database which supports escape character.

Data in oracle table
FirstName     Lastname     Designation
abc     xyz     mnz
def     ghío     pqr

Data should be exported like below
FirstnameíLastnameíDesignation
abcíxyzímnz
defígh/íoípqr

View 3 Replies View Related

Export/Import/SQL Loader :: How To Skip Database JOBS During Export And Import

Aug 7, 2012

1) Is there a way to skip database jobs while exporting (EXPDP) ?

2) Is there a way to skip database jobs while importing (IMPDP) ?

View 3 Replies View Related

Export/Import/SQL Loader :: Why Do Export-import Require Temporary Tablespace

Aug 9, 2012

Why do export-import require temporary tablespace? Since export-import do behave like DMLs, when does temporary tablespace be needed by datapump utility?

View 2 Replies View Related

Export/Import/SQL Loader :: Import Table Without Messing Up Existing Data In Table

Sep 6, 2012

table already exist & its little data too, may have to imp rest of lost data, is this the right command?

imp SYSTEM/password FILE=file.dmp FROMUSER=black TOUSER=blake TABLES=(vcr_mappings, tablename2) ignore=Y CONSTRAINTS=n

scenerio2 (if have to drop & recreate the entire table) is this the right command?

imp SYSTEM/password FILE=file.dmp FROMUSER=black TOUSER=blake TABLES=(vcr_mappings, tablename2) ignore=Y

just for single table imp

View 2 Replies View Related

Server Administration :: Cyrillic Character Set - Import A Table Of Data To Another Database

Aug 31, 2011

I'm try to import a table of data (character set: CL8ISO8859P5) to another database (character set: AL32UTF8) using exp/imp utility.After the import, all Cyrillic text was corrupted!

View 5 Replies View Related

Export Data-files To Another Linux Box To Import In Another?

Sep 11, 2012

I am a fairly new dba and we had one of our HP Array cluster's crash to the point where oracle will not startup or mount anymore. I can access the datafiles on the linux server however. Is there a way to export the datafiles to another linux box to import in another database or have I pretty much lost everything (we do have an RMAN backup however it is in a remote location and the only person that knows the password to it is unavailable.

View 4 Replies View Related

Export/Import/SQL Loader :: Data Pump For Migration

Sep 9, 2013

We should migrate our 10gR2 single-instance database with conventional file system to a two-node 11gR2 RAC on ASM (on same Windows Server platform…). 

How can I migrate my production database using data pump? I have full data pump export from target but I don’t know how to import, whether the scheme after scheme, full import, do I need to first create manually tablespaces on destination, whether to exclude the index, constraint, statistics?

View 3 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved