Server Utilities :: Character Set Error While Importing

Jun 9, 2011

I have oracle 10.2.0.4.0 installed on Window server 2008 [Machine A] and oracle 10.2.0.1.0 on windowx xp [Machine b]. Now I have taken the export of database on windows server 2k8 [Machine a] by puting the entry in the tnsname.ora file of Windox XP [Machine b].

Now when I am importing on the same machine I am getting the below mentioned error:

C:Documents and Settingsdsharma>IMP FROMUSER=SYSTEM TOUSER=ESCDBO FILE='D:sharevcc53_0106.dmp' LOG='D:sharevcc53_0106_IMP.LOG' ignore=y

Import: Release 10.2.0.1.0 - Production on Thu Jun 9 20:08:11 2011
Copyright (c) 1982, 2005, Oracle. All rights reserved.

Username: system
Password:

Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production..With the Partitioning, OLAP and Data Mining options

Export file created by EXPORT:V10.02.01 via conventional path
import done in WE8MSWIN1252 character set and UTF8 NCHAR character set
import server uses AL32UTF8 character set (possible charset conversion)
export client uses US7ASCII character set (possible charset conversion)
Import terminated successfully without warnings.

View 14 Replies


ADVERTISEMENT

Server Utilities :: Importing From 9i To 11g - Getting ORA-29339 Error Because Of Block Size?

Oct 5, 2010

We are working on migrating from 9.2.0.4 to 11.2 and we've set up a test machine so that we could test the install and the import (as well as test additional 11g features that we want to begin using).

So we created the database and created all of the tablespaces beforehand.

Our import command is

$ORACLE_HOME/bin/imp system/manager FULL=Y BUFFER=140000 FILE=/dbexport/Lhtech.exp VOLSIZE=2000M GRANTS=Y INDEXES=Y COMMIT=Y IGNORE=Y

However, when we run the import, we get the errors like so:

Import: Release 11.2.0.1.0 - Production on Tue Oct 5 15:01:19 2010
Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
Export file created by EXPORT:V09.02.00 via conventional path

[code]....

First of all, the block size in our "newly" created tablespaces is 8192...and these are obviously trying to recreate the tablespaces with a block size of 2048.

1) Why is it not ignoring these create tablespace commands when those tablespaces already exist?

2) how in the world do we get around the block size issue? We've tried nearly everything we could find, but we've still not had any luck.

View 8 Replies View Related

Server Utilities :: Error While Importing Dump File In Oracle 10g R1

Nov 1, 2012

While trying to import a schema using Data Dump, I am facing the following issue - UDI-00018 - Import utility version can not be more recent than the Data Dump server.Following is the version information of the source and target DB and the utilities :

Source DB server : 10.1.0.2.0
Export utility : 10.1.0.2.0
Import utility : 10.1.0.2.0

Target DB server : 10.1.0.2.0
Export utility : 10.2.0.1.0
Import utility : 10.2.0.1.0

View 5 Replies View Related

Server Utilities :: Importing 9i Dump Into 10g?

Jun 22, 2011

how can i import the oracle 9i dump file into 10g database, while iporting i get following error imp-00002 fail to open dump file

View 4 Replies View Related

Server Utilities :: Importing Data From 10G To 8i

Nov 9, 2010

I would like to know if a dump file taken from 10G could be imported into to an Oracle 8I database.

View 1 Replies View Related

Server Utilities :: Importing Table From 9i To 10g

May 31, 2012

I want to import my oracle 9i dump file into oracle 10g version. What should i do ???

View 2 Replies View Related

Server Utilities :: Exported Data Of One User Importing Into Another Schema At Another Server

Sep 15, 2010

I have exported data of one user an importing into another schema at another server. when i am trying to imoport it is working fine for quite no of imports into tables, but after some time it starts giving me below mention error...

IMP-00008: unrecognized statement in the export file:
<
IMP-00008: unrecognized statement in the export file:
<
IMP-00008: unrecognized statement in the export file:
<‏ے
IMP-00008: unrecognized statement in the export file:
+A
IMP-00008: unrecognized statement in the export file:
[code]...

View 6 Replies View Related

Server Utilities :: Importing Text Files?

Sep 16, 2011

I have a rather complicated process to import text files into my DB.I'm given thousands of files every day, separated by "," and with 80 fields each. With a bash script, I take the 45 fields I need and then split each file into x number of files grouping the rows by three fields.Then I use SQL Loader to insert them into de DB.

The problem is that now I must insert on two tables and the "WHEN" clause doesn't allow the use of > and <.

To make things a litle clearer take this text file (already splited and grouped and ready to be inserted):
...
1,1,135,1900,0,12,114,2011/08/25 17:19:00,135,...
1,1,135,1900,0,13,119,2011/08/25 17:19:00,136,...
1,1,135,1900,0,14,117,2011/08/25 17:19:00,137,...
1,1,135,1900,0,15,113,2011/08/25 17:19:00,138,...
1,1,135,1900,0,16,119,2011/08/25 17:19:00,139,...
...
When field 6 is higher or equal to 14, it must go to table a.When field 6 is lower than 14, it must go to table b.I can't use external tables as I'm in a different server.

View 1 Replies View Related

Server Utilities :: Importing Full Database With Different Name And Sid

Oct 20, 2011

Source:Oracle 11g 11.1.0.2 32 bit on Windows 2003,Destination: Oracle 11g 11.1.0.2 64 bit on Windows 2008

In our Source Database, we are using following to create full export

Source database_name=dbfour,sid=dbfour1

I am creating full export by using following

expdp system/psswd full=y dumpfile=expdp_%date:~0,2%-%date:~3,2%-%date:~6,4%.dmp logfile=explog_%date:~0,2%-%date:~3,2%-%date:~6,4%.log FLASHBACK_TIME="to_timestamp(to_char(sysdate,'yyyy-mm-dd hh24:mi:ss'),'yyyy-mm-dd hh24:mi:ss')"

In my destination DB with database_name=dbfive and sid=dbfive1, i am trying to import whole database by using the file created above and use following

impdp system/pwd@dbfive DIRECTORY=DATA_PUMP_DIR DUMPFILE=EXPDP_20-10-2011.DMP

Process started but after sometime it gave 1200 errors. Is it due to the Different Database name or Is it because i did not create table space in destination database.

View 3 Replies View Related

Server Utilities :: Tablespace For Importing Objects

Sep 20, 2010

Two different says in different documentation as below.

1) Import required same tablespace to fit the objects, otherwise it throws error.

2) Import by default uses the default tablespace to fit the objects, but some times it uses SYSTEM tablespace also.

confirm with the correct one.

View 8 Replies View Related

Server Utilities :: Getting Errors During Importing Data

Aug 14, 2012

I am getting following errors while importing data.

SCHEMA_EXPORT/TABLE/INDEX/DOMAIN_INDEX/INDEX
ORA-39083: Object type INDEX failed to create with error:
ORA-06550: line 2, column 1:
PLS-00201: identifier 'CTXSYS.DRIIMP' must be declared
ORA-06550: line 2, column 1:
[code]......

View 6 Replies View Related

Server Utilities :: Importing Into Lower Version Using (exp / Imp)

Jan 27, 2011

For importing higher version database exp dump e.g. 10g exp dump into 9i database it is suggested at many places "Always use a version of the EXPORT utility that is equal to the lowest version of either the source or the target database"

I understand from this is, I will be using 9i exp utility for exporting 10g database

I would like to know the ways it can be achieved.

1) Installing oracle 9i client on 10g database Do I still need to execute 9i catexp.sql after that? If answer to above is 'yes', I assume catexp.sql of 9i wont be created after 9i client is installed as such I will need to copy it from 9i database server and execute on 10g database server. Correct? For above question I have assumed catexp.sql has different contents for 9i and 10g

2) After configuring tns file in 9i database, can I use?

exp user/password@10gdbconnstring ....

(of course for small amount of data)

Is there any other option for achieving the same?

Also I understand export and import between any 9.* into any 9.* wont be issue e.g. import from 9.2 into 9.1 Similarly any 10.* to another 10.* wont be issue.

View 6 Replies View Related

Server Utilities :: Importing From DMP File Taken From Different Edition

Mar 2, 2010

i m having full export .dmp file taken from oracle 10.2.0.1 Enterprise Edition.

i need to reinstall 10.2.0.1 standard edition and want to create database using the dump file which is taken from the Enterprise edition.

View 3 Replies View Related

Server Utilities :: Importing Other Schema Object

Sep 19, 2010

I have 2 users (scott and Krishnan).

Krishnan wants to import the table, which was exported by scott.

1) but we should not run the script file of CATALOG.SQL.
2) we can't give the rights of imp_full_database from outside.

Also how user Krishnan can import that table to another user's schema?

View 3 Replies View Related

Server Utilities :: Importing Data From SQL Server To Oracle?

Mar 7, 2011

I'm importing data from SQL Server to Oracle. I used BCP to export the data from SQL Server. Below is the 1st record of table trlc from the csv file.

trlc.CSV

11032|100|Wman| | |2008-02-08| |

Using SQL Loader to import into Oracle:

TRLC table in Oracle database:

est_no varchar2(10) default ' '
right_no number(4) default 0
maj_auth varchar2(15) default ' '
weight varchar2(10) default ' '
idm_ht varchar2(8) default ' '
c_date date
P_tkt varchar2(5) default ' '

sqlldr user/pwd@db02 control=trlc.ctl log=trlc.log

trlc.ctL:

load data
infile 'trlc.csv'
replace
into table trlc
fields TERMINATED BY '|'
TRAILING NULLCOLS
(est_no,right_no,maj_auth,weight,idm_ht,c_date,P_tkt)

The rows get inserted successfully. But the result sets are different, for example: When I do a select in SQL Server,'select len(weight) from trlc;' , I get the length as 0. But when I do a select in oracle database, I get the length as 1. Also, the result set varies for the query below:

select * from trlc where weight=' ';

(SQL Server returns 1 row but Oracle returns no rows)

Do I need to mention any conversion code for the weight field to accept ' ' value?

View 12 Replies View Related

Server Utilities :: Too Many Errors While Importing Full DB Export?

Apr 22, 2013

i am trying to import full db export using datapump , i have too many errors for objects that is already exist . attached is the log file . thae steps i did so far

1- created the database .

2- imported the full db backup using

impdp system/xxxxxxx full=yes directory=datapump dumpfile=palbe_full_20130322.dp log=palbe_full_22042013.log

View 5 Replies View Related

Server Utilities :: Importing Limited Data In Oracle

Jun 29, 2010

I have received a dump, which i need to put on a newly created schema, there is a particular table with more then 4 million rows, and other tables have hardly few thousand rows.

I want to import it in a way where only 1000 rows get imported for this table and other tables do not get affected. Is there a way to do it?

Note: Tables in dump are more then 200.

View 7 Replies View Related

Server Utilities :: Exporting / Importing Partitioned Table

Sep 2, 2013

I am trying to export a partition of a table and import it to another database. I get the below error when I try to import.

ORA-14400: inserted partition key does not map to any partition

If I export the table(for that particular partition) and import the table(after dropping the table) in destination, the partitions and sub partitions are created without any problem.

The table is Range Partitioned and Sub partitioned in List. So I had to perform the below operation if I want to retain other data in the Destination table.

1. Drop the existing partition
2. Create the partition and sub partition, same as source
3. Execute imp

In fact I had to perform step#2, as if I split the partition also, the sub partition gets replicated in the new partition, which again throws the same error. Is there better way of managing the partitions and subpartition in destination with exp/imp utility, so that I need not perform step#1 and step#2 manually.

View 11 Replies View Related

Server Utilities :: Imp-00033 Table Not Found While Importing

Aug 5, 2010

while importing dump file into database,i received the following error.

imp -00033 table demo not found.

But i checked in dump file ,the table exist .

Below is the command i executed in unix box.

imp system/manager file=data.dmp log=data.log tables=DEMO feedback=100 buffer=1000 fromuser=TOM touser=maddy

View 3 Replies View Related

Server Utilities :: Importing Table Data Is Not Coming?

May 10, 2010

i take export of one table (export complet successfully without warnings) when i am going to import into prduction databae the data in the table no coming i past the table structure and import command and logfile for import.

import command

imp user/user@database file=C:Tempuserdrop_id10g _id_rac_ra_headertables.dmp log=C:Tempuserdrop_id10g _id_rac_ra_headertablesimp.log tables=t_id_rac_ra_header fromuser=user touser=user statistics=none ignore=yes

my table structure in production database is

CREATE TABLE t_id_rac_ra_header
(ra_company VARCHAR2(10) NOT NULL,
ra_key NUMBER NOT NULL,
ra_doc_type VARCHAR2(50) NOT NULL,
ra_doc_number VARCHAR2(25) NOT NULL,
ra_doc_date DATE DEFAULT SYSDATE NOT NULL,
ra_reserve_key NUMBER,

[code].....

View 4 Replies View Related

Server Utilities :: Importing Dump From Higher Database Version 2 Lower

Aug 3, 2010

We can Import dump from higher database version to lower database version.

View 26 Replies View Related

SQL & PL/SQL :: Importing To New Character Set And Its Affect On Encrypted Data

Nov 9, 2010

We are exporting from a 9i db to an 11g. During the migration, we are changing our character set from USASCII7 to AL32UTF8 so that the "extended" characters that our users like to put in text fields are stored and retrieved properly.

However, we've found a problem, and i'm not sure if Oracle has a method of dealing with. Searching this site and Oracle docs got me nowhere.

We store account #'s and credit card info in the DB encrypted with the dbms_obsfucation_tookit. We have an encryption key cross reference table that we use to store the key to un encrypt the data.

However, what we've found is that by importing these keys into our new character set database, the keys are no longer valid and can't be used with the DES3DECRYPT function to get the correct numbers out.

Is there a conversion utility or any tool that Oracle provides to maintain the encrypted datas "decrypt ability"? Worse comes to worse, we will have to write a script/procedure to decrypt everything on the 9i, import it to 11g, and then re-encrypt it.

View 8 Replies View Related

Server Utilities :: Difference In Character Sets Of Server And Client?

Mar 16, 2011

I've a question regarding difference of character sets, while taking a export(logical backup) of database on directly to server(linux RHEL 2.1 AS) and export on a client (windows xp prof machine, where only a oracle 9i client is installed). On server it seems to fine and okay, but on client node i'm getting following error for almost all tables.

EXP-00091: Exporting questionable statistics.

My question is :

[1] Is it creating any sort of problem, if later on i import the data which was taken from client node.

[2] Why there is a difference(marginal) in dump(.dmp) file size.

[3] Is there any way to overcome it, or it is the natural behave of it. Means not a problem.

[4] If i'm using a long or blob as datatype for some of my table,is they have any problem if i persist like above.

Additional Information about character sets On server node :

Export done in US7ASCII character set and AL16UTF16 NCHAR character set server uses WE8ISO8859P1 character set (possible charset conversion)

On client node :

Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set server uses US7ASCII character set (possible charset conversion)

View 7 Replies View Related

Server Utilities :: Getting Special Character While Loading?

Sep 19, 2013

I have a file which i am loading , the last column is being loaded with special character which looks like sqare shape character for all the rows.How can i prevent it to load as it is not present in the file

data in file

Department|2|Tranport for London|Rail & Underground|ER|ER|ER|ER|ER|ER|555555555|Owner 8|8

the data once loaded for the last column on which is for above example last column value 8 will be loaded with 8 and square share character.

View 5 Replies View Related

Server Utilities :: Importing From 9i To 11g With Compressed Extents Creating Extents On Target

Nov 2, 2010

We are migrating from a 9i db to 11g and we've been testing our apps on a similar (but not exact) machine as our production box.

Normally when we take a full export of the production data (on 9i) and import it into another 9i DB, the tables and indexes are created with the initial size large enough to hold the entire table. We also do our export with the compress extents param set to 'Y'.

However, we've noticed that when we import our data into the 11g DB, that tables are being created with multiple extents...sometimes up to 10 or 15. This seems to happen even with tables that don't even have extents on db that the export was taken from.

There ARE some differences in our 11g DB that i imagine might be the culprit, i've just been unable to narrow one of them down.

the differences i know of are:

a) the target DB has locally managed tablespaces while the source 9i DB had dictionary managed tablespaces
b) the block size is larger on the target 11g DB. 8192 vs 2048
c) the nchar character set on the source DB is AL16UTF16 and the target is UTF8 (we actually only have an nchar column in one of our tables...and also, the UTF8 setting was actually a mistake that we're correcting this weekend with a fresh DB and fresh import)

What would cause the import to produce all these extra tablespaces?

View 18 Replies View Related

Server Utilities :: Data Pump For Exporting And Importing Extremely Large Data Files

Sep 24, 2010

I am considering all of the capabilities and benefits of using Data Pump for exporting and importing extremely large data files. Would like to know if importing to tape is possible? If so, would the data be accessible if needed later?

View 4 Replies View Related

Server Utilities :: How To Ignore Rows With A Specific Character

Mar 10, 2010

If i have a CSV file that is in the following format

"fd!","sdf","dsfds","dsfd",
"fd!","asdf","dsfds","dsfd",
"fd","sdf","rdsfds","dsfd",
"fdd!","sdf","dsfds","fdsfd",
"fd!","sdf","dsfds","dsfd",
"fd","sdf","tdsfds","dsfd",
"fd!","sdf","dsfds","dsfd",

Is it possible to exclude any row where the first column has an exclamation mark at the end of the string.

i.e. it should only load the following rows

"fd","sdf","rdsfds","dsfd",
"fd","sdf","tdsfds","dsfd",

I am using Oracle 10.2

View 9 Replies View Related

Server Utilities :: IMP-00037 - Character Set Marker Unknown?

Mar 2, 2011

When I am trying to import a dump file which was exported from Oracle 10.2.0.4(installed on Solaris) in to Oracle 11.1.0.6.0(Installed on Windows 2003), I am getting the following error

IMP-00037: Character set marker unknown
IMP-00000: Import terminated unsuccessfully

So , I have couple of questions on this

1. Is the above import operation is supported?

2. If yes , which is actually causing the problem?

3. How to rectify this problem? - any solution or workarounds?

4. What should I check in both the Oracle instances - any specific parameters or configurations

View 2 Replies View Related

Server Utilities :: How To Use Escape Character For Loading Data Via Sql Loader

Mar 7, 2012

I have to load a fixed width file using sql loader utility. But the records have multiple special characters. writing / modifying the loader utility to load the data.

--Script to create the table
create table t1 (
ip1 varchar2(2),
ip2 number,
ip3 number);

--loader utility
LOAD DATA
INFILE 'c:inputfile.dat'
BADFILE 'c:adfile.bad'
REPLACE
INTO TABLE t1
FIELDS TERMINATED BY '' OPTIONALLY ENCLOSED BY '°'
(
ip1POSITION(1:2) CHAR,
ip2POSITION(3:17) INTEGER EXTERNAL ":ip2/100",
ip3POSITION(18:32) INTEGER EXTERNAL ":ip3/100",
)

--data file
9900000000000000000000059762160°
9900009694635473¶00009693856712-
99000024383898654000025664467904

--sql version i am using
SQL*Loader: Release 9.2.0.1.0 - Production on Wed Mar 7 18:32:33 2012
Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.

In the above mentioned data file, records has multiple special characters like '°','¶' ,'-'.
All these special characters have some meaning.
eg: '°' specifies the above column needs to be multiplied by -1
'¶' specifies the above column needs to be multiplied by -0.1

what changes need to be made in loader utility for the same? Also, will there be any change in the utility if I am using higher version of oracle?

View 13 Replies View Related

Server Utilities :: Escape Character For Loading Data Via Sql Loader

Dec 1, 2006

I have a text file which is comma separated with values enclosed in double quotes.

In my text file which I have to load into database, one of the field have the value like

Your "offspring"

When I run my normal sqlloader ctl file, it gives the error as

Record 304: Rejected - Error on table BUYER, column BUYERS_NAME.
no terminator found after TERMINATED and ENCLOSED field

Is there any way I can use some escape character for loading this type of data.

View 16 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved