Export/Import/SQL Loader :: Impdp - Should Network_link And Remap_data Work Together

Oct 7, 2013

Should I be able to use network_link and remap_data together?I am trying to do three things at once:- copy data from one db to an other (prod to test)- change schema- mask sensitive dataI would like to avoid using dump files, import from network link would be nice. My initial testing is suggesting the answer is no.

When I use impdp with the following parameter file, sourcetable.column_to_remap is transferred without modifications, no data remapping is

done.userid=importinguser/XX@orcltables=sourceuser.sourcetabletable_exists_action=REPLACEremap_data=sourceuser.sourcetable.column_to_remap:remappgk.remapfnnetwork_link=source_linkremap_schema=sourceuser:importinguserdirectory=dumpdirlogfile=log_of_imp If i dump the data into file first and then import from file, data remapping is done as expected so I believe my remappkg is correct.

t there is something wrong with my parameter file and network_link&remap_data work together! 

Edit: Testing is done with 11.2.0.1.0 at Windows Server 2008 R2 
Edit2: More testing: 11.1.0.7@linux - Success!,
Edit3: When writing "Edit2" I thought it worked at 11.2.0.1@linux but I was wrong (source data was already remapped). ==> So the answer is yes, they work together at version 11.1.0.7 and they don't seem to work at 11.2.0.1.

View 0 Replies


ADVERTISEMENT

Export/Import/SQL Loader :: Impdp With System But Doesn't Work With Another User?

Oct 10, 2012

We have troubles with import of transportable tablespace, when i try to import it with system the import is done currectly, when i try to use another user we receive this message :

impdp bmcese/***** directory=TTS_DIR dumpfile=RMCCO_RMC_ANA_STS_ABB_CO121001.dmp TRANSPORT_DATAFILES=/data/TTS/RMCCO_RMC_ANA_STS_ABB_CO121001.dbf logfile=tts_imp_proc.log

ORA-31626: il job non esiste
ORA-31633: impossibile creare la tabella principale "BMCESE.SYS_IMPORT_TRANSPORTABLE_05"
ORA-06512: a "SYS.DBMS_SYS_ERROR", line 95
ORA-06512: a "SYS.KUPV$FT", line 1020
ORA-00959: tablespace 'RMCCO_RMC_UTZ_PP080531' inesistente

this happens only on development server and it work into test server, so i don't think that the problem is in grants(i controlled that are the same and i have tried with DBA grant too and we received the error).Another strange thing is that tablespace 'RMCCO_RMC_UTZ_PP080531' is not included into the dump that i try to import and it doesn't exist in the database.

View 4 Replies View Related

Export/Import/SQL Loader :: Should Db_link Be Recreated After Impdp

Sep 17, 2013

I have imported schemas (impdp) from production database (10gR2, RHEL 64bit). One of the schemas has a db_link. The db_link points to a database that exists on the same server - both in production server and also on the new server where I imported the schemas. When I run a simple query in production using this db_link, it works but when I run the same query on the test server (where I imported the schemas), it gives me following error: ORA-02019: connection description for remote database not found I run this in prod database: Select count(1) from SOME_TABLE@my_link; when I run it in new database, it gives the above ORA error - even if I qualify the table and db_link with the schema owner like this: Select count(1) from the_owner.some_table@the_owner.my_link; NOTE: I am not running these queries as schema owner - I do not know the password. I am able to connect to both databases like this from the command prompt: $ sqlplus user@/password@db1$ sqlplus user@/password@db2 Does this mean that I need to recreate the db_link - perhaps every time I import?

View 6 Replies View Related

Export/Import/SQL Loader :: Impdp Custom Log Path

Oct 18, 2012

I am using the following command for importing data in to oracle database 11g in linux box.

impdp system/test-123 directory=test_dir dumpfile=test.dmp logfile=test.log full=yes remap_schema=test1:test

Here test_dir = /u01/app/oracle/product/11.2.0/db_1/DATABASE/test/dmp

By default the test.log file is getting created in the above said dir (test_dir) but my requirement is I want to have it in a separate dir.

View 13 Replies View Related

Export/Import/SQL Loader :: Use Impdp To Restore To Different Database Name

Jun 20, 2012

Can I use impdp to restore to a different database name? If yes what is the syntax? This is 10.2 on linux.

View 2 Replies View Related

Export/Import/SQL Loader :: IMPDP Mapping To A Single Tablespace

Oct 30, 2012

I have a dumpfile from a database with hundreds of tablespaces. Do I need to remap all of them on impdp or is there a way to point all tables to a default tablespace? I mean, the source database has 200 tablespaces. The target database just 1.

View 2 Replies View Related

Export/Import/SQL Loader :: IMPDP Into Schema With Truncated Tables

Aug 2, 2012

I've got a schema that I've truncated all tables. I have a full schema export I took awhile back, and I'm wanting to import this into the schema to basically 'reset' it.

First time run, I got the :

ORA-39151: Table "xyz.tablename" exists. All dependent metadata and data will be skipped due to table_exists_action of skip

I've been reading through, and see suggestions to add to the par file:

CONTENT=DATA_ONLY TABLE_EXISTS_ACTION=APPEND

And I've seen others use the option for:

table_exists_action=replace

I basically want to put the data back into the tables, and have the indexes rebuilt.....

View 1 Replies View Related

Export/Import/SQL Loader :: IMPDP Is Slow And Receives Error ORA-39126

May 30, 2013

I've been using datapump for a long time now but I have not come across this problem before.

Importing just two tables: Table1 data=100Mb=11 million rows
Table2 data=4.2Gb =19.6 million rows

Table1 ran for approx. 5 hours
Table2 ran for approx. 15 hours

If I run the impdp importing both tables in the same par file the default tablespace of the users the import is running as runs out of space due to ORA-01691: unable to extend lob segment <owner>.SYS_LOG0001175799C00045$$ by 512 in tablespace USERS. I do not understand why it is creating objects in order to import tables into someone elses schema.

The environment is Red Hat LINUX 4.1.2-51 running Oracle 11.2.0.1 of Oracle11gR2. This is a 9 node RAC using ASM.

View 4 Replies View Related

Export/Import/SQL Loader :: Changing Object Names During Expdp Or Impdp?

Mar 11, 2013

I'm curious to know if expdp or impdp is able to change object names during the process. What I mean by this is... can I export out procedures:

procedure1
procedure2
procedure3

Then import them like this:

test_procedure1
test_procedure2
test_procedure3

I'm not sure the expdp or impdp has that ability, but I could have missed it. I know how to remap a schema, but that only changes the schema name.

View 3 Replies View Related

Export/Import/SQL Loader :: Can Impdp Perform Checks To See If Dumps Are Valid

Jul 20, 2012

i'm writing a job to export data on a weekly basis and archive those data in-case needed to be re-imported in future, its important that we able to import successfully if ever needed.

can impdp perform checks to see if dumps are valid ?how about old imp?if not, will it be safe enough to rely on the export log file? in another word is it safe to assume its all safe if there is no error or warning in the exp log?

View 7 Replies View Related

Export/Import/SQL Loader :: Script To Execute Impdp Using SILO Concept

Apr 4, 2013

write a shell script to perform impdp using dbms_datapump using SILO concept .

View 9 Replies View Related

Export/Import/SQL Loader :: Impdp Network Mode Can Use Bequeath Connection?

Feb 12, 2013

I'm going to import a single database using FS to a RAC on ASM both are in same server running oracle 11.2.0.3. So I was wondering if can I use network link mode of impdp without setting up a listener using 10gb Ethernet interface ?

I don’t want to incur the overhead of the tcp network layer because both DB single and Rac are on the same server.

View 4 Replies View Related

Export/Import/SQL Loader :: Impdp Index-organized Table And Temp Usage?

Aug 1, 2012

Env: RHEL 5.8 RAC 11.2.0.2

I'm currently moving an IOT from one database to another using expdp/imdp. The IOT is non-partitioned and about 100GB in size containing ~1,1 billion rows.

The dumpfile contains nothing else but the IOT. I'm importing with no special parameters, no pre-created IOT, just ordinary dumpfile import. (impdp username/password dumpfile=impdp:iot.dmp nologfile=y )

During import I got unable to extend TEMP errors from impdp.

ORA-39171: Job is experiencing a resumable wait.
ORA-01652: unable to extend temp segment by 128 in tablespace TEMP
ORA-39171: Job is experiencing a resumable wait.
ORA-01652: unable to extend temp segment by 128 in tablespace TEMP

I had to add 2 additional files to my Temp tablespace (total 96GB of temp) before the import could finish off.

Is this temp usage to be expected when importing IOT's ?

View 4 Replies View Related

Export/Import/SQL Loader :: How To Skip Database JOBS During Export And Import

Aug 7, 2012

1) Is there a way to skip database jobs while exporting (EXPDP) ?

2) Is there a way to skip database jobs while importing (IMPDP) ?

View 3 Replies View Related

Export/Import/SQL Loader :: Why Do Export-import Require Temporary Tablespace

Aug 9, 2012

Why do export-import require temporary tablespace? Since export-import do behave like DMLs, when does temporary tablespace be needed by datapump utility?

View 2 Replies View Related

How Streams Export / Import Parameters Work

Nov 30, 2011

I have a test database with tables that occasionally has to be updated to the state of the production database tables. These databases both stream several tables to separate databases. (ie database A streams tables to database X and B to Y).

I export the needed schema daily from the production database "A". To update the test database "B", I import the needed tables from this export file. I then export the tables to be streamed from database B with the object_consistent=y parameter. I then import those tables to database Y using the streams_instantiation=y parameter.

So this is the process ...

1. Turn off capture on B
2. Turn off apply on Y
3. Drop the tables in B
4. Import tables from A to B
5. Prepare instantiation on tables in B (DBMS_CAPTURE_ADM.PREPARE_TABLE_INSTANTIATION)
6. Export from B with object_consistent=y
7. Drop tables in Y
8. Import from B to Y with streams_instantiation=y
9. Start apply on Y
10. Start capture on B

I ran this overnight through an OEM job and it appeared to work, just not how I expected. When I came in there was a lot of activity on both databases. All sessions were from the streams admin account. This continued for some time (long after the job had finished). During this time the "captured scn", "applied scn" & "last enqueued scn" values were quite different on the source side (didn't jot those down). I compared these ot the "last applied message number" on the target side and the valuse were quite a bit off.

After the activity subsided, the SCN values were all in the same ballpark. So I figure all the activity was changes getting applied from database B to database Y. I tested out the streaming and it's working fine.

I thought the parameters in the export and import statements were supposed to sync up the SCNs. That is I thought they would start off in the same state and there would be no need for applying changes. No one changed anything on the source database overnight.

I must not be understanding how the streams export/import parameters work.

View 5 Replies View Related

SQL & PL/SQL :: Export A Schema Using Expdp And Import Using Impdp

Apr 20, 2011

I have a process to export a schema using expdp and import using impdp. Everything creates successfully except for a trigger. The trigger gives and error that the table or view does not exist. The account that I use to import the schema is different than the schema user but is a highly privileged account. I notice that the schema in the create or replace trigger line of code is remapped (I am using remapping in the impdp syntax) and the rest of the syntax of the trigger (which is just a sequence trigger for a primary key column) does not have the schema. In order to fix the issue, I have my bash script log into oracle as the schema user after the import of the schema and execute the trigger code. why do I have to do this for trigger code but not for other objects like views that create just fine.

View 2 Replies View Related

Export/Import/SQL Loader :: How To Map One Field To Another In Control File Via SQL Loader

Mar 18, 2013

I have a flat file (student.dat delimiter %~| ) using control file (student.ctl) through sql loader. Here are the details.

student.dat

student_id, student_firstname, gender, student_lastName, student_newId
101%~|abc%~|F %~|xyz%~|110%~|

Corresponding table
Student (
Student_ID,
Student_FN,
Gender,
Student_LN
)

How do i map student_newId field to student_id field in STUDENT DB table so that new id should be inserted in student_id column. How do i specify the mapping in control file. I dont want to create a new column in student table. In control file i will specify the below, Is this a best approach?. Do we have any othe way?

STUDENT_ID *(:STUDENT_NEWID)*,
STUDENT_FN,
GENDER,
STUDENT_LNAME,
STUDENT_NEWID BOUNDFILLER

View 1 Replies View Related

Export/Import/SQL Loader :: Table Import Takes Long Time And Still Running?

Jun 22, 2012

MY DB Version: 10.2.o

OS: Windows Server 2003

I am trying to import on table which i have the export dump file which i take using expdp previously when i load that table on the same host

by using below command:

expdp scott/tiger@db10g tables=EMP,DEPT directory=TEST_DIR dumpfile=EMP_DEPT.dmp logfile=expdpEMP_DEPT.log

after that i zip that dump and move it to external usb and now i need that table i copy that table and unzip that that dump

Command i am using to do the import is :

impdp scott/tiger@db10g tables=EMP,DEPT directory=TEST_DIR dumpfile=EMP_DEPT.dmp logfile=impdpEMP_DEPT.log

But the query of import is still runing even not showing any amount of rows to be imported.

i already make the tablespace in which the table was previosuly before dropping but when i check the sapce of tablespace that is also not consuming one error i got preiviously while performing this task is:

Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Production
With the Partitioning, OLAP and Data Mining options
Master table "CDR"."SYS_IMPORT_TABLE_03" successfully loaded/unloaded
Starting "CDR"."SYS_IMPORT_TABLE_03":  cdr/********@tsiindia directory=TEST_DIR dumpfile=CAT_IN_DATA_042012.DMP tables=CAT_IN_DATA_042012 logfile=impdpCAT_IN_DATA_042012.log

[code]....

i check streams_pool_size it will show zero and then i make it to 48M and after that

SQL> show parameter streams_pool_size;
NAME                                 TYPE        VALUE
-----------
streams_pool_size                    big integer 48M

But still it takes time

View 13 Replies View Related

Export/Import/SQL Loader :: Import Scott To Scott1 Without Affecting Existing Records

Sep 29, 2012

My database version
SQL> select * from v$version;

BANNER
----------------------------------------------------------------
Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Prod
PL/SQL Release 10.2.0.1.0 - Production
CORE    10.2.0.1.0      Production
TNS for 32-bit Windows: Version 10.2.0.1.0 - Production
NLSRTL Version 10.2.0.1.0 - ProductionOS version:

Windows 7 64bit I have schema(scott) export with schema level option and imported with different name as (scott1).At regular period of time i need to import the scott to scott1 without affecting existing records.such as
*1. Need to append new created records.*
*2. Need to append updated records.*

for the above requirement I did in the following way
expdp xxxx/******** schemas=SCOTT directory=dumpdir dumpfile=SCOTT_28-SEP-2012.dmp logfile=exp_SCOTT_28-SEP-2012.log imported in the following way impdp xxxx/******** AS SYSDBA REMAP_SCHEMA=SCOTT:SCOTT1 directory=DUMPDIR dumpfile=SCOTT_28-SEP-2012.dmp logfile=imp_SCOTT2_28-09-2012.log TRANSFORM=SEGMENT_ATTRIBUTES:n TABLE_EXISTS_ACTION=APPEND.

The problem is i couldn'table to append the records to existing tables the log error show such ways.

ORA-31684: Object type USER:"SCOTT1" already exists
Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
Processing object type SCHEMA_EXPORT/ROLE_GRANT
Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
[code].....

View 5 Replies View Related

Export/Import/SQL Loader :: Error / Unable To Create Metadata And Import Fails

May 10, 2013

I am trying to run impdp over network to import tables only but i am getting an error saying not able to create metadata and the import fails

Here are the steps below,

1. Source database created a user and granted select on certain tables to the user.

2. Created a user in the Target database.

3. Created a public link as sys user in the target database.

4. granted imp and exp full database to both users and all the other privs.

5. Started the impdp from the target server.

The import fails with

$impdp abc/xyz directory=DATA_PUMP_DIR network_link=TESTAR logfile=net_import_proddev.log TABLES=impdb.abc parallel=12 REMAP_SCHEMA=IMPDB:ABC

Import: Release 11.2.0.3.0 - Production on Tue Apr 23 13:10:51 2013
Copyright (c) 1982, 2011, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
Starting "IMPDB"."SYS_IMPORT_TABLE_01": abc/******** directory=DATA_PUMP_DIR network_link=TESTAR logfile=net_import_proddev.log TABLES=impdb.abc parallel=12 REMAP_SCHEMA=IMPDB:ABC

[code]....

View 14 Replies View Related

Export/Import/SQL Loader :: How To Import Indexes Separately / Skipping Tables And Other Objects

Jul 9, 2013

aix 6.111.2.0.3 I have an expdp dump from prod  to be imported to our test database.I have imported it using impdp, but to my surprise the tables were imported but  lots of indexes were not created? even If I have used TRANSFORM=SEGMENT_ATTRIBUTES:N just to use the default USERS tablespace. How do I import the indexes separately, skipping the tables and other objects?

View 14 Replies View Related

Export/Import/SQL Loader :: How To Import Data From Excel File To Table Through Procedure

Jul 2, 2012

How to import data from excel(.xls) file to data base table

I have excel sheet(.xls) data details, I neet to upload details to data base table using procedure

excel sheet is not CSV file, so SQL Loader is not using

any alternative solution for this issue

View 3 Replies View Related

Export/Import/SQL Loader :: Expdp Command To Export Table By Specifying Query Parameter

Aug 16, 2013

I am using expdp command to export the table by specifying Query parameter. But i am unable to export the table based on the condition. 

Ex:EXPDP username/password dumpfile=employee.dmp logfile=emp.log directory=DATADIR_EXP TABLES=EMPLOYEE query=EMPLOYEE:"UPDATED_TIME >= '04-JUN-13' AND UPDATED_TIME >= '05-AUG-13'" Estimate in progress using BLOCKS method...Processing object type TABLE_ EXPORT/ TABLE/ TABLE_ DATATotal estimation using BLOCKS method: 3 GBORA-31693: Table data object "<username>"."EMPLOYEE" failed to load/unload and is being skipped due to error:ORA-00933: SQL command not properly endedMaster table "<username>"."EMPLOYEE" successfully loaded/unloaded...Dump file set for <username>.SYS_EXPORT_TABLE_01 is:  E:IMPDPemployee.dmpJob "<username>"."SYS_EXPORT_TABLE_01" completed with 1 error(s) at 12:34:45  Oracle 11g,

View 6 Replies View Related

Export/Import/SQL Loader :: Selective Export Failing - Solaris Sparc - Oracle 10g

May 11, 2013

I am trying to export selective data from one of my prod database tables. But not succeeding. I was keep on trying for the past 2 hours.

OS : SOLARIS SPARC
ORACLE - 10G
Query --> WHERE E3RECV_DT LIKE '201305%' (I need to export this query data)

Below Script i am using
===============
exp E3USER@SGEBAPU2 statistics=none consistent=n buffer=100000000 file=exp_pipe_file TABLES=IFDATA query="WHERE E3RECV_DT LIKE '201305\%'" log=PGTB_IFDATA_conditional.log

View 1 Replies View Related

Export/Import/SQL Loader :: Oracle Traditional Import Overrides Password?

Nov 23, 2012

I had just successfully finished a full importing from Oracle 9i DB to Oracle 11gR2 DB. My export was a full db export.

Prior to this importing, my 11g was a newly created DB with the default SYS, System etc.. schema. Their passwords is different from those in 9i.

However, i realised that after importing... their passwords in 11g was replaced by those passwords in 9i, including SYS and SYSTEM user...

View 5 Replies View Related

Export/Import/SQL Loader :: How To Filter Some Illegal Rows When Import Data

May 24, 2013

I want to import data in a csv file by SQL Loader.

but , I don't want to import some illegal rows when the column 'name' is null

how can I modify the SQL Loader ctrl file?

View 1 Replies View Related

Export/Import/SQL Loader :: Error During Data Pump Import With Developer

Sep 17, 2012

I try to transfer data from one database to another one through data pump via SQL Developer (data amount is quite important) exporting several tables. Tables export is doing fine, but I encounter the following error when I import the file (I try data only and data + DDL).

"Exception: ORA-39001: argument value invalid dbms_datapump.get_status(64...=
ORA-39001: argument value invalid
ORA-39000: ....
ORA-31619: ...

The file is in the right place, data pump folder of the new database. User is the same on both base, database version are similar.

View 4 Replies View Related

Export/Import/SQL Loader :: Schema Level Export For User Sample 1

Jul 11, 2012

I have schema level export for user SAMPLE1(Default tablespace USERS) on oracle 9.2.0.1 production database. I want to import into another 9i database on another server, so do i nneed to Create SAMPLE1 user and USERS tablespace in new database again.

View 5 Replies View Related

Export/Import/SQL Loader :: How To Export Full Dump And Metadata Of Particular Table

Apr 9, 2013

let us consider mytest schema is having 6 tables

tname tabtype
myt table
myaxpertlog table
abb table
ccc table
ddd table
xxx table

now from this schema i want full dump and also from myaxpertlog table i required metadata only not records.

c:> export mytest/log file=20130409mytest0904pm.dmp tables=(myaxpertlog) rows=n

if i tried i am get only one table but it does have records.

View 6 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved