Server Utilities :: How To Find Estimated Time To Complete Import (imp)
Apr 17, 2012
I am using imp / Schema level. Dump file contains 3 schemas with 120 GB size.Imp is running now. How to find completion time. So that I can inform to application developer about activity completion time.
View 4 Replies
ADVERTISEMENT
Feb 13, 2012
When i do a table export , it got over in 30 mins.When i do import using same dump file (that was created in 30 mins), its taking more than 30 mins .
why the import is taking more time than the export time ?
View 14 Replies
View Related
Sep 13, 2012
I am try to import 4G dump in Oracle 11R2 version, in that we have around 9000+ Package Body which is taking huge time than other objects (about 8 to 12 hrs) and also it is expecting lots of system space (roughly about 10GB).
I have tried both parallel and non-parallel.how to improve speed of the package body import.
Details about the Schema & Import No. of objects in Schema
SQL> select object_type,count(1) from user_objects GROUP BY ROLLUP( object_type);
OBJECT_TYPE COUNT(1)
------------------- ----------
FUNCTION 248
INDEX 5161
JAVA CLASS 471
JAVA RESOURCE 1
JAVA SOURCE 16
LIBRARY 1
ORA-00933: SQL command not properly ended
View 3 Replies
View Related
Jan 16, 2012
I use the following query to find out the remaining time to complete the table statistics which is running currently.
SELECT SID, SERIAL#, opname, SOFAR, TOTALWORK,username,context,
ROUND(SOFAR/TOTALWORK*100,2) COMPLETE
FROM V$SESSION_LONGOPS
WHERE
TOTALWORK != 0
AND SOFAR != TOTALWORK
order by 1;
SOFAR column shows 9325 and totalwork column shows 12287.How to calculate the columns in terms of hours and minutes ?
And also sometime the query never shows the output of the query for the current running query.
View 1 Replies
View Related
Jul 7, 2010
We will create a new instance in our production server, but this time, part of it's table structure has a BLOB data type (re: <column name> blob(3000)). It's our first time to handle this kind of Oracle data type. What would be my estimate size for it's default tablespace?
View 5 Replies
View Related
Jun 19, 2010
I have oracle 9i dump...now i want to import this dump into oracle 10g.....is it possible.
View 4 Replies
View Related
Dec 28, 2010
i am working on oracle 11g r2. is there any way to go back to oracle 9i with use of import/export utility. Should i take downgraded export from oracle 11g for oracle 9i.
View 2 Replies
View Related
Sep 6, 2010
While doing import got the below errors.
Failing sql is:
BEGIN DBMS_JOB.ISUBMIT( JOB=> 361, NEXT_DATE=> TO_DATE('2010-09-06 21:18:27', 'YYYY-MM-DD:HH24:MI:SS'), INTERVAL=> 'SYSDATE
+ 45/86400', WHAT=> 'PK_MONITORS.SP_OVERDUE_JOB;', NO_PARSE=> TRUE); END;
ORA-39083: Object type JOB failed to create with error:
[code]...
View 2 Replies
View Related
May 22, 2011
I do not import a full database.
View 3 Replies
View Related
Nov 29, 2006
We have a table in Oracle9i database with around 14 million records and we would like to import that table into 10g database with similar structure. We have exported the table from 9i database and would like to import the table into 10g database within same schema name with different table name as we already have the table with same name in 10g database in same schema. Is it possible to import a table with different table name?
We have a way around to import the table into 10g database in another schema and then push the data into our main table but want to know whether the above requirement is possible.
View 7 Replies
View Related
Jul 2, 2011
How to import a dump file into different tablespace.
normally when i am importing a dump file it goes to the table space where it was exported but i need to import in to different tablespace
View 9 Replies
View Related
Dec 31, 2011
I have a backup of Oracle Database 10g (business_bk.dmp). Now I would like to import it into Oracle DB 9i.
View 9 Replies
View Related
Dec 14, 2010
ORA 12899 prompts when we are doing import from 10.2.0.3 -> 11g R2. Also it is cross platform, AIX -> windows 2003.
View 7 Replies
View Related
Oct 4, 2011
I have Live Server with 11g R2 with Dataguard on server 2008 server 64 bit. I exporting the DMP from live and importing in 10g on Server 2003 32 bit. All tables are not imported.
View 3 Replies
View Related
Nov 6, 2013
I have taken an export using expdp of schema, data of the schema spread across different tablespaces , now i want to import the data to only one tablespace.
View 1 Replies
View Related
Jan 26, 2011
I am getting the below error when I import a table from Prod to Dev. I understand this error will be occured if length of the datatype is low. First I got the error when the datatype(length) which is 25 for the column PASSWORD column.Then I increased the length of this column to 45, then it was imported successfully.
why am facing the error when the datatye and length for this table is same in prod and dev? What are the possible ways to import the data without increasing the PASSWORD column length?
IMP-00019: row rejected due to ORACLE error 12899
IMP-00003: ORACLE error 12899 encountered
ORA-12899: value too large for column "ANEES"."SALSA_WEB_ACCESS"."PASSWORD" (actual: 28, maximum: 25)
[code]....
View 7 Replies
View Related
Apr 26, 2011
I was asked to do export/import of some schemas from 10g(linux) to 11g(AIX) using original expor/import method. I did not consider the character set and started doing export and import. while exporting, I get questionable statistics error in export log file. In the import log, I see the error like CREATE DATABASE LINK "xxxxxxxxxxxxx" CONNECT TO "xxxx" IDENTIFIED BY...
What can be done with these errors?
View 4 Replies
View Related
May 7, 2013
I'm unable to do an import of a *.dmp file.
[oracle@oracledbserver ASG1]$ cd /media/volume-01/u01/app/oracle/product/
[oracle@oracledbserver product]$ ls
11.2.0 20-04-2013full_backup.dmp full01-03-2013_backup.dmp new.dmp today.dmp
[oracle@oracledbserver product]$
[oracle@oracledbserver product]$
[oracle@oracledbserver product]$
[oracle@oracledbserver product]$ impdp full=Y directory=agge_dir dumpfile=/media/volume-01/u01/app/oracle/product/new.dmp NOLOGFILE=y;
Import: Release 11.2.0.1.0 - Production on Tue May 7 16:51:47 2013
Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
Username: sys as sysdba
Password:
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
ORA-39001: invalid argument value
ORA-39000: bad dump file specification
ORA-39088: file name cannot contain a path specification
[oracle@oracledbserver product]$
This is Oracle 11g hosted on an eucalyptus cloud instance.
View 10 Replies
View Related
Apr 30, 2012
I am trying to import .dmp file to my database
I am trying to run the imp command from dos prompt.
Here is the error I got .
C:Documents and SettingssairammMy Documentsartmsexp_AUDT2_04292012>imp
sairamm/mypassword@aws fromuser=AUDT2 toUSER=SAIRAMM file=exp_AUDT2_04292012.dmp
log=imp_AUDT2_04292012.log BUFFE
R=10000000 GRANTS=y
Import: Release 9.2.0.1.0 - Production on Mon Apr 30 14:23:29 2012
Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production With the Partitioning, Oracle Label Security, OLAP, Data Mining,Oracle Database Vault and Real Application Testing option
IMP-00010: not a valid export file, header failed verification
IMP-00000: Import terminated unsuccessfully
View 2 Replies
View Related
Jul 26, 2013
The way to do IMPORT of a schema's ALL the TABLES only. Through Data Pump. I don't want to import any other objects like 'package',procedure etc...
Is it the only way is specifying them within EXCLUDE parameter?
View 3 Replies
View Related
May 11, 2010
Whats is the usage of log file in Import/export .If i use following command ,it exports successfully
exp scott/tiger file=check.dmp log=empc.log tables=emp
and if i remove .log from here it will also export successfully So why do we use .log in import/export.
View 4 Replies
View Related
May 12, 2010
How to skip one table while import in traditional exp/imp not in DP.
View 4 Replies
View Related
Feb 7, 2011
I have exported a schema dump with schema name as 'A'.I want to import that dump in to schema 'B'.how ?
View 5 Replies
View Related
Feb 23, 2012
I have the problem with import in Oracle 8.1.7.The size of import file is 29600 kb and tablespace size is 16gb and when I try to make import oracle back this message:
IMP-00003: ORACLE error 1659 encountered
ORA-01659: unable to allocate MINEXTENTS beyond 7 in tablespace DATA
The data tablespace is full. I think that the import file contains information about the original tablespace from which has made export. But I don't now how to resolve the problem
View 10 Replies
View Related
Feb 11, 2013
I am trying to use NETWORK_LINK option in datapump and import a table from one server to another. I gave the below command :
C:>impdp example/example@db DIRECTORY=DATA_PUMP_DIR
NETWORK_LINK=db.legal.regn.net remap_schema=BI:example
tables=BI.BI_DIRECT dumpfile=BI.dmp logfile=BI.log
Got the following errors :
ORA-39002: invalid operation
ORA-39070: Unable to open the log file.
ORA-29283: invalid file operation
ORA-06512: at "SYS.UTL_FILE", line 536
ORA-29283: invalid file operation
Is this error related to the permission in the OS level (windows 7 in my case)? I manually created the folder 'DATA_PUMP_DIR' in the specified directory path. Though the directory I created (DATA_PUMP_DIR) shows read-only in the general tab of the property, I am able to create files under the folder 'DATA_PUMP_DIR'.
View 25 Replies
View Related
Apr 28, 2012
how can i monitor the export and import job and how increase the export and import job performance.
can i monitor the export and import job by checking the log and dump file created by export and import and can its performance increase by configure parallism. m i right or not?
View 2 Replies
View Related
Jan 18, 2010
I wanted to export a table "emp_production" from Production database then import it as "emp_datawarehouse" in Data warehouse database.Both tables has same structure. I have granted IMPORT FULL DATABASE & EXPORT FULL DATABASE privileges to both schema
I tired with the following syntax
$ Expdp u1/p1@h1[/email] tables= emp_production directory=test dumpfile=test1.dmp
$ Impdp u1/p2@h2[/email] directory=test dumpfile=test1.dmp remap_schema=u1.emp_production:u2.emp_datawarehouse
remap_tablespace=Example1:Example2
But I am getting the following error
ORA-31631: privileges are required
ORA-39122: Unprivileged users may not perform REMAP_SCHEMA remapping.
Why this ? "emp_production" table has 150 million rows, every week importing this table then inserting into "emp_datawarehouse" table takes long time.
View 8 Replies
View Related
Jun 24, 2008
I would like to run a daily job that will export the table data from SQL server table and Import back into Oracle table. I might need to run the query to update the flag back into sql server table once job is done. How can i do this using either sql server or oracle?
We have oracle 9.2 and sql server 2005.
Normally i do from flat file or csv file which is generated by developer or user from source destination (not me) and i dump into oracle using sql*loader but this time I have to directly extract/export the data from MS Sql server and load into Oracle table, mostly it will reload so i might doing any massaging data during the load.
Is it sql sql*loader has any function that i can use the datasource to connect the MS Sql server and fetch the data and insert back data into oracle? I have access to Sql server but i don't how to use sql server to do this or using oracle as a daily job even because have to schedule the job for this as it will be a daily job.
View 4 Replies
View Related
Feb 24, 2012
I am facing a problem importing DMP file in 11g. While importing it gives me error not responding. I have to attached the jpg file for that to clear you my point whats wrong is going during import. My Dump is on 9i i want to import that on 11G R2.
View 4 Replies
View Related
Apr 6, 2010
I have a text file of 175G with the delimeter as '|*'. I want to import it using sql *loader. I know that sql *loader can load data parallely. I can employ two schemes for importing it using sql *loader
a) with the text file being imported all at once using option parallel = true
b) with the single text file being broken down into files of smaller size(say 5G each or more) and then imported in parallel using parallel = true.
I can import it in somewhat powerful server with 128G of ram, 16 intel CPus and abundant hard disk space.Now my question is which option would be better in performance (in terms of time) it takes to import data?I donot have sufficient time to test both of these approaches.
View 5 Replies
View Related