Server Utilities :: Running Export Utility Online
May 26, 2010Just want to confirm if I can run the export (command line) even if the database is up (online)? Are there any risks/impact while doing so?
View 7 RepliesJust want to confirm if I can run the export (command line) even if the database is up (online)? Are there any risks/impact while doing so?
View 7 RepliesI have a From which take Logical Backup through oracle export utility. This Form work fine when i Start OCJ4 but when i want to take backup after running oracle Forms & Report services its not take backup.
My Question is dose Oracle Forms & Report Services (Standalone) Support data backup through Export (exp) utility. I have install Oracle Forms and Report Services (Standalone) on Window XP (SP-3).
We are migratingfrom [IBM P5 - AIX 5.3 - Oracle DB 10.2.0.1 64bit]to [IBM P7 - AIX 6.1 - Oracle DB 10.2.0.5 64bit]
The new database is already up and running, our next step now is the following:
1. Create Tablespaces (done)
2. Export the Database from P5
3. Import the dump to P7.
We have this following options:
I. Use a Windows XP workstation 32bit with Oracle 10.2.0.1 Database software to export from P5, then import the dump to P7. (Export/Import Utility 10.2.0.1 Windows 32bit)
II. Use the Export Utility of the P7 (10.2.0.5 AIX 64bit) to do the export and import.
III. Use the Export Utility of P5 (10.2.0.1 AIX 64bit) then use the Import Utility of P7 (10.2.0.5 AIX 64bit)
BTW, our colleague tried to do the following:
A. Use Windows - Export Utility 10.1.0.2.0 - 32bit to make a dump of our database (Oracle 10.2.0.1.0 AIX 64bit)
B. Use Windows - Import Utility 10.1.0.2.0 - 32bit to import the dump file from step A to our new database (Oracle 10.2.0.1.0 AIX 64bit)
But after issuing imp system/password@NEWDB file=(a.dmp, b.dmp... c.dmp) full=y ignore=y statistics=none
The import seems to hang here...importing SYSTEM's objects into SYSTEM
I have executed the exp utility to take schema backup as below.
DB_NAME=name
DAY=`date +%a_%d-%m-%Y`
gzip < exp_test_pipe > exp_test_${DAY}.dmp.gz &
exp username@test.cis.cat.com file=exp_test_pipe log=exp_test_dumps.log parfile=export_options.par
more export_options.par
compress=n
statistics=none
CONSISTENT=n
indexes=Y
owner=test
direct=Y
my password is starts with $
example: $12300$ahhh and hence its not allowing me to execute when given like below..
exp username/$12300$ahhh@test.cis.cat.com file=exp_test_pipe log=exp_test_dumps.log parfile=export_options.par
I met some problems with data pump tools. We have a large db(oracle 10.2.0.5, single instance on hpux v11.11, data about 8TB), while we wanna migrate to Linux ia 64bit, 10.2.0.5 RAC. The migrate window is around 20 hours, so the window can NOT be assign more hours.
As we consider if impdp & expdp work can start together, that will be save a lot of time. So I wonder if there are any ways to implement this or any other ways can speed up and make the data integrate better?
I exported three databases : servicedesk, report and mostcmdb on server A to a dumpfile. Moved them to similar dump location on server B. Now, I want to import them into B by using:
[SQL> impdp system/OSS_MOS7100 dumpfile=MOSTCMDB_0211.DMP schemas=mostcmdb
The error was:
SP2-0734: unknown command beginning "impdp syst..." - rest of line ignored.
I am trying to restore a package only from a dump file that was exported by original exp command. I know you can do it with data pump, but unfortunately this is dump file is not exported by expdp. Can I do that. I really do not want to import the whole database.
View 5 Replies View RelatedIs it possible to import a dump file using impdp data pump utility on oracle 10g where the export dump was taken using traditional exp utility and vice versa.
View 1 Replies View RelatedI ran "exp" command to take a back of Oracle Db based on user and later imported(using "imp" command) the dump into another db. Its seen that some the tables are not exported during exp command run. Can I use exp command on Oracle 11.2 version?
or should I always be using expdp command?
Why duplicated rows when a PK is present running Sqlldr?
View 1 Replies View RelatedI am using oracle 11g release2 and I am not able to find data pump utility in SQL developer.if I need to install it. I am new to this utility.
View 3 Replies View RelatedI am trying to export schemas from 10g to 11g. The NLS_CHARACTERSET for 10g is WE8ISO8859P1 and the NLS_CHARACTERSET for 11g is WE8MSWIN1252. Is it fine or do I need to change the character set, so that I will be able to successfully do the export/import?
View -1 Replies View Relatedi am having a problem when trying to export my DB,i could run an import fine,i have ran the catalog.sql,catproc.sql,catexp.sql and utlrp.sql again.Is it because the client and DB are different?How can i solve this problem
exp usr/pass file=exp_full.dmp log=exp_full.log full=y consistent=y
Export: Release 8.1.7.0.0 - Production on Wed Nov 28 13:40:04 2007
(c) Copyright 2000 Oracle Corporation. All rights reserved.
Connected to: Oracle8i Enterprise Edition Release 8.1.6.0.0 - Production
With the Partitioning option
[code]...
I would like to export specific tables(not entire schema) including metadata. I am using a parameter file for expdp.
Tables=emp,dept
Does this also include all metadata or should i also add the below Include in the parfile ?
INCLUDE =Indexes,Sequences,Procedures,Views
I was asked to do export/import of some schemas from 10g(linux) to 11g(AIX) using original expor/import method. I did not consider the character set and started doing export and import. while exporting, I get questionable statistics error in export log file. In the import log, I see the error like CREATE DATABASE LINK "xxxxxxxxxxxxx" CONNECT TO "xxxx" IDENTIFIED BY...
What can be done with these errors?
i need to export master data in excel sheets to our database and we use toad too. How i can export the data with the use of macros in excel. how i can export data from excel to oracle.
View 6 Replies View RelatedWhats is the usage of log file in Import/export .If i use following command ,it exports successfully
exp scott/tiger file=check.dmp log=empc.log tables=emp
and if i remove .log from here it will also export successfully So why do we use .log in import/export.
how can i monitor the export and import job and how increase the export and import job performance.
can i monitor the export and import job by checking the log and dump file created by export and import and can its performance increase by configure parallism. m i right or not?
I had specified the below:
Q1: Can we combine the 2 parameters together (owner and tables)? If not, then what is the way to specify it....
Exp scott/tiger owner=scott tables=(T1)
Error msg is: conflicting modes specified.
Q2: what is the privilege need for exporting other schema's tables?
Q3: what is the use of export table with index and many, but without ROWS?
send me the command for exporting multiple tables(1000+) in Linux env. 9i db, i know we can do using spool command but dont know exactly how to put it. i know using Datapump but this is 9i.
View 7 Replies View RelatedAs a part of our back up we used to export the production data every day using Original export Utility but from 11g original export Utility is de supported and also datapump doesn't support XML Objects so is there any other way to export the full database else any option to export xml Object using datapump.
View 2 Replies View RelatedI would like to run a daily job that will export the table data from SQL server table and Import back into Oracle table. I might need to run the query to update the flag back into sql server table once job is done. How can i do this using either sql server or oracle?
We have oracle 9.2 and sql server 2005.
Normally i do from flat file or csv file which is generated by developer or user from source destination (not me) and i dump into oracle using sql*loader but this time I have to directly extract/export the data from MS Sql server and load into Oracle table, mostly it will reload so i might doing any massaging data during the load.
Is it sql sql*loader has any function that i can use the datasource to connect the MS Sql server and fetch the data and insert back data into oracle? I have access to Sql server but i don't how to use sql server to do this or using oracle as a daily job even because have to schedule the job for this as it will be a daily job.
when i try to export schema using expdp i got error Like
Estimate in progress using BLOCKS method...
Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
Total estimation using BLOCKS method: 6.25 MB
Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
[code]...
HOw to solve this issue?
We are doing daily cold backup. Due to lack of disk space,we couldn't Hot backup. We want our database to be up when doing backups. Since only export/import is possible in our scenario, clarify few queries:
1) Export was done during off period from the live server.
2) We have a development server, in which we have to update our database daily. Can i overwrite the Development server using IMPORT daily? Since this import might show lots of errors (Object already exist), what parameters can i use for import.
i am getting "snapshot too old " error while take in export backup of a database(with CONSISTENT=y), it actually runs for 3 hours.
it always fails for table1 with snapshot error
i pulled the awr for that 3 hours, to see any long running SQL hitting table1 . and i found 3 , Two SELECT and one INSERT.
I assume it is INSERT not letting me take a consistent export backup of Table1 .
When i do a table export , it got over in 30 mins.When i do import using same dump file (that was created in 30 mins), its taking more than 30 mins .
why the import is taking more time than the export time ?
Is it possible to identify what level of export by looking at export dumpfile .. whether it is a schema export,full export,table export,..
If yes.. how ?
any other utilities that we can use to load data from our PROD server (10g) to DEV server (9i)? I've read some related topics here that it's not possible to import from a HIGHER to LOWER versions of Oracle. We've tried (many times) EXPorting selected tables from the 10g dB, then IMPort it to the 9i dB and we've haven't succeeded anyhow. PROD & DEV have a different schema/owner but the same table structures.
View 4 Replies View RelatedWe have two databases running on 10.2.0.4 and 9.2.0.8. Both are having the same unpartitioned table of size 80G. I am exporting the table on 10g by using parallel=8 and dumpfile with %U option. That took around 4 hours to export the table.
And on 9.2.0.8, i am exporting using below parameters, taking around 5 hours.
buffer=2000000
recordlength=64000
options i can try to speed up the export in both versions.
I want to create two or three sachems on my production server which should be the same copy of my another second production server. And I access this second server through VPN connection on toad9.0.1. And I access my production server through VNC viewer and database through toad.
How cloud I create schema on my first prod. server from second server.