Ideally we need to include description in tnsnames.ora but its taking time to contact DBA here. Hence we tried with this work around. The same is working fine if we are using sqlplus but sqlloader gives this error.
LRM-00116: syntax error at 'ADDRESS' following '('
SQL*Loader: Release 11.2.0.2.0 - Production on Thu Jan 10 21:31:32 2013
i want to perform full export + import of an oracle 11g database as fast as possible. i was thinking to perform the exp+imp on the same command.in exp i can perform something like this :
i know that i can do both action in impdp when using a dblink, but the problem is that some objects in the database cannot be copied via a dblink. the question is if there's a corresponding datapump command to the old exp+imp command i presented.
When I do the import the of succeeding dump, I drop the existing schema "SQL> drop user username cascade;" and import dump by " impdp system .... ". I would like to import a dump to an existing instance but only data import and will leave the current packages and other metadata untouched and unchanged on the said existing instance.
1. Do i need to drop user before the import if my requirements are the above?
2. If i need to drop user, what should be script.
3. For the import itself, what parameter should i use?
4. What are the necessaries I need to consider before doing the import.
While making a full db export i have got this error even though my export was completed with this warning. What should i need to do regarding to this error. My oracle version is 10.1.0.2 and Server is windows 2003.
I am trying to export a full DATABASE using the command...EXP [username/password]@[CS]FILE=PATH\[filename.dmp] LOG=PATH\[logname.log] INDEXES=n STATISTICS=none COMPRESS=Y
the database begins to export as shown below, but the export terminates with the below error.
About to export specified users ... . exporting pre-schema procedural objects and actions . exporting foreign function library names for user [username] . exporting PUBLIC type synonyms . exporting private type synonyms [code],,,,
considering that i have exported full databases successfully before using the mentioned command above.
One of my Friend gets error While datapump Export backup of Full database.
Pfa below error details:-
ORA-31693: Table data object "RADIOMIRCHI_PIP_HRMS"."GM_DEPT" failed to load/unload and is being skipped due to error: ORA-02354: error in exporting/importing data ORA-00604: error occurred at recursive SQL level 3 ORA-21780: Maximum number of object durations exceeded. [code]......
I have a long running transaction (more than 3 hours), and at the same time other operations are occurring on different tables, using the same UNDO space.
Sometimes we see ORA-30036, but this error occurs very late in the process. The transaction normally takes 3 hours, but when UNDO space is full, we do not get ORA-30036 upto 8 hours or 9 hours of process.
I am wondering what could be happening in the background, when UNDO space is full, which makes the transaction to extend upto 8 hours or 9 hours (pl. note, this transaction gets completed within 3 hours normally). This is in 11g, UNDO space is managed manually.
My oracle is sitting on UNIX, i have a sql loader scripts which load the data in oracle at every 10 min and bad files is written into a directory. since the file names are same it overwrite the badfiles in case of error record. i can devise a code to write the bad file with different name. I want to write error record into oracle table, is this possible and how can i achieve ?
I am loading content of an XML file into a table using SQL loader.Below is my Control file script -
LOAD DATA INFILE * INTO TABLE xx_cc_response_xml_stg TRUNCATE xmltype(XML_DATA) FIELDS ( COLUMN_ID constant 1, file_name filler char(4000), XML_DATA LOBFILE(file_name) TERMINATED BY EOF) BEGINDATA B2B_MasterDataUpdate_20120906152137.xml ------------------------------------------------------------------------------------
The file B2B_MasterDataUpdate_20120906152137.xml is correct and XML is well formed.When i try to query for XML_DATA (datatype XMLType) column in the table, i cannot see any content in the record, and it appears as (XMLTYPE)When I parse this XML using the below,
select value(d) from xxnbn_cc_response_xml_stg a, table(xmlsequence(extract(a.xml_data,'/InventorySearch'))) d; ------------------------------------------------------------------------------------ I get this error: ------------------------------------------------------------------------------------ ORA-00600: internal error code, arguments: [qmcxdsSelf4], [], [], [], [], [], [], [], [], [], [], [] 00600. 00000 - "internal error code, arguments: [%s], [%s], [%s], [%s], [%s], [%s], [%s], [%s]" *Cause: This is the generic internal error number for Oracle program exceptions. This indicates that a process has encountered an exceptional condition. *Action: Report as a bug - the first argument is the internal error number ------------------------------------------------------------------------------------
SQL> SELECT * FROM V$VERSION; BANNER ---------------------------------------------------------------- Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Prod PL/SQL Release 10.2.0.1.0 - Production CORE 10.2.0.1.0 Production TNS for Linux: Version 10.2.0.1.0 - Production NLSRTL Version 10.2.0.1.0 - Production
after i backup my database,i check the alert log ,i found the following errror:
Mon May 14 09:19:42 2012 Errors in file /u01/app/oracle/admin/szcargo/udump/szcargo_ora_26967.trc: Mon May 14 09:19:42 2012 Errors in file /u01/app/oracle/admin/szcargo/udump/szcargo_ora_26967.trc: Mon May 14 09:19:42 2012 Errors in file /u01/app/oracle/admin/szcargo/udump/szcargo_ora_26967.trc:
[code]....
the trace number 26967 :
[oracle@shenzhengair archivelog]$ cat /u01/app/oracle/admin/szcargo/udump/szcargo_ora_26967.trc /u01/app/oracle/admin/szcargo/udump/szcargo_ora_26967.trc Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production With the Partitioning, OLAP and Data Mining options ORACLE_HOME = /u01/app/oracle/product/10.2.0/db_1
I have a table which is being load by sqlloader, when i load the table without direct path set to TRUE IT Works well , but when DIRECT path set to TRUE ,it comes out with the following error
SQL*Loader-702: Internal error - Unknown column for OCI_ATTR_COL_COUNT SQL*Loader-2026: the load was aborted because SQL Loader cannot continue.
control file looks like below.
load data BADFILE '/backup/temp/rajesh/RIO/BadFiles/FILENAME' append into table TEMP_rio_RESP_TIME_LND TRAILING NULLCOLS ( INSTALLATION_ID CHAR
Since there is an extra double quote (denoting inch) in the third column, im getting an error. Is there any way to avoid this error without modifying the csv file.
I have a data base oracle 10g (10.2.0.3). I am exporting the Schema using Grid Control 10g.
Servers:
Source: Oracle 10.2.0.3.0
To: Oracle RAC 11.2.0.3
I am trying to import the Schema for the server Oracle RAC 11gR2 11.2.0.3 using the Cloud Control 12c and is presenting the following error:
Error: O job IMPORTA_EGESTAO_ESQUEMA_NOVO foi reaberto em Terca-Feira, 05 Marco, 2013 11:02 Reiniciando "SYSTEM"."IMPORTA_EGESTAO_ESQUEMA_NOVO": Processando o tipo de objeto SCHEMA_EXPORT/USER Processando o tipo de objeto SCHEMA_EXPORT/ROLE_GRANT Processando o tipo de objeto SCHEMA_EXPORT/DEFAULT_ROLE Processando o tipo de objeto SCHEMA_EXPORT/TABLESPACE_QUOTA Execute Failed: ORA-06502: PL/SQL: numeric or value error: character string buffer too small ORA-06512: at line 18 ORA-06512: at line 44 (DBD ERROR: OCIStmtExecute)
I am trying to load the data from .csv file into the table using SQL Loader.
The table has the following schema: src_id : number, dest_id : number, range: intsys.interval_typ --- > a type containing (lower,upper) payload : varchar2(100)
The loader.ctl file is :
load data infile * append into table sb_packet fields terminated by "," optionally enclosed by ' " ' (src_id,dest_id,range,payload) BEGINDATA 3,32,intsys.interval_typ(10043,142703),"misc data"
When I use the following this ctl file to transfer the data, i get the following error:
SQL*Loader-418: Bad datafile datatype for column RANGE
Below function has been used to transfor data and callled in sql loader control file CREATE OR REPLACE function return_domain( domain_name varchar2) return varchar2 as v_dmn varchar2(100)
[code]...
sql loader control file is as below:
load data BADFILE '/backup/temp/rajesh/CERNER/BadFiles/FILENAME' append into table TEMP_CERNER_RESP_TIME_LND WHEN CLINICAL_TRANSACTION_ID = 'USR:ERM PMSEARCH ENCOUNTER RESULTS DISPLAY' TRAILING NULLCOLS
[code]...
function takes the parameter as 'DOMAIN50_LPAR5002_slainterval051712_rj35cmi102_08_45_00.csv '
FILENAME in control file will be replace by DOMAIN50_LPAR5002_slainterval051712_rj35cmi102_08_45_00.csv when i run the the the loader i get the below error.
Record 1: Rejected - Error on table TEMP_CERNER_RESP_TIME_LND. ORA-00604: error occurred at recursive SQL level 1 ORA-01843: not a valid month
We are getting the below errors while migrating partitioned tables using expdp.
The source and target databases are both running on 10.2.0.5 and the main thing is source database doesn't have any active sessions. This is a clone of a Prod Database and no one is accessing it.
ORA-31693: Table data object "DPMMGR"."WHSE_CTNR_EVNT_W":"MSG_PRCS_N"."MSG_PRCS_N_DC556" failed to load/unload and is being skipped due to error: ORA-02354: error in exporting/importing data ORA-01555: snapshot too old: rollback segment number 31 with name "_SYSSMU31$" too small ORA-31693: Table data object "DPMMGR"."RLTM_PRDCT_LOG":"RPL_20120814" failed to load/unload and is being skipped due to error: ORA-02354: error in exporting/importing data ORA-01555: snapshot too old: rollback segment number 14 with name "_SYSSMU14$" too small
Undo Tablespace has enough space but still the expdp is failing.
SQL>/ TABLESPACE Totalspace(MB) Used Space(MB) Freespace(MB) % Used % Free --------------- --------------- -------------- ------------- ---------- ---------- UNDO01 145096 115338 29758 79.49 20.51 SQL> show parameter undo [code]....
I am recieving errors when trying to load the control file. The errors are as follows:
SQL*Loader-500 Unable to open file (homework.ctl) SQL*Loader-553 file not found SQL*Loader-559 SYstem error: The system cannot find the file specified.
My control file is located directly in the C drive (C:homework.ctl). The control file contains the following
LOAD DATA INFILE 'c:country.dat' APPEND INTO TABLE homework fields terminated by ',' optionally encloded by '"' (country, month, day) WHEN (month='April')
The command I am entering is:
sqlldr system/password control=homework.ctl
I've tried c:homework.ctl, 'c:homework.ctl', and placing the file in the BIN folder of Oracle.
I have several environment that are physically not connected to each other (except network). Each environment has its own tnsnames.ora. I have downloaded every tnsnames.ora into C:\OraHome1\network\ADMIN, so TOAD opens connect dialog using the file currently named 'tnsnames.ora'. Every time I need to switch to a different environment (and to different connection string) I have to manually rename the 'tnsnames.ora' into something like 'tnsnames.ora.environmentA' and rename another 'tnsnames.ora.environmentB' into 'tnsnames.ora'.
This is inconvenient and time consuming task as I have to restart TOAD after every rename.if there is an easier way to use multiple tnsnames.ora without renaming those.
Possible to call a tnsnames file from master tnsnames like the principle of ifile in the init.ora...My primary tnsnames.ora has a lot of development entries which are repointed frequently, it would work with logistics if I could manage these in a second tns file and load from primary so when my dev instances are repointed I'm not editing the master tnsnames
LOAD DATA INFILE "gateway.csv" truncate INTO TABLE GATEWAY Fields terminated by "," Optionally enclosed by '"' trailing nullcols
[code]....
and I got the following error:
zcyds891:/opt/oracle> sqlldr gwcem/gwcem@pfs control=gateway.ctl log=/tmp/ldr.log bad=/tmp/bad.log SQL*Loader: Release 9.2.0.8.0 - Production on Tue Dec 7 05:07:59 2010 Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved. SQL*Loader-350: Syntax error at line 12. Expecting "," or ")", found "INTERGER". GATEWAYPROTOCOL INTERGER, ^
I'm working with sqlldr and i try to insert data from a csv file to a CTL file. One field of my table contains 5 characters but one row has 6 characters in this field, so it's rejected by oracle. (Logical, you can't insert 6 chars in a 5 chars field)
an error is visibly returned, so i wondered how you could catch the value of this error?is it a code? a message?
I'd like to add to my script a condition so that the end of the script would continue even if this error code is returned for that CTL execution.
I've been using datapump for a long time now but I have not come across this problem before.
Importing just two tables: Table1 data=100Mb=11 million rows Table2 data=4.2Gb =19.6 million rows
Table1 ran for approx. 5 hours Table2 ran for approx. 15 hours
If I run the impdp importing both tables in the same par file the default tablespace of the users the import is running as runs out of space due to ORA-01691: unable to extend lob segment <owner>.SYS_LOG0001175799C00045$$ by 512 in tablespace USERS. I do not understand why it is creating objects in order to import tables into someone elses schema.
The environment is Red Hat LINUX 4.1.2-51 running Oracle 11.2.0.1 of Oracle11gR2. This is a 9 node RAC using ASM.
I am trying to export a Schema from 11.2.0.2 and getting the below error on export ORA-39127: unexpected error from call to export_string
:=SYS.DBMS_RMGR_GROUP_EXPORT.GRANT_EXP(12425,1,...) ORA-06502: PL/SQL: numeric or value error: NULL index table key value ORA-06512: at "SYS.DBMS_RMGR_GROUP_EXPORT", line 154 ORA-06512: at line 1 ORA-06512: at "SYS.DBMS_METADATA", line 7049 ORA-39127: unexpected error from call to export_string :=SYS.DBMS_RMGR_GROUP_EXPORT.GRANT_EXP(12424,1,...) ORA-06502: PL/SQL: numeric or value error: NULL index table key value ORA-06512: at "SYS.DBMS_RMGR_GROUP_EXPORT", line 154 ORA-06512: at line 1 ORA-06512: at "SYS.DBMS_METADATA", line 7049 ORA-39127: unexpected error from call to export_string :=SYS.DBMS_RMGR_GROUP_EXPORT.GRANT_EXP(12423,1,...) ORA-06502: PL/SQL: numeric or value error: NULL index table key value ORA-06512: at "SYS.DBMS_RMGR_GROUP_EXPORT", line 154 ORA-06512: at line 1 ORA-06512: at "SYS.DBMS_METADATA", line 7049 ORA-39127: unexpected error from call to export_string := SYS.DBMS_SCHED_JOB_EXPORT.GRANT_EXP(2003412,1,...) ORA-01031: insufficient privileges ORA-06512: at "SYS.DBMS_SCHED_MAIN_EXPORT", line 2296 ORA-06512: at "SYS.DBMS_SCHED_JOB_EXPORT", line 52 ORA-06512: at line 1 ORA-06512: at "SYS.DBMS_METADATA", line 7049.
I have searched the forum and found similar errors, but I am struggling to understand the cause and solution.Forums seem to indicate it may be related to bug 4358907, Re: Datapump export/import object grants given by another user but not sure it is relevant as I can see a ORA-01031: insufficient privileges in this log.
Way to share or manage the TNSNAMES.ORA file for all oracle clients in same place?I already have a TNSNAMES.ORA configured and running in the machines (locally).The problem is that any change done( ip, port etc), imply in update all files in all the Oracle clients