I am using sql loader to load data to tables from tab delimeted files.
Here the problem is that the sql loader is not handling null values. If there is any null value in the flat file it is moving the field values to left and loading to the table. I am using NVL function to handle the null values, but it is not working.
My control file is: LOAD DATA INFILE 'C: da_poc_filesSQL_scriptsSourcefilesTRADEGLOBNODE1.TXT' BADFILE 'C: da_poc_filesSQL_scriptsBadfilesTRADEGLOBNODE1.bad' DISCARDFILE 'C:C: da_poc_filesSQL_scriptsDiscardfilesTRADEGLOBNODE1.dsc'
[Code]....
Source file is attached to this link For the attached file in the first record, tradedate value is coming in to tradeprice field
create a new DB service and apply the load , then I need to verify the service load from Enterprise Manager or table V$SERVICEMETRIC .
Below are what I did
Quote:I tried to add service in tnsnames.ora like below
SERVICE = (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(PORT=25057)(HOST=dadvmc))(CONNECT_DATA=(SID=tgc4)))
/************ Also I tried the serviceName like below, but not working when I conneted SALES1 from sqlplus SALES1 = (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(PORT=25057)(HOST=dadvmc))(CONNECT_DATA=(SERVICE_NAME=host.us.xxx.com))) **********/
and login as
sqlplus scott/tiger@SERVICE[/email][/email]
then lock table EMP in exclusive mode in two session.
I can see the load is added to service SYS$USERS , but I can't see the new service from V$SERVICEMETRIC or Enterprise Manager.
I have oracle8i database on windows 2008 but not a domin server now i have domin server and i want to join this server on domin network i try to join domin it fine but my oracle8i service did not start what to do ?
old hostname=bagmati afterchange hostname=bagmati.alkatel.org.np
On last Saturday we encountered error 'TNS-12519: TNS:no appropriate service handler found" in our listener log for some moment of time let say around 5mins and later it got all right itself.
I created a script to make my admin server a windows service. I successfully made it a service but when it starts it then shuts itself down. i ran a debug and found a error. I dont get what to do. I have tried changing the variables in the script and try it over and over again .....here is my script code and here is the error i got from debug output:
script
SETLOCAL set DOMAIN_NAME=FRClassicDomain set USERDOMAIN_HOME=C:/Oracle/Middleware/user_projects/domains/FRClassicDomain set SERVER_NAME=FRweblogic
I have a problem some user modified the objects and now objects became invalid .How to trace the ip address and operating system username and service ip who modified the objects?
I have deinstalled Oracle database, ASM and oracle software. I installed the oracle software, ASM and database. I have seen the error in the ASM trace called
Errors in file /opt/oracle/admin/+ASM/bdump/+asm_gmon_10518.trc:
ORA-29702: error occurred in Cluster Group Service operation ORA-29702: error occurred in Cluster Group Service operation
I want to load lakhs of records into a table. My problem is when after loading the ¼ of records my process is abend due to the size of my rollback segment area. I don't have an option to increase it. So, Is there any way to go for intermediate commits when I am using the imp or sqlldr utilities to load the entire data without abend?
I am familiar with tool Netca. However there is one more utility exist for the same functionatlty which is netmgrI checked with many DBAs for the exact difference, however I did not get the best answer from them. I also have checked in google but not exactly got the difference. list the exact difference between those 2 tools (netca, netmgr)
I am attempting to use an rman backup form our live server to restore live data over an existing dev database with a different name. I cant use the duplicate command as the versions of oracle are different (10.2.0.4 EE on live, 10.2.0.5 SE on dev) So I copied the backup sets including an autobackup control file onto the dev server. I set the ORACLE_SID to the DEV version (SUN43DEV) and startup nomount and restore controlfile from the backup location. This works and restored the live controlfile. However, I then wanted to mount the database and catalog the backupsets before making a note of the last archivelog etc and restoring the backup with set newname options. But when I try to mount the database I am told that the database has a different name (SUN43 is not SUN43DEV). I read somewhere that I need to set the ORACLE_SID to the live value (SUN43) and restore as that, then use nid TO RESET THE DBID and DATABASE NAME. However, my tnsnames on DEV has no entry for the live database, and we do not have any service defined through ORADIM for the live instance on the dev server. Should I be creating an entry in DEV tnsnames that actually looks at the LIVE server? or do I need to create a service and tns entry etc for SUN43 on my DEV server pointing at the DEV server I have not seen any mention in instructions of creating a service or adding the entry to tnsnames, or indeed adding an entry to the listener for the original live sid. I am worried that if I actually point to the live server then the restore would be attempted on the live server not dev even though I am running it from Dev.
Also, do I need to drop the existing database on DEV before I restore or will RMAN just restore over the top of it?
Finally, if I try to add an entry to the LISTENER.ORA for the new sid, when I stop and start the listener, or reload, the entry is not recognised. I also noticed that the existing SUN43DEV entry is not in there either, yet I can still use that SID - I am assuming that is because the listener is not being used and the sid is recognised from the existence of the service in windows services instead?
I have exported data of one user an importing into another schema at another server. when i am trying to imoport it is working fine for quite no of imports into tables, but after some time it starts giving me below mention error...
IMP-00008: unrecognized statement in the export file: < IMP-00008: unrecognized statement in the export file: < IMP-00008: unrecognized statement in the export file: <ے IMP-00008: unrecognized statement in the export file: +A IMP-00008: unrecognized statement in the export file: [code]...
I have a requirement to read flat text file(around 15000 lines) residing at a client location from DB server and write into a table in One cell.
I tried UTL_FILE and DBMS_LOB but, i am not able to access client location to read the file as it reads path from Oracle Directory.
eg. my client path is 198.168.1.1 and my DB server is in unix say 192.168.1.10. file location is: \192.168.1.1shareabc.txt So I created One Oracle directory as MY_DIR having DIRECTORY_PATH as '\192.168.1.1share'. But both UTL_FILE and DBMS_LOB is not able to access the file.
Error Message: ------------- Unable to process CLOB -22288 ~ ORA-22288: file or LOB operation FILEOPEN failed No such file or directory
Few Details for reference: ------------------------- File Location: \192.168.1.1shareabc.txt Unix DB Server location: 192.168.1.10 Table : Test (filename varchar2(30), Content CLOB) Oracle Dir: MYDIR Directory_Path: \192.168.1.1share
I have a stored procedure that does a "select name into v_name" SQL statement, which works fine. The only problem is when the query finds no data (the procedure will error because there is no value to put into the variable). Now i have a work around to this by running the query first with a count statement (which will always have results) and then if it is not equal to 0, then i will run the select into.
My question is, is there a better way to handle this kind of issue?
The following code is working fine,But the thing is if column already exists in the table,then also the other statements should be executed instead of coming out of procedure.SO how can I handle that exception??
SQL> CREATE OR REPLACE PROCEDURE sp_execparameters(tname IN VARCHAR2, colname IN VARCHAR2,datatype IN VARCHAR2) 2 AS 3 v_sqlstr1 VARCHAR2(1000); 4 BEGIN 5 v_sqlstr1 := 'alter table '||tname||' add '||colname ||' '|| datatype ; [code].........
we are using oracle database. We hold the company records in the database. The records of the company should be available at anytime but over years the database keeps growing. Now how to handle the old data. All the data is important but if this goes for few more years then we need more and more disk space to handle. Is there any efficient methodologies to handle the old data? For us old mean the data that is 10 year old.
I've a question regarding difference of character sets, while taking a export(logical backup) of database on directly to server(linux RHEL 2.1 AS) and export on a client (windows xp prof machine, where only a oracle 9i client is installed). On server it seems to fine and okay, but on client node i'm getting following error for almost all tables.
EXP-00091: Exporting questionable statistics.
My question is :
[1] Is it creating any sort of problem, if later on i import the data which was taken from client node.
[2] Why there is a difference(marginal) in dump(.dmp) file size.
[3] Is there any way to overcome it, or it is the natural behave of it. Means not a problem.
[4] If i'm using a long or blob as datatype for some of my table,is they have any problem if i persist like above.
Additional Information about character sets On server node :
Export done in US7ASCII character set and AL16UTF16 NCHAR character set server uses WE8ISO8859P1 character set (possible charset conversion)
On client node :
Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set server uses US7ASCII character set (possible charset conversion)
1.Header(Contains the File Name,Branch Name,MIS date) 2.Body(Customer Details) 3.Footer (File Name,Contians Total Number of Records and Number of Customers)
from the above code I want to execute both the inner block exception and outer block exception and is there any way to pl/sql engine that execute the outer exception first and inner next