I have table which contains huge data. around 12 lakhs records. when I use sum function on accountname and docdate it gives wrong value. once I restart the server it gives the correct value. one or two days it gives correct value after that again I get the same problem. If I restart again it gives correct value.
I use Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 64 bit server on Linux.
I am using the following query with like 'T_%', i am getting 80 rows out of which the first table_name doesn't even have a beginning part 'T_%'.
the first table name has not started with 'T_', why is it appearing.
********************************************************************* SELECT 'Truncate table epic500.'||table_name FROM user_tables where table_name like 'T_%' order by table_name; ********************************************************************* output: Truncate table epic500.TEMP_ENC_DEL Truncate table epic500.T_ACCOMMODATION_CODE
I am actually having problems with a report that is generating wrong data in Windows 7 only. I have the same report working correctly on many Windows XP Computers.
I have one prod server (11.1.0.6.0) servers on Windows 2003 R2 64 bit.
Server Name (PRODDB)
I do not have access to that prod server , i want to take one export data pump from my client machine and due to space issue in prod server , i want to keep dump file in my client machine itself. i can take traditional export and keep the dump file in my client machine but i do not know how to achieve the same via data pump ...
How to generate dump file in client machine itself via data pump ?
export data program "ociuldr" can not run in 64bits win2008 environment. Where can i download 64 bit version of "ociuldr" program. I have read some article . The article mention that the export data program ociuldr.exe need to recompile into 64 bits version. Finally I also want to ask the import data program sql loader "sqlldr.exe". Can it run in 64 bits environment. Where can i downlaod 64 bits version of "sqlldr" program.
I am creating physical standby database through Rman duplicate command from 2 node rac cluster. rman do all its work. now am try to start the mrp process on physical standby database. I am getting following errors
------------------------------------------------------------ Check that the primary and standby are using a password file and remote_login_passwordfile is set to SHARED or EXCLUSIVE, and that the SYS password is same in the password files. returning error ORA-16191 ------------------------------------------------------------ ORA-16191: Primary log shipping client not logged on standby ------------------------------------------------------------
I copied the same pass file from primary to standby and many times verify the same but i got the same error.
Export and import of data in oracle forms...i have created 02 boutons one for export his trigger like this:
eclare alrt number; v_directory varchar2(200) := 'c:ackup'; --- that if the C Drive not the Drive that the windows had installed in it. path varchar2(100):='back_up' ||to_char(sysdate,'dd_mm_yyyy-hh24_mi_ss'); v_exp varchar2(200) := 'exp hamada/hamada2013@orcl file = ' ||v_directory ||'' ||path ||'.dmp'; [code]....
this code is correct he expot not only the data but also the creation of the table ....for exemple i do export and everything is good until now and i find the .dmp in the folder backup .. but when i deleted all data from my app and try to import this .dmp iit show me error it tell me thet the table phone is already created...just export the data of phone not the creation of table and data ???? or how can i import just the data from this .dmp ??
I have a requirement to import text files which are generated from 3d modelling software xsteel where it records all geometric information and i want to import this information into oracle table.
I need to export only the data from schemas or tables, how to do that with Oracle Data Pump? when we use schemas parameter this export all schema, not only the data right?
I would like to ask if there is the possibility using the data pump export utility to export my full database plus some partitions tables by selecting specific partitions. Can i have all these criteria in only one data pump export? If yes any example?
I'm stuck up with a requirement, where I have to export data of all tables from a schema into separate table specific CSV files.
I tried using SQL Plus 'SPOOL' but it gives me additional queries executed on top of CSV generated. Also, i'm not getting the headers (coz i'm using pagesize 0 )
I am new to ORACLE and i am having a table with images stored in that.Now i want to export only data into a another server with out creating the table again.Because table is already available in another server.Can any one give me a best example to finish this
Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production With the Partitioning, OLAP and Data Mining options ORA-31626: job does not exist ORA-31637: cannot create job SYS_EXPORT_SCHEMA_01 for user KAILAS ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95 ORA-06512: at "SYS.KUPV$FT_INT", line 600 ORA-39080: failed to create queues "KUPC$C_1_20111001165007" and "KUPC$S_1_20111001165007" for Data Pump job ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95 ORA-06512: at "SYS.KUPC$QUE_INT", line 1555 ORA-00832: no streams pool created and cannot automatically create one [oracle@localhost dbs]$
i have problem with export data from SQL to excel format my question do i can export directly data from sql to excel or i have to write sql statement in form report and from form i can export it to excel
I have large data in the one table(approx. 2GB and above). I want to load the data in one flat file.
When i use spool - it is loading half data and remaining it is corrupting. In UNIX, i kept .sql file and with that i am exporting in .dat file. How to export the large data into flat file(.dat). is there any way to load command which wil be used in UNIX.
I need to take only backup of schema objects with out data using exp (export) into .sql file and need to run that .sql file in the target.because I dont have exp/imp privs on target database.
I am trying to export data into excel (office 2007 and above- .xlsx) using the calls mentioned below, but i am getting an error. i am able to upload data and open the document in IE succesfully using office 2003 format(.xls)
The following are the calls made in the package and it works for .xls format
OWA_UTIL.MIME_HEADER('application/vnd.ms-excel');
The following are the calls made in the package and it is not working even thow i can see that the document getting opened .xlsx format
I have a set of store procedures that will construct the header and data dynamically. This procedure will return a CURSOR.Now, I will write a new procedure to export data file by calling the above store procedure.
a) is it possible for me to retain the dynamic header when I export the data out ?
b) use 1 export data file procedure to handle it without coding for each data file I want to export.
I have been testing manually creating the header. I am assigning the header string myself.
UTL_FILE.PUTF(fHandler, header_string);
and then use a cursor to loop through the data for each store procedure.
I have a specific requirement. Currently in our system a SQL script is automated trough TOAD DATA ANALYST (Contains 50 sql Select statement) and all the output are stored in a single excel file in multiple worksheet. All these sql statements are running sequentially. Now my requirement is
1) All the sql queries will be executed in database in parallel . 2) All the output of select statement will be stored in a single excel. 3) Idea using any Client but Toad is preferred.
If I am using the DESFORMAT=DELIMITED then I am getting unformated data and some time I am getting less data as per PDF output.
There are about 25 queries with interrelated way.
How can I get the output in excel from a RDF.
In Reports 6i Functionality was working while we were running reports because of reports were running in Previewer mode but in 10g reports output coming in PDF format.
I need to create a new schedule to export schema every day for Oracle on Linux Server. I am using OEM, But in Job Schedule of Export Schedule I can not see the repeat option.
i need to export master data in excel sheets to our database and we use toad too. How i can export the data with the use of macros in excel. how i can export data from excel to oracle.