create directory TEST13 as 'C:\temp';
create table test102(id number(4),aaa bfile);
insert into test102 values(1234,bfilename('TEST13','sg1.pdf'));
select * from test102;
1234 (BFILE)
select id, DBMS_LOB.GETLENGTH(aaa) from test102;
ORA-22288: file or LOB operation GETLENGTH failed
[code]...
I stored some pdf files in c:\temp in local machine. DBA suggests not to put the files on DB server. Is there a way to read the pdf files without storing them in DB server?
I was trying to load data from XML files to an Oracle database table.I followed these below steps to load that file data into a table. Created XML_DIR1 as oracle directory where i have kept all XML files.
Create table import_rpt_xml of xmltypexmltype store as binary xml; insert into import_rpt_xmlvalues (xmltype (bfilename('XML_DIR1','I-Yamanouchi-20040525-501.xml'),nls_charset_id('AL32UTF8')));
This insert shows below error: Error starting at line 80 in command:insert into import_rpt_xmlvalues(xmltype(bfilename('XML_DIR1', 'I-Yamanouchi-20040525-501.SGM'), nls_charset_id('AL32UTF8')))
Error report:SQL Error: ORA-31061: XDB error: XML event errorORA-19202: Error occurred in XML processingIn line 69 of orastream:LPX-00217: invalid character 142 (U+008E) I tried to look into my XML and got that it has some Japanese characters in it.
this to deal with japanese characters in XML. I don't want to miss those characters. My databse NLS_CHARACTERSET is 'AL32UTF8'.
I am trying to load multiple XML files into Oracle DB using SQL Loader. The filenames of the XML files starts with a description and then numbers, where the numbers are different each time.
Here's my CTL file:
LOAD DATA INFILE * INTO TABLE XML_TABLE TRUNCATE xmltype(XML_TABLE) FIELDS (
[code]....
I don't want to keep having to go into the ctl file and change the numbers of the xml file. Is there a way where I could just load all .xml files that begins with 'description'? Like maybe
I have a bunch of data in 50 excel files. I need to load all these 50 files into 50 different tables. I would like to do this in one script. I went through the forum to get this information, people suggested create a shell script etc or list the sqlldr command multiple times etc.
provide some clarity on this as to what's the best approach.If it is through shell scripting provide the shell script and instructions to execute it. Iam new to shell scripting.
the following situation, I have a directory named /dat/global/stock/ inside this i will get files named differently for example below.abcdef.112dfgrt.2......
Here i want to load this file one by one into the external tables and generate one more file based on some enrichment.
Step 1. Have to take first file and to load into the ext table. Step 2. Enrichment Step 3.File generation.
Now here i am facing a problem that in that particular directory i usually get 1000 files so i need to get file one by one and to put in one more directory. how can i get file one by one and generate file by using oracle loader
Is there any way to upload a file to APEX as BFILE instead of BLOB? I need it to be BFILE since I will be only using XE which means I have limited data storage and unfortunately the system will allow users to upload and download files of different formats (.exe,.bat ) etc. which can be as big as 10MB. Also, how to download these files considering the authentication of the application.
I have 1M Records coming from an External Data source as a Flat File (using ETL). Now I need only Yesterday's data only to load in my Database Table.
this can be done using Bulk Load and Filter.
write the CODE.
Second Part:-
Hint: if I need to update only those records been updated Say the Address1 field is updated. So this records need to update in my Master Customer Table.
If I have many fields in table and any records that are modified (coming to me from External Datasource as a Flat file) how to identify and update that record in my Master Customer Table?
I successfully loaded a record into table with BFILE data type. However I don't understand how to verify if I did this correctly. I am also suspicious of the value FSSLPB//I. FSSLPB is the directory object name. When I delete the record from the table the BFILE column value defaults to (null) so it seems the table is initialized for accepting bfiles.
I am trying to:
1. Load an external LOB into a staging table using SQLLDR, then 2. Using PL/SQL insert the ext LOB into a BLOB column from the staging table to an end user table (or application table).
I am overwhelmed with reading docs and not getting the complete picture for different approaches to loading into Oracle.
I have written make files that compile .pc files in unix. This was for several projects that use an oralib source code directory.Just running proc on one target .pc file works fine on unix. I am trying to use proc - Oracle 10.2.0 - in windows and I keep getting:
Quote:unable to open include file #include <stdio.h> and other C library headers.
I am doing all development under cygwin, this way I can write a makefile just like under unix instead of using nmake.All C library headers are in /usr/include When I run proc on Solaris as that:
proc program.pc No problems, and I do get program.c.
However in windows I get the previous error message. I have tried to do proc include=/user/include program.pc and proc include=/user/include parse=full program.pc but I still get the same error message.
I have a console application that needs to be scheduled in task manager. My system is 32-bit operating system and I’m using Oracle.DataAccess.dll in my application to establish connection to the oracle db. The version is 2.112.1.0 and the processor architecture of this dll in C:WindowsAssembly is x86. In my local m/c this dll works fine with all 3 build and target platforms – x64, x86 or AnyCPU. But when I copy the files to my staging server which is a 64-bit OS I’m getting the following exception. (Note: I’m also having Oracle.DataAccess version 10.2.0.100 which is also x86 available in C:WindowsAssembly)
System.BadImageFormatException: Could not load file or assembly 'Oracle.DataAccess, Version=2.112.1.0, Culture=neutral, PublicKeyToken=89b483f429c47342' or one of its dependencies. An attempt was made to load a program with an incorrect format. File name: 'Oracle.DataAccess, Version=2.112.1.0, Culture=neutral, PublicKeyToken=89b483f429c47342' at
Assembly manager loaded from: C:WindowsMicrosoft.NETFramework64v4.0.30319clr.dll Running under executable D:ProjectFolderMyExecutable.exe Assembly manager loaded from: C:WindowsMicrosoft.NETFramework64v4.0.30319clr.dll Running under executable D: ProjectFolderMyExecutable.exe
--- A detailed error log follows. === Pre-bind state information === LOG: User = UserId LOG: DisplayName = Oracle.DataAccess, Version=2.112.1.0, Culture=neutral, PublicKeyToken=89b483f429c47342 (Fully-specified)
[Code] ......
I have tried to build the application to target AnyCPU / x64 / x86. It fails in all 3 scenarios. There are other applications in the staging server where Oracle connection can be established. So ODP.Net should be registered in the server. So looks like problem with my console app.
Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production PL/SQL Release 11.2.0.2.0 - Production CORE 11.2.0.2.0 Production TNS for IBM/AIX RISC System/6000: Version 11.2.0.2.0 - Production NLSRTL Version 11.2.0.2.0 - Production
PARAMETER VALUE
NLS_LANGUAGE AMERICAN NLS_CHARACTERSET AL32UTF8 NLS_NCHAR_CHARACTERSET UTF8
When i am trying to load an XSD document its failing with Error-15. When i removed the below line from the xsd doc
It went through fine. The issue is its not supporting ;А-я NCS of Cyrillic alphabet(Russian) . Guess its some kind of character set issue. But i cant identify. The same doc is working fine in 10g. issue is in 11gr2.
Loading of xsd happens from a product interface. header details
Though the count in the Table after load is the same as the input file, they are all NULL.
I tried adding FIELDS TERMINATED BY X'A' for new line and also FIELDS TERMINATED BY X'D' for carriage return. Both times bad file was created and the records that were loaded were again NULL.
The input file has a list of emails: iatraveler2008@aol.com iaz65@aol.com 2blue2brown@comcast.net 2c3mwilson@embarqmail.com abigailolschan@comcast.net imisskoco@aol.com
I tried FIELDS TERMINATED BY X'10' and FIELDS TERMINATED BY X'13' too for new line and carriage return respectively. This time there were no bad file created, but the Table has Null values.
Log File contents:
Number to load: ALL Number to skip: 0 Errors allowed: 10000 Bind array: 64 rows, maximum of 256000 bytes Continuation: none specified Path used: Conventional Silent options: FEEDBACK, ERRORS and DISCARDS
we have a problem with forms containing webutil - they start very slow, after like a timeout, while forms without start ok. since i think it has something to do with name resolution i'll first describe our environment.
we have an external net with users, ip range 10.48.x.x, connected to a cisco router which is resolved in that network with the name 'cro'. so the users from that external net access us via 'http: //cro/.....'.
behind that router we have our production net, ip range 192.168.x.x, with forms-, reports- and http-servers. external users accessing our forms server will have to klick a link "http: //cro:8888/forms......'. that way they acces the cisco router which, depending on the port (here 8888) will NAT them to the correct forms server, say http: //forms1:7778/forms....' - up to here everything works well and fast with 'normal' forms (eg. without webutil).
forms *with* webutil on the other hand take like 30-60 seconds to load from that external net (via the router) while they load fast when called from a test-client in the internal 192.168.x.x net. after searching i found [URL] with a hint to alter the clients hosts file.
what i've found out so far is that the clients from the external 10.18.x.x net try to name resolve the *forms servers* in the internal net 192.168.x.x - i can cheat them by providing a fake hosts file which resolves the server named forms1 to the ip of the cisco router cro, but thats a temporarily and bad solution since i cant do that to real clients because i have no access to them.
so... how can i tell a webutil loaded form not to try to connect to the server they are started from (forms1) but to use the host they have been given to start from (cro)? or is there any other solution (which ofc cant include fiddling with the external net or external clients)?
We are getting the below error frequently from the application while doing insertion/dataloading to a table. The mentioned error is in the Primary key index
Error: 'ORA-01502: index 'INDEX_NAME' or partition of such index is in unusable state'.
I set the value SKIP_UNUSABLE_INDEXES = TRUE using the command 'ALTER SYSTEM SET SKIP_UNUSABLE_INDEXES = TRUE' to avoid this. Again we are getting the same error and Every time Iam rebuilding('alter index INDEX_NAME rebuild') the index and doing the DML Operation.
We have requirement such that whenever stored procedure is executed, their resultant records has to be stored in excel file ( Just like an reports ).No third party tool or reporting tools are used.
is there any option in oracle (Stored procedure or built in packages ) which can create excel file with the resultant records.
am trying to load java resource in oracle llgr2 using load java. It shows resource loaded but I am not able find it in database dictionary. I configured JVM also on database and gave all the neccessary grants. whenever I am trying to encrpt it give me
SQL> select banner from v$version 2 ; Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production PL/SQL Release 11.2.0.1.0 - Production CORE 11.2.0.1.0 Production TNS for IBM/AIX RISC System/6000: Version 11.2.0.1.0 - Production NLSRTL Version 11.2.0.1.0 - Production
I have a requirement like, I received .dat files and placed in my local directory and there is a process/method where this data which is there in .dat file will insert into my oracle external tables?Any link of the example with clear steps?
The manual work around on populating child tables for testing purpose are taking long time and its very painful work. So I am trying for a tool that takes parent table name and child table name as input and produce insert statements for child table with foreign keys as output.
Is it Possible doing SQL LOAD into Varray table having two inner objects.
find the structure of the table and its types for your reference
CREATE OR REPLACE TYPE OB_TEST_INFO AS OBJECT ( AGE NUMBER ( 3 ), NAME VARCHAR2 ( 14 ) ); /
CREATE OR REPLACE TYPE OB_TEST_INFO_VARRAY AS VARRAY( 400 ) OF OB_TEST_INFO ; /
CREATE OR REPLACE TYPE OB_TEST_MAINTENANCE AS OBJECT ( BREAKOUT_TYPE NUMBER ( 1 ), EXISTING_STRIPS_FLAG NUMBER ( 1 ), OB_TEST_INFO OB_TEST_INFO_VARRAY ); /
what empty blocks are, and how to remove them.What I'd like to do is not have empty blocks in the first place on loading a table. I load a lot of "static" tables and would like to not have any wasted space at the end, with minimal shinanigans.
I've set pctfree 0 I"ve set initial to close to the end table size I've set next to 1M I've set pctincrease 0 blocksize is 8k
Yet I still need to at least do an alter table deallocate unused
For example, we have a table ACCOUNT (snowflake dimension containing other dimension keys) and I have many fact tables based on this dimension. Normally data warehouse load happens like first dimensions needs to be loaded and then facts. Our frequency of loads is 30 mins.
To increase the rate in which the data will be available in the facts (as its a financial application), am considering to have two batches one with dimension and another one with fact (came to this conclusion as there is no dependency like first dimensions to be loaded then only fact) just the update might get missed sometimes. But if I do that, when dimension gets loaded, it will be read in the facts in another session. Will this affect the performance ?
LOADING (insert/update) and selecting data from table at the same time. Will it affect the performance in any way.
i am trying to load data into a table in a remote database schema, and my files are residing on another remote Server, Server having the files does not have a DB installed, i just need to know that if its possible or not..
ERROR at line 1: ORA-29913: error in executing ODCIEXTTABLEOPEN callout ORA-29400: data cartridge error KUP-00554: error encountered while parsing access parameters KUP-01005: syntax error: found "badfile": expecting one of: "column, enclosed,