SQL & PL/SQL :: Avoid Encoding Dealing When Storing CLOB To File System?
Oct 21, 2010
I am working with oracle 10g / 11. I need to operate with CLOB values of some table, unknown for me: I read them from .Net (3.5), and have to save them to files (Windows) to be loaded after this with DBMS_LOB. Lets say that table contains 3 columns (1 - CLOB) and several rows.I want to store each CLOB value to separate file, and to store it after this into another DB table (another Instance, also).I cannot use dblinks and other techniques like this.
The point is - how to avoid dealing with encoding?
I just want to save the CLOB values to the files, with no meter about Database Instance character set, language, and so on. If I use default for .Net StreamReader UTF-8 format, the loading with DBMS_LOB failed.
what will be the best way to determine the encoding, and to convert Oracle used encoding to .Net one?
I am working with oracle 10g / 11. I need to operate with CLOB values of some table, unknown for me: I read them from .Net (3.5), and have to save them to files (Windows) to be loaded after this with DBMS_LOB. Lets say that table contains 3 columns (1 - CLOB) and several rows.
I want to store each CLOB value to separate file, and to store it after this into another DB table (another Instance, also). I cannot use dblinks and other techniques like this. The point is - how to avoid dealing with encoding?
I mean, that in the best scenario, I just want to save the CLOB values to the files, with no meter about Database Instance character set, language, and so on. How this can be accomplished? If I use default for .Net StreamReader UTF-8 format, the loading with DBMS_LOB failed.
If this is not possible, what will be the best way to determine the encoding, and to convert Oracle used encoding to .Net one?
Oracle procedure was working fine with other XML files. Today I got new XML file and when I try to load the XML,I am getting below error.
ERROR at line 1: ORA-31011: XML parsing failed ORA-19202: Error occurred in XML processing LPX-00283: document encoding is UTF-8-based but default input encoding is not Error at line 1 ORA-06512: at "SYS.XMLTYPE", line 295 ORA-06512: at line 1
XML header is same as previous ones. <?xml version="1.0" encoding="utf-8" ?>
Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options
I'm creating a file using UTL_FILE.FOPEN and UTL_FILE.PUTF
But I don't know the file I created is in which encoding ASCII, UTF-8, EBCDIC etc.
1.) How can I create files in my desired encoding using UTL_FILE? 2.) Does UTL_FILE use database encoding? If yes then how to find out database encoding? 3.) Which encoding is used by UTL_FILE by default?
I am storing customer's snaps in a table ( column's data type as LONG RAW) using oracle forms Webutil. Now there are 250 snaps in the table. The file type of these snaps is JPG with the average size 30KB.
I made a backup using export utility before storing these snaps and the exported DMP file's size was 36MB. Now after storing these just 250 snaps of 30KB the DMP file's size is gone over 300MB.
i need to change column's datatype? or some where in oracle forms's image item. Because on window's file system the size of these files is just 8MB.
The basic requirement is to send an email to the user about order information, with specific subject and the corresponding e-mail body.
Here the E-mail body is very exhaustive, so we are asked to store the content of the e-mail in a word document and store this word document in a tables column of type BLOB.
Once we store this file in the BLOB column, when we send the e-mail to the customer we need to publish the content of the file in the e-mail body and send the e-mail to the user (This E-mail body has some dynamic content as well, that needs to be generated on the fly). This e-mail sending part is done by the shell script.
How can i publish the content of the BLOB object in the e-mail body.
I found there is clear password in the xml file in the process of Oracle instance creating. The process like below
1, when using the following process to create the instance
Login as user Oracle /home/Oracle11home/bin/dbca -createDatabase -silent -templateName /home/Oracle11home/assistants/dbca/templates/Small.dbt -gdbName testDB -sid testSID -sysPassword [b]testpwd[/b] -systemPassword [b]testpwd[/b]
2. Found the file in $ORACLE_HOME/checkpoints/dbca/OraDb11g_home2_oracle_creation_checkpoint.xml , it listed the sys and system passwd (like testpwd for sys and system password) directly.
I am dealing with a bunch of tables containing sales information for an New Zealand organisation. The sale datetime has been recorded as UTC.
New Zealand operates Daylight Savings, so twice a year it changes its clocks.
When New Zealand is on standard time it is UTC+12.
When New Zealand is on daylight savings time it is UTC+13.
Thus an event which actually occurred when New Zealand was on standard time at 2011-08-31 15:20:52 local time, is recorded in the database as having occurred at 2011-08-31 03:20:52. However, an event that actually occurred when New Zealand was on daylight savings time at 2011-10-06 15:20:52 local time, is recorded in the database as having occurred at 2011-10-06 02:20:52.
I want to be able to read the sales dates from my table and convert them to the actual time in New Zealand when the event occurred. The table will contain data for sales that occurred in both standard and daylight savings times.
I do not think that the data has been stored with time zone information, simply that the application writing the data to the Oracle database, calculated the event time as UTC when it occurred and wrote that time to the table.
Does Oracle only know about what UTC-offset is in force right now or is it capable of determining what offset from UTC is required for any given historical date ?
Just wanted to export a clob field to .txt file, the maximum length of the clob field exceeds the limit 32767. So only partial data is exported to flat file. is there any way to export the entire data available in clob field irrespective of the size or lenght.
Is it possible to pass the file into CLOB column in real-time without DIRECTORY in oracle sp. I have not come across but would like to know is there anyway to pass the file directly from .net to oracle.
I have read the following article:[URL] 11070I want to know wherer if there exists a possibility of write a clob or an xml into a file on disk, if we do not have the CREATE ANY DIRECTORY privilege. Many functions, like UTL_FILE.FOPEN, or dbms_xslprocessor.clob2file, or dbms_xmldom.writetofile, need an Oracle directory to be created (with CREATE OR REPLACE DIRECTORY...). But if we don't have this privilege, is there a possibility to export a clob into a file as xml (the clob contains 100% xml, but this is the column data type, CLOB) if we don't have that privilege?The clob data contains 48200 characters.
I am migrating data from DB2 to Oracle. I used DB2 export to extract the data specifying lobsinfile clause. This created all the CLOB data in one file. So a typical record has a column with a reference to the CLOB data. "OUTFILE.001.lob.0.2880/". where OUTFILE.001.lob is the name specified in the export command and 0 is the starting position in the file and 2880 is the length of the first CLOB.
When I try to load this data using sqlldr I'm getting a file not found.Attached is a copy of the control file and output from testing
PS. I cant use the DB2 option LOBSINSEPFILES which creates a separate file for each CLOB column because the table has over 14 million rows....and creating 14 mil files causes OS inode problems...
Attached File(s) sqlldr.txt ( 2.05K ) Number of downloads: 3
I am trying to retrieve images from the file system, but the open dialog box is not popping up for me to choose which file i want. It does nothing gives no error. Just says record 1/1 on the message bar at runtime.
I have two blocks on the module. One is the lobs_table block and the other is the control block. On the control block i have a when-button-pressed trigger on an item. And here is the code below.
I would like to spool a clob column to a flag file, however some of the clob are greater than 32k, and I have to have the same record in a single line in the file. Is there any way to achieve this through spooling?
set heading off set feedback off set term off set long 1000000 set longchunksize 500000 set line 32767 set trimspool on set pagesize 50000 spool file.txt @--this is my select statement. spool off exit
We got a request from Customer to migrate a RAC database of size 1.8TB from HP unix raw file system to AIX ASM with a min downtime. I have seen lot of methods of doing cross plat form migration and non-ASM to ASM but not together.
Do we have any proven method for migration such cross platform migration with raw file system to ASM conversion in a single go with min downtime?
I mistakenly added a datafile to a tablespace which is asm, however the datafile was created in a default location and not the asm location:
alter tablespace pdaiidata1 add datafile '<filename>' size 2048M;
What I should have done:
alter tablepsace <tablespace_name> add datafile '+DATA1' size 2048M;
Is there any way to move this filesystem datafile into the asm tablespace? In previous Oracle versions, I've taken a tablespace offline, moved a datafile, renamed it, then brought the tablespace back online. Can I do something similar here in this situation?
We are setting RAC between two machines(Desktop PC connected wirth LAN ....have storage Local to them and no Shared storage..Only NFS).We have shared the mount point by NFS.We do not have any shared storage but we need to share the RAW files to setup the OCR and VOTING DISKS.
The problem is:
1. We need to keep the RAW devices for OCR and voting disk common to the machines (Desktop PC..both Linux OELU5). I am able to create the raw file systems but that is local to one machine. I am not able to understand how to share the raw devices between two machines.
2. Tried to use ISCSI utility but that did not work. (How_to_use_iSCSI_Targets_on_Linux)
Scenario : One Machine has 500 GB Storage. Second machine has 80 GB Storage. Made private and public networks(Used two Network cards in LAN). All are communicating.
Problem : Need to make the storage 300GB (Out of 500 GB) of Machine 1 as shared storage. How can we do that? (Do we need any virtual software...ex .. vmware for that?)
We have two database servers(10g with OEL 5) to be clustered and two visible disk groups to each of those nodes. So question is can we choose only one disk group as a shared storage leaving the rest one as the redundant copy during database a creation window while installing the database. Because some of us argue that oracle database has a built-in capability to decide on what level of RAID we store our data.
Windows 2003, DB 10.2.0.3...I have to read some files from remote folder, so on remote Windows machine I shared folder c:est (grant Everyone first, and Administrator then) and on DB Server I mapped the remote folder with letter T..All these operations were made with administrator privileges.
In Oracle I create a directory object with this command: CREATE OR REPLACE directory T_DIR AS 'T:';
Then I granted read and write privileges to my user GRANT READ, WRITE ON DIRECTORY T_DIR TO <user>; (I tried also to set PUBLIC and SYSTEM)
But when I try to open a file with UTL_FILE with this command
the error is Exception: ORA-29283: operazione file non valida ORA-06512: a "SYS.UTL_FILE", line 475 ORA-29283: operazione file non valida
If I change my directory object in 'C: est' (local folder) it works correctly.I tried also restart DB while shared folders were connected but with the same result.I tried, from sqlplus, the command host (dir t:) and it works (folder is accessible)