I scheduled a RMAN backup job to take FULL database backup to a shared network drive.
RMAN Script used:
RMAN> run 2> { 3> allocate channel ch1 device type disk format '\\BACKUP1635\X$\ARC%U'; 4> backup database plus archivelog; 5> release channel ch1; 6> }
Environment: Windows server 2003 OS and Oracle 10.2.0.4.0
Iam facing below error.
RMAN-00571: =========================================================== RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS =============== RMAN-00571: =========================================================== RMAN-03009: failure of backup command on ch1 channel at 02/20/2011 18:19:46 ORA-19504: failed to create file "\\BACKUP1635\X$\ARC1BM55JL8_1_1" ORA-27040: file create error, unable to create file OSD-04002: unable to open file O/S-Error: (OS 5) Access is denied.
I checked shared network drive and found READ/WRITE access for me. I am able to take RMAN backup in local system without any issue.
I am using the HOST command within Oracle Forms 10g on Windows server to copy a pdf file from the application server to a network folder on another Windows server.
The command works fine if I logon to Windows app server and type it. But it doesn't work when it is launched within Oracle forms using HOST. The command looks as follow
net use \\FileServer\e-file /user:usergroup\username password & copy D:\RepTmp\myReport.pdf \\FileServer\e-file\test\myReport.pdf
Oracle forms doesn't give any error message at all but the pdf file never gets there. Even when I map the network folder as "R:" drive on the Windows app server, a simple copy command like this does not work with HOST:
copy D:\RepTmp\myReport.pdf r:\test\myReport.pdf
Oracle Forms server is running as "SYSTEM" on the Windows server. If HOST won't work, is there any other way to copy the file other than using fpt?
I take a backup on PROD and want to use that backup on test to duplicate the PROD db.I ran this first on PROD:
catalog rman11cv/password@metarep connect target / RUN { ALLOCATE CHANNEL disk1 DEVICE TYPE DISK FORMAT 'G: estore\%U'; BACKUP DATABASE PLUS ARCHIVELOG;
I runs fine.map the network drive on my TEST machine. I run the script:
connect target sys/password@pdm; connect auxiliary /; run { allocate auxiliary channel 'ch1' type disk; SET NEWNAME FOR DATAFILE 1 TO 'D:oracleoradatadatabasedataSYSTEM01.DBF'; SET NEWNAME FOR DATAFILE 2 TO 'D:oracleoradatadatabasedataUNDOTBS01.DBF'; SET NEWNAME FOR DATAFILE 3 TO 'D:oracleoradatadatabasedataCWMLITE01.DBF';
[code]....
Its trying to read from the right backupset.But Im not sure it can see it:
channel ch1: reading from backup piece G:RESTOREN3MKSOMQ_1_1 ORA-19870: error reading backup piece G:RESTOREN3MKSOMQ_1_1 ORA-19505: failed to identify file "G:RESTOREN3MKSOMQ_1_1" ORA-27041: unable to open file OSD-04002: unable to open file O/S-Error: (OS 3) The system cannot find the path specified.failover to previous backup
Net view of the PROD host gives this:
D:oracleadminduplicatePDMTEST>net view myprod Shared resources at myprod
Share name Type Used as Comment ------------------------------------------------------ restore Disk G: The command completed successfully.
How can i shift my datafiles to other drive like d: , they have become very big. and i also want to take backup on other drive like d: and oracle is installed on c:
I have a type to table where only few columns are SCD 2. Like address, city, zip.I wrote a code using cursor. But now the requirement is to replace the cursor .how to perform type 2 using Merge in oracle 10g or any better way to perform it without using cursor. Look the code which i have attached with cursor.
CREATE OR REPLACE PROCEDURE Proc_cp_jci_contract_dim_hist IS l_fun_ret_value NUMBER; v_src_id NUMBER; rec_jci_contract_dim jci_contract_dim%ROWTYPE; rec_jci_contract_dim_hist jci_contract_dim_hist%ROWTYPE; err_row VARCHAR2(4000); ind NUMBER(1) := 1; [code]....
Iam usin 11.2.0.3 database on linux 64 bit machine
I have got the report from the users that the dataload operation which they perform in the database is taking normal than the expected time
I rebooted the server and started the database i can see that the database startup is even taking more time
The memory allocated is sufficient and when i tried to perform the logswitch the database simply hangs
alertlog contents after restart ---------------------------------------
minact-scn: useg scan erroring out with error e:12751 Tue Oct 16 11:02:03 2012 minact-scn: useg scan erroring out with error e:12751 Suspending MMON action 'Block Cleanout Optim, Undo Segment Scan' for 82800 seconds Tue Oct 16 11:12:47 2012
Active Session History (ASH) performed an emergency flush. This may mean that ASH is undersized. If emergency flushes are a recurring issue, you may consider increasing ASH size by setting the value of ASHSIZE to a sufficiently large value. Currently, ASH size is 50331648 bytes. Both ASH size and the total number of emergency flushes since instance startup can be monitored by running the following query:
select total_size,awr_flush_emergency_count from v$ash_info; Tue Oct 16 11:17:53 2012 Suspending MMON action 'undo usage' for 82800 seconds Tue Oct 16 11:28:05 2012 Suspending MMON action 'metrics monitoring' for 82800 seconds
I have prepared shell scripts to do the parallel inserts on my DB table (LEGACY_SYSTEM).
There is a trigger (AFTER INSERT ON EACH ROW) associated with the above table. I am calling a package.function inside the trigger to do the required operation and finally it will insert records into my target table (PRICE_CHANGE).
Expectation: ------------ If I insert 10 rows into LEGACY_SYSTEM table, it should do few updates and finally insert 10 rows into PRICE_CHANGE table.
Result: ------- 10 rows got inserted into LEGACY_SYSTEM. All the updates are successful but I could see only 4 rows in PRICE_CHANGE table. If I run it for the second or third time, all the results will be perfect.
Instead of these shell script, if I insert one by one rows manually into LEGACY_SYSTEM table, I am getting all the expected results and the results are consistent. If you look at my scripts below, you will understand the problem better..
I am calling test_global.sh through the UNIX session and all the records got inserted into LEGACY_SYSTEM table and few rows are missing from PRICE_CHANGE table.
If I remove the '&' symbol and execute, the results are perfect. But the requirement is not to remove the '&' symbol. I have been facing this problem for the past 1 month.
I was just wondering how Oracle manages multiple sessions in a database performing DML. I believe this is related to 'Read Consistency' and I tried to search for the same but could not get any satisfactory online documents.
CASE 1: user A logs in to a database1 issues select on table A and then inserts 4 rows user B logs in to databse1 issues select on table A and then inserts 5 rows issues rollback user C logs in to a database1 issues select on table A and then inserts 6 rows issues commit
How many rows can user C see in the table A when he issues select?
CASE 2: user A logs in to a database1 issues select on table A and then inserts 4 rows user B logs in to databse1 issues select on table A and then inserts 5 rows user C logs in to a database1 user B issues rollback user C issues select on table A and then inserts 6 rows issues commit
How many rows can user C see in the table A when he issues select?
NOTE: All the users are currently logged in to the same database and none has logged out.
In my form i have a multi record block which is based on a table in which i am performing insertion, deletion, updation.
My requirement is as below,
While saving a records i want to check distinct value of specific item can not be greater than some value say 2. Then only my records should be saved. I am planning to populate those values in a collection table type and take the distinct values from it. How to populate record values in a collection table type.
We are facing serious problem while refreshing materialized views using fast refresh option in ORACLE..For the very first time we are performing Complete refresh of data from DB1 to DB2 for few tables.Ongoing we are performing Fast Refresh.Sometimes the fast refresh works fine without any error and sometimes it fails with the below error.
ERROR at line 1: ORA-32320: REFRESH FAST of "CIR"."C_BO_COMM" unsupported after cointainer table PMOPs ORA-06512: at "SYS.DBMS_SNAPSHOT", line 803 ORA-06512: at "SYS.DBMS_SNAPSHOT", line 860 ORA-06512: at "SYS.DBMS_SNAPSHOT", line 841 ORA-06512: at line 1
let me know when will this error occur during Fast Refresh.
I want to open any drive or specific folder in forms 6i in when button press trigger like we open any drive and folder in my computer or folder in windows
In my drive E: I'll have a space issue soon because one file is taking all space. SYSTEM01. DBF size is around 25GB. I want to know if I delete some data in my database I'll gain space.
is it possible to get information about free space on drive/path for example with some DBMS_*, UTL_* package or with stored function based on Java
(e.g. CREATE OR REPLACE FUNCTION GetFreeSpace(driveLetter IN VARCHAR2) RETURN VARCHAR2 AS LANGUAGE JAVA NAME 'DiskSpace.getFreeBytes(java.lang.String) return java.lang.String') but for JDK 1.3
I installed oracle 11g successfully in my windows7 operating system. After installation it is not allowing me to access C drive. And in the properties of the drive the memory of C drive is shown 0 bytes. And other applications installed in C drive are also not working. Even after uninstalling oracle 11g access to C drive is denied.
Tape streaming is not happening while performing RMAN tape backup. On investigation, you find that it is not because of the incremental backup or the empty file backup and that RMAN is sending data blocks to the tape drive fast enough.What could be a solution to make tape streaming happen during the backup?
A.Configure backup optimization
B.Configure the channel to increase MAXOPENFILES
C.Configure the channel to increase the capacity with the RATE parameter
D.Configure the channel to adjust the tape buffer size with the BLKSIZE option
I am creating a cross platform database application using Mono. It uses oracle database as back end. I used ODBC drive to connecting database. I can connect to database using this following connection string in my windows machine
Driver={Oracle in OraDb11g_home1};Server=localhost;Uid=system;Pwd=manager
For Linux I uses Linux Oracle OS for Database server. I installed oracle client in my fedora OS. and cofigure the client using net cofiguration assistant tool with oracle client. client connection working OK. Then I try to connect with my mono application using the following connection string
First of all, I'm not a DBA, and when I try to import the dmp file generated on AIX with Oracle 10G on a RHEL6 11G machine I got a lot of issues, how to do this?
i am trying to export table using datapump in oracle 10g, this expdp takes 5 hours time, so i want use use parallel keyword in expdp, my question is how should i know number of parallels can i use...?
I have a rather hard to trace problem which is keeping me up several days now. We want to implement a package for querying devices via SNMP in our database. There are no packages shipped with Oracle that come handy. UTL_TCP does not work because SNMP works via UDP. After all we came up with the idea to use URL.... Locally on my development machine and on the command line of the Oracle database our first prototype of the SNMP query function works very well, but inserted into the database all I'm getting is a "Request timed out". Firewall mechanisms can not be the problem because the java code runs smoothly on the command line and also a snmpwalk on the command line suceeds, so the database is not blocked by any firewall. First I thought some permission via Java are missing, but I took a look in the USER_JAVA_POLICY view and worked this with SYSDBA out. A java application in Oracle connection via TCP to port 80 and fetching a website works allright, the thing not working seems to be UDP.
After upgrading 11gR1 database (11.1.0.7.0) to 11gR2 (11.2.0.3.0), the datapump exports have been taking quite a bit longer. When database was 11gR1, a full expdp took approx. 40-45 minutes. After upgrade, it takes approx. 1 hour 40-50 minutes. These times were with parallel=4. I tried with parallel=8 and parallel=12, both of these took around 1 hour 5-10 minutes, better but still quite a bit slower than pre-11gR2 upgrade. I tried with exclude=statistics, index_statistics, indexes; it still took approx. 1 hour 40-45 minutes. This is a PeopleSoft database so there are many, many objects to be exported. The database was upgraded using dbua.