Streams :: Apply Archivelog / Generated Before Reset-log Of Source Database In Destination Database
Feb 9, 2013
Database Version - 10.2.0.4.0
OS - Red Hat Linux
We have configured oracle one way stream between two databases. Source database is capturing the changes (No downstream configured). Configuration was working fine but destination database was lagging behind very much i.e about 15 days behind the source database. We are ok with this but the problem is now that , as per client request we have restored previous backup and open the database with resetlog option in source database. After resetlog , archivelog sequence has been changed and stream is not working.
Can I apply the previous archivelog (before resetlog archivelogs ) in destination database anyway.Source database is a production database.
View 1 Replies
ADVERTISEMENT
Jul 20, 2009
We are using Oracle Streams for replication.
Column datatypes in some of the tables on Source and Destination are different (Number on Source and Varchar2 on destination).
Do we have to create any rule or dml handler to handle this or Streams will automatically take care of it?We are oracle 10g.
View 1 Replies
View Related
Jan 11, 2011
I want to replicate single table data/modification from source db to destination DB.procedure or steps how to create replication between source and dest db for single table.
View 18 Replies
View Related
Jul 15, 2013
I created a standby database yesterday and everything is working correctly. I need to make some tweaks, however, and one of those is the directory that the standby database expects the archivelogs to be.
I found that the archivelogs were being shipped to $ORACLE_HOME/dbs and named arch*.arc. Interestingly a log switch ships the archivelog minus the "arch" at the front and all archivelogs now do not have that format. I moved the archivelogs to the correct location and renamed them. I have been able to set the correct location using standby_archive_dest and can see that the archivelogs are shipped to there.
Also, v$archived_log shows the correct path and filename. The problem I have is that when I come to apply the archivelogs it still seems to think that they should be in $ORACLE_HOME/dbs and named with "arch" at the front. what parameter I need to change to tell oracle the correct path and filename to use when applying the archivelogs?
View 9 Replies
View Related
Nov 1, 2013
A streams apply process which applies to a sql sever database is increasing its pga use continually until i stop the process and restart it. I need to stop it once every week or it will use too much of the pga and the database will hand causin paging etc.
View 1 Replies
View Related
Dec 16, 2012
RMAN-00571: ===========================================================
RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
RMAN-00571: ===========================================================
RMAN-03009: failure of backup command on t1 channel at 12/15/2012 06:15:11
RMAN-10015: error compiling PL/SQL program
RMAN-10033: error during compilation of job step 2: ORA-00603: ORACLE server session terminated by fatal error
ORA-04030: out of process memory when trying to allocate 400 bytes (PL/SQL DIANA,LISIMA: Add cells)
We received the above error message after the following event:
rman failed to backup because archivelog destination was out of space
We moved SOME of the archivelog file manually using operating system command mv to another location so that our application can run first.When the engineers restarted the rman to backup, we receive the above error.
View 7 Replies
View Related
Mar 27, 2013
My setup is following:
CODEhost22 (Oracle Linux 5.9) | host30 (Windows XP) with SQL Server Express 2008
DB22 | Gateway to MSSQL | DB30
Local Local | |
Capture --> Apply ------> Gateway -----------------> DUMMYUSR.DUMMYTBL
Process Process | |
Capture process on DB22 captures changes made to local DUMMYUSR.DUMMYTBL table, enqueues it to capture queue and propagates to apply queue for local apply process, which, using database link DB30 supposed to apply these changes to table DUMMYUSR.DUMMYTBL on remote SQL Server database DB30.
The problem is that when apply process is executing transaction, it aborts with error:
QUOTE ORA-26713: remote object does not exists or is inaccessible.
ORA-26712: remote object is "DUMMYUSR"."DUMMYTBL"@"DB30".
In trace log I can see the following message:
QUOTE kniNoPrimaryKeyError: ORA-26713, remote obj doesn't exist,
DUMMYUSR.DUMMYTBL@DB30
[code]...
what can cause ORA-26713 error of apply process applying changes to SQL Server database through Oracle Database Gateway, if object to be modifyed exists and accessible from Oracle database for strmadmin user?
By the way, database link and tables were created with following commands: CODEDB22(Oracle): (connected as strmadmin) create database link DB30 connect to "dummyusr" identified by "dummypwd" using 'DB30';
DB22(Oracle): (connected as dummyusr)
create table DUMMYTBL (no number, name char(6), ddate date, payload varchar2(2013));
DB30(MSSQL): (connected as dummyusr)
create schema DUMMYUSR authorization dummyusr;
create table DUMMYUSR.DUMMYTBL (no float(53), name char(6), ddate datetime2(0), payload varchar(2013));
View 1 Replies
View Related
Oct 10, 2013
,I am trying to set up Streams on a 11.2.0.3 on a Windows 2008R2 server. Due to an error in running propagation, i a, trying to delete both Capture and Apply queues. I have deleted the Apply queues but unable to drop the Catpure,
SQL>EXEC DBMS_APPLY_ADM.STOP_APPLY(apply_name =>'LAO_NLPG73_BLPU_APPLY'); SQL> select * from dba_apply; LAO_NLPG73_BLPU_APPLY NLPG73_BLPU_APPLY_Q STRMADMIN YES RULESET$_732 STRMADMIN STRMADMIN 301355 ABORTED 09/10/2013 17:34:21 1013 ORA-01013: user requested cancel of current operation CAPTURED STREAMS APPLY SQL> select * from dba_queues; STRMADMIN NLPG73_BLPU_APPLY_Q
[code]...
View 5 Replies
View Related
Apr 22, 2011
I got a primary database with a logical standby database running Oracle 11g. I got two client applications, one is the production site pointing to the primary one, another one is just a backup site pointing to the logical one.Things will only be written into the primary database every mid night and client applications can only query the database but not add, update nor delete.And now, I want to apply the latest patch on both of my databases. I am also the DNS administrator, I can make the name server pointing to the backup site instead of the production one.I want to firstly apply the patch on the logical one, and then the physical one.
I found some reference which explains how to apply patches by adopting "Rolling Upgrade Method". however, I want to avoid doing any "switch over" mentioned in the reference because I can make use of name server. Can I just apply patches as the following way?
1)Stop SQL apply
2)Apply patches on logical standby database
3)let the name server point to the backup site
4)Apply patches on the primary database
5)Start SQL apply
6)Let the name server point back to the production site
View 2 Replies
View Related
Apr 29, 2013
i have 2 tables in two different sources.
I have loaded data from source to destination.
some rows are missed while loading.
i want to know the missing rows
View 11 Replies
View Related
Oct 31, 2011
how do i copy an image or a text file from ex: directory c to directory d
View 5 Replies
View Related
Jul 10, 2010
when I am trying to send report for prints then I received this error.
"
REP-57054: In-Process Job terminated: The report generated successfully but distribution to destination failed.
REP-50159:The report generated Successfully but distribution to destination failed.
REP-50151: Destination ID Currency failed with error An error occured while sending the file to the printer :HPofficejetpro7700Series- Cannot run program "COrcaleMiddlewareasinst_5Config
eportsin
wlpr":
create process error=2, The system cannot fine the file specified."
View 6 Replies
View Related
Jun 10, 2013
I turned on ARCHIVELOG in my main database, anticipating GoldenGate replications and standing up a physical standby database. However, during the process I forgot to create a new SPFILE for my database, meaning that the next time the database cycled, the archive log destination would revert to the Oracle default (rather than the mount point I had set up to receive them) $ORACLE_BASE/fast_recovery_area
. I discovered this because my database hung..... In looking at the ALERT LOG, I discovered that the logs were going to fast_recovery_area, and it had run out of room. OK, so I did an alter system set db_recovery_file_dest_size= 5G scope=both; at which point in time I was notified that no active SPFILE existed! I moved out all of the logfiles that were in the fast_recovery_area, but the database seemed to still think that it was full.
I've managed to get everything fixed and put back together, but my remaining question is: After I freed up all sorts of space in the fast_recovery_area, why did/does the database still think that area is full?
View 3 Replies
View Related
Jun 2, 2010
this Topic was already discussed in "Physical standby on oracle 11g - Reporting needs" but without finishing.
View 5 Replies
View Related
May 27, 2011
We have three unix servers with four databases (10gR2) containing "HP Operation Management Unix" (OMU) server messages for monitoring purpose, and we now want to transfer these data to one new database on a new server for reporting purpose.
The message table in each OMU database keeps the message row until it is "Acknowledged" or for maximum fourteen days, then it is moved to an historic table where it stays for another three days. Keeping data for only seventeen days are a performance issue.The new "Reporting database" is intended to hold messages data for the last 90 days.
I wonder which method to use to move/replicate data against the databases? Materialized view using database link, with view on top of the MVs. How to keep rows longer than the master (source) table, avoiding deletion when master row is deleted
Oracle Streams, with local capture and remote apply. How will this influence on the master database performance. There are about 10000 new messages in each OMU database every day. Is it possible having four streams connections against the reporting database ?
Or should I simply use database triggers which fires after insert and update and applies changes to the reporting database using database links ?
View 2 Replies
View Related
Apr 2, 2013
i have one question regarding archivelog and backup if i take just backup excluding archivelogs like = RMAN>backup database; (not plus archivelog ) now when even i recover database using backup , will it also recover changes that was in archives?
View 7 Replies
View Related
Aug 27, 2007
I have an Oracle 9i developer database on my Windows workstation. I havent used it in a couple weeks and forgot my Sysdba login password. how I can reset it or get into the database?
View 6 Replies
View Related
Sep 27, 2011
I wanted to copy some archive log files between 2 DB Server.Steps:
1) Edited the Local and Remote DB's TNSNAMES file and modifyied like below.
bookstore =
(DESCRIPTION =
(ADDRESS_LIST =
(ADDRESS = (PROTOCOL = TCP)(HOST = 10.80.100.28)(PORT = 1521))
)
(CONNECT_DATA =
(SID = BOOKSTORE)
)
)
2) Created a Source Directory --> DR1
3) Created a Target Directory --> DR2
4) Created a DB Link
CREATE DATABASE LINK LINK1 CONNECT TO <username of the Remote DB> IDENTIFIED BY <Password> USING '<Remote DB Name taken in the TNSFILE>.
5) In the Local Server I had written the below command.
create or replace procedure proc1 is
cursor c1 is select recid,substr(name,37) "ARCH_FILE" from v$archived_log;
var1 c1%rowtype;
begin
open c1;
fetch c1 into var1;
[code]...
It is working, but not coping the Files from Local To the Remote.
View 22 Replies
View Related
Jan 2, 2013
I'm trying to write short program to reset password on databases. I'm trying to use OCINewPassword property with following
public void ResetPassword(String sOldPass, String sNewPass) {
this.sCurrent = sOldPass;
this.sNewPassword = sNewPass;
Properties props;
Connection conn = null;
[code]....
View 4 Replies
View Related
Jun 4, 2013
Just now my database is hang due to the archivelog full. I try to use RMAN to delete the expired one and free space but encounter the error:
RMAN>crosscheck archivelog all;
archive log filename=/geodata/oraclelog/archive_log/ORAGF45/archivelog/2013_06_04/o1_mf_1_1444_8tsxvwc2_.arc recid=1350 stamp=817182450
validation succeeded for archived log
archive log filename=/geodata/oraclelog/archive_log/ORAGF45/archivelog/2013_06_04/o1_mf_1_1445_8ttx2q6j_.arc recid=1351 stamp=817214416
[code].......
So I try to back it up first by using the below command:
RMAN> backup archivelog all;
Starting backup at 04-JUN-13
using channel ORA_DISK_1
channel ORA_DISK_1: starting archive log backupset
channel ORA_DISK_1: specifying archive log(s) in backup set
[code]......
I know that I have to extend the db_recovery_file_dest directory for "limit exceed" error but how can I safely delete the archivelog in order to bring up the database?
View 3 Replies
View Related
Mar 3, 2012
I need to migrate a 500 GB 10.2 database from windows to linux. Of course I found Alejandro Vargas paper on how to do that ("Database Migration From Windows to Linux Using RMAN"). My problem is that the source database will be up again for 7 days and I need to apply all changes made to the source database to the migrated linux database.
Can I just take all redo log files generated from the source database and apply those to the destination database? I might have a problem with a) file conversion and that the redo log are not in sync with the target database (different check points?).I just found that on an exotic website: "The contents of a redo log file depend on a combination of Oracle version, operating system and server architecture. In general redo logs written on one architecture cannot be read on another. There are a few exceptions to this rule. For example, in Oracle 10.2 a redo log written in Linux can be read by a Windows database."
So if this is true, does it also work the other way AND will the database migration from window to linux create a perfect clone of the database where those redo log files actually can be applied to?
View 1 Replies
View Related
Jul 20, 2012
We need to apply the the 11.2.0.2 database server patch and this is a 12i Ebusiness suite env. Do we need to take any precautions(other than stopping the apps) on the APPS side when we are applying the patch on the database?
View 2 Replies
View Related
Jan 24, 2011
We have faced database(10.2.0) issue cause incomplete recovery and have performed open resetlogs. This DB is of 12Tb. In 10g opening database with reset logs do not invalidate previous backups. we have another replica(no reset state)of this database which we sync using archives.how we can apply those archives(reset database)to previous database?
View 7 Replies
View Related
Mar 25, 2010
can we apply 11.1.0.7 Patch Set on a 10.2.0.3 database server? i mean are patch sets cumulative or do i have to install the patch specific to my release the issue is i have the below error with rman: RMAN-03009: failure of delete command on ORA_DISK_1 channel at ORA-19633: control file record is out of sync with recovery catalog
View 7 Replies
View Related
Jul 22, 2013
I have a Sap Primry Database and also a standby db that was working perfectly.. We migrated the primary db from windows 2003 to windows 2008 and brought the primary db up.. I had to create a controlfile and do a system copy and had to reset the logs on the Primary.. All came up and when i checked the standby it was receiving the logs but after a month. i see that it was not applying the logs as I think because of the sequence number .. it stopped.
I did the ffg as per the attachement...My logs have been shipped across but not applied, But What worrries me is the log sequence number on my Primary
SQL> select max(sequence#),thread# from gv$archived_log group by thread#;
MAX(SEQUENCE#) THREAD#
-------------- ----------
27727 1
View 15 Replies
View Related
Aug 20, 2008
How to make the automatic deletion of all the archivelogs when it is applied successfully on the standby database? We have 1 primary database and 1 standby database using oracle physical dataguard.
In the primary database, we have daily backup job so that the applied archivelogs are deleted automatically, no problem
But in the standby database ( physical standby) I have to go there and delete manually because I don't issue backup on this database. Is there any solution for this problem.
View 3 Replies
View Related
Nov 23, 2010
1) The Primary Database is UP. The Physical Stand By Database is DOWN. The Current Archive Log Sequence is 99 in Primary.
We have to apply Archive Log from 51 to 99 to the Standby Database. But Unfortunately, there is no backup of those Archivelogs and the ArchiveLogs from 51 to 98 have got deleted at Primary end.
Now how will you apply these Archive Logs from Primary Database to Physical Standby Database?
Note : The Physical StandBy Database is DOWN.
View 2 Replies
View Related
Nov 27, 2012
I want to apply patch 13 to 11.2.0.3 database for windows server 2008 R2. I saw the pre requisites and found that opatch should be 11.2.0.1.9 or later. My optach is 11.2.0.1.7. I downloaded the patch # 11846294 and unzipped under the opatch directory. In one of the step it asked me to run the following command
$ORACLE_HOME/OPatch/oplan/oplan generateApplySteps <bundle patch location>
but then it says oplan is not recognized. I have set the Oracle_home and path variable
my path is C:Oraclegrid_homein;C:Oracledbhome_1in;C:Oraclegridin;%SystemRoot%system32;%SystemRoot%;%SystemRoot%System32Wbem;%SYSTEMROOT%System32WindowsPowerShellv1.0;C:Program FilesArcGISArcSDEora11gexein;C:Program FilesXIVhost_attachin;C:Oracledbhome_1oplan
Note that my database version is 11.2.0.3 with ASM
View 7 Replies
View Related
Sep 13, 2011
I have to apply monitoring template on my databases. Databases are already added and configured into OEM. They are visible under Databases tab and also under Groups where I have added them.
But when I try applying monitoring template and look for database in search window, it doesn't appear there.
View 1 Replies
View Related
Jun 28, 2012
does the source database have to be open in read-only? or can it be online at the time of exporting?
*do not take mean-while running transactions into consideration, I will do it at night when nobody's working.
View 10 Replies
View Related