Data Guard :: How To Disable REAL TIME QUERY In DG 11gr2
Dec 11, 2012
With 11gr2, by default, on STARTUP, the standby database is open
-> in READ ONLY mode
-> with Intended State: APPLY-ON
so the ACTIVE DATAGUARD option is in use ....
is there a way to deactivate REAL TIME QUERY permanently, so whether on STARTUP or STARTUP MOUNT, the standby stay only mounted with Intended State: APPLY-ON
The only way i found is to do the following :
SQL> startup
DGMGRL> edit database 'PHNXENT' set state='APPLY-OFF';
then
SQL> startup mount
DGMGRL> edit database 'PHNXENT' set state='APPLY-ON';
I have set up a cross platform (Microsoft Windows IA (32-bit) -> Linux x86 64-bit) data guard and it worked fine.Then I did a switch over (which again worked) and found out the data is not getting replicated at all.. checked the data files available from the new primary database and found out they are in the windows format as below..
SQL> select name from v$datafile;
NAME -------------------------------------------------------------------------------- D:ORACLEAPPADMINISTRATORORADATAMFSSYSTEM01.DBF D:ORACLEAPPADMINISTRATORORADATAMFSSYSAUX01.DBF D:ORACLEAPPADMINISTRATORORADATAMFSUNDOTBS01.DBF D:ORACLEAPPADMINISTRATORORADATAMFSUSERS01.DBF D:ORACLEAPPADMINISTRATORORADATAMFSRMANRMAN_TS01.DBF
and physically they were created at '/home/app/oracle/product/11.2.0/db_1/dbs/' and as
We are planning to have new installation of 11GR2 and would like to create data guard for this system. I have one server in Houston and other in Austin TX. the OS will be Windows 2008 R2.
How can i set up data guard and what is the requirements?
We have oracle 9i database named A with 2 schema A1 and A2.User enters data at A1 schema and the incremental data moved to A2 schema after some verfications performed by some scheduled job on A1 schema.
we want to move this incremental data to B1 schema of another oracle 9i database B at the real time when data gets entered to A2 schema of A oracle 9i database. We have access to A2 schema of A database and B1 schema of B database.way to do it or best practice to do this activity or if there is any third party tool or any available oracle utility to perform it?
Note: A2 schema of A database and B1 schema of B database has one to one mapping.We want to avoid using trigger on A2 schema, how data gets populated from A1 to A2 schema .
I'd like to stop Recovery in specific time. For instance ; My db works Day-1 and Every day Recovery starts 20:00 and finish 05:00. If it is possible that when DB time reach to sync 03:00, I'd like to stop recovery.
I have a deadlock trace file to analyse and i used to be able to see the rowid as a 16 bit hex value in the trace file, which i could then query on to get the actual real world row name.I see in the 11GR2 deadlock trace the formatting is different and for the life of me i am unable to see the rowid. Has Oracle stopped reporting this now, or is their another way to get this value? The deadlock graph shows me
I am trying to apply 11.2.0.2.5 psu on a 2-node cluster running on RHEL 5.5 vms. Followed ORACLE-BASE examples when installed this laptop RAC. I am not using any ACFS and none of GI and DB homes are shared. But, on node 2, cluvfy THINKS database home is shared.
I've just patched a development oracle environment (RAC 11GR2) with temporary patch 8730312. I would like to do same thing on Production RAC. Is it necessary to test all applications on dev environment with this kind of patch before deploy it on Production ?
I have recently setup a 2 node Rac on oracle linux 5.4 with oracel 11gR2, the installation went smoothly and all the cluster resources are up and running however the data is not syncing across the nodes, when I create a table it shows up on the other node but when I insert rows into the tables they don't show up on the other node and when I restart the cluster the inserted rows are completely gone even from the node where I inserted them.
I want to know that what will happen when I will resize database file from 32G to 5G on primary server or add file to tablespace. Will standby server do the same automatically or I have to to something with standby server?
I have created 2 PCs (1 for Primary DB (IT_SERVER), 2 for Standby DB (IT_SERVER2) on Oracle VM Virtual Box in order to check my Data Guard configuration.
I have created a Shared Folder on 2nd PC (IT_SERVER2) for Oracle Standby DB. Now when I issue my query from 1st PC (Primary DB) to Create a Parameter file on 2nd PC Shared Oracle Folder, it returns following errors
SQL> CREATE PFILE='\IT_SERVER2E:ORACLEPFILESTLDB2.ORA' FROM SPFILE; CREATE PFILE='\IT_SERVER2E:ORACLEPFILESTLDB2.ORA' FROM SPFILE * ERROR at line 1: ORA-09210: sftopn: error opening file OSD-04002: unable to open file O/S-Error: (OS 67) The network name cannot be found.
Then I try this
SQL> CREATE PFILE=\IT_SERVER2'E:ORACLEPFILESTLDB2.ORA' FROM SPFILE; CREATE PFILE=\IT_SERVER2'E:ORACLEPFILESTLDB2.ORA' FROM SPFILE * ERROR at line 1: ORA-00911: invalid character
I want to know if there is any reflection on the standby database states when it's primary is down or not reachable? I want to use this information from standby node to determine if I can programmatically execute (dgmgrl) failover operation or not? This is considering that the broker configuration has no fast-start failover and no observer enabled/configured.
What is the need of Temp table in Oracle ? what is the advantage of Temp table over normal table ? Is temp table a log operation one. What is the scenario of using temp table in real time ?
I have a primary database orcl and logical standby database orcl_std.
Real time apply is enabled. I have standby redologs in both primary and standby sides and I`ve started recovery with below command:
ALTER DATABASE START LOGICAL STANDBY APPLY IMMEDIATE;
When I create a new table in primary database, I am unable to see it on standby database (Although real time apply is enabled) However, when I switch log in primary, I can see the new table in standby database.
My question is, why realtime apply is not working in my scenerio ? I was expecting to see the new table immediately in standby database once it is created in primary database. Why am I supposed to wait for log switch in real time apply ?
CREATE OR REPLACE TYPE sample_object IS OBJECT (id NUMBER ,name VARCHAR2(30)); / CREATE OR REPLACE TYPE sample_table IS TABLE OF sample_object;
I have read some docs ,but I didn't get any information where exactly we use this.provide one real time scenario with an example.How this is different from record.
We have physical data guard configured version (10.2.0.4). We are in need to upgrade primary & standby database to 11G R2. Can we perform rolling upgrade.
I've got a query running a select count (*) over a table. The default plan takes in the order of 15 minutes to return, a hinted plan to use a different index takes 3 minutes to return.
Unfortunately I cant get at the index stats and a few other areas which I suspect may be key here.When running autotrace against the two queries I see fairly different values as one would expect.
Query
select count (*) from fulfilmentitem bfi where created >= sysdate-30 AND bfi.status = 'FA' AND bfi.fulfilmentmethod = 'D' Slow run PLAN_TABLE_OUTPUT ----------------------------------------------------------------------------------------------- ---------------------------------------------------------------------------------------------- | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| ---------------------------------------------------------------------------------------------- | 0 | SELECT STATEMENT | | 1 | 15 | 33119 (1)| | 1 | SORT AGGREGATE | | 1 | 15 | | |* 2 | TABLE ACCESS BY INDEX ROWID| FULFILMENTITEM | 12525 | 183K| 33119 (1)| |* 3 | INDEX RANGE SCAN | IDX_FULFIL_METHODSTATUS | 250K| | 1786 (1)| ---------------------------------------------------------------------------------------------- [code]....
IDX_FULFIL_METHODSTATUS is across FULFILMENTMETHOD & STATUS in that order. IDX_BFI_CREATED is on CREATED and is approx 70% of the size of the other index
The row counts estimated in the explain plan are out, the count(*) comes in at 32.8k rows.As you will have seen, the fast run shows a pretty significant consistent get increase compared to the slow run and a decent though not dramatic physical read drop.
My uncertainty is around if these changes in consistent get/phys read values would typically be enough to suggest the real time improvements I'm observing or if other (albeit perhaps temporary) factors are involved. It is a prod OLTP environment so the data will be rapidly changing and that may be a factor.
I know it can never be an exact science without intimately knowing the hardware/current loads etc but I also know that there's enough experience on these boards to have a loose handle on if the time shifts between queries are likely (or not) to be reflective of the stat changes or if those differences alone shouldn't (or typically wouldn't account) for it.
I'm thinking about instructing the query to ignore its original plan but am hesitant to do so without being a little more confident that it's not just a timing thing or something other than the change of index approach which may be causing the improvement. the autotrace stat changes observed I couldn't put my hand on heart say "yup - that change is good, ignore the default index all the time for this job".
I am trying to restore 11gr2 database to point in time. using following steps:
Control files are fine, that are not restored.
RMAN> run { set until time "to_time('2012-12-13 12:12:00','YYYY-MM-DD HH24:MI:SS')"; } RMAN> restore database; RMAN> recover database; RMAN> alter database open;
Here my problem is it is restore complete data what ever i was having, but i don't was it, i has to restore up to specified time only.
I Have three field and first field for START TIME ,Second END TIME & Third DURATION AND Putting START TIME AND END TIME i am getting duration in minutes by using code
I want to delete data from say 100 tables without disable the constraints.
Tables having foreign key.. which is giving error if I select table randomly to delete.
I want to know the query which will give me the sequence of all tables based upon dependencies, so that If i delete the data on that sequence, it wont give me child record exists error.