We have 2 Sun Solaris (Solaris 10) nodes each has an oracle db, and OID of the above versions.There is an ASR running between both nodes, and it has been setup as follow:
$ ./ldap/bin/remtool -asrsetup -v
------------------------------------------------------------------------------
ASR Setup for OID Replication
Make sure that the replication administrator that you enter below does not exist already in any of the nodes that will be part of the DRG to be created now. If the user exists, that user will be dropped and will be created newly.
------------------------------------------------------------------------------
Enter replication administrator's name : repadmin
Enter replication administrator's password :
Reenter replication administrator's password :
Enter Master Definition Site (MDS) details :
Enter global name of MDS : node1
[code].....
We ran a perl script which is using ldapmodify to change the value of one attribute in all records (7 Millions records).That has been done by creating many of LDIF files each of 10,000 records in 26.04, and by mistake this script has ran 3 times in parallel.Currently the replication between both nodes are stuck, and when I query the ods_chg_log on the master node (node1):
SQL> select count(*) from ods_chg_log;
COUNT(*)
----------
6690594
And the number is not decreasing.
And I queried on the asr_chg_log on the master node (node1):
SQL> select count(*) from asr_chg_log;
[code]....
i need to set up a central server with all the master tables and two other local database which will hold the updatable materialized view of the master table...the databases must be synchronized with central server..and user will work on the materialized view database...
I have a multimaster advance replication environment and we have less than 10 transaction per day...I want to propagate data as soon as possible like sync if I setup the schedule push to propagate transaction every 5 seconds as below
in other hand ,book(I mean Advance Replication) has written that for simulating continuous push we should setup it as below so it will propagate transaction every 1 minute.
My second question is:I know interval >= 'SYSDATE + (1/144)' means every 10 minutes a job will start and delay_seconds => 1200 means each job remain aware for 20 minutes, but I can't understand the logic?why it can simulate 1 minute propagation?
[oracle@server03 bin]$ ./runInstaller -updateNodeList ORACLE_HOME=$ORACLE_HOME CLUSTER_NODES="" -local Starting Oracle Universal Installer... Checking swap space: must be greater than 500 MB. Actual 20127 MB Passed The inventory pointer is located at /etc/oraInst.loc The inventory is located at /u01/app/oraInventory 'UpdateNodeList' was successful.then I executed:
To deinstall the Oracle home from the node you are deleting, run the following command from the Oracle_home/oui/bin directory:
Checking swap space: must be greater than 500 MB. Actual 20127 MB Passed Preparing to launch Oracle Universal Installer from /tmp/OraInstall2012-11-06_12-05-36PM. Please wait ...[oracle@server03 bin]$ Oracle Universal Installer, Version 11.1.0.7.0 Production Copyright (C) 1999, 2008, Oracle. All rights reserved.
[code]....
When I arrive to the point 5 I see one resource ora.DPCTTest.db ONLINE on one server (server02 or server03 that are the nodes I want to remove). If I execute:
crs_relocate ora.db_name.db
I obtain: [oracle@server03 bin]$ ./crs_relocate ora.DPCTTest.db Attempting to stop `ora.DPCTTest.db` on member `server02` Stop of `ora.DPCTTest.db` on member `server02` succeeded. Attempting to start `ora.DPCTTest.db` on member `server03`
I've configured a replication environment with Oracle 11g (Multi-master synchronous replication). Apparently it works fine, when replication is started any change in a table that belongs in a node is replicated in the rest of the nodes.
One of the tests I'm making is the following: Try to see what happens if one of the nodes drops. Meanwhile one of the nodes is no avalaible, any modification in an available node is not allowed (as I expected). I'm trying to see if the database service can be available meanwhile a node is missing, although the replications is not running.
What I've experienced is replication cannot be suspended until the the node was down is restarted.
Is this the usual behaviour or can I set any parameters to avoid this? In other words, is it possible the database service is running meanwhile one of the nodes is down? The application I'm working on needs this type of replication, buy also needs the service is running always?
What can I do to make possible the service is running even when a node falls?
Actually am trying to replicate two db servers from one in hong kong and another in china. when am trying to establish the replication, am getting error 'ORA-04052: error occurred when looking up remote object' like this...
but the same way i have tried in my local network, it is working fine.i have tried schema replication through enterprise manager grid control..
I want to set up advance replication for 3 master site (multimaster) I created 3 master site named orc1,orc2,orc3 and followed up oracle replication management of API book instruction I created 2 tables(tes1,test2) in hr schema in all 3 master site with the same data. then I created the following steps
1-CONNECT repadmin/repadmin@orc1
2-Create the master group named hr_test_repg
BEGIN DBMS_REPCAT.CREATE_MASTER_REPGROUP( gname => 'hr_test_repg'); END; /
4-add tables test1 and test2 to the group
BEGIN DBMS_REPCAT.CREATE_MASTER_REPOBJECT( gname => 'hr_test_repg', type => 'TABLE', oname => 'test1',
[code]....
I could create DBMS_REPCAT.GENERATE_REPLICATION_SUPPORT for test2 but not for test1 and it produces error
RROR at line 1: RA-23309: object hr.test1 of type TABLE exists RA-06512: at "SYS.DBMS_SYS_ERROR", line 105 RA-06512: at "SYS.DBMS_REPCAT_MAS", line 2552 RA-06512: at "SYS.DBMS_REPCAT", line 562 RA-06512: at line 2
Initial situation: Oracle 10g database single instance on Windows 2008 (datafile within NTFS partition) Final situation: same database upgraded to Oracle 11g r2 on a two nodes RAC on Windows 2008 (datafile within ASM)
In your opinion, is this the best way to achieve the job ?
1) on the two nodes install grid 11gr2, asm, rdbms 11g r2 (SE) 2) export full from 10g 3) create a new empty 11g r2 database in the RAC 11g r2 infrastructure with the same tablespace layout of 10g database 4) import full from 10g to 11g r2
I'm trying to install 2 nodes oracle RAC 11gR2 on SLES11. I configured DNS for public,virtual, private and scan IPs. when I check with nslookup, everything seems ok. but when I run runcluvfy.sh, it says that:
" PRVF-5636 : The DNS response time for an unreachable node exceeded 15000 ms on following nodes: rac1,rac2"
i'm trying to duplicate my RAC database with ASM and 2 nodes to a single node, non ASM with rman duplicate database for standby with active database on windows. The problem is i dont know how to remove RAC configuration when duplicating the database.
Here is the procedure:
On standby i've: - installed database software, - configured listener and tnsnames, - added instance service (oradim -new -sid STANDBY -intpwd PASSWORD -startmode M), - configured initora (with only db_name = STANDBY) - configured orapwd file - startup nomount
on RAC node1:
rman target sys/PASSWORD@RAC connected to target database RMAN> connect auxiliary sys/password@STANDBY connected to auxiliary database: STANDBY (not mounted)
now for the duplicate command:
RMAN> duplicate target database 2> for standby 3> from active database 4> spfile 5> parameter_value_convert = '+DATA/RAC','D:\oracle\11.2.0\DATAFILES\DATA\STANDBY' 6> set db_unique_name='STANDBY' 7> set log_file_name_convert = '+DATA/RAC','D:\oracle\11.2.0\DATAFILES\DATA\STANDBY' 8> set db_file_name_convert = '+DATA/RAC','D:\oracle\11.2.0\DATAFILES\DATA\STANDBY';
i have tried adding this parameter - set cluster_database='false' but no luck.
I have one 10g database in other country. I want part of their db (selected tables or tablespaces) and import that data to my 10g DB and i want keep to date this data.
I know two ways
1. Data Pump Imp/Emp via FTP, but i can't send only data that have changed (incremental), i must pumping whole selected part of database (i want only new data from their DB, but consistanse with my DB)
2. RMAN etc. or other archivisation tool, i can do incremental achivisation, but can i send files to another instances (my db) and load only that data? Can i do that with SQL*Loader?
For using replication in our production, here i am testing golden gate as replication tool. I tested all scenario in Uni direction ( source to destination). Now have to test replication with DDL support in Bi-Direction. Not getting any Doc for doing replication in Bi-direction( Two Way). If any one has done the same, then please share limitation of replication in Bi-direction through Golden Gate.
I Have configured the replication between two database's at table level. After few miniuts of successful configuration of replication between two db's at table level, I am getting the ORA-00001: unique constraint (%s_PK) violated error in dba_apply_error.
I checked constraint name,type and status on table replicated is same on both source and destination db.
We are trying to increase ASM disk space and with respect to it when we are trying to allocate more space this question came across my mind. Now this was previously configured by my previous SA.
[root@oracledbtest1 ~]# /etc/init.d/oracleasm querydisk -d `/etc/init.d/oracl cut -f2,10,11 -d" " | perl -pe 's/"(.*)".*[(.*), *(.*)]/$1 $2 $3/g;' | while read v_asmdisk v_minor v_major do v_device=`ls -la /dev | grep " $v_minor, *$v_major " | awk '{print $10}'` echo "ASM disk $v_asmdisk based on /dev/$v_device [$v_minor, $v_major]"
[code]....
why are my LUN's showing different sizes on RAC and also the best way to allocate the space to the disks in the above scenario.
I'm using ASM on LUNs from an EMC SAN, fronted by PowerPath. Right now I have only one fiber path to the SAN, so /dev/emcpowera3 maps directly to /dev/sda3, for example. Oracle had a typo in what they told me to do in /etc/sysconfig/oracleasm*, so the scan picks up both devices.
#/etc/init.d/oracleasm querydisk -p ASMVOL_01
Disk "ASMVOL_01" is a valid ASM disk /dev/emcpowera3: LABEL="ASMVOL_01" TYPE="oracleasm" /dev/sda3: LABEL="ASMVOL_01" TYPE="oracleasm"
But I don't think it can be using both. How do I see which one it's actually using?
i am have a problem with Oracle user equivalence. long time ago, we have created Two node RAC. that time, we configured Oracle user equivalence.
today when i am trying to install Oracle Agents on this RAC pair, installation was failed due to Oracle user equivalence. not sure how it happened, and i was confused too.
if i configure Oracle user equivalence again, will it cause any problem to my RAC setup?
is there any way to find, when was the Oracle user equivalence was changed/ ssh keys got changed?
I read that Oracle RAC which is bundled in Standard Edition can support at max 4 sockets. One of my client has a proposal of using RAC but for 3 nodes each using single 1 quad core Intel processor. As far as i understood, an Intel quad core is a multi chip module and actually is a combination of 2 modules dual core, so each Intel quad core may be counted as 2 sockets. Which yield to the proposal of my client will be failed, as the total number of sockets in it will be: 2*3=6 that exceed the max 4 support.
i am on RAC 11G R2 RHEL 5 , when i install rac can i run root.sh on all nodes in parallel ? or do i have to wait for first node to finish root.sh completely and then i shud run root.sh on second node and so on on third node
I want to apply patch for the 3 node rac database. I am going to use rolling upgrade, one node at a time. For this i need to stop all the instances or any processes running out of $ORACLE_HOME. I will use srvctl stop home command. My concern is:
1) with the srvctl stop home command how the instances shutdowns, either shutdown immediate or shutdown abort. 2) if applications are connected to the shutdown node, will the sessions fail over to surviving nodes? 3) If it is shutdown abort, how session failover?how about long running queries ? will they also fail-over?
Oracle version 11.2.0.2 on RHEL 5.6 , planning to apply Db PSU 11.2.0.2.7
This shows user=1 has permission on Account=4 , user=2 does not have access priviledge on Account_id=5.Now my requirement is to write a SQL query such that If user has permissions on a given leaf node then all the parent / grand parent nodes should also be returned. Eg: SELECT * FROM (.......) where USER_ID=2;
To shutdown the entire crs stack, I ran crsctl stop cluster -all from one node. After the command execution, the below mentioned processes were still running on all nodes.
So , I had to manually run crsctl stop crs on all nodes.
[root@intsmdp01 ~]# /crs/product/11.2.0/bin/crsctl stop crsCRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'intsmdp01'CRS-2673: Attempting to stop 'ora.crf' on 'intsmdp01'CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'intsmdp01'CRS-2673: Attempting to stop 'ora.mdnsd' on 'intsmdp01'CRS-2677: Stop of 'ora.crf' on 'intsmdp01' succeededCRS-2673: Attempting to stop 'ora.gipcd' on 'intsmdp01'CRS-2677: Stop of 'ora.mdnsd' on 'intsmdp01' succeededCRS-2677: Stop of 'ora.drivers.acfs' on 'intsmdp01' succeededCRS-2677: Stop of 'ora.gipcd' on 'intsmdp01' succeededCRS-2673: Attempting to stop 'ora.gpnpd' on 'intsmdp01'CRS-2677: Stop of 'ora.gpnpd' on 'intsmdp01' succeededCRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'intsmdp01' has completedCRS-4133: Oracle High Availability Services has been stopped.[root@intsmdp01 ~]#
Is there a way to bring down the entire crs stack in all the nodes in the cluster from one node ?