RAC & Failsafe :: Migrating Single Node Database To New 2-node Environment
Jan 24, 2011
We have oracle database 10g(10.2.0.4) running on AIX 5.1. we have installed 2-node RAC on another server AIX 6.1 . Now we want to migrate our single node database to new 2-node rac environment. how we can move/migrate/cluster-aware our single node database to 2-node rac?
I would like get more details about single node RAC.. What is single Node RAC?..
1). just configure only one server instead of two node... same RAC installation.. cluster installation and shared disk configuration in single server itself.. in future we can add node to this ... 2). in 11gR2 having new feature Single Node RAC.. Multiple instances running on same node and configure using Cluster...
is there any advantage using single node rac(other than online upgrade to full RAC).
srvctl config PRCR-1119 : Failed to look up CRS resources of database type PRCR-1115 : Failed to find entities of type resource that match filters (TYPE == ora.database.type) and contain attributes DB_UNIQUE_NAME,ORACLE_HOME,VERSION Cannot communicate with crsd
I have implemented Rac One node with two machine(dbtest01,dbtest02) in 11.2.0.3 Redhat 5.3 one for online another one for offline. i have created TAF also.
dbtest01 is online and create one query DECLARE i number(36):=1; begin while(i<10000000) loop dbms_output.put_line(i); i:=i+1; end loop; end;
i executed above query for test and i made relocate database using srvctl utility to dbtest02 machine.
session is relocated successfully with new session id...above query getting stop in 25000 itself. i want to make it continue that process..also i tried to find Omotion utility in all the path.. couldn't find it... where will it be.. i want to relocate all the session without losing work.
from application i am getting disconnect after relocation.. application server need to restart to get work again...
here we have san server and rac database two node database.i want to know how can i create directory on rac database. i know how to create directory on single database but i do not know how can i create directory on rac database.
we encounter a problem with a 2-nodes RAC with 11gR2 SE : Database RAC is correctly running but we encounter problems using RAC TNS entry, obtaining the message ORA-01033.We can access to the DB using a TNS entry using each one of IP addresses but when we use the SCAN address, it fails. 2 weeks ago RAC2 node could connect perfectly with the SCAN address but RAC1 (used as a client) obtained error message almost 30% of tries.
We made a reboot of RAC1 node and now SCAN address works correctly from RAC1 and RAC2 (used as a client) fails systematically with SCAN address but works perfectly with direct access to the database.On both nodes, SCAN address is perfectly resolved by DNS server, cluster seems correct, listeners seems correct...We have opened an SR on Metalink 4 weeks ago but we don't progress.
We have to configure Dataguard for our 24x7 critical banking 2 Node RAC database(10.2.0.4). Before proceeding with the configuration we have to make sure what steps we should follow to have minimum or no downtime.
1) Document covering DG setup in RAC environment. 2) We have to perform switchover as well so need its steps too. Either the normal switchover steps would be used or have to stop/start rac services as well.
i'm trying to duplicate my RAC database with ASM and 2 nodes to a single node, non ASM with rman duplicate database for standby with active database on windows. The problem is i dont know how to remove RAC configuration when duplicating the database.
Here is the procedure:
On standby i've: - installed database software, - configured listener and tnsnames, - added instance service (oradim -new -sid STANDBY -intpwd PASSWORD -startmode M), - configured initora (with only db_name = STANDBY) - configured orapwd file - startup nomount
on RAC node1:
rman target sys/PASSWORD@RAC connected to target database RMAN> connect auxiliary sys/password@STANDBY connected to auxiliary database: STANDBY (not mounted)
now for the duplicate command:
RMAN> duplicate target database 2> for standby 3> from active database 4> spfile 5> parameter_value_convert = '+DATA/RAC','D:\oracle\11.2.0\DATAFILES\DATA\STANDBY' 6> set db_unique_name='STANDBY' 7> set log_file_name_convert = '+DATA/RAC','D:\oracle\11.2.0\DATAFILES\DATA\STANDBY' 8> set db_file_name_convert = '+DATA/RAC','D:\oracle\11.2.0\DATAFILES\DATA\STANDBY';
i have tried adding this parameter - set cluster_database='false' but no luck.
I tried to clone a 2 node rac database to single instance non rac database using existing backup. I have not used connectivity to target or catalog. rman duplicate finished with below messages:
rman auxiliary sys/******@dbracdup RMAN> duplicate database to dbrac spfile backup location '/oracle/backup'; ... ... Finished recover at 25-JUL-12 Segmentation fault
And the database was in mount stage, and when i tried to open database it failed with below error:
SQL> alter database open; alter database open * ERROR at line 1: ORA-19838: Cannot use this control file to open database .
We have 2 node cluster with 10.2.0.4 home and 11.2.0.2 home under 11gr2 grid infrastructure with ASM in LINUX environment. We are planning to add 3rd node to this cluster. Is there any issues to add 3rd node to existing environment or running fine.
Insert statements from Java are processing very slow i.e 1L records taking 15Mins on every row commit ( Batch insert is not applicable that's why they are committing on every row). - Only 1 index
I have 2 node 11gR1 running on Unix, I want to add 3 more RAC node to it. What are all the steps involved and should I first add 3rd node and then the 4th node or I should add both at the same time? is there any difference between these two methods?
Oracle VM 2 Node RAC cluster Oracle 11G 11.2.0.3.0 Oracle 11g rel 2 GI
on current 2 node cluster we have GI and RAC db configured Nodes: vmorarac1,vmorarac2
we shutdown vmorarac1 to clone it to vmorarac5
on new node I have changed the hostname to vmorarac5 In /etc/sysconfig/network-scripts/ifcfg-eth0 change ip to new ip and same for /etc/sysconfig/network-scripts/ifcfg-eth1 152.144.199.210,152.144.199.211
mad echanges to /etc/hosts on vmorarac1/2 for 2 new IP address assigned to vmorarac5
I believe, in a multiple RAC Physical Standby nodes only one node applies the archivelogs to the standby database. How you can find which node is the APPLY NODE using a SQL query?
I know you can veiw this information by querying the standby configuration using DGMGRL but how would you find out this information from a remote client using SQLPlus/PLSQL?
I have many other rac servers and sids setup this way and they all work,,I can login to sqlplus and do any changes .. no issues going into sqlplus in node 207 and node 208..I have checked the directory i.e mesg and teh files are there ..
I think the message is decieving as it complains about oracle home and sp1<lang>.msb not found ,, when i tis there.I have also checked my env variables .. they all there i.e
scrbdefrmr207:orarer 58> env |grep ORA dbms_type=ORA ORACLE_SID=RER001 ORACLE_HOME=/oracle/RER/102_64 ORACLE_BASE=/oracle
I am installing Oracle 10g Clusterware in RHEL 5 Server. When i run root.sh in the second node the last step Running vipca(silent) for configuring nodeapps failed. Then i run VIPCA manually in second node and configured vip configuration successfully,but when i checked the post-checks for cluster services the Checking existence of VIP node application,Checking existence of ONS node application and Checking existence of GSD node application in second node failed. I am able to ping both the servers one another with vip name successfully and the CSS,CRS and EVM appears healthy in both nodes. 1. Does the vip configuration is proper? If no what is causing this error and how to rectify it? Find the output below
I have 2 node 11gR2 RAC running on AIX 6.1, after I shutdown the database and restart crs using crsctl start crs, ASM instance comes back up but not the database and I had to start the database using srvctl, isn't that the database should come up when I start the crs?
I have a production database with 2 node Physical RAC ENV on two separate machines , and we are going to enable audit trail parameter to DB values to capture all failed login attempts on a database.
we have done this on our testing ENV but dont know the procedure to deploy this steps on 2 node RAC ENV . I have done the below steps on our standalone testing ENV
1.Log in as SYS and connect with the SYSDBA privilege. oUser Name: SYS oPassword: Enter your password. oConnect As: SYSDBA
2.To enable audit on DB use below command SQL>alter system set audit_trail=DB scope=SPFILE;
3.Then restart the database.
4.After successful startup use the below command to capture log SQL> audit create session whenever not successful;
steps to enable audit_trail to capture failed login attempts on two node RAC ENV.
i am configuring asm on clusterware when i create instance on node1 and node 2 a message screen appear " Cant start asm instance on node 2)When i start asm instance manually on node 2 i found the below error in alert log file.
ORA-27508 IPC error sending a message ORA-27300 OS system dependent operation:send msg failed with status: 101 ORA-27301 OS failure message: Network is unreachable ORA-27302 failure occurred at: sskgxpsnd1
during installtion of Oracle RAC 11g r1 on Linux using VMware server, all the pre cluster installation was successful, and root.sh on node 1 was successful too, but whenever I run the root.sh script on the second node I get error message "Failure at final check of oracle CRS stack. 10"
I have installed Oracle 10g RAC crs and asm on CentOS release 5.4. When I am rebooting the nodes and starting crs manually All the services on both nodes starting successfully.
1.In 2 node rac while adding datafile to tablespace if you forget to mention '+'then what will happen whether it is going to be create or it throws an error if it creates where exactly located and other node users how to work on that tablespace .what all steps to perform that datafile is useful for all node users.
2. In Rac environment how to check how many sessions connected to particular node.
3)In Rac any node is evicted due to network failure then after we rebuild the network .Is there any steps to do manually to access the failure node after rebuilding the network or it will automatically available in cluster group which service is perform this activity.
4.While configuring cluster ware you choose voting disk and ocr disk location and which redundancy you will choose suppose if you go for normal redundancy how many disks you can select for each file either one or two?
I am configuring two node RAC on oracle virtual box. Did its mandatory to configure DNS on both nodes before Rac installation? Whats the purpose of DNS in RAC. And if its mandatory did i have to configure DNS on each node separately?