RAC & Failsafe :: Single Node RAC - Configure Only One Server Instead Of Two Node?
Sep 13, 2012
I would like get more details about single node RAC.. What is single Node RAC?..
1). just configure only one server instead of two node... same RAC installation.. cluster installation and shared disk configuration in single server itself.. in future we can add node to this ...
2). in 11gR2 having new feature Single Node RAC.. Multiple instances running on same node and configure using Cluster...
is there any advantage using single node rac(other than online upgrade to full RAC).
We have oracle database 10g(10.2.0.4) running on AIX 5.1. we have installed 2-node RAC on another server AIX 6.1 . Now we want to migrate our single node database to new 2-node rac environment. how we can move/migrate/cluster-aware our single node database to 2-node rac?
We have to configure Dataguard for our 24x7 critical banking 2 Node RAC database(10.2.0.4). Before proceeding with the configuration we have to make sure what steps we should follow to have minimum or no downtime.
1) Document covering DG setup in RAC environment. 2) We have to perform switchover as well so need its steps too. Either the normal switchover steps would be used or have to stop/start rac services as well.
I have implemented Rac One node with two machine(dbtest01,dbtest02) in 11.2.0.3 Redhat 5.3 one for online another one for offline. i have created TAF also.
dbtest01 is online and create one query DECLARE i number(36):=1; begin while(i<10000000) loop dbms_output.put_line(i); i:=i+1; end loop; end;
i executed above query for test and i made relocate database using srvctl utility to dbtest02 machine.
session is relocated successfully with new session id...above query getting stop in 25000 itself. i want to make it continue that process..also i tried to find Omotion utility in all the path.. couldn't find it... where will it be.. i want to relocate all the session without losing work.
from application i am getting disconnect after relocation.. application server need to restart to get work again...
I have 2 node 11gR1 running on Unix, I want to add 3 more RAC node to it. What are all the steps involved and should I first add 3rd node and then the 4th node or I should add both at the same time? is there any difference between these two methods?
Oracle VM 2 Node RAC cluster Oracle 11G 11.2.0.3.0 Oracle 11g rel 2 GI
on current 2 node cluster we have GI and RAC db configured Nodes: vmorarac1,vmorarac2
we shutdown vmorarac1 to clone it to vmorarac5
on new node I have changed the hostname to vmorarac5 In /etc/sysconfig/network-scripts/ifcfg-eth0 change ip to new ip and same for /etc/sysconfig/network-scripts/ifcfg-eth1 152.144.199.210,152.144.199.211
mad echanges to /etc/hosts on vmorarac1/2 for 2 new IP address assigned to vmorarac5
I believe, in a multiple RAC Physical Standby nodes only one node applies the archivelogs to the standby database. How you can find which node is the APPLY NODE using a SQL query?
I know you can veiw this information by querying the standby configuration using DGMGRL but how would you find out this information from a remote client using SQLPlus/PLSQL?
I have many other rac servers and sids setup this way and they all work,,I can login to sqlplus and do any changes .. no issues going into sqlplus in node 207 and node 208..I have checked the directory i.e mesg and teh files are there ..
I think the message is decieving as it complains about oracle home and sp1<lang>.msb not found ,, when i tis there.I have also checked my env variables .. they all there i.e
scrbdefrmr207:orarer 58> env |grep ORA dbms_type=ORA ORACLE_SID=RER001 ORACLE_HOME=/oracle/RER/102_64 ORACLE_BASE=/oracle
I am installing Oracle 10g Clusterware in RHEL 5 Server. When i run root.sh in the second node the last step Running vipca(silent) for configuring nodeapps failed. Then i run VIPCA manually in second node and configured vip configuration successfully,but when i checked the post-checks for cluster services the Checking existence of VIP node application,Checking existence of ONS node application and Checking existence of GSD node application in second node failed. I am able to ping both the servers one another with vip name successfully and the CSS,CRS and EVM appears healthy in both nodes. 1. Does the vip configuration is proper? If no what is causing this error and how to rectify it? Find the output below
I have 2 node 11gR2 RAC running on AIX 6.1, after I shutdown the database and restart crs using crsctl start crs, ASM instance comes back up but not the database and I had to start the database using srvctl, isn't that the database should come up when I start the crs?
I have a production database with 2 node Physical RAC ENV on two separate machines , and we are going to enable audit trail parameter to DB values to capture all failed login attempts on a database.
we have done this on our testing ENV but dont know the procedure to deploy this steps on 2 node RAC ENV . I have done the below steps on our standalone testing ENV
1.Log in as SYS and connect with the SYSDBA privilege. oUser Name: SYS oPassword: Enter your password. oConnect As: SYSDBA
2.To enable audit on DB use below command SQL>alter system set audit_trail=DB scope=SPFILE;
3.Then restart the database.
4.After successful startup use the below command to capture log SQL> audit create session whenever not successful;
steps to enable audit_trail to capture failed login attempts on two node RAC ENV.
here we have san server and rac database two node database.i want to know how can i create directory on rac database. i know how to create directory on single database but i do not know how can i create directory on rac database.
i am configuring asm on clusterware when i create instance on node1 and node 2 a message screen appear " Cant start asm instance on node 2)When i start asm instance manually on node 2 i found the below error in alert log file.
ORA-27508 IPC error sending a message ORA-27300 OS system dependent operation:send msg failed with status: 101 ORA-27301 OS failure message: Network is unreachable ORA-27302 failure occurred at: sskgxpsnd1
during installtion of Oracle RAC 11g r1 on Linux using VMware server, all the pre cluster installation was successful, and root.sh on node 1 was successful too, but whenever I run the root.sh script on the second node I get error message "Failure at final check of oracle CRS stack. 10"
I have installed Oracle 10g RAC crs and asm on CentOS release 5.4. When I am rebooting the nodes and starting crs manually All the services on both nodes starting successfully.
1.In 2 node rac while adding datafile to tablespace if you forget to mention '+'then what will happen whether it is going to be create or it throws an error if it creates where exactly located and other node users how to work on that tablespace .what all steps to perform that datafile is useful for all node users.
2. In Rac environment how to check how many sessions connected to particular node.
3)In Rac any node is evicted due to network failure then after we rebuild the network .Is there any steps to do manually to access the failure node after rebuilding the network or it will automatically available in cluster group which service is perform this activity.
4.While configuring cluster ware you choose voting disk and ocr disk location and which redundancy you will choose suppose if you go for normal redundancy how many disks you can select for each file either one or two?
we encounter a problem with a 2-nodes RAC with 11gR2 SE : Database RAC is correctly running but we encounter problems using RAC TNS entry, obtaining the message ORA-01033.We can access to the DB using a TNS entry using each one of IP addresses but when we use the SCAN address, it fails. 2 weeks ago RAC2 node could connect perfectly with the SCAN address but RAC1 (used as a client) obtained error message almost 30% of tries.
We made a reboot of RAC1 node and now SCAN address works correctly from RAC1 and RAC2 (used as a client) fails systematically with SCAN address but works perfectly with direct access to the database.On both nodes, SCAN address is perfectly resolved by DNS server, cluster seems correct, listeners seems correct...We have opened an SR on Metalink 4 weeks ago but we don't progress.
I am configuring two node RAC on oracle virtual box. Did its mandatory to configure DNS on both nodes before Rac installation? Whats the purpose of DNS in RAC. And if its mandatory did i have to configure DNS on each node separately?
I have the following problem on a 10.2.0.4 Oracle on AIX 1.6...
I have 2 RAC nodes as the node # 2 goes down, the VIP migrated correctly to the node # 1, but the listener # 1 do not see the change, and just resolve to the local ip.
The same does not apply in the reverse case, when it drops the node N 1, the listener attends to the fallen VIP node and migrated.The configuration is correct and identical in both nodes, according to the documentation, but I can not find a reasonable answer.
IP node # 1. scm1or01 192.68.60.11 scm1or01-VIP 192.68.60.13
IP node # 2 scm1or02 192.168.60.12 scm1or01-VIP 192.168.60.14
Below I write as I detect the problem: when scm1or02 is down:
I have two node RAC database running on linux machine (Oracle10G SE), our requirnment is to install new oracle binaries for Oracle10G EE on diffrent home location.after install Oracle10G EE binaries on new oracle home , i want to point our existing database to new Oracle10G EE home to adotp a features of Oracle10G EE features.
But at the same time i dont want to deinstall our Oracle10G SE binaries.So it is possible to rollback our process to point our database back to Oracle10G SE from Oracle10G EE in case anything going wrong.
Normally if node fails, VIP & resources are transferred to another node (clients need to failover).
if node is back again CRS detect this, make it as 'family member' and try transfer VIP & resources back to this node (clients will force to failover again).
this failover (despite of failover method I use) make each uncommitted transaction to be rolled back. client need to be disconnected from the service and VIP through they are connected to this service and they need reconnect to the new node - during this stage all DMLs modifications are gone, and need to be re execute.
srvctl config PRCR-1119 : Failed to look up CRS resources of database type PRCR-1115 : Failed to find entities of type resource that match filters (TYPE == ora.database.type) and contain attributes DB_UNIQUE_NAME,ORACLE_HOME,VERSION Cannot communicate with crsd