RAC/ASM Clusterware Installation :: Patching Cluster From 11.2.0.1 To 11.2.0.3?
Aug 15, 2012
I'm a long time Oracle DBA but have only recently started with RAC and have never done any patching or upgrades in a RAC environment.
My environment consists of a 2 node RAC cluster, and 2 physical standbys at separate remote locations.
The database, GI, ASM, and clusterware are all at version 11.2.0.1.
I want to upgrade all components to the latest patchset which is 11.2.0.3. Eventually I also want to apply the latest PSU.
Can I do this without down time? It is a mission critical database where downtime could conceivably endanger lives. If it can't be done with zero downtime, then downtime must be kept to an absolute minimum (minutes, not hours).
I suspect the clusterware needs to be patched first and I do not know whether I can do it on one node at a time.
I am using newstart cluster software. currently I have two blade servers in my system and I have configured cluster as well. right now I have one issue. I am using the ip address of 10.34.14.0/28 network for cluster but this can not be accessed from all network so I want to use different range of IP of 10.68.1.0/28 network in the same blade servers. Can I configure two different floating (cluster) IP? the first IP range is for signalling and second IP range 10.68.1.0/28 network is for OMM. I need to use both IP in two blade servers.
After installing a 4 node cluster 11.2.0.3 with 16 CPU's (4 on each node) on IBM 795 with aix 6.1 each server is using 0.5 CPU with no user load on the system. Running SIHA on one server typically uses 0.05 CPU with no user load on the system.
im having problems with CRS in a Rac installation. Because of this i cant start the database.I installed Oracle Rac 11.2.0.1 with ASM in a Red Hat 5.8 linux server. Everything was ok until a server reboot. I have 2 nodes, nodo1 and nodo2. Both got rebooted and after that im not able to start the cluster.
All the other resources seem to be just fine .
At nodo1: [root@nodo1 oracle]# crsctl start cluster -all CRS-2672: Attempting to start 'ora.cssdmonitor' on 'nodo1' CRS-2672: Attempting to start 'ora.cssdmonitor' on 'nodo2' CRS-2676: Start of 'ora.cssdmonitor' on 'nodo1' succeeded CRS-2672: Attempting to start 'ora.cssd' on 'nodo1' CRS-2672: Attempting to start 'ora.diskmon' on 'nodo1' CRS-2676: Start of 'ora.cssdmonitor' on 'nodo2' succeeded CRS-2672: Attempting to start 'ora.cssd' on 'nodo2'
I have been asked to look at making an existing Oracle server resilant. The technology I need to use are:
- Windows 2008 R2 Enterprise SP2 - Microsoft Cluster Services - Oracle Enterprise 11g release 2 - Oracle Fail-Safe (latest version available)
Unfortunately ApEX 4.1 and Mod_plsql are install on the existing stand-alone server. I'm afraid i know very little about Oracle and ApEX but i need to make a design decision and check fiesablilty.
I believe the HTTP service and Mod_plsql can be moved to a independant server which should work but ApEX itself must remian on the database server (i'm told).Will ApEX work in a MSCS setup? I can only find references to RAC support, RAC is out of scope.
We have 2 servers running Oracle 11gR2 Window server 2008 64 bit RAC. Today I add new shared disk visible from both servers. All our data placed on ASM partitions, but we need Windows volume for exchanging some data between servers & users;Actions:then i have prepare disk on shared storage from 1st node, run diskmfmt.msc & rescan disks when windows recognizes new hdd, initialize it.create an Extended partition & Logical drive & Assign drive letter. from another nodes, make sure no drive letter is assigned.
Sometimes Windows automatically assign drive letter. If drive letter is assigned, right click & Change drive letter & Remove. from 1st node, format this logical drive using ocfsFormat.exe after execute below scripts i have face some issues
E:app11.2.0gridcfs>ocfsformat /m P: /c 1024 /v DATA /f /d /aReg Get Cluster Name(): Reg Query ValueEx for CFS_CLUSTER_NAME failed with error 203
This indicates that the Cluster Name has not been configured on this node for OCFSThe volume formatted in this condition will be seen by all nodes running OCFS.
Oracle VM 2 Node RAC cluster Oracle 11G 11.2.0.3.0 Oracle 11g rel 2 GI
on current 2 node cluster we have GI and RAC db configured Nodes: vmorarac1,vmorarac2
we shutdown vmorarac1 to clone it to vmorarac5
on new node I have changed the hostname to vmorarac5 In /etc/sysconfig/network-scripts/ifcfg-eth0 change ip to new ip and same for /etc/sysconfig/network-scripts/ifcfg-eth1 152.144.199.210,152.144.199.211
mad echanges to /etc/hosts on vmorarac1/2 for 2 new IP address assigned to vmorarac5
We install grid using Grid Installer GUI. When doing a fresh installation of Grid Infrastructure 11.2.0.2 or 11.2.0.3 in a new server , I want to apply its latest PSU as well. When should I do this PSU patching ?
A. Should I apply the GI PSU after the entire Grid installation is successfully completed ?
or
B. Should I apply the GI PSU just before running the root.sh ?
i.e. when the installer prompts you to run root.sh (see the screenshot below)
[URL].........
At this stage(before running root.sh) the GI binaries are in place . So , I can apply the PSU on these binaries? Right?Should I follow option A or B above ?
I have a RAC working with SAN raw devices presented to ASM and working fine. My situation has changed since the installation of the RAC. Now I need a shared storage (with a filesystem) to upload some files (so DB can read these as external tables) for this I have used OCFS2.
As you can see I have Oracle RAC on raw ASM + shared storage on OCFS2 and for the time been all seems to be well and working perfectly. Can these coexist with out effecting the other?
I would like to build a 2-node cluster without RAC. I desire to operate 2 Oracle database instances for use by 2 separate applications. I plan to run Linux. I would like node1 to act as the primary server for the databases in application1 and node2 to act as the failover server for the databases in application1. Conversely, I want node2 to act as the primary server for databases in application2 and node1 to act as the failover server for databases in application2. Can Oracle SE and Oracle Clusterware be configured in this this way?
I am setting up a RAC cluster in Oracle 11.2.0.3 on two Windows 2008R2 nodes.I have completed all the pre-installation tasks would like to verify this by running:cluvfy comp sys -p database -n allusing the standalone download of cluvfy - I haven't installed any Oracle software as yet. This completes successfully on the remote node but fails on the local node (whichever one I run it from).
There is a lot of output but these seem to be the key problems: [1852@PRDAT217] [Worker 3] [ 2013-10-01 14:44:45.190 BST ] [WindowsSystem.deleteService:876] _WS_ deleteService2: node PRDAT217 Service OracleRemExecService result: 0|Access is denied.[1852@PRDAT217] [Worker 3] [ 2013-10-01 14:44:45.190 BST ] [NativeResult.<init>:91] NativeResult: The String obtained is0|Access is denied.[1852@PRDAT217] [Worker 3] [ 2013-10-01 14:44:45.190 BST ] [NativeResult.<init>:99] The status string is: 0[1852@PRDAT217] [Worker 3] [ 2013-10-01 14:44:45.190 BST ] [NativeResult.<init>:112]
I am trying to install Oracle RAC installation w/ guest OS OracleLinux6 on Virtualbox/Mac OS.But getting problem to set network & ip setting and yum is not able to reach oracle repository.
if anyone tried installing Oracle RAC installation w/ guest OS OracleLinux6 on Virtualbox on intel based Mac OS.
I have to manage an active passive cluster on SLES10 SP4. Oracle 10.2.0.4. All oracle binaries are on a DRBD Device on mountpoint /oradata. This mountpoint includes the binaries and all datafiles.
I havent installed these systems. At one stage on node died and got reinstalled by a system administrator. After a switch from the running node to the standby node, the cluster is not starting.
Jan 3 14:36:01 a logger: Cluster Ready Services waiting on dependencies. Diagnostics in /tmp/crsctl.6114. Jan 3 14:36:14 a logger: Waiting for Oracle CSS service to be available before starting Jan 3 14:36:14 a logger: ASM instance +ASM. Wait 2.
After issuing $ORACLE_HOME/bin/localconfig add the cluster is starting, the asm instance is coming up, and also the database instance is starting.
Shortly after that the db instance dies. Alert log is showing something like recursive sql error. Looks like writing to the datafiles is not working. In asmcmd the diskgroups are showing DISMOUNTED (show MOUNTED before)
so what to do ? a) Why is it necessary to issue "localconfig add" to make the cluster starting ? b) What could be a reason for not being able to write to the database ? c) How to install an active/passive cluster ? c1) Install OS on both nodes c2) Install oracle binaries c2.1) on both nodes in the local files system and do the mapping to the drbd afterwards? c2.2) only on one node and just switch the mountpoint ?
We have 3 node RAC. In one of the node clusterware is down and also CSSD process is down. I got to know that we have to clear socket files in that node.what needs to be done instead ofclearing socket files.
database Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production PL/SQL Release 11.2.0.2.0 - Production CORE 11.2.0.2.0 Production TNS for Linux: Version 11.2.0.2.0 - Production NLSRTL Version 11.2.0.2.0 - Production [code]...
We're trying to utilize Oracle RAC 11.2. We want to have two servers in our grid. These two server are installed on a VM using Hyper-V software.
common storage for asm is also developed on a VM using FreeNAS software.Now, when we try to install Grid Infrastructure the disk that is recognized as a extra drive in both servers are disabled and we can't select it as ASM storage. The OS we're using is Windows Server 2008 R2.I have to say the IP requirements are considered and the servers have two IPs.
I want to install Oracle Grid software 11g R2. What is the Latest Oracle Grid software 11g R2 version available?Is it 11.2.0.1 or 11.2.0.3?If it is 11.2.0.3, is it available in metalink download, because I see only 11.2.0.1 version in metalink download?
My application consists of Oracle RAC and Oracle Enterprise Manage (OEM) on same two nodes. I am exploring the idea of utilizing Oracle Clusterware (which is already installed with Oracle RAC) to support OEM in active passive mode.
Here is the scenario.
Node A running Solaris 10.9 on SPARC and Node B running Solaris 10.9 on SPARC. Node A and Node B are both Oracle RAC Nodes ( Active Active ) which use Shared Storage for Oracle using ASM.
Node A and Node B will run Oracle Enterprise Manager (OEM) Application in active passive scenario which will be using Oracle RAC for its database. So, both Oracle RAC and OEM are sharing same server A and B for its clusters.
Now, OEM application need a third party clusterware to fail it over. It needs clusterware to provide floating VIP and shared folder of the size of 20GB which will keep software libraries. If node A goes down, node B have access to same libraries and will come up as active.
Do you know if oracle clusterware which comes with Oracle RAC can support OEM for failing over i.e. it can provide floating VIP and shared file system?
Am planning to install oracle RAC 11g on linux server. Have doubt on installation of clusterware or grid infrastructure first .?I think grid infrastructure covers clusterware and ASM both and then oracle 11g software to be installed. But how to download grid insfrastructure software. Will it come with oracle 11g software..? or need to download it seperately.? Because when i try to download oracle RAC 11g which gives link to download the 11g software not grid.
I am getting "ORA-03114: not connected to ORACLE" while installing "Oracle Database 11GR2" on "Oracle RAC 11R2" cluster setup. I am getting this error at the time of "Oracle Database Configuration Assistant" configuration reaches till 39% on HP-UX 11.31 IA64 platform.
I am installaing oracle RAC 11.2.0.1. After I installed the grid infrastructure, when I ran root.sh, I got below errors;
ASM failed to start. Check /d1/app/grid/cfgtoollogs/asmca/asmca-1109068AM4612.log for details.
Configuration of ASM failed, see logs for details Did not succssfully configure and start ASM CRS-2500: Cannot stop resource 'ora.crsd' as it is not running
I have installed Oracle Grid on a standalone server and setup Oracle db 11.2.0.2 on Oracle Linux 6.2 64 bit server.When I reoot the server and run crs_stat -t,several daemons havent started thus the ASM and db instances are also down as below
I am forced to manually start the daemons via command crsctl start resource -all then I manually start the ASM and db instances. Yet when I run the commands
crsctl config has CRS-4622: Oracle High Availability Services autostart is enabled.
crsctl check has CRS-4638: Oracle High Availability Services is online.
Thus I would assume the daemons would start automatically during boot.
Is it possible to upgrade GI 11.2.0.1 Enterprise Edition version to 11.2.0.3 STANDARD EDITION?
We have 2 Node RAC with 10.2 DB SE. Last time I think by mistake or not knowing the licensing information GI was upgraded to EE 11.2.0.1 by somebody or may be whatever version of GI is used, it has nothing to do with the license?
I have heard that I can upgrade Grid Infrastructure, but to Standard Edition. Does GI also comes with SE and also just want to know that can I perform this upgrade?
I want to upgrade (out of place) Grid Infrastructure to latest version of 11.2.0.3 and then PSU on top of that.
GI HOME and DB HOME are separated. Database I want to upgrade later on, not with this clusterware upgrade. I have download p10404530_112030_Linux-x86-64_3of7.zip from MOS for upgrading GI.
grid@db1p [rec]: /home/grid > crsctl query crs softwareversion Oracle Clusterware version on node [db1p] is [11.2.0.1.0] grid@db1p [rec]: /home/grid > crsctl query crs activeversion Oracle Clusterware active version on the cluster is [11.2.0.1.0]grid@db1px [ec]: /home/grid > sqlplus / as sysasm
[code]....
1) Is this the correct p10404530_112030_Linux-x86-64_3of7.zip to upgrade GI from 11.2.0.1 to 11.2.03?
2) Do I need a COMPLETE DOWNTIME to upgrade GI on 2 node RAC where DB is STANDARD EDITION but Clusterware is 11.2 as you have seen above..
3) Will this ASM version will also be upgraded along with GI upgrade installer or it must be upgrades separately after GI Upgrade?
When I use 'ONSCTL PING' to test the ons status, it says: configure localport and remoteport directives and attempt to start ONS server again.ons is not running.The content of D:app11.2.0gridopmnconfons.config is:
localport=6100 # line added by Agent allowgroup=true usesharedinstall=true remoteport=6200 # line added by Agent nodes=rac1:6200,rac2:6200 # line added by Agent
The content of D:app11.2.0gridopmnconfons.config.rac2 is: localport=6100 # line added by Agent allowgroup=true usesharedinstall=true remoteport=6200 # line added by Agent nodes=rac1:6200,rac2:6200 # line added by Agent
The configuration file is almost the same in rac1.But the output of 'srvctl status nodeapps' is :
ons is enabled ons is running on node1:rac1 ons is running on node2:rac2
RAC installation on vmware is failing with following errors. I followed below link for installation. URL..... Found the following errors:
[root@rac2 ~]# cd /crs/oracle/bin/ [root@rac2 bin]# ./vipca PRKR-1062 : Failed to find configuration for node rac1 PRKR-1062 : Failed to find configuration for node rac1 [code]....
The "/crs/oracle/cfgtoollogs/configToolFailedCommands" script contains all commands that failed, were skipped or were cancelled. This file may be used to run these configuration assistants outside of OUI. Note that you may have to update this script with passwords (if any) before executing the same.The "/crs/ oracle/cfgtoollogs/configToolAllCommands" script contains all commands to be executed by the configuration assistants. This file may be used to run the configuration assistants outside of OUI. Note that you may have to update this script with passwords (if any) before executing the same.
[root@rac2 bin]# /crs/oracle/bin/racgons add_config rac1.localdomain:6200 rac2.localdomain:6200 [root@rac2 bin]# /crs/oracle/bin/oifcfg setif -global eth0/192.168.2.0:public eth1/192.168.0.0:cluster_interconnect PRIF-10: failed to initialize the cluster registry [root@rac2 bin]# /crs/oracle/bin/oifcfg setif -global eth0/192.168.2.0:public eth1/192.168.0.0:cluster_interconnect PRIF-10: failed to initialize the cluster registry
tried to mount a big endian asm disk (AIX for example) on little endian env (linux x86)? I have a disk that KFED is able to read but ASM is unable to mount it due to endianess incompatibility.
We try to upgrade the 10g to 11g oracle instance and asm instance.Below steps we r done;
1)First we are install the 11.2.0.3 database and grid infrastructure in separate home.
2)after that i try to upgrade asm instance so i drop old 10g asm localconfig cd $ORACLE10g_HOME./localconfig deletecd $ORACLE11g_HOME./localconfig add but here i can not find localconfig Note:database and grid installed without error....
3)Even we are try dbua but it run 10g dbua we are not able to run dbua for 11g
4)inventory is corrupted how to recreate inventory
5)provide the steps for upgrade 10g asm to 11g asm