I have a table called dept table with nested collection as below. Nested Table Collection:
create or replace
type courselist as table of varchar2(64) Table Using Nested Table Collection:
CREATE TABLE "FCSDWH_STG"."DEPT"
( "NAME" VARCHAR2(20 BYTE),
"DIRECTOR" VARCHAR2(20 BYTE),
"OFFICE" VARCHAR2(20 BYTE),
"COURSES" "FCSDWH_STG"."COURSELIST"
) Content Of Table:NameDirectorofficecoursesAccountingJames CharlesUNOFCSDWH_STG.COURSELIST('natural science','chemistry','Computer Science','Computer Science')
I am trying to select and print an element from nested table collection using below plsql block.
MY_COURSE VARCHAR2(64 CHAR);
BEGIN
SELECT courses INTO MY_COURSE FROM TABLE(SELECT COURSES FROM DEPT) where courses='chemistry';
DBMS_OUTPUT.PUT_LINE(MY_COURSE);
END; Error Message:
PL/SQL: ORA-00904: "COURSES": invalid identifier
We had a massive jump in cluster waits specifically; gc buffer busy acquire during an RMAN backup. We identified the cause of the waits to a few hot blocks with a table that may well need re-building in terms of ITLs and PCTFREE (Although I thorght ASSM would manage PCTFREE and PCTUSED..)
What happens during an RMAN backup that may cause huge cluster waits on hot tables? Is there some crazy redo issue going on or maybe flashback?
Either way the spike is there and we can pinpoint the activity on the database but just dont understand why RMAN would cause the issue.We have just found that the LARGE_POOL has not been set, in fact there is no SGA_TARGET either! Could this have an effect on RMAN and cluster waits?
I tried to run the duplicate script after transferring new rman backup to standby. Even after transferring the backup, duplicate command and was unable to open the database.
Yesterday I was installing 11.2.0.2 patch to our Oracle 11gR2(11.2.0.1.0) Two node RAC cluster on OEL 5.4 x86 Linux.Before installing I have read [URL] support note which says "Note: All Oracle Grid Infrastructure patch set upgrades must be out-of-place upgrades, in which you install the patch set into a new Grid home. In-place patch set upgrades are not supported. "
During Installation I found following Installation Option :
1) Install and configure Oracle grid Infrastructure for a cluster
2) Configure Oracle Grid Infrastructure for a Standalone Server
3) Upgrade Oracle Grid Infrastructure or Oracle Automatic Storage Management
4) Install Oracle Grid Infrastructure Software Only
i want to install oracle real application cluster using shared file system and not using ASM. we are not creating LUN for our data to kept and SAN admin will give us NFS share where we will place the database.
when i'm creating ASM instance, i'm getting an error
Quote:ora-29701: unable to connect to cluster manager
for this there were some workaround given to start css.
i tried everything from /etc/init.d/init.cssd start/enable but nothing worked, then i tried with executing
localconfig delete localconfig add====>it failed and recommended to go for reset option
when i executed localconfig reset i get this following error:
Quote: [root@rac2 bin]# sh localconfig add Successfully accumulated necessary OCR keys. Creating OCR keys for user 'root', privgrp 'root'.. Operation successful. Configuration for local CSS has been initialized
I am trying to create a rac setup on oracle 10.2. through NFS cluster configuration .i succeeded with the installtion of cluster configuarion and installation of oracle software ,but when i am trying to create database ,
I am installaing oracle RAC 11.2.0.1. After I installed the grid infrastructure, when I ran root.sh, I got below errors;
ASM failed to start. Check /d1/app/grid/cfgtoollogs/asmca/asmca-1109068AM4612.log for details.
Configuration of ASM failed, see logs for details Did not succssfully configure and start ASM CRS-2500: Cannot stop resource 'ora.crsd' as it is not running
It appears that the Oracle RAC cluster does not recognize the 'dba' group membership assignment to a user, unless it is assigned as the user's primary group.This also only affects newly created users, older users do not have this issue.
This affects us in the following way:
We are unable to create new cluster resources with the SAPCTL tool with the <sid>adm user.
When we try to create a resource using sapctl as the <sid>adm user we receive the following error:
CRS-0259: Owner of the resource does not belong to the group. SAP ABAP VIP creation failed.
The description of this error is that the resource owner (<sid>adm) is not a member of the group defined by 'osdbagrp' (what the cluster expects group membership for os users to control cluster resources), in our case this group is 'dba'.This is despite the user definitely being a member of the group.
Resource creation workaround:We are able to get around the resource creation by creating the resources either as root user, or by assigning 'dba' as <sid>adm's primary group.However, although we are able to control the VIP cluster resource, we are unable to control the other SAP enqueue resources as we assume that they need to be created by a user with 'sapsys' as a primary group (i.e. the resources do not start or it appears that there is not even an attempt to start the resources as there are no SAP startup logs created).
We have attempted multiple variations of creating and controlling cluster resources: changing <sid>adm primary group, changing cluster resource permissions, with nothing working for us.
The bottom line is that we need to be able to create these sapctl resources with a <sid>adm user that has 'sapsys' as a primary group while also being a member of the 'dba' group.
The installation of CRS is successfull & at end it ask to execute 2 scripts on both nodes i.e. orainstRoot.sh and root.sh
When the scripts are executed,the execution of orainstRoot.sh is successfull , but root.sh give following error:
[root@rac2host crs]# ./root.sh WARNING: directory '/home/oracle/product' is not owned by root WARNING: directory 'home/oracle' is not owned by root Checking to see if Oracle CRS stack is already configured /etc/oracle does not exist. Creating it now.
[code]....
Is this concerned with permissions on OCR, Voting Disk.
My application consists of Oracle RAC and Oracle Enterprise Manage (OEM) on same two nodes. I am exploring the idea of utilizing Oracle Clusterware (which is already installed with Oracle RAC) to support OEM in active passive mode.
Here is the scenario.
Node A running Solaris 10.9 on SPARC and Node B running Solaris 10.9 on SPARC. Node A and Node B are both Oracle RAC Nodes ( Active Active ) which use Shared Storage for Oracle using ASM.
Node A and Node B will run Oracle Enterprise Manager (OEM) Application in active passive scenario which will be using Oracle RAC for its database. So, both Oracle RAC and OEM are sharing same server A and B for its clusters.
Now, OEM application need a third party clusterware to fail it over. It needs clusterware to provide floating VIP and shared folder of the size of 20GB which will keep software libraries. If node A goes down, node B have access to same libraries and will come up as active.
Do you know if oracle clusterware which comes with Oracle RAC can support OEM for failing over i.e. it can provide floating VIP and shared file system?
RAC installation on vmware is failing with following errors. I followed below link for installation. URL..... Found the following errors:
[root@rac2 ~]# cd /crs/oracle/bin/ [root@rac2 bin]# ./vipca PRKR-1062 : Failed to find configuration for node rac1 PRKR-1062 : Failed to find configuration for node rac1 [code]....
The "/crs/oracle/cfgtoollogs/configToolFailedCommands" script contains all commands that failed, were skipped or were cancelled. This file may be used to run these configuration assistants outside of OUI. Note that you may have to update this script with passwords (if any) before executing the same.The "/crs/ oracle/cfgtoollogs/configToolAllCommands" script contains all commands to be executed by the configuration assistants. This file may be used to run the configuration assistants outside of OUI. Note that you may have to update this script with passwords (if any) before executing the same.
[root@rac2 bin]# /crs/oracle/bin/racgons add_config rac1.localdomain:6200 rac2.localdomain:6200 [root@rac2 bin]# /crs/oracle/bin/oifcfg setif -global eth0/192.168.2.0:public eth1/192.168.0.0:cluster_interconnect PRIF-10: failed to initialize the cluster registry [root@rac2 bin]# /crs/oracle/bin/oifcfg setif -global eth0/192.168.2.0:public eth1/192.168.0.0:cluster_interconnect PRIF-10: failed to initialize the cluster registry
I have 2 seperate rman backup one backup is of only datafile ,spfile and controlfile which i am able to restore and recover without any problem say bkp1 taken at 10 am
other set of backup is of only archive log files of same day but later time than datafile backup say bkp2 taken at 8pm...if i restore and recover bkp1 and try to restore bkp2 it gives error datafile exist (possibly due to fact both backup have control file)
if i just restore bkp1 and try to restore bkp2 so that i can do one recover at time it gives error datafile permission issue (possibly due to fact both backup have control file)
I want to restore database upto 8pm time how can i use both bkp1 and bkp2 to do it(restore datafiles and apply all archive logs on it)
We have a script which uses the duplicate functionality of RMAN. It was designed for a standalone database. Now I need to make it work for a cluster DB (RAC installation).
1. Does RMAN support this? 2. Do I need to make separate executions for both the nodes of the cluster?
We are currently blocked because we have no way of getting our system running without this! ******************************8 First attempt:
3808 11:52:49 ----- Starting Cloning Procedure ----- 3808 11:52:50 INFO: Client host 'sf25' says, it is 'sf25'. 3808 11:52:51 Getting Original DB Name on the source host (referred by ADVFRW_sf25 TNS-name)... ... 3808 RMAN: GROUP 8 ( '/export/home/oracle/dev/ADVFRW/ADVFRW.redo181' ) SIZE 134217728 REUSE 3808 RMAN: DATAFILE 3808 RMAN: '/export/home/oracle/dev/ADVFRW/ADVFRW.system' 3808 RMAN: CHARACTER SET WE8ISO8859P1 [code]....
5548 12:27:05 ----- Starting Cloning Procedure ----- 5548 12:27:06 INFO: Client host 'sf25' says, it is 'sf25'. 5548 12:27:07 Getting Original DB Name on the source host (referred by ADVFRW_sf25 TNS-name)... 5548 12:27:07 Original DB is 'ADVFRW' 5548 12:27:08 INFO: BGROUP=Srv [code].......
Now we have contrary error messages. Target DB name ADVFRW does not work, because on the target nodes there are configuration files for the node specific DB names (ADVFRW1.ora on first node resp. ADVFRW2.ora on second node). But with the node specific DB name the script fails, because it does not match with the DB name of the source system.
We are doing RMAN Duplicate set until time to refresh daily our test database for our developers and it taking long time to finish. We noticed on the restore log that RMAN was using a day old old backup pieces to refresh the test database and don't immediately use the latest backup pieces instead.
For additional details here's the rman duplicate command we are using which we run daily(mon-sat) at 4am once daily full backup on production completed.
RMAN Duplicate commands: run { set until time "to_date(to_char(sysdate,'Mon DD YYYY') || ' 04:00:00', 'Mon DD YYYY HH24:MI:SS')"; allocate auxiliary channel ch1 type disk; duplicate target database to testdb; } exit
Is there a way on how to let RMAN use the latest backup pieces instead?
I am tryign to run a split onlike full backup from the os useign the sap command which is linked to rman..Command that i am useing is ...
brbackup -u / -c force -t online_split -m full -p initBR1_onlinefull.sap
The backup goes thru but i get this erro on the end...
BR0522I 57 of 57 files/save sets processed by RMAN BR0280I BRBACKUP time stamp: 2011-01-26 12.23.26 BR0505I Full database backup (level 0) using RMAN successful BR0280I BRBACKUP time stamp: 2011-01-26 12.23.28
[code]...
Copyright (c) 1982, 2007, Oracle. All rights reserved.
Is this script is enough to delete the backup pieces older than 2 days.
CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 2 DAYS; configure device type disk parallelism 1 backup type to compressed backupset;
run { allocate channel d1 type disk ; backup filesperset 5 format '/u06/backup/EBSDEV/EBSDEV_data_t%t_s%s_p%p' database ; sql 'alter system archive log current' ; backup filesperset 20 format '/u06/backup/EBSDV3/EBSDV3_arch_t%t_s%s_p%p' archivelog all delete input ;