Server Administration :: Switch Between Disks In A Diskgroup?
Aug 31, 2011
I have an Oracle database 11.1.0.6 with ASM storage.
SQL> select * from v$version;
BANNER
--------------------------------------------------------------------------------
Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - 64bit Production
PL/SQL Release 11.1.0.6.0 - Production
CORE 11.1.0.6.0 Production
TNS for Solaris: Version 11.1.0.6.0 - Production
NLSRTL Version 11.1.0.6.0 - Production
There is a disk group named +FRA which has only 1 disk (call it disk 0 for later references).
+FRA disk group is in type of EXTERNAL redundancy
I'd like to add another disk to +FRA disk group and copy all files stored in disk 0 to the new disk (call it disk 1 for ex.), then take off disk 0 (not use for +FRA disk group any more).
Expected result: there is only one disk 1 left in the +FRA disk group, with all the data from disk 0 and keep their original file names
you have given 500gb of harddisk. how will u implement the size or how to assing the size for disks in the diskgroup.give me the best practice that is used in production
I've been ASM quite a while. I've two diskgroups extdg and nordg which are of external and normal redundancy respectively. Total diskgroup size is 9G, usuable space came down to 3mb when i tried to resize it to itz maximum. the value under usable column shows the same.
FYI... i've totally three disks in nordg, two in one failgroup and one in another.The disk sized equally 3g each.
I have a database started using pfile and spfile parameter shows NULL from command prompt as below:
****************************************************************** Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production With the Partitioning, OLAP and Data Mining options
SQL> show parameter spfile
NAME TYPE VALUE ------------------------------------ ----------- ------------------------------ spfile string SQL> *****************************************************************
Now when I want to change/switch database to start using SPFILE, can I do it with out restarting (shutdown and startup)the database.
I'm new on 11g, due to my limit storage with mounted internal disk, so I think, may I can store archivelog in ASM without FRA? However, I did not find any document to verify.
We have Exadata -V2 quarter rack with High Performance Disks. We applied EHCC's various compression methods on some of the table's partitions.
Now we are setting up Exadata Expansion Rack - High Capacity Disks. Post implementation, we would be moving older partitions to new Expansion Rack wherein compressed partitions are also included.
In this case, would there be any impact on the compression ratio as the expansion rack is having high capacity disks.
And, moving partition method would be same as it is for non-exadata database i.e. "alter table <table_name> move partition
I configure exadata + 2 switch between application server and and exadata. And we have problem to connect to switch and load balancing between Ethernet switch to exadata.
We might end up that is 'How about remove ethernet switch and directly connect to exadata?' I don't know which idea comes up with Ethernet gigabit switch between application server and exadata, exadata might be gigabit.
but from application server to ethernet switch is still 100M network, it is only fast in that load, other load is still 100m , it might useless. and about load balancing, exadata is RAC cluster node, exadata is doing load balancing, why we want another load balancing feature of ethernet switch?
Is it good idea to have dual switch between application server and exadata? exadata iti is what might be pro and con of this setting? exadata is handled by oracle support team, but dual switch is only cared by application operation team, so if something wrong with switch, nobody except operation team can manage .
I am a new Oracle DBA that has been tasked with creating some single node databases for our programmers to connect to. I have never built a database, only upgraded. I will be using a Dell 610 with 10.2.0.4 software. I have been reading the install guides but cannot find a definitive answer for how to create the raw partitions. I have a SAN that I can utilize and my current db is set up with RAID 6 like so:
Disk 2 300 GB 1 Extended 300 GB and 5 logical, 100MB, 30 MB, 100 GB, 100 GB, 100 GB Disk 3 100GB 1 Extended 100 GB and 3 logical 100 MB, 30 MB, 100 GB
why the old DBA partitioned this way? I am not good with storage and know that I need to understand before I build new databases so I do not duplicate something that is wrong. I was reading articles that recommend SAME but would not know how to do that on the SAN.
I will be using data pump to export the database and import to the new server. I was looking at using transportable tablespaces to do this. Would you recommend that from the cluster to the standalone?
Regarding 11gR2 RAC, ASM, and NFS. Installing 11gR2 Grid Infrastructure and 11gR2 RAC on OEL 5.5 servers with storage supplied by NetApp NFS shares over TCP/Ethernet.
How do I mount NFS file systems as block devices so ASM will use them?
It is my understanding that Linux mounts NFS file systems as anonymous file systems rather than block devices. ASM requires block devices. The only way I know to have Linux treat an anonymous file system mounted over NFS as a block device is to subsequently mount it as a loop device.The Oracle 11gR2 documentation states, "ASM does not support loop devices". Without defining NFS mounts as loop devices, oracleasm createdisk complains the given file on the NFS mounted file system is not a block device when i run
/usr/sbin/oracleasm createdisk DATA /oradata/prod/data/disk01
where /oradata is an NFS mount from a NetApp volume defined in /etc/fstab and disk01 was created with the dd command.I followed the steps in section 3.3.1.2 'Creating Files on a NAS Device for Use with Oracle ASM' of the Oracle guide..
When I use asmca, it does not find any candidate disks with the search path set accordingly.If successfully implemented ASM using NFS shares for database disk groups from a NAS, Steps I followed are below.
In /etc/fstab: server-ip-address:/vol/data /oradata nfs rw,bg,hard,nointr,tcp,nfsvers=3,timeo=600,rsize=32768,wsize=32768,actimeo=0 0 0 -- mount and verify. /oradata is mounted with specified options. # mount -a # mount [code]...
Command fails with error message "File /oradata/prod/data/disk01 is not a block device." None of the OUIs (grid installer, asmca, or dbca) identify any disks as ASM candidate disks with the search string '/oradata/prod/data/*'. I also tried using '/*', '/oradata/*', '/oradata/prod/data/*' to no avail.
I am using oel 5.8, oracle 11.2.0.3. The raw disk permissions are getting changed when the node gets rebooted.I am trying to use udev rule to change the permission permanently but its not working.
I am on H1 visa and working as chemical engineer...but I am bored of this job... I wanted to switch to IT and particularly oracle dba..
How much is the risk for me as I am coming from non computer science background..there are various IT consulting companies who can provide training and do h1 transfer......currently (year 2010) the market for oracle is booming a lot...so I was thinking what can go wrong if I do this switch.
I recently started working with legacy code and noticed that some huge tables (5 years worth of data, don't have more details on me right now but can post later if needed) are partitioned based on time sequence number column while majority of queries are done based on time (different column). Queries performance is degrading and I'd like to try to modify partitioning and run some tests to evaluate performance improvement.
My only concern is with so much live data I have to come up with solution on how to switch partitioning with the least impact on applications running 24 x 7. Something you have done in the same situation and it worked?
I have seen in many RMAN backup scripts that a manual log switching is performed in first line of the script? What is it purpose, just to write changes needed to reconstruct the DML/DDL statements to online redo logfiles and start the archiving process if db is running in archive log mode?
I am trying to install 2-node 11gR2 RAC on CentOS 5.6 using VMWare Fusion. I'm in the middle of Grid Infrastructure installation and while choosing the ASM diskgroups I don't see any candidate disks but when I click on all disks i can see them but can't select them.
I tried installing the grid before and ended up with and issue running the root.sh on all the nodes so had to deconfigure the grid binaries using the rootcrs.pl and then deinstalled the grid software using the ./desintall utility.
when tried installing again ...I ran the cluvfy utility ..it doesn't show any errors.
We are in preparation for the purchase of equipment for two- node Oracle10g RAC (on Windows Server 2008) and I need to configure the disks on servers. We have an existing EMC VNX5500 storage and SAN 8Gb switches.
My question is with how many disks to configure servers?Originally we planned four SSD discs (two in RAID 1 for OS and two in RAID 1 for locally Oracle home) for each node/server, and everything else on storage disks (Shared Disks Storage).
Whether it is necessary to have more local disks and for what? Where should be redologs files, archivelogs files, backup files…
Whether it is better server to be configured with six SSD disks: - two in RAID 1 for OS, - four in RAID 10 (or RAID 1) for locally Oracle home, archivelogs files, or something else…?
ASM instance SID = +ASM and the database installation was under SID =rdbms112
Im just in the process of running dbca and have specified which ASM disks to use for data, but then on the next screen of the dbca, im asked for the ASMNMP password,
the password for 'oracle' user is 'oracle' but im getting this message:
"Could not validate ASMNP password due to the following error: "ORA-01031-insufficient privilages" Do you want to continue? If you contiue ASM will not be configured to be moitored by database controll "
when I was asked for privilaged operating system groups when setting up ASM
OSDBA, OSOPER and OSASM ... were all set to group 'dba'
ORA-00202: control file: '+DATA/xa21/control01.ctl' ORA-17503: ksfdopn:2 Failed to open file +DATA/xa21/control01.ctl ORA-15001: diskgroup "DATA" does not exist or is not mounted ORA-15077: could not locate ASM instance serving a required diskgroup
The alert log +ASM instance, said:
ORA-27091: unable to queue I/O ORA-27070: async read/write failed OSD-04006: ReadFile() failure, unable to read from file O/S-Error: (OS 23) Data error (cyclic redundancy check).
I tried to resolve a physical error in file systems, with CHKDSK tool, unsuccessfully.
Normally ASMlib/oracleasm is used to prepare the disks for ASM. I just wonder, besides this tool, is there any other GUI based tool to use for preparing the ASM disks?
My exadata quarter rack machine has two asm diskgroups, DATA1 with 5TB and RECO with 3TB. I'd like to resize RECO to 1TB and DATA1 to 7TB.
I know ALTER DISKGROUP RESIZE command, but my question is about resize RECO volume from 3 to 1 TB: is it supported by Oracle?? Let me know risks /issues with this resize?
Then it archives the current redo and switches to next redo.
If again switching is done, then just wanted to know if redo is archived when its size is less than log_buffer, then is it resumed when it is switched into next time or after archiving it creates a empty redo file and then use it?
I have restored all datafile to new location (that is ASM) earlier it was on file system using set new name command (RMAN). Now when i'm trying to update control file with swicth command it is prompting error. Version is 11g R2.
RMAN> SWITCH DATAFILE ALL;
RMAN-00571: =========================================================== RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS =============== RMAN-00571: =========================================================== RMAN-00558: error encountered while parsing input commands RMAN-01009: syntax error: found "all": expecting one of: "double-quoted-string, integer, single-quoted-string" RMAN-01007: at line 1 column 17 file: standard input