We have Exadata -V2 quarter rack with High Performance Disks. We applied EHCC's various compression methods on some of the table's partitions.
Now we are setting up Exadata Expansion Rack - High Capacity Disks. Post implementation, we would be moving older partitions to new Expansion Rack wherein compressed partitions are also included.
In this case, would there be any impact on the compression ratio as the expansion rack is having high capacity disks.
And, moving partition method would be same as it is for non-exadata database i.e. "alter table <table_name> move partition
One of my databases which is running on exadata x2-2 , has been restored to non-exadata machine in order to test few things. I had a sub-partition table in the exadata , compressed for query high. test machine (NOT EXADATA) , after uncompressing this subpartiton tables i am getting the following err message :
ORA-64307: hybrid columnar compression is not supported for tablespaces on this storage type I have executed the following commands : alter table crm.cm_ncd modify partition P01_CM_NCD nocompress; alter table crm.cm_ncd modify partition P02_CM_NCD nocompress; alter table crm.cm_ncd modify partition P03_CM_NCD nocompress;
[code]...
ERROR at line 1: ORA-12801: error signaled in parallel query server P005 ORA-64307: hybrid columnar compression is not supported for tablespaces on this storage type If all the partitions are uncompressed why i am getting this error message ?
Does RMAN backupset (backup to ) on DBFS is supported. I can find ACFS supported but is not mentioned for DBFS. My current customer is thinking backup to DBFS then copy to tape as the interim solution before getting ZFS next year.
I am having an Oracle 11g R2 RAC setup. My database size is 22TB in normal servers. I have checked my database in Exadata X2-2 and found out that the HCC was good and as per my assumption the 22TB can come down to 10TB .
My challenge is i need to transfer the 22TB data to the exadata quater rack which is having space constraint. Is there any way other than export & Import as data export and import will use entire 22TB. Is there any way to transfer the data to Exadata with compressed form from the source server.
I am having doubt on the no_index concept in oracle. I am using oracle exadata server.It is basically data warehouse project.I am in the situation to join some tables and get the result set for reporting purpose.
Among the tables, 2 tables having huge count.1st table has more than - 1,000,000,000 rows2nd table has more than - 200,000,000 rows when i join these 2 tables with some small set of tables, it is taking long time (around 20 MIN) to retrieve the result set. The final result set is around 100 rows only.
But, when i force NO_INDEX hint in the same query, is giving the same result in very fast manner(around 5 MIN). Because it is working based on cell smart scan. So, can i force the NO_INDEX hint to all tables? I forced the NO_INDEX hint only the table which is contain 200,000,000 rows not for others.
Query Plan : Normal Query using the the range scan based on the key. No_INDEX Query going full table scan.
1) Post Installation check list ie: How can I see every things are installed correctly. 2) upgrade the exadata OS version . 3) Upgrade the Database. 4) Test Migration from Oracle database to Exadata. 5) Right now there RAC servers are installed without domain, I need to add domain name to existing RAC.
I am a new Oracle DBA that has been tasked with creating some single node databases for our programmers to connect to. I have never built a database, only upgraded. I will be using a Dell 610 with 10.2.0.4 software. I have been reading the install guides but cannot find a definitive answer for how to create the raw partitions. I have a SAN that I can utilize and my current db is set up with RAID 6 like so:
Disk 2 300 GB 1 Extended 300 GB and 5 logical, 100MB, 30 MB, 100 GB, 100 GB, 100 GB Disk 3 100GB 1 Extended 100 GB and 3 logical 100 MB, 30 MB, 100 GB
why the old DBA partitioned this way? I am not good with storage and know that I need to understand before I build new databases so I do not duplicate something that is wrong. I was reading articles that recommend SAME but would not know how to do that on the SAN.
I will be using data pump to export the database and import to the new server. I was looking at using transportable tablespaces to do this. Would you recommend that from the cluster to the standalone?
Regarding 11gR2 RAC, ASM, and NFS. Installing 11gR2 Grid Infrastructure and 11gR2 RAC on OEL 5.5 servers with storage supplied by NetApp NFS shares over TCP/Ethernet.
How do I mount NFS file systems as block devices so ASM will use them?
It is my understanding that Linux mounts NFS file systems as anonymous file systems rather than block devices. ASM requires block devices. The only way I know to have Linux treat an anonymous file system mounted over NFS as a block device is to subsequently mount it as a loop device.The Oracle 11gR2 documentation states, "ASM does not support loop devices". Without defining NFS mounts as loop devices, oracleasm createdisk complains the given file on the NFS mounted file system is not a block device when i run
/usr/sbin/oracleasm createdisk DATA /oradata/prod/data/disk01
where /oradata is an NFS mount from a NetApp volume defined in /etc/fstab and disk01 was created with the dd command.I followed the steps in section 3.3.1.2 'Creating Files on a NAS Device for Use with Oracle ASM' of the Oracle guide..
When I use asmca, it does not find any candidate disks with the search path set accordingly.If successfully implemented ASM using NFS shares for database disk groups from a NAS, Steps I followed are below.
In /etc/fstab: server-ip-address:/vol/data /oradata nfs rw,bg,hard,nointr,tcp,nfsvers=3,timeo=600,rsize=32768,wsize=32768,actimeo=0 0 0 -- mount and verify. /oradata is mounted with specified options. # mount -a # mount [code]...
Command fails with error message "File /oradata/prod/data/disk01 is not a block device." None of the OUIs (grid installer, asmca, or dbca) identify any disks as ASM candidate disks with the search string '/oradata/prod/data/*'. I also tried using '/*', '/oradata/*', '/oradata/prod/data/*' to no avail.
I am using oel 5.8, oracle 11.2.0.3. The raw disk permissions are getting changed when the node gets rebooted.I am trying to use udev rule to change the permission permanently but its not working.
I have an Oracle database 11.1.0.6 with ASM storage.
SQL> select * from v$version; BANNER -------------------------------------------------------------------------------- Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - 64bit Production PL/SQL Release 11.1.0.6.0 - Production CORE 11.1.0.6.0 Production TNS for Solaris: Version 11.1.0.6.0 - Production NLSRTL Version 11.1.0.6.0 - Production
There is a disk group named +FRA which has only 1 disk (call it disk 0 for later references).
+FRA disk group is in type of EXTERNAL redundancy
I'd like to add another disk to +FRA disk group and copy all files stored in disk 0 to the new disk (call it disk 1 for ex.), then take off disk 0 (not use for +FRA disk group any more).
Expected result: there is only one disk 1 left in the +FRA disk group, with all the data from disk 0 and keep their original file names
I am trying to install 2-node 11gR2 RAC on CentOS 5.6 using VMWare Fusion. I'm in the middle of Grid Infrastructure installation and while choosing the ASM diskgroups I don't see any candidate disks but when I click on all disks i can see them but can't select them.
I tried installing the grid before and ended up with and issue running the root.sh on all the nodes so had to deconfigure the grid binaries using the rootcrs.pl and then deinstalled the grid software using the ./desintall utility.
when tried installing again ...I ran the cluvfy utility ..it doesn't show any errors.
We are in preparation for the purchase of equipment for two- node Oracle10g RAC (on Windows Server 2008) and I need to configure the disks on servers. We have an existing EMC VNX5500 storage and SAN 8Gb switches.
My question is with how many disks to configure servers?Originally we planned four SSD discs (two in RAID 1 for OS and two in RAID 1 for locally Oracle home) for each node/server, and everything else on storage disks (Shared Disks Storage).
Whether it is necessary to have more local disks and for what? Where should be redologs files, archivelogs files, backup files…
Whether it is better server to be configured with six SSD disks: - two in RAID 1 for OS, - four in RAID 10 (or RAID 1) for locally Oracle home, archivelogs files, or something else…?
ASM instance SID = +ASM and the database installation was under SID =rdbms112
Im just in the process of running dbca and have specified which ASM disks to use for data, but then on the next screen of the dbca, im asked for the ASMNMP password,
the password for 'oracle' user is 'oracle' but im getting this message:
"Could not validate ASMNP password due to the following error: "ORA-01031-insufficient privilages" Do you want to continue? If you contiue ASM will not be configured to be moitored by database controll "
when I was asked for privilaged operating system groups when setting up ASM
OSDBA, OSOPER and OSASM ... were all set to group 'dba'
you have given 500gb of harddisk. how will u implement the size or how to assing the size for disks in the diskgroup.give me the best practice that is used in production
Normally ASMlib/oracleasm is used to prepare the disks for ASM. I just wonder, besides this tool, is there any other GUI based tool to use for preparing the ASM disks?
I have 11.1.0.7 instance running and got more 300G of storage available for the database. I would like to know what are the steps to add these disks/storage to the existing DATA asm disk group?
My exadata quarter rack machine has two asm diskgroups, DATA1 with 5TB and RECO with 3TB. I'd like to resize RECO to 1TB and DATA1 to 7TB.
I know ALTER DISKGROUP RESIZE command, but my question is about resize RECO volume from 3 to 1 TB: is it supported by Oracle?? Let me know risks /issues with this resize?
How to compress sub partition on exadata, using 'for query high' and pctfree 10 options? I used this statement, but I get only ORA-14160:this physical attribute may not be specified for a table subpartition.
alter table table_name move subpartition subpartition_name PCTFREE 10 compress for query high;
My management is wanting to know the serial numbers of all the components of our two Exadata machines: one quarter-rack V2 and one half-rack V1
I can use dmidecode to get the appropriate information for comp nodes and storage cells, but not for the Cisco/Voltaire switch, nor the IB switches. I read MOS 1299791.1 and the thought of asking a DOC operative to pull out the label ‘which could damage the switch’ worries me quite a bit.
Is it still true that we are unable to obtain the serial number for the IB switches and the Cisco/Voltaire switches from the CLI of the switches themselves? Sad face.
I am trying to enable OLTP compression on tables and at tablespace level for the tables
Steps I am following are:
1. Move indexes to its own tablespace 2. enable OLTP compression at table level: alter table table_name move compress for OLTP 3. Rebuild indexes 4. Issue I have is what to do with tables with LOB columns ALTER TABLE lob_table MOVE LOB (LOB_COL) STORE AS (TABLESPACE index_tbsp); -- Is this correct? 5. alter tablespace data_tablespace default compress for OLTP;
I have a question, is the sequence of steps correct. For tables with LOB columns do we needto move lobindex to index tablespace. Beacuse lobsegment and lobindex are created in data tablespace?
I configure exadata + 2 switch between application server and and exadata. And we have problem to connect to switch and load balancing between Ethernet switch to exadata.
We might end up that is 'How about remove ethernet switch and directly connect to exadata?' I don't know which idea comes up with Ethernet gigabit switch between application server and exadata, exadata might be gigabit.
but from application server to ethernet switch is still 100M network, it is only fast in that load, other load is still 100m , it might useless. and about load balancing, exadata is RAC cluster node, exadata is doing load balancing, why we want another load balancing feature of ethernet switch?
Is it good idea to have dual switch between application server and exadata? exadata iti is what might be pro and con of this setting? exadata is handled by oracle support team, but dual switch is only cared by application operation team, so if something wrong with switch, nobody except operation team can manage .