Despite it being one of the major selling points of Exadata (especially from X3 onwards), I'm struggling to find much information on our usage of the Exadata Smart FlashCache (I'm running RDBMS 11.2.0.2 BP7 on a V2 quarter-rack).
I can verify usage of the FlashCache by checking whether the object has been 'pinned' to the FlashCache via DBA_SEGMENTS and I can check for FlashCache usage by querying gv$sysstat (and even v$mystat), but are there other views that I could use? It seems a bit strange for Oracle not to provide the DBA all that much insight into their usage of this feature...
Design stage of New SAN. I can put a lot of the database, including redo on SSD.
1. Hardware sub tiering enabled. Has anyone ever gauged performance with enabling 11g Smart flash cache using the SSD as the flash device even with the hardware subtiering on? ie, if the hardware subtiering is moving hotblocks to most performant disk, is there much significant benefit having the flash cache on. Im guessing yes as the flash cache is working at the SGA level.
2. To redo or not to redo on SSD. Many notes say not to so as to avoid log file sync waits. Is there a way to keep it on the SSD and avoid the waits? Is there any risk of enhanced degradation of the disks over time with having so much sequential writes from the redo if I leave it on there?
How to Minimise Waits for 'Log File Sync' [ID 857576.1] (discusses keeping redo off ssd but only generally)
I have a table "tl" which is partitioned--say 30 partition and for each partition there is a seperate view like as follows
create view view_t130 as select * from tl partition (p30); create view view_t129 as select * from t1 partition (p29); . . . create view view_t101 as select * from t1 partition (p01);
my question is how to use hints on this table if your are using view to access the data from internal table.
Normal structure is if i don't wrong:-
index( <<view name|view alia name>> <<table name|table alias name>> name of index)
Consider my case
select * from view_t130 where <index_column> --not picking up index
i want to give expicit index hint.so i used the same structure that i specified above but it didn't work.
select /*+ index( view_t130 t1 <index_name) */ * from view_t130 where <index_name>
how to give explicit index hint..but one constraint is i cannot give any alias names for internal tables because those(view structure) are generated by predefined scripts..so it's not possible to change it.
One of my databases which is running on exadata x2-2 , has been restored to non-exadata machine in order to test few things. I had a sub-partition table in the exadata , compressed for query high. test machine (NOT EXADATA) , after uncompressing this subpartiton tables i am getting the following err message :
ORA-64307: hybrid columnar compression is not supported for tablespaces on this storage type I have executed the following commands : alter table crm.cm_ncd modify partition P01_CM_NCD nocompress; alter table crm.cm_ncd modify partition P02_CM_NCD nocompress; alter table crm.cm_ncd modify partition P03_CM_NCD nocompress;
[code]...
ERROR at line 1: ORA-12801: error signaled in parallel query server P005 ORA-64307: hybrid columnar compression is not supported for tablespaces on this storage type If all the partitions are uncompressed why i am getting this error message ?
At a customers site I see lot of long lasting library cache locks during a complex ETL run. Several Sessions run in parallel and create Database tables with dynamic sql ( CREATE TABLE AS SELECT ....) .
Sometimes these procedures wait for each other with wait event 'Library Cache' . I presume that this is a side effect of the dynamic DDL in the stored procedures. Is that possible even when the Procs create different Tables but reference the same tables in their SELECT clauses ?
I presume that this is plausible but I need some arguments to convince my colleagues.
I am having an Oracle 11g R2 RAC setup. My database size is 22TB in normal servers. I have checked my database in Exadata X2-2 and found out that the HCC was good and as per my assumption the 22TB can come down to 10TB .
My challenge is i need to transfer the 22TB data to the exadata quater rack which is having space constraint. Is there any way other than export & Import as data export and import will use entire 22TB. Is there any way to transfer the data to Exadata with compressed form from the source server.
I am having doubt on the no_index concept in oracle. I am using oracle exadata server.It is basically data warehouse project.I am in the situation to join some tables and get the result set for reporting purpose.
Among the tables, 2 tables having huge count.1st table has more than - 1,000,000,000 rows2nd table has more than - 200,000,000 rows when i join these 2 tables with some small set of tables, it is taking long time (around 20 MIN) to retrieve the result set. The final result set is around 100 rows only.
But, when i force NO_INDEX hint in the same query, is giving the same result in very fast manner(around 5 MIN). Because it is working based on cell smart scan. So, can i force the NO_INDEX hint to all tables? I forced the NO_INDEX hint only the table which is contain 200,000,000 rows not for others.
Query Plan : Normal Query using the the range scan based on the key. No_INDEX Query going full table scan.
1) Post Installation check list ie: How can I see every things are installed correctly. 2) upgrade the exadata OS version . 3) Upgrade the Database. 4) Test Migration from Oracle database to Exadata. 5) Right now there RAC servers are installed without domain, I need to add domain name to existing RAC.
I have a created a materialized view which is based on a view on remote database. Now how do I refresh the view.
Materialized view is created by
CREATE MATERIALIZED VIEW mv_employee_name AS SELECT EMPLID, EMPL_NAME FROM VEMPDATA@REMOTEDB WHERE REGION = 'US';
I am wondering how the refersh happens or how do I specify the refresh clause.REFRESH FAST option is looking for VIEW LOG on the master table but in this case its a remote view, so I cannot create any object on remote db.
I am removing sal column from table tab_emp; i want to check whether any materialized view or view using this column by querying using data dictionary :- if i use like condition against query column of all_mviews it is throwing error sicne it is long data type. is there a way to search it without creating any function and use it in a query.
I have to convert some existing materialized views (fast refresh) to partition materialized views.
Database version is oracle 10.1.0.4. I have decided to use on prebuilt table option to do the partitioning as it minimizes the time to transfer from the master site.
1) stop replication 1) create interim tables with similar structure as the materialized views 2) transfer all data from the materialized views to the interim tables 4) script out the materialized views structure and add in on prebuilt table option in the scripts 5) drop the materialized views 6) rename the interim tables with the same name as the materialized views 7) run the scripts to create the materialized views with on prebuilt table option 8) refresh the newly created materialized views -> it should take a short time since I am using on prebuilt table option
But I am facing one major issue. That is if I drop the materialized views, the materialized view logs of the master tables are purged. When the materialized views are refreshed fast, there are some data missing. the data that are purged out when the materialized view are dropped.
Do you happen to know other ways that existing materialized views can be converted to partitioned materialized views? Do you have any workaround to prevent the materialized view logs from being purged?
I am moving an Oracle DB (10.2.0.1) to a new server, by creating a blank DB and importing the data by user. I chose to create the new database with Flash Recovery activated (I haven't previously used this), but I'm having a nightmare when importing the data.
I encountered issues with the DB_RECOVERY_FILE_DEST_SIZE which I resolved by increasing the size of this parameter.
Next the import crawled along because of redo.log file errors "Checkpoint cannot complete". I'm assuming that I need to add redo.log files to accommodate this?
SYS@boston>select name, value, issys_modifiable from v$parameter where name='db _recovery_file_dest_size';
NAME -------------------------------------------------------------------------------- VALUE -------------------------------------------------------------------------------- ISSYS_MOD --------- db_recovery_file_dest_size 9663676416 IMMEDIATE
I just noticed on apex.oracle.com (4.2) that I couldn't find the region source for flash charts. I expected to find it under the Region Definition tab.
Has this moved or will it be included in a later release?
I'm trying to use 11gR2 Grid Control to manage several 10gR2 database targets. I'm unable to determine the correct mix of Adobe software plug-ins to use for displaying SQL execution plans in a more graphical style. My client OS is Windows 7 - I have tried to research plug-ins for Internet Explorer 8, I upgraded (and subsequently downgraded!) to IE 9; I installed Firefox 6.x, tried with Google Chrome and all with the same results.
I'm not seeing any relevant information from MyOracle support or on Google, but may not be asking the correct question. As far as I can determine, SVG 3.0 is needed, Adobe stopped development in 2009 and there is no variant available therefore for Windows 7.
The following query is used to generate a flash chart:
select null link 'Available' as "Available", to_number(max_licenses - consumed_licenses) from software_detail where capture_date = trim(GET_CAPTURE_DATE) and software_product = :p1 and software version = :p2 [code].....
But the 2d doughnut flash chart shows:
Available = 39 Consumed = 1
How is this possible?
Oracle 11.2.0.3 RAC on Windows 2008 R2 APEX 4.2.1.00.0 with Apex Listener 2.0 on Apache Tomcat 6.
Is there any way I could customize tooltip (when you go over the bar in the report, small box with the value appears).I would like to display different values there.