I have a package when it runs, it terminates with "unable to allocate 4120 bytes of shared memory"error at a particular INSERT statement. So everytime we had this error, we had to flush the shared pool area, after which it would run without issues(for 3-4 weeks).
It had a dynamic subpartition clause framed, which made it to repharse the query every time.So to reduce this, we removed the subpartition clause(made it to static query), which did benefit of not getting shared pool error.
I would like to find the current shared pool memory usage of my program.Is there a way I could find the live consumption of shared pool area of my program(pkg).
I am working on Tuning a Materialized view. I was getting an error saying ORA-12008: error in materialized view refresh path
ORA-12801: error signaled in parallel query server P002, instance hrms-stg-db01:HRSTG1 (1) ORA-01652: unable to extend temp segment by 16 in tablespace TEMP1.
so i increased the temp size. I am trying out various hints to get it rectified. But i am not sure if i should go for less cost or less bytes.
difference between V$ and V_$ views. It looks same to me. lets take the example of V$LOG and V_$LOG , both views returns same set of columns when described.
SQL> desc V$LOG Name Null? Type ----------------------------------------- -------- ----------------- GROUP# NUMBER THREAD# NUMBER SEQUENCE# NUMBER BYTES NUMBER MEMBERS NUMBER ARCHIVED VARCHAR2(3) STATUS VARCHAR2(16) FIRST_CHANGE# NUMBER FIRST_TIME DATE
SQL> desc V_$LOG Name Null? Type ----------------------------------------- -------- ----------------- GROUP# NUMBER THREAD# NUMBER SEQUENCE# NUMBER BYTES NUMBER MEMBERS NUMBER ARCHIVED VARCHAR2(3) STATUS VARCHAR2(16) FIRST_CHANGE# NUMBER FIRST_TIME DATE
I've migrated from Oracle 9i to Oracle 11gR2, when i'm checking my application on Oracle 11gR2, i found that some of the query result are differed from the previous version of Oracle 9i, which is very illogical to me. I've checked the data for all related tables which are involved in one of the query. I've checked all indexes also. Still i'm not convinced that Oracle is doing anything wrong. I cannot recreate the whole scenario here, because of many tables involved in that query.
One of the sample query code.
select 'ProductMaster' producttype,e.schemecd,e.SCHEMESHORTDESC as SCHEMEDESC,to_char(e.schemefrdate,'dd/MM/yyyy')as schemefrdate, to_char(e.schemetodate,'dd/MM/yyyy') as schemetodate,e.mkid ,e.countrycd,e.statecd,e.districtcd,f.schemetype, a.productcd,a.packsizecd,a.packtypecd,a.sortorder1, a.countrycd,a.statecd,a.districtcd,a.p_uniqueid,a.mrp, (a.productcd ||'-'||b.countryname||'-'||c.statename||'-'||d.districtname|| ' -pkt.prc:'||a.pktprice ||' -mrp:'||a.mrp) as pcode from wb.wbproductdetails a
We want to find out difference of data for some tables between current day & previous day. We can use query with minus operation but it will take lot of time since table size is in range from 200 to 500 GB. We have to do this exercise every day.
I am working on a Script in Perl where i have to find out which Tablespaces are autoextensible and how many space is in use at this spaces - i've found a Standardscript and wanted to customize this with this Autoextensible-clause - but it doesn't work...here ist my script to find out the raw Data.
Datafile name is +DISK_GROUP_1/oratst1/datafile/system.1122.764387443
displayed for the SYSTEM tablespace in dba_data_files.Infact, datafile name starts with +DISK_GROUP_1 for tablesspaces. Where is the reference to +DISK_GROUP_1.
What file name should I use to resize or to add new datafile to a tablespace?
I've a question regarding difference of character sets, while taking a export(logical backup) of database on directly to server(linux RHEL 2.1 AS) and export on a client (windows xp prof machine, where only a oracle 9i client is installed). On server it seems to fine and okay, but on client node i'm getting following error for almost all tables.
EXP-00091: Exporting questionable statistics.
My question is :
[1] Is it creating any sort of problem, if later on i import the data which was taken from client node.
[2] Why there is a difference(marginal) in dump(.dmp) file size.
[3] Is there any way to overcome it, or it is the natural behave of it. Means not a problem.
[4] If i'm using a long or blob as datatype for some of my table,is they have any problem if i persist like above.
Additional Information about character sets On server node :
Export done in US7ASCII character set and AL16UTF16 NCHAR character set server uses WE8ISO8859P1 character set (possible charset conversion)
On client node :
Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set server uses US7ASCII character set (possible charset conversion)
When i try to extract create statement of table i get below error message.
SQL> SQL> set heading off SQL> set pagesize 0 SQL> set long 1000000 SQL> set feedback off SQL> select 2 dbms_metadata.get_ddl( 'TABLE','CATALOG_TBL','JACK') 3 from 4 dual 5 /
ERROR:
ORA-04031: unable to allocate 104 bytes of shared memory ("shared pool",
"SELECT /*+rule*/ SYS_XMLGEN(...","SQLA^fc8f5280","qecsub : qkxrPXformQbc") ORA-06512: at "SYS.DBMS_METADATA", line 2625 ORA-06512: at "SYS.DBMS_METADATA", line 2668 ORA-06512: at "SYS.DBMS_METADATA", line 2983 ORA-06512: at "SYS.DBMS_METADATA", line 3897 ORA-06512: at "SYS.DBMS_METADATA", line 5678 ORA-06512: at line 1
Note: ---- I am executing this command from another user 'DEMO' which has got DBA privilege.
how to fix this error and extract the create table statement.
I am trying to build an XML document in a CLOB PLSQL variable. We are using Oracle 11gr2 database.
But when I am reaching more than 32767 bytes my code is failing.
Is there anyway we can store more than 32767 bytes of data in a PLSQL variable of type CLOB.
I am capturing the below error message
(ORA-06512: at "SCMSA_HIST.SCMSA_POC_HANDSET_MOBILITY_PKG", line 1480 AND LENGTH OF xmlfile -> 33078 )
I am adding my code also here for further clarification
PROCEDURE GET_HANDSET_DATA_PRC (p_ntlogin_id IN VARCHAR2, p_handset_data OUT NOCOPY CLOB) IS /****************************************************************************** NAME: GET_HANDSET_DATA_PRC PURPOSE: Date Ver By Description ---------- --- --- -----------
******************************************************************************/ CURSOR c_region_data IS SELECT NVL2 (T.ntlogin, T.ntlogin, pos.ntlogin) AS ntlogin, NVL2 (T.first_name, T.first_name, pos.first_name) AS first_name,
Recently i've migrated from Oracle 9i SE(9.2.0.1.0) to Oracle 11gR2 EE(11.2.0.1.0). Previously i'm taking export of some of my schema and it's file size was around 1g.(with exp utility of Oracle 9i). As per earlier practice now i'm taking export of same schema with same no of objects and same data volume, the size of export file size on Oracle 11gR2 database is significantly gone down , actual size around 825mb(with expdp utility of Oracle 11g).
So i would like to know why there is a difference in file size(.dmp files) of export files between two oracle versions. I have crosschecked objects and rows of data tables. It is perfectly same.
Currently I am doing one Oracle installation.But very much confused where to start from. Since I am installing oracle on my personaly Laptop,do I need to install Oracle database server and Oracle client both or only installing database server is sufficient.
What is the exact difference between oracle database server and client.Only knowledge I have is that client is used to access database server remotely.are Oracle database server and client installable to be downloaded separately?
I am seeing the following error message in alert.log (ORA-04031: unable to allocate 32 bytes of shared memory ) and my database goes into hung state and wont let any new sessions to connect and old sessions just hung with no response. i have to kill the instance to come out of this problem.
I am running 11.2.0.3 on AIX and have memory_target=35GB. I would like to know what approach to take to find out the root cause of this? btw - does AWR show if memory was under pressure and what SQL's were consuming more memory just before the problem surface?
I have a stored procedure which uses BULK COLLECT and the table has 16 nested tables within it. When I limit the number of rows processed it works fine, if I let it run with all the 10,000 rows I get this:
ERROR: ORA-04030: out of process memory when trying to allocate 16408 bytes (koh-kghu call ,pmuccst: adt/record)
Single Laptop Machine of 16GB RAM. Assigned 13GB to VM.
I have installed Oracle Database 11g. 11.2.00. When I try to access my application many times I received below error
• Error: “”Caused by: java.sql.SQLException: Could not retrieve datasource via JNDI url 'jdbc/oimJMSStoreDS' weblogic.jdbc.extensions.ConnectionDeadSQLExceptio n: weblogic.common.resourcepool.ResourceDeadException : Could not create pool connection. The DBMS driver exception was: ORA-00604: error occurred at recursive SQL level 3 ORA-04031: unable to allocate 352 bytes of shared memory ("shared pool","select /*+ rule */ bucket_cn...","SQLA^337fc737","qertbs:qertbIAlloca te") [code]......
Even when above error occur I have 7GB free JVM so JVM is not an Issue.
•I have updated Max Capacity of oimOperationsDB =100,oimOperdationDB, oimJMSStoreDS=80,SOADATAStore =80. But still I have error
• ERROR:: • ““Aug 30, 2012 3:24:10 PM EST> <Warning> <oracle.integration.platform.blocks.event.saq> <SOA-31013> <Error handling message (rolling back). java.sql.SQLException: ORA-04031: unable to allocate 352 bytes of shared memory ("shared pool","select /*+ rule */ bucket_cn...","SQLA^337fc737","qertbs:qertbIAlloca te") ORA-06512: at "SYS.DBMS_AQ", line 335 ORA-06512: at "DEV_SOAINFRA.EDN_DEQUEUE_OAOO_DELIVERY", line 14” [code]......
I have modified Shared_pool_size,large_pool_size. Content of my modified Pfile is as below.
o orcl.__db_cache_size=822083584 o orcl.__java_pool_size=2777216 o orcl.__large_pool_size=56777216 o orcl.__oracle_base='/home/oracle/app/oracle'#ORACLE_BASE set from environment o orcl.__pga_aggregate_target=805306368 o orcl.__sga_target=1174405120 o orcl.__shared_io_pool_size=0 o orcl.__shared_pool_size=601989888 o orcl.__streams_pool_size=0
What’s value of parameter I need to set for datastore, pfile ? (I have 13GB of memory).
Windows 2003 Server Enterprise Edition, Service Pack 2. 3.5GB RAM.Two Oracle Databases running on this server, one not heavily used. On the database i am concerned with it is 10.2.0.3.0.
I have just started to look at this database and am seeing various things in the alert log and have had user error messages as well.Initially when i started looking at this server it was running with quite a small SGA (600MB) and the memory_target_advice was saying to take it to 1000MB.(PGA is around 200MB)
In January, before i looked at it, it started to throw:
ORA-04031 unable to allocate 16 bytes of shared memory ("shared pool","........", "sql area", "tmp") These would occur regularly and mostly during nightly processes, it seems, and thus not bubbling up to users.Having switched the SGA_MAX_SIZE to 1GB a new set of errors starts to appear in the alert log:
ORA-27301: ..... Not enough storage is available to process this command ORA-27302: ..... ssthrddcr
At the same time the ORA-04031 errors start to bubble through to the users. Before i could return to the scene of the crime someone else set the SGA back to 600MB and the error messages stopped. However the ORA-04031 message start to reappear in the alert log. My questions are many and, mostly, from a lack of knowledge ;) In particular:
1. Why would increasing the SGA lead to a different memory error? I would assume this would have something to do with the amount of memory a 32bit OS will allow a single process to address but the PGA+SGA is only 1200MB.
2. How does a ORA-273xx error end up with a ORA-04031 being shown to the users? I would look at the listener log but its over 2GB!!!
We had the following error in alertlog repeatedly from 5 PM to 10 PM :
ORA-04031: unable to allocate 32 bytes of shared memory ("shared pool","SELECT VERSION FROM PSVERSIO .....Since MEMORY_TARGET is defined in spfile then shared pool is managed automatically.