Server Administration :: How To Read Dump Contents

Jan 4, 2012

How to read the treedump contents of a index IDX_TB_TEST_N1?

SQL> select object_id,object_name from dba_objects where owner='HXL';

OBJECT_ID OBJECT_NAME
---------- ---------------------------------------------
51786 IDX_TB_TEST_N1

alter session set events 'immediate trace name treedump level 51786'

/u01/app/oracle/admin/oracl/udump/oracl_ora_2679.trc
Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
With the Partitioning, OLAP and Data Mining options
ORACLE_HOME = /u01/app/oracle/product/10.2.0/db_1

[code]...

View 6 Replies


ADVERTISEMENT

Server Administration :: Contents Of Redo Log Files

Feb 1, 2011

I have some doubts about redo log files,

1) Can we fetch 'select statements' from redo log files through the use of log miner utility or any other?
(I think redo log file contains only insert,update,delete and DDL/DCL commands only)

2) If "No" to the above answer then how can i fetch all select statements fired on the system for a day or particular time.
(setting of sql_trace may be the one of them, but can it be possible for system level)

View 4 Replies View Related

Server Administration :: Cannot Audit Entry In Dump

Jul 26, 2012

I am trying to enable auditing as

SQL> alter system set audit_trail=OS SCOPE=SPFILE;

System altered.

SQL> STARTUP FORCE
ORACLE instance started.

Total System Global Area 171966464 bytes
Fixed Size 2019320 bytes
Variable Size 113246216 bytes
Database Buffers 50331648 bytes
Redo Buffers 6369280 bytes
Database mounted.
Database opened.

SQL> show parameter audit

NAME TYPE VALUE
------------------------------------ ----------- ------------------------------
audit_file_dest string /u01/app/oracle/admin/orcl/adu
mp
audit_sys_operations boolean FALSE
audit_syslog_level string
audit_trail string OS
SQL>

SQL> create user apexos identified by abc1;

User created.

SQL> grant connect, resource to apexos;

Grant succeeded.

SQL> audit select table, insert table by apexos by access;

Audit succeeded.

SQL> audit table by apexos by access;

SQL> SELECT audit_option, failure, success, user_name
FROM dba_stmt_audit_opts;

AUDIT_OPTION FAILURE SUCCESS USER_NAME
---------------------------------------- ---------- ---------- ------------------------------
TABLE BY ACCESS BY ACCESS APEXOS
SELECT TABLE BY ACCESS BY ACCESS APEXOS
INSERT TABLE BY ACCESS BY ACCESS APEXOS

cONN APPOS/ABC1

SQL> CREATE TABLE TAB1 (ID NUMBER, NAME VARCHAR2(20));

Table created.

SQL> insert into tab1 values (10, 'Michel');

1 row created.

SQL> insert into tab1 values (30, 'Andrew');

1 row created.

SQL> select * from tab1;

ID NAME
---------- --------------------
10 Michel
30 Andrew

SQL> /

ID NAME
---------- --------------------
10 Michel
30 Andrew

SQL>

SQL> select username, timestamp, action_name, action, SES_ACTIONs, sql_text
2 from USER_audit_trail where username='APEXOS';

no rows selected

SQL>

I also did not find any file contiaing the above statement as audit record in
/u01/app/oracle/admin/orcl/adump.

There are numerous old file in the /u01/app/oracle/admin/orcl/adump locaton. But When I executed the sql statement then that time no audit file was not generated in the location.

how to find audit record.

View 5 Replies View Related

Server Administration :: Core Dump Under $ORACLE_HOME / DBs

Oct 9, 2012

I saw big size .core file is generated in $ORACLE_HOME/dbs folder even when no dump dest parameter is set to dbs folder. How to check what causing this genrarating of these files.

PROD1 /oravl01/oracle/11.1.0.7/dbs > ls
1 core.11528 core.15973 core.22231 core.26792 core.29496 core.5262 initCCLCLMP.ora
controlfile.binary.PROD1.Fri core.12504 core.15981 core.2262 core.27129 core.29802 core.5847 initdw.ora
controlfile.binary.PROD1.Mon core.12520 core.1617 core.23115 core.27196 core.30793 core.6176 init.ora
controlfile.binary.PROD1.Sat core.1256 core.1626 core.23276 core.27531 core.30877 core.6221 initTESTCRM.ora
controlfile.binary.PROD1.Sun core.12594 core.16364 core.23681 core.27560 core.30885 core.6535 orapwPROD1

[code]....

View 7 Replies View Related

Server Administration :: User Dump / Trace Files With Huge Size In MBs

Mar 22, 2012

I am facing problem in user_dump_dest directory...I have noticed that there are a lot of trace files with huge size in MBs.I clean it and after 4 days there are 40G of size..

View 1 Replies View Related

Server Administration :: No Data To Read From Socket

Feb 5, 2010

I am getting an error through by application when i am trying to access a screen.

Error:No data to read from socket.

What could be the reasons for this error.

View 1 Replies View Related

Server Administration :: Fractured Block During Buffer Read?

Nov 17, 2011

I am in charge of several instances located on a Linux server CentOS, virtualized on a ESX 3.5 environment.

From time to time (every 4 to 5 days), I have some errors in the alert.log. Last occurence was last night :

Corrupt block relative dba: 0x01004e12 (file 4, block 19986)
Fractured block found during buffer read
Data in bad block:
type: 6 format: 2 rdba: 0x01004e12
last change scn: 0x0000.131aaa5b seq: 0x2 flg: 0x04
[Code] ........

We are doing user manual backup (with BEGIN/END BACKUP) every night at 8PM, ending at 9PM approx. Then, fractured blocks never occur during backups. At 1AM, the maintenance window is opening, thus explaining the GATHER_STATS_JOB job.

When I check corruption on early morning, I am always unable to reproduce the problem. DBV is OK without issues. We never had a problem with the data itself, whatever it is a table or an index in the reported failed block.

I would like to know what could cause these logical corruption, and how to stop them ?

View 7 Replies View Related

Server Administration :: Check Status Of Listener And Read Only From DB Query?

Jun 13, 2010

I need to display the parameter and status of DB for listener and Read Only.

I know those value could be get from command line , but could we get the values of Listener and Read only by SQL/PlSQL? So I can get it through the query of DB.

View 8 Replies View Related

Server Administration :: ORA-16000 / Database Open For Read-only Access

Apr 5, 2010

I am facing a strange issue on 11gR2 (OEL 5.4) standby readonly with apply database.It's throwing 16000: database open for read-only access during SELECT's .

Here is snapshot of errors.

ORA-00604: error occurred at recursive SQL level 1
ORA-16000: database open for read-only access

SQL> l
1* SELECT t0.airportID, t0.archived, t0.assetTag, t0.bluetoothID, t0.cmBundle, t0.createdDate,
t0.currentProductTaskID, t0.ethernetID, t0.failOrReworkCount, t0.highestCompletedTaskTypeID, t0.lastModDate, t0.lastStationID,
t0.modCount, t0.modelID, t0.oemSerialNumber, t0.orgSerialNumber, t0.pdmVersion, t0.preburnComplete,
t0.productID, t0.reworked, t0.secondaryEthernetID,
t0.serialNumber, t0.shipped, t0.specialBuildTypeID,

[code]....

View 7 Replies View Related

Server Administration :: Taking Tablespace Offline / Cannot Read Or Write

Nov 24, 2012

I have created an user named "Raja" with a default tablespace as "Raja_TBS" along with a datafile "rajadata.dbf". I have taken the tablespace offline

SQL> alter tablespace raja_tbs offline;

Tablespace Altered. when I take a tablespace offline, which means I cannot read or write and the tablespace is currently unavailable for users. I am still able to create a table on the "Raja_TBS" while it is offline.

View 39 Replies View Related

Server Administration :: ORA-27070 / Async Read / Write Failed

Jan 6, 2013

i have tried to create the index in diff tbs also but same error is there

SQL> create index inx_tbl_voicechat_unsub_ani on tbl_voicechat_unsub (ani) tablespace ideadb_index;
create index inx_tbl_voicechat_unsub_ani on tbl_voicechat_unsub (ani) tablespace ideadb_index
*
ERROR at line 1:
ORA-01115: IO error reading block from file 201 (block # 144265)
ORA-27070: async read/write failed
OSD-04016: Error queuing an asynchronous I/O request.
O/S-Error: (OS 23) Data error (cyclic redundancy check).
ORA-01115: IO error reading block from file 201 (block # 144265)
ORA-27070: async read/write failed
OSD-04016: Error queuing an asynchronous I/O request.
O/S-Error: (OS 23) Data error (cyclic redundancy check).

View 10 Replies View Related

Server Administration :: Ora-00333 / Redo Log Read Error Block 203 Count 8192

Apr 17, 2010

When i was starting my database .there was an error

ora-00333:redo log read error block 203 count 8192.

View 1 Replies View Related

Server Utilities :: How To Know Dmp File Contents

Sep 24, 2011

Is there a way to list contents of a dmp file(using expdp generate),which objects(tables/vies/procedures/packages) does it contain?

View 4 Replies View Related

Server Utilities :: Importing 9i Dump Into 10g?

Jun 22, 2011

how can i import the oracle 9i dump file into 10g database, while iporting i get following error imp-00002 fail to open dump file

View 4 Replies View Related

Server Utilities :: Reading From Oracle Dump

Jun 14, 2006

Is there is any way of reading from oracle dump file?

View 16 Replies View Related

Server Utilities :: Dump Table From One To Other Schema

Mar 4, 2010

i need to dump the table from A schema to B schema with different table name.

Suppose i have TABLE A IN "A" SCHEMA i need to dump the table with DATA+sTRUCTURE in " B"SCHEMA WITH TABLE NAME AS B.

View 3 Replies View Related

Server Utilities :: Dump File Determination

Mar 29, 2013

Is it possible to determine whether the dump file is created using data pump export or normal export method by just looking at dump file, If yes, how ?

Why i am asking such question is...normal export and data pump export would create a dump file with an same extension filename.dmp. So to avoid confusion during import, i would want to determine by what method the dump file was created.

Also this would be useful for me at the scenario when the customer sends me only the dumpfile and ask to import into target database. ( may be the customer don't know in what method the dump file was created ).

View 23 Replies View Related

Server Utilities :: How To Import Dump In Schema

Feb 7, 2011

I have exported a schema dump with schema name as 'A'.I want to import that dump in to schema 'B'.how ?

View 5 Replies View Related

Server Utilities :: FTP Dump File Over Network

Apr 19, 2010

I have A Daily hot backup using Expdp Command On oracle 10g R2 installed on the Linux server. And I'm trying to move this Dump File to Another directory on Windows server 2003 over network using Ftp script which will be run after the export process finished Automatically.

View 9 Replies View Related

Server Utilities :: Dump File Generation

Jan 18, 2012

I have a question on export dump file generation.

select sum(bytes)/(1024*1024*1024) "GB" from dba_segments where owner='JACK';

The above select query give the output of Schema size with 15 GB. When i perform the same schema export, the dump file size generating is 2 GB. What is the difference between the two scenarios as how come there could be a variation in file size?

View 6 Replies View Related

Server Utilities :: How To Reimport 9i Dump File To 10g

May 29, 2010

I am in the process of upgrading our 9i DB to 10g . As they are on different servers, I have installed 10g on the new server and applied the latest patchset 10.2.0.4.

I am creating the production database and importing th e9i dump file into this.Now I will be testing the whole application that uses this database.After a week, I need to take the latest 9i dump and export to the new 10g DB.

Do I need to just import the latest 9i dump into the 10g db or do I need to do anything else?

View 3 Replies View Related

Server Utilities :: Import Dump File In 11g

Feb 24, 2012

I am facing a problem importing DMP file in 11g. While importing it gives me error not responding. I have to attached the jpg file for that to clear you my point whats wrong is going during import. My Dump is on 9i i want to import that on 11G R2.

View 4 Replies View Related

Server Utilities :: Export Dump File

Jul 29, 2011

Is it possible to identify what level of export by looking at export dumpfile .. whether it is a schema export,full export,table export,..

If yes.. how ?

View 3 Replies View Related

Server Utilities :: Export Dump Of Large Table

Apr 9, 2010

We have two databases running on 10.2.0.4 and 9.2.0.8. Both are having the same unpartitioned table of size 80G. I am exporting the table on 10g by using parallel=8 and dumpfile with %U option. That took around 4 hours to export the table.

And on 9.2.0.8, i am exporting using below parameters, taking around 5 hours.

buffer=2000000
recordlength=64000

options i can try to speed up the export in both versions.

View 2 Replies View Related

Server Utilities :: Import Dump File Without 2 Tables

Jan 3, 2012

I want to import dump file (without 2 tables) .The dump file contains 100 tables,indexes and constraints. So out of 100 tables i want to import 98 tables from dump file (without 2 tables).

View 13 Replies View Related

Server Utilities :: Import Oracle 10g Dump Into 9i Database

Mar 31, 2010

If I want to import 10g export dump file in to the 9i database, I am connecting to 10g database from 9i database using
exp user/password@10gdb ....

However is there any option like executing 9i catexp.dat on 10g database and do the export from 10g database itself to be imported into 9i?

View 6 Replies View Related

Server Utilities :: Evaluate Export Dump Size

Jan 11, 2012

I want to take a schema level export .The schema size is 115 GB size . Do we require same amount of space to be available in server side (where we are taking a dump) as the schema size or less or more space is required in server side ?

View 6 Replies View Related

Server Utilities :: Procedures To Import A Dump From 9i To 11g Directly?

May 31, 2011

I tried to import a dump in 11g that was taken in oracle 9i. The import started but it hangs after some time. Exactly say it check only the character set of the DB's then it hangs. let me know if there are any specific procedures to import a dump from 9i to 11g directly.

View 8 Replies View Related

Server Utilities :: IMPDP (ORA-31619 Invalid Dump File)

May 29, 2012

I need to recreate/ clone my database to a new machine. The two machine are not connected in the network.

Step 1. (Oracle 10.2.0.5 AIX 64-bit)
expdp username/password@db1 full=y dumpfile=dp:fpac052912_dp%U.dmp logfile=dp:fpac052912_expdp.log job_name=full_exp

Step 2.
FTP dump files to Windows

Step 3. (Oracle 10.1.0.2 Windows 32-bit)
impdp username/password@db1 dumpfile=dp:fpac052912_dp%U.dmp logfile=dp:fpac052912_impdp.log full=y

I got:
ORA-39001: invalid argument value
ORA-39000: bad dump file specification
ORA-31619: invalid dump file "C:P7DBfpac052912_dp01.dmp"

Done in AIX:
create directory dp as '/bak'
grant read, write on directory dp to public;
grant exp_full_database to username;

Done in Windows:
create directory dp as 'C:P7DB';
grant read, write on directory dp to public;
grant exp_full_database to username;
grant imp_full_database to username;

View 8 Replies View Related

Server Utilities :: How To Export One Table To Few Separated Dump Files

Mar 24, 2010

I have a table with 9 regions. How can I export them in 9 seperated dump file?

View 4 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved