I have created an user named "Raja" with a default tablespace as "Raja_TBS" along with a datafile "rajadata.dbf". I have taken the tablespace offline
SQL> alter tablespace raja_tbs offline;
Tablespace Altered. when I take a tablespace offline, which means I cannot read or write and the tablespace is currently unavailable for users. I am still able to create a table on the "Raja_TBS" while it is offline.
We have a table (call it PROBLEM table) that varies between 60 and 100 million records from one data archival process to the next.The table has data from all 50 states and is range partitioned by each states unique code and has a primary key based on the following:
STATE the data is from BMRK YEAR the year the data was last benchmarked AREA location the data is from SERIES industry the data is located in YEAR the data was reported MONTH the data was reported DATA TYPE the type of data that was reported ESTIMATE TYPE the type of statistical estimate that was produced from the data REPORT ID report id unique to each state REPT SFC state the reporter resided in
We have developed a job that will process data from other tables and create some statistics resulting in approx 2.2 million inserts, .5 million updates and .5 million deletes on PROBLEM table during each job run. Once this table is loaded another process (call this PROBLEM_PROCESS)takes off that reads this updated information to produce some statistical data that is stored in another table. This data is produced by summing anywhere between 2 and 5000 records per query from the PROBLEM table.Here is the query (call it PROBLEM_QUERY)
select round(sum(cm_value*sample_weight),3),round(sum(pm_value*sample_weight),3) from matched_sample where state_fips_code = :1 and area_fips_code = :2 and series_type_code = 'B' and year = :3 and month = :4 [code]...
When we run this job it is very fast (20 minutes) until it gets to the PROBLEM_QUERY, at this point we get DB file scattered read waits, DB file sequential read waits and async IO waits and the job has run for as long as 2 days. When I look at grid, the PROBLEM query is doing a full table scan and ignoring the index.
Our systems people advised us to re-organize and re-index that data in the PROBLEM table after all of the DML and before the PROBLEM_PROCESS takes off. If we do this, it works and cuts our time down alot. However, the re-organize and re-index process adds eight hours to this process. Note: When we re-roganized the data we create a copy table on another set of table spaces and insert the data into the copy sorting the data on the primary key columns. After this we re-index the new table and drop the old table renaming the new table. When we do this later, we move the data back to another set of table spaces. So, we move the data back and for between table spaces A and B we can say.
We have Oracle installed on a M5000 server with Solaris as the OS the binaries and data files are stored on a NetApp Storage array(model 3160) of 500 1TB SATA 7200 RPM drives. However, there are 128 other databases on the NetApp filer as well.
IMO the array and the slow disks are the problem. I believe this because we are catering to the slow disks by re-organizing and re-indexing the PROBLEM table during each run. I don't believe this should be neccessary. We normally re-organize and re-index our data each week in our production system after many more transactions than this.
Our systems people state it is our application. Oracle Support tells us the statistics are out of date but have not answered us on why the statistics are out of date and the index is abandoned after 1 run.
I want to export all db using data pump. I got this error when using it:
Export: Release 10.2.0.4.0 - Production on Thu Nov 25 11:46:48 2010 Copyright (c) 1982, 2007, Oracle. All rights reserved. Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options Export done in US7ASCII character set and AL16UTF16 NCHAR character set server uses WE8ISO8859P1 character set (possible charset conversion)
About to export the entire database ...
. exporting tablespace definitions EXP-00068: tablespace TISDATA is offline EXP-00008: ORACLE error 1187 encountered ORA-01187: cannot read from file 203 because it failed verification tests ORA-01110: data file 203: '/u03/app/oracle/product/10.2.0/oradata/unidev/TISTEMP01.dbf' EXP-00000: Export terminated unsuccessful
Trying to learn about Oracle Database Mobile Server.Installed the server, installed the client on WIN 32, created a publication and successfully synchronized.Now need to write a client application for mobile users.
Qs 1) Can I write a simple JSP that reads a table from device SQLITE database?or, the JSP needs to read off a subscrption?If it needs a subscription as the source, then how do I mention it in the JSP?
I want to know if there are any freeware's available to read and write .sql files. I hava a .sql file with set of insert statements and I need to identify the values that are pushed into table, the table name, the inner query clauses etc., from the sql statement.
After identifying, need to translate and push those changes back to the .sql file. Is there any way to do this with parsers rather reading the whole of file line by line.
I did come across few parsers like ZQL, sqljep etc
I want to read and write data from a file in | seprated mode. file READ and WRITE using Oracle Forms6i. I have a knowledge of file handling in C++ but not use it in Oracel Form before this.
I will have to provide read/write access to an oracle user. What privileges should i grant to the user so as to enable the user to read from and write to the files?
alter session set events 'immediate trace name treedump level 51786'
/u01/app/oracle/admin/oracl/udump/oracl_ora_2679.trc Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production With the Partitioning, OLAP and Data Mining options ORACLE_HOME = /u01/app/oracle/product/10.2.0/db_1
Is it possible to create trigger on the various tables and views exists (i.e. dynamic performance views) in data dictionary, when ever any DML operations performs by Oracle it self?
I am in charge of several instances located on a Linux server CentOS, virtualized on a ESX 3.5 environment.
From time to time (every 4 to 5 days), I have some errors in the alert.log. Last occurence was last night :
Corrupt block relative dba: 0x01004e12 (file 4, block 19986) Fractured block found during buffer read Data in bad block: type: 6 format: 2 rdba: 0x01004e12 last change scn: 0x0000.131aaa5b seq: 0x2 flg: 0x04 [Code] ........
We are doing user manual backup (with BEGIN/END BACKUP) every night at 8PM, ending at 9PM approx. Then, fractured blocks never occur during backups. At 1AM, the maintenance window is opening, thus explaining the GATHER_STATS_JOB job.
When I check corruption on early morning, I am always unable to reproduce the problem. DBV is OK without issues. We never had a problem with the data itself, whatever it is a table or an index in the reported failed block.
I would like to know what could cause these logical corruption, and how to stop them ?
I issued a SHUTDOWN IMMEDIATE command on my database and for over 2 hours and got no response. I did this because I wanted to restart the database due to many hanging sessions which had led to high CPU usage on the server by oracle process.
Oracle Service<dbname> instance using the Windows Services screen which failed with an error 1053 message. At the same time I got an ORA-01013 message saying user requested cancel of operation. See below:
SQL> SHUTDOWN IMMEDIATE ORA-01013: user requested cancel of current operation SQL> SHUTDOWN IMMEDIATE ORA-01089: immediate shutdown in progress - no operations are permitted SQL> SHUTDOWN ABORT ORA-24324: service handle not initialized ORA-01041: internal error. hostdef extension doesn't exist
What I want to know is how do I proceed? Also how do I check to see the status of the Shutdown operation?
I have installed executable only (database option only) when i have installed oracle11 that time it is not asking any option for password. So is there any option to login in this type of condition?
I tried more to write Bangla in Oracle 10gR1, or 10gR2 or Express but failed. When I write bangla in the "emp_name" field for employee name this can be save, But when I execute to see the data again then the data comes with a big change like some symbol for some character.
I need to display the parameter and status of DB for listener and Read Only.
I know those value could be get from command line , but could we get the values of Listener and Read only by SQL/PlSQL? So I can get it through the query of DB.
I am facing a strange issue on 11gR2 (OEL 5.4) standby readonly with apply database.It's throwing 16000: database open for read-only access during SELECT's .
Here is snapshot of errors.
ORA-00604: error occurred at recursive SQL level 1 ORA-16000: database open for read-only access
When I am issuing below command from auxiliary database, getting ORA-00554.My tnsping is okay from both the side, remote login is exclusion. Password file is also okay.I ran UTLRP.sql twice to compile invalid packages.
From Auxiliary Database [ashish@localhost dbs]$ rman target sys/oracle@suman auxiliary / Recovery Manager: Release 10.2.0.1.0 - Production on Thu Apr 11 19:04:03 2013 Copyright (c) 1982, 2005, Oracle. All rights reserved.
From alert log, <msg time='2010-09-03T11:37:46.548+06:00' org_id='oracle' comp_id='rdbms' client_id='' type='UNKNOWN' level='16' host_id='DC-DB-01' host_addr='192.168.100.101' module='' pid='24310'> <txt>Error 12160 received logging on to the standby
[code]...
From trace file,
[oracle@DC-DB-01 ~]$ less /u01/app/oracle/diag/rdbms/bddipdc/bddipdc1/trace/bddipdc1_arcq_24310.trc Redo shipping client performing standby login OCIServerAttach failed -1 .. Detailed OCI error val is 12160 and errmsg is 'ORA-12160: TNS:internal error: \ [code]...
I have 3 sites which want to be updated as soon as the other one updates its data.(in 5 seconds). So I have chosen multi master solution,sync regardless of the resource consideration.
1-what is the difference between multimaster async and updatable materialized view regardless of resource consuming in multimaster?because in both of them each site can receive and also propagate the data...I just know that in materialized view we can work offline and then apply it manually is there any other thing?
2-I can not understand the difference between multimaster sync and async (considering the 5 seconds,because 5 second is a little time)I mean in this case are they different from each other or not.
In statspack I find a lot of wait events "ARCH wait on SENDREQ". The database is configured with the LGWR ASYNC attribute. In the documentation I can read that the "ARCH wait"-events may happen in case the database is configured with the ARCH attribute.How is it possible I get these "ARCH wait"-envents and do these wait events have any effect on the "on line" performance of the database? Is it possible that a user that saves data in the database has to wait for thest "ARCH wait"-events?
i have files a.text which is parent file and another one is child one called file b.txt . Both files are linked together by common field called "id". Interesting part child file have multiple layers name associated with ids. (we are only aware that in b.txt for each id there could be max 3 layers)
So they needs to get loaded into Table called PARENT_TBL
So PARENT_TABLE looks like ID NAME SUBJECT LAYER LAYERNO
My loader start and say commit reach logical records 8 as there are 8 records but donot load them and write them into badfile. there is no logs file generation happing so unable to trace.