How To Apply Patchset With No Downtime
Sep 20, 2012How to apply patchset (10.2.5) with no downtime?
My db version is 10.2.4 and OS is hp-ux
How to apply patchset (10.2.5) with no downtime?
My db version is 10.2.4 and OS is hp-ux
We are trying eliminate/minimize the downtime for our application. As part of new code deployments sometimes we need to modify DB Structure also. As it is taking time to backup current DB and apply new DDL, the application is down.
Is there a way to eliminate the downtime, if I can leverage Data Guard, Golden Gate or RAC concepts?
Is there any way/ tool to migrate from oracle to sybase with Zero Downtime?
View 2 Replies View RelatedIs there any way I can find out what caused the database to crash; either a history of commands executed within the database, I lost my bdump directory before the scheduled backup ran and the only logs available are after I re-created the directory.
SQL> startup
ORA-00444: background process "PMON" failed while starting
ORA-07446: sdnfy: bad value '' for parameter .
SQL>
[code]....
We are facing a project where it is mandatory that the migration (from 9i to 11g) happens without any downtime. We thought about using Goldengate do to this migration. But i would like to listen to somebody who already did such kind of migration (i never used goldengate before). The basic steps to do such migration would be:
1) Install the Goldengate client on both source and target
2) Export only the metadata (structure of the table, for example) from source to target (here is one point of doubt of mine. This export can be only done using exp/imp?)
3) Perform the initial load from source to target (here i have another doubt: It it possible to perform an initial load from a whole database?)
4) Configure manager, export and replicat to perform the migration with the source database open in read-write
With the steps above, would i be able to perform a migration without downtime? What other considerations do you have?
I have one query regarding ORACLE 10g RAC.
I have installed 10gRAC [10.2.0.1] on VMware.
now I have requirement to upgarde it this existing RAC setup to 10.2.0.4.
options for me the steps or sequence I need to follow in upgrade path? like first needs to upgarde crs,then fro ASM and lastly for DB something like this.
Platform: windows 64 bit
Oracle database 11.2.0.3
I successfully upgraded my database from oracle 11.2.0.1.0 to 11.2.0.3.0 and the upgrade worked fine. I installed the 11.2.0.3.0 in a new oracle home.
Now my task is to install the latest patch set release on the new oracle home(Make database upto date with latest fixes indeed). I am confused with the available documentations in my oracle support.
As CPU is now considered as a subset of PSU patch, i believe i have to apply the latest PSU only.Is this correct? Is there a step by step guide for applying latest patch?I saw there are 5 PSUs released (latest 11.2.0.3.5), do i need to apply each of them or just this final one?
Give me the latest available patchset/patch(Still not clear) install guide and latest available patch download links.
1)What is the use of inventory?
2) Global inventory is found by checking the oraInst.loc file. How to find the location of Local inventory?
3)Why is every patchset(11.2.0.2) now a full release?
Oracle version : 10g, 11g (applicable to both) Consider a schema named 'UNIVERSE' present in the database A running on Linux platform where this schema needs to be moved to another database B running on windows platform or AIX platform with no downtime provided and data needs to be consistent. Is this practically possible?
View 13 Replies View Relatedwhere I can download patchset to upgrade my database from 10.2.0.3 to 10.2.0.5?
View 1 Replies View RelatedI am testing upgrade from 11.1.0.6 to 11.2.0.2 patchset on linux X86-64 version. And facing the following component related issue.
The following components cannot be upgraded as the corresponding upgrade scripts are missing:
OLAP Analytic Workspace ( /u01/app/oracle/product/11.2.0/dbhome_3/olap/admin/apsdbmig.sql )
OLAP Catalog ( /u01/app/oracle/product/11.2.0/dbhome_3/olap/admin/amddbmig.sql )
Spatial ( /u01/app/oracle/product/11.2.0/dbhome_3/md/admin/sdodbmig.sql )
Oracle OLAP API ( /u01/app/oracle/product/11.2.0/dbhome_3/olap/admin/xoqdbmig.sql )
If you want to upgrade these components, you must first install them, using the OUI, and then re-run the Upgrade Assistant. Do you want to continue with upgrade? Note: If you continue, the above components will not be considered as part of the upgrade process.*
Following is the status of the specified components from 11.1.0.6
COMP_ID COMP_NAME VERSION STATUS
AMD OLAP Catalog 11.1.0.6.0 VALID
SDO Spatial 11.1.0.6.0 VALID
APS OLAP Analytic Workspace 11.1.0.6.0 VALID
Should I install these components while installing 11.2.0.2 patchset prior upgrading? I was unable to find any component installation on OUI.
Also DBUA pre-upgrade asked to remove existing UltraSearch from 11.1.0.6 prior upgrading. Though I removed it, but not sure is that going to hamper the upgraded DB, as I'm not sure of the usage of this component as well.
Know the Patch set number to upgrade a Oracle 10.2.0.1 to 10.2.0.5 on Solaris 10 Spark 64bits?
View 1 Replies View RelatedDoes we need to apply the patches in sequence. i installed oracle 11g r2 today only, then which patch i can install. if i like to install CPU, then does i need to install all the previous or only the latest one. if only latest one only, then how can i find the CPU patch number.
of course, instead of cpu i can install psu patch, then which one has to install. only latest one or all previous also. if latest one only how to find latest psu patch that has to apply.
I need to migrate a 500 GB 10.2 database from windows to linux. Of course I found Alejandro Vargas paper on how to do that ("Database Migration From Windows to Linux Using RMAN"). My problem is that the source database will be up again for 7 days and I need to apply all changes made to the source database to the migrated linux database.
Can I just take all redo log files generated from the source database and apply those to the destination database? I might have a problem with a) file conversion and that the redo log are not in sync with the target database (different check points?).I just found that on an exotic website: "The contents of a redo log file depend on a combination of Oracle version, operating system and server architecture. In general redo logs written on one architecture cannot be read on another. There are a few exceptions to this rule. For example, in Oracle 10.2 a redo log written in Linux can be read by a Windows database."
So if this is true, does it also work the other way AND will the database migration from window to linux create a perfect clone of the database where those redo log files actually can be applied to?
We need to apply the the 11.2.0.2 database server patch and this is a 12i Ebusiness suite env. Do we need to take any precautions(other than stopping the apps) on the APPS side when we are applying the patch on the database?
View 2 Replies View RelatedI have created Replication process for a single table using the below mentioned script.
connect sys/afccv@afccv as sysdba
show parameter open_cursor
create user STRMADMIN identified by STRMADMIN;
ALTER USER STRMADMIN DEFAULT TABLESPACE USERS TEMPORARY TABLESPACE TEMP QUOTA UNLIMITED ON USERS;
GRANT CONNECT, RESOURCE, AQ_ADMINISTRATOR_ROLE,DBA to STRMADMIN;
execute DBMS_STREAMS_AUTH.GRANT_ADMIN_PRIVILEGE('STRMADMIN');
connect sys/vxmldb@vxmldb as sysdba
[code]....
According to Alert log file Replication process started( Refer Alert log content below
Wed Oct 05 12:45:53 2011
Streams CAPTURE CP01 for STRMADMIN_CAPTURE started with pid=256, OS id=8692
Starting persistent Logminer Session with sid = 41 for Streams Capture STRMADMIN_CAPTURE
Wed Oct 05 12:46:20 2011
LOGMINER: Parameters summary for session# = 41
[code]....
Now When I am checking the apply status then it is showing Dequeue message
SQL> l
1* select apply_name,state,DEQUEUE_TIME from V$STREAMS_APPLY_READER
SQL> /
APPLY_NAME STATE DEQUEUE_TIME
------------------------------ ----------------- --------------------
STRMADMIN_APPLY DEQUEUE MESSAGES
when I am checking the number of rows on both tables (source and destination) , at destination it is not applying any thing.
How To Apply Group by Command in Forms 6i at block level for example in data block a field name is code,quantity and code replicate many times it is not working with out commit but i want to calculate sum of quantity in data block by code group before commit
if we apply this command at sql level then command is
select code,sum(nvl(quantity,0)) from sal_detail group by code and its work very well
I will begin to install Oracle Clusterware (11gR2), but I need to know how much space should I apply for the ASM instance? What size should I apply the shared disk to the nodes?
View 5 Replies View RelatedCan we apply LPAD in NUMBER datatype?
View 11 Replies View RelatedMy setup is following:
CODEhost22 (Oracle Linux 5.9) | host30 (Windows XP) with SQL Server Express 2008
DB22 | Gateway to MSSQL | DB30
Local Local | |
Capture --> Apply ------> Gateway -----------------> DUMMYUSR.DUMMYTBL
Process Process | |
Capture process on DB22 captures changes made to local DUMMYUSR.DUMMYTBL table, enqueues it to capture queue and propagates to apply queue for local apply process, which, using database link DB30 supposed to apply these changes to table DUMMYUSR.DUMMYTBL on remote SQL Server database DB30.
The problem is that when apply process is executing transaction, it aborts with error:
QUOTE ORA-26713: remote object does not exists or is inaccessible.
ORA-26712: remote object is "DUMMYUSR"."DUMMYTBL"@"DB30".
In trace log I can see the following message:
QUOTE kniNoPrimaryKeyError: ORA-26713, remote obj doesn't exist,
DUMMYUSR.DUMMYTBL@DB30
[code]...
what can cause ORA-26713 error of apply process applying changes to SQL Server database through Oracle Database Gateway, if object to be modifyed exists and accessible from Oracle database for strmadmin user?
By the way, database link and tables were created with following commands: CODEDB22(Oracle): (connected as strmadmin) create database link DB30 connect to "dummyusr" identified by "dummypwd" using 'DB30';
DB22(Oracle): (connected as dummyusr)
create table DUMMYTBL (no number, name char(6), ddate date, payload varchar2(2013));
DB30(MSSQL): (connected as dummyusr)
create schema DUMMYUSR authorization dummyusr;
create table DUMMYUSR.DUMMYTBL (no float(53), name char(6), ddate datetime2(0), payload varchar(2013));
i have attended an interview there i got a question like--->There are 100 archives generated in standby in that a series of five archives are missing and those are not available in the primary then how will you do.....
View 2 Replies View RelatedHow can i apply composite key on a table?
View 2 Replies View RelatedI want to upgrade 10.2.0.1 to 10.2.0.5.
Can I straightly apply 10.2.0.5 or need to sequentially apply one by one?
can we apply 11.1.0.7 Patch Set on a 10.2.0.3 database server? i mean are patch sets cumulative or do i have to install the patch specific to my release the issue is i have the below error with rman: RMAN-03009: failure of delete command on ORA_DISK_1 channel at ORA-19633: control file record is out of sync with recovery catalog
View 7 Replies View RelatedI want to apply divide and conquer approach on the field value, For example I have the field value like this
Value
32,36,12,65,85,9663
43,36,89,65,112,9663
36,63,12,65,95,123
32,36,12,85, 85,9663
Remember that these value are taken from ONLY one field i.e Value. Apply the divide and conquer approach in such a manner that take the value of a single field and each row and divide the value into two half. Continue this process until single value until we reach the single value or single sub problems.
Apply the divide and conquer approach on all the value of field. When we get final value then match the value of two different rows. And check how much field value are similar (match or duplicated). If field values are match more then 80% then delete the duplicated value and keep the original entity.
I have two tables like below-
CREATE TABLE prod_vendor_record (
vendor_record_seq_no NUMBER NOT NULL,
study_seq_no NUMBER NOT NULL,
vendor_subject_seq_no NUMBER NULL,
control_dataset_seq_no NUMBER NOT NULL,
checksum NUMBER NOT NULL,
processing_flag VARCHAR2(1) NULL,
[code]....
and executing below query on those tables-
insert into prod_temp_vendor(vendor_record_seq_no,checksum,rownumber,transaction_type,iu_flag)
select vr.vendor_record_seq_no, tvr.checksum, tvr.rownumber, tvr.transaction_type, 'U' from
prod_vendor_record vr, prod_temp_vendor_record_20000 tvr
where vr.study_seq_no=25707
and vr.control_dataset_seq_no=3910
and vr.key_hash=tvr.key_hash
and dbms_lob.compare(vr.key_col_val, tvr.key_col_val) = 0
and tvr.error_flag is null;
let me know on which columns of PROD_VENDOR_RECORD table to apply index to make processing faster. As I tried to build index like below-
CREATE INDEX idx_prod_vendor_record
ON prod_vendor_record (
study_seq_no,
control_dataset_seq_no,
key_hash
)
/
But it is not being used by above query (see execution plan)
--------------------------------------------------------------------------------------------------------------------------------------
| Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | TQ |IN-OUT| PQ Distrib |
--------------------------------------------------------------------------------------------------------------------------------------
| 0 | INSERT STATEMENT | | | | 5168 (100)| | | | |
| 1 | PX COORDINATOR | | | | | | | | |
| 2 | PX SEND QC (RANDOM) | :TQ10002 | 212 | 445K| 5168 (1)| 00:01:03 | Q1,02 | P->S | QC (RAND) |
[code]....
Oracle 11.2.0.3.0
I have a primary database orcl and logical standby database orcl_std.
Real time apply is enabled. I have standby redologs in both primary and standby sides and I`ve started recovery with below command:
ALTER DATABASE START LOGICAL STANDBY APPLY IMMEDIATE;
When I create a new table in primary database, I am unable to see it on standby database (Although real time apply is enabled) However, when I switch log in primary, I can see the new table in standby database.
My question is, why realtime apply is not working in my scenerio ? I was expecting to see the new table immediately in standby database once it is created in primary database. Why am I supposed to wait for log switch in real time apply ?
,I am trying to set up Streams on a 11.2.0.3 on a Windows 2008R2 server. Due to an error in running propagation, i a, trying to delete both Capture and Apply queues. I have deleted the Apply queues but unable to drop the Catpure,
SQL>EXEC DBMS_APPLY_ADM.STOP_APPLY(apply_name =>'LAO_NLPG73_BLPU_APPLY'); SQL> select * from dba_apply; LAO_NLPG73_BLPU_APPLY NLPG73_BLPU_APPLY_Q STRMADMIN YES RULESET$_732 STRMADMIN STRMADMIN 301355 ABORTED 09/10/2013 17:34:21 1013 ORA-01013: user requested cancel of current operation CAPTURED STREAMS APPLY SQL> select * from dba_queues; STRMADMIN NLPG73_BLPU_APPLY_Q
[code]...
Can I apply Referential Integrity to only selected rows of a particular column? This is the reference key to the primary key to another table. But, the issue is, this reference column is not having mandatory data for all the rows. So, whenever this is null, I dont want it to be referred by parent table.
no, and it makes no sense to do so either
We have an appointment form in our HIS, where patients take telephonic appointment. That time they may not know their user id given by hospital. So, it remains blank & name is entered manually. But if the user-id is entered, it must fetch patient name from the master.
The candidate key must be unique within its domain.The candidate key can not hold NULL values.
ALTER TABLE HLTHCHKAPPOINTHD ADD (
CONSTRAINT SYS_C007145
FOREIGN KEY (N_PATIENTMR_ID)
REFERENCES MRREGISTRTNHD (N_PATIENTMR_ID));
MRREGISTRTNHD is a patient master & has a primary key named N_PATIENTMR_ID.
HLTHCHKAPPOINTHD is a appointment table & has a foreign key named N_PATIENTMR_ID which references N_PATIENTMR_ID of MRREGISTRTNHD.
Also, N_PATIENTMR_ID of HLTHCHKAPPOINTHD is not a unique key & it can contain null values also. I want to define constraint or any other method such that only the not null values are referenced to the master i.e. it should validate in the master. And null values should skip this reference.
Now, what happens is due to this constraint, when I'm trying to edit & update the rows having null value in N_PATIENTMR_ID, it gives the following error.
ORA-02291: Integrity constraint (SYS_C007145) violated - parent key not found
So, can I give some condition in the above constraint saying, apply this constraint to table HLTHCHKAPPOINTHD only having the not-null values in N_PATIENTMR_ID coulmn?
1) The Primary Database is UP. The Physical Stand By Database is DOWN. The Current Archive Log Sequence is 99 in Primary.
We have to apply Archive Log from 51 to 99 to the Standby Database. But Unfortunately, there is no backup of those Archivelogs and the ArchiveLogs from 51 to 98 have got deleted at Primary end.
Now how will you apply these Archive Logs from Primary Database to Physical Standby Database?
Note : The Physical StandBy Database is DOWN.