Migrate Current Production Database To Test 10.2.0.4 On Windows
Feb 1, 2011
I have to migrate the current production database to test 10.2.0.4 on windows. Any non-export way to upgrade 9i to 10 g?
i have following steps
1) ALTER DATABASE BACKUP CONTROLFILE TO TRACE
2) shutdown oracle 9i database on server A
3) copy database file, controlfile, redolog, and other files to new server B
4) alter the controlfile backup with new location of bdump, udump, and log file and data file locations
5) user oradim ORADIM -NEW -SID SID [-INTPWD PASSWORD ]-MAXUSERS USERS
-STARTMODE AUTO -PFILE ORACLE_HOME\DATABASE\INITSID.ORA
6) start dabase in upgrade mode
7) run catpat.sql and util102.sql
8) take backup
9) open database for users
I would like to know that how can i automate the export from production to test server. I need direction to create process to import data from production (server A) to test server (server B).
We are planning to consolidate our Oracle Production DB into one server. We are basically a windows shop. Is it feasible to run 8 production Oracle DB in one windows server. All the DB are not really transaction intensive DB. 2 DB in the size of 300GB and others all DB falls under average size of 40GB.
I can take care of the HD slicing so Oracle does not enter into IO bottleneck. We are planning to go for external NAS or SAN for storage.
My main concern is on processor usage. The processor we are thinking about is Intel Xeon Quad Core x 2nos. Will there be a processor bottleneck or is there way in Oracle to assign processor usage(I belive there is no much tweaking options here)
I need to migrate database from Windows to Linux. The current size is ~50GB.
Current Env: OS = Windows 2003 DB version = 10.2.0.3
Proposed Env: OS = Linux DB version = 11.2.0.2
Would using datapump be a correct choice for this migration? Also do the step below seem correct?
01. Pre-create tablespaces on target 11g database 02. Export full database of source 10g database 03. Copy dumpfile to Destination Server 04. Grant IMPORT_FULL_DATABASE system privilege to user SYSTEM of target 11g database 05. Import full database to target 11g database
I want to automate the import from production to test.
1) export the production schema 2) import in to test server?
How can i automate that currently i am doing it manually as follow:
1) expdb the production schema 2) kill all connection on the test server to test schema 3) drop test user cascade; 4) recreate user; 5) impdb the production schema to test:
but i want it to automated or scheduled so i don't; have to log in every night!!
I have to migrate production database version 10gR2 on windows 2003 server to oracle 11gR2 on windows 2008 server...So far i have just installed the oracle software on 64bit windows 2008 server.how to move the data or migrate from old server to new server..
I have to migrate production database version 10gR2 on windows 2003 server to oracle 11gR2 on windows 2008 server . So far i have just installed the oracle software on 64bit windows 2008 server. How to move the data or migrate from old server to new server .
I have been asked to create a new instance of an application server. As part of this, my management has asked me to replace any old JREs or JDKs.
We run 10g, so I installed the 10203_vista_w2k8_x86_production_client on my new server. I would like to replace the 1.4.2 JDK that is bundled with the Oracle Production Client with 1.6.0.
We just upgraded one of our databases to 11g on a new server. The 3rd party application that accesses the db is 32 bit. The software needs to be running on the db server in order for some transactions to occur. Since the oracle client on the db server is 64bit do we need to install a 32 bit client as well. As of right now we are getting an error in the software when trying to connect to the db that says "Test failed. provider cannot be found. it may not be properly installed." This is a live production server so uninstalling and reinstalling oracle is not really an option at this time (unless absolutely necessary). How will installing a 32 bit oracle client effect everything (if that is the route we need to take).
I still have a legacy apps built using Forms 6i and Reports 6i running against Oracle 10g database on a Windows XP client and Windows 2008 server.It seems working well on Windows 7 client (32-bit only), but I think that it needs to test this apps.
Question: to migrate from oracle 10gR2 in Windows to oracle 11gR2 in Linux . Steps are:
1. need to expdp all schemas/users from oracle 10gR2 in Windows, 2. then impdp each schema/user into oracle 10gR2 in Linux first, 3. then upgrade the oracle 10gR2 in Linux to oracle 11gR2 in Linux ? 4. Do I need to pre-create all the same tablespaces first in oracle 10gR2 in Linux ?
Is it correct ? or I can migrate straight all schemas/users from oracle 10gR2 in Windows to oracle 11gR2 in Linux using impdp ?
I'm trying to migrate our forms application from a Linux platform to Windows 7. We have a common pll which is attached to all forms and unfortunately relies heavily on globals.
Attempting to run the application with the standard library attached to the login form has the effect that the globals are lost between the login form and subsequent forms. If I incorporate the library code directly into the login form, it functions correctly. If I use the debugger to track where the clearing occurs, it functions correctly.
we have a production database 'X'. Now i have created a test database 'T' and did'nt configured another listener to it! The issue is when i cam connecting to oracle through sqlplus i am directly connecting to Test database 'T' but not the production database 'X'----ofcourse i can login to production DB afterwards. but initially i want to access the production database 'X'.
I need to refresh a PROD database into TEST database. The PROD and TEST runs on 10g. I need a full refresh. Is there any pre req's which i should keep in mind ?.
I am trying to refresh the validation database with the old production backup i.e. our requirement. I have given the rman script i have executed and the output error message amd the rman configuration setting. Plus in the next post ill post the current sucessfully running RMAN script for your reference.
rman TARGET sys/passwd@Production CATALOG rman_pmxp/rman@catlog AUXILIARY sys/passwd @validation RMAN> show all 2> ; RMAN configuration parameters are: CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 90 DAYS; CONFIGURE BACKUP OPTIMIZATION OFF; # default CONFIGURE DEFAULT DEVICE TYPE TO DISK; CONFIGURE CONTROLFILE AUTOBACKUP ON; CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO 'H:\backups\DB_Sunday_Backup\Prod_%d_%F.rman.ctl'; CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE 'SBT_TAPE' TO '%F'; [code]...
I got following error when i am gathered stats on Schema level.
SQL> EXEC dbms_stats.gather_schema_stats(ownname=>'KDB', estimate_percent=>dbms_stats.auto_sample_size,cascade=>TRUE, force=>TRUE); BEGIN dbms_stats.gather_schema_stats(ownname=>'MFDB', estimate_percent=>dbms_stats.auto_sample_size,cascade=>TRUE, force=>TRUE); END; * ERROR at line 1: ORA-20003: Specified bug number (9196440) does not exist ORA-06512: at "SYS.DBMS_STATS", line 15342 ORA-06512: at "SYS.DBMS_STATS", line 15688 ORA-06512: at "SYS.DBMS_STATS", line 15766 ORA-06512: at "SYS.DBMS_STATS", line 15725 ORA-06512: at line 1
BEGIN dbms_stats.gather_schema_stats(ownname=>'CDOMIG_DATA', estimate_percent=>dbms_stats.auto_sample_size,cascade=>TRUE,DEGREE=>10); END; * ERROR at line 1: ORA-20003: Specified bug number (9196440) does not exist ORA-06512: at "SYS.DBMS_STATS", line 15342 ORA-06512: at "SYS.DBMS_STATS", line 15688 ORA-06512: at "SYS.DBMS_STATS", line 15766 ORA-06512: at "SYS.DBMS_STATS", line 15725 ORA-06512: at line 1
I need to migrate data from Mysql database to Oracle11g database.
a) is there any method available to import the all the sqls like table script,constraint scripts,data(insert ) script from Mysql.so that we can apply the sql directly to the oracle schema after making necessary changes(like datatype).
b) Is there any free tool available for the migration.
i've perform a stress test of a database in 2 configuration :
1) Single instance 11.2 with 8gb Sga e 4gb Pga on Linux, 6core with instance caging =4, resource_manager=defaut_plan
2) 2 Nodes Rac 11.2 with 8gb Sga e 4gb Pga in both nodes on Linux, 32core with instance caging =4, resource_manager=defaut_plan.
I've used Swingbench and the transaction per minute are equal, Avg transaction per s in rac are Greater then single instance (4ms vs 40ms).I know that rac is not for high performance.My question is: is rac config best than single instance and if yes what is the %?
I want to create a database trigger which will test the database link, if it is ok then it will use dblink and do its work.
If it fails then it will send the data into its own server logfile.
I Wrote:
CREATE OR REPLACE TRIGGER PERMIT.TESTTRG AFTER INSERT OR UPDATE ON PERMIT.TR_LP_M_H_COMPANY_25072012 REFERENCING NEW AS New OLD AS Old FOR EACH ROW Declare l number; nIgn PLS_INTEGER; nRows PLS_INTEGER := 0;
[code]....
When I execute it it is giving error:
LINE/COL ERROR -------- ----------------------------------------------------------------- 4/11 PLS-00201: identifier 'EXEC_SQL.CONNTYPE' must be declared 4/11 PL/SQL: Item ignored 8/2 PLS-00320: the declaration of the type of this expression is incomplete or malformed
I need to use Data Pump for the first time on my production Database.Currently on Testing Database, when i am taking schema level export there are no errors or warnings in the log file but when i importing it gives fallowing ORA in the import log file. i searched on google,the only way i found is to recompile the invalid objects. how to avoid this warnings in log file.
"ORA-39082: Object type ALTER_PROCEDURE:"QUANTISV4"."P_CTM_ABN_INVST_EQUITY" created with compilation warnings"
I need to migrate a 500 GB 10.2 database from windows to linux. Of course I found Alejandro Vargas paper on how to do that ("Database Migration From Windows to Linux Using RMAN"). My problem is that the source database will be up again for 7 days and I need to apply all changes made to the source database to the migrated linux database.
Can I just take all redo log files generated from the source database and apply those to the destination database? I might have a problem with a) file conversion and that the redo log are not in sync with the target database (different check points?).I just found that on an exotic website: "The contents of a redo log file depend on a combination of Oracle version, operating system and server architecture. In general redo logs written on one architecture cannot be read on another. There are a few exceptions to this rule. For example, in Oracle 10.2 a redo log written in Linux can be read by a Windows database."
So if this is true, does it also work the other way AND will the database migration from window to linux create a perfect clone of the database where those redo log files actually can be applied to?