Recently we are planning to migrate oracle 10g to 11g. We are migrating PHASE WISE (Schema wise) as this is a 50TB of Production instance. We are using Oracle Golden gate for this purpose.
Is there is any way that we can migrate oracle jobs from 10g to 11g?
1) to define a job in oracle9i for solaris 2) to schedule this job
Remember i have prior experience deploying jobs at windows platform but when i try it on (9i for ) solaris, the script does not work as it does not accept a submit job request.
We have to migrate our database from current OS (AIX 5.3) to Solaris 10. We'll do it through expdp / impdp. Is there any other way to do it as well?
Also, we have dataguard configured at current setup. So is there any way we can take backup of existing database at DR & restore it at new server (Solaris) or we'll have to take a fresh backup from DC (after migration) & ship it to DR & create a new standby?
I am migrating an oracle 10g to 11gR2. i am a bit confused. The developers in my work place have a tablespace for storing table and a different tablespace for storing indexes.
For the migration i am using exp and imp. My question is how i am going to migrate the indexes from 10g to 11gR2?
We will be having a meeting with our client regarding their Database Server Migration (They are planning to buy a new server). Their current database is Oracle 10gR2, they will not upgrade to 11g, they just plan to migrate to a new more powerful machine.
I was planning to ask the following questions.
1. Specifications of the current server and the new one. 2. Operation system (I think they will use same OS, just an updated one) 3. Can the business afford full downtime on current servers? 4. Size of the DB, because it can take hours to move large files.
And is there documentation regarding Server Migration (Change of machine only, not database upgrade or anything,
"How to Migrate Oracle 11gR2 32bit to 11gR2 64bit on Microsoft Windows"
Currently we are having Microsoft Windows Server 2003 standard Edition service pack 2 and we are trying to migrate it to microsoft windows server 2008 Release 2 standard Edition
I have scheduled few jobs (scheduled for each day) which has not run today, which was running successfully till yesterday. When I query dba_scheduler_jobs, i could see last_run_date as yesterday's date 6am and next_run_date is still showing today's date and time (it should show tomorrow's date/time).
I cannot run a Scheduler remote external job on Windows. Here's what I'm doing:
--Check that XDB isinstalled: desc resource_view --Check that MTS is running sho parameters dispatcher sho parameters shared_Server --Setup an http listening port: exec dbms_xdb.sethttpport(8080)
[code].....
This is the error:
orcl> select job_name, status, error#, actual_start_date, additional_info 2 from user_scheduler_job_run_details where job_name='TRYIT'; TRYIT FAILED -1.074E+09 15-JUN-11 02.16.54.583000 EUROPE/LISBON
[code].....
I am certain that the communications and the credentials are correct, I've tried variants and get different errors. I think the problem is the job action. I've tried running batch files as well as OS commands, same result. THere is nothing useful in the core dump. Is there perhaps a Windows specific technique for running external jobs? Some way of nominating the batch file, or specifying a command interpreter?
When performing test recoveries of a database we often wish to prevent database jobs from running application processes such as (a) sending emails; or (b) logging events to a central logging database. In order to accomplish this, we set the job_queue_processes to zero before opening the database.
This means that no dbms_jobs run on the open instance, but it does not prevent dbms_scheduler jobs from running.How can we prevent any dbms_scheduler jobs from running?
I have looked at setting the MAX_JOB_SLAVE_PROCESSES attribute using DBMS_SCHEDULER.SET_SCHEDULER_ATTRIBUTE, but this cannot be set to zero.
I had a task to reschedule database jobs to a different date as they were Running Statistics job running during business hours..
I used dbms_job.change procedure for the same On Thursday i.e. 17th June i changed a job as follows exec dbms_job.change(11, NULL, TRUNC(SYSDATE+2), 'TRUNC(SYSDATE+7)');
When i checked the status today querying user_jobsi found that the job was started today at 00:57 hrs.
I checked the alert logs as well but couldn't find any errors. How to proceed with the troubleshooting for this issue.
i have to migrate the current production database to test 10.2.0.4 on windows. Any non-export way to upgrade 9i to 10 g?
i have following steps me know if i am wrong? 1) ALTER DATABASE BACKUP CONTROLFILE TO TRACE 2) shutdown oracle 9i database on server A 3) copy database file, controlfile, redolog, and other files to new server B 4) alter the controlfile backup with new location of bdump, udump, and log file and data file locations 5) user oradim ORADIM -NEW -SID SID [-INTPWD PASSWORD ]-MAXUSERS USERS -STARTMODE AUTO -PFILE ORACLE_HOMEDATABASEINITSID.ORA 6) start dabase in upgrade mode 7) run catpat.sql and util102.sql take backup 9) open database for users
I want to migrate our production database from 9.2.0.8.0 to 10.2.0.4.0 through startup migrate ,so request you to kindly provide me a step by step doc or some reference link pertaining with same.
currently holding 9i database on win2000 and front end is dev4.5 so the thing is want to upgrade by win2008 pn 32bit with 10gdatabase want to keep the front end application as it is 4.5...
will it be compiatble 10g database on 32bit 2008 and front end 4.5developer forms and reports?
can i migrate 9i database to 10g database, what are the steps for full import??
We are planning to migrate data from an application called clintrace to another application called argus safety. Both the applications are related to pharmacovigilence safety operation. Both the applications functionality are similar. So both the database are having the same data though the table structures might be different. Both the database are oracle clintrace db is 9i and argus db is 11g.
We had a new outsourcer who completed migration of 1 of our system running under Oracle10g-Solaris 10. We created an inventory script that will list all objects counts of SYS, SYSTEM and ORACLE and all other schemas. This script will be ran before (old server) and after migration (to new server). After migration to the new & bigger server, we noticed from the output of the inventory script for SYS/SYSTEM object counts had been reduced compared from the original/actual value (from the old server). Our outsourcer said that this is not an issue since SYS/SYSTEM objects are not transactional and cannot be controlled.
As we will conform that the migration was completed successfully by our new outsourcer.
I have Windows server working Oracle 8.0.5.I have to migrate to another server that is Oracle 11gR2 on Windows 2008.Some say I can't do Export and Import directly.And some say DBlink is not working from 11Gr2 to 8.
What can be fewest step to migrate?Somebody try to use 'Oracle directory' and VB and Powerbuilder program to send data and DDL from 8 to 11Gr2.
We are having oracle 9i as database and migrating to sql server 2008. Our oracle 9i database connects to another oracle databases and these databases are having different oracle version(9.2.0.8.0, 9.2.0.7.0,10.2.0.5.0,11.2.0.2.0,9.2.0.6.0,10.2.0.4.0,8.1.7.4.0 )
I do have two queries :
1) how to connect from Sql server 2008 R2 to different Oracle databases
2) How will other oracle databases connect to Sql server 2008 R2 database
I have migrated from postgres to oracle entire databases however some tables have spaces how can they be removed and how can i found out all the tables in a schema with spaces?
Migrating the LIVE server to a NEW virtual machine.2 databases reside on the LIVE server, one runs in archivelog mode &other no archivelog modeOld server-Windows 2003 32 bit, oracle databases 11.1.0.7.0New-Windows 2008 64 bit,Oracle databases should be 11.2.0.3.0Just want to clarify the following1)When installing oracle, what would be the best option-create and configure a database / install software only(later create database using dbca) ?why? Just want to know if there will be a difference at all in these 2 methods.2)What is the best way to migrate the data in the database, datapump or rman (rman takes full backup in the present "live" so this can be moved to "new live" and can be recovered)? What would be the method with least downtime?(Once the data is transferred, the database in the "old server"will be deleted)3)The structure of the data files (folder path) are going to be different in the old and new servers, so I have to precreate the tablespaces, so will the impdp fail in this case?
im trying to migrate Sybase 12.5 to oracle 11G R2 on linux all tables are fine expect one which fails with error
"SQL Error: ORA-01841: (full) year must be between -4713 and +9999, and not be 0 01841. 00000 - "(full) year must be between -4713 and +9999, and not be 0" *Cause: Illegal year entered *Action: Input year in the specified range"
i have change the data type from CHAR to VARCHAR2 etc as for the datatype date shows as date on the tool - I'm using sqldeveloper from oracle.
I have a procedure that needs to be scheduled as a job. This procedure takes an average of 20-25 minutes to complete. The constraint is that these jobs are submitted using dbms_job.submit with an interval of 5 minutes.
So, will Job2 execute only after Job1 completes [even though Job1 takes 25 minutes to complete] or is there any other mechanism for setting incompatibility for a job with itself.
I have a requirement where I need to group sessions which takes time as given below when ran individually .There are a total of 32 sesions.I want to divide them into 7 groups.I want to submit each one of the 7 groups in a job.So there will be seven jobs on the whole.Execution flow will be group 1 completes ,then group 2 starts like that.Actually I don't want to bombard the database at a time with 32 jobs ,hence the idea of grouping 32 jobs among 7 groups.grouping the sessions so that on the whole ,resources are equally allocated among sessions in a group and which in turn makes the process complete faster. I want to achieve parallelism but at the same time database should not be loaded too much.I just want to know is there any good practice from oracle side while doing this type of grouping. I am planning to group the most time taking one with least time taking jobs.
My Oracle jobs are not running. I can run them manually but they don't start automatically as per interval. The job_queue_processes=100 and CJQ process is not running. I started CJQ0 process by resetting it to 0 and then to 100. But as soon as I enabled a job this background process stopped.
The job also did not execute in its next execution time. I have created the database manually without using DBCA. Is the problem because of this? I have set crontab immediately to get rid of it.