Server Administration :: Move Database From 10.2.0.4 (Solaris 9) To 11.2.0.2
Sep 17, 2011I have to move a database from 10.2.0.4(solaris 9) to a 11.2.0.2(Solaris 10) . 11g shell database is ready .Its a 400GB database.
View 7 RepliesI have to move a database from 10.2.0.4(solaris 9) to a 11.2.0.2(Solaris 10) . 11g shell database is ready .Its a 400GB database.
View 7 RepliesAre there any good information about on how to move a 10g database to a new server? We will keep the same disks on the SAN, but make a new installation on the new server.
View 3 Replies View RelatedI tried to upgrade oracle 10.2.0.4 to 11.2.0.2 on the Solaris 10 server, but after 40 minutes it hangs bei oracle XML Database.
View 5 Replies View RelatedWe have to migrate our database from current OS (AIX 5.3) to Solaris 10. We'll do it through expdp / impdp. Is there any other way to do it as well?
Also, we have dataguard configured at current setup. So is there any way we can take backup of existing database at DR & restore it at new server (Solaris) or we'll have to take a fresh backup from DC (after migration) & ship it to DR & create a new standby?
our system admin is changing the time zone on solaris system, how can it effect my DB or as DBA what should i check before he do that and after he change the time zone. he is changing the time zone because of an issue with the DST (Daylightsaving time).
View 1 Replies View RelatedHow to
1) to define a job in oracle9i for solaris
2) to schedule this job
Remember i have prior experience deploying jobs at windows platform but when i try it on (9i for ) solaris, the script does not work as it does not accept a submit job request.
My problem is only on Some Saturday's my Oracle server's Load average goes high (more than 300 ).
-- oracle version 11.2.0.1.0
-- runs on Sun solaris 10
-- Sun fire V 440
-- Sun storEdge 3315 connected to the server.
Same setup is working find with higher volumes without any problem but only on saturday's, that too not on all saturdays, some specific saturdays the load average goes high.
At the time nothing will be processed from the application side and the cpu utilisation goes high upto 95 %.
I am sending necessary information as follows.
vmstat 6 10
kthr memory page disk faults cpu
r b w swap free re mf pi po fr de sr s0 s1 s3 s4 in sy cs us sy id
3 0 0 28660024 6513408 285 212 2044 2 2 0 0 0 17 0 42 830 3469 1380 22 5 74
125 0 0 29027480 6156256 2 6 34 0 0 0 0 0 5 0 10 791 53770 30278 83 17 0
125 0 0 29027680 6157496 29 140 21 1 1 0 0 0 5 0 9 786 52756 30309 83 17 0
116 0 0 29031600 6159896 24 125 0 0 0 0 0 7 3 0 5 819 54081 31069 83 17 0
[code]....
Erroneously created datafile, re: "/path/../large_rbs_03.dbf" was created under the SYSTEM tablespace which is supposedly be in the LARGE_RBS tablespace.
How do I make the said datafile be under LARGE_RBS?
How to move a partition of a table to db_keep_cache ?
View 8 Replies View RelatedI mistakenly added a datafile to a tablespace which is asm, however the datafile was created in a default location and not the asm location:
alter tablespace pdaiidata1 add datafile '<filename>' size 2048M;
What I should have done:
alter tablepsace <tablespace_name> add datafile '+DATA1' size 2048M;
Is there any way to move this filesystem datafile into the asm tablespace? In previous Oracle versions, I've taken a tablespace offline, moved a datafile, renamed it, then brought the tablespace back online. Can I do something similar here in this situation?
To one of my server i can see listener log are on normal location i.e. $ORACLE_HOME/network/log and on other server listener log is DIAG home. how to move listener log from ORACLE_HOME to DIAG_HOME.
oracle > lsnrctl status LISTENER
LSNRCTL for HPUX: Version 11.2.0.2.0 - Production on 29-JUN-2012 10:20:14
Copyright (c) 1991, 2010, Oracle. All rights reserved.
Connecting to (ADDRESS=(PROTOCOL=tcp)(HOST=)(PORT=1521))
STATUS of the LISTENER
------------------------
Alias LISTENER
Version TNSLSNR for HPUX: Version 11.2.0.2.0 - Production
Start Date 03-MAY-2012 08:09:57
Uptime 57 days 0 hr. 10 min. 16 sec
[code]...
I am using expdp/impdp to migrate 4 TB database from solaris to Linux.But the import process is taking forever.
View 13 Replies View RelatedI need to move my indexes (around 300) from data tablespace to index tablespace. What's the best way to do it?
View 9 Replies View RelatedI want to move 350 database objects from one database to another database schema.
View 16 Replies View RelatedI was about to move some tables from one table space to another but it seems it is not possible to move partitioned tables between table spaces of different block sizes.
So far the only option I have is to export and then import back the data.
know if there is any way to move a partitioned table between table spaces of different block size?
I am part of a upgradation and migration proj where an Oracle 9i db on a HP-UX OS is going to be upgraded and migrated as 10g db on a solaris OS environment.I am from the application team(Data warehouse team) and we are yet to allocate a specific DBA for taking care of this activity.
We are still in the requirements stage...different requirements and challenges that should be taken care and would occur in the roadmap of this project from a DBA perspective.All your inputs would be very valuable to me.
move the tables with data present in the user scott(full) to another schema named test. In my case scott is in user tablespace and for test schema i have created different tablespace named test_tbs.
View 14 Replies View RelatedWe have an Oracle 9i (9.2.0.7) two-node RAC database running on Solaris 5.8. Recently we have switched our production database from no archive log mode to archive log. We changed the parameter CLUSTER_DATABASE=FALSE on one of the instances before converting and reverted back.why this parameter needs to be changed to false?
View 2 Replies View RelatedWe are planning to migate Oracle database from Linux 10gR2 (non-ASM) to Solaris (ASM) and then Oracle upgrade that database from 10gR2 to 11gR2 on Solaris.
View 4 Replies View RelatedWe've been battling with very slow performance for some time. Herewith a detail description of the problem:
Solaris-11/ZFS/Oracle problem
We purchased Oracle T4-2 servers, and are experiencing some weird performance problems.
Hardware:
T4-2
2 x 600GB HDD per server
128GB memory per server
2 x dual port QLE2562 FO HBAs
IBM V7000 StorWyse data array
2 x CISCO MDS 9148 fibre optic fabric switches
Software:
Solaris 11.1
MPXIO
Solaris 10 branded local zones
ORACLE 10g Enterprise edition
Project for the oracle user:
user.oracle:100::oracle::process.max-file-descriptor=(basic,8192,deny);process.max-stack-size=(priv,32768,deny);project.max-shm-memory=(priv,21474836480,deny)
We received the first server and wanted to migrate our APPB application system and Oracle 10g Standard Edition database from our SUN T5240 to the T4-2.
T4-2 setup � Disk 0:
Global zone: Solaris 11.1 ( ZFS - whole disk for the root pool)
Local zones: On the Solaris 11 environment we built two branded Solaris 10 zones using an Oracle template provided on the Oracle website - solaris-10u11-sparc.bin.
Our complete database resides on the IBM data array, in UFS LUNs.The UFS LUNs were mounted onto ZFS mount-points in the root partition (/), and then LOFS mounted into the zones.
We started with the Solaris 11.1 environment.
1)After a day or two the performance of the database starts deteriorating rapidly. We then stop the database and reboot the machine. After the reboot the performance level is restored.
2)Another huge deterioration in performance happens when we unmount the V7000 LUNs, and reboot to the alternate Solaris, and re-mount the LUNs.
3)What further compounds the issue is that when we start another the database in the second zone, we see another huge performance degradation.
4)We have logged a call with ORACLE. They requested us to gather information which was analysed by them. They did not find anything wrong with the way ORACLE was installed and the setup of the instances.
On Disk 1 we did a Solaris 10 8/11 (Update 10) installation, which we patched with the April 2013 CPU patchset. In this Solaris 10 global zone we built two native Solaris 10 local zones. The Oracle 10g databases were built in the zones (same configurations settings) not in the global zone, onto UFS LUNs. The database in its entirety lives on the IBM V7000 data array. This works fine.
We then received our next T4-2 server.Loaded it again with Solaris 11.1, and upgraded to ORACLE 12c SE Release 12.1.0.1.0 - 64bit12xxx seeing that Oracle 10 is not certified on Solaris 11. To keep things simple, we built two small databases in the ZFS root pool. The complete system now resides on one disk no UFS LUNs to consider, no Fibre Optic fabric, no CISCO switches, no IBM data array, BUT we get the same problem. The system will run for some time and then slow down drastically. Starting the second database slows the system down abnormally.
Oracle Solaris 11
Oracle Database 11.2.0.3
Create a new database using dbca and at the end dbca reports it could not start database control.
The dbca warning window asks to set ORACLE_UNQNAME and to start database control again.
su - oracle
$ . oraenv
rcat
export ORACLE_UNQNAME=rcat
emctl stop dbconsole
emctl start dbconsole
failed...
I tried that and dbconsole still cannot start.
I have oracle 10g up and running on Solaris 10, from windows I would like to connect to sql plus through windows authentication, for that I have already made sure that remote_auth = true and have created user in oracle with OPS$. But still I cannot connect.
I have the same setup but with oracle on windows server, the os authentication from windows clients works just fine.
does oracle 10g on solaris 10 supports windows os authentication?
which is really a HA zone clustered in VCS(not important for installation of the database).
What parameters do I need to set or change in the Solaris zone, there seems to be no /etc/system file.
From some day I have this error during export data pump:
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options
ORA-31626: job does not exist
ORA-31687: error creating worker process with worker id 1
ORA-31687: error creating worker process with worker id 1
ORA-31688: Worker process failed during startup.
This error is random, if I retry after few minutes the expdp work correctly.
I have a question: Did someone have in production two or more RAC 11gR2 on different Solaris LDOMs on the same physical server?.
View 0 Replies View RelatedWe plan to move the database to a new host with much more RAM, CPU. The new database server will have exactly the same OS version, Kernel version and the same version of database / grid standalone (ASM). I will have to install OS, Database software, Grid Standalone on the new server.
The current version of the database is 11.2.0.3.0 Enterprise Edition and the OS version is Oracle Enterprise Linux 6.3. However, we plan to use the same disks (ASM disks). I believe I don't need to use RMAN to do the migration, Is it correct?
i am installing oracle database 8.1.7 on dell server power edge 2650 first time database successfully installed but when i want to crate new database by Database Configuration Assistant it is not working for new database creation.
View 1 Replies View RelatedI wanted to move all db objects from one schema to another schema.
View 4 Replies View RelatedThe requirement details are as below,
1) The database contains static as well as transactional data for telecom domain.
2) We have to move region-wise data from one database to another.
3) There are around 10 regions.
4) The region wise data extraction from source db is based on around 40 to 50 tables. Rest of the tables contains static data & it will not change.
5) The volume is around 1 million subscribers per region.
6) The migration is required because
a) The client is upgrading its base product which uses this database
b) There is a change in structure of static tables
c) Hardware upgrade
d) The client want to start with single region on new database & rest of the regions will be operated from old database.
7) Keep execution time to very minimum.
I am thinking to have solution as below,
1) Create destination database with upgraded db structure (as mentioned in 6b)
2) Create database links to access source db in destination db.
3) Write SQL queries to fetch data from all the respective tables for a specific region
4) Write separate PL/SQL blocks for each table to fetch data from source db & insert into respective table in destination db
a) Use FOR ALL & bulk collect to improve the performance
b) Divide table data into multiple chunks & execute parallel batches (around 10) to insert the data
5) Validate pre & post counts to verify the success of migration
I want to move a oracle job from one database to another database. i tried to open the job from toad by pressing f4 on it in editor , im getting message like...
"unsupported type for describe popups (Without DB admin module)