Server Utilities :: Downgrade From Oracle11g To 9i Via Exp / Imp?
Dec 28, 2010
I want to downgrade my database from Oracle 11g (11.2.0.1.0) to Oracle 9i (9.2.0.1.0).
I've done trial of taking export of higher version(Oracle11g) through lower version(Oracle9i) exp utility and trying to import same exp.dmp files into lower version(Oracle9i), but not succeeded and getting some parameter related error.
On Oracle9i database client- Here i'm connecting to Oracle11g via use of 'wbdata.wbh-db1' service
[oracle1@WBDevelop smbshare]$ exp lab/lab@wbdata.wbh-db1
Export: Release 9.2.0.1.0 - Production on Tue Dec 28 15:45:32 2010
Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Produc
tion
With the Partitioning, OLAP, Data Mining and Real Application Testing options
Enter array fetch buffer size: 4096 >
[code].....
On Oracle9i database client - Here i'm importing the lab.dmp
(Export taken from Oracle11g via use of 'wbdata.wbh-db1' service)
[oracle1@WBDevelop smbshare]$ imp lab/lab
Import: Release 9.2.0.1.0 - Production on Tue Dec 28 15:49:51 2010
Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
Connected to: Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
With the Partitioning, OLAP and Oracle Data Mining options
JServer Release 9.2.0.1.0 - Production
My boss told me he wanted to know the QPS(Query Per Second) and TPS(Transaction Per Second) of current Oracle Database. The version is 11.2.0.3. I found the TPS in the section of Load Profile of AWR report. But how can I get the QPS? I have the OEM12C. If it can provide the info, which page?
I like to switch from a rac to a standalone database including downgrading the license to standard edition one. Actually SE1 is not an install option for 10g.
Would it be enough to reconfigure the 2nd rac node ?
Actually I have to modify set SQLNET.INBOUND_CONNECT_TIMEOUT,TRACE_LEVEL_CLIENT in SQL net .ora . But I cant fine that file in $ORACLE_ HOME /network/admin directory.
But I found this other files like listener.ora and tnsname.ora in the same directory.
When I give sho parameter nalue in listener prompt it shows some values. Is it showing default values without sqlnet file (is it fauture of Oracle11g).
I have started to migrate 11.2.0.3 database from Enterprise to Standard edition with expdp/impdp (RHEL box). I've managed to transfer database without any problem (ok i had little difficulties with bitmap indexes but never mind).
But now i do not have all public synonyms in place as i already knew. So i have to create sql script and i have found several good ones. My question is do i need to recreate all public synonyms or there is a exceptions in my case because i now have Standard Edition database (old one was Enterprise)?
What I was trying to do is duplicate my 10.2.0.4 DB and overwrite my 11.2.0.2 DB. Is there a way to downgrade my 11.2.0.2 DB. my 11Gr2 DB I have already set the compatibility parameter to *compatible='11.2.0.2' My 11gr2 is a test DB I don't have a valid backup..
I installed oracle11g on my computer ,when im going to build demo its not working.its giving a error like this.
sys>conn u1/u1 error accesing PRODUCT_USER_PROFILE warning product user profile information not loaded you may need to run PUPBLD.SQL AS SYSTEM connected u1>@?/sqlplus/demo/demobld unable to open file "/oraeng/app/oracle/product/11.1.0/sqlplus/demo/demobld.sql"
I am checking out licenses. We all know that EE is much more expensive than SE. But many customers do have EE installed - unsure if they need all the features at all. After several years of production, a downgrade is considered 'risky' and we continue to pay the full EE.
How can we check and be sure that a downgrade to SE would not be any problem?
Some checks include: * partitioning used in user schemas? --> no downgrade to EE * bitmap indexes in user schemas? --> no downgrade to EE
How can we complete this list, or is there some script to make this easy?
am trying to downgrade my oracle database 10g release 10.2.0.2.0 to 10.2.0.1.0the reason for this is that i had to copy /oracle_home/bin utilities from 10.2.0.1 to 10.2.0.2 because of this most utilities like rman,dbca were not working.
i think i need to downgrade my oracle database 10g release to 10.2.0.1.0 right?i read some oracle documentations of database downgrade instruciting me to
>take full backup of oracle DB >ORACLE_HOME/bin/emctl stop dbconsole SQL> STARTUP DOWNGRADE DROP USER sysman CASCADE; >SQL> SPOOL downgrade.log
We have Forms application created in Oracle Forms Builder version 9.0.4.0.19. We have done modification in some forms and few forms created using Oracle Forms Builder 10g version 10.1.2.0.2.
Now we need to convert those forms which are around 5 or 6 in number to Oracle Forms 9.
srvctl config PRCR-1119 : Failed to look up CRS resources of database type PRCR-1115 : Failed to find entities of type resource that match filters (TYPE == ora.database.type) and contain attributes DB_UNIQUE_NAME,ORACLE_HOME,VERSION Cannot communicate with crsd
I am using oracle 11g express edition and MY Sql database How to establish database link between them using ODBC gateway. I am using ODBC gateway Mysqlodbc connector.
I want to load lakhs of records into a table. My problem is when after loading the ¼ of records my process is abend due to the size of my rollback segment area. I don't have an option to increase it. So, Is there any way to go for intermediate commits when I am using the imp or sqlldr utilities to load the entire data without abend?
I am familiar with tool Netca. However there is one more utility exist for the same functionatlty which is netmgrI checked with many DBAs for the exact difference, however I did not get the best answer from them. I also have checked in google but not exactly got the difference. list the exact difference between those 2 tools (netca, netmgr)
I have exported data of one user an importing into another schema at another server. when i am trying to imoport it is working fine for quite no of imports into tables, but after some time it starts giving me below mention error...
IMP-00008: unrecognized statement in the export file: < IMP-00008: unrecognized statement in the export file: < IMP-00008: unrecognized statement in the export file: <ے IMP-00008: unrecognized statement in the export file: +A IMP-00008: unrecognized statement in the export file: [code]...
I have a requirement to read flat text file(around 15000 lines) residing at a client location from DB server and write into a table in One cell.
I tried UTL_FILE and DBMS_LOB but, i am not able to access client location to read the file as it reads path from Oracle Directory.
eg. my client path is 198.168.1.1 and my DB server is in unix say 192.168.1.10. file location is: \192.168.1.1shareabc.txt So I created One Oracle directory as MY_DIR having DIRECTORY_PATH as '\192.168.1.1share'. But both UTL_FILE and DBMS_LOB is not able to access the file.
Error Message: ------------- Unable to process CLOB -22288 ~ ORA-22288: file or LOB operation FILEOPEN failed No such file or directory
Few Details for reference: ------------------------- File Location: \192.168.1.1shareabc.txt Unix DB Server location: 192.168.1.10 Table : Test (filename varchar2(30), Content CLOB) Oracle Dir: MYDIR Directory_Path: \192.168.1.1share
I've a question regarding difference of character sets, while taking a export(logical backup) of database on directly to server(linux RHEL 2.1 AS) and export on a client (windows xp prof machine, where only a oracle 9i client is installed). On server it seems to fine and okay, but on client node i'm getting following error for almost all tables.
EXP-00091: Exporting questionable statistics.
My question is :
[1] Is it creating any sort of problem, if later on i import the data which was taken from client node.
[2] Why there is a difference(marginal) in dump(.dmp) file size.
[3] Is there any way to overcome it, or it is the natural behave of it. Means not a problem.
[4] If i'm using a long or blob as datatype for some of my table,is they have any problem if i persist like above.
Additional Information about character sets On server node :
Export done in US7ASCII character set and AL16UTF16 NCHAR character set server uses WE8ISO8859P1 character set (possible charset conversion)
On client node :
Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set server uses US7ASCII character set (possible charset conversion)
load data infile 'trlc.csv' replace into table trlc fields TERMINATED BY '|' TRAILING NULLCOLS (est_no,right_no,maj_auth,weight,idm_ht,c_date,P_tkt)
The rows get inserted successfully. But the result sets are different, for example: When I do a select in SQL Server,'select len(weight) from trlc;' , I get the length as 0. But when I do a select in oracle database, I get the length as 1. Also, the result set varies for the query below:
select * from trlc where weight=' ';
(SQL Server returns 1 row but Oracle returns no rows)
Do I need to mention any conversion code for the weight field to accept ' ' value?
I've inherited a 10.2.0.1.0 instance running on a windows 2003 box; running fine, no problems other than system has been in production since 2005 and has gotten pretty old and tired. This old box has one tablespace on it... called "gateway".
I've installed 11.2.0.1 on a new (Windows 2008 R2 Enterprise 64-bit) server and created an empty database also called "gateway".
Now to move the data and views, objects, everything.
I've read up on a variety of migration techniques (oops, I mean "upgradation" LOL) and can follow the steps...
In short, I want to pull everything off of server a (10.2) and put it into production on server b (11.2). There seems to be quit a few options.
1. install 10.2 on my NEW server (server b), move the data over and get everything running, then install 11.2 and have it upgrade the database as part of the installation process. 2. drop the empty tablespace on server b, stop the database on server a, copy the files over from the old to the new home, run DBUA or set the compatibility attribute... 3. run some type of server a to server b utility that can bridge the two over the network. Some type of mirroring technique? 4. run file export scripts on server a, copy files to server b and run various import scripts
I tend to think that option 3 would be the best because both instances are in great health and are running right now. Is there a mechanism that allows the 11.2 instance to see and upgrade from a different server?
any other utilities that we can use to load data from our PROD server (10g) to DEV server (9i)? I've read some related topics here that it's not possible to import from a HIGHER to LOWER versions of Oracle. We've tried (many times) EXPorting selected tables from the 10g dB, then IMPort it to the 9i dB and we've haven't succeeded anyhow. PROD & DEV have a different schema/owner but the same table structures.
I have lot of problem using 11G and i dont want to continue trying to repair them all. I just want to do a full DMP Export from 11G DB and Import into 10G.
The reason? No problems when using 10G before.
I try using the oracle export tool on the 11G DB, but when i try to import into 10G there's a header problem. I guess that, like in other oracle DB version, the 11G DMP isnt compatible whit the 10G import tool.
For migrating the database version 10.2.0.3 on Windows to version 11.2.0.3 on Linux i am using the transportable tablespace option.After the import of metadata and copy of tablespace datafiles to destination the sequences and functions are not available. Does it work if i later export and import the sequences and functions separately ?