SQL & PL/SQL :: Creating New Instance Compared To New Schema?
May 17, 2011
I have SQL Server database I would like to migrate into Oracle. The database supports a large application. It is around 10GB. I requested a new instance but was advised I would have to pay for that but if I asked for a new Schema it could go in our current company instance. I am fine with that since it wont cost more money if I just add a new schema to our Company Oracle instance. Just curious what is the advantage of getting a new instance compared to creating a new schema for 10 GB of data?I assume the advantage of creating a new instance is our Schema (in new instance) and work will have its own space/house and can grow in size without any issues?
creating an ASM instance. I've never done this before and searched all known documentation. I just can't seem to figure this out. This is an 11g database on a WinXp machine for testing purpose before attempting to do this on a live production environment
H want to Create a view which reside in one instance for example xxx Instance . I want to create the same view which reside in xxx instance to another instance for example YYY Instance How to acheive this ?I tried desc the View take the view script from XXX instance and copied the view script to YYY instance when I compiled Iam getting an Error
For Example My view in XXX Instance CREATE OR REPLACE FORCE VIEW xxx.sample ( First_name, last_name ) as select "F_NAME", "L_NAME" FROM EMP;
When try to copy and compile the same view in my YYY Instance i am getting error
I have been asked to improve the speed of our extract feeds to another team. It was using an excel file as csv created with C# reading the csv into a global temporary table and a procedure to insert rows from global table to main table one at a time.
Around 1 million rows That was taking around 25 min. After trying sqlldr and an intermediate table I managed to reduce timing to 15 sec with sqlldr and Direct Path Load. I tried all that in one of our own DEV databases. Now we want to try the approach in one of DEV databases belong to downstream group (the one we feed data). They are rather reluctant to give me access to their DEV database! I have asked them to create a new login and grant object creation (DDL) to this user.They keep saying that they are prepared to give me a role but NOT schema owner. What is the best way of enabling me to create tables somewhere and test loading times with sqlldr.
I need to do full DDL (table/proc creation) and DML on a certain schema. So my question: What is the best way of doing so.. Do I need a schema created for user to do this workI they already have a general purpose schema, can they create a role to have full DDL: rights in that schema?
I have installed Weblogic Server 10.3.6 in my PC .While in installation i have chosen 'Custom' option instead of 'Typical' and Unchecked software Components 'Database Evaluation' and 'Oracle Coherence' and Installed 'Node Manager Service' then continued with Weblogic server installation .Weblogic Server installed Succesfully.
(In weblogic server home folder nodemanager.domains file also not created after weblogic server installation)
Now i started installation of Forms and Reports 11g R2 . While Configuring Instance 'Domain Creation is Failed'.
create an ASM (Automatic Storage Management) instance using dbca. At first try I mistakenly skipped the part where I click on ASM Parameters to add /dev/raw/raw* as my asm_diskstring so it created an ASM instance. When I get to the next page ASM Disk Groups, I selected Create New and I did not have the desired display, because of the step I missed.
I went back to do some cleaning up as follows:
1) Edit the oratab file to remove the ASM entry
2) backed up these files by renaming to old-<filename> old-ab_+ASM.dat old-hc_+ASM.dat old-lk+ASM old-orapw+ASM old-spfile2+ASM.ora
Now when I try to go through the steps to create an ASM instance, at the point after I enter my ASM Parameters, I get the following prompt:
"DBCA will now create and start the ASM instance. After the ASM instance is started you can create the diskgroups to be used as storage for your database."
We have a vendor-supplied 11g database where records are split between two schemas -- an ACTIVE schema and an ARCHIVE schema. Each object has both a corresponding object in both of the ACTIVE and ARCHIVE schemas.
The vendor also has a third schema where each object is merely a UNION ALL of the associated ACTIVE and ARCHIVE schema objects. For the sake of example, I'll call that schema COMBO.Over the years, we've created queries and reports that reference both the COMBO and ARCHIVE schemas and that has worked just fine.
The vendor has now set up a secondary database for us that we can use when the primary database is offline for patching/upgrades/etc. The trouble is, this secondary database only has the ACTIVE schema and records. The vendor will not be writing any ARCHIVE records to it.
Primary DB: ACTIVE, ARCHIVE, and COMBO schemas Secondary DB: only the ACTIVE schema.
is there a way to set up the missing ARCHIVE and COMBO schemas on the secondary DB such that we won't have to rewrite our SQL to accomodate the lack of an ARCHIVE schema when we move reports over to the backup database?
Of course, no records would need to be returned from the virtual ARCHIVE schema, but I'd love for the untouched SQL to run without error.
I'm supposed to create a database for an application. The server where this will be running previously had a database for a pretty similar app. I don't know much about Oracle so I reused the ZFS filesystems and left them how they were created (because honestly, I didn't knew why they were created that way, but I'm pretty sure it was for a good reason).
Since I reused the filesystems I created my database and placed the controlfiles in the same places where the old database files were (/oradata/SMART/ora1,/oradata/SMART/ora2,/oradata/SMART/ora3). Thinking like MySQL works I created CODEapp/oradata_smart_ora4 60.6K 400G 60.6K /oradata/SMART/ora4 specifically to store the database there.
The databases startups and mounts no problem. Note that this server will manage with millions/billions records throughout its lifetime.
1. Now that my Database is created, whats the next step? Create the Schema or Tablespace? 2. Tablespace Questions: Tablespace datafile(s) is where actual data from tables is stored? how many are needed? Default or Temporary? How much space will I need for it? Autoextend?
I need to create a structure DATABASE=>SCHEMA=>TABLE as
DB=>SC=>EMPLOYEE ...but after connecting database i could create table only user my user schema(own schema)only . I want to create a new schema called SC as public and need to create a table .
I have a list of materialized views in schema A. I want to create a refresh group and then refresh it from Schema B (Dynamically-run time based on some criteria). What Grants are necessary on schema B in order for it to be able to create and refresh the groups on Materialized views in Schema A.
I know that one of the Options is to, GRANT ALTER ANY MATERIALIZED VIEW as a SYS user. But I do not have any SYS privileges.
Recently we have upgraded to 11g(11.2.0.3) from 10.2.0.2 on AIX 5.1 enviornment. After upgrade, it has been noticed that CPU utilization is more while executing a program.
CPU is crossing 85% in 11g, in 10g it is 60% during the program execution. I have gone through the SQLs which are getting executed, elapsed time for all SQLs are less than in 10g. i.e. elaspsed for all SQLs in 11 is very less. in AWR report, DB CPU stood in the top.
I am using toad to output the details of the database. (I need these details in a format which i can then import into excel) However, When I run sporadic individual records through my query, the results come back correct and as expected. But when I run the full data set through my query a lot of the records have different values than what they do when run individually!
Ive spent all day on this script and its really a hack of various other cursors and scripts brought together. Its driving me mad as my code doesn't seem to be wrong (since the results are correct when a single record is run) yet there must be something wrong as my full data set does not return correctly!
Primary RAC and Standby RAC databases are using version 10.2.0.4. They are configured & controlled via Data Guard Broker. Currently, the observer runs from one of the Standby nodes. All are on Solaris x86 64bit OS version 10.
Now, I have to move Observer to 3rd DC and I am looking If I could use another OS for observer instead of same as DB OS version. Why? Our new infrastructure is on RHEL Linux, therefore, I would like to deploy observer once and then wait for existing databases to be upgraded and moved to RHEL. Also, i understand that database and observer software should be of same version e.g. 10g, 11g.
My question is what if i install & configure observer on RHEL ? The end result would be:
1. Primary DB: Solaris 10, Oracle Software binaries version 10.2.0.4 2. Secondary/Standby DB: Solaris 10, Oracle Software binaries version 10.2.0.4 3. Observer: RHEL 5.8, Oracle Software binaries version 10.2.0.4
Is it accepted mix configuration say e.g. to run for six months? Also, it would be grateful to know if there any license implications when running observer on a different node running no databases at all.
I am creating a table from another existing table in another schema. The existing table contains data. When I am using the query- create table m _voucher as select * from ipm.m_voucher,I am getting the whole data of m_voucher but I want empty m_voucher table, so what will be the query to get the empty m_voucher table?
Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options Master table "MVANMANNEKES"."SYS_IMPORT_SCHEMA_01" successfully loaded/unloaded Starting "MVANMANNEKES"."SYS_IMPORT_SCHEMA_01": mvanmannekes/******** schemas=cmsstagingb remap_tablespace=cmsliveb_data:cmslivea_data
- one ASM instance - X DB instances - each DB instance uses 2 or more dedicated diskgroups from the ASM instance - there is one diskgroup named FREEDISK that contains spare disks
On each DB instance you can see:
- the list and global parameters of all diskgroups using v$asm_diskgroup view - the list and parameters of all disks the instance is using with v$asm_disk view
So my question is: how (if this is possible) to know the list of (spare) disks in FREE DISK disk group?
There is an interface table and there is an normal transcational table..interface table is being compared with normal table and if match found the result is dumped into another normal table.
I am using two cursors one is to query the interface table and in a for loop pass the results to the second cursor..The interface table is having 5000 + rows and the transcation table is having more than 3.7 millions ..and the program is taking lots of time to execute..took almost 35-45 minutes..
A single master schema where many developers are accessing. all share same password.
now i would like to trace all the changes made by each users. so i create a individual users for all and grant permission to access that schema.do i have a possibility of auditing the changes did by each user for that particular schema
We have an application with many separate databases (one per customer). Given they share the same business requirements (service hours, change mgmt etc), we're interested in potentially consolidating the separate DBs (which are relatively small) into separate schemas within a fewer no of databases to reduce the overhead.
Our issue is that the application is hard-coded to use a specific administrator and application connection user name. Changing this is unfortunately not an option.
Given this limitation, is there any possibility to map a generic user into a customer-specific schema based on the database service that they connect to? Each customer connects to different database services but may use the same user name. We considered using private synonyms but this seems to acheive the opposite (i.e. many different users could connect and map to a single users schema). One thing to point out is that where there is a single user name, it is acceptable for a single password to be used across the different customer DBs as they will be a single admin/user.
I would like to create a table in another schema(CBF) as already exist in my schema(TLC) without data but related indexes,synonyms and grants should be include.
How could I do this without using export import. I am using TOAD 9.0.1.
move the tables with data present in the user scott(full) to another schema named test. In my case scott is in user tablespace and for test schema i have created different tablespace named test_tbs.
A user is using an ad hoc tool similar to SQL Developer called PeopleSoft Application Designer.
He creates a connection to the db, then issues an alter session set current_schema = 'restricted_schema'. The connected user does not have direct privileges on the "restricted_schema" which they call SYSADM.
After changing the schema context in that manner he creates objects in SYSADM. A schema trigger is then fired and grants privileges on the new objects created in SYSADM. Doing the same in either SQL Plus or SQL Developer does not fire the schema trigger.
I think SQL Plus and SQL Dev are working as they should. Altering the session like that does not change your identity - just the schema context. But, when you examine v_$session, the connection with this other tool looks exactly the same as one from SQL Plus or SQL Dev when changing the schema context in the session.
Instead of trying to figure out what this other tool is doing, is there any way for that schema trigger to fire when using this process from one of our tools?