Attack Databases While Performing Backup?
Apr 1, 2011are there any possibility to attack oracle databases while performing database backup!
View 3 Repliesare there any possibility to attack oracle databases while performing database backup!
View 3 RepliesTape streaming is not happening while performing RMAN tape backup. On investigation, you find that it is not because of the incremental backup or the empty file backup and that RMAN is sending data blocks to the tape drive fast enough.What could be a solution to make tape streaming happen during the backup?
A.Configure backup optimization
B.Configure the channel to increase MAXOPENFILES
C.Configure the channel to increase the capacity with the RATE parameter
D.Configure the channel to adjust the tape buffer size with the BLKSIZE option
I have the task to migrate the total databases(Exact copy to be moved to another server).The current server is going for format.After I did the following steps I am getting the tablespaces(databases)-4 sizes same ,but I am facing issue like some default tablespaces i.e temp,system are not matching.
temp tablespace
***************
current server - 4.0(approximately)
Migrating server - 160 MB
System tablespace
*****************
current server - 580 MB
Migrating server - 220 MB
Also I checked the tables are also matching for the 4 databases.Also Provide the solution or method which is correct.
steps done for migrating(By me)
********************************
EXPORTING DATA USING DATAPUMP
*********************************
1 From command prompt MKDIR 'c:oraclexeapp mp';
2 From SQL prompt conn system/kotak;
3 create or replace directory dmpdir as 'c:oraclexeapp mp';
4 grant read,write on directory dmpdir to kotak;
5 From command prompt
expdp system/kotak@xe full=Y directory=dmpdir dumpfile=xe.dmp logfile=expdpxe.log;
IMPORTING DATA USING DATAPUMP
*****************************
in another server machine
1 From SQL prompt conn system/kotak;
2 create or replace directory dmpdir as 'c:oraclexeapp mp';
3 grant read,write on directory dmpdir to kotak;
4 From command prompt set ORACLE_SID=xe;
5 impdp system/kotak@xe full=Y directory=dmpdir dumpfile=xe.dmp logfile=impdpxe.log;
IN OUR PROCESS we created the below tablespaces and user before IMPORTING created 4 tablespaces
1. kotak
2. kotakdb
3. wired_data
4. ferrari
Created Users which are there in 219 server
1. KOTAK
2. KOTAKDB
3. FC_80
4. DEMOINTERNETBANK
5. DEMOINTERNETBANKDB
6. CREDITCARD
if you are using ShadowProtect to backup your Oracle databases. What are your experiences and the pros and cons in using ShadowProtect as the one and only backup/recovery tools for Oracle databases? Can the potential of ShadowProtect replaces the commonly use Oracle backup/recovery tools?
The reason why I raise this question is because my boss has the idea that ShadowProtect is THE best backup/recovery tools. Personally I don't think so, because there can be cases where we only need to recover the database and not the whole OS.
When we run a backup of a database [oracle7.3.4.5] using EBU [2.2.1.0.0] on Solaris 7, error message appears:
Statement processed.
BFS in progress "1134684df1de4a" cancelled on 10-JUN-2011 11:05:34
BFS in progress "1134694df1de4a" cancelled on 10-JUN-2011 11:05:35
Ending backup to tape
This backup spawns archive sequence numbers 1731 to 1731
This backup comprises SCN 8725981 to 8725802
EBU-2003: Job 24267 failed due to an internal error
on 10-JUN-2011 11:05:52 [ 7178 : brcdodbfil ]
[cod]....
In EBU catalog, are registered 2 databases with the same name but different hosts.
what type of storage systems used for saving archive files and backup files "especially for oracle databases"
View 5 Replies View Relatedwhile try to expdp on network drive. getting below error.how can we perform the expdp on network drive
Network location: \\tsclient\p\expdp
1.)SQL> show user
USER is "SYS"
SQL> create or replace directory exp as '\\tsclient\p\expdp';Directory created.
SQL>Grant read,write on directory exp to system;
2.)expdp system/xxxxx@orcl directory=exp dumpfile=EXP_orcl_072013.dmp logfile=EXP_72013_1.log schemas=('IIMS','CMMN')
Export: Release 11.2.0.2.0 - Production on Tue Jul 23 13:48:07 2013
Copyright © 1982, 2009, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Release 11.2.0.2.0 - 64bit Production
ORA-39002: invalid operation
ORA-39070: Unable to open the log file.
ORA-29283: invalid file operation
ORA-06512: at "SYS.UTL_FILE", line 536
ORA-29283: invalid file operation
I have a type to table where only few columns are SCD 2. Like address, city, zip.I wrote a code using cursor. But now the requirement is to replace the cursor .how to perform type 2 using Merge in oracle 10g or any better way to perform it without using cursor. Look the code which i have attached with cursor.
CREATE OR REPLACE PROCEDURE Proc_cp_jci_contract_dim_hist
IS
l_fun_ret_value NUMBER;
v_src_id NUMBER;
rec_jci_contract_dim jci_contract_dim%ROWTYPE;
rec_jci_contract_dim_hist jci_contract_dim_hist%ROWTYPE;
err_row VARCHAR2(4000);
ind NUMBER(1) := 1;
[code]....
Iam usin 11.2.0.3 database on linux 64 bit machine
I have got the report from the users that the dataload operation which they perform in the database is taking normal than the expected time
I rebooted the server and started the database i can see that the database startup is even taking more time
The memory allocated is sufficient and when i tried to perform the logswitch the database simply hangs
alertlog contents after restart
---------------------------------------
minact-scn: useg scan erroring out with error e:12751
Tue Oct 16 11:02:03 2012
minact-scn: useg scan erroring out with error e:12751
Suspending MMON action 'Block Cleanout Optim, Undo Segment Scan' for 82800 seconds
Tue Oct 16 11:12:47 2012
Active Session History (ASH) performed an emergency flush. This may mean that ASH is undersized. If emergency flushes are a recurring issue, you may consider increasing ASH size by setting the value of ASHSIZE to a sufficiently large value. Currently, ASH size is 50331648 bytes. Both ASH size and the total number of emergency flushes since instance startup can be monitored by running the following query:
select total_size,awr_flush_emergency_count from v$ash_info;
Tue Oct 16 11:17:53 2012
Suspending MMON action 'undo usage' for 82800 seconds
Tue Oct 16 11:28:05 2012
Suspending MMON action 'metrics monitoring' for 82800 seconds
I have prepared shell scripts to do the parallel inserts on my DB table (LEGACY_SYSTEM).
There is a trigger (AFTER INSERT ON EACH ROW) associated with the above table. I am calling a package.function inside the trigger to do the required operation and finally it will insert records into my target table (PRICE_CHANGE).
Expectation:
------------
If I insert 10 rows into LEGACY_SYSTEM table, it should do few updates and finally insert 10 rows into PRICE_CHANGE table.
Result:
-------
10 rows got inserted into LEGACY_SYSTEM. All the updates are successful but I could see only 4 rows in PRICE_CHANGE table. If I run it for the second or third time, all the results will be perfect.
Instead of these shell script, if I insert one by one rows manually into LEGACY_SYSTEM table, I am getting all the expected results and the results are consistent. If you look at my scripts below, you will understand the problem better..
DRIVING SCRIPT(test_global.sh):
---------------
#!/bin/sh
./test_script.sh &
./test_script2.sh &
exit;
test_script.sh:
---------------
#!/bin/sh
./test.sh 100052051 10974 &
./test.sh 100052051 8649 &
./test.sh 100052060 10974 &
./test.sh 100052060 8649 &
./test.sh 100052078 10974 &
./test.sh 100052078 8649 &
./test.sh 100052086 10974 &
./test.sh 100052086 8649 &
exit;
test_script2.sh:
---------------
#!/bin/sh
./test.sh 100052051 11698 &
./test.sh 100052051 11777 &
./test.sh 100052060 11698 &
./test.sh 100052060 11777 &
./test.sh 100052078 11698 &
./test.sh 100052078 11777 &
./test.sh 100052086 11698 &
exit;
test.sh:
--------
#!/bin/sh
sqlplus HRUAMGR/HRUAMGR@MALDEV_DW1<<EOF
ALTER SESSION ENABLE PARALLEL DML;
insert into legacy_system (product, item, supp, type, g_ind, d_ind, o_num) values (7896544,$1,$2,'TEST',NULL,NULL,NULL);
commit;
exit;
EOF
I am calling test_global.sh through the UNIX session and all the records got inserted into LEGACY_SYSTEM table and few rows are missing from PRICE_CHANGE table.
If I remove the '&' symbol and execute, the results are perfect. But the requirement is not to remove the '&' symbol. I have been facing this problem for the past 1 month.
I was just wondering how Oracle manages multiple sessions in a database performing DML. I believe this is related to 'Read Consistency' and I tried to search for the same but could not get any satisfactory online documents.
CASE 1:
user A logs in to a database1
issues select on table A and then inserts 4 rows
user B logs in to databse1
issues select on table A and then inserts 5 rows
issues rollback
user C logs in to a database1
issues select on table A and then inserts 6 rows
issues commit
How many rows can user C see in the table A when he issues select?
CASE 2:
user A logs in to a database1
issues select on table A and then inserts 4 rows
user B logs in to databse1
issues select on table A and then inserts 5 rows
user C logs in to a database1
user B issues rollback
user C issues select on table A and then inserts 6 rows
issues commit
How many rows can user C see in the table A when he issues select?
NOTE: All the users are currently logged in to the same database and none has logged out.
In my form i have a multi record block which is based on a table in which i am performing insertion, deletion, updation.
My requirement is as below,
While saving a records i want to check distinct value of specific item can not be greater than some value say 2. Then only my records should be saved. I am planning to populate those values in a collection table type and take the distinct values from it. How to populate record values in a collection table type.
We are facing serious problem while refreshing materialized views using fast refresh option in ORACLE..For the very first time we are performing Complete refresh of data from DB1 to DB2 for few tables.Ongoing we are performing Fast Refresh.Sometimes the fast refresh works fine without any error and sometimes it fails with the below error.
ERROR at line 1:
ORA-32320: REFRESH FAST of "CIR"."C_BO_COMM" unsupported after
cointainer table
PMOPs
ORA-06512: at "SYS.DBMS_SNAPSHOT", line 803
ORA-06512: at "SYS.DBMS_SNAPSHOT", line 860
ORA-06512: at "SYS.DBMS_SNAPSHOT", line 841
ORA-06512: at line 1
let me know when will this error occur during Fast Refresh.
understanding rman scn based Backup Algorithm .. In SCN based backup do rman generate backup from archivelogs ???
View 1 Replies View RelatedI have 2 seperate rman backup one backup is of only datafile ,spfile and controlfile which i am able to restore and recover without any problem say bkp1 taken at 10 am
other set of backup is of only archive log files of same day but later time than datafile backup say bkp2 taken at 8pm...if i restore and recover bkp1 and try to restore bkp2 it gives error datafile exist (possibly due to fact both backup have control file)
if i just restore bkp1 and try to restore bkp2 so that i can do one recover at time
it gives error datafile permission issue (possibly due to fact both backup have control file)
I want to restore database upto 8pm time how can i use both bkp1 and bkp2 to do it(restore datafiles and apply all archive logs on it)
In my company we have few 10.2.0.2 databases and few 10.2.0.4 databases.Currently we are monitoring using OEM 10g (10.1.0.5.0) installed on a seperate server.
Planning to install OEM on a new server as part of server room renovation. let me know the process of installing OEM on a different server without loosing the historical data?Also wondering if the OEM 11g will support the current databases?
I have a windows 2008 server.
I already installled oracle software for oracle_home C:/oracle/product/11.2.0/dbhome_1 for database DEV.
Now, I want to create new database with different oracle_home on same server with oracle_home as C:/oracle/product/11.2.0/dbhome_2 and SID as TEST.
Do I have to istall oracle software again ? if yes, what about inventoryfiles in C:\Program Files\Oracle\Inventory ? wont they be overwritten ?
We are getting a consultant to upgrade an Oracle 9i installation to 11g R2. The current installation has 6 different databases installed on the same server. Each database is a different customer so for reasons of security we have requested that this be split into 6 virtual machines with one database per machine.
The consultant suggested that they could install the 11g database once and then just make copies (which would all have the same instance name. We are told that the TNS names can be configured so clients are directed to the right database.
As we can give grant on sequence and can use in different schema of same database, can same sequence be used across databases?
View 2 Replies View RelatedI want copy some tables from one database to another. Which of following is the fastest way:
- INSERT /*+APPEND*/ via database link
- Transportable Tablespaces
- datapump IMP/EXP (I supposed it's the slowest way)
I'm just wondering because i didn't find any example / test databases for Oracle. Thus MS has its "northwind", mysql its "world" or "sakila".
Additionally, the tables view is quiet confusing, referring to all those generated tables (
[URL]........
I tried right-click + apply filter *remove generated* but it didn't work.
I have cloned the database into another using DBMS_METADTA API (export the metadata in xml form and recreate it on destination). I need to synch these two periodically. I need to update the XML to synchronize the databases.
I don't have 11g to use DBMS_METADATA_DIFF API.
As part of our project, we need to perform table comparisons in two different databases. I am currently looking for various options to accomplish this.
One of them is doing minus operation between these two tables. Also, i have looked at the data compare option in toad utility.
I have a query that queries an Oracle 10g database see below
SELECT PLOGDBY, COUNT(PLOGDBY) AS "Calls Logged"
FROM LCC_TPROBLEM WHERE
(PLOGDBY = 'WOODWARD_A' OR
PLOGDBY = 'BASSETT_A' OR
PLOGDBY = 'MISTRY_R' OR
PLOGDBY = 'MEADOWS_K' OR
PLOGDBY = 'SKINNER_R' OR
PLOGDBY = 'DAVIS_Z' OR
PLOGDBY = 'BLOUNT_L' OR
[Code]...
This gives me the result:
PLOGDBY Calls Logged
DAVIS_Z 11
MEADOWS_K 26
MISTRY_R 47
PATEL_RI 53
SKINNER_R 21
THANDI_D 58
WOODWARD_A 30
I also have a query that queries a SQL Server database:
Select Agent, SUM([acd calls which have rung the agent])As CallsRung
FROM Dashboard_stats
Where Date = DATEADD(DAY,DATEDIFF(DAY,'20000101',GETDATE())-1,'20000101')
Group by Agent
This gives me the result:
Davis_Z22
Meadows_K31
Mistry_R54
Patel_R65
Skinner_R35
Thandi_D89
Woodward_A34
What I want to do is write a query that returns the result:
Query 1(Oracle database) as a percentage of Query 2 (SQL Server database)
Results would be:
Davis_Z50%
Meadows_K83%
Mistry_R87%
Patel_R81%
Skinner_R60%
Thandi_D65%
Woodward_A88%
I have 2 databases( A and .. which shares common ORACLE_HOME....
i configured 2 listeners through netca in different ports....
and both listener names are different...
But when I stop LISTENER.. both the listeners are getting stop and when i start both the listeners are up and running.. and I could see database B's services.. and vice versa...and when i tnsping , i am able to see port 1521 default port for database A.
when I run I bash, i have included ORACLE_HOME path... so should i add oracle_SID also there..if so, which SID should I add A or B?
I have two tables having same name in two different databases.i have to create a new table with these two tables in a new database.
But if one table have 2 columns and another table having three columns I have to create a new table with three columns..
I've just found out that 12cR1 will not (in all likelihood) allow "flashback database" for pluggable DBs. Am I the only one disappointed by that ? I use flashback db (+replay) a lot to revert and replay automated tests and I had plans to consolidate tenths of test environments into PDBs.
View 3 Replies View Relatedwhile using top command it shows multiple processes usage.Actually we have 3 databases in the same server.how to find out which database process and user session hogging the cpu resources
View 2 Replies View RelatedI have a production DB that has the oracle tuning pack option turned on, I occasionally need to clone the DB to a test server where I do not use the tuning pack, If I select from DBA_feature_usage_statistics it tells me that the Options pack is turned on for the Prod db DBID but not for the cloned db dbid. Do I need to worry about this old data being left around in my database? If so is there anyway to remove it? I am running 10.2.0.4
View 5 Replies View RelatedI have a requirement to move the databases to a new server.
Existing set up:
database1-40GB,
database2-40GB,
database3-180GB
Oracle version - 11.2.0.2
OS - UNIX AIX-5.3
New Server (setup):
database1-40GB,
database2-40GB,
database3-180GB
Oracle version - 11.2.0.3
OS - UNIX AIX-6.1
My query is what is the best way to move the database to new server(UNIX AIX 6.1)?
a) can i upgrade the database to 11.2.0.3 before moving to new server or its better to move the database in 11.2.0.2 and then upgrade there?
b)is export and import of database can work for moving this?
c) RMAN backup and restoration can be done for moving databases?
d)simply moving all the related files to new server and start upgrading it from there?