Server Administration :: Why Archive Should Be Turned ON Before Turning On Flashback
Aug 5, 2010
I have read it in books that flashback uses undo data to create the flashback data or to flashback the database to a time in the past.Then, what is the role of archive files in flashback operation. Why it is mandatory to turn on archiving before turning on flashback. Also, if you remove the latest archive files, you can NOT flashback the data to a time in past (Oracle complains of missing archive files).
I need store history for two tables in my system. I thought that Flashback Data Archive will be the best option. There is also another ways to do this but don't focus on this. I need to to this by FDA (Flashback Data Archive);
So my prerequisite was to create tablespace and flash back archive, and alter table to be archived.
alter table teta_admin.t_prac flashback archive audit_flash_archive;
and everything works fine but on sys user. i can query this table using "as of timestamp" clause
select prac_id, imie, imie_2, nazwisko, nr_ew from teta_admin.t_prac as of timestamp to_timestamp('2011-08-23 08:20:00','yyyy-mm-dd hh24:mi:ss')
but final construction of idea was to create additional user (interface), grant select on teta_admin.t_prac object and query archive data from interface user. and this is point of my failure. this don't work on new user.
interface user have such sys privs:
SQL> SELECT * FROM dba_sys_privs 2 WHERE grantee = 'INTERFACE'; GRANTEE PRIVILEGE ADM ------------------------------ ---------------------------------------- --- INTERFACE CREATE SESSION NO
and table privs:
SQL> SELECT * FROM dba_tab_privs 2 WHERE grantee = 'INTERFACE';
Is there anyway to backup Flashback Data Archive (FBDA) data and can be restore on new database. I cannot find Oracle's document or any document explain about backing up this data.
SYS@boston>select name, value, issys_modifiable from v$parameter where name='db _recovery_file_dest_size';
NAME -------------------------------------------------------------------------------- VALUE -------------------------------------------------------------------------------- ISSYS_MOD --------- db_recovery_file_dest_size 9663676416 IMMEDIATE
on weekends we have too many archive logs generated .i have taken the data of a week and found that average archive log generated from monday to friday is 7 files per day but on satuarday and sunday the average is 60 files and FG1 gets full. on weekends we have all type of backups running like incremental,archival and logical backups and on sunday we have full physical backup
what is the reason of too many archive log files generations at weekends. is it due to hot and logical backups , if yes then how ?
Our database was generating archivelog(50MB) every 30seconds! I think this is not normal because what I did is open our database, I was the only one who is connected, I'm not running anything, but our database is still generating archivelogs!
Our redo logs: 6groups 3members.
This are the things I saw on our alert logs: - advanced to log sequence - cannot allocate new log, sequence - checkpoint not complete - private strand flush not complete
What I did is change the log mode of our database to noarchivelog then open the database, then returned it to archivelog mode then that fixed the problem. But the thing is after 6hours its abnormal behavior goes back again.
We had a database (DB A)that is having version 9.2.0.6.0.This DB is having multiple tables and volume of 6 million in individual tables.Another database is also 9.2.0.6.0 (DB B), this DB has Mviews pointing to DB A. Mviews are refreshed in every 15 mins, with fast refresh option in 90% cases and remaining having complete refresh.
Last weekend we have migrated DB 2 to version 10.2.0.4.0 - 64bi and on another server.After version upgrade and DB migration complete refresh was done once for all mviews.
Now DB A is generating huge amount of archive log and also it's UNDO space is getting fully consumed causing performance issue and DB hang.what has gone wrong and what we can do to improve response of DB A and also to reduce size of Archive log ?
ora-00257 archiver error. connect internal only until freed
when we tried to remove the unwanted arc files thro ASMCMD,we are getting below error:
ASMCMD> rm -ef 2011_04_05/ Unknown option: e usage: rm [-rf] <name1 name2 . . .> ASMCMD> rm -rf 2011_04_05/ ORA-15032: not all alterations performed ORA-15028: ASM file '+XCOM_BACKUP_DG/TXCOM/ARCHIVELOG/2011_04_05/thread_2_seq_27215.1143.747641143' not dropped; currently being accessed (DBD ERROR: OCIStmtExecute) ORA-15032: not all alterations performed ORA-15028: ASM file '+XCOM_BACKUP_DG/TXCOM/ARCHIVELOG/2011_04_05/thread_3_seq_21762.826.747641143' not dropped; currently being accessed (DBD ERROR: OCIStmtExecute) ORA-15032: not all alterations performed ORA-15177: cannot operate on system aliases (DBD ERROR: OCIStmtExecute)
I'm facing problem with archive log file size, Archive logs are generated with only of 90m or 92m or 94m(Variable sizes of less than 100m), Although i had set 100m for each of my redo log file. Here i'm providing my create db script for your reference. I want to know why the log switches before it reaches 100m.Is there any connection of intial 10m for my .dbf files.
We are facing a different issue in our database. From yesterday night, the archive log generated with 5 digit. But it supposed to be 6 digit. Hence we are not able to apply the logs in DR Location.
In normal days size of archives generated in a day is 14-15GB. But since yesterday morning, almost 150GB of archives have been generated and are still getting generated(200MB every 1-2 minutes).
There was a sudden reboot of server yesterday morning. At that time there was heavy load of transactions on database. Can it be a reason that smon is still doing recovery? (I am not sure on this). Also, Undo tablespace is increased from 18 GB to 50 GB since yesterday (autoextend on).
Now we are running out of space for archive file system (can't delete them also until they are transferred to DR) Size of redo log is 200MB. This database supports around 2500 users.
performance wise I don't see any hit. Also wait events are normal. (only few db file sequential read) finding the query/session which are causing this much huge amount of archives?
i am using oracle 10g on solaris 10 os.currently archived log is generated by size wise 52 mb.i want to know whar is the best practice for archive log generation . it should be time interval or size wise.
Application Express 4.1.1.00.23 Internet Explorer - 8 Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
I am about to embark on my first APEX development and have got far as the login screen. I disabled JavaScript just to see what would happen if someone had it disable, it turned out by disabling javascript the whole login region dissapeared. I am wondering how to develop the application so it works when javascript is disabled. I am using the cloudy theme. Not sure how to develop the application as its got to work with JavaScript disabled.
We use apex and my professor wants us to import a table model into apex. everytime i try and import my table it says "your export file is not supported". i asked my professor what to do and he said "You should gen a ddl file from designer. Then use that to upload into apex and then run.
Can I disable tracing just on deadlocks events on 11gR2?One of our applications provoke several deadlocks. I've warned the developers team several times, but it's taking time to solve it. However, these deadlocks generate trace files of several sizes (from 300M to 3G ..)The Deadlock Graph on them are very useful but the PROCESS STATE section are too long, and until the developers team fix the problem they are just filling up my user dump dest. Can I disable them or disable the PROCESS STATE section from them?
I created a new form for Oracle Apps, At first when I ran the form from the application all the fields backgrounds were black, so I changed the background in the property palette to white and foreground to black.
Now it shows fine but when I close this form and open another those fields are now blacked out. What should I do, I know the problem comes from the new form.
I am not able to delete one month old archive log file manually on windows which doesn't having info about the standby on v$archived_log view of primary database. the sequence were already applied to the standby database. It shows the status as deleted in v$archived_log. while deleting the file manually. it showing an error like another program or person is using it.
Developer is importing some large volumes of data. As the filesystem was filling up fast, I removed the all the archive log files. Will this affect the functioning of database? It is a development environment.
Can I use flashback in application, especially DBMS_FLASHBACK.TRANSACTION_BACKOUT. I import some information (from files) in my database and after that I want to revert that transaction(delete imported data). I did some test I took xid of the transaction from v$transaction and then executed
DBMS_FLASHBACK.TRANSACTION_BACKOUT
with that XID for parameter. Everything is fine but that time there was only one transaction, when there are many transaction how can I get the exact XID. I can use dbms_transaction.local_transaction_id to get information for transaction which is running. And when i tried to use flashback that way a deadlock occur and db killed the other transaction.
I've just found out that 12cR1 will not (in all likelihood) allow "flashback database" for pluggable DBs. Am I the only one disappointed by that ? I use flashback db (+replay) a lot to revert and replay automated tests and I had plans to consolidate tenths of test environments into PDBs.