i have batch file name exptab,to export all table in database
for example
@echo off
exp scott/tigger tables=emp file=C:impemp.dmp LOG =C:implog1
batch file working fine when i click on it,
I need to use the Scheduled Tasks in windows 2003 so when the computer startup batch file will run automatically i follow all the step but its not working i try the same story with windows xp still problem exist .
As I was observing the space issues on my db server. I found that there is lots of Trm and Trc file which is being created very much frequently. Due to this its consuming lots of space even the size of each files is not more than 1Mb.
For cleaning I am deleting all the trm and trc files mannully using DEL command Os level. How can i schedule the purging of trm and trc files.
I have a scenario where the requirement is to create a "SEQ NO". Based on the 'Tasks' & it's creation date. Where there are multiple 'tasks' for a single SR.
The following example will make it clear SR TASK Created on SEQ NO 11009 2345 14/10/1988 12:15:17 1 11009 2346 14/10/1988 12:15:57 2 11234 - - -
If there is a SR which has no task then the SEQ NO should be 0. And the seq no should be in the order of the created on field & if 2 tasks are created at the same time i.e suppose 2345 & 2346 have the same time stamp then the SEQ no should be 1,2 respectively n not both as 1.
I am using expdp command to export the table by specifying Query parameter. But i am unable to export the table based on the condition.
Ex:EXPDP username/password dumpfile=employee.dmp logfile=emp.log directory=DATADIR_EXP TABLES=EMPLOYEE query=EMPLOYEE:"UPDATED_TIME >= '04-JUN-13' AND UPDATED_TIME >= '05-AUG-13'" Estimate in progress using BLOCKS method...Processing object type TABLE_ EXPORT/ TABLE/ TABLE_ DATATotal estimation using BLOCKS method: 3 GBORA-31693: Table data object "<username>"."EMPLOYEE" failed to load/unload and is being skipped due to error:ORA-00933: SQL command not properly endedMaster table "<username>"."EMPLOYEE" successfully loaded/unloaded...Dump file set for <username>.SYS_EXPORT_TABLE_01 is: E:IMPDPemployee.dmpJob "<username>"."SYS_EXPORT_TABLE_01" completed with 1 error(s) at 12:34:45 Oracle 11g,
I want to export a table ( using exp or expdp ) from client machine. Dump file should be created to client machine. Is this possible ? How to do this ?
I want to make report of processing time for tasks. Every task has several stages and connected with them modification dates. I need calculate total processing time for every task.
In my calculations i have to consider only "working" hours. There is time window table (For example working hours: from 8 a.m to 4 p.m, but can be different thats why i use table with id - different id = different time window). Scripts i used to simulate problem:
create table t ( id number, s_date date, e_date date ); declare [code]........
table t constrains "working" hours:
12011-04-10 08:00:002011-04-10 16:00:00 12011-04-11 08:00:002011-04-11 16:00:00 ... but could be also
table w contains tasks history processing by team (8 hours for tenet, rest time (out of hours defined in table t) of working trough incident should be ignored/omitted in calculations):
At the beginning I tried this query, but I don't know how consider/dose time window from table t
select id, change_date, round( to_number(change_date-lag(change_date) over (partition by id order by change_date))*24*60*60) sec from w;
12011-04-10 10:00:00 12011-04-11 02:08:1658096 - sholud be 21600 (sum time only for 2011-04-10 16:00:00) 12011-04-11 16:23:1051294 - should be 28800 (sum time only from 2011-04-11 08:00:00 to 2011-04-11 16:00:00) 22011-04-12 08:12:30 22011-04-13 00:40:1559265 - should be 28050 (sum time only from 2011-04-12 08:12:30 to 2011-04-12 16:00:00) 32011-04-14 09:37:54
we have 2 table same structure one of them is empty and the pther one is contain data the vendor do the insert as select but i found he is wrong due to there duplicated ,now i want to use export and then rename the table and then import but i need with export do a condition
exp user/pass tables=MTR_EPPC_CALLED_DATA file=MTR_EPPC_CALLED_DATA.dmp query="where callstarttime >=to_date('01122012','ddmmyyyy') and callstarttime <=to_date('31122012','ddmmyyyy')"
but it's seem the query take one condition how can i use this above condition in export ???also my friend say there is way to insert with rowid is this possible ??
I'm having a problem with performing backups from EM DB Control on two of my Oracle 10g 10.2.0.4 systems. When I configure and submit any backup job from EM (either online or offline), the following happens:
- When the job is due to run a Pre backup task has status of 'Scheduled'
- Prebackup task completes and status changes to 'Succeeded'
- Backup task appears with status of 'Scheduled' and value of started time within the next minute
- Status of backup task never changes from 'Scheduled', and the started time keeps updating to the current time
- This goes on indefinitely until I abort the job
At the moment I have scheduled tasks configured to run an RMAN job every night, but I need to get this working through Enterprise Manager Database control.
Something else which has happended on one of these systems is that after applying some Microsoft Critical updates the database control service would not start. The only way I could find to resolve this was to recreate the Enterprise Manager repository. Could there be a link here at all?
I've also posted this here: [URL] but with no response so far.
i have installed oracle 11gr2 on centos 4.7,,but after that when i invoke command for startup in sqlplus for sys user it shows an insufficient privileges error. and what are the post installation tasks? is the database installed also?what is the name of database and where is it?
the database (11gR2) is located on Linux server. A business application is installed on a Windows server with an Oracle client 11g.The application is able to start a datapump export, but as matter of fact the dumpfiles are always written to the Linux-server. The directoryobject is defined as DATA_PUMP_DIR (which is the default directory).
Now we are supposed to change the datapump export in a matterthat the dumpfiles get written to the Windows server. Creating a new directory (e.g. c:datapump) and starting than the datapump from theclient always raises the errors
ORA-39002: ...ORA-39070: ...ORA-29283: ...ORA-06512: in "SYS.UTL_FILE", Zeile 536ORA-29283: ...
Is it possible at all to start a datapump export from a Windows client and writing the dumpfiles to the Windows server itself? Or do the dumpfiles always written to the database-server?
I have set up two scheduled jobs in oracle 11g for executiong a procedure in specific time intervals. One as Hourly and other as minutely.But the scheduled jobs are not running as per schedule. i.e the minutely job is not running every minute. But If i see the run logs for some period it would have run every minute, and suddenly it will not execute and give a gap of 30 mins or 1 hour and then run again.
P.s. The job priority is set as 3. and except for these two jobs there are no other jobs set. But this is an important job which needs to be run every minute.
JOB_QUEUE_PROCESSES is set to 1000, which is max. And there are only 3 scheduled jobs in the whole DB. BEGIN SYS.DBMS_SCHEDULER.CREATE_JOB ( job_name => 'LOGTBLPURGE'
A procedure in a package uses a database link. The database link is defined without a username and a password. The user that uses the database link is supposed to be present in the opposite database as wel.
When the procedure is started in SQL*Plus or TOAD it runs perfect. But the procedure has to run in a scheduled job. And that doesn't work. The procedure fails because of: ORA-01017 invalid username/password. The user that is used to run the scheduled job is the right one.
I have created a materialized view log on table testing_mview_rev in 11g:
SQL> CREATE MATERIALIZED VIEW LOG ON testing_mview_rev WITH ROWID;
Materialized view log created.
And a MV on it in 10g:
SQL> create materialized view testing_mview_1_rev REFRESH Force start with (sysdate) next (sysdate+1/1440) with rowid as select * from testing_mview_rev@S58_TO_S56; Materialized view created.
Now this is suppose to automatically refresh itself every minute but refresh is failing. However manual refresh is working fine.
what I have missed here. is there any other way to schedule the MV to run every minute?
I have a job running under HUMADM user and would like disable or drop it.I do not have the login credentials of HUMADM user but can log into the database as system/sys user.
When I run the below commands as sys/system user it says humadmjob must be a job or it is not running.This is because it is looking for the job under sys/system user.
am facing an issue with scheduled jobs. The below script will create a scheduled job to insert a record for every 2 mins into the table "job_table".
--Table script CREATE TABLE job_table (now DATE); --Stored procedure run by scheduled job CREATE OR REPLACE PROCEDURE test1 IS BEGIN
[code]...
The scheduled job does not require any special grant privillege as the PUBLIC has access to run it. Anyhow the same set of instruction working for my friend on his machine but not on my environment.
BEGIN dbms_scheduler.create_schedule( schedule_name => 'INTERVAL_EVERY_1_MINUTES', start_date => trunc(sysdate)+18/24, repeat_interval => 'freq=MINUTELY;interval=1', comments => 'Runtime: Every day all 1 minutes'); END;-- program code BEGIN
[code]....
all above three code compiled successfully . but it is not running automatically and not run every 1 minute.
I try to run this manuall by using following code
BEGIN dbms_scheduler.run_job('JOB_FEED_UK_DATA',TRUE); END;
It works fine..
create or replace PROCEDURE LOAD_CUSTOMERS_UK IS BEGIN Insert into CUSTOMERS_UK (ID, NAME, CITY) VALUES (1,'SMITH','LONDON'); Insert into CUSTOMERS_UK (ID, NAME, CITY) VALUES (2,'JONES','LONDON'); Insert into CUSTOMERS_UK (ID, NAME, CITY) VALUES (3,'BROWN','BRIGHTON');
We are using one application. This application is having 2types of jobs. Under each job again there are each 5 sub jobs. All the jobs are scheduled jobs. Now the requirement is The main 2 jobs we are going to run on some specified days( for this we need to mention on what days we want to run.once we gave those day the next steps will start).
for example 1 main job is scheduled on Monday ( we need to schedule it on monday-- we need to write a program for this) under this every sub job will run.these sub jobs are dependent jobs.if sub job completed successfully it should send an email to the distributed mail group which says like " <Job_name> completed successfully" if it fails also it needs to send an email to the distributed group. for each sub job it should send either success/falure mail to the distriuted mail group.
once after completing all the sub jobs successfully it should send a mail to the distributed mail group saying that The main job got completed successfully.these two main jobs will run paralelly but the the second main job will run after 30 mins of 1st job starts.
in the second job also will work as same as job1 process.it should also send mails to the distributed mail groups.The jobs are already scheduled.No need to scheduled the jobs.need to write a pl/sql program like it will check all the jobs and respected sub jobs if any thing got completed/failed it should send an email to the distributed mail group automatically. For that it should monitor all the jobs..that's the requirement.
I have already scheduled all the jobs. now want to write a script for the above requiement.
Yesterday, there were performance issue at server. So today, when i am generating report for that particular period, found snapshot id sequence is serially but with skipped hourly timed. Instead of generating report at 15:30, it generated at 16:30.
Enter value for num_days: 2
Listing the last 2 days of Completed Snapshots Snap Instance DB Name Snap Id Snap Started Level ------------ ------------ --------- ------------------ ----- tagidev TAGIDEV 2857 10 Sep 2013 00:30 1 2858 10 Sep 2013 01:30 1 2859 10 Sep 2013 02:30 1 2860 10 Sep 2013 03:30 1 2861 10 Sep 2013 04:30 1 2862 10 Sep 2013 05:31 1
[code]....
Below are the details at alert log -
Tue Sep 10 14:28:20 2013 Thread 1 cannot allocate new log, sequence 7029 Checkpoint not complete Current log# 2 seq# 7028 mem# 0: E:APPORACLEORADATATAGIDEVREDO02.LOG Thread 1 advanced to log sequence 7029 (LGWR switch)
[code]....
1) why snap didn't started at 15:30?
2) since database just started at the scheduled time of AWR snap time. But generated at 16:32 instead of 16:30, though last services "SMCO" is started at 16:42. How it snap id generated for this particular time?
3) what does "kewastUnPackStats(): bad magic 1 (0x000000001B3CE48D, 0)" mean?
I copied a sqlplus scheduled job procedure. Procedure successfully created .I fount it here [How to Schedule RMAN Daily Backup? [URL]
The procedure is successfully created. Now, doing the following operations:
1) running that job 2) checking whether that job is running or not 3) listing all other jobs 4) unschedule a scheduled job 5) delete a job 6) edit a job
There is a readonly user on our reporting server. Developers want to use global temporary tables with this user. I don't want the user to have permissions other than readonly.I can grant the user CREATE TABLE privilege and did not grant quota on any permanent tablespace, therefore user would not be unable to create permanent tables but still should be able to create global temporary tables.
Question is would a user with such permissions still be able to utilize temp tables as part of a scheduled job?