i need to delete a record from table but it showed a error for foreign key constraint so i disabled the constraint and again deleted, now the row is deleted.
Again inserted another values instead the deleted value. after that i tried to alter the constraint but i received the error ORA-02298 cannot validate - Parent keys not found
I have two oracle 9i databases A and B with complete equal schema. Only the data is different. I want to import all table data from A to B. The problem is that there are duplicate primary keys. Therefore I want to insert data with new primary keys (all referencing tables are concerned too).
I have taken database backup using exp command and when I try to import in other pc the foreign keys are not imported. It saying error message that no matching unique key or primary key for this column.
how will i take backup including with primary keys?
I have csv file which having parent and child related data.i need to load parent data in to parent table and get parent reference id and store in child table with child data. i am not able to find how to get parenet refernce id using control file. In my csv file i have 2 parent name rows. I need to create one A record in parent table and get that parent primary(P_ID) for 'A' record and put into child table for c_name test1 and test2 records.
My controll file ------------------- LOAD DATA INFILE 'mydata.data' APPEND INTO TABLE parent when (1:1) = 'p' fields terminated by ',' optionally enclosed by '"' trailing nullcols
[code].....
able to load parent data and generate P_ID, but i am not able to get P_ID for child records.
i am running the query from which i am getting below mentioned error how can i find the record which is not there in parent table ora02291 integrety constrain violated and parent key not found.
I'm new to oracle and solaris.got a dmp file that I need to restore from. I've managed to successfully restore it in windows but not in solaris. the following command:
D:>impdp TESTIMP/TESTIMP TABLES=ENVDB DIRECTORY=Rini DUMPFILE=envdb_11g.dmp REM AP_SCHEMA=TESTIMP:STP11G REMAP_TABLESPACE=TBS_TESTIMP:TBS_STP11G; Import: Release 11.2.0.1.0 - Production on Thu Sep 9 15:16:45 2010 Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved. Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production With the Partitioning, OLAP, Data Mining and Real Application Testing options ORA-39002: invalid operation ORA-39166: Object TESTIMP.ENVDB was not found.
I have a parent table EMPLOYEE which includes columns (sysid, serviceno,employeename...) sysid is Primary key, serviceno is Unique key and I have child table DEPENDENT includes columns (sysid,employee_sysid,name,dob...) here again SYSID is primary key for DEPENDENTS table, employee_sysid is Foreign key of EMPLOYEE table.
Now I want to change SYSID (using sequence) in EMPLOYEE table which need to be update in DEPENDENTS table as well
Note: I have 10000 records in EMPLOYEE table as well as I have 5 more child tables which need to update new SYSID.
I was reading the documentation for oracle 11gr2, with reference to URL>.....
The following examples show how to correctly choose the cluster key and set the HASH IS, SIZE, and HASHKEYS parameters. For all examples, assume that the data block size is 2K and that on average, 1950 bytes of each block is available data space (block size minus overhead).Note that 34 hash keys are assigned for each data block
how they arrive at 34 hash keys because another portion of the document states
This space determines the maximum number of cluster or hash values stored in a data block. If SIZE is not a divisor of the data block size, then Oracle Database uses the next largest divisor.
if that is the case, then number of hash keys should be 1900/55 = 34.55 which should have rounded up to 35.
I have an Image Type on a forum page. I want a default "not-found" image to display if the BLOB column value is null or if there is no data for that search value. The image is stored with the app: #APP_IMAGES#not-found.png
We transferred our Oracle database 11.1.0.7 from windows 2003 enterprise edition 32 bit to windows 2008 enterprise edition server 64 bit.Database is working fine but we have 53 uncompiled objects which are related to OLAPSYS and public as follows
i have installed oracle 10g. normally with that URL address
we can navigate to this URL in our browser and log in as the SYS user with the associated password, and connect as SYSDBA. Enterprise Manager can be used to perform common database administration tasks... but..,in my system when this procedure is run i'm receiving an error stating that the server cannot be found...
Instead of using localhost I used the hostname or the IP address of the box without success.The whole installation process went fine from scratch. No errors, no warnings, nothing.I am using the following environment:
CentOS 6.3 Oracle 11.2.0.3 EE Apex 4.1.1
With exec dbms_xdb.sethttpport(7778); I enabled the above port number. If I look at the output of the lsnrctl status I see that my local listener is also listening on 7778. I have no errors in the alert.log and no other trace files where I can have a look at.
I previously received REP-51002; searched over the forums,& made changed to the rwservlet properties file by setting my Server,setting the trace on,& changing my host file in windows.
Errors received now: REP-52266: The in-process Reports Server rep_fsprod failed to start.oracle.reports.RWException: IDL:oracle/reports/RWException:1.0
I also tried these commands:
rwclient server=rep_CC-BA0C7240E38E report=test destype=cache desformat=html GIVES ERROR:- REP-0178: can not connect to reports server
rwdiag -findall Broadcast mechanism used to locate servers Channel address: 228.5.6.7 Channel port: 14021 REP-50503 No server found in the network.
I am not able to get the relevant segment from the above information
SQL> select segment_name, segment_type, owner 2 from dba_extents 3 where file_id = 4 4 and 756652 between block_id 5 and block_id + blocks -1;
no rows selected
DBVERIFY Summary DBVERIFY - Verification complete Total Pages Examined : 3932160 Total Pages Processed (Data) : 3119107 Total Pages Failing (Data) : 0 Total Pages Processed (Index): 755048
[code]....
I have uploaded the complete logfile.
Below is a part of logfile
DBVERIFY - Verification starting : FILE = /prd/dvp/ora/oradata/LHF/disk06/gds_t01_01.dbf Block Checking: DBA = 21728172, Block Type = KTB-managed data block **** kdxcoavs = -84 < 0, avail = 3129 ---- end index block validation Page 756652 failed with check code 6401 ##not here that 756652 is the same block# mentioned in v$database_block_corruption
On last Saturday we encountered error 'TNS-12519: TNS:no appropriate service handler found" in our listener log for some moment of time let say around 5mins and later it got all right itself.
don't know if I can repeat it. It occured while converting a database to archivelog mode, after completing a Data Pump export/import upgrade from 11.1.0.7. I fixed it like this: juno> startup mount;
ORACLE instance started.
Total System Global Area 1.7108E+10 bytes Fixed Size 2175440 bytes Variable Size 8858373680 bytes Database Buffers 8220835840 bytes Redo Buffers 26886144 bytes Database mounted.
juno> alter database archivelog; alter database archivelog
ERROR at line 1: ORA-19821: an intentionally corrupt log file was found
juno> recover database until cancel; Media recovery complete.
juno> alter database open resetlogs; Database altered. juno> shu immediate; Database closed. Database dismounted. ORACLE instance shut down. juno> startup mount; ORACLE instance started. [code]...
I want to load lakhs of records into a table. My problem is when after loading the ΒΌ of records my process is abend due to the size of my rollback segment area. I don't have an option to increase it. So, Is there any way to go for intermediate commits when I am using the imp or sqlldr utilities to load the entire data without abend?