Server Utilities :: Tablespace Management?
May 22, 2012how we can perform the tablespace management? monitor the tablespace and tune the tablespace space.
View 1 Replieshow we can perform the tablespace management? monitor the tablespace and tune the tablespace space.
View 1 RepliesHow to import dump into specific tablespace instead of default tablespace users.
I want to import my dump file to newly created tablespace ,so how can i do that . I have created new user called cvm and while creating it i mentioned default tablespace to newly created tablespace . But when i try to import my dumo file it goes to users tablespace .
i have 80 gb table d9040 which residing in separate tablespace in existing database(10g) OLTP .
i m going to import(impdp) an export(expdp) dump on 11r2 on which ASM is residing.
I have following two options available while importing
1.Put all the datafiles of that particular tabesace on a single disk .
or
2.Put half datafiles on disk 1 and remaining half on disk 2.
I am trying to drop 90 columns from a big partitioned table. I was trying a physical drop first and since it is taking longer time I decided make the columns unused state and drop them. However I was able to set them to unused state.
Now I am trying to drop those unused columns from the table, it is running since 22hours Apporox. I am keep increasing the undo tablespace to retain the undo data.
I also have decreased the undo_retention to 300 from 900.
My question is there any better way to drop these columns. And is there any way to flush out the data from undo.
C:UsersNeetesh>expdp system/*****@orcl2 dumpfile=temporary.dmp tables=testuser.test,testuser.test2
Export: Release 11.2.0.1.0 - Production on Fri Oct 19 16:39:06 2012
Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
Starting "SYSTEM"."SYS_EXPORT_TABLE_01": system/*****@orcl2 dumpfile=temporary.dmp tables=testuser.test,testuser.test2
[code]...
then i import it in new user
SQL> create user temp identified by temp;
User created.
SQL> grant create session ,create table to temp;
Grant succeeded.
SQL> alter user temp quota 10 m on users;
[code]...
it gives 2 errors both are same as -ORA-01950: no privileges on tablespace 'new_tbl'
exported table are exist in this 'new_tbl' tablespace but this is not exist in importing database.then is there any way to import these tables in 'users' tablespace or other tablespace other than 'new_tbl'?
How to import a dump file into different tablespace.
normally when i am importing a dump file it goes to the table space where it was exported but i need to import in to different tablespace
I want to transportable tablespace feature of EXP/IMP. I did it with no error. But when I check user by user I found that no Function Procedures were imported. How to deal with these objects using transportable tablespace feature.
View 8 Replies View RelatedI am trying to refresh tablespace containing indexes only. Can you let me know whether i am on the right track. Below is my query :
expdp directory=expdp DUMPFILE=user_index.dmp LOGFILE=user_index.log TABLESPACES=user_index
impdp directory=expdp dumpfile=user_index.dmp logfile=user_index_impdp.log tablespaces=user_index
I have been asked to take the export of a schema and then asked to remove all dependencies on tablespace. Provide me the syntax.
View 1 Replies View RelatedI have taken an export using expdp of schema, data of the schema spread across different tablespaces , now i want to import the data to only one tablespace.
View 1 Replies View RelatedI want to move all data from my users tablespace to anohter new tablespace i have created name test1
so how can i do this using expdp
i want to move all objects from this my users tablespace to this new tablespace
Two different says in different documentation as below.
1) Import required same tablespace to fit the objects, otherwise it throws error.
2) Import by default uses the default tablespace to fit the objects, but some times it uses SYSTEM tablespace also.
confirm with the correct one.
I am using transportable tablespace to migrate from 10g to 11g from Windows to Linux System.
The steps are as follow On 10g make the tablespace read only mode and export the metadata information and copy the tablespace datafiles to the 11g server. Now on 11g when i am importing the exported metadata it says that the user does not exist and if i create the user and tablespace it does not work as it says tablespace already exist.
For transportable tablespace do i have to create the user already on 11g ? If yes then i also need to create the tablespace which i need to assign to the user.
Is this possible to run a direct impdp with database link for transport tablespace instead of first creating export dump and then using impdp.
I tried the below command and it shows up error.I have gone through the entire Oracle docs but could not find the exact command with network link for transport tablespace.
$impdp system/passwd DIRECTORY=data_pump_dir transport_tablespaces=TBS1,TBS2,TBS3 LOGFILE=tbs.log network_link=dblink
UDI-00015: invalid context or job state for parameter, 'transport_tablespaces'
When I am importing, I get these errors
IMP-00019: row rejected due to ORACLE error 1
IMP-00003: ORACLE error 1 encountered
ORA-00001: unique constraint (XXXXXXXXXXXXXXXX) violated
Column 1 2
Column 2 OFFLINE
[code]....
I added a datafile in undo tablespace (its an ASM database). I doubt that since I added the datafile to undo tablespace, I am getting this error.
We load large amount of data into multiple tables using sqlldr. Amount of data that we need to load varies according to the situation. We want to estimate the tablespace usage growth due to this data load, so we can verify/extend the tablespaces before the data load. Though, setting to autoextend will work in this case, We want to avoid extending the tablespace during sqlldr executing due to performance.
Our initial attempt was to note the tablespace size before and after executing the sqlldr and use the delta. But this delta was not consistent in different environments for the same amount of data. Different environments mean different oracle servers, different existing sizes of tablespaces, One data file Vs multiple data files etc.
How do we reliably estimate how much tablespace we need for the given amount of data?
I am doing an import job and the following error occurs during Index import. is the reason for this error?
View 1 Replies View Relatedwhile importing data i got this error in my log file.and i cannot import my data successfuly
in my log file error i found like this
ORA-01658: unable to create INITIAL extent for segment in tablespace USERS
IMP-00017: following statement failed with ORACLE error 1658:
IMP-00003: ORACLE error 1658 encountered
ORA-01658: unable to create INITIAL extent for segment in tablespace USERS
IMP-00017: following statement failed with ORACLE error 1658:
i can import my data using imp utility using this syntax
C:UsersAdministrator>imp tiger/****@tcs file=E:DUMP s.
dmp log=E:DUMP s.log fromuser=tiger121 touser=tiger statistics=none
this my user tiger default tablespace its uses and its a auto extend on and locally managed tablespace,and i have enough space on my drive also.
i have a tablespace which contains 121 datafile(max limit reached) as a dba what we have to do?
creating a new tablespace with a datafile and assign the users to the current tablespace which i created now.iif the above process is correct,after some time the tablespace which was filled up got freed up.now can i give the access to the users previous (i.e. freed up tablespace) and current tablespaces
My database version is
Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
PL/SQL Release 11.1.0.7.0 - Production
CORE 11.1.0.7.0 Production
TNS for Linux: Version 11.1.0.7.0 - Production
NLSRTL Version 11.1.0.7.0 - Production
My os version is
Linux damdat01 2.6.18-128.7.1.el5 #1 SMP Wed
Aug 19 04:00:49 EDT 2009 x86_64 x86_64 x86_64
GNU/Linux
My database is OLP system.
My question is what are the advantages and disadvantages having one single tablespace versus multiple tablespace?
Easy to maintain when you have single tablespace. but hard to track the IO issues if you have one single tablespace.
i have install 0racle 9i its working fine .I can access data from launch standalone but when i tried with login to oracle management server i can't login .
how can login with oracle management server .
I have one database which is recently upgraded from oracle 8.1.5 to oracle 10.2.0.4.The database is having around 300 tablespace and total size of the database is 1.5 TB.
The database was created in oracle 8i and all the teblespace were DMT(Dictionary Managed Tablespace) .Usually after up gradation all the tablespace are in DMT mode. Now my requirement is to convert all the tablespace into LMT (Locally Managed Tablespace) so that I can AVAIL ALL THE FEATURES OF LMT.
This database is a mission critical database and very less downtime can be allowed.
am using Oracle 10.2.0.4 on win 2008 server SP2. I would like to know if we can set the Segment Space Management feature to AUTO for RBS and Temporary tablespaces. As the data is not permanent in these tablespaces, will it manage automatically?
Presently its Segment Space Mgmt is manual for System, RBS, Temporary tablespaces.
I have read this article:
[URL].........
And i see:
Quote:
If MEMORY_TARGET is set to non zero value:
SGA_TARGET, SGA_MAX_SIZE and PGA_AGGREGATE_TARGET are set to 0, 60% of memory mentioned in MEMORY_TARGET is allocated to SGA and rest 40% is kept for PGA.
So, I have set:
alter system set sga_target=0 scope=spfile;
alter system set pga_aggregate_target=0 scope=spfile;
alter system set sga_max_size=0 scope=spfile;
alter system set memory_max_target=512M scope=spfile;
alter system set memory_target=300M scope=spfile;
, and then bounced the instance.
After startup, I see:
SQL> startup
ORACLE instance started.
Total System Global Area 272027648 bytes
Fixed Size 1384012 bytes
Variable Size 100663732 bytes
Database Buffers 163577856 bytes
Redo Buffers 6402048 bytes
Database mounted.
Database opened.
SQL>
But Total System Global Area should't be in that case 60% of memory mentioned in MEMORY_TARGET? Memory mentioned in MEMORY_TARGET was 300M, and 60% of 300M is 180M, which is not 272027648 bytes.
I just want to use automatic memory management, so I've set the other parameters above to 0. My instance is on my local machine, so just for my own personal use.
when i want to do backup in oracle management server for my database (repository)that i created give me an error message
<< using perferred credentials , the following error occurred: ORA-01031 :insufficent privileges>>
the syntax to make a tablespace's segment space mnagement auto which is segement management is *manual* and exent management is *local autoallocate*.
View 2 Replies View RelatedI have been reading various articles about the undo management. This basic concept of undo management is simple but how oracle implements it is bit harder for me to grasp.
What i have read and understood is that whenever a DML(Update, Delete, Insert) statement is issued by a user, the data is fetched from datafile to database buffer cache and at the same time a copy of the original data is saved in undo segment. Now if other users requests the same data, they are presented with the unchanged copy in the undo segment.
Now I have the following questions:
1) In case of Insert statement, what data is saved in undo segment. Is it the complete data in the table to which we want to insert the new row?
2)When the user issues DML statement, there are three copies of the same data, one in Memory (which is changed and not the same as original data), second in Undo segment (Which is unchanged copy of original data) and third in datafile file ( which is original data). What is the difference in the data in undo segment and data in datafile at this stage. Why are the other users presented with the data from undo segment rather than original data from the datafile to maintain read consistency.
3)When the user issues rollback, the changes made to the copy of data in memory are undone.The copies of data in memory and undo segment are now same?. What happens to the before change copy in undo segment. Is it still there or deleted.
the erelationship between sga_max_size,sga_targt,shared_pool_size,pga_aggregate_target and the server memory.
In short how shud i choose the above parameters for a server with a fixed RAM.
I have one generic question about space management. I have one table with size of 1TB. This table stored in ORC1 tablespace. This tablespace contains 70 datafiles.
Since it's 10.2.0.4 database. I have dropped this table by using purge
drop table <<table_name>> purge;
Once table drop was completed. When I check the tablespace space it was 100% free but due to HWM was unable to resize the datafile from current size to small size. What was the reason behind this. Is there any process needs to follow when dropping big tables ? like instead of dropping the tables do I need to truncate first & then drop .
I have one tablespace called U01. This tablepspace contains 31 data files. Due to high water mark I was unable to most datafiles. Since my database running onair application they will not provide me downtime to move the tables. Is there anyway to fix the high water mark without getting downtime window? almost 700+g space unused. I need to reuse them asap because running out of space with in asm diskgroup.
SQL> SELECT A.TABLESPACE_NAME,round(SUM(A.TOTS)/1024/1024) "Tot size MB",
2 round(SUM(A.SUMB)/1024/1024) "Tot Free MB",
3 round(SUM(A.SUMB)*100/SUM(A.TOTS)) "%FREE",
100-round(SUM(A.SUMB)*100/SUM(A.TOTS)) "%USED",
round(SUM(A.LARGEST)/1024/1024) MAX_FREE,SUM(A.CHUNKS) CHUNKS_FREE
4 5 6 FROM (
7 SELECT TABLESPACE_NAME,0 TOTS,SUM(BYTES) SUMB,
[code]...