Datapump Job
Hi,
I moved the datapump export file to different directory when still export job is running.I thought that it will throw an error but the expdp job is still running .
I am wonder how the expdp is running inspite of moved the file to different directory .
Is it feature of the datapump job ? Could you please explain me ?
Thanks in advance.
Regards,
Mac Li
http://download.oracle.com/docs/cd/E14072_01/server.112/e10701/dp_export.htm
Edited by: Suhail Faraaz on Aug 7, 2010 12:08 AM
Similar Messages
-
How can I change default OEM DataPump job?
Greetings,
I have successfully created a DataPump job to run out of OEM GC by choosing the target database I wish to export, then under the Data Movement tab selecting the Export to Export Files option. I then create the DataPump job and it runs successfully.
Now I would like to add the parameter REUSE_DUMPFILES=Y to the job but am not able to find an option in OEM allowing me to add this parameter. Is this possible and if so can someone explain to me how to do it.
Thanks.
Bill WagmanBill,
I don't see a way to do that. For sure, not in the UI while creating the "Export to Export Files". The only place you may be able to do this is add it to the command line for the job, as follows:
expdp scott/tiger directory=dir dumpfile=scott.dmp logfile=scott.log tables=emp REUSE_DUMPFILES=Y
Doesn the creation of the job save the job and the commands in the job library?
I would create a job in EM that does an OS command vs creating a job on the DB job queue and that way, you can ose the above.
let me know.
Thanks -
Recommended/Widely followed way to kill a Datapump job
DB version: 11.2
OS : AIX/Solaris
What is the recommeded way to kill a currently running expdp or impdp job (dba_datapump_jobs.state = 'EXECUTING') ?
When I searched in google and OTN , i see different opinions and options.Refer this
how to kill datapump jobs?
And
http://blog.oracle48.nl/killing-and-resuming-datapump-expdp-and-impdp-jobs/ -
Oracle dba datapump job low space
my server has 10gb free space ,but datapump job requires 100gb how can i export with export
With plain old unix compression I'm getting about 80% compression on my exports.
Another idea is to use the FILESIZE parameter to limit files to something less than available space, then move the file while expdp stalls with the out of space error. I've never tried that, so I don't know if it works. It's usually easier for me to move other stuff around to make space first, or use original exp into a pipe. -
Starting the datapump jobs from client side
sir
please make me understand
datapump is server side utility means we can not invoke it from client. but in documents (oracle).
they r talking that we can start the datapump job from client(Sqlplus) but the dmp files will be saved and the process will run on the server side
so how can invoke the datapump from the client side (SQL Plus) ?user13014926 wrote:
sir
please make me understand
datapump is server side utility means we can not invoke it from client.That's wrong to say and understand. The correct understanding should be that it's configured only for server side which means, that the folder where the dump files would be created would be available on a folder which must be there on the server side and over which, proper permissions must be available to you in the form of Directory object. There is no such thing that it can't be invoked sitting on the client side. It surely can but unlike the traditional export, there is no facility available anymore to give absolute file paths which would store the file on the local client machine.
but in documents (oracle).
they r talking that we can start the datapump job from client(Sqlplus) but the dmp files will be saved and the process will run on the server sideAll what they are saying is that since the datapump's API, DBMS_DATAPUMP , is completely exposed and is documented, it can be used by the client and as like any other package, either stored or created, can be invoked from the sqlplus session. The dump files location would indeed be on the server side only.
so how can invoke the datapump from the client side (SQL Plus) ?As mentioned already by Srini, using the API .
Aman.... -
How to kill datapump jobs?
hi all,
i tried to export data using expdp. i read somewhere that using expdp creates a background process so that even if i turn off my client machine, it won't stop the export.
problem is i need to stop the expdp process. how do i kill it?
when i query the database using
select *
from DBA_DATAPUMP_JOBS;
the jobs are still there with a value of NOT RUNNING in the STATE column. i think the jobs are already done in this case but how do i remove them from the list? also how do i kill a currently running datapump job?
any suggestions?
thankshi,
thanks for the link. although, i was not able to find a way on how to delete the records showing in my dba_datapump_jobs view.
unfortunately i'm not in the expdp> prompt anymore since i turned my machine off. is there a way to get back there?
here's the record in the dba_datapump_jobs
OWNER_NAME JOB_NAME OPERATION JOB_MODE STATE DEGREE ATTACHED DATAPUMP
SYSTEM SYS_IMPORT_FULL_01 IMPORT FULL NOT RUNNING 0 0 0
TIME_OWNER_QA IMPTIME IMPORT FULL NOT RUNNING 0 0 0
thanks -
Check the status expdp datapump job
All,
I have started expdp (datapump) on size of 250GB schema.I want to know when the job will be completed.
I tried to find from views dba_datapump_sessions and v$session_longops,v$session.But but all in vain.
Is it possible to find completion time of expdp job?
Your help is really appreciated.Hi,
Have you started the Job in "Interactie mode".. ??
If yes then you can get the status of Exectuion with "STATUS"- defauly is zero (If you setting to not null value then based on that it will show the status - seconds)
Second, thing check "dba_datapump_jobs"
- Pavan Kumar N -
Starting failed imp datapump job again....
Dear Team,
I am importing about 220 gb data through datapump, it is failing while importing statistics so I am going to restart the job with exclude=statistics...
Already 212 gb data is imported. so i want to ignore the imported objects.. Datapump will skip the imported objects or any command we have to add in syntax...
Pls suggest...Sure.
The RTFM command.
More specially, the impdp help=y command will show you the attach_job clause of the impdp command.
Kindly do use available resources, so as not to clutter up this forum with doc questions further.
Sybrand Bakker
Senior Oracle DBA -
Hi all, I have a problem. In next SQL query SELECT owner_name, job_name, operation, job_mode, state, attached_sessions FROM dba_datapump_jobs;
OWNER_NAME JOB_NAME OPERATION JOB_MODE STATE ATTACHED_SESSIONS
SYS exp_sys_stats1 EXPORT TABLE NOT RUNNING 0
SYS exp_sys_stats EXPORT TABLE NOT RUNNING 0
2 rows selected. Then I try
DROP TABLE SYS.exp_sys_stats1;
DROP TABLE SYS.exp_sys_stats1
Error at line 1
ORA-00942: table or view does not exist
SELECT o.status, o.object_id, o.object_type,
o.owner||'.'||object_name "OWNER.OBJECT"
FROM dba_objects o, dba_datapump_jobs j
WHERE o.owner=j.owner_name AND o.object_name=j.job_name
AND j.job_name NOT LIKE 'BIN$%' ORDER BY 4,2;
STATUS OBJECT_ID OBJECT_TYPE OWNER.OBJECT
VALID 113753 TABLE SYS.exp_sys_stats
VALID 113788 TABLE SYS.exp_sys_stats1
2 rows selected.
DECLARE
job1 NUMBER;
BEGIN
job1 := DBMS_DATAPUMP.ATTACH('exp_sys_stats1','sys');
DBMS_DATAPUMP.STOP_JOB (job1);
END;
also say ORA-31626: job does not exist I trying attach job, but got a error job does not exist. What can I do to delete this jobs?Yes I create this jobs with DBMS_SCHEDULER, and then delete, but records of them remained in view.
SELECT * FROM sys.obj$ where name like ('exp_sys_stats%');
OBJ# OWNER# NAME CTIME
113753 0 exp_sys_stats 11/17/2011
113788 0 exp_sys_stats1 11/17/2011
2 rows selected.*2oradba* I cannot atach job, I got error job does not exist :)
Edited by: 807831 on 17.11.2011 6:01 -
When I run the datapump export (Oracle 10g) from my Solaris box, the dumpfile and logfile created in the directory are owned by OS user oracle and group dba with a permission mask of -rw-r--r--.
Is there a way I can provide a different user/group/permission for the files generated by export?Pl see if this related thread can help
Re: file permissions on data pump exports
HTH
Srini -
IMPDP: The job SYS_IMPORT_FULL_01 has halted due to a fatal error
Hi,
I'm having problems importing a database whose export was successful.
The tool I'm using is expdp and impdp.
The source schema is in a tablespace different from the target schema, but in the same instance of the Oracle database.
I granted the role of EXP_FULL_DATABASE for the source user and granted the role IMP_FULL_DATABASE to the target user.
Specifically, the following happens: I can export the schema and can successfully import the first four tables to the target schema. So the impdp provides the following message:
Processing object type SCHEMA_EXPORT / TABLE / TABLE_DATA
. . imported "BDE_INEA_DES". "GPL_DECLIV_FRAGIL" 328.4 MB 668 706 lines
. . imported "BDE_INEA_DES". "GLN_CURVA_NIVEL" 336.3 MB 124 324 lines
. . imported "BDE_INEA_DES". "GPL_APP_10" 2920 lines 238.7 MB
. . imported "BDE_INEA_DES". "GLN_CURVA_NIVEL_10" 200.8 MB 15 344 lines
The job "BDE_INEA_DES". "SYS_IMPORT_SCHEMA_01" was halted due to a fatal error at 11:52:41
I've tried exporting using SYSTEM user and the user's schema source owner. I tried the same procedures with the import, but without success.
Information of my OS:
Windows Server 2008 R2 x64
Information of Oracle database:
SQL> SELECT * FROM V $ VERSION;
BANNER
Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
PL / SQL Release 11.2.0.1.0 - Production
CORE 11.2.0.1.0 Production
TNS for 64-bit Windows: Version 11.2.0.1.0 - Production
NLSRTL Version 11.2.0.1.0 - ProductionFollow these steps. You will get exactly what you are trying to find.
1) spool invalidobj_and_registry.txt
SELECT SUBSTR(comp_id,1,15) comp_id, status, SUBSTR(version,1,10)
version, SUBSTR(comp_name,1,30) comp_name
FROM dba_registry
ORDER BY 1;
SELECT status, object_id, object_type, owner||'.'||object_name "OWNER.OBJECT"
FROM dba_objects
WHERE status != 'VALID'
ORDER BY 4,2;
spool.off
2) $ sqlplus "/as sysdba"
sql>EXEC DBMS_STATS.GATHER_DICTIONARY_STATS;
sql>EXEC DBMS_STATS.GATHER_FIXED_OBJECTS_STATS;
3) Run your import with these two extra parameters along with your parameters
CONTENT=METADATA_ONLY METRICS=Y TRACE=480300
4) As soon as impdp job started:
-- In SQL*Plus, obtain Data Pump processes info:
CONNECT / as sysdba
select to_char(sysdate,'YYYY-MM-DD HH24:MI:SS') "DATE", s.program, s.sid, s.status, s.username, d.job_name, p.spid, s.serial#, p.pid
from v$session s, v$process p, dba_datapump_sessions d
where p.addr=s.paddr and s.saddr=d.saddr;
-- Get the sid and serial# for DM00 and DW01 and execute:
exec DBMS_SYSTEM.SET_EV([SID],[SERIAL#],10046,12 ,'');
for ***both*** of them.
After hang is noticed please leave the import running for one more hour and kill it.
Please check:
- alert log
- impdp log file
- trace files generated during import time in bdump directory
++ before restarting the impdp check if there are orphan Data Pump jobs left in database. Use Note 336014.1 - "How To Cleanup Orphaned DataPump Jobs In DBA_DATAPUMP_JOBS ? -
Hi all,
I'm having problems with my 10.2.0.3 Enterprise Oracle version to kill some old jobs. These were datapump jobs.
I've all these jobs when I query the dba_datapump_jobs:
OWNER_NAME JOB_NAME OPERATION JOB_MODE STATE ATTACHED_SESSIONS
SYSTEM EXPORT_FULL_20111118 EXPORT FULL DEFINING 1
SYSTEM SYS_EXPORT_SCHEMA_01 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_02 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_03 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_04 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_05 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_06 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_TABLE_01 EXPORT TABLE DEFINING 0
If I try to drop it, I've this error:
SYSTEM@SPA3> exec dbms_scheduler.drop_job('EXPORT_FULL_20111118');
BEGIN dbms_scheduler.drop_job('EXPORT_FULL_20111118'); END;
ERROR at line 1:
ORA-27475: "SYSTEM.EXPORT_FULL_20111118" must be a job
ORA-06512: at "SYS.DBMS_ISCHED", line 178
ORA-06512: at "SYS.DBMS_SCHEDULER", line 544
ORA-06512: at line 1
I've tried doing expdp commands but was impossible:
[oracle@serverpro ~]$ expdp system/xxxx attach=EXPORT_FULL_20111118
Export: Release 10.2.0.3.0 - 64bit Production on Viernes, 18 Noviembre, 2011 11:59:13
Copyright (c) 2003, 2005, Oracle. All rights reserved.
Connected to: Oracle Database 10g Release 10.2.0.3.0 - 64bit Production
With the Real Application Clusters option
ORA-31626: job does not exist
ORA-06512: at "SYS.DBMS_SYS_ERROR", line 79
ORA-06512: at "SYS.KUPV$FT", line 430
ORA-31638: cannot attach to job EXPORT_FULL_20111118 for user SYSTEM
ORA-31632: master table "SYSTEM.EXPORT_FULL_20111118" not found, invalid, or inaccessible
ORA-00942: table or view does not exist
Any ideas about how to kill them?
Best regards,
dbajugHi Fran,
Yes I tried it.
These jobs doesn't appears on mgmt_job table:
SYS@SPA3> select job_id, job_name, job_owner from sysman.mgmt_job where job_owner = 'SYSTEM';
no rows selected
But yes if I make a query over dba_datapump_jobs:
SELECT owner_name, job_name, operation, job_mode, state, attached_sessions
FROM dba_datapump_jobs WHERE job_name NOT LIKE 'BIN$%' ORDER BY 1,2;
OWNER_NAME JOB_NAME OPERATION JOB_MODE STATE ATTACHED_SESSIONS
SYSTEM EXPORT_FULL_20111118 EXPORT FULL DEFINING 1
SYSTEM SYS_EXPORT_SCHEMA_01 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_02 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_03 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_04 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_05 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_SCHEMA_06 EXPORT SCHEMA DEFINING 0
SYSTEM SYS_EXPORT_TABLE_01 EXPORT TABLE DEFINING 0
SYSTEM SYS_EXPORT_TABLE_02 EXPORT TABLE DEFINING 1
I found many DMxx processes (Coordinates the Data Pump job tasks performed by Data Pump worker processes and handles client interactions) but I can't kill the sessions or OS processes.
How could I kill all these datapump jobs?
Regards,
dbajug -
Error while importing schemas using datapump
Hi,
I am trying to import schema from qc to development. after importing i got the following error attached below:
Processing object type SCHEMA_EXPORT/TABLE/GRANT/WITH_GRANT_OPTION/OBJECT_GRANT
Processing object type SCHEMA_EXPORT/TABLE/GRANT/CROSS_SCHEMA/OBJECT_GRANT
Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
ORA-39065: unexpected master process exception in RECEIVE
ORA-39078: unable to dequeue message for agent MCP from queue "KUPC$C_2_20090421161917"
Job "SYS"."uat.210409" stopped due to fatal error at 20:15:13
ORA-39014: One or more workers have prematurely exited.
ORA-39029: worker 2 with process name "DW02" prematurely terminated
ORA-31671: Worker process DW02 had an unhandled exception.
ORA-39078: unable to dequeue message for agent KUPC$A_2_20090421161934 from queue "KUPC$C_2_20090421161917"
ORA-06512: at "SYS.KUPW$WORKER", line 1397
ORA-06512: at line 2
ORA-39029: worker 3 with process name "DW03" prematurely terminated
ORA-31671: Worker process DW03 had an unhandled exception.
ORA-39078: unable to dequeue message for agent KUPC$A_2_20090421162030 from queue "KUPC$C_2_20090421161917"
ORA-06512: at "SYS.KUPW$WORKER", line 1397
ORA-06512: at line 2
ORA-39029: worker 4 with process name "DW04" prematurely terminated
ORA-31671: Worker process DW04 had an unhandled exception.
ORA-39078: unable to dequeue message for agent KUPC$A_2_20090421162031 from queue "KUPC$C_2_20090421161917"
ORA-06512: at "SYS.KUPW$WORKER", line 1397
ORA-06512: at line 2
Is my import completed successfully or not??. please help...When a datapump job runs, it creates a table called the master table. It has the same name as the job name. This is used to keep track of where all of the information in the dumpfile is located. It is also used when restarting a job. For some reason, this table got dropped. I'm not sure why, but in most cases, datapump jobs are restartable. I don't know why the original message was reported, but I was hoping the job woudl be restartable. You could always just rerun the job. Since the job that failed already created tables and indexes, if you restart the job, all of the objects that are dependent on those objects will not be created by default.
Let's say you have table tab1 with an index ind1 and both table and index are anaylized. Since tab1 is already created, the datapump job will mark all of the objects dependent on tab1 to skip. This includes the index, table_statistics, and index_statistics. To get around this, you could say
table_exists_action=replace
but this will replace all tables that are in the dumpfile. Your other options are:
table_exists_action=
truncate -- to truncate the data in the table and then just reload the data, but not the dependent objects
append -- to append the data from the dumpfile to the existing table, but do not import the dependent objects
skip -- skip the data and dependent objects from the dumpfile.
Hope this helps.
Dean -
Datapump through grid control 12C
hi,
I have created a schema in database which has exp_full_database privilege to run datapump jobs (verified by running a datapump API job too). I have a os user(non dba group) which has ability to run expdp command.
I also created an administrator in grid 12C which has connect to target privilege on database target and execute any command privilege on host. However when I submit a datapump job using these credentails, the submission of job fails with error message as "user doesn't have privilege for this operation".
I am able to execute datapump API successfully from within database schema; using the on dba group os user, I am able to execute expdp command too successfully..
What am i missing? please help.It sounds like your Data Pump command is referencing the EM user instead of the schema owner. Look at the output log of the job; if you see "ORA-01031: insufficient privileges" then this is likely the case. The EM admin only needs sufficient privileges to exec the job; Data Pump is responsible for the database access and privileges.
Edited by: pschwart on Mar 14, 2013 5:59 PM -
Datapump import error on 2 partioned tables
I am trying to run impdp to import two tables that are partioned and use the LOB types...for some reason it always errors out. Anyone seen this issue in 11g?
Here is the info:
$ impdp parfile=elm_rt.par
Master table "ELM"."SYS_IMPORT_TABLE_05" successfully loaded/unloaded
Starting "ELM"."SYS_IMPORT_TABLE_05": elm/******** parfile=elm_rt.par
Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE
Processing object type DATABASE_EXPORT/SCHEMA/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/INDEX
Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
Processing object type DATABASE_EXPORT/SCHEMA/TABLE/AUDIT_OBJ
Processing object type DATABASE_EXPORT/SCHEMA/TABLE/STATISTICS/TABLE_STATISTICS
ORA-39014: One or more workers have prematurely exited.
ORA-39029: worker 1 with process name "DW01" prematurely terminated
ORA-31671: Worker process DW01 had an unhandled exception.
ORA-04030: out of process memory when trying to allocate 120048 bytes (session heap,kuxLpxAlloc)
ORA-06512: at "SYS.KUPW$WORKER", line 1602
ORA-06512: at line 2
ORA-39014: One or more workers have prematurely exited.
ORA-39029: worker 2 with process name "DW01" prematurely terminated
ORA-31671: Worker process DW01 had an unhandled exception.
ORA-04030: out of process memory when trying to allocate 120048 bytes (session heap,kuxLpxAlloc)
ORA-06512: at "SYS.KUPW$WORKER", line 1602
ORA-06512: at line 2
Job "ELM"."SYS_IMPORT_TABLE_05" stopped due to fatal error at 13:11:04
elm_rt.par_
$ vi elm_rt.par
"elm_rt.par" 25 lines, 1340 characters
DIRECTORY=DP_REGRESSION_DATA_01
DUMPFILE=ELM_MD1.dmp,ELM_MD2.dmp,ELM_MD3.dmp,ELM_MD4.dmp
LOGFILE=DP_REGRESSION_LOG_01:ELM_RT.log
DATA_OPTIONS=SKIP_CONSTRAINT_ERRORS
CONTENT=METADATA_ONLY
TABLES=RT_AUDIT_IN_HIST,RT_AUDIT_OUT_HIST
REMAP_TABLESPACE=RT_AUDIT_IN_HIST_DAT01:RB_AUDIT_IN_HIST_DAT01
REMAP_TABLESPACE=RT_AUDIT_IN_HIST_IDX04:RB_AUDIT_IN_HIST_IDX01
REMAP_TABLESPACE=RT_AUDIT_OUT_HIST_DAT01:RB_AUDIT_OUT_HIST_DAT01
PARALLEL=4Read this metalink note 286496.1. (Export/Import DataPump Parameter TRACE - How to Diagnose Oracle Data Pump)
This will help you generate trace for the datapump job.
Maybe you are looking for
-
Adding systems with same SID in SMSY - solution manager
Dear All, We are having systems with same SID's which belongs to different landscape created for various customers. Currently We are having only one solution manager. I consider this example for explanation. Customer 1 : DEV --> QAS --> PRD ( SID's)
-
Would someone give me a link or model number for the "micro USB cable" that specifically goes from the Apple TV 2/3 gen to a MacBook Pro 13 2012? Thanks!
-
HTML link in column ... Need Help!!!!!
I need to use a column in my viewobject to create an html link to an external resource on another server, a simple webpage. I've used the RowImpl for my view object to build the HREF but when the rows are rendered the link is fully escaped and isn't
-
27" iMac as Mac Pro Display?
I currently use the 24" LED as a display for my Mac Pro, but give the increased pixels on the 27" iMac and it's MiniDisplay Port in capability, and given it's cheaper than the 30" cinema display, with a much better panel, on only slightly lower verti
-
Airplay works differently from ATV vs Airport
So let me try to explain this properly. When I use an AppleTV to Airplay, and someone else wants to airplay to the same appletv, they automatically kick me and their stream starts to run. When I do the same with an Airport device, the new user either