Datapump in RAC

Hi,
I have a 2 nodes RAC 11.2g installation on Solaris 10.
I don't use Cluster Solaris.
I would like to plan to run, every night, a full datapump for my database.
I want an idea about this, because I don't want to put in crontab the datapump
only on one node and I would like that in case of a temporary down of a node the datapump
should run on the other.
I'm open to all suggestions...
I can use ASM to store datapump file (not log files i know this),
I can create a resource in crs... so please tell me something
Thanks in advance
c.

I already use rman.
We need datapump for different reasons...
for example we don't want to use flashback and other (imho) amazing solutions for
restore single tables or singles schema and because often we have to migrate
schema or table from production to test environment.
Thanks

Similar Messages

  • Encrypt datapump in oracle 10g

    Hi, I have an Oracle RAC SE 10gR2 (10.2.03) with 3 nodes on SLES 10 on ibm ppc.
    Every night I have some oracle exports (expdp) of some schemas. I want to encrypt one of various datadumps throught a shell bash script. I see that exists TDE for encrypt colums in a table, and I have to create a wallet for encrypt. However, I'm thinking that I preferred clear text (no encrypt in my tables and schemas) and only encrypt the datapumps files. For example, I have my database same as now, and I should specify one options in my expdp for encrypt.
    It's possible?
    Can I encrypt those datapumps without use TDE (in Oracle 10g)?
    Thanks you very much!!

    Rober wrote:
    Hi, I have an Oracle RAC SE 10gR2 (10.2.03) with 3 nodes on SLES 10 on ibm ppc.
    Every night I have some oracle exports (expdp) of some schemas. I want to encrypt one of various datadumps throught a shell bash script. I see that exists TDE for encrypt colums in a table, and I have to create a wallet for encrypt. However, I'm thinking that I preferred clear text (no encrypt in my tables and schemas) and only encrypt the datapumps files. For example, I have my database same as now, and I should specify one options in my expdp for encrypt.
    It's possible?
    Can I encrypt those datapumps without use TDE (in Oracle 10g)?
    Thanks you very much!!TDE perform encryption at storage level.
    data pump has encryption clause in 11g
    http://www.oracle-base.com/articles/11g/data-pump-enhancements-11gr1.php#encryption_parameters

  • How unhealthy is this RAC?

    Here's is the contents of v$system_event..
    Is this
    EVENT     TOTAL_WAITS     TIME_WAITED     AVERAGE_WAIT
    enq: TX - index contention     40564851     214701526      5.29
    enq: TX - row lock contention     188846     12454614     65.95
    enq: SQ - contention     141971     70568     0.5
    cause for concern?
    EVENT     TOTAL_WAITS     TIME_WAITED     AVERAGE_WAIT
    SQL*Net message to client     6015051449     607254     0
    SQL*Net message from client     6015048542     178177969892     29.62
    gcs remote message     2948555287     2633481757     0.89
    CGS wait for IPC msg     1517805027     634397     0
    db file sequential read     1500615188     816364485     0.54
    ges remote message     1247679701     1407300224     1.13
    gc cr multi block request     778432813     9913464     0.01
    gc current block 2-way     747852637     38030616     0.05
    db file scattered read     709428365     460939295     0.65
    rdbms ipc message     708473316     37650068633     53.14
    gc buffer busy acquire     671285134     1033621285     1.54
    PX Deq: reap credit     667784615     484449     0
    gcs log flush sync     592376026     171712257     0.29
    gc cr block 2-way     530861847     19607062     0.04
    library cache pin     437937120     15126237     0.03
    log file sync     379523272     797193932     2.1
    DIAG idle wait     359607166     2822108755     7.85
    log file parallel write     351225436     259263769     0.74
    LNS ASYNC end of log     350170653     1398410516     3.99
    LNS wait on SENDREQ     321652621     3209301     0.01
    PX qref latch     297396661     94308     0
    read by other session     289140108     148440270     0.51
    buffer deadlock     163505781     983055     0.01
    gc current block busy     119223825     467716658     3.92
    PX Deq: Table Q Normal     117332841     23574867     0.2
    ksxr poll remote instances     110480324     90333     0
    buffer busy waits     106938153     19933900     0.19
    direct path read     93429599     108427028     1.16
    SQL*Net more data from client     86471785     23026529     0.27
    gc current grant busy     84978157     28215346     0.33
    control file sequential read     82646297     23694583     0.29
    PX Deq Credit: send blkd     78641669     9569299     0.12
    latch: cache buffers chains     74218671     690277     0.01
    gc current grant 2-way     72557796     1920419     0.03
    library cache: mutex X     71106697     75993     0
    DFS lock handle     70722498     2716407     0.04
    gc cr grant 2-way     64558237     1633004     0.03
    PX Deq: Execution Msg     61706261     314222076     5.09
    gc cr block busy     61469863     119850802     1.95
    library cache lock     52428649     3773354     0.07
    PX Deq: Slave Session Stats     48040224     1886805     0.04
    db file parallel read     46415188     118467902     2.55
    IPC send completion sync     46250594     965101     0.02
    enq: TX - index contention     40564851     214701526     5.29
    PX Deq: Execute Reply     39689685     17243721     0.43
    gc buffer busy release     36976909     242714774     6.56
    SQL*Net more data to client     36627952     44167     0
    PX Deq: Msg Fragment     30501244     343397     0.01
    rdbms ipc reply     29725302     1352370     0.05
    RMAN backup & recovery I/O     28824547     37722662     1.31
    reliable message     27892263     3082134     0.11
    PX Idle Wait     27356097     4651277341     170.03
    ASM file metadata operation     25098749     8850323     0.35
    gc object scan     22705857     7485     0
    db file parallel write     19896252     52152606     2.62
    latch: ges resource hash list     19336183     427451     0.02
    enq: PS - contention     19143961     707455     0.04
    PX Deq: Parse Reply     19093356     895799     0.05
    gc cr disk read     17816846     448909     0.03
    ASM background timer     16101806     1383957874     85.95
    PX Deq: Slave Join Frag     16044789     233149     0.01
    wait for unread message on broadcast channel     15056320     1413552546     93.88
    cursor: mutex X     13435193     24140     0
    KJC: Wait for msg sends to complete     13268497     11397     0
    PX Deq: Signal ACK RSG     13214824     101941     0.01
    KSV master wait     13206286     4235645     0.32
    direct path read temp     12617694     5487608     0.43
    PX Deq Credit: need buffer     11675868     879967     0.08
    row cache lock     11536185     398216     0.03
    PX Deq Credit: Session Stats     9480862     78910     0.01
    SQL*Net message to dblink     9312894     1538     0
    SQL*Net message from dblink     9312894     6279631     0.67
    control file parallel write     7760982     11854435     1.53
    pmon timer     7558889     1412576090     186.88
    PX Deq: Join ACK     7548816     498931     0.07
    gc current multi block request     6035173     155898     0.03
    PING     5706961     1413230267     247.63
    enq: XR - database force logging     4662671     198813     0.04
    class slave wait     4561877     7097429006     1555.81
    Streams AQ: waiting for messages in the queue     4495828     1543411682     343.3
    SQL*Net more data from dblink     3696582     444575     0.12
    LGWR wait for redo copy     3655353     17840     0
    log file sequential read     3387305     6610414     1.95
    Log archive I/O     2990486     276772     0.09
    SQL*Net break/reset to client     2971976     2385935     0.8
    direct path write temp     2839390     2522114     0.89
    Space Manager: slave idle wait     2827526     1412987186     499.73
    latch: shared pool     2808517     298150     0.11
    latch: gc element     2421717     24688     0.01
    SGA: MMAN sleep for component shrink     2336447     2458094     1.05
    latch: enqueue hash chains     2279645     15435     0.01
    latch free     2089418     78732     0.04
    gc current split     2044784     1864009     0.91
    PX Deq: Signal ACK EXT     1976164     19263     0.01
    enq: FB - contention     1473469     61036     0.04
    cursor: pin S wait on X     1313129     1464789     1.12
    SQL*Net more data to dblink     1232891     986     0
    Streams AQ: RAC qmn coordinator idle wait     1211300     788     0
    enq: HW - contention     1175390     2077008     1.77
    latch: session allocation     1167768     21883     0.02
    Streams AQ: qmn coordinator idle wait     1144699     1412546634     1233.99
    Streams AQ: qmn slave idle wait     1031585     2227183681     2158.99
    lock deadlock retry     962937     4698     0
    enq: CF - contention     956154     609647     0.64
    latch: cache buffers lru chain     902764     37552     0.04
    latch: object queue header operation     817911     27717     0.03
    global enqueue expand wait     768633     654105     0.85
    Data file init write     756191     329758     0.44
    latch: gcs resource hash     647021     4147     0.01
    local write wait     603007     286191     0.47
    latch: row cache objects     599358     6453     0.01
    ges lmd/lmses to freeze in rcfg - mrcvr     481759     156345     0.32
    shared server idle wait     471190     1413238589     2999.3
    enq: RF - DG Broker Current File ID     469833     23209     0.05
    smon timer     432383     1411851085     3265.28
    SGA: allocation forcing component growth     363333     379008     1.04
    gc current retry     341104     1121252     3.29
    enq: RF - synch: DG Broker metadata     319143     588290     1.84
    enq: PG - contention     313659     14830     0.05
    enq: TT - contention     260134     11207172     43.08
    enq: KO - fast object checkpoint     236745     820808     3.47
    dispatcher timer     236637     1413242481     5972.2
    direct path write     231382     191008     0.83
    cursor: pin S     229011     394     0
    Streams AQ: waiting for time management or cleanup tasks     199981     1413148548     7066.41
    enq: TX - row lock contention     188846     12454614     65.95
    enq: TX - allocate ITL entry     153703     54252     0.35
    enq: SQ - contention     141971     70568     0.5
    ksdxexeother     141885     56     0
    latch: redo allocation     138912     1858     0.01
    recovery area: computing applied logs     126415     45925     0.36
    gc current block congested     126318     21768     0.17
    resmgr:cpu quantum     123074     151384     1.23
    jobq slave wait     120678     35574221     294.79
    Datapump dump file I/O     90431     9127     0.1
    ges inquiry response     89402     4041     0.05
    os thread startup     83809     222586     2.66
    cr request retry     80062     71896     0.9
    PX Deq: Table Q Sample     79665     133402     1.67
    gc cr block congested     79026     14792     0.19
    gc cr failure     77521     25019     0.32
    enq: WF - contention     73983     825388     11.16
    enq: TQ - TM contention     72871     3319     0.05
    lock escalate retry     65714     1574     0.02
    buffer exterminate     59775     64919     1.09
    fbar timer     47136     1413183353     29980.98
    log file switch completion     46911     452097     9.64
    recovery area: computing obsolete files     45699     8547     0.19
    enq: US - contention     40401     8805     0.22
    enq: TM - contention     39149     5435032     138.83
    library cache load lock     36311     382575     10.54
    kjbdrmcvtq lmon drm quiesce: ping completion     31668     47443     1.5
    enq: TD - KTF dump entries     31468     1424     0.05
    enq: RO - fast object reuse     28422     31772     1.12
    parallel recovery slave wait for change     27558     3163     0.11
    name-service call wait     23694     181533     7.66
    control file single write     22375     1624     0.07
    kksfbc child completion     21239     106926     5.03
    PX Deq: Table Q qref     19325     245     0.01
    enq: TX - contention     18805     113253     6.02
    latch: messages     17203     181     0.01
    enq: RS - prevent file delete     16913     1013     0.06
    enq: RS - prevent aging list update     15682     642     0.04
    PX Deq: Table Q Get Keys     14322     42935     3
    gc current grant congested     14292     2192     0.15
    cursor: mutex S     13285     8     0
    log file single write     13164     5371     0.41
    latch: undo global data     12649     178     0.01
    kksfbc research     11894     12680     1.07
    parallel recovery slave idle wait     11193     5872     0.52
    wait list latch free     11026     11794     1.07
    enq: CT - state     11001     417     0.04
    latch: checkpoint queue latch     10526     132     0.01
    enq: PE - contention     10506     1139     0.11
    ARCH wait on SENDREQ     9957     216480     21.74
    gc cr grant congested     9465     1584     0.17
    wait for scn ack     9377     3155     0.34
    enq: TA - contention     8856     324     0.04
    log buffer space     8777     89323     10.18
    enq: TK - Auto Task Serialization     8542     343     0.04
    enq: DR - contention     7842     323     0.04
    process diagnostic dump     7707     2072     0.27
    JOX Jit Process Sleep     7612     11286431     1482.72
    enq: TC - contention     7357     340817     46.33
    ges global resource directory to be frozen     7140     12299     1.72
    enq: CO - master slave det     6850     312     0.05
    enq: JS - job run lock - synchronize     6704     397     0.06
    gcs drm freeze in enter server mode     6542     40742     6.23
    enq: TS - contention     5959     89332     14.99
    ARCH wait for archivelog lock     5600     36     0.01
    PX Nsq: PQ load info query     5377     104798     19.49
    db file single write     5373     3452     0.64
    gc remaster     5315     50625     9.52
    latch: parallel query alloc buffer     4939     1906     0.39
    enq: TO - contention     4799     143     0.03
    enq: AF - task serialization     4395     161     0.04
    enq: PI - contention     4251     163     0.04
    ges2 LMON to wake up LMD - mrcvr     4210     28     0.01
    enq: DL - contention     3889     239     0.06
    kjctssqmg: quick message send wait     3408     22     0.01
    LNS wait on DETACH     3275     741     0.23
    ksfd: async disk IO     3274     1     0
    LNS wait on ATTACH     3273     51940     15.87
    ARCH wait on DETACH     3231     714     0.22
    ARCH wait on ATTACH     3226     43238     13.4
    enq: BR - file shrink     2787     116     0.04
    write complete waits     2631     1070     0.41
    enq: MD - contention     2596     67     0.03
    enq: WL - contention     2198     266518     121.25
    single-task message     2098     25896     12.34
    enq: OD - Serializing DDLs     2054     66     0.03
    resmgr:internal state change     2001     14735     7.36
    ARCH wait on c/f tx acquire 2     1751     175230     100.07
    enq: WR - contention     1636     69     0.04
    latch: cache buffer handles     1610     29     0.02
    statement suspended, wait error to be cleared     1497     22626     15.11
    Streams AQ: qmn coordinator waiting for slave to start     1214     678966     559.28
    enq: PD - contention     1182     33     0.03
    JS kgl get object wait     1096     4922     4.49
    undo segment extension     1070     10065     9.41
    PL/SQL lock timer     949     8739819     9209.5
    enq: AE - lock     937     28     0.03
    LGWR-LNS wait on channel     832     913     1.1
    ges DFS hang analysis phase 2 acks     816     495     0.61
    latch: redo writing     729     9     0.01
    gc quiesce     665     564     0.85
    enq: JS - queue lock     482     2111     4.38
    PX Deq: Test for credit     442     13     0.03
    enq: SS - contention     386     274     0.71
    recovery area: computing dropped files     328     1400     4.27
    recovery area: computing backed up files     328     496     1.51
    ksdxexeotherwait     279     10592     37.97
    log switch/archive     250     137570     550.28
    gc domain validation     223     39964     179.21
    auto-sqltune: wait graph update     195     96514     494.95
    wait for a undo record     170     1214     7.14
    parallel recovery coord send blocked     168     4     0.02
    enq: JS - wdw op     168     3741     22.27
    enq: KT - contention     165     5     0.03
    switch logfile command     156     6290     40.32
    gcs resource directory to be unfrozen     149     12839     86.17
    Data Guard Broker Wait     139     10906     78.46
    enq: SK - contention     129     4     0.03
    enq: JS - job recov lock     128     4     0.03
    gc cr block lost     125     6772     54.17
    virtual circuit wait     122     3     0.03
    ges LMON to get to FTDONE      100     187     1.87
    enq: CU - contention     80     242     3.02
    enq: JQ - contention     78     7     0.09
    cursor: pin X     73     83     1.14
    parallel recovery coord wait for reply     70     510     7.29
    PX Deq: Txn Recovery Start     67     3436     51.29
    SQL*Net break/reset to dblink     60     0     0
    gc current block lost     57     2869     50.33
    ges LMD suspend for testing event     51     709     13.89
    inactive session     46     4550     98.91
    recovery read     45     5     0.11
    JS kill job wait     41     3548     86.53
    enq: AS - service activation     40     1     0.03
    enq: TL - contention     35     2     0.05
    enq: UL - contention     34     524     15.42
    gcs enter server mode     33     1559     47.24
    wait for stopper event to be increased     30     218     7.27
    enq: TQ - DDL contention     24     300     12.52
    enq: MR - contention     21     1     0.03
    ges reconfiguration to start     20     54     2.72
    ges enter server mode     20     502     25.08
    buffer latch     18     1337     74.26
    enq: SR - contention     18     1     0.05
    Streams: RAC waiting for inter instance ack     18     3748     208.21
    enq: PR - contention     17     46     2.72
    kupp process wait     16     166     10.39
    checkpoint completed     15     3678     245.19
    PX Deque wait     14     68     4.87
    enq: BF - allocation contention     14     1     0.08
    enq: XL - fault extent map     14     51     3.66
    enq: FU - contention     14     17     1.18
    enq: TH - metric threshold evaluation     13     114     8.78
    enq: MW - contention     12     0     0.04
    enq: DD - contention     10     0     0.04
    process terminate     8     41     5.08
    ges cgs registration     8     151     18.9
    buffer resize     7     0     0
    ktm: instance recovery     7     698     99.66
    LNS wait on LGWR     6     0     0
    ASM background starting     6     381     63.43
    gc cr block 3-way     5     0     0.08
    enq: PV - syncstart     5     9     1.74
    Global transaction acquire instance locks     4     4     1.09
    enq: RS - read alert level     4     0     0.04
    LGWR wait on LNS     3     0     0
    gc recovery     3     540     179.85
    Streams AQ: enqueue blocked on low memory     3     544     181.2
    DBWR range invalidation sync     3     17     5.83
    enq: DM - contention     3     0     0.03
    enq: RF - FSFO Observer Heartbeat     3     0     0.03
    enq: JS - q mem clnup lck     3     0     0
    DG Broker configuration file I/O     2     0     0
    enq: RC - Result Cache: Contention     2     493     246.6
    enq: KM - contention     2     0     0.03
    enq: RT - contention     2     0     0.04
    instance state change     2     0     0.12
    kkdlgon     2     10     5.11
    enq: TQ - INI contention     2     292     146.07
    enq: JS - contention     2     0     0
    ARCH wait for netserver start     1     400     400.02
    log file switch (checkpoint incomplete)     1     3     3.44
    JS coord start wait     1     50     50.09
    ges lmd and pmon to attach     1     1     1.26
    wait for tmc2 to complete     1     3     3.03
    control file heartbeat     1     400     400.02
    enq: SW - contention     1     0     0.04
    enq: PW - perwarm status in dbw0     1     0     0.09
    enq: FS - contention     1     0     0.04
    enq: XR - quiesce database     1     0     0.04
    enq: RS - write alert level     1     0     0.02
    enq: CN - race with init     1     0     0.03
    enq: FE - contention     1     4     3.77
    Wait for shrink lock2     1     10     10.03
    enq: IA - contention     1     0     0.02
    enq: RF - atomicity     1     0     0.05
    enq: RF - synchronization: aifo master     1     0     0.02
    enq: RF - RF - Database Automatic Disable     1     0     0.06
    enq: WP - contention     1     0     0.02
    enq: TB - SQL Tuning Base Cache Load     1     0     0.05
    enq: JS - evt notify     1     0     0.02Edited by: steffi on Mar 20, 2011 12:21 AM
    Edited by: steffi on Mar 20, 2011 8:18 AM
    Edited by: steffi on Mar 20, 2011 8:19 AM

    Text can be formatted by tagging the beginning and end of the block of text with the code ta
    \Formatted text goes here.
    \Example:
    This is formatted.When cutting and pasting text such as execution plans, excerpts from AWR reports, etc, it will maintain spacing and formatting, and make for much easier reading.
    As to the content, well, dumping the contents of v$system_event is pretty close to useless.
    As to the first three events you listed, 'enq: TX - index contention', 'enq: TX - row lock contention', 'enq: SQ - contention', well, all of those are easily tunable.
    First, for 'enq: SQ - contention', check your sequences. Do you have any NOCACHE sequences? Or sequences with small caches?
    As for 'enq: TX - row lock contention', well that's fairly self-explanatory. You have multiple sessions trying to lock the same row in the same table at the same time.
    Last, 'enq: TX - index contention', that's non-row level contention on an index. For example, if you have a unique index, insert a row w/ column value 1, but don't commit, then try to insert that same value from another session.
    But, before you do any of that, I think you need to clearly understand where the bottlenecks are. Try taking some AWR snapshots, about 5 minutes apart, when you're having performance problems. Look at the AWR report for that 5 minute period. In particular, look at your Top 5 timed events.
    Hope that helps,
    -Mark

  • DBMS_DATAPUMP with degree (parallel), causes worker run on wrong RAC node

    Hi @ all,
    I have a very very strange problem with DBMS_DATAPUMP :
    - I have a procedure that export (every night) some objects from my database on a partition mounted only in the node 1 in +/exports+ mount point
    - This Procedure is inside a job, inside a job group, inside a service that force this job to run ONLY on node 1 of my RAC (Enterprise 11.2) , and the service works fine:
    #srvctl status service -d ORCL
    Service EXPDPNODE01_SERVICE is running on instance(s) ORCL1- These objects are grown and the exp require too much time. I have introduced the parallelism
    Now with parallelism enabled my problem is that the exp raise many ORA-31617 apparently for no reason but the problem seems described here: [ http://oraclehandson.wordpress.com/2012/07/17/oracle-11-2-datapump-ora-31617-ora-19505-ora-27037/] and I have confirmed this problem because i have found one Worker run on second node ...
    With expdp (application) i can use option introduced in 11.2, cluster=N but i want to use DBMS_DATAPUMP ...
    can you help me?
    Edited by: iaio on 26-mar-2013 15.31

    Works! Thank you ... I was reading 11.1 but in 11.2 oracle have added many additional parameters!!!
    I have tested and works good! This is my (new) START_JOB :
    DBMS_DATAPUMP.START_JOB (
                             HANDLE       => HNDLR,
                             SKIP_CURRENT => 0,
                             ABORT_STEP   => 0,
                             CLUSTER_OK   => 0,
                             SERVICE_NAME => 'EXPDPNODE01_SERVICE'
    );Thank you again!

  • Datapump + 10.1

    Hi all,
    I have a problem with my export Datapump.
    It starts à 20:00 and stops directly. You can see the log here :
    Starting "SYSTEM"."SYS_EXPORT_FULL_22": USERID=system/********@bas FULL=Y direc
    tory=EXPORT_BASR1 dumpfile=expBAS_%U.dmp logfile=expBAS.log
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    Do you have an idea ?*
    Thanks.

    I'm not on rac.
    My directory is ok and tnsping too.
    But my export stay in this status :
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - Produc
    tion
    With the Partitioning, OLAP and Data Mining options
    Starting "SYSTEM"."SYS_EXPORT_FULL_23": USERID=system/********@bas FULL=Y direc
    tory=EXPORT_BASR1 dumpfile=testbas.dmp logfile=testbas.log
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA

  • Full database exp/imp  between RAC  and single database

    Hi Experts,
    we have a RAC database oracle 10GR2 with 4 node in linux. i try to duplicate rac database into single instance window database.
    there are same version both database. during importing, I need to create 4 undo tablespace to keep imp processing.
    How to keep one undo tablespace in single instance database?
    any experience of exp/imp RAC database into single instance database to share with me?
    Thanks
    Jim
    Edited by: user589812 on Nov 13, 2009 10:35 AM

    JIm,
    I also want to know can we add the exclude=tablespace on the impdp command for full database exp/imp?You can't use exclude=tablespace on exp/imp. It is for datapump expdp/impdp only.
    I am very insteresting in your recommadition.
    But for a full database impdp, how to exclude a table during full database imp? May I have a example for this case?
    I used a expdp for full database exp. but I got a exp error in expdp log as ORA-31679: Table data object "SALE"."TOAD_PLAN_TABLE" has long columns, and longs can not >be loaded/unloaded using a network linkHaving long columns in a table means that it can't be exported/imported over a network link. To exclude this, you can use the exclude expression:
    expdp user/password exclude=TABLE:"= 'SALES'" ...
    This will exclude all tables named sales. If you have that table in schema scott and then in schema blake, it will exclude both of them. The error that you are getting is not a fatal error, but that table will not be exported/imported.
    the final message as
    Master table "SYSTEM"."SYS_EXPORT_FULL_01" successfully loaded/unloaded
    Dump file set for SYSTEM.SYS_EXPORT_FULL_01 is:
    F:\ORACLEBACKUP\SALEFULL091113.DMP
    Job "SYSTEM"."SYS_EXPORT_FULL_01" completed with 1 error(s) at 16:50:26Yes, the fact that it did not export one table does not make the job fail, it will continue on exporting all other objects.
    . I drop database that gerenated a expdp dump file.
    and recreate blank database and then impdp again.
    But I got lots of error as
    ORA-39151: Table "SYSMAN"."MGMT_ARU_OUI_COMPONENTS" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    ORA-39151: Table "SYSMAN"."MGMT_BUG_ADVISORY" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    ......ORA-31684: Object type TYPE_BODY:"SYSMAN"."MGMT_THRESHOLD" already exists
    ORA-39111: Dependent object type TRIGGER:"SYSMAN"."SEV_ANNOTATION_INSERT_TR" skipped, base object type VIEW:"SYSMAN"."MGMT_SEVERITY_ANNOTATION" >already exists
    and last line as
    Job "SYSTEM"."SYS_IMPORT_FULL_01" completed with 2581 error(s) at 11:54:57Yes, even though you think you have an empty database, if you have installed any apps or anything, it may create tables that could exist in your dumpfile. If you know that you want the tables from the dumpfile and not the existing ones in the database, then you can use this on the impdp command:
    impdp user/password table_exists_action=replace ...
    If a table that is being imported exists, DataPump will detect this, drop the table, then create the table. Then all of the dependent objects will be created. If you don't then the table and all of it's dependent objects will be skipped, (which is the default).
    There are 4 options with table_exists_action
    replace - I described above
    skip - default, means skip the table and dependent objects like indexes, index statistics, table statistics, etc
    append - keep the existing table and append the data to it, but skip dependent objects
    truncate - truncate the existing table and add the data from the dumpfile, but skip dependent objects.
    Hope this helps.
    Dean

  • Data pump export full RAC database  in window single DB by network_link

    Hi Experts,
    I have a window 32 bit 10.2 database.
    I try to export a full rac database (350G some version with window DB) in window single database by dblink.
    exp syntax as
    exdpd salemanager/********@sale FULL=y DIRECTORY=dataload NETWORK_LINK=sale.net DUMPFILE=sale20100203.dmp LOGFILE=salelog20100203.log
    I created a dblink with fixed instance3. It was working for two day and display message as
    ORA-31693: Table data object "SALE_AUDIT"."AU_ITEM_IN" failed to load/unload and is being skipped due to error:
    ORA-29913: error in executing ODCIEXTTABLEPOPULATE callout
    ORA-01555: snapshot too old: rollback segment number with name "" too small
    ORA-02063: preceding line from sale.netL
    I stoped export and checked window target alert log.
    I saw some message as
    kupprdp: master process DM00 started with pid=16, OS id=4444
    to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_FULL_02', 'SYSTEM', 'KUPC$C_1_20100202235235', 'KUPC$S_1_20100202235235', 0);
    Tue Feb 02 23:56:12 2010
    The value (30) of MAXTRANS parameter ignored.
    kupprdp: master process DM00 started with pid=17, OS id=4024
    to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_FULL_01', 'SALE', 'KUPC$C_1_20100202235612', 'KUPC$S_1_20100202235612', 0);
    kupprdp: worker process DW01 started with worker id=1, pid=18, OS id=2188
    to execute - SYS.KUPW$WORKER.MAIN('SYS_EXPORT_FULL_01', 'SALE');
    In RAC instance alert.log. I saw message as
    SELECT /*+ NO_PARALLEL ("KU$") */ "ID","RAW_DATA","TRANSM_ID","RECEIVED_UTC_DATE ","RECEIVED_FROM","ACTION","ORAUSER",
    "ORADATE" FROM RELATIONAL("SALE_AUDIT"."A U_ITEM_IN") "KU$"
    How to fixed this error?
    add more undotbs space in RAC instance 3 or window database?
    Thanbks
    Jim
    Edited by: user589812 on Feb 4, 2010 10:15 AM

    I usually increate undo space. Is your undo retention set smaller than the time it takes to run the job? If it is, I would think you would need to do that. If not, then I would think it would be the space. You were in the process of exporting data when the job failed which is what I would have expected. Basically, DataPump want to export each table consistent to itself. Let's say that one of your tables is partitioned and it has a large partition and a smaller partition. DataPump attempts to export the larger partiitons first and it remembers the scn for that partition. When the smaller partitions are exported, it will use the scn to get the data from that partition as it would have looked like if it exported the data when the first partiiton was used. If you don't have partitioned tables, then do you know if some of the tables in the export job (I know it's full so that includes just about all of them) are having data added to them or removed from them? I can't think of anything else that would need undo while exporting data.
    Dean

  • How to upgade Oracle from 9.2 on SUN to Oracle 10g R2 RAC on Oracle EL4

    Hi all,
    I have production database Oracle 9i EE Release 9.2.0.5.0 - 64bit on SunOS 5.9 cca 100GB. I would like to upgrade Oracle 10g Relase2 RAC with 2 nodes on Oracle Enterprise Linux. Database availability is 24x7 and system breakdown must be minimal. Upgrade could be done in several independent steps. What is the best form of upgrade realisation?
    Thx for some idea
    Libor Veverka
    [email protected]

    My 2 cents:
    1. de-RAC 9i on solaris
    2. upgrade to 10g on solaris (why? take advantage of datapump, much faster than exp/imp)
    3. expdp from 10g solaris
    4. create single instance 10g on linux
    5. impdp on linux
    Please note there are a number of issues with 9ir2.10gr2 upgrade (check metalink for details).
    HTH,
    [email protected]

  • RAC instance won't start: ORA-00443: background process "VKRM" did not star

    I've logged a SR with Oracle, but while waiting for response from them...I'm stumped I can't find more out here about this article.
    On Oracle knowledgebase...any search for VKRM, gives about the same 3 articles relating to RDA (Remote Diagnostic Assistant).
    Not sure what went on here.
    I have a 5 node RAC cluster. All other instances seem to be running just fine.
    On one instance...some applications were getting an error like:
    ORA-01033: ORACLE initialization or shutdown in progress
    I looked in GRID...and it indicated that only two of the 5 nodes had this instance running...which was strange in that with srvctl, it showed ll 5 up and running:
    [oracle@server2 bin]$ ./srvctl status database -d INSTANCE
    Instance INSTANCE1 is running on node server1
    Instance INSTANCE2 is running on node server2
    Instance INSTANCE3 is running on node server3
    Instance INSTNANCE4 is running on node server4
    Instance INSTANCE5 is running on node server5
    Anyway, thought I'd poke around. I started with trying to get srvctl to stop instance #2...in GRID it seemed that instance 2,4 and 5 weren't working.
    srvctl stop instance -D INSTANCE -i INSTANCE2
    This just hung...
    I thought I'd cycle all the nodes...so, did a ctl-c out of that one, and did:
    [oracle@server2 bin]$ ./srvctl stop database -d INSTANCE -o abort
    PRCD-1124 : Failed to stop database INSTANCE and its services
    PRCR-1065 : Failed to stop resource (((((NAME STARTS_WITH ora.instance.) && (NAME ENDS_WITH .svc)) && (TYPE == ora.service.type)) && ((STATE != OFFLINE) || (TARGET != OFFLINE))) || (((NAME == ora.instance.db) && (TYPE == ora.database.type)) && (STATE != OFFLINE)))
    CRS-2675: Stop of 'ora.instance.db' on 'server5' failed
    CRS-2675: Stop of 'ora.instance.db' on 'server4' failed
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    Linux-x86_64 Error: 2: No such file or directory
    Process ID: 0
    Session ID: 0 Serial number: 0
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    Linux-x86_64 Error: 2: No such file or directory
    Process ID: 0
    Session ID: 0 Serial number: 0
    The ORA-01034 messages repeats a number of times...and the one with server4 and server5 repeated again too.
    I also got this:
    CRS-2680: Clean of 'ora.instance.db' on 'server2' failed
    CRS-2675: Stop of 'ora.instance.db' on 'server5' failed
    CRS-2675: Stop of 'ora.instance.db' on 'server4' failed
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    Linux-x86_64 Error: 2: No such file or directory
    Process ID: 0
    Session ID: 0 Serial number: 0
    I get similar messages when I try to restart.
    clusters seem to be up, and other instances seem to be ok
    Looking at alert log...I found some strangeness in the traces
    Starting background process VKRM
    Errors in file /u01/app/oracle/diag/rdbms/instance/INSTANCE2/trace/INSTANCE2_dbrm_26982.trc:
    ORA-00443: background process "VKRM" did not start
    Errors in file /u01/app/oracle/diag/rdbms/instance/INSTANCE2/trace/INSTANCE2_ora_27467.trc:
    ORA-00450: background process '' did not start
    Errors in file /u01/app/oracle/diag/rdbms/instance/INSTANCE2/trace/INSTANCE2_ora_27467.trc:
    ORA-00450: background process '' did not start
    Error 450 happened during db open, shutting down database
    USER (ospid: 27467): terminating the instance due to error 450
    LGWR waiting for instance termination
    Instance terminated by USER, pid = 27467
    ORA-1092 signalled during: ALTER DATABASE OPEN...
    opiodr aborting process unknown ospid (27467) as a result of ORA-1092
    Looking at the trace listed above:
    2011-05-09 12:17:18.305726 :84271119:db_trace:ksb.c@2157:ksbs1p_real(): [10254:6:464] KSBS1P: process DBRM trying to start background VKRM
    2011-05-09 12:17:18.305729 :8427111A:db_trace:ksb.c@2220:ksbs1p_real(): [10254:6:464] KSBS1P: process DBRM obtained PR enqueue to start background VK
    RM
    2011-05-09 12:17:18.306021 :8427111D:db_trace:ksb.c@2354:ksbs1p_real(): [10254:6:464] KSBS1P: creation error posted OER(1089)
    2011-05-09 12:17:18.306029 :8427111F:db_trace:ksb.c@2424:ksbs1p_real(): [10254:6:464] KSBS1P: out of loop: process did not start
    Trace Bucket Dump End: default bucket for process 6 (osid: 26982, DBRM)
    ORA-00443: background process "VKRM" did not start
    kskdbrmpa: reply error 450
    Any ideas? Again...I can't seem to find much of ANY information searching out there for the VKRM background process not starting...
    Thanks in advance,
    cayenne
    Edited by: cayenne on May 9, 2011 11:34 AM

    Anyone? Anyone? Bueller?
    Ok..have been on phone with Oracle support, and have them stumped so far.
    I've checked..other instances are running (except one other that failed to cleanly shut down with srvctl and same error messages).
    I've check...ASM is running on all 5 nodes. I've used crsctl to check CRS on all nodes..clustering seems ok.
    Memory while somewhat high...should have enough room...the system has never complained before, and this has been up on this config for over a year.
    I was able on node one...to fire up the first nodes instance using the pfile there...it came up. I started it restrict..and promptly ran a datapump export, and then shut back down.
    While Oracle support is going through logs and trace files I sent...I've got another 11Gr2 environment (3 node RAC) I've been using as a test environment...and am recreated the instance there temporarily to allow my developers to test and get past the upcoming deadline.
    But, I'm still puzzled as to the solution on the main cluster. This seem to be normal, except these two instances....
    any ideas on where and what to look for?
    cayenne

  • Datapump TTS from filesystem to ASM migration

    Hi Experts,
    I have to move the 1.5 TB of data which is on filesystem running on Solaris SPARC (single instance) to Linux x86 RAC (10.2.0.4) with ASM. I am considering the Transportable tablespaces option using datapump. But I have done the filesystem to filesystem datapump for TTS. Is there any way that I can use datapump TTS for filesystem to ASM directly ? Any suggestion and comments please.
    Thanks
    Shashi

    I'm not sure what your question is. Data Pump works with ASM and supports transportable. You would have to copy your data files from your file system to your asm file system, then do the export on your source, copy the dumpfile to your asm target and then run data pump. If you have more specific questions, please list them.
    Hope this helps.
    Thanks
    Dean

  • Datapump between two databases

    I have two databases, by usıng datapump I would lıke to move package,procedures and views to other database.
    expdp live/live@rac DIRECTORY=my_dir DUMPFILE=plsql.dmp INCLUDE=VIEW,PROCEDURE,PACKAGE,FUNCTION
    impdp test/test@orcl DIRECTORY=my_dir DUMPFILE=plsql.dmp remap_schema=LIVE:TEST
    but I couldnt manage to do the import,because I thınk second server dont know the dırectory that I created ın fırst machıne.

    >> I thınk second server dont know the dırectory that I created ın fırst machıne.
    Yes, you have to have directory created in Second server as well. Otherwise, it will not recognize as existing directory.
    Create the directory i.e. my_dire using CREATE DIRECTORY commands, and grant read and write privilege to test user, and finally do import.
    Regards,
    Sabdar Syed.

  • Datapump 에서 Alert.log 메시지 관련 ~ 좀 봐주세요 ^^

    안녕하세요 ~ OTN 여러분들 ^^
    바야흐로 봄이네요...^^ 그래서 그런지 싱숭생숭 하네요 ㅋㅋ
    춘곤증의 계절이라고 하던데...ㅋ 점심 먹고 나면 잠이 솔솔 ^^;
    오늘도 한가지 질문을 드릴려고 합니다 ~ 혹시라도 경험해 보신 분이나 아시는 분은 알려주시길 바랍니다 ^^
    환경은 오라클 10g RAC 로 구성이 되어 있으며 OS 는 HP-UX 입니다.
    헌데 Datapump 을 통해서 .... expdp 를 사용하였죠~
    $expdp system/manager directory=dumpfile job_name=datapump logfile=expdp_pump.log schemas=TEST estimate=statistics estimate_only=y
    ← 이런식으로 말이죠 ^^
    그랬더니 Alert.log 에 다음과 같이 메시지가 떳더라구요 ~
    Wed Feb 28 09:24:53 2007
    ALTER SYSTEM SET service_names='HLCDBK2','HLCDBK','SYS$SYS.KUPC$C_2_20070228092450.HLCDBK' SCOPE=MEMORY SID='HLCDBK2';
    Wed Feb 28 09:24:53 2007
    ALTER SYSTEM SET service_names='SYS$SYS.KUPC$C_2_20070228092450.HLCDBK','HLCDBK2','HLCDBK','SYS$SYS.KUPC$S_2_20070228092450.H
    LCDBK' SCOPE=MEMORY SID='HLCDBK2';
    kupprdp: master process DM00 started with pid=746, OS id=26313
    to execute - SYS.KUPM$MCP.MAIN('DATAPUMP', 'SYSTEM', 'KUPC$C_2_20070228092450', 'KUPC$S_2_20070228092450', 0);
    kupprdp: worker process DW01 started with worker id=1, pid=215, OS id=26440
    to execute - SYS.KUPW$WORKER.MAIN('DATAPUMP', 'SYSTEM');
    Wed Feb 28 09:25:55 2007
    ALTER SYSTEM SET service_names='SYS$SYS.KUPC$S_2_20070228092450.HLCDBK','HLCDBK2','HLCDBK' SCOPE=MEMORY SID='HLCDBK2';
    Wed Feb 28 09:25:55 2007
    ALTER SYSTEM SET service_names='HLCDBK2','HLCDBK' SCOPE=MEMORY SID='HLCDBK2';
    이런식으로 Alter system set 문이 발생하더라구요~ 그래서 저의 개인적인 생각으로는 RAC 환경하고 관련이 있다고 생각하는데요... single instance 상황에서 테스트를 해본 결과는 alter system set 문이 발생하지를 않더라구요 ~
    혹시라도 이와 관련하여 아시는 분은 답변 부탁드립니다 ~^^

    우선 이 문제는 해결을 하였습니다 ~
    일종의 Oracle code issue(일종의 메시지 버그) 인 것 같습니다.
    오라클에서도 11g 에서 해결해야 할 문제라고 합니다.

  • How to Enhance or Speedup Datapump Schema Refresh

    Hi All,
    My environment is Oracle 10g Rel2 on RAC (Linux 5), central Storage... We are doing Datapump Schema refresh(Export) with parallel and excluding only grants; parameters. Is there any possibility to enhace the import process by importing metadata and data separately. We tried Parallel option, but still import takes much longer to complete. Is it anything to do with breaking import in two parts. But i dont know how to break the import task.
    Kindly reply with appropriate examples..Would appreciate your response.
    Thanks,
    Abdul Mannan

    Hi Hitgon,
    Please find the following details:
    how many physical cpu/core? - 16 CPU
    whether the os is 64bit/32bit? - 64 bit (64 bit Grid Servers)
    how many physical memory in os? - 12 Servers with (256 GB) and 4 Servers (64GB)
    how many memory assign to db? - SGA - 12GB & PGA - 6GB
    what the size of database? more than 1 TB
    what the size of schema you are going to import? around 700 - 950 GB
    import process running from remote client/server or it's the server import - we are initiating the job on server using (Putty).
    Using the following Export and Import Commands:
    Datapump Export parfile:
    JOB_NAME=Schema1_20120621_100122_27198
    DIRECTORY=DPUMP_Export
    PARALLEL=4
    DUMPFILE=Schema1_20120621_100122_27198_%U.dmp
    LOGFILE=DPUMP_EXPORT:DPEXPORT_20120621_100122_27198.log
    CONTENT=ALL
    FILESIZE=10000M
    Datapump Import parfile:
    SCHEMAS=Schema1
    JOB_NAME=Schema1_20120118_213551
    DIRECTORY=DPUMP_EXPORT
    PARALLEL=4
    DUMPFILE=SCHEMA1_20120114_025007_19273_%U.dmp
    LOGFILE=DPUMP_EXPORT:DPIMP_20120118_213551.log
    REMAP_TABLESPACE=Schema1DATA:Schema2DATA
    EXCLUDE=GRANT
    EXCLUDE=STATISTICS
    Would appreciate your reply..
    Thanks,
    Abdul Mannan

  • Impdp in RAC : can not specify dump file in node 1

    I have oracle cluster RAC in 2 nodes : rac1, rac2.
    I have created directory in 2 nodes, and copied file dump to directory of rac1.
    When i run impdp from rac1, the impdp did not use file dump in rac1, it found file dump in rac2, so it thrown error : ORA-39000: bad dump file specification
    ORA-31640: unable to open dump file "/u01/datapump/aa.DMP" for read
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory.
    I must copied file dump to rac2, the impdp run normally.
    What the master with cluster RAC, how to config RAC so the impdp read file dump in login node.

    File dump has been exported in another standalone server.
    I create dir in OS level first then create dir in database, then copy file pump.
    I think this problem relate to config in oracle cluster, the impdp is not check file dump in 2 nodes.
    I found this problem when I not create the datapump in rac2. It allways return the error log file create fail. I must created datapump dir in rac2, then impdp just used path in rac2, file log was created in rac2, impdp not use datapump in rac1.

  • RAC & Db upgrade

    Hi,
    I have 10.1.0.4 rac & db on aix with 2 nodes.I will upgrade these to latest version 11.2.0.3. does this action plan have true approach for upgrade ? By this action plan i think that i do not upgrade cluster site.
    1. Take db to non-rac
    2. Upgrade db to 11.2.0.3 without rac upgrade.
    3. fresh install 11.2.0.3 of cluster
    4. register upgraded db to new cluster
    Thanks.

    Hi;
    I agree Dmorgan. It seems best way to install 11.2.0.3 wiht lastest PSU than use rman or datapump.
    You may want to review below note also:
    Minimizing Downtime During Production Upgrade [ID 478308.1]
    Master Note For Oracle Database Upgrades and Migrations [ID 1152016.1]
    Different Upgrade Methods For Upgrading Your Database [ID 419550.1]
    PS:Please dont forget to change thread status to answered if it possible when u belive your thread has been answered, it pretend to lose time of other forums user while they are searching open question which is not answered,thanks for understanding
    Regard
    Helios

Maybe you are looking for