Job aborted! (11570)

Hello,
can some one help me with this:
'Failed to retrieve data from the database.
Details:  [Database Vendor Code: 942 ]^Failed to retrieve data from the database.
Details:  [Database Vendor Code: 942 ]Failed to retrieve data from the database.Failed to export the report.
^Error in File RPT23.tmp {3CD2D088-1FF5-41BA-85FC-F64537081183}.rpt:
Failed to retrieve data from the database.
Details:  [Database Vendor Code: 942 ]'. - Job aborted! (11570)
Thanks

is this when you run the report? locally or through the viewer or application?
Error in File RPT23.tmp {3CD2D088-1FF5-41BA-85FC-F64537081183}.rpt
what is RPT23.tmp?
what is your dataset?  it may need to be verified, if it is a stored procedure it may not exist in the database.

Similar Messages

  • How to send mail after job abort

    Hi,
    how can I send an email when a job is aborted? Or is there a system event if a job aborts?
    Thanks for your help
    Lutz

    You might want to look into ABAP.  Basically a generic job that would get called, to send email, whenever a certain job "aborts", but then you'd also need a wrapper job for detecting the bad return code. help.sap.com has some info on that (using events to trigger jobs), like this:
    http://help.sap.com/saphelp_47x200/helpdata/en/fa/096f33543b11d1898e0000e8322d00/frameset.htm
    or this (the full-control method)
    http://help.sap.com/saphelp_47x200/helpdata/en/fa/096f33543b11d1898e0000e8322d00/frameset.htm
    CCMS can also monitor job status too, as the other poster said,
    http://help.sap.com/saphelp_47x200/helpdata/en/fa/096f33543b11d1898e0000e8322d00/frameset.htm
    however it's very limited.  Especially if you want to use logic for the alerts (ie only certain jobs should alert upon failure, certain times of day, recipients, etc).  It's all possible with CCMS, I've seen it done before using scripts that read CCMS external alerting but it's probably easier just adding event triggers to your critical jobs.

  • Catching job abort/system stop in batch report

    Is there any way in a report running in background to catch a job abort request for the report's job or that the system is being stopped?  This would allow for an elegant exit.  Thanks.

    Hello Ralph,
    If I have understood your requirement correctly, then I'm afraid what you want cannot be accomplished.
    Let us say, for example, that the application server has suddenly begun to shutdown , say due to power failure. How can an event like this ever communicate with a background job?
    I know you would say that you did not mean a case like this, but System Shutdown , Job Abort etc., are all pre-emptive. I do not think you will be able to determine that within your program. The fcat that a program continues to be running is in itself a proof that no such an event has occurred.
    Even if you were to be able to check for a particular condition in your code, let us say , for every 10 seconds, the system shutdown event might occur just immediately after the execution of the logic. And you will not be able to have the "graceful" exit any more.
    Hope I have understood your problem correctly and if yes, my explanation has been satisfactory.
    Regards,
    Anand Mandalika

  • Background job ABORTED in ECC

    Hi,
    I am working for an upgrade project. We have a program in 4.7 which start background job using the FM: JOB_OPEN and submit7 standard programs with variant using FM: JOB_SUBMIT. Please find the program details below.
    1. RMCP6FLS
    2. RMCP6RES
    3. RMCA9026
    4. RMCP6BIN
    5. RMCP6RES
    6. RMCA9026
    7. RMCP6BIN
    After this close the job using FM: JOB_CLOSE and find the job status using the FM SHOW_JOBSTATE.
    For the same variant it is working fine in 4.7 but in ECC FM SHOW_JOBSTATE aborted = 'X'.
    Thanks and Regards
    Joby Jacob

    Please investigate yourself first, if you are still stuck, open another thread describing what you have analysed so far, providing all necessary details right away.
    Thread locked.
    Thomas

  • Contact sheet print job aborted in snow leopard

    Recently upgraded to snow leopard. When printing a large contact sheet of over 50 raw files the print job is aborted. Printing to a HP Z3100. Nailed it down to amount of data 50 pics at 20mb each seems to be limit. No problem under Leopard. I can convert the files to Jpg and print 80 of them, file size is 1/3. Anyone run into this issue?

    Hmm.. Can't say as I've seen that, and I've printed more than 120 photos 12+meg at a time (including some 80meg tif files).  Don't know what it might be... how's your scratch space, or the space on your main disk?
    Cheers!

  • DS 4.1 job aborts with Access Violation

    Hi Friends,
    I have DS 4.1 Jobs running on prod server, it has been running fine for the last 1 year. Today it aborted with below error. To be honest nothing changed in the job or the environment.
    System Exception <ACCESS_VIOLATION> occurred. Process dump option is off. Process is not dumped.
    Call stack:
                                                        (null)
    Registers:
    RAX=0000000000000000  RBX=000000000B7241F8 RCX=0000000083EF8368  RDX=0000000000000000  RSI=0000000001CFFD68
    RDI=0000000000000000  RBP=000000000043EFB0 RSP=0000000001CFFB40  RIP=00000000801AC561  FLG=0000000000010246
    R8=000000000AC85920   R9=0000000002C2F989 R10=0000000000000000  R11=0000000000000202  R12=0000000000000000
    R13=0000000000000000  R14=000000000000FFFF  R15=0000000077054610
    Exception code: C0000005 ACCESS_VIOLATION
    Fault address:  00000001801AC561 01:00000000001AB561 C:\Program Files (x86)\SAP BusinessObjects\Data Services\bin\acta.dll
    The only change i observed this morning was the license key of the BI 4 server expired. we were running on temp licenses because we are awaiting for a perm license key from SAP.
    After i removed the temp key and added the new perm keys, i validated and reran the job and to my surprise the job ran successfully.
    We have a BI with DI edge license and BI & DS are integrated installs but we have 2 separate keys i.e. 1 key for BI and 1 for DS. All i want to understand is why did the job fail given that only the BI key expired and not the DS key.
    Has anyone seen anything similar?
    Thanks,

    Hi Gleomar Perez,
    For the above issue,please check with the oracle server name that was given  while adding the repository in the Central Management Console.It should be given as the fully qualified name ( machine.domain.com format).
    Regards,
    Mubashir Hussain

  • Trigger Workflow for Background Job Abort

    Hi,
    I want to start a workflow when a Background Job fails.
    I could find the BOs BPJOB & BPJOB_SMAN and the event BPJobSchedMan.Aborted, but it is not getting triggered when a background job cancelled.
    Let me know if any specific settings need to be made.
    Appreciate your help. Thanks in advance.
    Rgds,
    Kumar

    Hi,
    Did  you manage to get this working?  i am also using BPJOB abort even but its not getting triggered.
    Thanks.
    Rgds,
    Pankaj

  • Regarding Background Job Abort using

    Hi ,
    I am using FM BP_JOB_ABORT to abort background jobs through program,but it taking longtime to cancel background jobs in SM37 after exectuing FM BP_JOB_ABORT in program..
    will i have to write  any statement in program after this  FM BP_JOB_ABORT to cancel the background job immediately through program..
    Awaiting for valuable suggestion..
    Regards
    Anil

    Hi,
    I looking for FM to cancel background jobs quikly through program..
    Regards
    Anil

  • CNVMBTDTL_CLUSTER_DELETE job aborted - missing required key field

    Hi Experts,
    I believe this problem has been reported before.
    The CNV_MBT_CNVMBTDTL_CLUSTER_DELETE (Postprocessing: PC002_DTL_CLUSTER_DELETE_SND) was aborted.
    In the job log, the failure was due to missing required fields.
    The recommendation is to run the SGEN transaction to regenerate the DMIS components.
    But what I would llike understand is what is causing this problem? Why do the DMIS component need to be regen?
    Thank you!
    Best Regards,
    Bee Huang

    Hi Niraj and all,
    Thank you for your response!
    I have not run the SGEN but would like to understand the reason. Any note to address this problem? We just upgraded to SP14 for DMIS/DMIS_CNT. We did not see this problem earlier on in a lower SP. Is the bug related to SP14?
    Thank you again.
    Best Regards,
    Bee Huang

  • FM or a method to abort a batch job

    Hi All,
    I am looking for a way in which I can abort a background job which is either scheduled or currently running programmatically. Is there a way to do this even if the job currently running is waiting due to either a wait statement or for a response from a RFC.
    Immediate Responses would be appreciated.
    Thanks & Regards,
    Pawan

    Hi Rich,
    Thanks for your reply.
    I have been using this Function Module earlier.
    There is a problem using ths FM.
    Please see below the snippet of the code I am using to abort the job using the FM 'BP_JOB_ABORT'.
    REPORT  yumt_trigger_abort                      .
    TYPES: BEGIN OF ty_joblist,
             jobname  TYPE tbtco-jobname,
             jobcount TYPE tbtco-jobcount,
           END OF ty_joblist.
    DATA: lv_jobname  TYPE tbtcjob-jobname,
          lv_count    TYPE tbtcjob-jobcount,
          lv_repid    TYPE sy-repid,
          wa_joblist  TYPE ty_joblist,
          lit_joblist TYPE TABLE OF ty_joblist.
    CONSTANTS  : lc_jobclass TYPE tbtcjob-jobclass VALUE 'A',
                 gc_true     TYPE flag VALUE 'X'.
    REFRESH : lit_joblist.
    Schedule the jobs
    DO 2 TIMES.
      CLEAR:  lv_repid , lv_jobname , lv_count.
      IF sy-index = 1.
        lv_repid = 'YUMT_TRITEST1'.
        lv_jobname = 'YUMT_TRITEST1'.
      ELSE.
        lv_repid = 'YUMT_TRITEST2'.
        lv_jobname = 'YUMT_TRITEST2'.
      ENDIF.
    Pass the other parameters and schedule a job.
      CALL FUNCTION 'JOB_OPEN'
        EXPORTING
          jobname          = lv_jobname
          jobclass         = lc_jobclass
        IMPORTING
          jobcount         = lv_count
        EXCEPTIONS
          cant_create_job  = 1
          invalid_job_data = 2
          jobname_missing  = 3
          OTHERS           = 4.
      IF sy-subrc <> 0.
       MESSAGE s000(zumt_gen) WITH text-007." Background Job Creation Error.
        RETURN.
      ENDIF.
    Message to be included in Job description as a step
    Background job <JOB NAME> <JOB ID> scheduled.
      MESSAGE s000(zumt_gen) WITH text-002 lv_jobname
                                  lv_count text-003.
      CALL FUNCTION 'JOB_SUBMIT'
        EXPORTING
          authcknam               = sy-uname
          jobcount                = lv_count
          jobname                 = lv_jobname
          report                  = lv_repid
        EXCEPTIONS
          bad_priparams           = 1
          bad_xpgflags            = 2
          invalid_jobdata         = 3
          jobname_missing         = 4
          job_notex               = 5
          job_submit_failed       = 6
          lock_failed             = 7
          program_missing         = 8
          prog_abap_and_extpg_set = 9
          OTHERS                  = 10.
      IF sy-subrc <> 0.
    Message to be included in Job description as a step
      MESSAGE s000(zumt_gen) WITH text-008. "Background Job Submitting Error
        RETURN.
      ENDIF.
      CALL FUNCTION 'JOB_CLOSE'
        EXPORTING
          jobcount             = lv_count
          jobname              = lv_jobname
          strtimmed            = gc_true
        EXCEPTIONS
          cant_start_immediate = 1
          invalid_startdate    = 2
          jobname_missing      = 3
          job_close_failed     = 4
          job_nosteps          = 5
          job_notex            = 6
          lock_failed          = 7
          invalid_target       = 8
          OTHERS               = 9.
      IF sy-subrc <> 0.
    Message to be included in Job description as a step
        MESSAGE s000(zumt_gen) WITH text-009. " Background Job Closing Error
        RETURN.
      ENDIF.
      CLEAR: wa_joblist.
      wa_joblist-jobname = lv_jobname.
      wa_joblist-jobcount = lv_count.
      APPEND wa_joblist TO lit_joblist.
    ENDDO.
    Abort the jobs
    WAIT UP TO 2 SECONDS.
    CLEAR: wa_joblist.
    LOOP AT lit_joblist INTO wa_joblist.
      CALL FUNCTION 'BP_JOB_ABORT'
        EXPORTING
          jobcount                   = wa_joblist-jobcount
          jobname                    = wa_joblist-jobname
        EXCEPTIONS
          checking_of_job_has_failed = 1
          job_abort_has_failed       = 2
          job_does_not_exist         = 3
          job_is_not_active          = 4
          no_abort_privilege_given   = 5
          OTHERS                     = 6.
      IF sy-subrc <> 0.
        MESSAGE s000(zumt_gen) WITH text-010 wa_joblist-jobname sy-subrc.
        " Job Abortion Failed
      ELSE.
        MESSAGE s000(zumt_gen) WITH text-001 wa_joblist-jobname.
        " Job Abortion Success
      ENDIF.
    ENDLOOP.
    However when the program runs the FM runs successfully for both the jobs but when we check the status of the job in sm36 We see that they are still running.
    Thanks & Regards,
    Pawan

  • Job SEND_SYSTEM_RELATIONSHIP_TO_SUPPORT aborts

    Hi all,
    for some reason the Job SEND_SYSTEM_RELATIONSHIP_TO_SUPPORT (Step: AI_SC_SEND_SYSTEM_RELATIONSHIP) aborts at every execution.
    The Job Log looks like:
    - Job started
    - Step 001 started (Program AI_SC_SEND_SYSTEM_RELATIONSHIP, Variant , Username xyz)
    - SLM set as administrating Solution-Manager-System
    - Following Errormessages returned from SAP Support Portal:
    - Job aborted after System-Exception ERROR_MESSAGE
    Does anyone know what could be the reason?
    Regards,
    Benjamin Ziegler

    Hi Afi,
    I have deleted the job and recreated it manually; now it runs fine!
    Thanks for the quick help?
    I only don't understand the note you are referring to; are you sure it's the right note number?
    Best regards,
    Koen

  • Running job will not abort

    Hey guysDoes anyone know how to force the abort of an job? I have a job that hung for more than 10 days, last week i cancelled the running job but now it's still on "cancelling". I also tried to abort the job over the command line withdfbm job abort 197850the result isError: Job 197850 is already being aborted.Since 6 days I have this message, actually I cannot take a backup of the desired dataset as the dataset waits for the job to finish.Can anyone help me?Regards.sal

    and if you are using Windows, after identified job number (GUI or dfpm job list -v jobs-running), then start the Task Manager, "Processes" tab, make sure you "show processes from all users", View>Select column... , check the "Command Line" column, you must now see the ...\dfpm.exe" job run job_id, the you can end this one only.

  • TSPITR - Recover a Dropped Tablespace aborting due dependency over XDB.

    I found an issue using TSPITR to Recover a Dropped Tablespace when XML DB (XDB) is installed on a different tablespace from "SYSAUX".
    During execution of TSPITR RMAN use Data Pump to export/import Tablespace Metadata as Data Pump has dependency over XDB views and related objects and XDB is installed on a set of tablespace not included on dummy database then EXPDP fails.
    Errors raised ORA-39123, ORA-00376, ORA-01110.
    I didn't found any limitation about XDB location on documentation or MoS.
    This looks like a Oracle Bug.
    See this:
    XDB installed on TABLESPACE XDB.
    SQL> select owner,tablespace_name
        from dba_segments
        where owner='XDB'
         GROUP BY owner,tablespace_name ;
    OWNER                          TABLESPACE_NAME
    XDB                            XDBDropping Tablespace SOE.
    SQL> select to_char(sysdate,'dd-mm-yyyy hh24:mi:ss') actual_date from dual;
    ACTUAL_DATE
    17-12-2012 11:57:44
    SQL> drop tablespace SOE including contents and datafiles;
    Tablespace dropped.Restoring Dropped Tablespace.
    RMAN> list backup tag TAG20121217T112226;
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    69      Full    995.61M    DISK        00:02:41     17-12-2012 11:25:07
            BP Key: 69   Status: AVAILABLE  Compressed: YES  Tag: TAG20121217T112226
            Piece Name: +DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947
      List of Datafiles in backup set 69
      File LV Type Ckp SCN    Ckp Time            Name
      1       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/system.349.799325249
      2       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/sysaux.348.799325249
      3       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/undotbs1.347.799325251
      4       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/users.335.799325251
      5       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/audit_aux.287.799330083
      6       Full 1628648    17-12-2012 11:22:26
      7       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/xdb.311.801410833
    RMAN>
    run{
    SET NEWNAME FOR DATAFILE 6 TO "+DS8000_DG";
    RECOVER TABLESPACE SOE
    UNTIL TIME "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')"
    AUXILIARY DESTINATION '/u01/app/oracle/flash_recovery_area01/stage_area_tspitr';
    executing command: SET NEWNAME
    Starting recover at 17-12-2012 12:01:29
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=18 device type=DISK
    RMAN-05026: WARNING: presuming following set of tablespaces applies to specified point-in-time
    List of tablespaces expected to have UNDO segments
    Tablespace SYSTEM
    Tablespace UNDOTBS1
    Creating automatic instance, with SID='FCox'
    initialization parameters used for automatic instance:
    db_name=PROD
    db_unique_name=FCox_tspitr_PROD
    compatible=11.2.0.0.0
    db_block_size=8192
    db_files=200
    sga_target=280M
    processes=50
    db_create_file_dest=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr
    log_archive_dest_1='location=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr'
    #No auxiliary parameter file used
    starting up automatic instance PROD
    Oracle instance started
    Total System Global Area     292278272 bytes
    Fixed Size                     2220880 bytes
    Variable Size                109055152 bytes
    Database Buffers             176160768 bytes
    Redo Buffers                   4841472 bytes
    Automatic instance created
    List of tablespaces that have been dropped from the target database:
    Tablespace SOE
    contents of Memory Script:
    # set requested point in time
    set until  time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
    # restore the controlfile
    restore clone controlfile;
    # mount the controlfile
    sql clone 'alter database mount clone database';
    # archive current online log
    sql 'alter system archive log current';
    # avoid unnecessary autobackups for structural changes during TSPITR
    sql 'begin dbms_backup_restore.AutoBackupFlag(FALSE); end;';
    executing Memory Script
    executing command: SET until clause
    Starting restore at 17-12-2012 12:01:46
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=16 device type=DISK
    channel ORA_AUX_DISK_1: starting datafile backup set restore
    channel ORA_AUX_DISK_1: restoring control file
    channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/autobackup/2012_12_17/s_802265113.276.802265115
    channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/autobackup/2012_12_17/s_802265113.276.802265115 tag=TAG20121217T112513
    channel ORA_AUX_DISK_1: restored backup piece 1
    channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:00:05
    output file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/controlfile/o1_mf_8dy9ghpz_.ctl
    Finished restore at 17-12-2012 12:01:52
    sql statement: alter database mount clone database
    sql statement: alter system archive log current
    sql statement: begin dbms_backup_restore.AutoBackupFlag(FALSE); end;
    contents of Memory Script:
    # set requested point in time
    set until  time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
    # set destinations for recovery set and auxiliary set datafiles
    set newname for clone datafile  1 to new;
    set newname for clone datafile  3 to new;
    set newname for clone datafile  2 to new;
    set newname for clone tempfile  1 to new;
    set newname for datafile  6 to
    "+DS8000_DG";
    # switch all tempfiles
    switch clone tempfile all;
    # restore the tablespaces in the recovery set and the auxiliary set
    restore clone datafile  1, 3, 2, 6;
    switch clone datafile all;
    executing Memory Script
    executing command: SET until clause
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    renamed tempfile 1 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_temp_%u_.tmp in control file
    Starting restore at 17-12-2012 12:01:58
    using channel ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: starting datafile backup set restore
    channel ORA_AUX_DISK_1: specifying datafile(s) to restore from backup set
    channel ORA_AUX_DISK_1: restoring datafile 00001 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_%u_.dbf
    channel ORA_AUX_DISK_1: restoring datafile 00003 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_%u_.dbf
    channel ORA_AUX_DISK_1: restoring datafile 00002 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_%u_.dbf
    channel ORA_AUX_DISK_1: restoring datafile 00006 to +DS8000_DG
    channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947
    channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947 tag=TAG20121217T112226
    channel ORA_AUX_DISK_1: restored backup piece 1
    channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:09:15
    Finished restore at 17-12-2012 12:11:13
    datafile 6 switched to datafile copy
    input datafile copy RECID=5 STAMP=802267873 file name=+DS8000_DG/prod/datafile/soe.487.802267321
    datafile 1 switched to datafile copy
    input datafile copy RECID=6 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_8dy9gs09_.dbf
    datafile 3 switched to datafile copy
    input datafile copy RECID=7 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_8dy9ov17_.dbf
    datafile 2 switched to datafile copy
    input datafile copy RECID=8 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_8dy9hxsc_.dbf
    contents of Memory Script:
    # set requested point in time
    set until  time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
    # online the datafiles restored or switched
    sql clone "alter database datafile  1 online";
    sql clone "alter database datafile  3 online";
    sql clone "alter database datafile  2 online";
    sql clone "alter database datafile  6 online";
    # recover and open resetlogs
    recover clone database tablespace  "SOE", "SYSTEM", "UNDOTBS1", "SYSAUX" delete archivelog;
    alter clone database open resetlogs;
    executing Memory Script
    executing command: SET until clause
    sql statement: alter database datafile  1 online
    sql statement: alter database datafile  3 online
    sql statement: alter database datafile  2 online
    sql statement: alter database datafile  6 online
    Starting recover at 17-12-2012 12:11:14
    using channel ORA_AUX_DISK_1
    starting media recovery
    archived log for thread 1 with sequence 15 is already on disk as file +DG_16M/prod/archivelog/2012_12_17/thread_1_seq_15.277.802267317
    channel ORA_AUX_DISK_1: starting archived log restore to default destination
    channel ORA_AUX_DISK_1: restoring archived log
    archived log thread=1 sequence=14
    channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/backupset/2012_12_17/annnf0_tag20121217t112512_0.264.802265113
    channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/backupset/2012_12_17/annnf0_tag20121217t112512_0.264.802265113 tag=TAG20121217T112512
    channel ORA_AUX_DISK_1: restored backup piece 1
    channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:00:01
    archived log file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/1_14_801941068.dbf thread=1 sequence=14
    channel clone_default: deleting archived log(s)
    archived log file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/1_14_801941068.dbf RECID=141 STAMP=802267877
    archived log file name=+DG_16M/prod/archivelog/2012_12_17/thread_1_seq_15.277.802267317 thread=1 sequence=15
    media recovery complete, elapsed time: 00:00:01
    Finished recover at 17-12-2012 12:11:19
    database opened
    contents of Memory Script:
    # make read only the tablespace that will be exported
    sql clone 'alter tablespace  SOE read only';
    # create directory for datapump import
    sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''
    /u01/app/oracle/flash_recovery_area01/stage_area_tspitr''";
    # create directory for datapump export
    sql clone "create or replace directory TSPITR_DIROBJ_DPDIR as ''
    /u01/app/oracle/flash_recovery_area01/stage_area_tspitr''";
    executing Memory Script
    sql statement: alter tablespace  SOE read only
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''
    Performing export of metadata...
       EXPDP> Starting "SYS"."TSPITR_EXP_FCox":
       EXPDP> ORA-39123: Data Pump transportable tablespace job aborted
    ORA-00376: file 7 cannot be read at this time
    ORA-01110: data file 7: '+DS8000_DG/prod/datafile/xdb.311.801410833'
    Removing automatic instance
    shutting down automatic instance
    database closed
    database dismounted
    Oracle instance shut down
    Automatic instance removed
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_temp_8dyb0qok_.tmp deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_3_8dyb0kfp_.log deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_2_8dyb0fm9_.log deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_1_8dyb07xv_.log deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_8dy9hxsc_.dbf deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_8dy9ov17_.dbf deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_8dy9gs09_.dbf deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/controlfile/o1_mf_8dy9ghpz_.ctl deleted
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of recover command at 12/17/2012 12:12:19
    RMAN-06962: Error received during export of metadata
    RMAN-06965: Datapump job has stopped
    RMAN-06960:    EXPDP> Job "SYS"."TSPITR_EXP_FCox" stopped due to fatal error at 12:12:02
    RMAN>If I drop tablespace SOE and XDB on Target database and try to restore both (SOE,XDB) export will succeeds but during IMPORT Metadata RMAN will faill due XDB dependency.
    In my view, they (mean Oracle) need find way to include XDB Tablespace at DUMMY database only to export (EXPDP) Metadata purposes.
    If someone get a workaround for this error will be grateful.
    P.S: This is not a Production environment and will soon open a SR on Oracle Support.

    Hi Michael,
    I did not find solution to use TSPITR to recover dropped tablespace if XML DB is installed in a non-default tablespace (i.e outside of SYSAUX).
    As far I could understand, the only way to perform this RESTORE is using the old fashion. By creating clone database excluding all unnecessary tablespace, exporting and importing metadata from desired tablespace manually.
    Since Oracle has not solved the issue I created my own workaround.
    @All
    If someone is able to use TSPITR to restore dropped tablespace when XDB is installed in a non-default tablespace, please let me know!!!
    The procedure below is supported on All Oracle Version since 10.1 until 11.2.
    I DON'T RECOMMEND to create AUXILIARY (clone) database using the same server from TARGET because you can damage the TARGET database.
    On this example I'm using same server.
    See this example:
    ### ON TARGET DATABASE
    SQL> select count(1) SEG_COUNT from dba_segments where tablespace_name='SOE';
      SEG_COUNT
            32
    SQL> ! date
    Wed Dec 19 15:08:11 GMT-02:00 2012
    SQL> drop tablespace soe including contents and datafiles;
    Tablespace dropped.
    ### CHECK TABLESPACE TO BE EXCLUDED FROM RESTORE
    SQL> select tablespace_name
        from dba_tablespaces
         where tablespace_name not in ('SYSTEM','SYSAUX','UNDOTBS1','SOE','XDB','TEMP');
    TABLESPACE_NAME
    USERS
    AUDIT_AUX
    ### CHECK BACKUP ON TARGET DATABASE
    RMAN> list backup completed between "sysdate-1" and "sysdate";
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    95      Full    986.56M    DISK        00:03:18     19-12-2012 14:41:59
            BP Key: 95   Status: AVAILABLE  Compressed: YES  Tag: TAG20121219T143840
            Piece Name: /fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp
      List of Datafiles in backup set 95
      File LV Type Ckp SCN    Ckp Time            Name
      1       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/system.349.799325249
      2       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/sysaux.348.799325249
      3       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/undotbs1.347.799325251
      4       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/users.335.799325251
      5       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/audit_aux.287.799330083
      6       Full 1675258    19-12-2012 14:38:42
      7       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/xdb.311.801410833
    BS Key  Size       Device Type Elapsed Time Completion Time
    96      1.16M      DISK        00:00:00     19-12-2012 14:42:08
            BP Key: 96   Status: AVAILABLE  Compressed: YES  Tag: TAG20121219T144208
            Piece Name: /fra/PROD/backupset/2012_12_19/o1_mf_annnn_TAG20121219T144208_8f3vm0k3_.bkp
      List of Archived Logs in backup set 96
      Thrd Seq     Low SCN    Low Time            Next SCN   Next Time
      1    8       1675198    19-12-2012 14:38:25 1675365    19-12-2012 14:42:07
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    97      Full    7.52M      DISK        00:00:02     19-12-2012 14:42:11
            BP Key: 97   Status: AVAILABLE  Compressed: NO  Tag: TAG20121219T144209
            Piece Name: /fra/PROD/autobackup/2012_12_19/o1_mf_s_802449729_8f3vm2cc_.bkp
      SPFILE Included: Modification time: 19-12-2012 14:36:30
      SPFILE db_unique_name: PROD
      Control File Included: Ckp SCN: 1675376      Ckp time: 19-12-2012 14:42:09
    RMAN>
    ### PREPARING ENV
    cp $ORACLE_HOME/dbs/orapwprod $ORACLE_HOME/dbs/orapwtspitr
    cd $ORACLE_HOME/dbs
    ## Creating INIT from TSPITR Instance
    vi inittspitr.ora
    DB_NAME=prod
    DB_UNIQUE_NAME=tspitr
    CONTROL_FILES=/stage_area_tspitr/PROD/controlfile/control01.ctl
    REMOTE_LOGIN_PASSWORDFILE=exclusive
    COMPATIBLE =11.2.0
    DB_BLOCK_SIZE=8192
    memory_target=700M
    streams_pool_size=100M
    ### creating stage area to deploy database
    mkdir -p  /fra/stage_area_tspitr/PROD
    cd /fra/stage_area_tspitr/PROD
    mkdir controlfile datafile onlinelog
    #### RESTORING DATABASE TSPITR
    export ORACLE_SID=tspitr
    $ rman target /
    RMAN> startup nomount
    Oracle instance started
    Total System Global Area     730714112 bytes
    Fixed Size                     2225088 bytes
    Variable Size                553651264 bytes
    Database Buffers             167772160 bytes
    Redo Buffers                   7065600 bytes
    RMAN> restore controlfile from '/fra/PROD/autobackup/2012_12_19/o1_mf_s_802449729_8f3vm2cc_.bkp';
    Starting restore at 19-12-2012 15:16:27
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=146 device type=DISK
    channel ORA_DISK_1: restoring control file
    channel ORA_DISK_1: restore complete, elapsed time: 00:00:03
    output file name=/fra/stage_area_tspitr/PROD/controlfile/control01.ctl
    Finished restore at 19-12-2012 15:16:31
    RMAN> startup mount
    database is already started
    database mounted
    released channel: ORA_DISK_1
    ### REPORT SCHEMA TO GET DATAFILE NUMBER TO BE RESTORED
    RMAN> report schema;
    RMAN-06139: WARNING: control file is not current for REPORT SCHEMA
    Report of database schema for database with db_unique_name TSPITR
    List of Permanent Datafiles
    ===========================
    File Size(MB) Tablespace           RB segs Datafile Name
    1    700      SYSTEM               ***     +DS8000_DG/prod/datafile/system.349.799325249
    2    600      SYSAUX               ***     +DS8000_DG/prod/datafile/sysaux.348.799325249
    3    200      UNDOTBS1             ***     +DS8000_DG/prod/datafile/undotbs1.347.799325251
    4    5        USERS                ***     +DS8000_DG/prod/datafile/users.335.799325251
    5    100      AUDIT_AUX            ***     +DS8000_DG/prod/datafile/audit_aux.287.799330083
    6    0        SOE                  ***     +DS8000_DG/tspitr/datafile/soe.368.802445265
    7    100      XDB                  ***     +DS8000_DG/prod/datafile/xdb.311.801410833
    List of Temporary Files
    =======================
    File Size(MB) Tablespace           Maxsize(MB) Tempfile Name
    1    385      TEMP                 32767       +DS8000_DG/prod/tempfile/temp.300.799325437
    RMAN>
    ### GET LOGFILE NAME to RENAME to new location.
    $ sqlplus / as sysdba
    SQL> select member from v$logfile;
    MEMBER
    +DS8000_DG/prod/onlinelog/group_1.330.802352807
    +DS8000_DG/prod/onlinelog/group_2.331.802352815
    +DS8000_DG/prod/onlinelog/group_3.321.802352821
    run {
    set newname for datafile 1 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 2 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 3 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 7 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 6 to "+DS8000_DG";
    restore datafile 1,2,3,7,6 from tag 'TAG20121219T143840' ;
    switch datafile 1;
    switch datafile 2;
    switch datafile 3;
    switch datafile 7;
    switch datafile 6;
    switch tempfile all;
    set until time "to_date('19-12-2012 14:42:00','dd-mm-yyyy hh24:mi:ss')";
    recover database SKIP TABLESPACE  "USERS", "AUDIT_AUX" delete archivelog;
    ## Renaming ONLINELOG
    sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_1.330.802352807'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group1.rdo''";
    sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_2.331.802352815'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group2.rdo''";
    sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_3.321.802352821'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group3.rdo''";
    alter database open resetlogs;
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    Starting restore at 19-12-2012 15:32:13
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting datafile backup set restore
    channel ORA_DISK_1: specifying datafile(s) to restore from backup set
    channel ORA_DISK_1: restoring datafile 00001 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1
    channel ORA_DISK_1: restoring datafile 00002 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSAUX_FNO-2
    channel ORA_DISK_1: restoring datafile 00003 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-UNDOTBS1_FNO-3
    channel ORA_DISK_1: restoring datafile 00007 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-XDB_FNO-7
    channel ORA_DISK_1: restoring datafile 00006 to +DS8000_DG
    channel ORA_DISK_1: reading from backup piece /fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp
    channel ORA_DISK_1: piece handle=/fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp tag=TAG20121219T143840
    channel ORA_DISK_1: restored backup piece 1
    channel ORA_DISK_1: restore complete, elapsed time: 00:12:35
    Finished restore at 19-12-2012 15:44:48
    datafile 1 switched to datafile copy
    input datafile copy RECID=6 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1
    datafile 2 switched to datafile copy
    input datafile copy RECID=8 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSAUX_FNO-2
    datafile 3 switched to datafile copy
    input datafile copy RECID=10 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-UNDOTBS1_FNO-3
    datafile 7 switched to datafile copy
    input datafile copy RECID=12 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-XDB_FNO-7
    datafile 6 switched to datafile copy
    input datafile copy RECID=14 STAMP=802453489 file name=+DS8000_DG/tspitr/datafile/soe.304.802452733
    renamed tempfile 1 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-TEMP_FNO-1 in control file
    executing command: SET until clause
    Starting recover at 19-12-2012 15:44:49
    using channel ORA_DISK_1
    Executing: alter database datafile 4 offline
    Executing: alter database datafile 5 offline
    starting media recovery
    archived log for thread 1 with sequence 8 is already on disk as file +DS8000_DG/prod/onlinelog/group_2.331.802352815
    archived log file name=+DS8000_DG/prod/onlinelog/group_2.331.802352815 thread=1 sequence=8
    media recovery complete, elapsed time: 00:00:02
    Finished recover at 19-12-2012 15:44:55
    sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_1.330.802352807'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group1.rdo''
    sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_2.331.802352815'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group2.rdo''
    sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_3.321.802352821'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group3.rdo''
    database opened
    ## Exporting Tablespace Metadata
    RMAN >
    run {
    sql 'alter tablespace  SOE read only';
    # create directory for datapump export
    sql  "create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''";
    # grant on directory to SYSTEM
    sql "grant read,write on directory TSPITR_DIROBJ_DPDIR to system";
    # exporting metadata
    host "expdp system/oracle@localhost:1521/tspitr DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_FULL_CHECK=Y TRANSPORT_TABLESPACES=SOE";
    using target database control file instead of recovery catalog
    sql statement: alter tablespace  SOE read only
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''
    sql statement: grant read,write on directory TSPITR_DIROBJ_DPDIR to system
    Export: Release 11.2.0.3.0 - Production on Wed Dec 19 15:53:02 2012
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, Automatic Storage Management, Oracle Label Security, OLAP,
    Data Mining, Oracle Database Vault and Real Application Testing options
    Starting "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01":  system/********@localhost:1521/tspitr DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_FULL_CHECK=Y TRANSPORT_TABLESPACES=SOE
    Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
    Processing object type TRANSPORTABLE_EXPORT/TABLE
    Processing object type TRANSPORTABLE_EXPORT/INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK
    Master table "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01" successfully loaded/unloaded
    Dump file set for SYSTEM.SYS_EXPORT_TRANSPORTABLE_01 is:
      /fra/stage_area_tspitr/exp_transport_soe.dmp
    Datafiles required for transportable tablespace SOE:
      +DS8000_DG/tspitr/datafile/soe.304.802452733
    Job "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01" successfully completed at 15:54:28
    host command complete
    ## Stop Database and destroy it to make sure that Datafile 6 will not be used by this database;
    RMAN>
    run {
    shutdown immediate;
    ## removing controlfile and SYSTEM Tablespace to make sure that database will not be opened
    host "rm -f /fra/stage_area_tspitr/PROD/controlfile/control01.ctl /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1";
    database closed
    database dismounted
    Oracle instance shut down
    host command complete
    ## TARGET Database
    ## Plugging Tablespace on TARGET database.
    $ export ORACLE_SID=prod
    $ rman target /
    run {
    # create directory for datapump import
    sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''";
    # grant on directory to SYSTEM
    sql "grant read,write on directory TSPITR_DIROBJ_DPDIR to system";
    host "impdp system/oracle@localhost:1521/prod DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_DATAFILES=''+DS8000_DG/tspitr/datafile/soe.304.802452733''";
    sql "alter tablespace soe read write";
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''
    sql statement: grant read,write on directory TSPITR_DIROBJ_DPDIR to system
    Import: Release 11.2.0.3.0 - Production on Wed Dec 19 16:09:30 2012
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, Automatic Storage Management, Oracle Label Security, OLAP,
    Data Mining, Oracle Database Vault and Real Application Testing options
    Master table "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully loaded/unloaded
    Starting "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01":  system/********@localhost:1521/prod DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_DATAFILES=+DS8000_DG/tspitr/datafile/soe.304.802452733
    Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
    Processing object type TRANSPORTABLE_EXPORT/TABLE
    Processing object type TRANSPORTABLE_EXPORT/INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK
    Job "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully completed at 16:10:03
    host command complete
    sql statement: alter tablespace soe read write
    ### Recompiling Invalid Objects and Checking objects from  tablespace SOE
    $ sqlplus / as sysdba
    SQL>  @$ORACLE_HOME/rdbms/admin/utlrp.sql
    TIMESTAMP
    COMP_TIMESTAMP UTLRP_BGN  2012-12-19 16:10:34
    OBJECTS WITH ERRORS
                      0
    ERRORS DURING RECOMPILATION
                              0
    Function created.
    PL/SQL procedure successfully completed.
    Function dropped.
    PL/SQL procedure successfully completed.
    SQL> select count(1) SEG_COUNT from dba_segments where tablespace_name='SOE';
      SEG_COUNT
            32
    SQL> SELECT COUNT(1) OBJ_INVALID FROM DBA_OBJECTS WHERE STATUS != 'VALID';
    OBJ_INVALID
              0
    ## PERFORM A FULL BACKUP AFTER RESTORE TABLESPACE.Regards,
    Levi Pereira

  • Reg: Function Module thats used to check the status of JOBS and mail it

    Hi Experts,
    Is there any Standard Function module thats going to find the status of the jobs and mail this status to specied email ID in .xls format. While checking status function module should give whether job is - Aborted, Successful, Running, Waiting, Locked, Error, Warning, Starting Date, Ending Date of the Job.
    Please help me,

    Hello Nagaraj,
    You can find out the status of a job with the SHOW_JOBSTATE function module. You provide this function module with a job name and job number. It returns one of the six possible statuses of the job:
    ABORTED TYPE C,
    FINISHED TYPE C,
    PRELIMINARY TYPE C,
    READY TYPE C,
    RUNNING TYPE C,
    SCHEDULED TYPE C.
    Following is the link for the same:
    http://help.sap.com/saphelp_sm32/helpdata/en/fa/096e10543b11d1898e0000e8322d00/content.htm
    Warm Regards,
    Riki Dash

  • TSPITR abort

    Hi!!!
    I'm under Oracle Linux 5 and Oracle 11.2 Database.
    I'm trying to test a tablespace drop case and to recover it with TSPITR but I wasn't success! it aborted as follows:
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of recover command at 06/08/2012 13:48:18
    RMAN-06965: Datapump job has stopped
    RMAN-06961: IMPDP> Job "SYS"."TSPITR_IMP_jqip" stopped due to fatal error at 13:48:11
    RMAN-06961: IMPDP> ORA-39123: Data Pump transportable tablespace job aborted
    ORA-01565: error in identifying file '+DATA/orcl/datafile/dw_datos_m01.260.785419471'
    ORA-17503: ksfdopn:2 Failed to open file +DATA/orcl/datafile/dw_datos_m01.260.785419471
    ORA-15012: ASM file '+DATA/orcl/datafile/dw_datos_m01.260.785419471' does not exist
    I don't know what to do, could anyone help me please please!
    Thanks
    José

    But how do I restore this datafile could you help me please!? I think as I dropped, including datafile, it deleted control file entry...may be I have to
    restore controlfile before?
    Regards
    José
    Edited by: jamv on Jun 11, 2012 12:15 PM

Maybe you are looking for