Schema reporting

Back when I worked with SQL/DS on a VM (mainframe) system, I was able to type a command and a nice little report of the entire schema was produced. Is there a way to do this with Oracle 9i via OEM? Thanx.

Most of the information displayed is available in data dictionary views:
user_objects
user_tables
user_views
user_indexes
user_tab_columns
user_ind_columns
user_constraints
user_cons_columns
user_tab_comments
etc...
It would be possible to use these to create a script to generate a report on the schema which could be very close to the one you posted (possibly not in a word format though).

Similar Messages

  • Opening SQL tab of other's user schema reports error

    Hi,
    I just try to open the SQL table tab to get the SQL script of a table, and it only works when the schema is owned by the user. If you try to open SQL tab on another user table it reports:
    ORA-31603: object "COM_BANK" of type TABLE not found in schema "PROVIS_OWN"
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.DBMS_METADATA", line 653
    ORA-06512: at "SYS.DBMS_METADATA", line 1260
    ORA-06512: at line 1
    I can access that table through a synonym and grants.
    Server is Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit Production
    I am using 2.1 of Oracle SQL Developer.
    Thanks in advance,

    This is a reported issue, but I'm not sure it will be fixed in the upcoming patch. You can try searching the forum.
    As workaround, you'd need more permissions (don't know which exactly; the DBA role works for me).
    Regards,
    K.

  • Authorization Scheme - Report

    I was looking for a report that would show the utilization of the authorization schemes, defined in an application. Where do I find or how do I create such a report?
    Thanks,
    Denes Kubicek

    Thanks Scott,
    next time I will keep my eyes open while searching for that stuff. Your application - HTML db - is a good thing. Thanks for that as well.
    Denes Kubicek

  • Authentication Scheme Report  "Bug"

    Hi -
    Since htmldb.oracle.com is down, I am working locally on my XE. Version 2.1.0.0.0.39. Stumbled across a strange bug I've never seen before - just wanted to report it if you dont all know about it before. Couldnt find it referenced in the search.
    I imported an application successfully, but when I go to Shared Components => Authentication Themes, the report says there are not any. The default display type is Icons. However switching over to Details lists all of them. Its page 801 if you are keeping track at home.
    I also tried creating a theme from scratch in a brand new application. Same behavior - I can only see them when selecting details instead of icons. I did not notice this problem on the hosted site, but its possible there are different versions of the software.
    Any one else run across this?

    We are aware of that bug in the XE version but thanks for the details.
    Scott

  • How to Serial No Printing in Report - group Above Report

    In Scott Schema , report data query - select rownum, emp, ename, job, deptno from emp;
    Reprot type : group Above.
    If I am not using group above, serial no like 1,2,3 is printing correctly. But if i use group above, serial no are not printing on order. it is printing like 5,4,3, 8.. li
    How to print serially with group above.
    Thanks in advance for the help

    O.Developer wrote:
    In Scott Schema , report data query - select rownum, emp, ename, job, deptno from emp;
    Reprot type : group Above.
    If I am not using group above, serial no like 1,2,3 is printing correctly. But if i use group above, serial no are not printing on order. it is printing like 5,4,3, 8.. li
    How to print serially with group above.Avoid rownum and use summary column at detail block and set following property.
    Summary function :count
    summary column : ename
    reset at : <group name>Your serial problem solved.
    Hope this helps...
    Hamid

  • Report with dynamic order by (binding of parameter)

    hy,
    my problem is the following:
    i want to let the user customize the ordering of a portal report.
    but the binding does not really work, the bind variable seems to
    have no effect in the order by clause.
    code sample:
    select * from SCOTT.EMP order by :p_order
    then, portal gets the right value for the parameter:
    :p_order binding with HIREDATE
    but: this results in an emp table sorted by the empno.....
    and not by the hiredate
    do you have any ideas how to get this working ???
    thanx, martin allhoff

    Hi, the way to do this is to pass in a different set of
    parameters. Try not using the :bind parameter in the report
    query, but instead, call your report like this:
    http:// ... /pls/portal30/Schema.report.show?
    p_arg_names=_ORD_BY_COL&P_arg_values=HIREDATE
    then try calling
    http:// ... /pls/portal30/Schema.report.show?
    p_arg_names=_ORD_BY_COL&P_arg_values=EMPNO
    to see if indeed it is dynamically selecting the order by clause
    based on the info passed in. Also, check my syntax, I cant
    remember if ORDBY_COL is correct or not. But any possible
    options should be listed for you in the "Display Call Interface"
    screen for the report.

  • XML Schema for Java Version 1.0

    Hi everybody,
    1.) in my opinion, the sample schema report.xsd seems to be incorrect with respect to the keyref definition (selector element), but the parser doesn't care anyway. I have modified the report.xml sample to include duplicates of the zip code and invalid keyrefs, but the parser still doesn't care. Despite the notice that unique, key and keyref doesn't work with SAX, it obviously doesn't work at all.
    2.) The "types" contained within the Schema-Definition are not accessible from outside the "oracle" packages, and so far are rather useless. It would be an obvious advantage to access the type- and validation classes of the Schema processor in order to allow for interactive validation of user input, when documents are composed by gathering user input from a http-request. Will this change, as far as XMLSchema is becoming a standard?
    TIA
    Achim

    I'm glad someone else has noticed that unique keyref and key don't seem to be working with the Dom Parser. If they don't work WHY ARE THEY (key, keyref, unique) IN THE EXAMPLE'S THAT ARE DOWNLOADED WITH THE SCHEMA PARSER. In report.xsd, a file downloaded with the example, it makes clear usage of unique, key, and key ref. But if you violate the schema definitions in the corresponding file report.xml the parser doesn't complain whatsoever. The only time it barfs is if you change the keyref refer attribute to something other than "pNumKey". It obviously has to work. No bone head would send example files along with their product that didn't work.
    So, if anyone at ORACLE or elsewhere has figured out how to use unique, key, or keyref please respond with an explination of how to correctly use them with the parser. Your name will be blessed throughout the ages as a most kind and venerable person. You will be a hallmark, a standard, a shining light for all future generations of what a human being should be! Okay maybe I'm going a little overboard but I'm DESPERATE. With no books or collateral on how this stupid thing works all I can do is hack.
    -Thanks

  • PDF printing--tomcat/cocoon installed, but report server not found------

    As I understood, APEX 3.1 requires either BI/FO tools to enable PDF printing.
    As I do not have BI license, and as I happened to check a very good link on a blog which had a detailed viewlet on apache-cocoon route for PDF print enabling, I downloaded the tomcat, cocoon source files, built the cocoon .WAR, put it in required place as instructed, started tomcot (ran into pronblems because of port conflits, resolved these by configuring different ports for 8005 and 8080 ports).
    Now, when I want to configure the external report server, I am stuck.
    I am not able to locate 'Instances' tab, or 'manage environment settings ' tab in 'manage services'
    I am able to find only following tabs in 'manage services' under HOME.--> Administration ----> Manage services
    1) session state
    2) caching
    3) workspace preferences
    4) applications
    5) workspace
    6) workspace schema reports
    7) Logs
    Can anyone help me to find out where to configure the 'report server' now?
    Thanks,
    suresh.

    Sorry, it was a silly problem:) You have to be logged in as admin to 'see' the 'instances' tab. maybe in all the tomcat/cocoon build/config issues, this simply got off my head!
    now, the problem of print server config is over, I have configured the same. the 'print' link is available now in report region, when I run the report. but when I click the 'print' link, it is taking ages...or rather it is still processing...!!
    Can anyone please point out to me now what is the problem...?
    thanks.

  • Non OBIEE Publisher reports and dashboards

    I have developed several Publisher reports that source data from direct SQL queries against an Oracle DB (They're not star schema reports that are appropriate for OBIEE). The reports are parameterized and direct SQL requests are used to populate the lists of values for the parameters. My task is to make them available on a dashboard.
    I have a report attached to a dashboard but I cannot figure out how to prompt for the parameters in the dashboard. Everything I've read about dashboard prompts and presentation variables assumes the report is sourced by OBIEE.
    Is there a way to do this?

    You have to create parameter in reports and as well create prompt in dashboard with same name.
    With this link, the values will be passed into report parameter.
    Use that parameter in report to filter

  • Q: How to create reports with no login required?

    I would like to make some reports available to internal people but I dont want/need login security. Is this possible? how?

    Daryl,
    Change the application's authentication scheme to DATABASE. You should see that scheme on the authentication scheme report, just click the "make current" link and confirm. Then the whole application will be public. To make selected pages public in your application, leave the authentication scheme as is and change the authentication attribute of those pages to "Page is Public" on the page attributes form.
    Scott

  • Authorization scheme (using {not} Scheme)

    I have build a change password page and every user, except user with a Guest role (= GUEST SCHEME) have access to that page.
    I defined a scheme GUEST for users with the GUEST role. When I define the page with Authorization scheme {not}GUEST this isn't working everyone has access to the page, also the guest users.
    am I misunderstanding the {not}scheme choice or is something else wrong.
    Fred.

    Fred,
    I have solved it with the work around I mentioned before:I read what you said very carefully but thought it reckless to conclude that the workaround was successful because you just said "To work around the problem, I did xyz" without indicating the outcome.
    The authorization schemes on navigation tabs fire also on the default login pageYes they do, they fire on every page whether or not the page template accommodates a navigation bar. This looks like a bug to me.
    Is there a "authorization scheme report" which shows all the objects where the authorization scheme is defined.Shared Components > Authorization Schemes > Utilization (slightly different in each version).
    Scott

  • Workspace to parse SQL statements from multiple database schema's

    Hi ,
    When go through the following link of workspace admin user :
    Home>HTML DB Workspace Administration>Manage Services>Schema Reports>Workspace Schemas there I saw "Your workspace has the privilege to parse SQL statements using the following database schemas. Note that the terms database schema and database user can be used interchangeably.
    " at the right pane.
    So how can i make my workspace to parse SQL statements from multiple database schema's ?That must be listed there .But i have only one schema in view ? Can i make it multiple ?If then How ?
    Anybody have an idea ?
    ROSY

    Assign as many schemas to your workspace as you want using the administration app. Read about it in the doc.
    Scott

  • Multiple MapTriggers allowed on the same cache?

    I'm trying to register a pair of trigger listeners on the same cache (Coherence 3.5.1)
    (See below for xml scheme reported in storage node log).
    It appears as if only the first listener is active.
    Both triggers work if configured in isolation.
    We could partially solve the problem by creating a "super" trigger that encapsulates both functions, but this would prevent us from removing one of them at runtime by unregistering the listener.
    In summary:
    1) If multiple triggers are not supported, then shouldn't the config below be rejected?
    2) If multiple triggers are supported, could anyone suggest what I'm doing wrong?
    3) If multiple triggers are supported when added programatically, but not via config, is there a recommended way of managing them?
    Thanks,
    Philip
    <distributed-scheme>
      <scheme-name>dist-default</scheme-name>
      <backing-map-scheme>
        <local-scheme/>
      </backing-map-scheme>
      <autostart>true</autostart>
      <serializer>
        <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
        <init-params>
          <init-param>
            <param-type>string</param-type>
            <param-value>java_pof_config.xml</param-value>
          </init-param>
        </init-params>
      </serializer>
      <thread-count>2</thread-count>
      <listener>
        <class-scheme>
          <class-factory-name>com.psinventory.cache.EntryVersioningTrigger</class-factory-name>
          <method-name>createMapTriggerListener</method-name>
        </class-scheme>
      </listener>
      <listener>
        <class-scheme>
          <class-factory-name>com.psinventory.cache.EntryLoggingTrigger</class-factory-name>
          <method-name>createMapTriggerListener</method-name>
        </class-scheme>
      </listener>
    </distributed-scheme>

    Hi Philip
    Multiple triggers do work if you add them using NamedCache.addMapListener(...). If you configure the trigger using the xml cache config file then only the first trigger is added. I will enter an internal bug for this issue.
    Please be aware of 2 items:
    1) The order in which the trigger is called is not specified in the API or any doc (that I know of). I did some testing and the last trigger added was the first called. However, you cannot rely on this behavior unless it is documented.
    2) The output of the first trigger is passed as input to the second trigger. If you cannot rely on the ordering of the triggers then this could be a real problem.
    Paul

  • TSPITR - Recover a Dropped Tablespace aborting due dependency over XDB.

    I found an issue using TSPITR to Recover a Dropped Tablespace when XML DB (XDB) is installed on a different tablespace from "SYSAUX".
    During execution of TSPITR RMAN use Data Pump to export/import Tablespace Metadata as Data Pump has dependency over XDB views and related objects and XDB is installed on a set of tablespace not included on dummy database then EXPDP fails.
    Errors raised ORA-39123, ORA-00376, ORA-01110.
    I didn't found any limitation about XDB location on documentation or MoS.
    This looks like a Oracle Bug.
    See this:
    XDB installed on TABLESPACE XDB.
    SQL> select owner,tablespace_name
        from dba_segments
        where owner='XDB'
         GROUP BY owner,tablespace_name ;
    OWNER                          TABLESPACE_NAME
    XDB                            XDBDropping Tablespace SOE.
    SQL> select to_char(sysdate,'dd-mm-yyyy hh24:mi:ss') actual_date from dual;
    ACTUAL_DATE
    17-12-2012 11:57:44
    SQL> drop tablespace SOE including contents and datafiles;
    Tablespace dropped.Restoring Dropped Tablespace.
    RMAN> list backup tag TAG20121217T112226;
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    69      Full    995.61M    DISK        00:02:41     17-12-2012 11:25:07
            BP Key: 69   Status: AVAILABLE  Compressed: YES  Tag: TAG20121217T112226
            Piece Name: +DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947
      List of Datafiles in backup set 69
      File LV Type Ckp SCN    Ckp Time            Name
      1       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/system.349.799325249
      2       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/sysaux.348.799325249
      3       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/undotbs1.347.799325251
      4       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/users.335.799325251
      5       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/audit_aux.287.799330083
      6       Full 1628648    17-12-2012 11:22:26
      7       Full 1628648    17-12-2012 11:22:26 +DS8000_DG/prod/datafile/xdb.311.801410833
    RMAN>
    run{
    SET NEWNAME FOR DATAFILE 6 TO "+DS8000_DG";
    RECOVER TABLESPACE SOE
    UNTIL TIME "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')"
    AUXILIARY DESTINATION '/u01/app/oracle/flash_recovery_area01/stage_area_tspitr';
    executing command: SET NEWNAME
    Starting recover at 17-12-2012 12:01:29
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=18 device type=DISK
    RMAN-05026: WARNING: presuming following set of tablespaces applies to specified point-in-time
    List of tablespaces expected to have UNDO segments
    Tablespace SYSTEM
    Tablespace UNDOTBS1
    Creating automatic instance, with SID='FCox'
    initialization parameters used for automatic instance:
    db_name=PROD
    db_unique_name=FCox_tspitr_PROD
    compatible=11.2.0.0.0
    db_block_size=8192
    db_files=200
    sga_target=280M
    processes=50
    db_create_file_dest=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr
    log_archive_dest_1='location=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr'
    #No auxiliary parameter file used
    starting up automatic instance PROD
    Oracle instance started
    Total System Global Area     292278272 bytes
    Fixed Size                     2220880 bytes
    Variable Size                109055152 bytes
    Database Buffers             176160768 bytes
    Redo Buffers                   4841472 bytes
    Automatic instance created
    List of tablespaces that have been dropped from the target database:
    Tablespace SOE
    contents of Memory Script:
    # set requested point in time
    set until  time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
    # restore the controlfile
    restore clone controlfile;
    # mount the controlfile
    sql clone 'alter database mount clone database';
    # archive current online log
    sql 'alter system archive log current';
    # avoid unnecessary autobackups for structural changes during TSPITR
    sql 'begin dbms_backup_restore.AutoBackupFlag(FALSE); end;';
    executing Memory Script
    executing command: SET until clause
    Starting restore at 17-12-2012 12:01:46
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=16 device type=DISK
    channel ORA_AUX_DISK_1: starting datafile backup set restore
    channel ORA_AUX_DISK_1: restoring control file
    channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/autobackup/2012_12_17/s_802265113.276.802265115
    channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/autobackup/2012_12_17/s_802265113.276.802265115 tag=TAG20121217T112513
    channel ORA_AUX_DISK_1: restored backup piece 1
    channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:00:05
    output file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/controlfile/o1_mf_8dy9ghpz_.ctl
    Finished restore at 17-12-2012 12:01:52
    sql statement: alter database mount clone database
    sql statement: alter system archive log current
    sql statement: begin dbms_backup_restore.AutoBackupFlag(FALSE); end;
    contents of Memory Script:
    # set requested point in time
    set until  time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
    # set destinations for recovery set and auxiliary set datafiles
    set newname for clone datafile  1 to new;
    set newname for clone datafile  3 to new;
    set newname for clone datafile  2 to new;
    set newname for clone tempfile  1 to new;
    set newname for datafile  6 to
    "+DS8000_DG";
    # switch all tempfiles
    switch clone tempfile all;
    # restore the tablespaces in the recovery set and the auxiliary set
    restore clone datafile  1, 3, 2, 6;
    switch clone datafile all;
    executing Memory Script
    executing command: SET until clause
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    renamed tempfile 1 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_temp_%u_.tmp in control file
    Starting restore at 17-12-2012 12:01:58
    using channel ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: starting datafile backup set restore
    channel ORA_AUX_DISK_1: specifying datafile(s) to restore from backup set
    channel ORA_AUX_DISK_1: restoring datafile 00001 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_%u_.dbf
    channel ORA_AUX_DISK_1: restoring datafile 00003 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_%u_.dbf
    channel ORA_AUX_DISK_1: restoring datafile 00002 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_%u_.dbf
    channel ORA_AUX_DISK_1: restoring datafile 00006 to +DS8000_DG
    channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947
    channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947 tag=TAG20121217T112226
    channel ORA_AUX_DISK_1: restored backup piece 1
    channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:09:15
    Finished restore at 17-12-2012 12:11:13
    datafile 6 switched to datafile copy
    input datafile copy RECID=5 STAMP=802267873 file name=+DS8000_DG/prod/datafile/soe.487.802267321
    datafile 1 switched to datafile copy
    input datafile copy RECID=6 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_8dy9gs09_.dbf
    datafile 3 switched to datafile copy
    input datafile copy RECID=7 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_8dy9ov17_.dbf
    datafile 2 switched to datafile copy
    input datafile copy RECID=8 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_8dy9hxsc_.dbf
    contents of Memory Script:
    # set requested point in time
    set until  time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
    # online the datafiles restored or switched
    sql clone "alter database datafile  1 online";
    sql clone "alter database datafile  3 online";
    sql clone "alter database datafile  2 online";
    sql clone "alter database datafile  6 online";
    # recover and open resetlogs
    recover clone database tablespace  "SOE", "SYSTEM", "UNDOTBS1", "SYSAUX" delete archivelog;
    alter clone database open resetlogs;
    executing Memory Script
    executing command: SET until clause
    sql statement: alter database datafile  1 online
    sql statement: alter database datafile  3 online
    sql statement: alter database datafile  2 online
    sql statement: alter database datafile  6 online
    Starting recover at 17-12-2012 12:11:14
    using channel ORA_AUX_DISK_1
    starting media recovery
    archived log for thread 1 with sequence 15 is already on disk as file +DG_16M/prod/archivelog/2012_12_17/thread_1_seq_15.277.802267317
    channel ORA_AUX_DISK_1: starting archived log restore to default destination
    channel ORA_AUX_DISK_1: restoring archived log
    archived log thread=1 sequence=14
    channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/backupset/2012_12_17/annnf0_tag20121217t112512_0.264.802265113
    channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/backupset/2012_12_17/annnf0_tag20121217t112512_0.264.802265113 tag=TAG20121217T112512
    channel ORA_AUX_DISK_1: restored backup piece 1
    channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:00:01
    archived log file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/1_14_801941068.dbf thread=1 sequence=14
    channel clone_default: deleting archived log(s)
    archived log file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/1_14_801941068.dbf RECID=141 STAMP=802267877
    archived log file name=+DG_16M/prod/archivelog/2012_12_17/thread_1_seq_15.277.802267317 thread=1 sequence=15
    media recovery complete, elapsed time: 00:00:01
    Finished recover at 17-12-2012 12:11:19
    database opened
    contents of Memory Script:
    # make read only the tablespace that will be exported
    sql clone 'alter tablespace  SOE read only';
    # create directory for datapump import
    sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''
    /u01/app/oracle/flash_recovery_area01/stage_area_tspitr''";
    # create directory for datapump export
    sql clone "create or replace directory TSPITR_DIROBJ_DPDIR as ''
    /u01/app/oracle/flash_recovery_area01/stage_area_tspitr''";
    executing Memory Script
    sql statement: alter tablespace  SOE read only
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''
    Performing export of metadata...
       EXPDP> Starting "SYS"."TSPITR_EXP_FCox":
       EXPDP> ORA-39123: Data Pump transportable tablespace job aborted
    ORA-00376: file 7 cannot be read at this time
    ORA-01110: data file 7: '+DS8000_DG/prod/datafile/xdb.311.801410833'
    Removing automatic instance
    shutting down automatic instance
    database closed
    database dismounted
    Oracle instance shut down
    Automatic instance removed
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_temp_8dyb0qok_.tmp deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_3_8dyb0kfp_.log deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_2_8dyb0fm9_.log deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_1_8dyb07xv_.log deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_8dy9hxsc_.dbf deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_8dy9ov17_.dbf deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_8dy9gs09_.dbf deleted
    auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/controlfile/o1_mf_8dy9ghpz_.ctl deleted
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of recover command at 12/17/2012 12:12:19
    RMAN-06962: Error received during export of metadata
    RMAN-06965: Datapump job has stopped
    RMAN-06960:    EXPDP> Job "SYS"."TSPITR_EXP_FCox" stopped due to fatal error at 12:12:02
    RMAN>If I drop tablespace SOE and XDB on Target database and try to restore both (SOE,XDB) export will succeeds but during IMPORT Metadata RMAN will faill due XDB dependency.
    In my view, they (mean Oracle) need find way to include XDB Tablespace at DUMMY database only to export (EXPDP) Metadata purposes.
    If someone get a workaround for this error will be grateful.
    P.S: This is not a Production environment and will soon open a SR on Oracle Support.

    Hi Michael,
    I did not find solution to use TSPITR to recover dropped tablespace if XML DB is installed in a non-default tablespace (i.e outside of SYSAUX).
    As far I could understand, the only way to perform this RESTORE is using the old fashion. By creating clone database excluding all unnecessary tablespace, exporting and importing metadata from desired tablespace manually.
    Since Oracle has not solved the issue I created my own workaround.
    @All
    If someone is able to use TSPITR to restore dropped tablespace when XDB is installed in a non-default tablespace, please let me know!!!
    The procedure below is supported on All Oracle Version since 10.1 until 11.2.
    I DON'T RECOMMEND to create AUXILIARY (clone) database using the same server from TARGET because you can damage the TARGET database.
    On this example I'm using same server.
    See this example:
    ### ON TARGET DATABASE
    SQL> select count(1) SEG_COUNT from dba_segments where tablespace_name='SOE';
      SEG_COUNT
            32
    SQL> ! date
    Wed Dec 19 15:08:11 GMT-02:00 2012
    SQL> drop tablespace soe including contents and datafiles;
    Tablespace dropped.
    ### CHECK TABLESPACE TO BE EXCLUDED FROM RESTORE
    SQL> select tablespace_name
        from dba_tablespaces
         where tablespace_name not in ('SYSTEM','SYSAUX','UNDOTBS1','SOE','XDB','TEMP');
    TABLESPACE_NAME
    USERS
    AUDIT_AUX
    ### CHECK BACKUP ON TARGET DATABASE
    RMAN> list backup completed between "sysdate-1" and "sysdate";
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    95      Full    986.56M    DISK        00:03:18     19-12-2012 14:41:59
            BP Key: 95   Status: AVAILABLE  Compressed: YES  Tag: TAG20121219T143840
            Piece Name: /fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp
      List of Datafiles in backup set 95
      File LV Type Ckp SCN    Ckp Time            Name
      1       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/system.349.799325249
      2       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/sysaux.348.799325249
      3       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/undotbs1.347.799325251
      4       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/users.335.799325251
      5       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/audit_aux.287.799330083
      6       Full 1675258    19-12-2012 14:38:42
      7       Full 1675258    19-12-2012 14:38:42 +DS8000_DG/prod/datafile/xdb.311.801410833
    BS Key  Size       Device Type Elapsed Time Completion Time
    96      1.16M      DISK        00:00:00     19-12-2012 14:42:08
            BP Key: 96   Status: AVAILABLE  Compressed: YES  Tag: TAG20121219T144208
            Piece Name: /fra/PROD/backupset/2012_12_19/o1_mf_annnn_TAG20121219T144208_8f3vm0k3_.bkp
      List of Archived Logs in backup set 96
      Thrd Seq     Low SCN    Low Time            Next SCN   Next Time
      1    8       1675198    19-12-2012 14:38:25 1675365    19-12-2012 14:42:07
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    97      Full    7.52M      DISK        00:00:02     19-12-2012 14:42:11
            BP Key: 97   Status: AVAILABLE  Compressed: NO  Tag: TAG20121219T144209
            Piece Name: /fra/PROD/autobackup/2012_12_19/o1_mf_s_802449729_8f3vm2cc_.bkp
      SPFILE Included: Modification time: 19-12-2012 14:36:30
      SPFILE db_unique_name: PROD
      Control File Included: Ckp SCN: 1675376      Ckp time: 19-12-2012 14:42:09
    RMAN>
    ### PREPARING ENV
    cp $ORACLE_HOME/dbs/orapwprod $ORACLE_HOME/dbs/orapwtspitr
    cd $ORACLE_HOME/dbs
    ## Creating INIT from TSPITR Instance
    vi inittspitr.ora
    DB_NAME=prod
    DB_UNIQUE_NAME=tspitr
    CONTROL_FILES=/stage_area_tspitr/PROD/controlfile/control01.ctl
    REMOTE_LOGIN_PASSWORDFILE=exclusive
    COMPATIBLE =11.2.0
    DB_BLOCK_SIZE=8192
    memory_target=700M
    streams_pool_size=100M
    ### creating stage area to deploy database
    mkdir -p  /fra/stage_area_tspitr/PROD
    cd /fra/stage_area_tspitr/PROD
    mkdir controlfile datafile onlinelog
    #### RESTORING DATABASE TSPITR
    export ORACLE_SID=tspitr
    $ rman target /
    RMAN> startup nomount
    Oracle instance started
    Total System Global Area     730714112 bytes
    Fixed Size                     2225088 bytes
    Variable Size                553651264 bytes
    Database Buffers             167772160 bytes
    Redo Buffers                   7065600 bytes
    RMAN> restore controlfile from '/fra/PROD/autobackup/2012_12_19/o1_mf_s_802449729_8f3vm2cc_.bkp';
    Starting restore at 19-12-2012 15:16:27
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=146 device type=DISK
    channel ORA_DISK_1: restoring control file
    channel ORA_DISK_1: restore complete, elapsed time: 00:00:03
    output file name=/fra/stage_area_tspitr/PROD/controlfile/control01.ctl
    Finished restore at 19-12-2012 15:16:31
    RMAN> startup mount
    database is already started
    database mounted
    released channel: ORA_DISK_1
    ### REPORT SCHEMA TO GET DATAFILE NUMBER TO BE RESTORED
    RMAN> report schema;
    RMAN-06139: WARNING: control file is not current for REPORT SCHEMA
    Report of database schema for database with db_unique_name TSPITR
    List of Permanent Datafiles
    ===========================
    File Size(MB) Tablespace           RB segs Datafile Name
    1    700      SYSTEM               ***     +DS8000_DG/prod/datafile/system.349.799325249
    2    600      SYSAUX               ***     +DS8000_DG/prod/datafile/sysaux.348.799325249
    3    200      UNDOTBS1             ***     +DS8000_DG/prod/datafile/undotbs1.347.799325251
    4    5        USERS                ***     +DS8000_DG/prod/datafile/users.335.799325251
    5    100      AUDIT_AUX            ***     +DS8000_DG/prod/datafile/audit_aux.287.799330083
    6    0        SOE                  ***     +DS8000_DG/tspitr/datafile/soe.368.802445265
    7    100      XDB                  ***     +DS8000_DG/prod/datafile/xdb.311.801410833
    List of Temporary Files
    =======================
    File Size(MB) Tablespace           Maxsize(MB) Tempfile Name
    1    385      TEMP                 32767       +DS8000_DG/prod/tempfile/temp.300.799325437
    RMAN>
    ### GET LOGFILE NAME to RENAME to new location.
    $ sqlplus / as sysdba
    SQL> select member from v$logfile;
    MEMBER
    +DS8000_DG/prod/onlinelog/group_1.330.802352807
    +DS8000_DG/prod/onlinelog/group_2.331.802352815
    +DS8000_DG/prod/onlinelog/group_3.321.802352821
    run {
    set newname for datafile 1 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 2 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 3 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 7 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
    set newname for datafile 6 to "+DS8000_DG";
    restore datafile 1,2,3,7,6 from tag 'TAG20121219T143840' ;
    switch datafile 1;
    switch datafile 2;
    switch datafile 3;
    switch datafile 7;
    switch datafile 6;
    switch tempfile all;
    set until time "to_date('19-12-2012 14:42:00','dd-mm-yyyy hh24:mi:ss')";
    recover database SKIP TABLESPACE  "USERS", "AUDIT_AUX" delete archivelog;
    ## Renaming ONLINELOG
    sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_1.330.802352807'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group1.rdo''";
    sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_2.331.802352815'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group2.rdo''";
    sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_3.321.802352821'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group3.rdo''";
    alter database open resetlogs;
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    Starting restore at 19-12-2012 15:32:13
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting datafile backup set restore
    channel ORA_DISK_1: specifying datafile(s) to restore from backup set
    channel ORA_DISK_1: restoring datafile 00001 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1
    channel ORA_DISK_1: restoring datafile 00002 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSAUX_FNO-2
    channel ORA_DISK_1: restoring datafile 00003 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-UNDOTBS1_FNO-3
    channel ORA_DISK_1: restoring datafile 00007 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-XDB_FNO-7
    channel ORA_DISK_1: restoring datafile 00006 to +DS8000_DG
    channel ORA_DISK_1: reading from backup piece /fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp
    channel ORA_DISK_1: piece handle=/fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp tag=TAG20121219T143840
    channel ORA_DISK_1: restored backup piece 1
    channel ORA_DISK_1: restore complete, elapsed time: 00:12:35
    Finished restore at 19-12-2012 15:44:48
    datafile 1 switched to datafile copy
    input datafile copy RECID=6 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1
    datafile 2 switched to datafile copy
    input datafile copy RECID=8 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSAUX_FNO-2
    datafile 3 switched to datafile copy
    input datafile copy RECID=10 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-UNDOTBS1_FNO-3
    datafile 7 switched to datafile copy
    input datafile copy RECID=12 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-XDB_FNO-7
    datafile 6 switched to datafile copy
    input datafile copy RECID=14 STAMP=802453489 file name=+DS8000_DG/tspitr/datafile/soe.304.802452733
    renamed tempfile 1 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-TEMP_FNO-1 in control file
    executing command: SET until clause
    Starting recover at 19-12-2012 15:44:49
    using channel ORA_DISK_1
    Executing: alter database datafile 4 offline
    Executing: alter database datafile 5 offline
    starting media recovery
    archived log for thread 1 with sequence 8 is already on disk as file +DS8000_DG/prod/onlinelog/group_2.331.802352815
    archived log file name=+DS8000_DG/prod/onlinelog/group_2.331.802352815 thread=1 sequence=8
    media recovery complete, elapsed time: 00:00:02
    Finished recover at 19-12-2012 15:44:55
    sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_1.330.802352807'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group1.rdo''
    sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_2.331.802352815'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group2.rdo''
    sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_3.321.802352821'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group3.rdo''
    database opened
    ## Exporting Tablespace Metadata
    RMAN >
    run {
    sql 'alter tablespace  SOE read only';
    # create directory for datapump export
    sql  "create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''";
    # grant on directory to SYSTEM
    sql "grant read,write on directory TSPITR_DIROBJ_DPDIR to system";
    # exporting metadata
    host "expdp system/oracle@localhost:1521/tspitr DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_FULL_CHECK=Y TRANSPORT_TABLESPACES=SOE";
    using target database control file instead of recovery catalog
    sql statement: alter tablespace  SOE read only
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''
    sql statement: grant read,write on directory TSPITR_DIROBJ_DPDIR to system
    Export: Release 11.2.0.3.0 - Production on Wed Dec 19 15:53:02 2012
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, Automatic Storage Management, Oracle Label Security, OLAP,
    Data Mining, Oracle Database Vault and Real Application Testing options
    Starting "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01":  system/********@localhost:1521/tspitr DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_FULL_CHECK=Y TRANSPORT_TABLESPACES=SOE
    Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
    Processing object type TRANSPORTABLE_EXPORT/TABLE
    Processing object type TRANSPORTABLE_EXPORT/INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK
    Master table "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01" successfully loaded/unloaded
    Dump file set for SYSTEM.SYS_EXPORT_TRANSPORTABLE_01 is:
      /fra/stage_area_tspitr/exp_transport_soe.dmp
    Datafiles required for transportable tablespace SOE:
      +DS8000_DG/tspitr/datafile/soe.304.802452733
    Job "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01" successfully completed at 15:54:28
    host command complete
    ## Stop Database and destroy it to make sure that Datafile 6 will not be used by this database;
    RMAN>
    run {
    shutdown immediate;
    ## removing controlfile and SYSTEM Tablespace to make sure that database will not be opened
    host "rm -f /fra/stage_area_tspitr/PROD/controlfile/control01.ctl /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1";
    database closed
    database dismounted
    Oracle instance shut down
    host command complete
    ## TARGET Database
    ## Plugging Tablespace on TARGET database.
    $ export ORACLE_SID=prod
    $ rman target /
    run {
    # create directory for datapump import
    sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''";
    # grant on directory to SYSTEM
    sql "grant read,write on directory TSPITR_DIROBJ_DPDIR to system";
    host "impdp system/oracle@localhost:1521/prod DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_DATAFILES=''+DS8000_DG/tspitr/datafile/soe.304.802452733''";
    sql "alter tablespace soe read write";
    sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''
    sql statement: grant read,write on directory TSPITR_DIROBJ_DPDIR to system
    Import: Release 11.2.0.3.0 - Production on Wed Dec 19 16:09:30 2012
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, Automatic Storage Management, Oracle Label Security, OLAP,
    Data Mining, Oracle Database Vault and Real Application Testing options
    Master table "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully loaded/unloaded
    Starting "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01":  system/********@localhost:1521/prod DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_DATAFILES=+DS8000_DG/tspitr/datafile/soe.304.802452733
    Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
    Processing object type TRANSPORTABLE_EXPORT/TABLE
    Processing object type TRANSPORTABLE_EXPORT/INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_INDEX/INDEX
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT
    Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
    Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK
    Job "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully completed at 16:10:03
    host command complete
    sql statement: alter tablespace soe read write
    ### Recompiling Invalid Objects and Checking objects from  tablespace SOE
    $ sqlplus / as sysdba
    SQL>  @$ORACLE_HOME/rdbms/admin/utlrp.sql
    TIMESTAMP
    COMP_TIMESTAMP UTLRP_BGN  2012-12-19 16:10:34
    OBJECTS WITH ERRORS
                      0
    ERRORS DURING RECOMPILATION
                              0
    Function created.
    PL/SQL procedure successfully completed.
    Function dropped.
    PL/SQL procedure successfully completed.
    SQL> select count(1) SEG_COUNT from dba_segments where tablespace_name='SOE';
      SEG_COUNT
            32
    SQL> SELECT COUNT(1) OBJ_INVALID FROM DBA_OBJECTS WHERE STATUS != 'VALID';
    OBJ_INVALID
              0
    ## PERFORM A FULL BACKUP AFTER RESTORE TABLESPACE.Regards,
    Levi Pereira

  • Cascading LOVs in discoverer desktop 10

    Hi
    Is it possible to create Cascading parameters in Discoverer Desktop 10?
    i know the possibility in discoverer plus but didn't find the "filter the lov based on the selected condition" radio button while creating the paramenter in Desktop10.
    Thanks

    Hi,
    You do not have to upgrade the desktop or administrator (you can if you like to).
    can i install Plus as client?No, the plus is installed on IAS or weblogic
    you can search this forum for links to download, and it will need installation and configuration.
    the same schema, reports, configuration will work on Plus?Yes it will as long as the configuration is right.
    Tamir

Maybe you are looking for