Schema reporting
Back when I worked with SQL/DS on a VM (mainframe) system, I was able to type a command and a nice little report of the entire schema was produced. Is there a way to do this with Oracle 9i via OEM? Thanx.
Most of the information displayed is available in data dictionary views:
user_objects
user_tables
user_views
user_indexes
user_tab_columns
user_ind_columns
user_constraints
user_cons_columns
user_tab_comments
etc...
It would be possible to use these to create a script to generate a report on the schema which could be very close to the one you posted (possibly not in a word format though).
Similar Messages
-
Opening SQL tab of other's user schema reports error
Hi,
I just try to open the SQL table tab to get the SQL script of a table, and it only works when the schema is owned by the user. If you try to open SQL tab on another user table it reports:
ORA-31603: object "COM_BANK" of type TABLE not found in schema "PROVIS_OWN"
ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
ORA-06512: at "SYS.DBMS_METADATA", line 653
ORA-06512: at "SYS.DBMS_METADATA", line 1260
ORA-06512: at line 1
I can access that table through a synonym and grants.
Server is Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit Production
I am using 2.1 of Oracle SQL Developer.
Thanks in advance,This is a reported issue, but I'm not sure it will be fixed in the upcoming patch. You can try searching the forum.
As workaround, you'd need more permissions (don't know which exactly; the DBA role works for me).
Regards,
K. -
I was looking for a report that would show the utilization of the authorization schemes, defined in an application. Where do I find or how do I create such a report?
Thanks,
Denes KubicekThanks Scott,
next time I will keep my eyes open while searching for that stuff. Your application - HTML db - is a good thing. Thanks for that as well.
Denes Kubicek -
Authentication Scheme Report "Bug"
Hi -
Since htmldb.oracle.com is down, I am working locally on my XE. Version 2.1.0.0.0.39. Stumbled across a strange bug I've never seen before - just wanted to report it if you dont all know about it before. Couldnt find it referenced in the search.
I imported an application successfully, but when I go to Shared Components => Authentication Themes, the report says there are not any. The default display type is Icons. However switching over to Details lists all of them. Its page 801 if you are keeping track at home.
I also tried creating a theme from scratch in a brand new application. Same behavior - I can only see them when selecting details instead of icons. I did not notice this problem on the hosted site, but its possible there are different versions of the software.
Any one else run across this?We are aware of that bug in the XE version but thanks for the details.
Scott -
How to Serial No Printing in Report - group Above Report
In Scott Schema , report data query - select rownum, emp, ename, job, deptno from emp;
Reprot type : group Above.
If I am not using group above, serial no like 1,2,3 is printing correctly. But if i use group above, serial no are not printing on order. it is printing like 5,4,3, 8.. li
How to print serially with group above.
Thanks in advance for the helpO.Developer wrote:
In Scott Schema , report data query - select rownum, emp, ename, job, deptno from emp;
Reprot type : group Above.
If I am not using group above, serial no like 1,2,3 is printing correctly. But if i use group above, serial no are not printing on order. it is printing like 5,4,3, 8.. li
How to print serially with group above.Avoid rownum and use summary column at detail block and set following property.
Summary function :count
summary column : ename
reset at : <group name>Your serial problem solved.
Hope this helps...
Hamid -
Report with dynamic order by (binding of parameter)
hy,
my problem is the following:
i want to let the user customize the ordering of a portal report.
but the binding does not really work, the bind variable seems to
have no effect in the order by clause.
code sample:
select * from SCOTT.EMP order by :p_order
then, portal gets the right value for the parameter:
:p_order binding with HIREDATE
but: this results in an emp table sorted by the empno.....
and not by the hiredate
do you have any ideas how to get this working ???
thanx, martin allhoffHi, the way to do this is to pass in a different set of
parameters. Try not using the :bind parameter in the report
query, but instead, call your report like this:
http:// ... /pls/portal30/Schema.report.show?
p_arg_names=_ORD_BY_COL&P_arg_values=HIREDATE
then try calling
http:// ... /pls/portal30/Schema.report.show?
p_arg_names=_ORD_BY_COL&P_arg_values=EMPNO
to see if indeed it is dynamically selecting the order by clause
based on the info passed in. Also, check my syntax, I cant
remember if ORDBY_COL is correct or not. But any possible
options should be listed for you in the "Display Call Interface"
screen for the report. -
XML Schema for Java Version 1.0
Hi everybody,
1.) in my opinion, the sample schema report.xsd seems to be incorrect with respect to the keyref definition (selector element), but the parser doesn't care anyway. I have modified the report.xml sample to include duplicates of the zip code and invalid keyrefs, but the parser still doesn't care. Despite the notice that unique, key and keyref doesn't work with SAX, it obviously doesn't work at all.
2.) The "types" contained within the Schema-Definition are not accessible from outside the "oracle" packages, and so far are rather useless. It would be an obvious advantage to access the type- and validation classes of the Schema processor in order to allow for interactive validation of user input, when documents are composed by gathering user input from a http-request. Will this change, as far as XMLSchema is becoming a standard?
TIA
AchimI'm glad someone else has noticed that unique keyref and key don't seem to be working with the Dom Parser. If they don't work WHY ARE THEY (key, keyref, unique) IN THE EXAMPLE'S THAT ARE DOWNLOADED WITH THE SCHEMA PARSER. In report.xsd, a file downloaded with the example, it makes clear usage of unique, key, and key ref. But if you violate the schema definitions in the corresponding file report.xml the parser doesn't complain whatsoever. The only time it barfs is if you change the keyref refer attribute to something other than "pNumKey". It obviously has to work. No bone head would send example files along with their product that didn't work.
So, if anyone at ORACLE or elsewhere has figured out how to use unique, key, or keyref please respond with an explination of how to correctly use them with the parser. Your name will be blessed throughout the ages as a most kind and venerable person. You will be a hallmark, a standard, a shining light for all future generations of what a human being should be! Okay maybe I'm going a little overboard but I'm DESPERATE. With no books or collateral on how this stupid thing works all I can do is hack.
-Thanks -
PDF printing--tomcat/cocoon installed, but report server not found------
As I understood, APEX 3.1 requires either BI/FO tools to enable PDF printing.
As I do not have BI license, and as I happened to check a very good link on a blog which had a detailed viewlet on apache-cocoon route for PDF print enabling, I downloaded the tomcat, cocoon source files, built the cocoon .WAR, put it in required place as instructed, started tomcot (ran into pronblems because of port conflits, resolved these by configuring different ports for 8005 and 8080 ports).
Now, when I want to configure the external report server, I am stuck.
I am not able to locate 'Instances' tab, or 'manage environment settings ' tab in 'manage services'
I am able to find only following tabs in 'manage services' under HOME.--> Administration ----> Manage services
1) session state
2) caching
3) workspace preferences
4) applications
5) workspace
6) workspace schema reports
7) Logs
Can anyone help me to find out where to configure the 'report server' now?
Thanks,
suresh.Sorry, it was a silly problem:) You have to be logged in as admin to 'see' the 'instances' tab. maybe in all the tomcat/cocoon build/config issues, this simply got off my head!
now, the problem of print server config is over, I have configured the same. the 'print' link is available now in report region, when I run the report. but when I click the 'print' link, it is taking ages...or rather it is still processing...!!
Can anyone please point out to me now what is the problem...?
thanks. -
Non OBIEE Publisher reports and dashboards
I have developed several Publisher reports that source data from direct SQL queries against an Oracle DB (They're not star schema reports that are appropriate for OBIEE). The reports are parameterized and direct SQL requests are used to populate the lists of values for the parameters. My task is to make them available on a dashboard.
I have a report attached to a dashboard but I cannot figure out how to prompt for the parameters in the dashboard. Everything I've read about dashboard prompts and presentation variables assumes the report is sourced by OBIEE.
Is there a way to do this?You have to create parameter in reports and as well create prompt in dashboard with same name.
With this link, the values will be passed into report parameter.
Use that parameter in report to filter -
Q: How to create reports with no login required?
I would like to make some reports available to internal people but I dont want/need login security. Is this possible? how?
Daryl,
Change the application's authentication scheme to DATABASE. You should see that scheme on the authentication scheme report, just click the "make current" link and confirm. Then the whole application will be public. To make selected pages public in your application, leave the authentication scheme as is and change the authentication attribute of those pages to "Page is Public" on the page attributes form.
Scott -
Authorization scheme (using {not} Scheme)
I have build a change password page and every user, except user with a Guest role (= GUEST SCHEME) have access to that page.
I defined a scheme GUEST for users with the GUEST role. When I define the page with Authorization scheme {not}GUEST this isn't working everyone has access to the page, also the guest users.
am I misunderstanding the {not}scheme choice or is something else wrong.
Fred.Fred,
I have solved it with the work around I mentioned before:I read what you said very carefully but thought it reckless to conclude that the workaround was successful because you just said "To work around the problem, I did xyz" without indicating the outcome.
The authorization schemes on navigation tabs fire also on the default login pageYes they do, they fire on every page whether or not the page template accommodates a navigation bar. This looks like a bug to me.
Is there a "authorization scheme report" which shows all the objects where the authorization scheme is defined.Shared Components > Authorization Schemes > Utilization (slightly different in each version).
Scott -
Workspace to parse SQL statements from multiple database schema's
Hi ,
When go through the following link of workspace admin user :
Home>HTML DB Workspace Administration>Manage Services>Schema Reports>Workspace Schemas there I saw "Your workspace has the privilege to parse SQL statements using the following database schemas. Note that the terms database schema and database user can be used interchangeably.
" at the right pane.
So how can i make my workspace to parse SQL statements from multiple database schema's ?That must be listed there .But i have only one schema in view ? Can i make it multiple ?If then How ?
Anybody have an idea ?
ROSYAssign as many schemas to your workspace as you want using the administration app. Read about it in the doc.
Scott -
Multiple MapTriggers allowed on the same cache?
I'm trying to register a pair of trigger listeners on the same cache (Coherence 3.5.1)
(See below for xml scheme reported in storage node log).
It appears as if only the first listener is active.
Both triggers work if configured in isolation.
We could partially solve the problem by creating a "super" trigger that encapsulates both functions, but this would prevent us from removing one of them at runtime by unregistering the listener.
In summary:
1) If multiple triggers are not supported, then shouldn't the config below be rejected?
2) If multiple triggers are supported, could anyone suggest what I'm doing wrong?
3) If multiple triggers are supported when added programatically, but not via config, is there a recommended way of managing them?
Thanks,
Philip
<distributed-scheme>
<scheme-name>dist-default</scheme-name>
<backing-map-scheme>
<local-scheme/>
</backing-map-scheme>
<autostart>true</autostart>
<serializer>
<class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
<init-params>
<init-param>
<param-type>string</param-type>
<param-value>java_pof_config.xml</param-value>
</init-param>
</init-params>
</serializer>
<thread-count>2</thread-count>
<listener>
<class-scheme>
<class-factory-name>com.psinventory.cache.EntryVersioningTrigger</class-factory-name>
<method-name>createMapTriggerListener</method-name>
</class-scheme>
</listener>
<listener>
<class-scheme>
<class-factory-name>com.psinventory.cache.EntryLoggingTrigger</class-factory-name>
<method-name>createMapTriggerListener</method-name>
</class-scheme>
</listener>
</distributed-scheme>Hi Philip
Multiple triggers do work if you add them using NamedCache.addMapListener(...). If you configure the trigger using the xml cache config file then only the first trigger is added. I will enter an internal bug for this issue.
Please be aware of 2 items:
1) The order in which the trigger is called is not specified in the API or any doc (that I know of). I did some testing and the last trigger added was the first called. However, you cannot rely on this behavior unless it is documented.
2) The output of the first trigger is passed as input to the second trigger. If you cannot rely on the ordering of the triggers then this could be a real problem.
Paul -
TSPITR - Recover a Dropped Tablespace aborting due dependency over XDB.
I found an issue using TSPITR to Recover a Dropped Tablespace when XML DB (XDB) is installed on a different tablespace from "SYSAUX".
During execution of TSPITR RMAN use Data Pump to export/import Tablespace Metadata as Data Pump has dependency over XDB views and related objects and XDB is installed on a set of tablespace not included on dummy database then EXPDP fails.
Errors raised ORA-39123, ORA-00376, ORA-01110.
I didn't found any limitation about XDB location on documentation or MoS.
This looks like a Oracle Bug.
See this:
XDB installed on TABLESPACE XDB.
SQL> select owner,tablespace_name
from dba_segments
where owner='XDB'
GROUP BY owner,tablespace_name ;
OWNER TABLESPACE_NAME
XDB XDBDropping Tablespace SOE.
SQL> select to_char(sysdate,'dd-mm-yyyy hh24:mi:ss') actual_date from dual;
ACTUAL_DATE
17-12-2012 11:57:44
SQL> drop tablespace SOE including contents and datafiles;
Tablespace dropped.Restoring Dropped Tablespace.
RMAN> list backup tag TAG20121217T112226;
List of Backup Sets
===================
BS Key Type LV Size Device Type Elapsed Time Completion Time
69 Full 995.61M DISK 00:02:41 17-12-2012 11:25:07
BP Key: 69 Status: AVAILABLE Compressed: YES Tag: TAG20121217T112226
Piece Name: +DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947
List of Datafiles in backup set 69
File LV Type Ckp SCN Ckp Time Name
1 Full 1628648 17-12-2012 11:22:26 +DS8000_DG/prod/datafile/system.349.799325249
2 Full 1628648 17-12-2012 11:22:26 +DS8000_DG/prod/datafile/sysaux.348.799325249
3 Full 1628648 17-12-2012 11:22:26 +DS8000_DG/prod/datafile/undotbs1.347.799325251
4 Full 1628648 17-12-2012 11:22:26 +DS8000_DG/prod/datafile/users.335.799325251
5 Full 1628648 17-12-2012 11:22:26 +DS8000_DG/prod/datafile/audit_aux.287.799330083
6 Full 1628648 17-12-2012 11:22:26
7 Full 1628648 17-12-2012 11:22:26 +DS8000_DG/prod/datafile/xdb.311.801410833
RMAN>
run{
SET NEWNAME FOR DATAFILE 6 TO "+DS8000_DG";
RECOVER TABLESPACE SOE
UNTIL TIME "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')"
AUXILIARY DESTINATION '/u01/app/oracle/flash_recovery_area01/stage_area_tspitr';
executing command: SET NEWNAME
Starting recover at 17-12-2012 12:01:29
allocated channel: ORA_DISK_1
channel ORA_DISK_1: SID=18 device type=DISK
RMAN-05026: WARNING: presuming following set of tablespaces applies to specified point-in-time
List of tablespaces expected to have UNDO segments
Tablespace SYSTEM
Tablespace UNDOTBS1
Creating automatic instance, with SID='FCox'
initialization parameters used for automatic instance:
db_name=PROD
db_unique_name=FCox_tspitr_PROD
compatible=11.2.0.0.0
db_block_size=8192
db_files=200
sga_target=280M
processes=50
db_create_file_dest=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr
log_archive_dest_1='location=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr'
#No auxiliary parameter file used
starting up automatic instance PROD
Oracle instance started
Total System Global Area 292278272 bytes
Fixed Size 2220880 bytes
Variable Size 109055152 bytes
Database Buffers 176160768 bytes
Redo Buffers 4841472 bytes
Automatic instance created
List of tablespaces that have been dropped from the target database:
Tablespace SOE
contents of Memory Script:
# set requested point in time
set until time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
# restore the controlfile
restore clone controlfile;
# mount the controlfile
sql clone 'alter database mount clone database';
# archive current online log
sql 'alter system archive log current';
# avoid unnecessary autobackups for structural changes during TSPITR
sql 'begin dbms_backup_restore.AutoBackupFlag(FALSE); end;';
executing Memory Script
executing command: SET until clause
Starting restore at 17-12-2012 12:01:46
allocated channel: ORA_AUX_DISK_1
channel ORA_AUX_DISK_1: SID=16 device type=DISK
channel ORA_AUX_DISK_1: starting datafile backup set restore
channel ORA_AUX_DISK_1: restoring control file
channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/autobackup/2012_12_17/s_802265113.276.802265115
channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/autobackup/2012_12_17/s_802265113.276.802265115 tag=TAG20121217T112513
channel ORA_AUX_DISK_1: restored backup piece 1
channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:00:05
output file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/controlfile/o1_mf_8dy9ghpz_.ctl
Finished restore at 17-12-2012 12:01:52
sql statement: alter database mount clone database
sql statement: alter system archive log current
sql statement: begin dbms_backup_restore.AutoBackupFlag(FALSE); end;
contents of Memory Script:
# set requested point in time
set until time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
# set destinations for recovery set and auxiliary set datafiles
set newname for clone datafile 1 to new;
set newname for clone datafile 3 to new;
set newname for clone datafile 2 to new;
set newname for clone tempfile 1 to new;
set newname for datafile 6 to
"+DS8000_DG";
# switch all tempfiles
switch clone tempfile all;
# restore the tablespaces in the recovery set and the auxiliary set
restore clone datafile 1, 3, 2, 6;
switch clone datafile all;
executing Memory Script
executing command: SET until clause
executing command: SET NEWNAME
executing command: SET NEWNAME
executing command: SET NEWNAME
executing command: SET NEWNAME
executing command: SET NEWNAME
renamed tempfile 1 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_temp_%u_.tmp in control file
Starting restore at 17-12-2012 12:01:58
using channel ORA_AUX_DISK_1
channel ORA_AUX_DISK_1: starting datafile backup set restore
channel ORA_AUX_DISK_1: specifying datafile(s) to restore from backup set
channel ORA_AUX_DISK_1: restoring datafile 00001 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_%u_.dbf
channel ORA_AUX_DISK_1: restoring datafile 00003 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_%u_.dbf
channel ORA_AUX_DISK_1: restoring datafile 00002 to /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_%u_.dbf
channel ORA_AUX_DISK_1: restoring datafile 00006 to +DS8000_DG
channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947
channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/backupset/2012_12_17/nnndf0_tag20121217t112226_0.286.802264947 tag=TAG20121217T112226
channel ORA_AUX_DISK_1: restored backup piece 1
channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:09:15
Finished restore at 17-12-2012 12:11:13
datafile 6 switched to datafile copy
input datafile copy RECID=5 STAMP=802267873 file name=+DS8000_DG/prod/datafile/soe.487.802267321
datafile 1 switched to datafile copy
input datafile copy RECID=6 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_8dy9gs09_.dbf
datafile 3 switched to datafile copy
input datafile copy RECID=7 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_8dy9ov17_.dbf
datafile 2 switched to datafile copy
input datafile copy RECID=8 STAMP=802267873 file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_8dy9hxsc_.dbf
contents of Memory Script:
# set requested point in time
set until time "to_date('17-12-2012 11:55:00','dd/mm/yyyy hh24:mi:ss')";
# online the datafiles restored or switched
sql clone "alter database datafile 1 online";
sql clone "alter database datafile 3 online";
sql clone "alter database datafile 2 online";
sql clone "alter database datafile 6 online";
# recover and open resetlogs
recover clone database tablespace "SOE", "SYSTEM", "UNDOTBS1", "SYSAUX" delete archivelog;
alter clone database open resetlogs;
executing Memory Script
executing command: SET until clause
sql statement: alter database datafile 1 online
sql statement: alter database datafile 3 online
sql statement: alter database datafile 2 online
sql statement: alter database datafile 6 online
Starting recover at 17-12-2012 12:11:14
using channel ORA_AUX_DISK_1
starting media recovery
archived log for thread 1 with sequence 15 is already on disk as file +DG_16M/prod/archivelog/2012_12_17/thread_1_seq_15.277.802267317
channel ORA_AUX_DISK_1: starting archived log restore to default destination
channel ORA_AUX_DISK_1: restoring archived log
archived log thread=1 sequence=14
channel ORA_AUX_DISK_1: reading from backup piece +DG_16M/prod/backupset/2012_12_17/annnf0_tag20121217t112512_0.264.802265113
channel ORA_AUX_DISK_1: piece handle=+DG_16M/prod/backupset/2012_12_17/annnf0_tag20121217t112512_0.264.802265113 tag=TAG20121217T112512
channel ORA_AUX_DISK_1: restored backup piece 1
channel ORA_AUX_DISK_1: restore complete, elapsed time: 00:00:01
archived log file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/1_14_801941068.dbf thread=1 sequence=14
channel clone_default: deleting archived log(s)
archived log file name=/u01/app/oracle/flash_recovery_area01/stage_area_tspitr/1_14_801941068.dbf RECID=141 STAMP=802267877
archived log file name=+DG_16M/prod/archivelog/2012_12_17/thread_1_seq_15.277.802267317 thread=1 sequence=15
media recovery complete, elapsed time: 00:00:01
Finished recover at 17-12-2012 12:11:19
database opened
contents of Memory Script:
# make read only the tablespace that will be exported
sql clone 'alter tablespace SOE read only';
# create directory for datapump import
sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''
/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''";
# create directory for datapump export
sql clone "create or replace directory TSPITR_DIROBJ_DPDIR as ''
/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''";
executing Memory Script
sql statement: alter tablespace SOE read only
sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''
sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/u01/app/oracle/flash_recovery_area01/stage_area_tspitr''
Performing export of metadata...
EXPDP> Starting "SYS"."TSPITR_EXP_FCox":
EXPDP> ORA-39123: Data Pump transportable tablespace job aborted
ORA-00376: file 7 cannot be read at this time
ORA-01110: data file 7: '+DS8000_DG/prod/datafile/xdb.311.801410833'
Removing automatic instance
shutting down automatic instance
database closed
database dismounted
Oracle instance shut down
Automatic instance removed
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_temp_8dyb0qok_.tmp deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_3_8dyb0kfp_.log deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_2_8dyb0fm9_.log deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/onlinelog/o1_mf_1_8dyb07xv_.log deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_sysaux_8dy9hxsc_.dbf deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_undotbs1_8dy9ov17_.dbf deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/datafile/o1_mf_system_8dy9gs09_.dbf deleted
auxiliary instance file /u01/app/oracle/flash_recovery_area01/stage_area_tspitr/PROD/controlfile/o1_mf_8dy9ghpz_.ctl deleted
RMAN-00571: ===========================================================
RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
RMAN-00571: ===========================================================
RMAN-03002: failure of recover command at 12/17/2012 12:12:19
RMAN-06962: Error received during export of metadata
RMAN-06965: Datapump job has stopped
RMAN-06960: EXPDP> Job "SYS"."TSPITR_EXP_FCox" stopped due to fatal error at 12:12:02
RMAN>If I drop tablespace SOE and XDB on Target database and try to restore both (SOE,XDB) export will succeeds but during IMPORT Metadata RMAN will faill due XDB dependency.
In my view, they (mean Oracle) need find way to include XDB Tablespace at DUMMY database only to export (EXPDP) Metadata purposes.
If someone get a workaround for this error will be grateful.
P.S: This is not a Production environment and will soon open a SR on Oracle Support.Hi Michael,
I did not find solution to use TSPITR to recover dropped tablespace if XML DB is installed in a non-default tablespace (i.e outside of SYSAUX).
As far I could understand, the only way to perform this RESTORE is using the old fashion. By creating clone database excluding all unnecessary tablespace, exporting and importing metadata from desired tablespace manually.
Since Oracle has not solved the issue I created my own workaround.
@All
If someone is able to use TSPITR to restore dropped tablespace when XDB is installed in a non-default tablespace, please let me know!!!
The procedure below is supported on All Oracle Version since 10.1 until 11.2.
I DON'T RECOMMEND to create AUXILIARY (clone) database using the same server from TARGET because you can damage the TARGET database.
On this example I'm using same server.
See this example:
### ON TARGET DATABASE
SQL> select count(1) SEG_COUNT from dba_segments where tablespace_name='SOE';
SEG_COUNT
32
SQL> ! date
Wed Dec 19 15:08:11 GMT-02:00 2012
SQL> drop tablespace soe including contents and datafiles;
Tablespace dropped.
### CHECK TABLESPACE TO BE EXCLUDED FROM RESTORE
SQL> select tablespace_name
from dba_tablespaces
where tablespace_name not in ('SYSTEM','SYSAUX','UNDOTBS1','SOE','XDB','TEMP');
TABLESPACE_NAME
USERS
AUDIT_AUX
### CHECK BACKUP ON TARGET DATABASE
RMAN> list backup completed between "sysdate-1" and "sysdate";
List of Backup Sets
===================
BS Key Type LV Size Device Type Elapsed Time Completion Time
95 Full 986.56M DISK 00:03:18 19-12-2012 14:41:59
BP Key: 95 Status: AVAILABLE Compressed: YES Tag: TAG20121219T143840
Piece Name: /fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp
List of Datafiles in backup set 95
File LV Type Ckp SCN Ckp Time Name
1 Full 1675258 19-12-2012 14:38:42 +DS8000_DG/prod/datafile/system.349.799325249
2 Full 1675258 19-12-2012 14:38:42 +DS8000_DG/prod/datafile/sysaux.348.799325249
3 Full 1675258 19-12-2012 14:38:42 +DS8000_DG/prod/datafile/undotbs1.347.799325251
4 Full 1675258 19-12-2012 14:38:42 +DS8000_DG/prod/datafile/users.335.799325251
5 Full 1675258 19-12-2012 14:38:42 +DS8000_DG/prod/datafile/audit_aux.287.799330083
6 Full 1675258 19-12-2012 14:38:42
7 Full 1675258 19-12-2012 14:38:42 +DS8000_DG/prod/datafile/xdb.311.801410833
BS Key Size Device Type Elapsed Time Completion Time
96 1.16M DISK 00:00:00 19-12-2012 14:42:08
BP Key: 96 Status: AVAILABLE Compressed: YES Tag: TAG20121219T144208
Piece Name: /fra/PROD/backupset/2012_12_19/o1_mf_annnn_TAG20121219T144208_8f3vm0k3_.bkp
List of Archived Logs in backup set 96
Thrd Seq Low SCN Low Time Next SCN Next Time
1 8 1675198 19-12-2012 14:38:25 1675365 19-12-2012 14:42:07
BS Key Type LV Size Device Type Elapsed Time Completion Time
97 Full 7.52M DISK 00:00:02 19-12-2012 14:42:11
BP Key: 97 Status: AVAILABLE Compressed: NO Tag: TAG20121219T144209
Piece Name: /fra/PROD/autobackup/2012_12_19/o1_mf_s_802449729_8f3vm2cc_.bkp
SPFILE Included: Modification time: 19-12-2012 14:36:30
SPFILE db_unique_name: PROD
Control File Included: Ckp SCN: 1675376 Ckp time: 19-12-2012 14:42:09
RMAN>
### PREPARING ENV
cp $ORACLE_HOME/dbs/orapwprod $ORACLE_HOME/dbs/orapwtspitr
cd $ORACLE_HOME/dbs
## Creating INIT from TSPITR Instance
vi inittspitr.ora
DB_NAME=prod
DB_UNIQUE_NAME=tspitr
CONTROL_FILES=/stage_area_tspitr/PROD/controlfile/control01.ctl
REMOTE_LOGIN_PASSWORDFILE=exclusive
COMPATIBLE =11.2.0
DB_BLOCK_SIZE=8192
memory_target=700M
streams_pool_size=100M
### creating stage area to deploy database
mkdir -p /fra/stage_area_tspitr/PROD
cd /fra/stage_area_tspitr/PROD
mkdir controlfile datafile onlinelog
#### RESTORING DATABASE TSPITR
export ORACLE_SID=tspitr
$ rman target /
RMAN> startup nomount
Oracle instance started
Total System Global Area 730714112 bytes
Fixed Size 2225088 bytes
Variable Size 553651264 bytes
Database Buffers 167772160 bytes
Redo Buffers 7065600 bytes
RMAN> restore controlfile from '/fra/PROD/autobackup/2012_12_19/o1_mf_s_802449729_8f3vm2cc_.bkp';
Starting restore at 19-12-2012 15:16:27
using target database control file instead of recovery catalog
allocated channel: ORA_DISK_1
channel ORA_DISK_1: SID=146 device type=DISK
channel ORA_DISK_1: restoring control file
channel ORA_DISK_1: restore complete, elapsed time: 00:00:03
output file name=/fra/stage_area_tspitr/PROD/controlfile/control01.ctl
Finished restore at 19-12-2012 15:16:31
RMAN> startup mount
database is already started
database mounted
released channel: ORA_DISK_1
### REPORT SCHEMA TO GET DATAFILE NUMBER TO BE RESTORED
RMAN> report schema;
RMAN-06139: WARNING: control file is not current for REPORT SCHEMA
Report of database schema for database with db_unique_name TSPITR
List of Permanent Datafiles
===========================
File Size(MB) Tablespace RB segs Datafile Name
1 700 SYSTEM *** +DS8000_DG/prod/datafile/system.349.799325249
2 600 SYSAUX *** +DS8000_DG/prod/datafile/sysaux.348.799325249
3 200 UNDOTBS1 *** +DS8000_DG/prod/datafile/undotbs1.347.799325251
4 5 USERS *** +DS8000_DG/prod/datafile/users.335.799325251
5 100 AUDIT_AUX *** +DS8000_DG/prod/datafile/audit_aux.287.799330083
6 0 SOE *** +DS8000_DG/tspitr/datafile/soe.368.802445265
7 100 XDB *** +DS8000_DG/prod/datafile/xdb.311.801410833
List of Temporary Files
=======================
File Size(MB) Tablespace Maxsize(MB) Tempfile Name
1 385 TEMP 32767 +DS8000_DG/prod/tempfile/temp.300.799325437
RMAN>
### GET LOGFILE NAME to RENAME to new location.
$ sqlplus / as sysdba
SQL> select member from v$logfile;
MEMBER
+DS8000_DG/prod/onlinelog/group_1.330.802352807
+DS8000_DG/prod/onlinelog/group_2.331.802352815
+DS8000_DG/prod/onlinelog/group_3.321.802352821
run {
set newname for datafile 1 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
set newname for datafile 2 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
set newname for datafile 3 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
set newname for datafile 7 to '/fra/stage_area_tspitr/PROD/datafile/datafiles_%U';
set newname for datafile 6 to "+DS8000_DG";
restore datafile 1,2,3,7,6 from tag 'TAG20121219T143840' ;
switch datafile 1;
switch datafile 2;
switch datafile 3;
switch datafile 7;
switch datafile 6;
switch tempfile all;
set until time "to_date('19-12-2012 14:42:00','dd-mm-yyyy hh24:mi:ss')";
recover database SKIP TABLESPACE "USERS", "AUDIT_AUX" delete archivelog;
## Renaming ONLINELOG
sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_1.330.802352807'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group1.rdo''";
sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_2.331.802352815'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group2.rdo''";
sql "alter database rename file ''+DS8000_DG/prod/onlinelog/group_3.321.802352821'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group3.rdo''";
alter database open resetlogs;
executing command: SET NEWNAME
executing command: SET NEWNAME
Starting restore at 19-12-2012 15:32:13
using channel ORA_DISK_1
channel ORA_DISK_1: starting datafile backup set restore
channel ORA_DISK_1: specifying datafile(s) to restore from backup set
channel ORA_DISK_1: restoring datafile 00001 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1
channel ORA_DISK_1: restoring datafile 00002 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSAUX_FNO-2
channel ORA_DISK_1: restoring datafile 00003 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-UNDOTBS1_FNO-3
channel ORA_DISK_1: restoring datafile 00007 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-XDB_FNO-7
channel ORA_DISK_1: restoring datafile 00006 to +DS8000_DG
channel ORA_DISK_1: reading from backup piece /fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp
channel ORA_DISK_1: piece handle=/fra/PROD/backupset/2012_12_19/o1_mf_nnndf_TAG20121219T143840_8f3vdn3q_.bkp tag=TAG20121219T143840
channel ORA_DISK_1: restored backup piece 1
channel ORA_DISK_1: restore complete, elapsed time: 00:12:35
Finished restore at 19-12-2012 15:44:48
datafile 1 switched to datafile copy
input datafile copy RECID=6 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1
datafile 2 switched to datafile copy
input datafile copy RECID=8 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSAUX_FNO-2
datafile 3 switched to datafile copy
input datafile copy RECID=10 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-UNDOTBS1_FNO-3
datafile 7 switched to datafile copy
input datafile copy RECID=12 STAMP=802453489 file name=/fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-XDB_FNO-7
datafile 6 switched to datafile copy
input datafile copy RECID=14 STAMP=802453489 file name=+DS8000_DG/tspitr/datafile/soe.304.802452733
renamed tempfile 1 to /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-TEMP_FNO-1 in control file
executing command: SET until clause
Starting recover at 19-12-2012 15:44:49
using channel ORA_DISK_1
Executing: alter database datafile 4 offline
Executing: alter database datafile 5 offline
starting media recovery
archived log for thread 1 with sequence 8 is already on disk as file +DS8000_DG/prod/onlinelog/group_2.331.802352815
archived log file name=+DS8000_DG/prod/onlinelog/group_2.331.802352815 thread=1 sequence=8
media recovery complete, elapsed time: 00:00:02
Finished recover at 19-12-2012 15:44:55
sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_1.330.802352807'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group1.rdo''
sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_2.331.802352815'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group2.rdo''
sql statement: alter database rename file ''+DS8000_DG/prod/onlinelog/group_3.321.802352821'' to ''/fra/stage_area_tspitr/PROD/onlinelog/group3.rdo''
database opened
## Exporting Tablespace Metadata
RMAN >
run {
sql 'alter tablespace SOE read only';
# create directory for datapump export
sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''";
# grant on directory to SYSTEM
sql "grant read,write on directory TSPITR_DIROBJ_DPDIR to system";
# exporting metadata
host "expdp system/oracle@localhost:1521/tspitr DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_FULL_CHECK=Y TRANSPORT_TABLESPACES=SOE";
using target database control file instead of recovery catalog
sql statement: alter tablespace SOE read only
sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''
sql statement: grant read,write on directory TSPITR_DIROBJ_DPDIR to system
Export: Release 11.2.0.3.0 - Production on Wed Dec 19 15:53:02 2012
Copyright (c) 1982, 2011, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
With the Partitioning, Automatic Storage Management, Oracle Label Security, OLAP,
Data Mining, Oracle Database Vault and Real Application Testing options
Starting "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01": system/********@localhost:1521/tspitr DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_FULL_CHECK=Y TRANSPORT_TABLESPACES=SOE
Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
Processing object type TRANSPORTABLE_EXPORT/TABLE
Processing object type TRANSPORTABLE_EXPORT/INDEX/INDEX
Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_INDEX/INDEX
Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT
Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS
Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT
Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK
Master table "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01" successfully loaded/unloaded
Dump file set for SYSTEM.SYS_EXPORT_TRANSPORTABLE_01 is:
/fra/stage_area_tspitr/exp_transport_soe.dmp
Datafiles required for transportable tablespace SOE:
+DS8000_DG/tspitr/datafile/soe.304.802452733
Job "SYSTEM"."SYS_EXPORT_TRANSPORTABLE_01" successfully completed at 15:54:28
host command complete
## Stop Database and destroy it to make sure that Datafile 6 will not be used by this database;
RMAN>
run {
shutdown immediate;
## removing controlfile and SYSTEM Tablespace to make sure that database will not be opened
host "rm -f /fra/stage_area_tspitr/PROD/controlfile/control01.ctl /fra/stage_area_tspitr/PROD/datafile/datafiles_data_D-PROD_TS-SYSTEM_FNO-1";
database closed
database dismounted
Oracle instance shut down
host command complete
## TARGET Database
## Plugging Tablespace on TARGET database.
$ export ORACLE_SID=prod
$ rman target /
run {
# create directory for datapump import
sql "create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''";
# grant on directory to SYSTEM
sql "grant read,write on directory TSPITR_DIROBJ_DPDIR to system";
host "impdp system/oracle@localhost:1521/prod DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_DATAFILES=''+DS8000_DG/tspitr/datafile/soe.304.802452733''";
sql "alter tablespace soe read write";
sql statement: create or replace directory TSPITR_DIROBJ_DPDIR as ''/fra/stage_area_tspitr''
sql statement: grant read,write on directory TSPITR_DIROBJ_DPDIR to system
Import: Release 11.2.0.3.0 - Production on Wed Dec 19 16:09:30 2012
Copyright (c) 1982, 2011, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
With the Partitioning, Automatic Storage Management, Oracle Label Security, OLAP,
Data Mining, Oracle Database Vault and Real Application Testing options
Master table "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully loaded/unloaded
Starting "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01": system/********@localhost:1521/prod DIRECTORY=TSPITR_DIROBJ_DPDIR DUMPFILE=exp_transport_soe.dmp TRANSPORT_DATAFILES=+DS8000_DG/tspitr/datafile/soe.304.802452733
Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
Processing object type TRANSPORTABLE_EXPORT/TABLE
Processing object type TRANSPORTABLE_EXPORT/INDEX/INDEX
Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_INDEX/INDEX
Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT
Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS
Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT
Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS
Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK
Job "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully completed at 16:10:03
host command complete
sql statement: alter tablespace soe read write
### Recompiling Invalid Objects and Checking objects from tablespace SOE
$ sqlplus / as sysdba
SQL> @$ORACLE_HOME/rdbms/admin/utlrp.sql
TIMESTAMP
COMP_TIMESTAMP UTLRP_BGN 2012-12-19 16:10:34
OBJECTS WITH ERRORS
0
ERRORS DURING RECOMPILATION
0
Function created.
PL/SQL procedure successfully completed.
Function dropped.
PL/SQL procedure successfully completed.
SQL> select count(1) SEG_COUNT from dba_segments where tablespace_name='SOE';
SEG_COUNT
32
SQL> SELECT COUNT(1) OBJ_INVALID FROM DBA_OBJECTS WHERE STATUS != 'VALID';
OBJ_INVALID
0
## PERFORM A FULL BACKUP AFTER RESTORE TABLESPACE.Regards,
Levi Pereira -
Cascading LOVs in discoverer desktop 10
Hi
Is it possible to create Cascading parameters in Discoverer Desktop 10?
i know the possibility in discoverer plus but didn't find the "filter the lov based on the selected condition" radio button while creating the paramenter in Desktop10.
ThanksHi,
You do not have to upgrade the desktop or administrator (you can if you like to).
can i install Plus as client?No, the plus is installed on IAS or weblogic
you can search this forum for links to download, and it will need installation and configuration.
the same schema, reports, configuration will work on Plus?Yes it will as long as the configuration is right.
Tamir
Maybe you are looking for
-
Can I put songs in my library onto another library on the same computer???
I have a song on my library, and someone in my family wants it on his library ON THE SAME COMPUTER. Is there a way to put it on his library without him having to buy it for himself??? Please help me. I don't want to have to buy one for his library.
-
Calling Web Service with Http Basic authentication in SOA 11g
I am calling a webservice which has http basic authentication attached to it. Thus i am adding 'oracle/wss_http_token_client_policy' OWSM policy to the WS refrence in my composite in Jdeveloper,but it doesn't showme the option of providing the http U
-
How to boost wifi extreme signal
My brother has a 2007 model Airport Extreme that I got for him back then. He's updated the firmware last autumn - and when he's in the same room of his long single-story home the signal is excellent for things like SKYPE audio, but only fair for SKY
-
Scroll text in layers stack...
I got problem with scrolling text in layers stack. Is there some solution?
-
Can someone help me with this issue. I can't install disk and restart my Mac it indicates verify permissions for Mac OSx install disc 1, the red error message states the underlying task reported failure on exit. I cannot start up my Mac and I am conc