Mount source datafile on target database

I have 1 user which store all db objects on 1 tablespace (1 file). Can I just copy this file to target database and mount it then in some magic way? Is this really possible to show Oracle additional datafile and configure db to read from it?

Hi Tutus,
tutus wrote:
I have 1 user which store all db objects on 1 tablespace (1 file). Can I just copy this file to target database and mount it then in some magic way? Is this really possible to show Oracle additional datafile and configure db to read from it?Well, although what you have just described does not work like this, it is not too far from the reality. Oracle call it transportable tablespaces.
For more info check the following: http://download.oracle.com/docs/cd/E11882_01/server.112/e17120/tspaces013.htm#i1007169
HTH.
Regards,
Jozsef

Similar Messages

  • Database Project Schema Compare Source Control to target database Not working

    I am trying to use the schema compare feature of database project / TFS.  I don't understand why it is trying to DELETE a stored procedure that exists in TFS.
    The Degenerate Dimension

    Hi MM,  
    Thanks for  your post.
    If you open this solution from a local folder(not stored in TFS), and try to compare that same file, what’s the result? We need confirm this issue relate to VS Schema Compare tool or relate to TFS Source Control first.
    http://blogs.msdn.com/b/ssdt/archive/2012/03/23/schema-compare-improvements.aspx
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Mapping 2 source structures to a single target database

    Hi Experts, My scenario is a Proxy to JDBC in which I need to send data from 2 source structures to 1 target table.
    Source structures :                                           Target:
    ABC                                                                 XYZ
      row1                                                                 InsertStatement
        Item1                                                                  DBTable
                                                                        Action
       (Fields)                                                     Table
                                                                                    acess
    DEF                                                                              -
      row2                                                                          -
       item2                                                                         -
      (Fields)
    (Fields)
    Please let me know how to map the 2 source nodes to the target access node to transfer data from the source tables to target database.
    Regards,
    Krishna

    You should make use of BPM to collect the two source structures.
    One of the BPM examples mentioned in IR --> SAP BASIS ---> SystemPatterns can be referred for this purpose.
    Once the messages are collected, perform a 2:1 mapping (two Proxies as source message and one JDBC message as target).
    If you want the mapping logic then please provide a proper format of both your Source and Target messages and the expected mapping, so that someone from SDN can help you out.
    The target structure provided is a bit confusing.
    Are you going to receive two different proxy messages (i.e. two different calls) or just one Proxy call and there you will have two different nodes within the same message??? ...... me confused
    Regards,
    Abhishek.
    Edited by: abhishek salvi on Sep 25, 2009 11:51 AM

  • Duplicate target database for standby failed

    Hi,
    Im trying to uplicate target database for standby and get error message bellow :
    RMAN> DUPLICATE TARGET DATABASE
    2>   FOR STANDBY
    3>   FROM ACTIVE DATABASE
    4>   DORECOVER
    5>   SPFILE
    6>     SET db_unique_name='dgtst' COMMENT 'Is standby'
    7>     SET LOG_ARCHIVE_DEST_2='SERVICE=dgtst ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=dgtst'
    8>     SET FAL_SERVER='dgtst' COMMENT 'Is primary'
    9>   NOFILENAMECHECK;
    Starting Duplicate Db at 24-FEB-12
    using target database control file instead of recovery catalog
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 02/24/2012 22:51:21
    RMAN-05501: aborting duplication of target database
    RMAN-03009: failure of allocate command on ORA_AUX_SBT_TAPE_1 channel at 02/24/2012 22:51:21
    ORA-19554: error allocating device, device type: SBT_TAPE, device name:
    ORA-27211: Failed to load Media Management Library
    Additional information: 2Please advice
    Thanks

    Hello mseberg,
    Actually you find the problem .....
    Adding the following softlink helped :
    ln -s /opt/omni/lib/libob2oracle8_64bit.so /u01/app/oracle/product/11.2.0/DG11gR2TST/lib/libobk.so
    I run the duplicate command and its look much better:
    $ rman TARGET sys/mysyspass@dgtst AUXILIARY sys/mysyspass@drtst
    Recovery Manager: Release 11.2.0.2.0 - Production on Sat Feb 25 07:43:15 2012
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    connected to target database: DGTST (DBID=244743902)    <<<=== connected
    connected to auxiliary database: DGTST (not mounted)    <<<=== connected
    RMAN> DUPLICATE TARGET DATABASE
    2> FOR STANDBY
    3> FROM ACTIVE DATABASE
    4> DORECOVER
    5> SPFILE
    6> SET db_unique_name='dgtst' COMMENT 'Is standby'
    SET LOG_ARCHIVE_DEST_2='SERVICE=dgtst ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=dgtst'
    7> 8> SET FAL_SERVER='dgtst' COMMENT 'Is primary'
    9> NOFILENAMECHECK;
    Starting Duplicate Db at 25-FEB-12
    using target database control file instead of recovery catalog
    allocated channel: ORA_AUX_SBT_TAPE_1
    channel ORA_AUX_SBT_TAPE_1: SID=156 device type=SBT_TAPE
    channel ORA_AUX_SBT_TAPE_1: Data Protector A.06.11/PHSS_40470/PHSS_40471/DPSOL_00391/DPLNX_
    allocated channel: ORA_AUX_SBT_TAPE_2
    channel ORA_AUX_SBT_TAPE_2: SID=189 device type=SBT_TAPE
    channel ORA_AUX_SBT_TAPE_2: Data Protector A.06.11/PHSS_40470/PHSS_40471/DPSOL_00391/DPLNX_
    allocated channel: ORA_AUX_SBT_TAPE_16
    channel ORA_AUX_SBT_TAPE_16: SID=129 device type=SBT_TAPE
    channel ORA_AUX_SBT_TAPE_16: Data Protector A.06.11/PHSS_40470/PHSS_40471/DPSOL_00391/DPLNX_
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=160 device type=DISK
    contents of Memory Script:
       backup as copy reuse
       targetfile  '/u01/app/oracle/product/11.2.0/DG11gR2TST/dbs/orapwdgtst' auxiliary format
    '/u01/app/oracle/product/11.2.0/DG11gR2TST/dbs/orapwdgtst'   targetfile
    '/u01/app/oracle/product/11.2.0/DG11gR2TST/dbs/spfiledgtst.ora' auxiliary format
    '/u01/app/oracle/product/11.2.0/DG11gR2TST/dbs/spfiledgtst.ora'   ;
       sql clone "alter system set spfile= ''/u01/app/oracle/product/11.2.0/DG11gR2TST/dbs/spfiledgtst.ora''";
    executing Memory Script
    Starting backup at 25-FEB-12
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=21 device type=DISK
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 02/25/2012 07:44:02
    ORA-17629: Cannot connect to the remote database server  <<<<<<====
    ORA-17627: ORA-12541: TNS:no listener                    <<<<<<====
    ORA-17629: Cannot connect to the remote database server
    continuing other job steps, job failed will not be re-run
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 02/25/2012 07:44:03
    RMAN-05501: aborting duplication of target database
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 02/25/2012 07:44:03
    ORA-17629: Cannot connect to the remote database server
    ORA-17627: ORA-12541: TNS:no listener
    ORA-17629: Cannot connect to the remote database serverAs you can see above rman successfully connected to the local and remote database.
    It also successfully allocated the DISK channel , but than it failed on :
    ORA-17627: ORA-12541: TNS:no listenerI cant understand why it happen:
    $ tnsping dgtst
    TNS Ping Utility for HPUX: Version 11.2.0.2.0 - Production on 25-FEB-2012 08:54:48
    Copyright (c) 1997, 2010, Oracle.  All rights reserved.
    Used parameter files:
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = hercules)(PORT = 1528))) (CONNECT_DATA = (SERVICE_NAME = dgtst)))
    OK (0 msec)
    $ tnsping drtst
    TNS Ping Utility for HPUX: Version 11.2.0.2.0 - Production on 25-FEB-2012 08:54:54
    Copyright (c) 1997, 2010, Oracle.  All rights reserved.
    Used parameter files:
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = pulse)(PORT = 1528))) (CONNECT_DATA = (SERVICE_NAME = dgtst)))
    OK (10 msec)And as you i show the rman successfully connect to both rempte and local as SYS user.

  • Restore single datafile from source database to target database.

    Here's my issue:
    Database Release : 11.2.0.3 across both the source and targets. (ARCHIVELOG mode)
    O/S: RHEL 5 (Tikanga)
    Database Storage: Using ASM on a stand-alone server (NOT RAC)
    Using Oracle GG to replicate changes on the Source to the Targets.
    My scenario:
    We utilize sequences to keep the primary key in tact and these are replicated utilizing GG. All of my schema tables are located in one tablespace and datafile and all of my indexes are in seperate tablespace (nothing is being partitioned).
    In the event of media failure on the Target or my target schema being completely out of whack, is there a method where I can copy the datafile/tablespace from my source (which is intact) to my target?
    I know there are possibilites of
    1) restore/recover the tablespace to a SCN or timestamp in the past and then I could use GoldenGate to run the transactions in (but this could take time depending on how far back I need to recover the tablespace and how many transactions have processed with GG) (This is not fool-proof).
    2) Could use DataPump to move the data from the Source schema to the Target schema (but the sequences are usually out of order if they haven't fired on the source, get that 'sequence is defined for this session message'). I've tried this scenario.
    3) I could alter the sequences to get them to proper number using the start and increment by feature (again this could take time depending on how many sequences are out of order).
    I would think you could
    1) back up the datafile/tablespace on the source,
    2)then copy the datafile to the target.
    3) startup mount;
    4) Newname the new file copied from the source (this is ASM)
    5) Restore the datafile/tablespace
    6) Recover the datafile/tablespace
    7) alter database open;
    Question 1: Do I need to also copy the backup piece from the source when I execute the backup tablespace on the source as indicated in my step 1?
    Question 2: Do I need to include "plus archivelog" when I execute the backup tablespace on the source as indicated in my step 1?
    Question 3: Do I need to execute an 'alter system switch logfile' on the Target when the recover in step 6 is completed?
    My scenario sounds like a Cold Backup but running with Archivelog mode, so the source could be online while the database is running.
    Just looking for alternate methods of recovery.
    Thanks,
    Jason

    Let me take another stab at sticking a fork into this myth about separating tables and indexes.
    Let's assume you have a production Oracle database environment with multiple users making multiple requests and the exact same time. This assumption mirrors reality everywhere except in a classroom where a student is running a simple demo.
    Let's further assume that the system looks anything like a real Oracle database system where the operating system has caching, the SAN has caching, and the blocks you are trying to read are split between memory and disk.
    Now you want to do some simple piece of work and assume there is an index on the ename column...
    SELECT * FROM emp WHERE ename = 'KING';The myth is that Oracle is going to, in parallel, read the index and read the table segments better, faster, whatever, if they are in separate physical files mapped by separate logical tablespaces somehow to separate physical spindles.
    Apply some synapses to this myth and it falls apart.
    You issue your SQL statement and Oracle does what? It looks for those index blocks where? In memory. If it finds them it never goes to disk. If it does not it goes to disk.
    While all this is happening the hundreds or thousands of other users on the database are also making requests. Oracle is not going to stop doing work while it tries to find your index blocks.
    Now it finds the index block and decides to use the ROWID value to read the block containing the row with KING's data. Did it freeze the system? Did it lock out everyone else while it did this? Of course not. It puts your read request into the queue and, again, first checks memory to see if it needs to go to disk.
    Where in here is there anything that indicates an advantage to having separate physical files?
    And even if there was some theoretical reason why separate files might be better ... are they separate in the SAN's cache? No. Are they definitely located on separate stripes or separate physical disks? Of course not.
    Oracle uses logical mappings (tables and tablespaces) and SANS use logical mappings so you, the DBA or developer, have no clue as to where anything physically is located.
    PS: Ouija Boards don't work either.

  • RMAN duplicate target database for standby from active fails to create newname for system tablespace/datafile

    When executing 'duplicate target database for standby from active'  the system tablespace/datafile (datafile 1)  is not cloned.  All other datafiles clone successfully.  The RMAN process aborts with the following errors while attempting to clone the system tablespace/datafile.
    ORA-19558: error de-allocating device
    ORA-19557: device error, device type: DISK, device name:
    ORA-17627: ORA-01041: internal error. hostdef extension doesn't exist
    ORA-17627: ORA-01041: internal error. hostdef extension doesn't exist
    ORA-03135: connection lost contact
    Here are the details:
    Primary is 11.2.0.2 RAC database  on an Exadata platform
    Standby is 11.2.0.2 Single Instance database (same patch level as primary) on a Red Hat Linux box
    This is an ASM to ASM duplication.
    This is not unique to this database.  We tried another database and go the same behavior - all datafiles clone successfully with the exception of the system tablespace/datafile.
    We have traced the RMAN execution and it seems to fail when it is trying to assign a NEWNAME to the system tablespace/datafile.
    We even issued an explicit SET NEWNAME command but RMAN ignored it.
    We also shutdown the primary and started is up in mount mode thinking that something had ahold of the System Tablespace/datafile.
    We also opened up the network firewall to allow permit any,any traffic.
    We increased the max_server_processes
    and added TCP.NODELAY=yes to the sqlnet.ora file.
    There seems to be some artifact present in our Primary System tablespace/data file that is preventing it form being cloned.
    checked all alert files grid, asm,  and dbhome - no abnormal messages.
    We are in the process of restoring the database from a backup but we would prefer to get this working using the 'Active Database' methodology

    I successfully created the standby database using RMAN backup and recovery.
    I started the managed recovery.  Archive logs are being sent from the primary to the standby ( I can see them in ASM), but the standby is not applying them.
    I get the following messages in the standby alert log...
    Fetching gap sequence in thread 2, gap sequence 154158-154257
    Tue Nov 26 16:19:58 2013
    Using STANDBY_ARCHIVE_DEST parameter default value as USE_DB_RECOVERY_FILE_DEST
    Using STANDBY_ARCHIVE_DEST parameter default value as USE_DB_RECOVERY_FILE_DEST
    Tue Nov 26 16:20:01 2013
    Fetching gap sequence in thread 2, gap sequence 154158-154257
    Tue Nov 26 16:20:11 2013
    Fetching gap sequence in thread 2, gap sequence 154158-154257
    Tue Nov 26 16:20:22 2013
    Fetching gap sequence in thread 2, gap sequence 154158-154257
    Tue Nov 26 16:20:32 2013
    Fetching gap sequence in thread 2, gap sequence 154158-154257
    I don't see any MRP processes:
    select process,
    status,
        thread#,
        sequence#,
       block#,
      blocks
      7     from v$managed_standby;
    PROCESS   STATUS          THREAD#  SEQUENCE#     BLOCK#     BLOCKS
    ARCH      CLOSING               2     154363          1        132
    ARCH      CONNECTED             0          0          0          0
    ARCH      CONNECTED             0          0          0          0
    ARCH      CONNECTED             0          0          0          0
    ARCH      CONNECTED             0          0          0          0
    ARCH      CONNECTED             0          0          0          0
    ARCH      CONNECTED             0          0          0          0
    ARCH      CONNECTED             0          0          0          0
    RFS       IDLE                  0          0          0          0
    RFS       IDLE                  1     145418        121          1
    RFS       IDLE                  0          0          0          0
    PROCESS   STATUS          THREAD#  SEQUENCE#     BLOCK#     BLOCKS
    RFS       IDLE                  0          0          0          0
    12 rows selected.
    SQL>  SELECT THREAD#, SEQUENCE#, APPLIED FROM V$ARCHIVED_LOG;
       THREAD#  SEQUENCE# APPLIED
             2     154356 NO
             2     154357 NO
             1     145411 NO
             2     154358 NO
             2     154360 NO
             2     154361 NO
             1     145414 NO
             1     145415 NO
             2     154362 NO
             2     154363 NO
             1     145416 NO
    11 rows selected.
    I do have the archive logs that cover sequences 154158-154257
    Crosschecked 38 objects
    Crosschecked 62 objects
    Finished implicit crosscheck backup at 26-NOV-13
    Starting implicit crosscheck copy at 26-NOV-13
    using channel ORA_DISK_1
    using channel ORA_DISK_2
    Crosschecked 2 objects
    archived log file name=+RECO_XORA/nmuasb00/archivelog/2013_11_26/thread_2_seq_154377.344.832521989 RECID=29 STAMP=832521990
    validation succeeded for archived log
    archived log file name=+RECO_XORA/nmuasb00/archivelog/2013_11_26/thread_2_seq_154378.346.832521991 RECID=31 STAMP=832521993
    Crosschecked 31 objects

  • Issue in OWB mapping - when changing source and target database

    Hi,
    I need help for resolution of the issue I am facing when moving mapping from development environment to QA.
    Here is situation,
    We develop ETL using one source, one staging and one target database.
    In development we use one control_center for Source to staging and another control center staging to target.
    All works fine in development.
    Now I have created new runtime repository and imported all OWB projects (with full dependencies, exact replica of development). Now I need to change source and staging and target as different database.
    I have created new database location connections and defined/attached DB connectors with stage and target location.
    Now issues are
    1.     Two Staging mapping are not able to bound with source table (giving different error
    a.     One mapping show error for source synonym translation no longer valid when deploying this mapping , but validation comes without any issue)
    b.     Other mapping show error for source table/object not bound to repository
    2.     All the target mapping show validation successful, but when deploying says “table or view does not exist”. But tables are exists on source stage and target. (also permission are set correctly for target user to read from staging tables).
    Not sure how to proceed from here.
    I have recreated new repository and re-imported all project/mappings and defined all connection but still same issue.
    Thanks in advance,
    Vipin

    1. Two Staging mapping are not able to bound with source table (giving different error
    a. One mapping show error for source synonym translation no longer valid when deploying this mapping , but validation comes without any issue)
    b. Other mapping show error for source table/object not bound to repository
    The above error were resolved when re-synchronized the table (for few I have to reimport the table) and mapping.
    2. All the target mapping show validation successful, but when deploying says “table or view does not exist”. But tables are exists on source stage and target. (also permission are set correctly for target user to read from staging tables).
    The above error still pending. My target mapping are not able to deployed/compiled.
    For the above I have defined one Staging location to one target target location and target location have connector to staging (not sure if I have to define connector name same as staging location, as I have created DB connector with different name but reference database is same as staging location).
    Mapping are assoicated with desired data locatoin and meta data.
    control center is also have that data location.
    Mapping are configured for the desired location.

  • One problem with constraints missing in source and target database....

    DB Gurus,
    In my database I am going to export and import data using datapump but dont specify CONTENT clause.
    so when I start working on my application, I came to know there are some constraints are missing. now how to check what are the constraints are missing from Source database and how to create those missing constraints on Target database?
    Suggest me if you have any idea.

    Create a database link between source and target schema.
    Use all_/dba_/_user_constraints to generate the difference using MINUS operator.
    something like this:
    select owner,constraint_name,constraint_type from dba_constraints@source
    MINUS
    select owner,constraint_name,constraint_type from dba_constraints

  • Relocating datafiles on standby database after mount point on stanby is ful

    Hi,
    We have a physical standby database.
    The location of datafiles on primary database are at /oracle/oradata/ and the location of datafiles on standby database are at /oracle/oradata/
    Now we are facing a situation of mount mount getting full on standby database so we need to move some tablespaces to another location on standby.
    Say old location is /oracle/oradata/ and new location is /oradata_new/ and the tablespaces to be relocated are say tab1 and tab2.
    Can anybody tell me whether following steps are correct.
    1. Stop managed recovery on standby database
    alter database recover managed standby database cancel;
    2. Shutdown standby database
    shutdown immediate;
    3. Open standby database in mount stage
    startup mount;
    4. Copy the datafiles to new location say /oradata_new/ using os level command
    4. Rename the datafile
    alter database rename file
    '/oracle/oradata/tab1.123451.dbf', '/oracle/oradata/tab1.123452.dbf','/oracle/oradata/tab2.123451.dbf',''/oracle/oradata/tab2.123452.dbf'
    to '/oradata_new/tab1.123451.dbf', '/oradata_new/tab1.123452.dbf','/oradata_new/tab2.123451.dbf',''/oradata_new/tab2.123452.dbf';
    5. Edit the parameter db_file_name_convert
    alter system set db_file_name_convert='/oracle/oradata/tab1','/oradata_new/tab1','/oracle/oradata/tab2','/oradata_new/tab2'
    6. Start a managed recovery on standby database
    alter database recover managed standby database disconnect from session;
    I am littelbit confused in step 5 as we want to relocate only two tablespaces and not all tablespaces so we have used.
    Can we use db_file_name_convert like this i.e. does this work for only two tablespaces tab1 and tab2.
    Thanks & Regards
    GirishA

    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/manage_ps.htm#i1010428
    8.3.4 Renaming a Datafile in the Primary Database
    When you rename one or more datafiles in the primary database, the change is not propagated to the standby database. Therefore, if you want to rename the same datafiles on the standby database, you must manually make the equivalent modifications on the standby database because the modifications are not performed automatically, even if the STANDBY_FILE_MANAGEMENT initialization parameter is set to AUTO.
    The following steps describe how to rename a datafile in the primary database and manually propagate the changes to the standby database.
    To rename the datafile in the primary database, take the tablespace offline:
    SQL> ALTER TABLESPACE tbs_4 OFFLINE;
    Exit from the SQL prompt and issue an operating system command, such as the following UNIX mv command, to rename the datafile on the primary system:
    % mv /disk1/oracle/oradata/payroll/tbs_4.dbf
    /disk1/oracle/oradata/payroll/tbs_x.dbf
    Rename the datafile in the primary database and bring the tablespace back online:
    SQL> ALTER TABLESPACE tbs_4 RENAME DATAFILE 2> '/disk1/oracle/oradata/payroll/tbs_4.dbf'
    3> TO '/disk1/oracle/oradata/payroll/tbs_x.dbf';
    SQL> ALTER TABLESPACE tbs_4 ONLINE;
    Connect to the standby database, query the V$ARCHIVED_LOG view to verify all of the archived redo log files are applied, and then stop Redo Apply:
    SQL> SELECT SEQUENCE#,APPLIED FROM V$ARCHIVED_LOG ORDER BY SEQUENCE#;
    SEQUENCE# APP
    8 YES
    9 YES
    10 YES
    11 YES
    4 rows selected.
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;
    Shut down the standby database:
    SQL> SHUTDOWN;
    Rename the datafile at the standby site using an operating system command, such as the UNIX mv command:
    % mv /disk1/oracle/oradata/payroll/tbs_4.dbf /disk1/oracle/oradata/payroll/tbs_x.dbf
    Start and mount the standby database:
    SQL> STARTUP MOUNT;
    Rename the datafile in the standby control file. Note that the STANDBY_FILE_MANAGEMENT initialization parameter must be set to MANUAL.
    SQL> ALTER DATABASE RENAME FILE '/disk1/oracle/oradata/payroll/tbs_4.dbf'
    2> TO '/disk1/oracle/oradata/payroll/tbs_x.dbf';
    On the standby database, restart Redo Apply:
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE
    2> DISCONNECT FROM SESSION;
    If you do not rename the corresponding datafile at the standby system, and then try to refresh the standby database control file, the standby database will attempt to use the renamed datafile, but it will not find it. Consequently, you will see error messages similar to the following in the alert log:
    ORA-00283: recovery session canceled due to errors
    ORA-01157: cannot identify/lock datafile 4 - see DBWR trace file
    ORA-01110: datafile 4: '/Disk1/oracle/oradata/payroll/tbs_x.dbf'

  • Integrate source db with target EBS database using ODI

    Hi
    We have a requirement to integrate our legacy systems to Oracle E-business suite 12.1.1 database 10g. We have several MS-SQL and oracle databases that has to be integrate with our Oracle EBS database. Oracle says to have separate schema for Master and Work repository, and also create a new schema when connecting to the data store and use the same schema as a WORK SCHEMA when creating a physical schema. My question is:
    1) What are all the privileges to be assigned to the new schema used as a WORK SCHEMA? Oracle EBS holds many FND schemas and APPS uses the synonyms of all the tables. So what oracle recommends when using EBS with ODI?
    2) Do we need to create a separate work schema for each and every SOURCE Database as well?
    3) We are not using separate oracle database for ODI we will be creating schemas in PROD to hold ODI data. Our management wants the the odi schema to hold the staging data. Can we use the the same Work schema mentioned in the first point as a staging area as well?
    Thank you
    Regards
    Shahrukh Yasin

    Hi Shahrukh ,
    There are two ways.(1st one is recommended)
    1.You can make source schema in one database and target schema, work schema in another database.
    For source schema only read only access privilege.
    You must have complete ownership in work schema and target schema(Minimum insert,update,delete).
    Here you have to always create one work schema in every target database
    2.if all schema present in 3 different database,you also need above criteria.
    But here you don't have to create work schema always.its fixed.
    This process takes much time for transformation and not an efficient way.
    Regards
    Bhabani

  • Exchange 2010 - 2013 Public Folder Migration target database gets much bigger than source database

    Hi,
    i have a question about the pf migration from Exchange 2010 to Exchange 2013. The source database ist about 370 GB. The migration process is still running. At the moment the target database is already 491GB big. Is that a normal behaviour during the migration
    process.
    Thanks

    Hi,
    After you have finished the public folder migration process, you can compare the item count of public folders to make sure all public folders have been moved to Exchange 2013.
    You can check this by comparing the snapshots before and after the migration. You can refer to the "How do I know this worked?" section in the following article.
    Migrate Public Folders to Exchange 2013 From Previous Versions
    http://technet.microsoft.com/en-us/library/jj150486(v=exchg.150).aspx
    Best regards,
    Belinda
    Belinda Ma
    TechNet Community Support

  • How to read the source file and copy the entire content to the target Database field

    Hi All,
    PI system extracts the actual business content from the incoming attachment file and copy the entire format to the target Database field which is a blob data type field.
    Sender Adapter: SOAP
    Receiver Adapter: JDBC
    JDBC Structure
    Can we achived the above requirement by using UDF with out Java Mapping?
    Thanks,
    Mahi.

    Ok, in this case yo can use the conten converion of the file adapter. Therefore you first need to create a Datatype in ESR with the required structure
    Then in your file adapter you need to use this datatype and its required fields:
    The Datatype then can be used as usual within you mapping.
    regards
    Christian

  • Duplicate target database not working

    Hi All,
    I want to clone my exisitng DB to other.
    I was refering the below link to do the same.
    [http://web.njit.edu/info/limpid/DOC/backup.102/b14191/rcmdupdb006.htm|http://web.njit.edu/info/limpid/DOC/backup.102/b14191/rcmdupdb006.htm]
    OS:Solaris 10
    Oracle version:10.2.0.4.0
    What I did tilll now:-
    1) Created a database MYDB1 in archive mode
    2) Created one tablespace, user and 1 table in DB
    3) Registered the MYDB1 in RMAN Db.
    4)Created a Pfile from MYDB1 database and changed db value and other values of MYDB1 to MYDB2(Database name to be duplicate)
    5)Created a password file
    6)Startup MYDB2 in NOMOUNT stage.
    7) At RMAN prompt I have done below things to but i have getting the error
    rman:rman:>rman target sys/mydb1@mydb1 catalog rman/rman auxiliary sys/mydb2@mydb2
    Recovery Manager: Release 10.2.0.4.0 - Production on Fri Sep 7 12:20:22 2012
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    connected to target database: MYDB1 (DBID=3867673817)
    connected to recovery catalog database
    connected to auxiliary database: MYDB2 (not mounted)
    RMAN> DUPLICATE TARGET DATABASE TO mydb2;
    Starting Duplicate Db at 07-SEP-12
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: sid=162 devtype=DISK
    WARNING: DB_FILE_NAME_CONVERT resulted in invalid ASM names; names changed to diskgroup only.
    contents of Memory Script:
       set until scn  911233;
       set newname for datafile  1 to
    "+disk2";
       set newname for datafile  2 to
    "+disk2";
       set newname for datafile  3 to
    "+disk2";
       set newname for datafile  4 to
    "+disk2";
       set newname for datafile  5 to
    "+disk2";
       restore
       check readonly
       clone database
    executing Memory Script
    executing command: SET until clause
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    Starting restore at 07-SEP-12
    using channel ORA_AUX_DISK_1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 09/07/2012 12:21:01
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-06026: some targets not found - aborting restore
    RMAN-06023: no backup or copy of datafile 5 found to restore
    RMAN-06023: no backup or copy of datafile 4 found to restore
    RMAN-06023: no backup or copy of datafile 3 found to restore
    RMAN-06023: no backup or copy of datafile 2 found to restore
    RMAN-06023: no backup or copy of datafile 1 found to restore
    List of Backups
    ===============
    Key     TY LV S Device Type Completion Time #Pieces #Copies Compressed Tag
    745933  B  F  A DISK        07-SEP-12       1       1       NO         FULL_BACKUP
    745934  B  F  A DISK        07-SEP-12       1       1       NO         FULL_BACKUP
    RMAN> list backup of database;
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    745933  Full    549.12M    DISK        00:01:19     07-SEP-12
            BP Key: 745935   Status: AVAILABLE  Compressed: NO  Tag: FULL_BACKUP
            Piece Name: /backup/oracle_backup/test_dir/rman_bkp/mydb1_0gnkjokf_1_1
      List of Datafiles in backup set 745933
      File LV Type Ckp SCN    Ckp Time  Name
      1       Full 980599     07-SEP-12 +DISK1/mydb1/datafile/system.256.793120373
      2       Full 980599     07-SEP-12 +DISK1/mydb1/datafile/undotbs1.258.793120373
      3       Full 980599     07-SEP-12 +DISK1/mydb1/datafile/sysaux.257.793120373
      4       Full 980599     07-SEP-12 +DISK1/mydb1/datafile/users.259.793120373
      5       Full 980599     07-SEP-12 +DISK1/mydb1/datafile/tbs_umesh.266.793207735
    RMAN> show all;
    RMAN configuration parameters are:
    CONFIGURE RETENTION POLICY TO REDUNDANCY 1; # default
    CONFIGURE BACKUP OPTIMIZATION OFF; # default
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP OFF;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '%F'; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE DISK FORMAT   '/backup/oracle_backup/test_dir/rman_bkp/mydb1_%U';
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '/software/oracle/product/10.2.0/database/dbs/snapcf_mydb1.f'; # default
    RMAN> exit
    Recovery Manager complete.
    @rman:rman:>Please Suggest where i am going wrong..
    Regards
    Umesh Gupta

    WARNING: DB_FILE_NAME_CONVERT resulted in invalid ASM names; names changed to diskgroup only.
    You must check DB_FILE_NAME_CONVERT parameter on MYDB2 database
    or I think you must add DB_FILE_NAME_CONVERT, LOG_FILE_NAME_CONVERT to DUPLICATE Command.
    DUPLICATE TARGET DATABASE TO MYDB2
    SPFILE
    SET
      "DB_FILE_NAME_CONVERT"="+DATA/MYDB1/","+DATA/MYDB2"
    ...  Regards
    Mahir M. Quluzade
    www.mahir-quluzade.com

  • Unable to connect my target database

    I am trying to create a clone of my database named proddb using my production RMNA backup,
    so when i export ORACLE_SID=proddb
    and then try to connect to my rman using
    rman target /
    its showing connected to target database clonedb
    i have checked the servicxes and there are no services named clonedb and their is no entry in my tnsnames or listner as clonedb......

    user9038728 wrote:
    oracle@localhost ~]$ export ORACLE_SID=proddb
    [oracle@localhost ~]$ export ORACLE_HOME=/home/oracle/app/oracle/product/11.2.0/dbhome_1/
    [oracle@localhost ~]$ rman target sys/password
    Recovery Manager: Release 11.2.0.1.0 - Production on Fri Jan 6 11:27:13 2012
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    connected to target database: CLONEDB (not mounted)
    RMAN>
    this is the output which i am getting.......Your target/source/Production database should be in OPEN status & your new/clone/auxiliary instance should be in started/not mounted status.
    You have to connect as below
    $rman target sys/*****@prod auxiliary /
    Then target will be production, you have to export SID of auxiliary and test the TNS connectivity to primary.

  • RMAN duplicate target database error

    Hello,
    I’m trying to use RMAN ‘duplicate target from active database’ to a remote host however I’m getting a confusing error:
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 11/19/2009 13:47:23
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-03009: failure of backup command on aux13 channel at 11/19/2009 13:47:23
    ORA-19504: failed to create file "+DATA2"
    RMAN-03009: failure of backup command on aux14 channel at 11/19/2009 13:47:23
    ORA-19504: failed to create file "+DATA2"
    Here’s my script:
    connect target sys/<password>@node1;
    connect auxiliary sys/<password>@node2;
    spool log to '/u01/app/oracle/dba/duplicate111909.log'
    run
    allocate auxiliary channel aux11 device type disk;
    allocate auxiliary channel aux12 device type disk;
    allocate channel aux13 device type disk;
    allocate channel aux14 device type disk;
    duplicate target database to node2 from active database DB_FILE_NAME_CONVERT='+DATA1','+DATA2';
    release channel aux14;
    release channel aux13;
    release channel aux12;
    release channel aux11;
    spool log off;
    exit;
    Node1 uses ASM disk group DATA1 and I intend to use ASM disk group DATA2 on node2. Has anyone seen this error before? I have tried various permutations of DB_FILE_NAME_CONVERT clauses but all error with ORA-19504. On a side note I was able to use ASMCMD cp command to copy a datafile from node1 DATA1 to node 2 DATA2. Also, if I use disk group DATA1 on node 2 then the RMAN duplicate command runs successfully, it only errors when using DATA2. And yes +DATA2 is mounted on node2:
    SQL> select name, state from v$asm_diskgroup;
    NAME STATE
    DATA1 MOUNTED
    DATA2 MOUNTED
    Any and all ideas are welcome. I’m running out of things to check.
    Thanks,
    RJL

    For a duplicate you don't need to give the DB_FILE_NAME_CONVERT especially if you go from a asm to asm diskgroup.
    The place where it'll put all the files will be automatically done taking care of the init.ora file parameter db_create_file_dest, (db_create_online_log_dest_1)

Maybe you are looking for

  • WSConfig problem when connecting to Portal with Netweaver Business client

    Hello everybody, I am trying to connect to an existing portal (7.0) with the new Netwevaer Business client. On one system, it works - on the other, it doesn't. The client tries to call the url: http://<host>:<j2ee-port>/NavigationWS/NavigationWSConfi

  • .mac email troubles on windows 2000 machine

    I posted this on another mac discussion board but didn't get any answers, so i thought i'd try here. During the time that I am unable to access the internet with my iMac, I'm using my Windows machine, WIndows 200 and Outlook Express to get my .mac em

  • What are the code completion keys in JDeveloper on Mac?

    Does anybody know which key combinations to use for code completion in the Java editor in JDeveloper on a Mac?

  • Iweb URL is the same for every page

    I have a small website built on iweb with a personal domain name, www.westonbennett.com and if you click on any of my pages the url doesn't change and is still www.westonbennett.com and I would like it to be a unique url for every page. Can anyone he

  • Using OR in Contains

    hi .. Using contains i want to search for record containing any word from the given string.... i.e if i use select count(*) title from docs where contains(text,'first',1) > 0 i want the records containing first or document... Where should i place thi