Steps for Data Guard with one primary and 2 standby

Hi,
Database :10.2.0.4, 11.2.0.1
Os: Windows , Unix
A ----------------> Primary database
B ----------------> Standby Database 1
C ----------------> Standby Database 2
I want to configure *2 standby* databases for single primary database.
Lets take, A ,B and C are my machines.My data guard configuration will be like,*archive logs will be moving* from A to B and A to C.
If i do any switchover in between A and B , now B is primary and remaining A and C are standby databases.At this stage also , archive logs should move from B to A and B to C. Also, same should happen from C to A and C to B,If i do switchover in between B and C.If everything is fine , then i will do switchback to main Primary database(A).
How do i have to mention PFILE in all machines ,the parameters like
LOG_ARCHIVE_DEST_1=LOCATION=<PATH> -- LOCAL ARCHIVE PATH
LOG_ARCHIVE_DEST_2=SERVICE=
LOG_ARCHIVE_DEST_3=SERVICE=
FAL_SERVER=
FAL_CLIENT=
STANDBY_FILE_MANAGEMENT=
In my tnsnames.ora , primary,standby1 and standby2 are my service entries and these are same in all of my machines.
Please suggest me , how do i can configure my pfiles in all machines ?.
Thanks,
Sunand

Not yet, but now you have me interested.
Please consider Flashback.
I still have to test but here's my take:
PRIMARY SETTINGS
*.FAL_SERVER=STANDBY
*.FAL_CLIENT=PRIMARY
*.STANDBY_FILE_MANAGEMENT=AUTO
*.DB_UNIQUE_NAME=PRIMARY
*.LOG_FILE_NAME_CONVERT='STANDBY','PRIMARY'
*.log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=PRIMARY'
*.log_archive_dest_2='SERVICE=STANDBY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY'
*.log_archive_dest_3='SERVICE=STANDBY2 LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY2'
*.LOG_ARCHIVE_DEST_STATE_1=ENABLE
*.LOG_ARCHIVE_DEST_STATE_2=ENABLE
*.LOG_ARCHIVE_DEST_STATE_3=ENABLE
*.LOG_ARCHIVE_MAX_PROCESSES=30
STANDBY 1 SETTINGS
*.FAL_SERVER=PRIMARY
*.FAL_CLIENT=STANDBY
*.STANDBY_FILE_MANAGEMENT=AUTO
*.DB_UNIQUE_NAME=STANDBY
*.LOG_FILE_NAME_CONVERT='PRIMARY','STANDBY'
*.log_archive_dest_1=LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=STANDBY'
*.log_archive_dest_2='SERVICE=PRIMARY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PRIMARY'
*.log_archive_dest_3='SERVICE=STANDBY2 LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY2'
*.LOG_ARCHIVE_DEST_STATE_1=ENABLE
*.LOG_ARCHIVE_DEST_STATE_2=DEFER
*.LOG_ARCHIVE_DEST_STATE_3=DEFER
*.LOG_ARCHIVE_MAX_PROCESSES=30
STANDBY2 SETTINGS
*.FAL_SERVER=PRIMARY
*.FAL_CLIENT=STANDBY2
*.STANDBY_FILE_MANAGEMENT=AUTO
*.DB_UNIQUE_NAME=STANDBY2
*.LOG_FILE_NAME_CONVERT='PRIMARY','STANDBY2'
*.log_archive_dest_1=LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=STANDBY2'
*.log_archive_dest_2='SERVICE=STANDBY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY'
*.log_archive_dest_3='SERVICE=PRIMARY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PRIMARY'
*.LOG_ARCHIVE_DEST_STATE_1=ENABLE
*.LOG_ARCHIVE_DEST_STATE_2=DEFER
*.LOG_ARCHIVE_DEST_STATE_3=DEFER
*.LOG_ARCHIVE_MAX_PROCESSES=30
Edited by: mseberg on Nov 29, 2010 9:39 AM
The first test slapped me. Looking at 409013.1 Cascaded Standby Databases
Edited by: mseberg on Nov 29, 2010 12:49 PM

Similar Messages

  • Data guard difference in primary and standby OS.

    Dear All,
    We are planning to setup a data guard in our landscape. But my question what are the prerequisites should be consider before getting in to that.
    Like my environment -
    1. The primary database is running with Red hat Linux 5 - IBM 64 bit and the standby database is Redhat Linux 2.1 IBM 32 bit system.
    2. The primary database is running with Red hat Linux 5 - IBM 64 bit with oracle 8.1.7.4 and the standby database is
       Linux  2.1 IBM 32 bit system with oracle 10g.
    Is the above thing is possible? or we should maintain the same OS / DB combination in the data guard setup.
    I am following the below link.
    http://download-west.oracle.com/docs/cd/B19306_01/server.102/b14239/standby.htm#i72053
    If anyone have thoughts on this please do advice me.
    Regards
    Vijay

    Please explain to us what exactly you want to do with the standby database?
    According to note 105047 - Support for Oracle functions in the SAP environment
    12. Data Guard
    You can use "Physical Standby", but Oracle provides support (not SAP).
    You cannot use "Logical Standby"
    So in short, no a data guard configuration with Red hat Linux 5 - IBM 64 bit with oracle 8.1.7.4 to Linux 2.1 IBM 32 bit system with oracle 10g is NOT possible.
    Regards, Michael

  • Data Guard with Oracle 9i and 10g -- have you done it?

    We have an ERP system (JDEdwards, db size approx 600 GB) and we'd like to copy the data in this system to a database on another computer. We'd use the remote database for reporting, so we need it to be available. We are looking at using Data Guard running in Logical Standby mode.
    The ERP database is Oracle 9i, and we are considering having the Logical Standby database be 10g. Have you done this and if so, what was your experience?
    Here's why we want to do this. There are many tables in the ERP database that contain descending indexes. We can't change this, and it is not supported with Data Guard under Oracle 9.2. Tables with descending indexes do not copy to the logical standby database.
    We were told that this is corrected in a newer release of Oracle, and we were wondering if a 10g database would properly copy tables with descending indexes.
    Thanks in advance for any experiences you can share.
    Best Regards,
    Mike

    What you are trying to do might not be possible. Because when you create logical standby you have to create a physical standby first and convert it. Primary and standby have to be same version.
    In furture it might be possible because 10G will support rolling upgrade of logical standby.
    However even it's possible you have to go through a lot of pain to setup and maintain it because Oracle don't support the setup.
    What you could try are Stream and Replication, I will say Stream is very interesting one. Because Oracle say : "Oracle Streams and Oracle Data Guard (including Data Guard SQL Apply) are independent features based on some common underlying infrastructure and technology. "
    http://www.oracle.com/technology/deploy/availability/htdocs/DataGuardStreams.html

  • Steps for data guard RAC primary RAC standby

    Hi,
    I have some doubt in configuring data guard for rac DR to rac primary.
    1) In RAC DR, physical standby will be in RAC before starting data guard for RAC Primary , right?.
    2) While configuring RAC DR , should one node be down in standby?.
    3) After creating the auxiliary database in standby, i mean restoring the rman backup from staging directory in DR,
    Shall we have to register the services(database,instance,asm,etc..) to CRS(OCR)?. If yes ,Why?.
    4) After DR configure over , Shall i shut/down one node?. Should all the nodes in DR be in up , while applying logs to
    standby ?.
    If anybody have made the setup and have the prepared document , requesting you to share with me.
    Thank you very much,
    Sunand

    Hi Sunand,
    Please follow the following My Oracle Support (MOS) Document ID:
    MAA - Creating a RAC Physical Standby for a RAC Primary [ID 380449.1]
    For further details, see the referred documents in previous post. Try practicing on VMWare because it will give you the flexibility for creating snapshots to save your work. Please see the answers for your question below:
    Q1. In RAC DR, physical standby will be in RAC before starting data guard for RAC Primary , right?.
    Answer: Yes, this is recommend but not mandatory. You can add a second node later on.
    Q2. While configuring RAC DR , should one node be down in standby?
    Answer: Yes, there are some database configuration commands that should be run in exclusive mode. You can start the remaining instance(s) after DB creation. Do not forget to set CLUSTER_DATABASE parameter to TRUE. Also, all instances except one should be close while performing switchover/failover.
    Q3. After creating the auxiliary database in standby, i mean restoring the rman backup from staging directory in DR, Shall we have to register the services(database,instance,asm,etc..) to CRS(OCR)?. If yes ,Why?
    Answer: Yes, this is recommended for RAC configuration to take advantage of high availability services.
    Q4. After DR configure over , Shall i shut/down one node?. Should all the nodes in DR be in up , while applying logs to
    standby ?
    Answer: You can start up all the instances after DR creation. In this case if your one node goes down log apply services will continue to apply changes in DR.
    Hope, this helps to clear your questions.
    Regards,
    Shahid

  • Tip for Data Merge with Multiple Records (labels, etc.)

    I have seen many InDesign Data Merge questions about how to create sheets of mailing labels, especially the problem of getting only one label per page, when you need 30 or more.
    Adobe's instructions are poor and incomplete in that InDesign doesn't step out the records from a data source - it steps out the FRAMES that contain the data field placeholders.
    That is why you only need to place a text or image frame once on a single page - during the datamerge, InDesign will create the additional FRAMES for each record, and it will create the pages required to hold them.
    You do have to set the desired spacing on the
    If you create the frame on a Master page, ID allows you to update the data source (when it changes) in the Data Merge tool panel.
    These are very nice and robust features, but the documentation for them is confusing to many people.
    You will find more great in-depth help  for Data Merge, with screen captures and attachments, here in the forum.

    For a multiple record merge you need one set of placeholders, then set the margins and spacing in the merge options so the positioning is correct.
    Warning: Using Preview in a multiple record merge will corrupt the file. If you press the preview button, Undo after looking at the preview, then merge without using preview.

  • Oracle data guard configuration for primary and standby db_name

    I am working on configuring an active data guard for one primary DB and one standby DB. I have a few questions:
    1. Can I use different db_name, db_unique_name and instance_name for primary and standby. For example: primary(db_name, db_unique_name and instance_name)=chicago. When I create standby DB with Rman backup and copy of pfile and control file from primary DB or use Grid control to create standby database. Oracle document or Grid control all keep standby db_name=chicago. Only make standby db_unique_name and instance_name=boston. Due to my application system condition, I want to make db_name=boston, not keep it as the same as primary=chicago. Is this valid configuration?
    2. In primary datafiles, application system generate datafile name like this: hr_chicago_01.dbf, fn_chicago_01.dbf. When I move datafiles to standby server, if I plan to use db_name=boston for standby DB, can I rename datafiles as
    hr_boston_01.dbf, fn_boston_01.dbf? In this way, datafile name match up with db_name. but I will create standby log group and members on primary and standby identically. If in future switching over, DB will not have problems.
    3. If I don't use primary DB backup. Instead, I copy all datafiles, redo_log files (no control files) to standby. Then "alter database backup controlfile to trace" from promary and also " create pfile='/xxx/initSTANDBY.ora' from primary. Then modify init.ora and controlfile. Then run control.sql to bring standby DB up. After that, configure redo log shipping and apply with data guard or SQL. Is this a acceptable way to create physical standby DB?
    Please advise your comments. Thanks in advance.

    I want to make db_name=boston, not keep it as the same as primary=chicago. Is this valid configuration?NO. DB_NAME must be the same ("chicago") at both sites. The Standby will be using a different DB_UNIQUE_NAME (e.g. "boston") and can be using a different Instance name / SID (e.g. "boston").
    can I rename datafiles Yes. The database file names can be changed.
    If I don't use primary DB backup. Instead, I copy all datafiles, redo_log files (no control files) to standbyWhat is the difference between the first sentence (a backup of the primary) and the second sentence (a copy of the primary) ? A Copy is a backup.
    Are you intending to differentiate between an RMAN Backup and a User-Managed (aka "scripted") backup ?
    Normally, for DataGuard, tou can use non-RMAN methods to copy the database but there's no value add in this.
    You'd still have to setup DataGuard ! (and I wonder if you'd have complications setting up Active DataGuard).
    But remember that you MUST create the Standby controlfile from the Primary and copy it over to your Standby -- particularly as you are planning to use DataGuard. This is not created by 'alter database backup controlfile to trace' , but by 'ALTER DATABASE CREATE STANDBY CONTROLFILE AS 'filename''
    Hemant K Chitale

  • Data Guard Failover after primary site network failure or disconnect.

    Hello Experts:
    I'll try to be clear and specific with my issue:
    Environment:
    Two nodes with NO shared storage (I don't have an Observer running).
    Veritas Cluser Server (VCS) with Data Guar Agent. (I don't use the Broker. Data Guard agent "takes care" of the switchover and failover).
    Two single instance databases, one per node. NO RAC.
    What I'm being able to perform with no issues:
    Manual switch(over) of the primary database by running VCS command "hagrp -switch oraDG_group -to standby_node"
    Automatic fail(over) when primary node is rebooted with "reboot" or "init"
    Automatic fail(over) when primary node is shut down with "shutdown".
    What I'm NOT being able to perform:
    If I manually unplug the network cables from the primary site (all the network, not only the link between primary and standby node so, it's like a server unplug from the energy source).
    Same situation happens if I manually disconnect the server from the power.
    This is the alert logs I have:
    This is the portion of the alert log at Standby site when Real Time Replication is working fine:
    Recovery of Online Redo Log: Thread 1 Group 4 Seq 7 Reading mem 0
      Mem# 0: /u02/oracle/fast_recovery_area/standby_db/onlinelog/o1_mf_4_9c3tk3dy_.log
    At this moment, node1 (Primary) is completely disconnected from the network. SEE at the end when the database (standby which should be converted to PRIMARY) is not getting all the archived logs from the Primary due to the abnormal disconnect from the network:
    Identified End-Of-Redo (failover) for thread 1 sequence 7 at SCN 0xffff.ffffffff
    Incomplete Recovery applied until change 15922544 time 12/23/2013 17:12:48
    Media Recovery Complete (primary_db)
    Terminal Recovery: successful completion
    Forcing ARSCN to IRSCN for TR 0:15922544
    Mon Dec 23 17:13:22 2013
    ARCH: Archival stopped, error occurred. Will continue retrying
    ORACLE Instance primary_db - Archival ErrorAttempt to set limbo arscn 0:15922544 irscn 0:15922544
    ORA-16014: log 4 sequence# 7 not archived, no available destinations
    ORA-00312: online log 4 thread 1: '/u02/oracle/fast_recovery_area/standby_db/onlinelog/o1_mf_4_9c3tk3dy_.log'
    Resetting standby activation ID 2071848820 (0x7b7de774)
    Completed:  ALTER DATABASE RECOVER MANAGED STANDBY DATABASE FINISH
    Mon Dec 23 17:13:33 2013
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE FINISH
    Terminal Recovery: applying standby redo logs.
    Terminal Recovery: thread 1 seq# 7 redo required
    Terminal Recovery:
    Recovery of Online Redo Log: Thread 1 Group 4 Seq 7 Reading mem 0
      Mem# 0: /u02/oracle/fast_recovery_area/standby_db/onlinelog/o1_mf_4_9c3tk3dy_.log
    Identified End-Of-Redo (failover) for thread 1 sequence 7 at SCN 0xffff.ffffffff
    Incomplete Recovery applied until change 15922544 time 12/23/2013 17:12:48
    Media Recovery Complete (primary_db)
    Terminal Recovery: successful completion
    Forcing ARSCN to IRSCN for TR 0:15922544
    Mon Dec 23 17:13:22 2013
    ARCH: Archival stopped, error occurred. Will continue retrying
    ORACLE Instance primary_db - Archival ErrorAttempt to set limbo arscn 0:15922544 irscn 0:15922544
    ORA-16014: log 4 sequence# 7 not archived, no available destinations
    ORA-00312: online log 4 thread 1: '/u02/oracle/fast_recovery_area/standby_db/onlinelog/o1_mf_4_9c3tk3dy_.log'
    Resetting standby activation ID 2071848820 (0x7b7de774)
    Completed:  ALTER DATABASE RECOVER MANAGED STANDBY DATABASE FINISH
    Mon Dec 23 17:13:33 2013
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE FINISH
    Attempt to do a Terminal Recovery (primary_db)
    Media Recovery Start: Managed Standby Recovery (primary_db)
    started logmerger process
    Mon Dec 23 17:13:33 2013
    Managed Standby Recovery not using Real Time Apply
    Media Recovery failed with error 16157
    Recovery Slave PR00 previously exited with exception 283
    ORA-283 signalled during:  ALTER DATABASE RECOVER MANAGED STANDBY DATABASE FINISH...
    Mon Dec 23 17:13:34 2013
    Shutting down instance (immediate)
    Shutting down instance: further logons disabled
    Stopping background process MMNL
    Stopping background process MMON
    License high water mark = 38
    All dispatchers and shared servers shutdown
    ALTER DATABASE CLOSE NORMAL
    ORA-1109 signalled during: ALTER DATABASE CLOSE NORMAL...
    ALTER DATABASE DISMOUNT
    Shutting down archive processes
    Archiving is disabled
    Mon Dec 23 17:13:38 2013
    Mon Dec 23 17:13:38 2013
    Mon Dec 23 17:13:38 2013
    ARCH shutting downARCH shutting down
    ARCH shutting down
    ARC0: Relinquishing active heartbeat ARCH role
    ARC2: Archival stopped
    ARC0: Archival stopped
    ARC1: Archival stopped
    Completed: ALTER DATABASE DISMOUNT
    ARCH: Archival disabled due to shutdown: 1089
    Shutting down archive processes
    Archiving is disabled
    Mon Dec 23 17:13:40 2013
    Stopping background process VKTM
    ARCH: Archival disabled due to shutdown: 1089
    Shutting down archive processes
    Archiving is disabled
    Mon Dec 23 17:13:43 2013
    Instance shutdown complete
    Mon Dec 23 17:13:44 2013
    Adjusting the default value of parameter parallel_max_servers
    from 1280 to 470 due to the value of parameter processes (500)
    Starting ORACLE instance (normal)
    ************************ Large Pages Information *******************
    Per process system memlock (soft) limit = 64 KB
    Total Shared Global Region in Large Pages = 0 KB (0%)
    Large Pages used by this instance: 0 (0 KB)
    Large Pages unused system wide = 0 (0 KB)
    Large Pages configured system wide = 0 (0 KB)
    Large Page size = 2048 KB
    RECOMMENDATION:
      Total System Global Area size is 3762 MB. For optimal performance,
      prior to the next instance restart:
      1. Increase the number of unused large pages by
    at least 1881 (page size 2048 KB, total size 3762 MB) system wide to
      get 100% of the System Global Area allocated with large pages
      2. Large pages are automatically locked into physical memory.
    Increase the per process memlock (soft) limit to at least 3770 MB to lock
    100% System Global Area's large pages into physical memory
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Initial number of CPU is 32
    Number of processor cores in the system is 16
    Number of processor sockets in the system is 2
    CELL communication is configured to use 0 interface(s):
    CELL IP affinity details:
        NUMA status: NUMA system w/ 2 process groups
        cellaffinity.ora status: cannot find affinity map at '/etc/oracle/cell/network-config/cellaffinity.ora' (see trace file for details)
    CELL communication will use 1 IP group(s):
        Grp 0:
    Picked latch-free SCN scheme 3
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =88
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    NUMA system with 2 nodes detected
    Starting up:
    Oracle Database 11g Enterprise Edition Release 11.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options.
    ORACLE_HOME = /u01/oracle/product/11.2.0.4
    System name:    Linux
    Node name:      node2.localdomain
    Release:        2.6.32-131.0.15.el6.x86_64
    Version:        #1 SMP Tue May 10 15:42:40 EDT 2011
    Machine:        x86_64
    Using parameter settings in server-side spfile /u01/oracle/product/11.2.0.4/dbs/spfileprimary_db.ora
    System parameters with non-default values:
      processes                = 500
      sga_target               = 3760M
      control_files            = "/u02/oracle/orafiles/primary_db/control01.ctl"
      control_files            = "/u01/oracle/fast_recovery_area/primary_db/control02.ctl"
      db_file_name_convert     = "standby_db"
      db_file_name_convert     = "primary_db"
      log_file_name_convert    = "standby_db"
      log_file_name_convert    = "primary_db"
      control_file_record_keep_time= 40
      db_block_size            = 8192
      compatible               = "11.2.0.4.0"
      log_archive_dest_1       = "location=/u02/oracle/archivelogs/primary_db"
      log_archive_dest_2       = "SERVICE=primary_db ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=primary_db"
      log_archive_dest_state_2 = "ENABLE"
      log_archive_min_succeed_dest= 1
      fal_server               = "primary_db"
      log_archive_trace        = 0
      log_archive_config       = "DG_CONFIG=(primary_db,standby_db)"
      log_archive_format       = "%t_%s_%r.dbf"
      log_archive_max_processes= 3
      db_recovery_file_dest    = "/u02/oracle/fast_recovery_area"
      db_recovery_file_dest_size= 30G
      standby_file_management  = "AUTO"
      db_flashback_retention_target= 1440
      undo_tablespace          = "UNDOTBS1"
      remote_login_passwordfile= "EXCLUSIVE"
      db_domain                = ""
      dispatchers              = "(PROTOCOL=TCP) (SERVICE=primary_dbXDB)"
      job_queue_processes      = 0
      audit_file_dest          = "/u01/oracle/admin/primary_db/adump"
      audit_trail              = "DB"
      db_name                  = "primary_db"
      db_unique_name           = "standby_db"
      open_cursors             = 300
      pga_aggregate_target     = 1250M
      dg_broker_start          = FALSE
      diagnostic_dest          = "/u01/oracle"
    Mon Dec 23 17:13:45 2013
    PMON started with pid=2, OS id=29108
    Mon Dec 23 17:13:45 2013
    PSP0 started with pid=3, OS id=29110
    Mon Dec 23 17:13:46 2013
    VKTM started with pid=4, OS id=29125 at elevated priority
    VKTM running at (1)millisec precision with DBRM quantum (100)ms
    Mon Dec 23 17:13:46 2013
    GEN0 started with pid=5, OS id=29129
    Mon Dec 23 17:13:46 2013
    DIAG started with pid=6, OS id=29131
    Mon Dec 23 17:13:46 2013
    DBRM started with pid=7, OS id=29133
    Mon Dec 23 17:13:46 2013
    DIA0 started with pid=8, OS id=29135
    Mon Dec 23 17:13:46 2013
    MMAN started with pid=9, OS id=29137
    Mon Dec 23 17:13:46 2013
    DBW0 started with pid=10, OS id=29139
    Mon Dec 23 17:13:46 2013
    DBW1 started with pid=11, OS id=29141
    Mon Dec 23 17:13:46 2013
    DBW2 started with pid=12, OS id=29143
    Mon Dec 23 17:13:46 2013
    DBW3 started with pid=13, OS id=29145
    Mon Dec 23 17:13:46 2013
    LGWR started with pid=14, OS id=29147
    Mon Dec 23 17:13:46 2013
    CKPT started with pid=15, OS id=29149
    Mon Dec 23 17:13:46 2013
    SMON started with pid=16, OS id=29151
    Mon Dec 23 17:13:46 2013
    RECO started with pid=17, OS id=29153
    Mon Dec 23 17:13:46 2013
    MMON started with pid=18, OS id=29155
    Mon Dec 23 17:13:46 2013
    MMNL started with pid=19, OS id=29157
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    starting up 1 shared server(s) ...
    ORACLE_BASE from environment = /u01/oracle
    Mon Dec 23 17:13:46 2013
    ALTER DATABASE   MOUNT
    ARCH: STARTING ARCH PROCESSES
    Mon Dec 23 17:13:50 2013
    ARC0 started with pid=23, OS id=29210
    ARC0: Archival started
    ARCH: STARTING ARCH PROCESSES COMPLETE
    ARC0: STARTING ARCH PROCESSES
    Successful mount of redo thread 1, with mount id 2071851082
    Mon Dec 23 17:13:51 2013
    ARC1 started with pid=24, OS id=29212
    Allocated 15937344 bytes in shared pool for flashback generation buffer
    Mon Dec 23 17:13:51 2013
    ARC2 started with pid=25, OS id=29214
    Starting background process RVWR
    ARC1: Archival started
    ARC1: Becoming the 'no FAL' ARCH
    ARC1: Becoming the 'no SRL' ARCH
    Mon Dec 23 17:13:51 2013
    RVWR started with pid=26, OS id=29216
    Physical Standby Database mounted.
    Lost write protection disabled
    Completed: ALTER DATABASE   MOUNT
    Mon Dec 23 17:13:51 2013
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE
             USING CURRENT LOGFILE DISCONNECT FROM SESSION
    Attempt to start background Managed Standby Recovery process (primary_db)
    Mon Dec 23 17:13:51 2013
    MRP0 started with pid=27, OS id=29219
    MRP0: Background Managed Standby Recovery process started (primary_db)
    ARC2: Archival started
    ARC0: STARTING ARCH PROCESSES COMPLETE
    ARC2: Becoming the heartbeat ARCH
    ARC2: Becoming the active heartbeat ARCH
    ARCH: Archival stopped, error occurred. Will continue retrying
    ORACLE Instance primary_db - Archival Error
    ORA-16014: log 4 sequence# 7 not archived, no available destinations
    ORA-00312: online log 4 thread 1: '/u02/oracle/fast_recovery_area/standby_db/onlinelog/o1_mf_4_9c3tk3dy_.log'
    At this moment, I've lost service and I have to wait until the prmiary server goes up again to receive the missing log.
    This is the rest of the log:
    Fatal NI connect error 12543, connecting to:
    (DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=node1)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=primary_db)(CID=(PROGRAM=oracle)(HOST=node2.localdomain)(USER=oracle))))
      VERSION INFORMATION:
            TNS for Linux: Version 11.2.0.4.0 - Production
            TCP/IP NT Protocol Adapter for Linux: Version 11.2.0.4.0 - Production
      Time: 23-DEC-2013 17:13:52
      Tracing not turned on.
      Tns error struct:
        ns main err code: 12543
    TNS-12543: TNS:destination host unreachable
        ns secondary err code: 12560
        nt main err code: 513
    TNS-00513: Destination host unreachable
        nt secondary err code: 113
        nt OS err code: 0
    Fatal NI connect error 12543, connecting to:
    (DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=node1)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=primary_db)(CID=(PROGRAM=oracle)(HOST=node2.localdomain)(USER=oracle))))
      VERSION INFORMATION:
            TNS for Linux: Version 11.2.0.4.0 - Production
            TCP/IP NT Protocol Adapter for Linux: Version 11.2.0.4.0 - Production
      Time: 23-DEC-2013 17:13:55
      Tracing not turned on.
      Tns error struct:
        ns main err code: 12543
    TNS-12543: TNS:destination host unreachable
        ns secondary err code: 12560
        nt main err code: 513
    TNS-00513: Destination host unreachable
        nt secondary err code: 113
        nt OS err code: 0
    started logmerger process
    Mon Dec 23 17:13:56 2013
    Managed Standby Recovery starting Real Time Apply
    MRP0: Background Media Recovery terminated with error 16157
    Errors in file /u01/oracle/diag/rdbms/standby_db/primary_db/trace/primary_db_pr00_29230.trc:
    ORA-16157: media recovery not allowed following successful FINISH recovery
    Managed Standby Recovery not using Real Time Apply
    Completed: ALTER DATABASE RECOVER MANAGED STANDBY DATABASE
             USING CURRENT LOGFILE DISCONNECT FROM SESSION
    Recovery Slave PR00 previously exited with exception 16157
    MRP0: Background Media Recovery process shutdown (primary_db)
    Fatal NI connect error 12543, connecting to:
    (DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=node1)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=primary_db)(CID=(PROGRAM=oracle)(HOST=node2.localdomain)(USER=oracle))))
      VERSION INFORMATION:
            TNS for Linux: Version 11.2.0.4.0 - Production
            TCP/IP NT Protocol Adapter for Linux: Version 11.2.0.4.0 - Production
      Time: 23-DEC-2013 17:13:58
      Tracing not turned on.
      Tns error struct:
        ns main err code: 12543
    TNS-12543: TNS:destination host unreachable
        ns secondary err code: 12560
        nt main err code: 513
    TNS-00513: Destination host unreachable
        nt secondary err code: 113
        nt OS err code: 0
    Mon Dec 23 17:14:01 2013
    Fatal NI connect error 12543, connecting to:
    (DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=node1)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=primary_db)(CID=(PROGRAM=oracle)(HOST=node2.localdomain)(USER=oracle))))
      VERSION INFORMATION:
            TNS for Linux: Version 11.2.0.4.0 - Production
            TCP/IP NT Protocol Adapter for Linux: Version 11.2.0.4.0 - Production
      Time: 23-DEC-2013 17:14:01
      Tracing not turned on.
      Tns error struct:
        ns main err code: 12543
    TNS-12543: TNS:destination host unreachable
        ns secondary err code: 12560
        nt main err code: 513
    TNS-00513: Destination host unreachable
        nt secondary err code: 113
        nt OS err code: 0
    Error 12543 received logging on to the standby
    FAL[client, ARC0]: Error 12543 connecting to primary_db for fetching gap sequence
    Archiver process freed from errors. No longer stopped
    Mon Dec 23 17:15:07 2013
    Using STANDBY_ARCHIVE_DEST parameter default value as /u02/oracle/archivelogs/primary_db
    Mon Dec 23 17:19:51 2013
    ARCH: Archival stopped, error occurred. Will continue retrying
    ORACLE Instance primary_db - Archival Error
    ORA-16014: log 4 sequence# 7 not archived, no available destinations
    ORA-00312: online log 4 thread 1: '/u02/oracle/fast_recovery_area/standby_db/onlinelog/o1_mf_4_9c3tk3dy_.log'
    Mon Dec 23 17:26:18 2013
    RFS[1]: Assigned to RFS process 31456
    RFS[1]: No connections allowed during/after terminal recovery.
    Mon Dec 23 17:26:47 2013
    flashback database to scn 15921680
    ORA-16157 signalled during: flashback database to scn 15921680...
    Mon Dec 23 17:27:05 2013
    alter database recover managed standby database using current logfile disconnect
    Attempt to start background Managed Standby Recovery process (primary_db)
    Mon Dec 23 17:27:05 2013
    MRP0 started with pid=28, OS id=31481
    MRP0: Background Managed Standby Recovery process started (primary_db)
    started logmerger process
    Mon Dec 23 17:27:10 2013
    Managed Standby Recovery starting Real Time Apply
    MRP0: Background Media Recovery terminated with error 16157
    Errors in file /u01/oracle/diag/rdbms/standby_db/primary_db/trace/primary_db_pr00_31486.trc:
    ORA-16157: media recovery not allowed following successful FINISH recovery
    Managed Standby Recovery not using Real Time Apply
    Completed: alter database recover managed standby database using current logfile disconnect
    Recovery Slave PR00 previously exited with exception 16157
    MRP0: Background Media Recovery process shutdown (primary_db)
    Mon Dec 23 17:27:18 2013
    RFS[2]: Assigned to RFS process 31492
    RFS[2]: No connections allowed during/after terminal recovery.
    Mon Dec 23 17:28:18 2013
    RFS[3]: Assigned to RFS process 31614
    RFS[3]: No connections allowed during/after terminal recovery.
    Do you have any advice?
    Thanks!
    Alex.

    Hello;
    What's not clear to me in your question at this point:
    What I'm NOT being able to perform:
    If I manually unplug the network cables from the primary site (all the network, not only the link between primary and standby node so, it's like a server unplug from the energy source).
    Same situation happens if I manually disconnect the server from the power.
    This is the alert logs I have:"
    Are you trying a failover to the Standby?
    Please advise.
    Is it possible your "valid_for clause" is set incorrectly?
    Would also review this:
    ORA-16014 and ORA-00312 Messages in Alert.log of Physical Standby
    Best Regards
    mseberg

  • Best Practice for monitoring database targets configured for Data Guard

    We are in the process of migrating our DB targets to 12c Cloud Control. 
    In our current 10g environment the Primary Targets are monitored and administered by OEM GC A, and the Standby Targets are monitored by OEM GC B.  Originally, I believe this was because of proximity and network speed, and over time it evolved to a Primary/Standby separation.  One of the greatest challenges in this configuration is keeping OEM jobs in sync on both sides (in case of switchover/failover).
    For our new OEM CC environment we are setting up CC A and CC B.  However, I would like to determine if it would be smarter to monitor all DB targets (Primary and Standby) from the same CC console.  In other words, monitor and administer DB Primary and Standby from the same OEM CC Console.   I am trying to determine the best practice.  I am not sure if administering a swichover from Cloud Control from Primary to Standby requires that both targets are monitored in the same environment or not.
    I am interested in feedback.   I am also interested in finding good reference materials (I have been looking at Oracle documentation and other documents online).   Thanks for your input and thoughts.  I am deliberately trying to keep this as concise as possible.

    OMS is a tool it is not need to monitor your primary and standby what is what I meant by the comment.
    The reason you need the same OMS to monitor both the primary and the standby is in the Data Guard administration screen it will show both targets. You also will have the option of doing switch-overs and fail-overs as well as convert the primary or standby. One of the options is also to move all the jobs that are scheduled with primary over to the standby during a switch-over or fail-over.
    There is no document that states that you need to have all targets on one OMS but that is the best method for the reason of having OMS. OMS is a tool to have all targets in a central repository. If you start have different OMS server and OMS repository you will need to log into separate OMS to administrator the targets.

  • Database is configured for Data Guard

    I am run a UTF8 conversion on a development database that has been cloned from Data Guard. There is a warning in Migration Status: "Database is configured for Data Guard" What is the DMU looking at to determine this? The database is open in read-write mode and it behaving like a primary database (I run the DMU scan and run updates to fix invalid representations). I would like to know what settings I need to update.
    Is this preventing me from converting tables using CTAS? When I try to select this for all tables I get the message, "The DMU does not support the conversion method "Copy data using CREATE TABLE AS SELECT" for tables that are involved in an Oracle Streams process, like capture or apply. Use another available conversion method for the table"
    Thanks,
    Ben

    The DMU checks if the parameter DG_BROKER_START is set to true.
    The problem with CTAS is independent from Data Guard. The DMU checks for tables that:
    - are source of asynchronous Streams capture, or
    - have update conflict handlers, or
    - have DML handlers, or
    - have conflict resolution parameters
    The above tables are considered configured for Oracle Streams and are not supported by CTAS conversion method. This is because the CTAS method creates a converted copy of the table and drops the original. The DMU is not capable of moving the Streams configuration information from the old table to the new one.
    Thanks,
    Sergiusz

  • I tried to sinc my photos and i ended up deleting off my phone all the pictures.  i only ended up with one folder, and i cant figure out how to put rest of folders from my computer back on my phone.....help

    i tried to sinc my photos and i ended up deleting off my phone all the pictures.  i only ended up with one folder, and i cant figure out how to put rest of folders from my computer back on my phone.....help
    and on a second problem.  I need to know for sure that my contacts are backed up on my computer

    Create a parent folder on your computer to store the named folders of photos that you want transferred to your iPhone. Under the Photos tab for your iPhone sync preferences with iTunes, select the parent folder. All named folders of photos within will be transferred to your iPhone as separate folders or albums.
    Regarding your 2nd question, this should be easy. Your contacts should be available on your computer with or without an iPhone or any cell phone, which can be lost or stolen.
    Although contacts are included with your iPhone's backup which is updated by iTunes as the first step during the iTunes sync process, not a good idea to depend on this either since the iPhone backup includes a significant amount of other data. Contacts are designed to be synced with a supported address book app on your computer. With Windoze, this can be with Outlook 2003, 2007, or 2010 along with syncing calendar events and notes, or with the address book app used by Outlook Express with XP and by Windows Mail with Vista and Windows 7 called Windows Contacts for syncing contacts only.

  • Steps for Data extraction from SAP r/3

    Dear all,
    I am New to SAP Bw.
    I have done data extraction from Excel into SAP BW system.
    that is like
    Create info objects > info area> Catalog
                                                --> Character catalog
                                                --> Key catalog
    Create info source
    Upload data.
    create info cube
    I need similar steps for data extraction for SAP R/3
    1. when data is in Ztables ( using Views/Infosets/Function etc)
    2. When data is with Standard SAP using Business Content.
    Thanks and Regards,
    Gaurav Sood

    hi,
    chk the links
    Generic Extraction
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/84bf4d68-0601-0010-13b5-b062adbb3e33
    CO-PA
    http://help.sap.com/saphelp_46c/helpdata/en/7a/4c37ef4a0111d1894c0000e829fbbd/content.htm
    CO-PC
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/fb07ab90-0201-0010-c489-d527d39cc0c6
    iNVENTORY
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f83be790-0201-0010-4fb0-98bd7c01e328
    Extractions in BI
    https://www.sdn.sap.com/irj/sdn/wiki
    LO Extraction:
    /people/sap.user72/blog/2004/12/16/logistic-cockpit-delta-mechanism--episode-one-v3-update-the-145serializer146
    /people/sap.user72/blog/2005/01/19/logistic-cockpit-delta-mechanism--episode-three-the-new-update-methods
    /people/sap.user72/blog/2005/02/14/logistic-cockpit--when-you-need-more--first-option-enhance-it
    /people/sap.user72/blog/2004/12/23/logistic-cockpit-delta-mechanism--episode-two-v3-update-when-some-problems-can-occur
    /people/sap.user72/blog/2005/04/19/logistic-cockpit-a-new-deal-overshadowed-by-the-old-fashioned-lis
    Remya

  • Question on db_unique_name in init.ora for Data Guard

    I need to set up only one physical standby on a different box (at a different location) for the primary db in production.
    OS: Sun Sparc Solaris 10
    Oracle: 10.2.0.3
    Can I use the same db_unique_name in init.ora for both primary and standby DBs?
    What are the minimal parameters required by Data Guard I have to specify in the init.ora in my case?
    Could anyone please post an example of init.ora for both primary and standby DBs?
    Thanks very much in advance.

    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/create_ps.htm#i63561

  • Separate listener for Data Guard

    I am setting up a best practice about using a dedicated listener for Data Guard. The idea is to maintain full functionality of Data Guard while application team is requesting to bring down listener service (according to business requirement). Need your opinion on these:
    1. I understand that there may be a very little chance when listener is required by Data Guard, but I find it no harm to do this. Do you agree with me?
    2. In RAC environment, we can only have 1 VIP to be used in listener.ora. I am thinking of using same IP but different port numbers for different listener. Any better idea than this?
    Many thanks

    It is never a bad practice to use separate listeners at the primary and at the Standby for Data Guard's use. A listener at the standby is required by Data Guard to make a connection to that standby. A listener at the Primary is required for Data Guard to make a reverse connection from the Standby to the Primary for some kinds of Gap resolution (missing log file the the Primary thinks it already sent or a corrupted log file etc). And of course, when you switch roles.
    To answer the second question could you please tell me what version of Oracle you are using and if you plan on using the Data Guard Broker or not?
    Thanks.
    Larry

  • HT5621 When I got my new iPhone and was trying to switch everything over I accidentally set up a new icloud account with one email and had one on my old iphone with another account. Does anyone know how I can merge the two accounts?

    When I got my new iPhone and was trying to switch everything over I accidentally set up a new icloud account with one email and had one on my old iphone with another account. Does anyone know how I can merge the two accounts?

    You cannot merge Apple IDs but you can go to Settings > iCloud and 'Delete Account'.  When prompted to turn off documents and data, choose the only option which is Delete from my iPhone, but on the other prompt for Contacts, Calendars, etc you can choose 'Keep on my iPhone'  Then once the Account is Deleted form the iPhone, log back in with the correct Apple ID and choose Merge when prompted.  This will merge your data from this iPhone with that iCloud account effectively putting your devices on the same account.

  • Mapping scenario with one customer and many invoices

    Hi!
    I am receiving a flat file with one customer and n-invoice records. The file could look like this:
    133133222051241480320842420500089
    13317350071000010806372131000                 0015890009876543X
    13317350071000010806372131000                 0015890009876543X
    13317350071000010806372131013                 0015890009876443X
    13317350071000010806372131014                 0015890009872343X
    where the first line is customer and the rest are invoices.
    My question is: How can I get this into a structure like this:
    <Customer>
      <...>
    <Invoice>
      <...>
    <Invoice>
      <...>
    When using content conversion (Customer,1,Invoice,*) there has to be specified a key field, which is not a possibility. Do I have to create a UDF for this to get the mapping I want?
    Thanks for your answers!
    regs Ole Mugaas

    Thanks for your answers!
    As I can see from the Krawcyk blog, he is only using one node and maybe that's why he don't have to use keyField. We have also come to the conclusion that we have to use keyfield, but then we have to give the fields out own values. This is not good enough since the values are predefined. Creating a "placeholder" in each data type would not be a good answer because then our customers have to send an additional field in the file which would cause a lot of work.
    Do you know how this can be solved with the keyfield? Is it possible to create a keyField and give it a value that is already there, ex:
    Invoice.keyFieldValue = Invoicenumber (where Invoicenumber is a field specified in the DT) ?
    regards Ole

Maybe you are looking for

  • IPod Nano 6th Gen Won't Sync After Getting iTunes Match

    I have a new iPod Nano 6th Generation, with the new clock faces, bought in december. I bought iTunes Match in early january. After syncing and using my ipod successfully, I was confident that iTunes Match would not effect my ipod syncing as I only us

  • Verizon, Does YOUR equipment need service and upgrade? Not OURS? Please Help

         Hello, I used to have a pretty stable dial-up, then DSL, then DSL enhanced connection. But late this summer, my service has gotten horrible. In the evenings, I can get disconnected every 1 to 2 minutes or even get disconnected when I am trying t

  • Samsung Printer ML 2010 doesn't work since Mountain Lion

    Upgraded to Mountian Lion today and now my Samsung printer (ML 2010) doesn't work.  I uninstalled it then added it back and still doesn't work. 

  • Peformace issue

    Hello Friends,       I have doubt in implementing a pgm, it goes like this,, i need to selecta a main materials which was created today and then i need to create a sub-material which has the same plant, sales org,Dist.ch.And some modifications should

  • Downloading iPhoto

    I have an iMac and have the downloaded the trial version of Elements 9. My version of iPhoto is 08 and the Elements Organiser states that I must update to version 8.0 to 8.1.2. Not sure if this means I must update to 09 version. Can any one advise? I