Standby db scenario

Hi,
Oracle version is 10.2.0.3. I am into the process of creating a physical standby. Both primary as well as standby are kept in same IDC connected by 1Gbps link.
Client doesnt want primary db to get affected in case if, because of any reason, standby goes down or link goes down. In this situation the only choice I have is to configure primary db in maximum performance mode.
In this scenario what is advisable for redo transport. ARCn or LGWR (with SYNC or ASYNC).
My confusion is we can have multiple ARCn processes in Database server but we have only one LGWR, so aint I putting extra load(which may lead to poor primary db performance)on primary db if I choose LGWR for log transfer. At the same time if I go for ARCn, I feel like I am not making use of high bandwidth between two db's.
Kindly have your recommendations/suggestions on this
Regards,
Amit

Hi Amit,
My suggestion would be LGWR with ASYNC.
In that case the impact to your primary database is lowest.
In this configuration, LGWR writes redo data from log buffer to online redo logs.
Network Server Process, LNS reads the online redo logs and ships the redo streams to Remote File Server Process (RFS) on standby site.
RFS takes the redo stream and updates the Active Standby Redo Log.
Be aware that Standby Redo Log is different than regular Redo Logs.
http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/standby.htm#i72459
One a Standby Redo Log logswitch occurs, it archives the active Standby Redo Log. Once archived it is ready to be used by MRP (Managed Recovery Process) process which applies it to the standby database.
LGWR (SYNC or ASYNC) configuration also reduces the possibility of loosing data in a case of failover.
Read the following documents to learn more about the redo transport services:
Oracle® Data Guard Concepts and Administration
http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/log_transport.htm
Data Guard Redo Transport & Network Best Practices Oracle Database 10g Release 2
(http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_10gR2_DataGuardNetworkBestPractices.pdf)
Cheers,

Similar Messages

  • Oracle Standby DB with SAP - BRARCHIVE and BRBACKUP

    Hi experts,
    I was requested to configure and standby Oracle scenario for with SAP. After reviewing many different documentation regarding this, I would like to confirm the following with youu2026How do the BRARCHIVE and BRBACKUP tools should be integrated in this standby/data guard configuration? Actually data guard configuration has been already completed and redo logs are being shipping and applying successfully from primary db to the standby db. Please note that no settings for BRACRHIVE or BRBACKUP have been done so far.
    From doc. u201COracle White Paperu2014Setting up Oracle 11g Data Guard for SAP customersu201D, it says that u201CThere is no change to BRARCHIVE configuration necessary. It can work like with any other standolane databaseu201D. What changes or considerations would be required on the current data guard configuration to implement these BRTOOLS? For exampleu2026If I start using BRARCHIVE to copy the offline redo log files from the production database to the standby database, its necessary to disable anything from the current configuration?
    Hope I was clear enoughu2026thanks for your help.
    Regards

    Hi Orkun,
    If the use of BRARCHIVE is not required or at least is not mandatory to configure an standby db for SAP environments...I've started thinking If I've followed the right process to complete this?
    As mentioned if my post the following documentation was used "Oracle White Paperu2014Setting up Oracle 11g Data Guard for SAP customers". Did I follwed the right process or is there any different set of steps to configure Oracle standby DB for SAP using BRARCHIVE and BRBACKUP?
    As you can see I'm confused about the Standby DB configuration process for SAP environments.
    Any clarification you can provide me regarding this would be highly appreciated.
    Regards

  • Modifying default tablespace + implications

    Suppose an Oracle user has his default tablespace set to Tablespace_Old and there are already objects stored there.
    Now suppose we want to alter the user such that this default tablespace is set to Tablespace_New. What happens to the objects which were stored in Tablespace_Old? Will they still be accessible from Tablespace_New or there will be the need to migrate the old data to the new tablespace?
    Moreover, in the case of a standby database scenario, will the standby database adopt such changes (i.e. will the user be assigned to tablespace Tablespace_New?)?
    thanks for your time.

    The default tablespace is only used to assign a location to build an object if a tablespace is not defined. The old objects will still be accessable. The more important thing is what quota's are assigned for the user on the tablespace. If the Quota is none, the user can't write in the tablespace. If it is unlimited, they can write anything and if the quota is 20 meg, then they can't create more then 20 meg of objects and rows.

  • Logical Standby Scenario..!!

    Hello,
    We are having dataguard setup with Primary, Physical standby and Logical Standby databases located on three different machines..
    Oracle Version : 10.2.0.1
    OS : RHEL 4 on (x86 32-Bit)
    Following are our initialization parameters for all the three instances..
    Primary Database Parameters--_
    *.db_unique_name='betapri'
    *.fal_client='BETAPRI'
    *.fal_server='BETAPHYSTDBY'
    *.instance_name='oracle'
    *.log_archive_config='DG_CONFIG=(betapri,betaphystdby,betalogstdby)'
    *.log_archive_dest_1='LOCATION=/u01/app/oracle/archive/oracle VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=betapri MANDATORY'
    *.log_archive_dest_2='SERVICE=BETAPHYSTDBY LGWR VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=betaphystdby'
    *.log_archive_dest_3='SERVICE=BETALOGSTDBY LGWR SYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=betalogstdby'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_dest_state_3='ENABLE'
    *.standby_archive_dest='/u01/app/oracle/archive/oracle'
    *.standby_file_management='AUTO'
    *.remote_login_passwordfile='EXCLUSIVE'
    Physical standby Database Parameters--_
    *.db_unique_name='betaphystdby'
    *.fal_client='BETAPHYSTDBY'
    *.fal_server='BETAPRI'
    *.instance_name='oracle'
    *.log_archive_config='DG_CONFIG=(betapri,betaphystdby,betalogstdby)'
    *.log_archive_dest_1='LOCATION=/u01/app/oracle/archive/oracle VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=betaphystdby'
    *.log_archive_dest_2='SERVICE=BETAPRI LGWR VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=betapri'
    *.log_archive_dest_3='SERVICE=BETALOGSTDBY VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=betalogstdby'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_dest_state_3='ENABLE'
    *.standby_archive_dest='/u01/app/oracle/archive/oracle'
    *.standby_file_management='AUTO'
    *.remote_login_passwordfile='EXCLUSIVE'*Logical Standby Database Parameters *_
    *.db_unique_name='betalogstdby'
    *.instance_name='oracle'
    *.log_archive_config='DG_CONFIG=(betapri,betaphystdby,betalogstdby)'
    *.log_archive_dest_1='LOCATION=/u01/app/oracle/archive/betalogstdby VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=betalogstdby'
    *.log_archive_dest_2='LOCATION=/u01/app/oracle/archive/oracle VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=betalogstdby'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.standby_archive_dest='/u01/app/oracle/archive/oracle'
    *.standby_file_management='AUTO'
    *.remote_login_passwordfile='EXCLUSIVE'Whole data guard setup is working properly as per the configuration.. But just for testing purpose i did FAILOVER_ and made our physical standby as new primary database and using Flashback i made my old primary as new physical standby and haven't made any changes to logical standby database since as per my init parameters i guess logical standby should remain in sync and keep applying the logs..
    But performing this scenario logical standby seems to be messed up.. My new primay database does generate archives to logical standby but logical standby logmnr doesn't apply the changes to the database.. So once i issue "ALTER DATABASE START LOGICAL STANDBY APPLY" following message gets displayed in alert log *"Fatal Error: LogMiner processed beyond new branch scn."*
    Tue Sep 29 18:55:14 2009
    alter database start logical standby apply
    Tue Sep 29 18:55:14 2009
    ALTER DATABASE START LOGICAL STANDBY APPLY (oracle)
    Tue Sep 29 18:55:14 2009
    No optional part
    Attempt to start background Logical Standby process
    LSP0 started with pid=16, OS id=18308
    LOGSTDBY status: ORA-16111: log mining and apply setting up
    Tue Sep 29 18:55:14 2009
    LOGMINER: Parameters summary for session# = 1
    LOGMINER: Number of processes = 3, Transaction Chunk Size = 201
    LOGMINER: Memory Size = 30M, Checkpoint interval = 150M
    Tue Sep 29 18:55:14 2009
    Fatal Error: LogMiner processed beyond new branch scn.
    LOGSTDBY status: ORA-01346: LogMiner processed redo beyond specified reset log scn
    Tue Sep 29 18:55:14 2009
    Errors in file /u01/app/oracle/admin/oracle/bdump/oracle_lsp0_18308.trc:
    ORA-01346: LogMiner processed redo beyond specified reset log scn
    LOGSTDBY status: ORA-16222: automatic Logical Standby retry of last action
    LOGSTDBY status: ORA-16111: log mining and apply setting up
    Tue Sep 29 18:55:14 2009
    LOGMINER: Parameters summary for session# = 1
    LOGMINER: Number of processes = 3, Transaction Chunk Size = 201
    LOGMINER: Memory Size = 30M, Checkpoint interval = 150M
    Tue Sep 29 18:55:14 2009
    Fatal Error: LogMiner processed beyond new branch scn.
    LOGSTDBY status: ORA-01346: LogMiner processed redo beyond specified reset log scn
    Tue Sep 29 18:55:14 2009
    Errors in file /u01/app/oracle/admin/oracle/bdump/oracle_lsp0_18308.trc:
    ORA-01346: LogMiner processed redo beyond specified reset log scn
    Tue Sep 29 18:55:14 2009
    ORA-16210 signalled during: alter database start logical standby apply... Note : I am having flashback enabled only on one database which was my old primary..
    Please any comments, help or suggestions would be great..
    Thanks - HP

    Appreciate your kind reply Robert..
    Would it be the only option left for me to re instantiate the logical standby..? Or what if i enable flashback on logical standby and then try to flashback till "standby_became_primary_scn" and then try to start logical standby apply.. Will this work in my scenario?
    Any suggestions or comments are most welcome.
    Thanks - HP

  • Test scenario for Physical Standby

    Dear Team,
    Could you please share the Test scenario for Physical Standby.
    1. Read only.
    2. Switch Over
    3. Fail Over.
    Many Thanks in Advance
    Arjun.

    You didnt mention the version.
    Oracle10g Release 2, Data Guard Switchover and Failover best practices
    http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_10gR2_SwitchoverFailoverBestPractices.pdfOracle9i, Data Guard Switchover & Failover best practices
    http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_9iSwitchoveFailoverBestPractices.pdfKhurram

  • Switchover scenario in a cascaded standby setup

    Hi,
    I have a setup of primary and standby databases in three locations A, B, C as below:
    Database version: 11.1.7.0
    In location A I have the primary database that ships redo log to a physical standby at location B alone.
    In location B I have a physical standby that replicates to site C.
    In site C I have the final physical standby that receives its redo logs directly from site B. In essence, this is a cascaded standby database.
    My question is about the switchover:
    Currently, both the instances in sites A and B have the LOG_ARCHIVE_DEST_3 set to point to site C and this is valid for standby role alone.
    Which means in case the two sites - A & B reverse roles, site C will continue receiving redo logs from either A or B, depending on which one is the new standby.
    However, the doubt comes in if site C were to become the new primary instance. I suspect that I would need to add two parameters LOG_ARCHIVE_DEST_2 and LOG_ARCHIVE_DEST_3 and point them to sites A and B respectively, and make them valid only for the primary role. In this case, if site C becomes the new primary, then both sites A and B can continue receiving redo logs from site C.
    Anyone who has ever implemented this?
    Regards,
    dula

    The scenario in the doc. may also help you:
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28294/cascade_appx.htm#i638620
    Example E-1 Sample Use of Initialization Parameters in Cascaded Destinations
    Boston Database (Primary Role)
    DB_UNIQUE_NAME=boston
    REMOTE_LOGIN_PASSWORDFILE=EXCLUSIVE
    LOG_ARCHIVE_CONFIG='DG_CONFIG=(chicago,boston,denver)'
    LOG_ARCHIVE_DEST_1='LOCATION=/arch1/boston/ VALID_FOR=(ALL_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=boston'
    LOG_ARCHIVE_DEST_2= 'SERVICE=denver VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=denver'
    LOG_ARCHIVE_DEST_3= 'SERVICE=chicago VALID_FOR= (ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=chicago'
    Chicago Database (Standby Role)
    DB_UNIQUE_NAME=chicago
    LOG_ARCHIVE_CONFIG= 'DG_CONFIG=(chicago,boston,denver)'
    REMOTE_LOGIN_PASSWORDFILE=EXCLUSIVE
    LOG_ARCHIVE_DEST_1= 'LOCATION=/arch1/chicago/ VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=chicago'
    LOG_ARCHIVE_DEST_2= 'SERVICE=denver VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=denver'
    LOG_ARCHIVE_DEST_3= 'SERVICE=boston VALID_FOR= (ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=boston'
    Denver Database (Standby Role)
    DB_UNIQUE_NAME=denver
    LOG_ARCHIVE_CONFIG= 'DG_CONFIG=(chicago,boston,denver)'
    REMOTE_LOGIN_PASSWORDFILE=EXCLUSIVE
    LOG_ARCHIVE_DEST_1= 'LOCATION=/arch1/denver/ VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=denver'
    LOG_ARCHIVE_DEST_2= 'LOCATION=/arch2/denver/ VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=denver'
    Both the Boston primary database and the Chicago physical standby database define the LOG_ARCHIVE_DEST_2 initialization parameter as SERVICE=denver VALID_FOR=(STANDBY_LOGFILES, STANDBY_ROLE). Hence, even if the Boston and Chicago databases switch roles, the redo data will continue to be forwarded to the Denver database. Remember, as part of the original setup of the physical standby database, you should define a local destination, VALID_FOR=(ALL_LOGFILES, PRIMARY_ROLE), that will be used for local archiving when the physical standby database transitions to the primary role.

  • Restore Scenarios when backup was taken on Standby exadata cluster through zfs

    I have a question here.
    We have a dataguard environment in our exadata clusters.  Primary resides on full Rack and standby on half rack. Both x4 and backup are disks to disks through zfs.
    The backups are full once a week and incremental afterwards , happens on the standby side per oracle  best practice.
    However we try to test restore scenarios now and how do we restore the backups to primary?
    Thanks in advance.

    Hi..
    To validate the RMAN backup you need to restore it at some new host server.You can go through
    [http://download.oracle.com/docs/cd/B19306_01/backup.102/b14191/rcmrecov.htm#BRADV908]
    You can even use the RMAN validation command.
    Anand

  • Physical Standby Database Lags Far Behind the Primary Database scenario

    Hi,
    Have anyone successfully implemented described in the Oracle Documentation following
    scenario?
    For me it does not work at all.
    After implementing all steps current_scn of standby database remains the same and when trying to push recovery using redo it asks for non-existing archived log which I can't restore on primary database as all backups containing that archived log already gone.
    The only option is to recreate standby database from backup again.
    Best Regards,
    Alex

    I already found the reason.
    Just for your information:
    Recovering a Standby database from a missing archivelog
    Unfortunately documentation missed that important step of standby controlfile recreation.
    To my luck I was copying the whole backup to the standby server when found this link.

  • Changing from Active to Standby in a DHCP Failover scenario (on the fly) without deconfiguring operation...

    Hi all;
    Suppose I have set up a scope in SRV01 to be in Active mode. Is it possible to change its state to Standby mode (on the fly) without deconfigurig it and then configuring it again with my desired configuration?
    Thanks
    Please VOTE as HELPFUL if the post helps you and remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading
    the thread.

    Hi all;
    Suppose I have set up a scope in SRV01 to be in Active mode. Is it possible to change its state to Standby mode (on the fly) without deconfigurig it and then configuring it again with my desired configuration?
    Thanks
    Please VOTE as HELPFUL if the post helps you and remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading
    the thread.

  • How Standby recover in this scenario

    Hi All,
    I have a question in my mind related to Dataguard.
    I have Dataguard Configuration containing one primary and one standby.
    Support i start the heavy update in primary which causes the heavy redo and undo to generated in primary. And after some time log get switch. redo data transfer to standby site and get applied.
    But after some time my primary database get crash and update statement didn't commit. So when i will open my primary db instance recovery will happen. and my uncommitted update statement will get rollback with the help of undo segment.
    But since standby doesn't have undo record how the standby recover this situation.
    Thanks
    Jamsher

    As Uwe said standby has the necessary undo data.
    When a crash recovery is performed on the primary, that crash recovery changes are also sent to standby via redo log data.
    Also in RAC, when one of the instances crashes, another instance performs crash recovery on the failed instance's redo and ships the final archivelog to standby.

  • Standby database upgrade scenario

    Hi all
    I have to databases. One 10.1.3 and second 10.1.5
    I created standby database for 10.1.3 database. And now want to create second standby database for my second 10.1.5 database...
    How I must do it?
    Can I convert (or how can I convert) my installed 10.1.3 standby database to preapre creating another 10.1.5 standby database?
    Thanks...

    I have to say that's not very clear for me.
    You have one 10.1.3 database with a standby.
    You have one 10.1.5 database you want to build a standby on.
    Right ?
    So, why do you need to "convert" the 10.1.3 standby to 10.1.5 ?
    Keep in mind, it's better having the standby database in same verion/release/patchset as the primary db.
    Nicolas.

  • Best practice on using Flashback and Logical Standby

    Hello,
    I'm testing a fail-back scenario where I first need to activate a logical standby, then do some dummy transactions before I flashback this db and resme the redo apply. Here is what the steps look like:
    1)     Ensure logical standby is in-sync with primary
    2)     Enable flashback on standby
    3)     Create a flashback guaranteed restore point
    4)     Defer log shipping from primary
    5)     Activate the logical standby so it’s fully open to read-write
    6)     Dummy activities against the standby (which is now fully open)
    7)     Flashback the database to the guaranteed checkpoint
    8)     Resume log shipping on primary
    9)     Resume redo apply on secondary
    In the end, i can see the log shipping is happening but the logical standby does not apply any of these..and there is no error in the alert log on Standby side. But the following query could explains why the standby is idle:
    SELECT TYPE, HIGH_SCN, STATUS FROM V$LOGSTDBY;
    TYPE HIGH_SCN STATUS
    COORDINATOR ORA-16240: Waiting for log file (thread# 2, sequence# 0)
    ORA-16240: Waiting for log file (thread# string, sequence# string)
    Cause: Process is idle waiting for additional log file to be available.
    Action: No action necessary. This informational statement is provided to record the event for diagnostic purposes.
    I dont understand why it's looking for sequence #0 after the flashback.
    Thanks for the help.

    Hello;
    I hesitate to answer your question because you are not doing a good job of keeping the forum clean :
    Total Questions: 13 (13 unresolved)
    Please consider closing some of you old answered questions and rewarding those who helped you.
    No action necessary.
    Do you really have a thread 2? ( Redo thread number )
    Quick check
    select applied_scn,latest_scn from v$logstdby_progress;Use the DBA_LOGSTDBY_LOG View if you don't have a thread 2 then the sequence# is meaningless.
    COLUMN DICT_BEGIN FORMAT A10;
    SELECT FILE_NAME, SEQUENCE#, FIRST_CHANGE#, NEXT_CHANGE#,
    TIMESTAMP, DICT_BEGIN, DICT_END, THREAD# AS THR# FROM DBA_LOGSTDBY_LOG
    ORDER BY SEQUENCE#;Logical Standby questions are difficult, not a lot of them out there I'm thinking.
    Check
    http://docs.oracle.com/cd/E14072_01/server.112/e10700/manage_ls.htm
    "Waiting On Gap State" ( However I still believe you don't have a 2nd thread# )
    OR
    http://psilt.wordpress.com/2009/04/29/simple-logical-standby/
    Best Regards
    mseberg
    Edited by: mseberg on Apr 26, 2012 5:13 PM

  • Data guard synchronization after link down b/w primary and physical standby

    Hi All,
    I have configured data guard on oracle 11gr2 db. Normally switchover between my primary and physical standby happens smoothly and the Apply lag would be zero. Recently We had to test a scenario when the network link between Primary and Physical standby is completely down and Physical standby is isolated completely for more than half an hour.
    When we brought up the link every thing worked smoothly but apply lag started increasing from 0 to around 3 hrs. And then it started reducing to 0. Currently Apply lag and transport lag shows 0.
    But is this normal behaviour of oracle data guard that when the link between primary and physical standby is completely down, It requires 3-4 hrs for resynchronization ??? Even when during isolation, there were very few transactions happend on primary database ??
    Are there any documents available for this scenario??
    Thanks

    Hi, after the link is up, if there were some transactions and produced archive logs it's normal to take some time for resync. To check if 3-4 hours is normal or not, you can repeat the scenario and this time check
    - how many archivelogs does primary produce in this period.
    - after the link is up, does archivelog transfer immediately starts from primary to standby? Is primary able to send these archivelogs parallel?
    - Is there anything wrong with the apply process?
    check primary & standby alert log files, and run this query on standby to check the transport and apply processes:
    SELECT PROCESS, STATUS, THREAD#, SEQUENCE#, BLOCK#, BLOCKS FROM V$MANAGED_STANDBY;
    regards

  • ORA-16191: Primary log shipping client not logged on standby.

    Hi,
    Please help me in the following scenario. I have two nodes ASM1 & ASM2 with RHEL4 U5 OS. On node ASM1 there is database ORCL using ASM diskgroups DATA & RECOVER and archive location is on '+RECOVER/orcl/'. On ASM2 node, I have to configure STDBYORCL (standby) database using ASM. I have taken the copy of database ORCL via RMAN, as per maximum availability architecture.
    Then I have ftp'd all to ASM2 and put them on FS /u01/oradata. Have made all necessary changes in primary and standby database pfile and then perform the duplicate database for standby using RMAN in order to put the db files in desired diskgroups. I have mounted the standby database but unfortunately, log transport service is not working and archives are not getting shipped to standby host.
    Here are all configuration details.
    Primary database ORCL pfile:
    [oracle@asm dbs]$ more initorcl.ora
    stdbyorcl.__db_cache_size=251658240
    orcl.__db_cache_size=226492416
    stdbyorcl.__java_pool_size=4194304
    orcl.__java_pool_size=4194304
    stdbyorcl.__large_pool_size=4194304
    orcl.__large_pool_size=4194304
    stdbyorcl.__shared_pool_size=100663296
    orcl.__shared_pool_size=125829120
    stdbyorcl.__streams_pool_size=0
    orcl.__streams_pool_size=0
    *.audit_file_dest='/opt/oracle/admin/orcl/adump'
    *.background_dump_dest='/opt/oracle/admin/orcl/bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='+DATA/orcl/controlfile/current.270.665007729','+RECOVER/orcl/controlfile/current.262.665007731'
    *.core_dump_dest='/opt/oracle/admin/orcl/cdump'
    *.db_block_size=8192
    *.db_create_file_dest='+DATA'
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='orcl'
    *.db_recovery_file_dest='+RECOVER'
    *.db_recovery_file_dest_size=3163553792
    *.db_unique_name=orcl
    *.fal_client=orcl
    *.fal_server=stdbyorcl
    *.instance_name='orcl'
    *.job_queue_processes=10
    *.log_archive_config='dg_config=(orcl,stdbyorcl)'
    *.log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST'
    *.log_archive_dest_2='SERVICE=stdbyorcl'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_format='%t_%s_%r.dbf'
    *.open_cursors=300
    *.pga_aggregate_target=121634816
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=364904448
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS'
    *.user_dump_dest='/opt/oracle/admin/orcl/udump'
    Standby database STDBYORCL pfile:
    [oracle@asm2 dbs]$ more initstdbyorcl.ora
    stdbyorcl.__db_cache_size=251658240
    stdbyorcl.__java_pool_size=4194304
    stdbyorcl.__large_pool_size=4194304
    stdbyorcl.__shared_pool_size=100663296
    stdbyorcl.__streams_pool_size=0
    *.audit_file_dest='/opt/oracle/admin/stdbyorcl/adump'
    *.background_dump_dest='/opt/oracle/admin/stdbyorcl/bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='u01/oradata/stdbyorcl_control01.ctl'#Restore Controlfile
    *.core_dump_dest='/opt/oracle/admin/stdbyorcl/cdump'
    *.db_block_size=8192
    *.db_create_file_dest='/u01/oradata'
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='orcl'
    *.db_recovery_file_dest='+RECOVER'
    *.db_recovery_file_dest_size=3163553792
    *.db_unique_name=stdbyorcl
    *.fal_client=stdbyorcl
    *.fal_server=orcl
    *.instance_name='stdbyorcl'
    *.job_queue_processes=10
    *.log_archive_config='dg_config=(orcl,stdbyorcl)'
    *.log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST'
    *.log_archive_dest_2='SERVICE=orcl'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_format='%t_%s_%r.dbf'
    *.log_archive_start=TRUE
    *.open_cursors=300
    *.pga_aggregate_target=121634816
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=364904448
    *.standby_archive_dest='LOCATION=USE_DB_RECOVERY_FILE_DEST'
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS'
    *.user_dump_dest='/opt/oracle/admin/stdbyorcl/udump'
    db_file_name_convert=('+DATA/ORCL/DATAFILE','/u01/oradata','+RECOVER/ORCL/DATAFILE','/u01/oradata')
    log_file_name_convert=('+DATA/ORCL/ONLINELOG','/u01/oradata','+RECOVER/ORCL/ONLINELOG','/u01/oradata')
    Have configured the tns service on both the hosts and its working absolutely fine.
    <p>
    ASM1
    =====
    [oracle@asm dbs]$ tnsping stdbyorcl
    </p>
    <p>
    TNS Ping Utility for Linux: Version 10.2.0.1.0 - Production on 19-SEP-2008 18:49:00
    </p>
    <p>
    Copyright (c) 1997, 2005, Oracle. All rights reserved.
    </p>
    <p>
    Used parameter files:
    </p>
    <p>
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.20.20)(PORT = 1521))) (CONNECT_DATA = (SID = stdbyorcl) (SERVER = DEDICATED)))
    OK (30 msec)
    ASM2
    =====
    </p>
    <p>
    [oracle@asm2 archive]$ tnsping orcl
    </p>
    <p>
    TNS Ping Utility for Linux: Version 10.2.0.1.0 - Production on 19-SEP-2008 18:48:39
    </p>
    <p>
    Copyright (c) 1997, 2005, Oracle. All rights reserved.
    </p>
    <p>
    Used parameter files:
    </p>
    <p>
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.20.10)(PORT = 1521))) (CONNECT_DATA = (SID = orcl) (SERVER = DEDICATED)))
    OK (30 msec)
    Please guide where I am missing. Thanking you in anticipation.
    Regards,
    Ravish Garg

    Following are the errors I am receiving as per alert log.
    ORCL alert log:
    Thu Sep 25 17:49:14 2008
    ARCH: Possible network disconnect with primary database
    Thu Sep 25 17:49:14 2008
    Error 1031 received logging on to the standby
    Thu Sep 25 17:49:14 2008
    Errors in file /opt/oracle/admin/orcl/bdump/orcl_arc1_4825.trc:
    ORA-01031: insufficient privileges
    FAL[server, ARC1]: Error 1031 creating remote archivelog file 'STDBYORCL'
    FAL[server, ARC1]: FAL archive failed, see trace file.
    Thu Sep 25 17:49:14 2008
    Errors in file /opt/oracle/admin/orcl/bdump/orcl_arc1_4825.trc:
    ORA-16055: FAL request rejected
    ARCH: FAL archive failed. Archiver continuing
    Thu Sep 25 17:49:14 2008
    ORACLE Instance orcl - Archival Error. Archiver continuing.
    Thu Sep 25 17:49:44 2008
    FAL[server]: Fail to queue the whole FAL gap
    GAP - thread 1 sequence 40-40
    DBID 1192788465 branch 665007733
    Thu Sep 25 17:49:46 2008
    Thread 1 advanced to log sequence 48
    Current log# 2 seq# 48 mem# 0: +DATA/orcl/onlinelog/group_2.272.665007735
    Current log# 2 seq# 48 mem# 1: +RECOVER/orcl/onlinelog/group_2.264.665007737
    Thu Sep 25 17:55:43 2008
    Shutting down archive processes
    Thu Sep 25 17:55:48 2008
    ARCH shutting down
    ARC2: Archival stopped
    STDBYORCL alert log:
    ==============
    Thu Sep 25 17:49:27 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-01017: invalid username/password; logon denied
    Thu Sep 25 17:49:27 2008
    Error 1017 received logging on to the standby
    Check that the primary and standby are using a password file
    and remote_login_passwordfile is set to SHARED or EXCLUSIVE,
    and that the SYS password is same in the password files.
    returning error ORA-16191
    It may be necessary to define the DB_ALLOWED_LOGON_VERSION
    initialization parameter to the value "10". Check the
    manual for information on this initialization parameter.
    Thu Sep 25 17:49:27 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-16191: Primary log shipping client not logged on standby
    PING[ARC0]: Heartbeat failed to connect to standby 'orcl'. Error is 16191.
    Thu Sep 25 17:51:38 2008
    FAL[client]: Failed to request gap sequence
    GAP - thread 1 sequence 40-40
    DBID 1192788465 branch 665007733
    FAL[client]: All defined FAL servers have been attempted.
    Check that the CONTROL_FILE_RECORD_KEEP_TIME initialization
    parameter is defined to a value that is sufficiently large
    enough to maintain adequate log switch information to resolve
    archivelog gaps.
    Thu Sep 25 17:55:16 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-01017: invalid username/password; logon denied
    Thu Sep 25 17:55:16 2008
    Error 1017 received logging on to the standby
    Check that the primary and standby are using a password file
    and remote_login_passwordfile is set to SHARED or EXCLUSIVE,
    and that the SYS password is same in the password files.
    returning error ORA-16191
    It may be necessary to define the DB_ALLOWED_LOGON_VERSION
    initialization parameter to the value "10". Check the
    manual for information on this initialization parameter.
    Thu Sep 25 17:55:16 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-16191: Primary log shipping client not logged on standby
    PING[ARC0]: Heartbeat failed to connect to standby 'orcl'. Error is 16191.
    Please suggest where I am missing.
    Regards,
    Ravish Garg

  • Bounce the standby database

    Hello,
    I am using oracle10g with standby database. Here is the database version.
    Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64bi
    PL/SQL Release 10.2.0.3.0 - Production
    CORE 10.2.0.3.0 Production
    TNS for Linux: Version 10.2.0.3.0 - Production
    NLSRTL Version 10.2.0.3.0 - Production
    Now i want to increase the number of process in primary database. Primary needs downtime. But in this circumstances, should i bounce the standby?
    Let me put all my questions here.
    1. Should i bounce the standby at this scenario?
    2. Here are the commands and sequence of steps to shutdown and start the database. Please correct me if i am wrong....
    2a) shutdown the standby database first. (run this command in secondary)
    connect / as sysdba
    SQL>ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;
    SQL>shutdown immediate
    2b) shutdown the primary(run this command in primary)
    SQL> ALTER SYSTEM SET PROCESS=999 SOPE=spfile;
    SQL> SHUTDOWN IMMEDIATE
    2c) startup the primary
    SQL> STARTUP
    2d) Start the secondary database
    connect / as sysdba
    startup nomount
    ALTER DATABASE MOUNT STANDBY DATABASE;
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;
    3) After start the database, how do i ensure, primary is shipping the archivelog to secondary?
    Any help is highly appreciated...

    +1. Should i bounce the standby at this scenario?+
    Not required.
    +2. Here are the commands and sequence of steps to shutdown and start the database. Please correct me if i am wrong....+
    +2a) shutdown the standby database first. (run this command in secondary)+
    connect / as sysdba
    SQL>ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;
    SQL>shutdown immediate
    +2b) shutdown the primary(run this command in primary)+
    SQL> ALTER SYSTEM SET PROCESS=999 SOPE=spfile;
    SQL> SHUTDOWN IMMEDIATE
    +2c) startup the primary+
    SQL> STARTUP
    +2d) Start the secondary database+
    connect / as sysdba
    startup nomount
    ALTER DATABASE MOUNT STANDBY DATABASE;------------------------------------------------------------------------------
    You can just mount
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;
    +3) After start the database, how do i ensure, primary is shipping the archivelog to secondary?+
    Check the alert log on the standby.
    Instead you need not shut down the standby at all.
    Edited by: user13403205 on Aug 31, 2010 9:04 AM

Maybe you are looking for