Rac到单实例的dataguard,想由物理standby转为逻辑standby,在执行转换的时候挂住,一直没有反应,详细见内,求教,谢谢

【primary db】 :
*.service_names='zxin'
*.log_archive_config='dg_config=(zxin_std,zxin)'
*.log_archive_dest_2='service=zxin_std lgwr async noaffirm valid_for=(online_logfiles,primary_role) db_unique_name=zxin_std'
*.db_file_name_convert='/zxindata/racdata','+DATA/zxin/datafile','/zxindata/racdata','+DATA/zxin/tempfile','/zxindata/racdata','+DATA'
*.log_file_name_convert='/zxindata/racdata','+DATA/zxin/onlinelog'
*.standby_file_management=auto
*.fal_server='zxin_std'
*.fal_client='zxin'
【standby db】
*.log_archive_config='dg_config=(zxin_std,zxin)'
*.standby_archive_dest='/zxinbak/arch'
*.log_archive_dest_1='location=/zxinbak/arch'
*.log_archive_dest_2='service=zxin valid_for=(online_logfiles,primary_role) db_unique_name=zxin'
*.db_file_name_convert='+DATA/zxin/datafile','/zxindata/racdata','+DATA/zxin/tempfile','/backup','+DATA', '/zxindata/racdata'
*.log_file_name_convert='+DATA/zxin/onlinelog','/zxindata/racdata'
*.standby_file_management=auto
*.fal_server='zxin'
*.fal_client='zxin_std'
*.db_name='zxin'
*.dispatchers='(PROTOCOL=TCP) (SERVICE=zxin_stdXDB)'
*.db_unique_name='zxin_std'
*.service_names='zxin_std'
转换的步骤:
1)在primary db上执行:
execute dbms_logstdby.build;
2)在standby db上执行:
alter database recover managed standby database cancel;
ALTER DATABASE RECOVER TO LOGICAL STANDBY zxin_std; --一直挂死                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

追加转换期间的alert日志:
Wed Nov 14 01:59:26 EST 2012
MRP0: Background Media Recovery process shutdown (zxin_std)
Wed Nov 14 01:59:27 EST 2012
Managed Standby Recovery Canceled (zxin_std)
Wed Nov 14 01:59:27 EST 2012
Completed: alter database recover managed standby database cancel
Wed Nov 14 01:59:29 EST 2012
RFS[4]: Archived Log: '/zxinbak/arch/2_110_793274616.dbf'
RFS[4]: Archived Log: '/zxinbak/arch/2_111_793274616.dbf'
Wed Nov 14 01:59:32 EST 2012
ALTER DATABASE RECOVER TO LOGICAL STANDBY zxin_std
Wed Nov 14 01:59:32 EST 2012
Media Recovery Start: Managed Standby Recovery (zxin_std)
Wed Nov 14 01:59:32 EST 2012
Managed Standby Recovery not using Real Time Apply
parallel recovery started with 7 processes
Media Recovery Waiting for thread 1 sequence 76
Fetching gap sequence in thread 1, gap sequence 76-76
Wed Nov 14 02:00:04 EST 2012
FAL[client]: Failed to request gap sequence
GAP - thread 1 sequence 76-76
DBID 1448565237 branch 793274616
FAL[client]: All defined FAL servers have been attempted.
Check that the CONTROL_FILE_RECORD_KEEP_TIME initialization
parameter is defined to a value that is sufficiently large
enough to maintain adequate log switch information to resolve
archivelog gaps.
-------------------------------------------------------------

Similar Messages

  • RAC & DataGuard (Physical Standby)

    Hello all,
    I'm trying to get a high level overview of how RAC & DataGuard would behave in the following configuration. I've written down my understanding of how things would work. Please correct me if I'm wrong.
    1) 2 node RAC (Primary Database) with a single instance physical standby.
    a) Same standby related init.ora parameters would have to configured on both primary rac nodes.
    b) The redo apply service at the standby would merge the redo from the 2 threads from the primary and apply it to the standby to keep it in sync.
    c) During switch over only one primary RAC instance should be up besides the standby instance.
    d) During switch back again only one primary RAC instance should be up besides the standby instance.
    e) During failover ofcourse both primary instances would be down which warrants the fail over.
    2) 2 node RAC (Primary) with a 2 node physical standby
    This where it gets complex and I'm not really sure how a,b,c,d,e would be in this scenario. Appreciate if you could shed some light on this and list what
    a,b,c,d,e would look like in this scenario.
    I'm assuming that only one instance in the standby RAC should be up when the standby is in RAC configuration for the redo apply to work.
    Also, if there is a white paper that details step by step procedure for setting up the above 2 scenarios, please let me know. So far I was able to find only the MAA white paper but that was not very helpful. If you can prescribe a good book for RAC & DataGuard that would be great too..
    Thanks for your help

    >
    1) 2 node RAC (Primary Database) with a single instance physical standby.
    a) Same standby related init.ora parameters would have to configured on both primary rac nodes.Usually rac nodes share their spfile on a shared diskvolume or through ASM. So they have identical DG parameter anyway
    b) The redo apply service at the standby would merge the redo from the 2 threads from the primary and apply it to the standby to keep it in sync.Correct
    c) During switch over only one primary RAC instance should be up besides the standby instance.
    Sounds logical to me
    Edit: In fact during normal operation both RAC nodes are up, so during switchover they are both active too
    It is the configuration that matters. As soon as the standby is becoming primary, it might be that it talks only to one RAC node for its log-apply files
    d) During switch back again only one primary RAC instance should be up besides the standby instance.That single was the only one running in step c.
    e) During failover ofcourse both primary instances would be down which warrants the fail over.Correct. Keep in mind that during a failover the RAC config would be deconfigured from the DG setup and has to be added back again as standby.
    You might have to take a look at: http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_10g_RACPrimarySingleInstancePhysicalStandby.pdf
    >
    2) 2 node RAC (Primary) with a 2 node physical standby
    This where it gets complex and I'm not really sure how a,b,c,d,e would be in this scenario. Appreciate if you could shed some light on this and list what
    a,b,c,d,e would look like in this scenario.
    I'm assuming that only one instance in the standby RAC should be up when the standby is in RAC configuration for the redo apply to work.
    No. RAC primary to RAC Standby would mean both nodes are up on both side. PrimNode1 send its redo logs to StandbyNode1 and the same for PrimNode2 and StandbyNode2
    Have a look here: http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_10g_RACPrimaryRACPhysicalStandby.pdf
    HTH,
    FJFranken
    My Blog: http://managingoracle.blogspot.com
    Edited by: fjfranken on 16-jul-2010 1:13

  • RAC physical standby in dataguard

    Hi All
    How many nodes need to be on the mount state in the RAC physical standby ? for example if we have RAC of 4 nodes in the physical standby . do I startup all of them in mount state or only one and what about the other they should be open or we shut them down ?
    Thanks

    Hi,
    -- assuming version 10.2
    If your primary database is a rac all the instances will send the redo information but even if the standby database is also a rac only one will receive/apply the redo information.
    How many nodes need to be on the mount state in the RAC physical standby ? for example if we have RAC of 4 nodes in the physical standby . do I startup all of them in mount state or only one and what about the other they should be open or we shut them down ?
    The instances will be mounted, if the working instance fails another one will be able to do the job.
    Best regards
    Phil
    Edited by: Philippe Florent on Sep 15, 2010 12:05 AM -- completing the answer

  • DataGuard Standby 環境で db が open できない.

    DataGuard Standby 環境で db が open できなくて困っています。
    詳しい方いれば、原因と対処方法を教えていただきたいと思います。
    create した spfile が正確に読み込まれていないことが原因のようなのですが。。
    なぜ読み込めないのかが分からない状態です。
    pfile を読み込んだ起動は可能ですが、spfile を読み込んだ起動の前提条件とは、
    何なのでしょうか。単に整合性が取れていないだけなのか、それとも、、。
    [grid@osaka1 shell]$ asmcmd
    ASMCMD> ls
    DATA/
    FRA/
    ASMCMD> cd data
    ASMCMD> ls
    ASM/
    WEST/
    ASMCMD> ls -l
    Type  Redund  Striped  Time             Sys  Name
                                            Y    ASM/
                                            N    WEST/
    ASMCMD>
    ASMCMD> cd west
    ASMCMD> ls
    CONTROLFILE/
    DATAFILE/
    ONLINELOG/
    PARAMETERFILE/
    TEMPFILE/
    spfilewest.ora
    ASMCMD> ls -l
    Type           Redund  Striped  Time             Sys  Name
                                                     N    CONTROLFILE/
                                                     N    DATAFILE/
                                                     N    ONLINELOG/
                                                     N    PARAMETERFILE/
                                                     N    TEMPFILE/
                                                     N    spfilewest.ora => +DATA/WEST/PARAMETERFILE/spfile.257.824236121
    ASMCMD> pwd
    +data/west
    ASMCMD>
    ASMCMD> cd para*
    ASMCMD> ls -l
    Type           Redund  Striped  Time             Sys  Name
    PARAMETERFILE  MIRROR  COARSE   AUG 23 18:00:00  Y    spfile.257.824236121
    ASMCMD>
    ASMCMD> pwd
    +data/west/PARAMETERFILE
    ASMCMD> quit
    [grid@osaka1 shell]$
    [oracle@osaka1 dbs]$ more initHPYMUSIC.ora
    SPFILE='+DATA/west/spfilewest.ora'
    [oracle@osaka1 dbs]$
    よろしくお願い致します。
    ps.
    ORA-12154 は整合性の問題であるので、それを合わせれば消えると思っています。
    そもそも RAC を前提としていたのですが、それを standalone に置き換えて検証始めた結果、
    こうなってしまっています。
    open できない原因が ORA-12154 だったりして。。
    ■ プライマリの場合
    ○ open するほうは、シンプルに以下だけで open することが確認できる。
       が、db_name を変更した関係で
       「ORA-12154: TNS: 指定された接続識別子を解決できませんでした」
       が出続けている。
       ORA-12154 は db が open できない原因とは無関係かもしれない。
    srvctl stop database -d east -f
    srvctl start database -d east -o open
    srvctl config database -d east
    srvctl status database -d east
    ○ 参考出力
    set linesize 500 pages 0
    col value for a90
    col name for a50
    select name, value
    from v$parameter
    where name in ('db_name','db_unique_name','log_archive_config', 'log_archive_dest_1','log_archive_dest_2',
                   'log_archive_dest_state_1','log_archive_dest_state_2', 'remote_login_passwordfile',
                   'log_archive_format','log_archive_max_processes','fal_server','db_file_name_convert',
                   'log_file_name_convert', 'standby_file_management');
    SQL>
    db_file_name_convert
    log_file_name_convert
    log_archive_dest_1
    log_archive_dest_2            SERVICE=HPYMUSIC SYNC NOAFFIRM VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=west
    log_archive_dest_state_1      enable
    log_archive_dest_state_2      enable
    fal_server
    log_archive_config
    log_archive_format            %t_%s_%r.dbf
    log_archive_max_processes     4
    standby_file_management       AUTO
    remote_login_passwordfile     EXCLUSIVE
    db_name                       HPYMUSIC
    db_unique_name                HPYMUSIC    ← ▼ db_name だけを変更したつもりが db_unique_name も変更されていた
    14行が選択されました。
    [oracle@tokyo1 shell]$ srvctl stop database -d east -f
    [oracle@tokyo1 shell]$ /u01/app/11.2.0/grid/bin/crsctl status resource -t
    NAME           TARGET  STATE        SERVER                   STATE_DETAILS
    Cluster Resources
    ora.east.db
          1        OFFLINE OFFLINE                               Instance Shutdown
    [oracle@tokyo1 shell]$ srvctl start database -d east -o open
    [oracle@tokyo1 shell]$ /u01/app/11.2.0/grid/bin/crsctl status resource -t
    NAME           TARGET  STATE        SERVER                   STATE_DETAILS
    Local Resources
    ora.DATA.dg
                   ONLINE  ONLINE       tokyo1
    ora.FRA.dg
                   ONLINE  ONLINE       tokyo1
    ora.LISTENER.lsnr
                   ONLINE  ONLINE       tokyo1
    ora.asm
                   ONLINE  ONLINE       tokyo1                   Started
    Cluster Resources
    ora.cssd
          1        ONLINE  ONLINE       tokyo1
    ora.diskmon
          1        ONLINE  ONLINE       tokyo1
    ora.east.db
          1        ONLINE  ONLINE       tokyo1                   Open ← ▼
    [oracle@tokyo1 shell]$
    [oracle@tokyo1 shell]$ srvctl config database -d east
    一意のデータベース名: east
    データベース名: east
    Oracleホーム: /u01/app/oracle/product/11.2.0/dbhome_1
    Oracleユーザー: grid
    spfile: +DATA/east/spfileeast.ora
    ドメイン:
    開始オプション: open
    停止オプション: immediate
    データベース・ロール: PRIMARY
    管理ポリシー: AUTOMATIC
    ディスク・グループ: DATA,FRA
    サービス:
    [oracle@tokyo1 shell]$ srvctl status database -d east
    データベースは実行中です。
    Fri Aug 23 19:44:10 2013
    Error 12154 received logging on to the standby
    Errors in file /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_arc2_7579.trc:
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    PING[ARC2]: Heartbeat failed to connect to standby 'HPYMUSIC'. Error is 12154.
    [oracle@tokyo1 dbs]$ pwd
    /u01/app/oracle/product/11.2.0/dbhome_1/dbs
    [oracle@tokyo1 dbs]$
    [oracle@tokyo1 dbs]$
    [oracle@tokyo1 dbs]$ more 2013.08.23_east_pfile.txt
    HPYMUSIC.__db_cache_size=301989888
    HPYMUSIC.__java_pool_size=4194304
    HPYMUSIC.__large_pool_size=8388608
    HPYMUSIC.__pga_aggregate_target=339738624
    HPYMUSIC.__sga_target=503316480
    HPYMUSIC.__shared_io_pool_size=0
    HPYMUSIC.__shared_pool_size=176160768
    HPYMUSIC.__streams_pool_size=0
    *.audit_file_dest='/u01/app/oracle/admin/east/adump'
    *.audit_trail='db'
    *.compatible='11.2.0.0.0'
    *.control_files='+DATA/east/controlfile/current.270.823277705','+FRA/east/controlfile/current.
    260.823277707'
    *.db_block_checking='TRUE'
    *.db_block_checksum='TRUE'
    *.db_block_size=8192
    *.db_create_file_dest='+DATA'
    *.db_domain=''
    *.db_name='HPYMUSIC'
    *.db_recovery_file_dest='+FRA'
    *.db_recovery_file_dest_size=3038773248
    *.diagnostic_dest='/u01/app/oracle'
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=HPYMUSICXDB)'
    *.log_archive_format='%t_%s_%r.dbf'
    *.memory_target=842006528
    *.nls_language='JAPANESE'
    *.nls_territory='JAPAN'
    *.open_cursors=300
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.standby_file_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    Fri Aug 23 19:49:38 2013
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 3
    Using LOG_ARCHIVE_DEST_1 parameter default value as USE_DB_RECOVERY_FILE_DEST
    ARCH: Warning; less destinations available than specified
    by LOG_ARCHIVE_MIN_SUCCEED_DEST init.ora parameter
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =27
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    Starting up:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options.
    Using parameter settings in server-side spfile /u01/app/oracle/product/11.2.0/dbhome_1/dbs/spfileHPYMUSIC.ora
    System parameters with non-default values:
      processes                = 150
      nls_language             = "JAPANESE"
      nls_territory            = "JAPAN"
      memory_target            = 804M
      control_files            = "+DATA/east/controlfile/current.270.823277705"
      control_files            = "+FRA/east/controlfile/current.260.823277707"
      db_block_checksum        = "TRUE"
      db_block_size            = 8192
      compatible               = "11.2.0.0.0"
      log_archive_dest_2       = "SERVICE=HPYMUSIC SYNC NOAFFIRM VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=west"
      log_archive_format       = "%t_%s_%r.dbf"
      db_create_file_dest      = "+DATA"
      db_recovery_file_dest    = "+FRA"
      db_recovery_file_dest_size= 2898M
      standby_file_management  = "AUTO"
      undo_tablespace          = "UNDOTBS1"
      db_block_checking        = "TRUE"
      remote_login_passwordfile= "EXCLUSIVE"
      db_domain                = ""
      dispatchers              = "(PROTOCOL=TCP) (SERVICE=HPYMUSICXDB)"
      audit_file_dest          = "/u01/app/oracle/admin/east/adump"
      audit_trail              = "DB"
      db_name                  = "HPYMUSIC"
      open_cursors             = 300
      diagnostic_dest          = "/u01/app/oracle"
    Fri Aug 23 19:49:39 2013
    PMON started with pid=2, OS id=8442
    Fri Aug 23 19:49:39 2013
    VKTM started with pid=3, OS id=8444 at elevated priority
    VKTM running at (10)millisec precision with DBRM quantum (100)ms
    Fri Aug 23 19:49:39 2013
    GEN0 started with pid=4, OS id=8448
    Fri Aug 23 19:49:39 2013
    DIAG started with pid=5, OS id=8450
    Fri Aug 23 19:49:39 2013
    DBRM started with pid=6, OS id=8452
    Fri Aug 23 19:49:39 2013
    PSP0 started with pid=7, OS id=8454
    Fri Aug 23 19:49:39 2013
    DIA0 started with pid=8, OS id=8456
    Fri Aug 23 19:49:39 2013
    MMAN started with pid=9, OS id=8458
    Fri Aug 23 19:49:39 2013
    DBW0 started with pid=10, OS id=8460
    Fri Aug 23 19:49:39 2013
    LGWR started with pid=11, OS id=8462
    Fri Aug 23 19:49:39 2013
    CKPT started with pid=12, OS id=8464
    Fri Aug 23 19:49:39 2013
    SMON started with pid=13, OS id=8466
    Fri Aug 23 19:49:39 2013
    RECO started with pid=14, OS id=8468
    Fri Aug 23 19:49:39 2013
    RBAL started with pid=15, OS id=8470
    Fri Aug 23 19:49:39 2013
    ASMB started with pid=16, OS id=8472
    Fri Aug 23 19:49:39 2013
    MMON started with pid=17, OS id=8474
    Fri Aug 23 19:49:39 2013
    MMNL started with pid=18, OS id=8478
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    NOTE: initiating MARK startup
    starting up 1 shared server(s) ...
    Starting background process MARK
    Fri Aug 23 19:49:39 2013
    MARK started with pid=20, OS id=8482
    NOTE: MARK has subscribed
    ORACLE_BASE not set in environment. It is recommended
    that ORACLE_BASE be set in the environment
    Reusing ORACLE_BASE from an earlier startup = /u01/app/oracle
    Fri Aug 23 19:49:39 2013
    ALTER DATABASE MOUNT
    NOTE: Loaded library: System
    SUCCESS: diskgroup DATA was mounted
    ERROR: failed to establish dependency between database HPYMUSIC and diskgroup resource ora.DATA.dg
    SUCCESS: diskgroup FRA was mounted
    ERROR: failed to establish dependency between database HPYMUSIC and diskgroup resource ora.FRA.dg
    Fri Aug 23 19:49:46 2013
    NSS2 started with pid=24, OS id=8572
    Successful mount of redo thread 1, with mount id 2951868947
    Database mounted in Exclusive Mode
    Lost write protection disabled
    Completed: ALTER DATABASE MOUNT
    ALTER DATABASE OPEN
    LGWR: STARTING ARCH PROCESSES
    Fri Aug 23 19:49:47 2013
    ARC0 started with pid=26, OS id=8574
    ARC0: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC0: STARTING ARCH PROCESSES
    Fri Aug 23 19:49:48 2013
    ARC1 started with pid=27, OS id=8576
    Fri Aug 23 19:49:48 2013
    ARC2 started with pid=28, OS id=8578
    ARC1: Archival started
    ARC2: Archival started
    ARC1: Becoming the 'no FAL' ARCH
    ARC1: Becoming the 'no SRL' ARCH
    ARC2: Becoming the heartbeat ARCH
    Fri Aug 23 19:49:48 2013
    ARC3 started with pid=29, OS id=8580
    LGWR: Setting 'active' archival for destination LOG_ARCHIVE_DEST_2
    ARC3: Archival started
    ARC0: STARTING ARCH PROCESSES COMPLETE
    Error 12154 received logging on to the standby
    Fri Aug 23 19:49:51 2013
    Errors in file /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_lgwr_8462.trc:
    ORA-12154: TNS: ?????????????????????
    Error 12154 for archive log file 2 to 'HPYMUSIC'
    LGWR: Failed to archive log 2 thread 1 sequence 8 (12154)
    Thread 1 advanced to log sequence 8 (thread open)
    Thread 1 opened at log sequence 8
      Current log# 2 seq# 8 mem# 0: +DATA/hpymusic/onlinelog/group_2.272.824213887
      Current log# 2 seq# 8 mem# 1: +FRA/hpymusic/onlinelog/group_2.262.824213889
    Successful open of redo thread 1
    Fri Aug 23 19:49:51 2013
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Fri Aug 23 19:49:51 2013
    SMON: enabling cache recovery
    Error 12154 received logging on to the standby
    Errors in file /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_arc2_8578.trc:
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    PING[ARC2]: Heartbeat failed to connect to standby 'HPYMUSIC'. Error is 12154.
    Archived Log entry 7 added for thread 1 sequence 7 ID 0xaff1210d dest 1:
    Error 12154 received logging on to the standby
    Errors in file /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_arc3_8580.trc:
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    FAL[server, ARC3]: Error 12154 creating remote archivelog file 'HPYMUSIC'
    FAL[server, ARC3]: FAL archive failed, see trace file.
    Errors in file /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_arc3_8580.trc:
    ORA-16055: FALリクエストが拒否されました。
    ARCH: FAL archive failed. Archiver continuing
    ORACLE Instance HPYMUSIC - Archival Error. Archiver continuing.
    Successfully onlined Undo Tablespace 2.
    Verifying file header compatibility for 11g tablespace encryption..
    Verifying 11g file header compatibility for tablespace encryption completed
    SMON: enabling tx recovery
    Database Characterset is AL32UTF8
    No Resource Manager plan active
    replication_dependency_tracking turned off (no async multimaster replication found)
    Starting background process QMNC
    Fri Aug 23 19:49:55 2013
    QMNC started with pid=32, OS id=8590
    Completed: ALTER DATABASE OPEN
    Fri Aug 23 19:49:59 2013
    Starting background process CJQ0
    Fri Aug 23 19:49:59 2013
    CJQ0 started with pid=33, OS id=8609
    Fri Aug 23 19:49:59 2013
    db_recovery_file_dest_size of 2898 MB is 6.38% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    [root@tokyo1 app]#
    [root@tokyo1 app]# more /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_arc2_8578.trc
    Trace file /u01/app/oracle/diag/rdbms/hpymusic/HPYMUSIC/trace/HPYMUSIC_arc2_8578.trc
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, Automatic Storage Management, OLAP, Data Mining
    and Real Application Testing options
    ORACLE_HOME = /u01/app/oracle/product/11.2.0/dbhome_1
    System name:    Linux
    Node name:      tokyo1.oracle11g.jp
    Release:        2.6.18-348.12.1.el5
    Version:        #1 SMP Wed Jul 10 05:28:41 EDT 2013
    Machine:        x86_64
    Instance name: HPYMUSIC
    Redo thread mounted by this instance: 1
    Oracle process number: 28
    Unix process pid: 8578, image: [email protected] (ARC2)
    *** 2013-08-23 19:49:51.707
    *** SESSION ID:(15.1) 2013-08-23 19:49:51.707
    *** CLIENT ID:() 2013-08-23 19:49:51.707
    *** SERVICE NAME:() 2013-08-23 19:49:51.707
    *** MODULE NAME:() 2013-08-23 19:49:51.707
    *** ACTION NAME:() 2013-08-23 19:49:51.707
    Redo shipping client performing standby login
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    *** 2013-08-23 19:49:51.972 4132 krsh.c
    Error 12154 received logging on to the standby
    *** 2013-08-23 19:49:51.972 869 krsu.c
    Error 12154 connecting to destination LOG_ARCHIVE_DEST_2 standby host 'HPYMUSIC'
    Error 12154 attaching to destination LOG_ARCHIVE_DEST_2 standby host 'HPYMUSIC'
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    *** 2013-08-23 19:49:51.973 4132 krsh.c
    PING[ARC2]: Heartbeat failed to connect to standby 'HPYMUSIC'. Error is 12154.
    *** 2013-08-23 19:49:51.973 2747 krsi.c
    krsi_dst_fail: dest:2 err:12154 force:0 blast:1
    *** 2013-08-23 19:50:49.816
    Redo shipping client performing standby login
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    *** 2013-08-23 19:50:50.070 4132 krsh.c
    Error 12154 received logging on to the standby
    *** 2013-08-23 19:50:50.070 869 krsu.c
    Error 12154 connecting to destination LOG_ARCHIVE_DEST_2 standby host 'HPYMUSIC'
    Error 12154 attaching to destination LOG_ARCHIVE_DEST_2 standby host 'HPYMUSIC'
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    *** 2013-08-23 19:50:50.070 4132 krsh.c
    PING[ARC2]: Heartbeat failed to connect to standby 'HPYMUSIC'. Error is 12154.
    *** 2013-08-23 19:50:50.070 2747 krsi.c
    krsi_dst_fail: dest:2 err:12154 force:0 blast:1
    *** 2013-08-23 19:51:51.147
    Redo shipping client performing standby login
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    *** 2013-08-23 19:51:51.403 4132 krsh.c
    Error 12154 received logging on to the standby
    *** 2013-08-23 19:51:51.403 869 krsu.c
    Error 12154 connecting to destination LOG_ARCHIVE_DEST_2 standby host 'HPYMUSIC'
    Error 12154 attaching to destination LOG_ARCHIVE_DEST_2 standby host 'HPYMUSIC'
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    *** 2013-08-23 19:51:51.403 4132 krsh.c
    PING[ARC2]: Heartbeat failed to connect to standby 'HPYMUSIC'. Error is 12154.
    *** 2013-08-23 19:51:51.403 2747 krsi.c
    krsi_dst_fail: dest:2 err:12154 force:0 blast:1
    [root@tokyo1 app]#
    [grid@tokyo1 shell]$ ./grid_info_east-x.sh
    + export ORACLE_SID=+ASM
    + ORACLE_SID=+ASM
    + LOGDIR=/home/grid/log
    + PRIMARYDB=east_DGMGRL
    + STANDBYDB=
    + PASSWORD=dataguard
    + mkdir -p /home/grid/log
    ++ date +%y%m%d,%H%M%S
    + echo 'asm info,130823,195709'
    + sqlplus / as sysasm
    SQL*Plus: Release 11.2.0.1.0 Production on 金 8月 23 19:57:09 2013
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Automatic Storage Management option
    に接続されました。
    SQL> SQL>
    SYSDATE
    13-08-23
    SQL> SQL> SQL>
    NAME                                 TYPE                VALUE
    asm_diskgroups                       string              FRA
    asm_diskstring                       string              /dev/sd*1
    asm_power_limit                      integer             1
    asm_preferred_read_failure_groups    string
    audit_file_dest                      string              /u01/app/11.2.0/grid/rdbms/audit
    audit_sys_operations                 boolean             FALSE
    audit_syslog_level                   string
    background_core_dump                 string              partial
    background_dump_dest                 string              /u01/app/grid/diag/asm/+asm/+ASM/trace
    cluster_database                     boolean             FALSE
    cluster_database_instances           integer             1
    cluster_interconnects                string
    core_dump_dest                       string              /u01/app/grid/diag/asm/+asm/+ASM/cdump
    db_cache_size                        big integer         0
    db_ultra_safe                        string              OFF
    db_unique_name                       string              +ASM
    diagnostic_dest                      string              /u01/app/grid
    event                                string
    file_mapping                         boolean             FALSE
    filesystemio_options                 string              none
    ifile                                file
    instance_name                        string              +ASM
    instance_number                      integer             1
    instance_type                        string              asm
    large_pool_size                      big integer         12M
    ldap_directory_sysauth               string              no
    listener_networks                    string
    local_listener                       string
    lock_name_space                      string
    lock_sga                             boolean             FALSE
    max_dump_file_size                   string              unlimited
    memory_max_target                    big integer         272M
    memory_target                        big integer         272M
    nls_calendar                         string
    nls_comp                             string              BINARY
    nls_currency                         string
    nls_date_format                      string
    nls_date_language                    string
    nls_dual_currency                    string
    nls_iso_currency                     string
    nls_language                         string              AMERICAN
    nls_length_semantics                 string              BYTE
    nls_nchar_conv_excp                  string              FALSE
    nls_numeric_characters               string
    nls_sort                             string
    nls_territory                        string              AMERICA
    nls_time_format                      string
    nls_time_tz_format                   string
    nls_timestamp_format                 string
    nls_timestamp_tz_format              string
    os_authent_prefix                    string              ops$
    os_roles                             boolean             FALSE
    pga_aggregate_target                 big integer         0
    processes                            integer             100
    remote_listener                      string
    remote_login_passwordfile            string              EXCLUSIVE
    remote_os_authent                    boolean             FALSE
    remote_os_roles                      boolean             FALSE
    service_names                        string              +ASM
    sessions                             integer             172
    sga_max_size                         big integer         272M
    sga_target                           big integer         0
    shadow_core_dump                     string              partial
    shared_pool_reserved_size            big integer         6081740
    shared_pool_size                     big integer         0
    sort_area_size                       integer             65536
    spfile                               string              +DATA/asm/asmparameterfile/registry.253.823204697
    sql_trace                            boolean             FALSE
    statistics_level                     string              TYPICAL
    timed_os_statistics                  integer             0
    timed_statistics                     boolean             TRUE
    trace_enabled                        boolean             TRUE
    user_dump_dest                       string              /u01/app/grid/diag/asm/+asm/+ASM/trace
    workarea_size_policy                 string              AUTO
    ++ date +%y%m%d,%H%M%S
    + echo 'asmcmd info,130823,195709'
    + asmcmd ls -l
    State    Type    Rebal  Name
    MOUNTED  NORMAL  N      DATA/
    MOUNTED  NORMAL  N      FRA/
    + asmcmd ls -l 'data/asm/*'
    Type              Redund  Striped  Time             Sys  Name
    ASMPARAMETERFILE  MIRROR  COARSE   AUG 11 19:00:00  Y    REGISTRY.253.823204697
    + asmcmd ls -l 'data/east/*'
    Type           Redund  Striped  Time             Sys  Name
    +data/east/CONTROLFILE/:
    CONTROLFILE    HIGH    FINE     AUG 12 15:00:00  Y    Current.260.823276231
    CONTROLFILE    HIGH    FINE     AUG 23 19:00:00  Y    Current.270.823277705
    +data/east/DATAFILE/:
    DATAFILE       MIRROR  COARSE   AUG 12 15:00:00  Y    SYSAUX.257.823276133
    DATAFILE       MIRROR  COARSE   AUG 23 19:00:00  Y    SYSAUX.267.823277615
    DATAFILE       MIRROR  COARSE   AUG 12 15:00:00  Y    SYSTEM.256.823276131
    DATAFILE       MIRROR  COARSE   AUG 23 19:00:00  Y    SYSTEM.266.823277615
    DATAFILE       MIRROR  COARSE   AUG 12 15:00:00  Y    UNDOTBS1.258.823276133
    DATAFILE       MIRROR  COARSE   AUG 23 19:00:00  Y    UNDOTBS1.268.823277615
    DATAFILE       MIRROR  COARSE   AUG 12 15:00:00  Y    USERS.259.823276133
    DATAFILE       MIRROR  COARSE   AUG 23 19:00:00  Y    USERS.269.823277615
    +data/east/ONLINELOG/:
    ONLINELOG      MIRROR  COARSE   AUG 12 15:00:00  Y    group_1.261.823276235
    ONLINELOG      MIRROR  COARSE   AUG 12 15:00:00  Y    group_2.262.823276241
    ONLINELOG      MIRROR  COARSE   AUG 12 15:00:00  Y    group_3.263.823276247
    +data/east/PARAMETERFILE/:
    PARAMETERFILE  MIRROR  COARSE   AUG 23 12:00:00  Y    spfile.265.823277967
    +data/east/TEMPFILE/:
    TEMPFILE       MIRROR  COARSE   AUG 12 15:00:00  Y    TEMP.264.823276263
    TEMPFILE       MIRROR  COARSE   AUG 23 19:00:00  Y    TEMP.274.823277733
                                                     N    spfileeast.ora => +DATA/EAST/PARAMETERFILE/spfile.265.823277967
    + asmcmd ls -l 'fra/east/*'
    Type         Redund  Striped  Time             Sys  Name
    +fra/east/ARCHIVELOG/:
                                                   Y    2013_08_12/
                                                   Y    2013_08_15/
                                                   Y    2013_08_19/
                                                   Y    2013_08_22/
                                                   Y    2013_08_23/
    +fra/east/CONTROLFILE/:
    CONTROLFILE  HIGH    FINE     AUG 12 15:00:00  Y    Current.256.823276231
    CONTROLFILE  HIGH    FINE     AUG 23 19:00:00  Y    Current.260.823277707
    +fra/east/ONLINELOG/:
    ONLINELOG    MIRROR  COARSE   AUG 12 15:00:00  Y    group_1.257.823276237
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_10.272.823535727
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_11.273.823535737
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_12.274.823535745
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_13.275.823535757
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_14.276.823535763
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_15.277.823535771
    ONLINELOG    MIRROR  COARSE   AUG 12 15:00:00  Y    group_2.258.823276245
    ONLINELOG    MIRROR  COARSE   AUG 12 15:00:00  Y    group_3.259.823276251
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_7.269.823535685
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_8.270.823535695
    ONLINELOG    MIRROR  COARSE   AUG 23 19:00:00  Y    group_9.271.823535703
    +fra/east/STANDBYLOG/:
                                                   N    standby_group_07.log => +FRA/EAST/ONLINELOG/group_7.269.823535685
                                                   N    standby_group_08.log => +FRA/EAST/ONLINELOG/group_8.270.823535695
                                                   N    standby_group_09.log => +FRA/EAST/ONLINELOG/group_9.271.823535703
                                                   N    standby_group_10.log => +FRA/EAST/ONLINELOG/group_10.272.823535727
                                                   N    standby_group_11.log => +FRA/EAST/ONLINELOG/group_11.273.823535737
                                                   N    standby_group_12.log => +FRA/EAST/ONLINELOG/group_12.274.823535745
                                                   N    standby_group_13.log => +FRA/EAST/ONLINELOG/group_13.275.823535757
                                                   N    standby_group_14.log => +FRA/EAST/ONLINELOG/group_14.276.823535763
                                                   N    standby_group_15.log => +FRA/EAST/ONLINELOG/group_15.277.823535771
    + asmcmd find +data 'group*'
    +data/EAST/ONLINELOG/group_1.261.823276235
    +data/EAST/ONLINELOG/group_2.262.823276241
    +data/EAST/ONLINELOG/group_3.263.823276247
    +data/HPYMUSIC/ONLINELOG/group_1.271.824213881
    +data/HPYMUSIC/ONLINELOG/group_2.272.824213887
    +data/HPYMUSIC/ONLINELOG/group_3.273.824213895
    + asmcmd find +data 'spf*'
    +data/EAST/PARAMETERFILE/spfile.265.823277967
    +data/EAST/spfileeast.ora
    + asmcmd ls -l data/east/CONTROLFILE
    Type         Redund  Striped  Time             Sys  Name
    CONTROLFILE  HIGH    FINE     AUG 12 15:00:00  Y    Current.260.823276231
    CONTROLFILE  HIGH    FINE     AUG 23 19:00:00  Y    Current.270.823277705
    + asmcmd find +fra 'group*'
    +fra/EAST/ONLINELOG/group_1.257.823276237
    +fra/EAST/ONLINELOG/group_10.272.823535727
    +fra/EAST/ONLINELOG/group_11.273.823535737
    +fra/EAST/ONLINELOG/group_12.274.823535745
    +fra/EAST/ONLINELOG/group_13.275.823535757
    +fra/EAST/ONLINELOG/group_14.276.823535763
    +fra/EAST/ONLINELOG/group_15.277.823535771
    +fra/EAST/ONLINELOG/group_2.258.823276245
    +fra/EAST/ONLINELOG/group_3.259.823276251
    +fra/EAST/ONLINELOG/group_7.269.823535685
    +fra/EAST/ONLINELOG/group_8.270.823535695
    +fra/EAST/ONLINELOG/group_9.271.823535703
    +fra/HPYMUSIC/ONLINELOG/group_1.261.824213883
    +fra/HPYMUSIC/ONLINELOG/group_2.262.824213889
    +fra/HPYMUSIC/ONLINELOG/group_3.263.824213897
    + asmcmd find +fra 'spf*'
    + asmcmd ls -l fra/east/CONTROLFILE
    Type         Redund  Striped  Time             Sys  Name
    CONTROLFILE  HIGH    FINE     AUG 12 15:00:00  Y    Current.256.823276231
    CONTROLFILE  HIGH    FINE     AUG 23 19:00:00  Y    Current.260.823277707
    ++ date +%y%m%d,%H%M%S
    + echo END,130823,195712
    [grid@tokyo1 shell]$
    ■ 以下、スタンバイ側 ■ ■ ■ ■ ■ ■ ■
    export ORACLE_SID=HPYMUSIC
    sqlplus / as sysdba
    startup nomount pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt'
    create spfile='+data/west/spfilewest.ora' from pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt';
    srvctl stop database -d west -f
    srvctl start database -d west -o open
    srvctl start database -d west -o mount
    srvctl start database -d west
    startup mount pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt'
    srvctl start database -d west -o open
    srvctl config database -d west
    srvctl status database -d west
    alter database recover managed standby database disconnect from session;
    select name, database_role, open_mode from gv$database;
    srvctl modify database -d west -s open
    ○ spfile を作成する
    export ORACLE_SID=HPYMUSIC
    sqlplus / as sysdba
    startup nomount pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt'
    create spfile='+data/west/spfilewest.ora' from pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt';
    ○ 落とす
    srvctl stop database -d west -f
    ○ open したいが起動しない( Mounted (Closed) 状態で起動することもある)。
    srvctl start database -d west -o open
    PRCR-1079 : リソースora.west.dbの起動に失敗しました
    CRS-2674: Start of 'ora.west.db' on 'osaka1' failed
    ○ open したいが起動しない( Mounted (Closed) 状態で起動することもある)。
    srvctl start database -d west -o mount
    PRCR-1079 : リソースora.west.dbの起動に失敗しました
    CRS-2674: Start of 'ora.west.db' on 'osaka1' failed
    ○ open したいが起動しない( Mounted (Closed) 状態で起動することもある)。
    srvctl start database -d west
    PRCR-1079 : リソースora.west.dbの起動に失敗しました
    CRS-2674: Start of 'ora.west.db' on 'osaka1' failed
    ○ 起動するがエラーあり( alert_HPYMUSIC.log )
    startup mount pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt'
    [oracle@osaka1 dbs]$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.1.0 Production on 金 8月 23 19:05:35 2013
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    アイドル・インスタンスに接続しました。
    SQL> startup mount pfile='/u01/app/oracle/product/11.2.0/dbhome_1/dbs/pfile_for_standby_HPYMUSIC.txt'
    ORACLEインスタンスが起動しました。
    Total System Global Area  839282688 bytes
    Fixed Size                  2217992 bytes
    Variable Size             515901432 bytes
    Database Buffers          314572800 bytes
    Redo Buffers                6590464 bytes
    データベースがマウントされました。
    Error 12154 received logging on to the standby
    FAL[client, ARC3]: Error 12154 connecting to HPYMUSIC for fetching gap sequence
    Errors in file /u01/app/oracle/diag/rdbms/west/HPYMUSIC/trace/HPYMUSIC_arc3_25690.trc:
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    Errors in file /u01/app/oracle/diag/rdbms/west/HPYMUSIC/trace/HPYMUSIC_arc3_25690.trc:
    ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    ○ open にならず Mounted (Closed) としかなってくれない。
    srvctl start database -d west -o open
    [oracle@osaka1 dbs]$ srvctl start database -d west -o open
    [oracle@osaka1 dbs]$ /u01/app/11.2.0/grid/bin/crsctl status resource -t
    NAME           TARGET  STATE        SERVER                   STATE_DETAILS
    Local Resources
    ora.DATA.dg
                   ONLINE  ONLINE       osaka1
    ora.FRA.dg
                   ONLINE  ONLINE       osaka1
    ora.LISTENER.lsnr
                   ONLINE  ONLINE       osaka1
    ora.asm
                   ONLINE  ONLINE       osaka1                   Started
    Cluster Resources
    ora.cssd
          1        ONLINE  ONLINE       osaka1
    ora.diskmon
          1        ONLINE  ONLINE       osaka1
    ora.west.db
          1        ONLINE  INTERMEDIATE osaka1                   Mounted (Closed)

    srvctl config database -d west
    srvctl status database -d west
    [oracle@osaka1 dbs]$ srvctl config database -d west
    一意のデータベース名: west
    データベース名: HPYMUSIC
    Oracleホーム: /u01/app/oracle/product/11.2.0/dbhome_1
    Oracleユーザー: grid
    spfile: +data/west/spfilewest.ora
    ドメイン:
    開始オプション: open
    停止オプション: immediate
    データベース・ロール: physical_standby
    管理ポリシー: AUTOMATIC
    ディスク・グループ: DATA,FRA
    サービス:
    [oracle@osaka1 dbs]$ srvctl status database -d west
    データベースは実行中です。
    ○ mrp プロセスが起動するが、Read Only ではない。
    alter database recover managed standby database disconnect from session;
    select name, database_role, open_mode from gv$database;
    [oracle@osaka1 dbs]$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.1.0 Production on 金 8月 23 19:33:08 2013
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, Automatic Storage Management, OLAP, Data Mining
    and Real Application Testing options
    に接続されました。
    SQL>
    SQL> alter database recover managed standby database disconnect from session;
    データベースが変更されました。
    SQL> select name, database_role, open_mode from gv$database;
    NAME                        DATABASE_ROLE
    OPEN_MODE
    HPYMUSIC                    PHYSICAL STANDBY
    MOUNTED
    [root@osaka1 app]# ps -ef |egrep -i mrp
    oracle   26269     1  0 19:33 ?        00:00:00 ora_mrp0_HPYMUSIC
    ○ modify しても open にならない。
    srvctl modify database -d west -s open
    [oracle@osaka1 dbs]$ srvctl modify database -d west -s open
    [oracle@osaka1 dbs]$ /u01/app/11.2.0/grid/bin/crsctl status resource -t
    NAME           TARGET  STATE        SERVER                   STATE_DETAILS
    Local Resources
    ora.DATA.dg
                   ONLINE  ONLINE       osaka1
    ora.FRA.dg
                   ONLINE  ONLINE       osaka1
    ora.LISTENER.lsnr
                   ONLINE  ONLINE       osaka1
    ora.asm
                   ONLINE  ONLINE       osaka1                   Started
    Cluster Resources
    ora.cssd
          1        ONLINE  ONLINE       osaka1
    ora.diskmon
          1        ONLINE  ONLINE       osaka1
    ora.west.db
          1        ONLINE  INTERMEDIATE osaka1                   Mounted (Closed)

    standby 側に、アーカイブログファイルを少し適用すれば open できそうに感じます。
    これを解決するためには、以下のエラーを解決するのがよいと思います。
    ・ORA-12154: TNS: 指定された接続識別子を解決できませんでした
    両ノードの tnsnames.ora を確認させて貰えますか?
    この推測が間違えていたら再検討しましょう。

  • RAC and standby database

    Hi,
    We have configured a RAC production system and an standalone standby database.
    Last week, we had trouble with our ESS subsystem and we have to failover.
    Rigth now we solve the problem with the ESS subsysytem and the RAC enviroment is up and running but the database is no updated because is at the moment that we had the problem.
    So, we want to back to the initial configuration (RAC primary database and standalone standby).
    I want to know is this can be possible with only a backup/restore? (standby database cold backup and then restore the database over the RAC environment) and then reconfigure our standby database?
    Or which steps should we follow to do this?
    Thanks in advance,
    Gerald

    My question is:
    Is it possible to setup a standby database on a RAC environment with 2 nodes? I've searched through MOS but didn't find any information about that...Yes, it is possible. Refer this
    *Data Guard Physical Standby 11.2 RAC Primary to RAC Standby using a second network [ID 1349977.1]*
    MAA / Data Guard 10g Setup Guide – Creating a RAC Physical Standby for a RAC Primary
    http://www.oracle.com/technetwork/database/features/availability/maa-wp-10g-racprimaryracphysicalsta-131940.pdf
    http://wiki.tapriuneclak.com/index.php?title=Oracle_-StepBy_Step_of_Configuring_Oracle_11gR2_(11.2.0.1)_RAC_to_RAC_Dataguard
    http://davidalejomarcos.wordpress.com/2010/02/17/creating-a-3-node-rac-database-as-standby/

  • Rebuild RAC phsical standby with RMAN

    Hi all,
    Can anyone give any advice as to rebuilding an ACTIVATED RAC physical standby using RMAN? We have been able to revert the db with flashback but I need to know the rebuild method.
    Is there any extra steps with the OCR?
    I am guessing the steps might be:
    1. Use DBCA to delete the newly activated standby.
    (The DBCA will clean up the OCR components)
    2. Use RMAN to duplicate a physical standby database.
    3. Register the instances and services with the OCR.
    Can any comment on this and suggest a decent task list from oracle as to the method.
    Thanks in advance,
    jd
    Edited by: jonnyd on 30-May-2010 13:42

    ...anyone...?

  • Configuring Non RAC physical standby fro RAC

    Hi,
    EBS-R12i , Oracle Database:11g
    We are in project design phase and following is the hardware setup for Critical banking domain:
    Database (2 node RAC)-Solaris 10
    Application server server (2 server) -Solaris 10
    Web server (2 server for Failover)= RHEL
    For DR
    Database (1 node)-Solaris 10
    need to know
    1.What are the best practice to implement this kind of setup and steps to configure DR as Non-RAC physical standby for RAC database server.
    2.What are implication if we go for single node except failover node unavailability for DR.
    3.What are the steps of Switchover in this case.
    4.What's the recommended backup solution for this setup of 8-10 Terabyte database with db growth rate of 50% per year.

    1.What are the best practice to implement this kind of setup and steps to configure DR as Non-RAC physical standby for RAC database server.There really aren't any differences in setting up your physical standby in a non-RAC configuration. It is after the physical standby is setup that it is converted to a clustered database, if desired. In your case, you do not desire the conversion.
    The best place to start for best practices in setting up your physical standby is the Oracle documentation on the subject.
    http://download.oracle.com/docs/cd/E11882_01/server.112/e17022/toc.htm
    2.What are implication if we go for single node except failover node unavailability for DR.The major implications are that you will be running with less horsepower after a failover. So depending on your DR resources, you may have to scale back application access. In addition to less resources, you will also have less availability because you only have one node. Should that node then fail, you will not have a node to failover to.
    3.What are the steps of Switchover in this case.The steps are identical if the physical standby is a single node or multiple nodes. These steps are outlined in the Oracle documentation.
    4.What's the recommended backup solution for this setup of 8-10 Terabyte database with db growth rate of 50% per year.What are your recovery requirements? Without knowing your recovery requirements, a valid backup solution cannot be determined. How much data loss can you afford in the event of a failure? Do you have data that can be read only? In the event of a failure, how long before you are required to be up and running? Are you required to restore and recover at an offsite location?
    Cheers,
    Brian

  • RAC & Dataguard

    Hi everyone, i hope somebody can help me.
    I have a configuration RAC (2 nodes) and 1 dataguard (standby physicall).
    Both run in Linux Red Hat 4 Update 7.
    My quesion is, Probably i need to do a flashback database in the RAC for some test so Must i do some in my dataguard? or it is transparent?
    Thank you very much for your time.
    Regards.

    Hi,
    I think you need to test. Whether it is RAC or single instance, asnwer will be same. In worst case you may need to setup the data guard again. Data guard supports DDL from primary to standby database, but I did not read anything about flashing back the primary in this setup. We may benefit from your experiment.
    Regards

  • Steps to sync RAC -dataguard

    Hi folks,
    I m having setup with 2 node RAC primary and 2 node RAC Dataguard on 10.2.0.4.0. Dataguard setup is working fine. Dataguard is setup with Standby Redo log group with managed recovery. Due to network issue network issue the transferring archives & applying on standby couldnot take place.The network issue is sorted but now the primary and standby are not in sync.
    I am new with RAC and dataguard combined.Can you proivde suitable document or steps how i can manually sync both.

    Hi;
    Pelase see below which could be helpful for your issue:
    RAC & DataGuard (Physical Standby)
    RAC & DataGuard (Physical Standby)
    Also see:
    http://www.bestremotedba.com/oracle-10g-dataguard-sync-standby-database/
    If those are not help you can move your issue to Forum Home » Grid Computing » Real Application Clusters which you can get more quick response
    Regard
    Helios

  • RAC + dataguard architecture/setup metalink notes

    Looking for RAC & dataguard configuration for one db ...

    Hi,
    I think it is rather worth to look around on the [Oracle Maximum Availability Architecture site|http://www.oracle.com/technology/deploy/availability/htdocs/maa.htm] for whitepapers than checking the Metalink (Well, at least I like MAA much better the Metalink)
    Hope this helps.
    Regards,
    Jozsef

  • Conflicting PROCESSES values on 10g RAC+Dataguard

    Hi peeps,
    So i've set up a RAC+Dataguard environment for testing, and have recently encountered ORA-12518 errors; quite a few Google results suggest checking V$RESOURCE_LIMITS for any resource problems. Upon checking i noticed that the MAX_UTILIZATION for resource PROCESSES is set to 71, with CURRENT_UTILIZATION set to 70;i found this strange 'coz i specifically increased the PROCESSES parameter to 300 when i set up the db; i ran SHOW PARAMETER PROCESSES and verified it was set to 300.
    Shouldn't i be getting the same values?
    tia

    Hello badkuk,
    The max_utilization column means the greater value of this resource since last instance startup. See it on http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/dynviews_2050.htm#REFRN30206
    Also see the description of error message at http://ora-12518.ora-code.com
    I think the root cause of this error was not processes parameter.
    Take a look on Metalink note ID 556428.1
    *"In general ORA-12518 occurs if there is a resource problem , may be on the listener side or on the Database server. It indicates a condition where the process is not able to make any more memory allocations either due to non-availability of resource or the process has reached the 32 bit (VM Size) limit"*
    Best regards,
    Rodrigo Mufalani
    http://www.mrdba.com.br/mufalani

  • Next log sequence to archive in Standby Database (RAC Dataguard Issue)

    Hi All,
    I just had implemented Data Guard in our server. My primary Database is RAC configured, but it is only single node. The other Instance was removed and converted it to Developement Instance. For the reason I kept the primary as RAC is when I will implement dataguard, my Primary Database is RAC with 7 nodes.
    The first test is successful, and I was able to "switchover" from my primary to standby. I failed in the 'FAILOVER" test.
    I restore my primary server and redo the setup.
    BTW, my standby DB is physical standby.
    When I try to switchover again and issue archive log list, below is my output.
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination USE_DB_RECOVERY_FILE_DEST
    Oldest online log sequence 38
    *Next log sequence to archive 0*
    Current log sequence 38
    SQL> select open_mode, database_role from v$database;
    OPEN_MODE DATABASE_ROLE
    MOUNTED PHYSICAL STANDBY
    ===============================================
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination USE_DB_RECOVERY_FILE_DEST
    Oldest online log sequence 38
    *Next log sequence to archive 38*
    Current log sequence 38
    SQL> select open_mode, database_role from v$database;
    OPEN_MODE DATABASE_ROLE
    READ WRITE PRIMARY
    In my first attempt to switchover before I failed in "FAILOVER" test, I also issue +archive log list+ in both primary and standby database, and if I remember it right, next log sequence on both should be identical. Am I right on this?
    Thanks in Advance.
    Jay A

    Or Am i just overthinking on this?
    Is dataguard only looking for the current and oldest log sequence?

  • Add datafile dataguard from physical  RAC to standby RAC

    Hello we have a 3 node cluster RAC 10g on the primary and RAC 10g on the standby.We use as db storage ASM wich is configured over raw devices and file management omf.The same groups DATA and FRA on both servers.STANDBY_FILE_MANAGEMENT is manual.
    We need to add a datafile on the primary safely.Can you offer me some ideeas?
    Thank you!

    Hello;
    Change STANDBY_FILE_MANAGEMENT to "auto" before you add the file and then leave it set to "auto". ( set this to AUTO on both Primary and Standby )
    STANDBY_FILE_MANAGEMENT enables or disables automatic standby file management.
    If you leave it set to MANUAL you may create unwanted file in $ORACLE_HOME/dbs and cause issues with the MRP process.
    Best Regards
    mseberg
    Edited by: mseberg on Jun 7, 2013 9:02 AM

  • Steps to upgrade RAC wit Standby Database ?

    Hi All
    I have RAC environment with 8 Nodes (Linux 4) , I have one node as physical standby for this RAC.
    My oracle is 10.2.0.3 , and i want to update it to 10.2.0.5
    Could you please advice me with steps , when i Google i found that i have to start with standby then Clusterware ? 
    Any recommendation ?
    Thanks in advance.

    Hello, look at the documentation of Dataguard Concepts and Admininstration http://docs.oracle.com/cd/B19306_01/server.102/b14239/upgrades.htm#SBYDB1200. There's a chapter "B Upgrading Databases in a Data Guard Configuration"

  • Rac with Standby non-RAC

    Hi hussein,
    EBS R12.1
    OEL 5.4
    Is it possible that a EBS RAC setup will have a Standby (dataguard) which is non-rac?
    Thanks a lot,
    Ms K

    Hi Mskl;
    Please also check below note which could be helpful
    Business Continuity for Oracle Applications Release 12 on Database Release 10gR2 - Single Instance and RAC [ID 452056.1]
    Regard
    Helios

Maybe you are looking for