Data guard - logical standby in no archive log

We are creating a logical standby for reporting purpose only. I see no reason why it should be in archivelog mode. Would this hamper our current situation?

the solution is already approved , the only thing i could do is wonder if backup is necessary for the logical.
Its not going to be used for any purpose other than reporting.
From what i understand, no archivemode should not affect it. Is this undertanding true

Similar Messages

  • Data Guard logical standby Versus Streams

    I'm referring to both Oracle 10g/9i
    If a Data Guard logical database uses similar technology to Streams (Log Mining and SQL apply), why can't you stand up a standby database on a different platform, or at least I have found nothing on the subject.
    But in 11g Oracle Data Guard (physical standby database) is a solution for same endianness platform migration.
    I will appreciate any insight on the subject.

    Yes...thats true...both uses same technology...
    REDO LOGMINER
    SQL------------------------>BLOCK LEVEL CHANGES---------------------------------->SQL
    But there are serious implemetation diff..
    1) Oracle Data Guard is designed for protecting from data failure and disasters.
    Streams is designed for information sharing and distribution but can also provide a very efficient high availability solution.
    2) Streams is configured from the bottom up — individual tables, schemas, capture processes, apply processes, queues.
    Logical Standby is configured from the top down — start with entire database, then specify only what you don’t want.
    As logical standby is top down and changes are capture at remote location (logical standby db) and for that archive log need to be shipped using FAL client/server to remote and to ship the archivelog in Data Guard configuration, all members should be running on same platform.
    As said before, Streams is configured from the bottom up, it start with tables--> Schemas--->Database and we can capture changes at local/remote location. If we capture changes locally, target streams db can be on diff platform. But for downstream capture must need same platform as logical standby database needed so that archivelogs can transport from source to downstream db.

  • Data guard real time apply vs archived log apply on physical standby

    Dear DBA's,
    last week i configuared DR , now the phyiscal stanby database is archive apply mode,
    i want to confirm is it better to apply the archived log or should i cahnge it to real time apply .
    give me sugesstions.
    Thanks and Regards
    Raja...

    One question are you using ARCH transport to move the redo? or have you configured standby redo logs and logwr transport (either async or syncronous), if you are using the archiver to transport the logs then you can not use real time apply.
    If you are using log writer to transpor the redo the realtime apply reduces the recovery time required if you need to failover as trher should be less redo to apply to bring the standby up to date, which mode you use to transport redo will depend on what is acceptable in terms of data loss and the impact on performance.

  • DATA Guard Logical Standby v.s. Streams Apply (10gR2, rac -- rac )

    Greetings -
    We currently have a 3 node cluster, doing a 'Schema' capture, with a 2 node cluster serving as the apply side. Both are on 10gR2 (solaris 10)
    We have been experiencing apply latency due to large transactions. The way logminer/streams evaluates the arch logs, it converts each updated row into a transaction as part of a transaction set, using 'decode (' statements.
    I am under the impression a physical standby will do the same thing. But what about a logical Standby ?
    (this from the Oracle Documentation)
    [10g Concepts|http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/concepts.htm#SBYDB00010]
    ...Contains the same logical information as the production database, although the physical organization and structure of the data can be different. The logical standby database is kept synchronized with the primary database though SQL Apply, which transforms the data in the redo received from the primary database into SQL statements and then executing the SQL statements on the standby database.
    Does anyone know firsthand if SQL APPLY re-issues the original sql, or if it relies on the 'decode' command as well ?
    Thanks
    The Nets Edge

    A Physical standby does not do anything with SQL. It is running media recovery under the covers. A Logical standby uses the Logminer to read the redo, converts it to SQL and Data and applies those transactions to the Logical standby.
    The both build logical change records, figure out order, dependencies etc etc and applies the transaction. Both are susceptible to long running transactions on the Primary (source) database.
    Streams uses Logminer, Logical Standby use part of the Streams apply so as Mr. Morgan said, they are very similar :^)
    If you are having apply performance issues you may want to look at Active data Guard.
    Larry

  • Data Guard Logical Standby DB Questions

    Hi,
    I am creating Oracle 9i Logical Standby database on AIX 5 on same server ( for
    testing only ).
    In Data Guard Manual ( PDF file page no : 86 ) it is said that
    Step 4 Create a backup copy of the control file for the standby database. On the
    primary database, create a backup copy of the control file for the standby database:
    SQL> ALTER DATABASE BACKUP CONTROLFILE TO
    2> '/disk1/oracle/oradata/payroll/standby/payroll3.ctl';
    My Questions regarding this are
    1. Does it missing "backup standby controlfile" word ? if not what is the use of
    this control file ? Because as per this manual section 4.2.2 step no 1 and 2 I
    have stopped the primary database and copied datafiles,logfiles and controlfiles
    as specified which are having consistent stage.
    2. On primary database I am mirroring 2 controlfiles at two diffrent locations.
    On Standby Database I am going to keep same policy. Then on standby database do I have to copy this controlfile to two locations and rename datafiles or I have
    to user controlfiles which I have backed up along with datafiles ?
    3. Suppose my primary database is "infod" and standby database is "standby".
    Then what should be value of log_file_format parameter of standby database ? On
    primary db I have defined "infod LOG_ARCHIVE_FORMAT=infod_log%s_%t.arc. Do I have to keep same or change it ? If on standby db I change the value do I
    require other parameters to be set ?
    regards & thanks
    pjp

    Q/A1) Its correct you dont need the standby keyword. Not sure abut the reason why but we have created logical standby running in prod using the oracle doc.
    Q/A2) Yo can specify any location and name for your controlfiles as it is instance specific and not depending on primary database.
    Q/A3) You can set any format for your archlogs on standby side. It doesnt bother the primary db or DG configuration.
    Regards,
    http://askyogesh.com

  • Data Guard Logical Standby Skip Handler bug

    Does anyone know of a documented bug concerning the skip handler not skipping a grant on a view in a schema with a SCHEMA_DDL skip handler defined?
    Here is my test case.
    On Primary:
    Create table scott.junk (x number) tablespace users; (created successfully on logical standby)
    On logical standby:
    exec dbms_logstdby.skip('SCHEMA_DDL','SCOTT','%'null,true,'\'); (created skip handler on all SCOTT objects)
    On Primary:
    grant select on scott.junk to public; (successfully skipped due to missing object)
    On Primary:
    create view scott.v_junk as select * from scott.junk; (successfully skipped on logical standby)
    On Primary:
    grant select on scott.v_junk to public; (received 'ora-00942 table or view does not exist' on logical standby and SQL Apply stopped).
    I'm running version 10.2.0.4 but have not applied the recommended patch bundle for logcal standby #7937113.8.
    If anyone knows whether this has been documented or not, please let me know. If not, I will submit an SR to get this documented.
    Thanks,
    Michael Anderson
    OCP - Bank of the West

    When I look at the syntax for the package
    http://www.morganslibrary.org/reference/dbms_logstdby.html
    what I see is:
    dbms_logstdby.skip(
    stmt        IN VARCHAR2,
    schema_name IN VARCHAR2 DEFAULT NULL,
    object_name IN VARCHAR2 DEFAULT NULL,
    proc_name   IN VARCHAR2 DEFAULT NULL,
    use_like    IN BOOLEAN  DEFAULT TRUE,
    esc         IN CHAR1    DEFAULT NULL);So looking at your code:
    exec dbms_logstdby.skip('SCHEMA_DDL','SCOTT','%'null,true,'\'); (created skip handler on all SCOTT objects)
    'SCHEMA_DDL' is your statement.
    'SCOTT' is your schema name.
    '%' indicates all objects
    and ... shouldn't there be a comma before the NULL?
    In the future don't provide values where the default's will suffice. It makes thing's far simpler.

  • Is the only way to import large amount of data and database objects into a primary database is to shutdown the standby, turn off archive log mode, do the import, then rebuild the standby?

    I have a primary database that need to import large amount of data and database objects. 1.) Do I shutdown the standby? 2.) Turn off archive log mode? 3.) Perform the import? 4.) Rebuild the standby? or is there a better way or best practice?

    Instead of rebuilding the (whole) standby, you take an incremental (from SCN) backup from the Primary and restore it on the Standby.  That way, if, for example
    a. Only two out of 12 tablespaces are affected by the import, the incremental backup would effectively be only the blocks changed in those two tablespaces (and some other changes in system and undo) {provided that there are no other changes in the other ten tablespaces}
    b. if the size of the import is only 15% of the database, the incremental backup to restore to the standby is small
    Hemant K Chitale

  • Logical Standby: WAITING FOR DICTIONARY LOGS status.

    Hi,
    I have just configured a Logical Standby from Physical Standby, I followed all steps explained in section 4.2.1 from the Oracle® Data Guard Concepts and Administration 10g Release 2 (10.2) documentation, but now my logical standby stay on WAITING FOR DICTIONARY LOGS more than 3 days, all archives from primary DB have been replicated and registered, I don't understand whats
    wrong.
    SQL> SELECT * FROM V$LOGSTDBY_STATE;
    PRIMARY_DBID SESSION_ID
    REALTIME_APPLY
    STATE
       144528764          1
    Y
    WAITING FOR DICTIONARY LOGS
    Fri Apr  8 18:05:47 2011
    RFS LogMiner: Registered logfile [/archive/dbp/581440892_1_0000157573.arc] to LogMiner session id [1]
    Fri Apr  8 18:10:55 2011
    RFS LogMiner: Client enabled and ready for notification
    Fri Apr  8 18:10:55 2011
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: Successfully opened standby log 4: '/redo2a/dbp/redo4a_stb.log'
    Fri Apr  8 18:10:58 2011
    RFS LogMiner: Registered logfile [/archive/dbp/581440892_1_0000157574.arc] to LogMiner session id [1]
    Fri Apr  8 18:15:54 2011
    RFS LogMiner: Client enabled and ready for notification
    Fri Apr  8 18:15:54 2011
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: Successfully opened standby log 3: '/redo1a/dbp/redo3a_stb.log'
    Fri Apr  8 18:15:57 2011
    RFS LogMiner: Registered logfile [/archive/dbp/581440892_1_0000157575.arc] to LogMiner session id [1]Thanks in advance.
    My Oracle version is: 10.2.0.4
    My Platform is: AIX 6.1
    Nataly.

    On standby alert no error messages found.
    RFS LogMiner: Client enabled and ready for notification
    Mon Apr 11 18:09:58 2011
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: Successfully opened standby log 4: '/redo2a/mefsf/redo4a_stb.log'
    Mon Apr 11 18:10:01 2011
    RFS LogMiner: Registered logfile [/archive/mefsf/581440892_1_0000157782.arc] to LogMiner session id [1]
    Mon Apr 11 18:23:16 2011
    RFS LogMiner: Client enabled and ready for notification
    Mon Apr 11 18:23:16 2011
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: Successfully opened standby log 3: '/redo1a/mefsf/redo3a_stb.log'
    Mon Apr 11 18:23:18 2011
    RFS LogMiner: Registered logfile [/archive/mefsf/581440892_1_0000157783.arc] to LogMiner session id [1]Continues
    On primary alertlog,
    There is a common message:
    ORACLE Instance bdprod - Archival Error. Archiver continuing.
    Mon Apr 11 18:22:57 2011
    Errors in file /oracle/app/oracle/admin/bdprod/bdump/bdprod_arc4_2818414.trc:
    ORA-00308: cannot open archived log '/archive/bdprod/581440892_1_0000157545.arc'
    ORA-27037: unable to obtain file status
    IBM AIX RISC System/6000 Error: 2: No such file or directory
    Additional information: 3
    Mon Apr 11 18:22:57 2011
    FAL[server, ARC4]: FAL archive failed, see trace file.
    Mon Apr 11 18:22:57 2011
    Errors in file /oracle/app/oracle/admin/bdprod/bdump/bdprod_arc4_2818414.trc:
    ORA-16055: FAL request rejected
    ARCH: FAL archive failed. Archiver continuing
    Mon Apr 11 18:22:57 2011
    ORACLE Instance bdprod - Archival Error. Archiver continuing.
    Mon Apr 11 18:22:57 2011
    Errors in file /oracle/app/oracle/admin/bdprod/bdump/bdprod_arc1_7078038.trc:
    ORA-00308: cannot open archived log '/archive/bdprod/581440892_1_0000157546.arc'
    ORA-27037: unable to obtain file status
    IBM AIX RISC System/6000 Error: 2: No such file or directory
    Additional information: 3
    Mon Apr 11 18:22:57 2011
    FAL[server, ARC1]: FAL archive failed, see trace file.
    Mon Apr 11 18:22:57 2011
    Errors in file /oracle/app/oracle/admin/bdprod/bdump/bdprod_arc1_7078038.trc:
    ORA-16055: FAL request rejected
    ARCH: FAL archive failed. Archiver continuing
    Mon Apr 11 18:22:57 2011
    ORACLE Instance bdprod - Archival Error. Archiver continuing.

  • Primary/standby and 3rd archive log destination?

    Running Oracle EE 10.2.0.4 linux 64 bit. I have a primary and standby configuration using data guard, with appropriate LOG_ARCHIVE_DEST_1, LOG_ARCHIVE_DEST_2 for the primary and standby - all works as expected. I want to multiplex the log files by adding a 3rd archive log destination to a different disk. I'm not clear on whether or not specifying a second 'primary' archive destination will write to both dest_1 and dest_3 or just alternate between the 2? I want to make sure that both DEST_1 and DEST_3 are written to (there can be a lag for DEST_3) -
    Is the following all I need or are there additional parameters to LOG_ARCHIVE_DEST_n I'm missing?
    LOG_ARCHIVE_DEST_3 = 'valid_for=(ONLINE_LOGFILE,ALL_ROLES)', 'location="/myThirdLoc"'
    Thanks -

    The database hang because of the redo log file is full. my test database is 16.7G and I gave the recovery filesystem 19 G for the redo log. Apparently it is not enough. After i set up the backup with daily incremental and run for 5 days, the recovery is full and db hang. How do you free up the space or set up the backup so that it can recycle the space? My DB is 10g r2 on RHEL3.

  • 11.2 DB & Data Guard : ORA-16014 how to archive a sequence?

    Hi,
    I've installed 11.2 Oracle Database in my laptop with Oracle Enterprise Linux 5.3 and I have created two standby databases, orcl (primary) and orclstby (physical standby). I performed a switchover to orclstby and consequently, orclstby was the new primary and orcl the physical standby, I checked those values with SQLPlus executing the select database_role from v$database statement, so, there were no problems during the switchover.I also shutteddown and started both primary and stanby databases to check all was fine.
    Today, I've tried to start the environment again and I've encountered the following problem during the startup of the primary database:
    [oracle@mredon-es ~]$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.1.0 Production on Wed Oct 21 17:34:21 2009
    Copyright (c) 1982, 2009, Oracle. All rights reserved.
    Connected to an idle instance.
    SQL> startup
    ORACLE instance started.
    Total System Global Area 839282688 bytes
    Fixed Size 2217992 bytes
    Variable Size 507512824 bytes
    Database Buffers 322961408 bytes
    Redo Buffers 6590464 bytes
    Database mounted.
    ORA-03113: end-of-file on communication channel
    Process ID: 5562
    Session ID: 9 Serial number: 3
    I've checked the newest log created and this is its content:
    Dump file /u01/app/oracle/product/11.2.0/dbhome_1/rdbms/log/orclstby_ora_5382.trc
    *** 2009-10-21 17:34:30.467
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORACLE_HOME = /u01/app/oracle/product/11.2.0/dbhome_1
    System name: Linux
    Node name: mredon-es.us.oracle.com
    Release: 2.6.18-128.el5
    Version: #1 SMP Wed Jan 21 08:45:05 EST 2009
    Machine: x86_64
    Instance name: orclstby
    Redo thread mounted by this instance: 0 <none>
    Oracle process number: 0
    Unix process pid: 5382, image: [email protected]
    *** 2009-10-21 17:34:30.526
    2009-10-21 17:34:30.457: [ default]utgdv:2:ocr loc file /etc/oracle/olr.loc cann
    ot be opened. errno 2
    2009-10-21 17:34:30.527: [ default]utgdv:2:ocr loc file /etc/oracle/ocr.loc cann
    ot be opened. errno 2
    I've searched in Metalink information about ocr.loc and olr.loc and they seem to be part of an Oracle RAC installation, I don't know why the database needs these files to start if I am using single instance...
    Any idea would be kindly appreciated because I'm a bit confused and I don't really know what steps I am suppposed to take in order to solve this problem.
    Thanks in advance.
    Edited by: mredon on Oct 26, 2009 5:22 AM
    Reason: title of the post changed

    sorry for the delay, here is the alert log:
    <msg time='2009-10-21T23:03:09.576+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>Data Guard: version check completed
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:09.642+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6486'>
    <txt>LGWR: STARTING ARCH PROCESSES
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:09.709+02:00' org_id='oracle' comp_id='rdbms'
    msg_id='ksbrdp:3833:3697353022' type='NOTIFICATION' group='process start'
    level='16' host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130'
    pid='6588'>
    <txt>RSM0 started with pid=23, OS id=6588
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:09.831+02:00' org_id='oracle' comp_id='rdbms'
    msg_id='ksbrdp:3833:3697353022' type='NOTIFICATION' group='process start'
    level='16' host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130'
    pid='6590'>
    <txt>ARC0 started with pid=24, OS id=6590
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:10.832+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6486'>
    <txt>ARC0: Archival started
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:10.832+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6486'>
    <txt>LGWR: STARTING ARCH PROCESSES COMPLETE
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:10.833+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6590'>
    <txt>ARC0: STARTING ARCH PROCESSES
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.180+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>ARCH: LGWR is scheduled to archive destination LOG_ARCHIVE_DEST_2 after log switch
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.183+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>ARCH: LGWR is scheduled to archive destination LOG_ARCHIVE_DEST_1 after log switch
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.258+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>Errors in file /u01/app/oracle/product/diag/rdbms/orclstby/orclstby/trace/orclstby_ora_6582.trc:
    ORA-16014: log 1 sequence# 27 not archived, no available destinationsORA-00312: online log 1 thread 1: &apos;/u01/app/oracle/product/oradata/orclstby/redo01.log&apos;
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.263+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>USER (ospid: 6582): terminating the instance due to error 16014 </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.429+02:00' org_id='oracle' comp_id='rdbms'
    msg_id='ksbrdp:3833:3697353022' type='NOTIFICATION' group='process start'
    level='16' host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130'
    pid='6592'>
    <txt>ARC1 started with pid=25, OS id=6592
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:12.728+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>Instance terminated by USER, pid = 6582
    </txt>
    </msg>
    I've marked in bold the errors which may be causing the problem, ¿what do you think?

  • 11g Data guard 建physical standby database配置相关问题

    General information
    OS:red hat Linux 2.6.32-200.13.1.el5uek x86_64(primary,standby)
    Home version:11.2.0.3(primary,standby)
    Situation:
    实验用库,想学习data guard,在另一个服务器上克隆了一个HOME并按照官方文档http://docs.oracle.com/cd/E11882_01/server.112/e25608/create_ps.htm#i1225703
    中去实验,步骤均按照官方文档来,最后在mount standby库时校验select sequence# from v$archived_log没有记录,出错日志如下(从mount开如截取)
    standby alert_dbacoe.log
    Error logs
    alter database mount
    Completed: alter database mount
    Error 604 received logging on to the standby
    +FAL[client, ARC2]: Error 604 connecting to PRIMARYSV for fetching gap sequence+
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo01_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 1 thread 1: '/u03/app/oracle/oradata/dbacoe/redo01.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo01_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 1 thread 1: '/u03/app/oracle/oradata/dbacoe/redo01.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 2 of thread 1
    ORA-00312: online log 2 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo02_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 2 thread 1: '/u03/app/oracle/oradata/dbacoe/redo02.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 2 of thread 1
    ORA-00312: online log 2 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo02_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 2 thread 1: '/u03/app/oracle/oradata/dbacoe/redo02.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 3 of thread 1
    ORA-00312: online log 3 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo03_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 3 thread 1: '/u03/app/oracle/oradata/dbacoe/redo03.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 3 of thread 1
    ORA-00312: online log 3 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo03_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 3 thread 1: '/u03/app/oracle/oradata/dbacoe/redo03.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ARC3: Archival started
    ARC0: STARTING ARCH PROCESSES COMPLETE
    Addition:
    因为是实验用库,所以结构有点怪,大体如下:
    Primary
    数据文件在BASE所在的oradata下 即/u03/app/oracle下,
    redo文件每组两份,一份在oradata下,另一份在同机的/u04/app/oracle/redundancy中(以_s结尾命名)
    archive文件在/u04/app/oracle/fast_recovery_area中
    standby
    数据文件在/u05/oracle/app/oracle/oradata下
    redundancy中的redo文件随着该目录放到oradata下的red中了
    archive文件目录则是/u05/oracle/app/oracle/fast_recovery_area中
    接下来是对应的init文件与TNS信息
    Primary端的init主要信息
    db_unique_name='PRIMARY'
    fal_client='PRIMARYSV'
    fal_server='STANDBYSV'
    log_archive_config='DG_CONFIG=(PRIMARY,STANDBY)'
    log_archive_dest_1='location=/u04/app/oracle/fast_recovery_area/DBACOE/archivelog VALID_FOR=(ALL_LOGFILES,ALL_ROLES) db_unique_name=PRIMARY'
    log_archive_dest_2='service=STANDBY ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) db_unique_name=STANDBY'
    log_archive_dest_state_1='enable'
    log_archive_dest_state_2='enable'
    db_file_name_convert='/u05/oracle/app/oracle/oradata/dbacoe','/u03/app/oracle/oradata/dbacoe',/u05/oracle/app/oracle/oradata/red','/u04/app/oracle/redundancy'
    log_file_name_convert='/u05/oracle/app/oracle/fast_recovery_area/DBACOE/archivelog','/u04/app/oracle/fast_recovery_area/DBACOE/archivelog'
    Standby端的init信息
    log_archive_config='DG_CONFIG=(PRIMARY,STANDBY)'
    db_unique_name='STANDBY'
    db_file_name_convert='/u03/app/oracle/oradata/dbacoe','/u05/oracle/app/oracle/oradata/dbacoe','/u04/app/oracle/redundancy','/u05/oracle/app/oracle/oradata/red'
    log_file_name_convert='/u04/app/oracle/fast_recovery_area/DBACOE/archivelog','/u05/oracle/app/oracle/fast_recovery_area/DBACOE/archivelog'
    log_archive_dest_1='location=/u05/oracle/app/oracle/fast_recovery_area/DBACOE/archivelog VALID_FOR=(ALL_LOGFILES,ALL_ROLES) db_unique_name=STANDBY'
    log_archive_dest_state_1=enable
    log_archive_format=log%t_%s_%r.arc
    log_archive_dest_2='SERVICE=PRIMARYSV ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PRIMARY'
    log_archive_dest_state_2=enable
    TNS信息如下(均能tnsping通)
    PRIMARYSV =
    +(DESCRIPTION =+
    +(ADDRESS = (PROTOCOL = TCP)(HOST = db1.ad.xxxxxx.com)(PORT = 1521))+
    +(CONNECT_DATA =+
    +(SERVER = DEDICATED)+
    +(SERVICE_NAME = dbacoe)+
    +)+
    +)+
    STANDBYSV =
    +(DESCRIPTION =+
    +(ADDRESS = (PROTOCOL = TCP)(HOST = db2.ad.xxxxx.com)(PORT = 1529))+
    +(CONNECT_DATA =+
    +(SERVER = DEDICATED)+
    +(SERVICE_NAME = dbacoe)+
    +)+
    +)+
    PS:请大家花费一点点时间阅读一下这冗长的段子,刚开学data guard,参数上设置都很生疏,(光照着文档做了,是否有redo log相关的操作遗漏?)敬请指教
    Edited by: 961394 on Dec 10, 2012 1:52 AM
    Edited by: 961394 on 2012-12-10 上午4:58

    结贴了,感谢maclean一群的Jesse Lui帮我解决问题

  • Logical standby stops deleting archives automatically

    We are ruining 11.1.0.7 on windows 2008 64 bit
    The issue I am facing is we have these settings implemented
    log_auto_delete = true
    log_auto_delete_retention_target = 5
    Logical standby works fine and deletes the archive logs coming from primary but after some day it stops working. I have to stop and start the database in order to make it work. Does anybody has this kind of situation? Please help.

    Hello;
    This might be worth a look:
    Bug 13448652 - LOG_AUTO_DELETE setting reverts to FALSE on Logical Standby (Doc ID 13448652.8)
    Best Regards
    mseberg

  • About Data Guard - Physical Standby Database

    Dear All,
    I have read many documents regading Data Guard.
    I am about to setup Data Guard in our current environment but want to clear few tings.
    I have a confusion between physical and logical standby database.
    What I have read from different documents is:
    -Physical Standby Database in (11g)
    1) It is the most efFicient
    2) I can be either in mount or open state
    3) Select queries and reporting can be done to improve performance of the primary database.
    4) Schema on primary and standby database is always the same
    -- Logical Standby Database
    1) You can create additional tables, indexes, etc.
    2) Always in open mode
    3) Select queries and reporting can be done to improve performance of the primary database
    Now our scenerio is, that we have one server at the moment, OS is linux and operating system is 11g. We want to setup an other server in another country, will also be 11g on Linux in a way that it acts as standby/backup server. Schema and data is always the same. In case due to unavailability of primary server, standby server acts as the primary server (This has to be automated). The reason for unavailability could be any like maintenance work on primary server, network or hardware failure at primary server. The last and the most important thing is that users from this country where we will setup standby database will insert/update data on the primary server BUT queries and reporting will be done from this newly created standby data.
    Kindly recomend the best Data Guard in this scenerio and kindly correct me where I am wrong.
    Thanks, Imran

    A logical standby has various limitations on things like data types. It's also a much more complex architecture, which makes it more likely that something will break periodically and require attention. Applying redo to a physical standby is code that has been around forever and is as close to bullet-proof as you'll get. And you would generally prefer to fail over to a physical standby-- if you do things like create new objects in the logical standby, you may have to get rid of those objects during a failover to get acceptable OLTP performance.
    Justin

  • Data Guard Physical Standby Failover

    I need clarification on Physical Standby Failovers. Say I have a Primary db (A) and a physical standby (B), (A) fails and (B) is now the primary, can (A) be setup to become a standby of (B)? I've read 2 different statements in Oracle 10g Data Guard Concepts and Admin
    "During failovers involving a physical standby database: In all cases, after a failover, the original primary database can no longer participate in the Data Guard configuration."
    vice
    "After a fast-start failover occurs, the old primary database will automatically reconfigure itself as a new standby database upon reconnection to the configuration."
    Much thanks.

    In all cases, after a failover, the original primary database can no longer participate in the Data Guard configuration.
    TRUE ! But this doesn't mean that you cannot make your failed primary a new standby database !.
    You can achieve this using FLASHBACK DATABASE option:
    From the documentation:
    After a failover occurs, the original primary database can no longer participate in the Data Guard configuration until it is repaired and established as a standby database in the new configuration. To do this, you can use the Flashback Database feature to recover the failed primary database to a point in time before the failover occurred, and then convert it into a physical or logical standby database in the new configuration.
    Read the following documents:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/role_management.htm#sthref995
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/scenarios.htm#i1049997
    In addition to this I'd like to mention that you can achieve the same with doing an incomplete recovery instead of using FLASHBACK DATABASE.
    "After a fast-start failover occurs, the old primary database will automatically reconfigure itself as a new standby database upon reconnection to the configuration."
    TRUE. If you enable Fast-Start Failover:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14230/sofo.htm#CHDDFFEC
    Cheers!

  • Data Guard phys. standby creation failing in 10g EM?

    All,
    I'm working on setting up Oracle Data Guard in a test environment at my company. It was suggested to me to go through the GUI to set up a physical standby database and automate the management of my DG setup.
    However, the several attempts that I have made result in the process failing during the "Create Standby Database" and the process fails with the following errors:
    dgcreate.recoverStby: ALTER DATABASE RECOVER CANCEL
    dgcreate.recoverStby: ALTER DATABASE OPEN READ ONLY
    SQL Error: ORA-16004: backup database requires recovery
    ORA-01152: file 1 was not restored from a sufficiently old backup
    ORA-01110: data file 1: '/u01/app/oracle/product/10.2.0/db_1/oradata/oratstdg/system01.dbf' (DBD ERROR: OCIStmtExecute)
    Can anyone pass along any ideas as to where the issue might be? I would really appreciate it - I need this up and going ASAP. Thanks!
    Matt Gordon

    Ugonic,
    I actually did create a new Physical Standby Database. I did not have an existing database configured as a standby database. In fact, the standby server that I used only had the Oracle binaries installed on it - there was no database on that system.
    Additionally, I did not have a saved configuration to use, so I generated completely new files and a new configuration.
    I left the screen alone for over an hour before checking the status and seeing these errors in the log - I am wondering if anybody else has encountered this same issue.
    On a side note, I've stumbled my way further into this process and now have a functional DG configuration. However, I can only switchover from SQL*Plus - the GUI way fails. When I click on Verify in the GUI I get the following log:
    Initializing.
    Connected to instance fcoracle1.cdps.cdp:oratest1
    Starting alert log monitor...
    Updating Data Guard link on database homepage...
    WARNING: Database oratest2.cdps.cdp is not discovered.
    Data Protection Settings:
    Protection mode : Maximum Performance
    Log Transport Mode settings:
    oratest1.cdps.cdp: ASYNC
    oratest2.cdps.cdp: ASYNC
    Checking standby redo logs.....OK
    Checking Data Guard status
    oratest1.cdps.cdp : Normal
    oratest2.cdps.cdp : Normal
    Checking Inconsistent Properties
    Checking agent status
    oratest1.cdps.cdp ... WARNING: Undefined subroutine &main::executeSQLPlusSYSDBA called at - line 5.
    WARNING: Switchover or failover may not succeed as a result.
    oratest2.cdps.cdp ... WARNING: Undefined subroutine &main::executeSQLPlusSYSDBA called at - line 5.
    WARNING: Switchover or failover may not succeed as a result.
    Switching log file 235.Done
    Checking applied log on oratest2.cdps.cdp....OK
    Processing completed.
    Have you experienced this problem and/or do you have any suggestions to resolve it?
    Matt

Maybe you are looking for

  • Issue with BBM 7 Compatibly with Blackberri​es with OS6 systems (BOLD 9780)

    I downloaded BBM 7 on my blackberry and the icon simply dissapeared, i dont know what to do, do i try to re-download it or do i simply try to re-install bbm 6? eitherways i have tired to do both and neither have worked, please help. URGENT

  • Parallel Approval In Quoting

    Hi All, Our customer requires Parallel Approval in quoting. (Approval mail to be sent to more than one approver). Customer will be using AME. Is there any way we can achieve this? Please confirm. - Mahesh Jayaraman

  • HT1296 syncing from computer to phone

    1.  Why does my iPhone 4 not show up on My Computer as a Device?  And yet it beeps accordingly, and iTunes starts up, so I must be connected? 2.  I want to move music from my computer onto my iPhone, but sync'ing does nothing, as though I am not conn

  • Kernel Panic When Printing -- Please Help

    I used to have the Kernel Panic problem all the time because I stupidly did not keep my computer on a hard surface. Once I started leaving it directly on my desk, the kernel panic subsided. Now it's back. Every time I try to print something, a few pa

  • Import stops after a few seconds

    I just bought a Canon Elura 80 a few months ago, and have just taped an hour of incredibly sweet baby stuff from our newborn onto a used tape (oh boy, I hope it's not the tape...) anyway, I had no trouble the first time importing using iMovie HD from