Dataguard archive logs

Hi,
in dataguard using oracle 10g r2
when i query on primary side:
select max(sequence#) from v$log where archived='YES';MAX(SEQUENCE#)
19
standby side:
select max(sequence#) from v$log where archived='YES';MAX(SEQUENCE#)
20
why it is showing different log sequences, it has to show similar sequences right?

venky wrote:
Hi,
in dataguard using oracle 10g r2
when i query on primary side:
select max(sequence#) from v$log where archived='YES';MAX(SEQUENCE#)
19
standby side:
select max(sequence#) from v$log where archived='YES';MAX(SEQUENCE#)
20
why it is showing different log sequences, it has to show similar sequences right?Do you realize there is an entire forum just for Data Guard: Data Guard

Similar Messages

  • Dataguard physical standby archive log question

    Hi all,
    I will try to keep this simple..
    I have a 4 node RAC primary shipping logs to a 2 node physical standby.
    On the primary when I run 'alter system archive log current' on an instance I only see 1 log being applied on the standby, that is by querying v$archived_log.
    If I run the following on the standby:
    select thread#,sequence#,substr(name,43,70)"NAME",registrar,applied,status,first_time from v$archived_log where first_time
    in
    (select max(first_time) from v$archived_log group by thread#)
    order by thread#
    I get:
    THREAD# SEQUENCE# NAME REGISTR APPLIED S FIRST_TIME
    1 602 thread_1_seq_602.2603.721918617 RFS YES A 17-jun-2010 12:56:58
    2 314 thread_2_seq_314.2609.721918627 RFS NO A 17-jun-2010 12:56:59
    3 311 thread_3_seq_311.2604.721918621 RFS NO A 17-jun-2010 12:57:00
    4 319 thread_4_seq_319.2606.721918625 RFS NO A 17-jun-2010 12:57:00
    Why do we only see the max(sequence#) having been applied and not all of them?
    This is the same no matter how many times I archive the current log files on any of the instances on the primary and also the standby does not have any gaps.
    Hope this is clear..
    any ideas?
    jd

    ok output from gv$archived_log on standby BEFORE 'alter system archive log current' on primary
    THREAD# SEQUENCE# NAME REGISTR APPLIED S FIRST_TIME
    1 679 thread_1_seq_679.1267.722001505 RFS NO A 18-jun-2010 11:58:22
    1 679 thread_1_seq_679.1267.722001505 RFS NO A 18-jun-2010 11:58:22
    2 390 thread_2_seq_390.1314.722001507 RFS NO A 18-jun-2010 11:58:23
    2 390 thread_2_seq_390.1314.722001507 RFS NO A 18-jun-2010 11:58:23
    3 386 thread_3_seq_386.1266.722001505 RFS YES A 18-jun-2010 11:58:22
    3 386 thread_3_seq_386.1266.722001505 RFS YES A 18-jun-2010 11:58:22
    4 393 thread_4_seq_393.1269.722001507 RFS NO A 18-jun-2010 11:58:23
    4 393 thread_4_seq_393.1269.722001507 RFS NO A 18-jun-2010 11:58:23
    Output from v$archived_log on standby AFTER 'alter system archive log current' on primary
    THREAD# SEQUENCE# NAME REGISTR APPLIED S FIRST_TIME
    1 680 thread_1_seq_680.1333.722004227 RFS NO A 18-jun-2010 11:58:29
    1 680 thread_1_seq_680.1333.722004227 RFS NO A 18-jun-2010 11:58:29
    2 391 thread_2_seq_391.1332.722004227 RFS NO A 18-jun-2010 11:58:30
    2 391 thread_2_seq_391.1332.722004227 RFS NO A 18-jun-2010 11:58:30
    3 387 thread_3_seq_387.1271.722004225 RFS NO A 18-jun-2010 11:58:28
    3 387 thread_3_seq_387.1271.722004225 RFS NO A 18-jun-2010 11:58:28
    4 394 thread_4_seq_394.1270.722004225 RFS YES A 18-jun-2010 11:58:29
    4 394 thread_4_seq_394.1270.722004225 RFS YES A 18-jun-2010 11:58:29
    as a reminder we have a 4 node RAC system shipping logs to a 2 node RAC standby. There are no gaps but only one log is ever registered as being applied.
    Why is that? Why arnt all logs registered as being applied?

  • Dataguard sync and applying archive log

    Just did a check on my primary and standby using:
    Primary:
    SELECT max(sequence#) AS "PRIMARY" FROM v$log_history;
    PRIMARY
    45527
    Standby:
    SELECT max(sequence#) AS "STANDBY", applied FROM v$archived_log GROUP BY applied:
    STANDBY APP
    45476 NO
    45528 YES
    My question is how do I fix my sync?

    SQL> select OPEN_MODE  from v$database;
    OPEN_MODE
    MOUNTED
    SQL> recover standby database until cancel;
    ORA-01153: an incompatible media recovery is active
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;
    Database altered.
    SQL> recover standby database until cancel;
    ORA-00279: change 4558008 generated at 08/21/2010 14:02:39 needed for thread 1
    ORA-00289: suggestion : /data_gpfs/oracle/archivelogs/1_112_726952005.dbf
    ORA-00280: change 4558008 for thread 1 is in sequence #112
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    */data_gpfs/oracle/archivelogs/2_118_726952005.dbf*
    ORA-00325: archived log for thread 1, wrong thread # 2 in header
    ORA-00334: archived log: '/data_gpfs/oracle/archivelogs/2_118_726952005.dbf'

  • *HOW TO DELETE THE ARCHIVE LOGS ON THE STANDBY*

    HOW TO DELETE THE ARCHIVE LOGS ON THE STANDBY
    I have set the RMAN CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON STANDBY; on my physical standby server.
    My archivelog files are not deleted on standby.
    I have set the CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default on the Primary server.
    I've checked the archivelogs with the FRA and they are not beign deleted on the STANDBY. Do I have to do something for the configuation to take effect? Like run a RMAN backup?
    I've done a lot ofresearch and i'm getting mixed answers. Please help. Thanks in advanced.
    J

    Setting the Policy will not delete the Archive logs on the Standby. ( I found a thread where the Data Guard product manager says "The deletion policy on both sides will do what you want" ). However I still
    like to clean them off with RMAN.
    I would use RMAN to delete them so that it can use that Policy are you are protected in case of Gap, transport issue etc.
    There are many ways to do this. You can simply run RMAN and have it clean out the Archive.
    Example :
    #!/bin/bash
    # Name: db_rman_arch_standby.sh
    # Purpose: Database rman backup
    # Usage : db_rman_arch_standby <DBNAME>
    if [ "$1" ]
    then DBNAME=$1
    else
    echo "basename $0 : Syntax error : use . db_rman_full <DBNAME> "
    exit 1
    fi
    . /u01/app/oracle/dba_tool/env/${DBNAME}.env
    echo ${DBNAME}
    MAILHEADER="Archive_cleanup_on_STANDBY_${DBNAME}"
    echo "Starting RMAN..."
    $ORACLE_HOME/bin/rman target / catalog <user>/<password>@<catalog> << EOF
    delete noprompt ARCHIVELOG UNTIL TIME 'SYSDATE-8';
    exit
    EOF
    echo `date`
    echo
    echo 'End of archive cleanup on STANDBY'
    mailx -s ${MAILHEADER} $MAILTO < /tmp/rmandbarchstandby.out
    # End of ScriptThis uses ( calls an ENV) so the crontab has an environment.
    Example ( STANDBY.env )
    ORACLE_BASE=/u01/app/oracle
    ULIMIT=unlimited
    ORACLE_SID=STANDBY
    ORACLE_HOME=$ORACLE_BASE/product/11.2.0.2
    ORA_NLS33=$ORACLE_HOME/ocommon/nls/admin/data
    LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib
    LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/lib
    LIBPATH=$LD_LIBRARY_PATH:/usr/lib
    TNS_ADMIN=$ORACLE_HOME/network/admin
    PATH=$ORACLE_HOME/bin:$ORACLE_BASE/dba_tool/bin:/bin:/usr/bin:/usr/ccs/bin:/etc:/usr/sbin:/usr/ucb:$HOME/bin:/usr/bin/X11:/sbin:/usr/lbin:/GNU/bin/make:/u01/app/oracle/dba_tool/bin:/home/oracle/utils/SCRIPTS:/usr/local/bin:.
    #export TERM=linux=80x25 wrong wrong wrong wrong wrong
    export TERM=vt100
    export ORACLE_BASE ORACLE_SID ORACLE_TERM ULIMIT
    export ORACLE_HOME
    export LIBPATH LD_LIBRARY_PATH ORA_NLS33
    export TNS_ADMIN
    export PATH
    export MAILTO=?? your email hereNote use the env command in Unix to get you settings.
    There are probably ten other/better ways to do this, but this works.
    other options ( you decide )
    Configure RMAN to purge archivelogs after applied on standby [ID 728053.1]
    http://www.oracle.com/technetwork/database/features/availability/rman-dataguard-10g-wp-1-129486.pdf
    Maintenance Of Archivelogs On Standby Databases [ID 464668.1]
    Tip I don't care myself but in some of the other forums people seem to mind if you use all caps in the subject. They say it shouting. My take is if somebody is shouting at me I'm probably going to just move away.
    Best Regards
    mseberg
    Edited by: mseberg on May 8, 2012 11:53 AM
    Edited by: mseberg on May 8, 2012 11:56 AM

  • Shell script for archive log transfer

    hi
    I dont want to reinvent the wheel.
    I am looking for shell script for log shipping to provide standby db.
    What I want to do is, get the last applied archived log number from alert.log
    Copy the files from archive destination according to this value.
    Cheers

    If you don't want to re-invent the wheel you use Dataguard, no scripts.
    And your script should use the dictionary, instead of some bs method to read the alert.
    v$archived_log has all information!
    Also as far as I know, the documentation describes manual standby.
    So apparently you not only don't want to reinvent the wheel, but you want the script on a silver plate on your doorstep!
    Typical attitude of most DBAs here. Use OTN for a permanent vacation.
    Sybrand Bakker
    Senior Oracle DBA

  • Archived log to standby server stopped

    Hi,
    I configured dataguard standby server few weeks ago,
    I set log_archive_dest_2='SERVICE=JHRD0107_OFS1 OPTIONAL'
    and log_archive_dest_state=ENABLE in primary server,
    and very go fine until yesterday, I shutdown my standby server and set the log_archive_dest_state=DEFER,
    when I tried to startup again the standby server, and set log_archive_dest_state=ENABLE again,
    the archive stop sending the archived log to log_archive_dest_2.
    No error was found in alert log of primary and standby server.
    It seem that the primary server don't know there is another archive destination..
    The database version is 10g 10.2.0.4
    Anyone what could be the cause of this?
    Thanks

    Hi, Below is the parameter related to archive in my primary server
    SQL> alter system set log_archive_dest_state_2=enable;
    System altered.
    SQL> show parameter archive
    NAME TYPE VALUE
    archive_lag_target integer 0
    log_archive_config string
    log_archive_dest string
    log_archive_dest_1 string LOCATION=I:\OFSDB\OFS1\archive
    MANDATORY REOPEN=300
    log_archive_dest_10 string LOCATION=USE_DB_RECOVERY_FILE_
    DEST OPTIONAL REOPEN=300
    log_archive_dest_2 string SERVICE=JHRD0107_OFS1 OPTIONAL
    log_archive_dest_3 string
    log_archive_dest_4 string
    log_archive_dest_5 string
    NAME TYPE VALUE
    log_archive_dest_6 string
    log_archive_dest_7 string
    log_archive_dest_8 string
    log_archive_dest_9 string
    log_archive_dest_state_1 string enable
    log_archive_dest_state_10 string enable
    log_archive_dest_state_2 string ENABLE
    log_archive_dest_state_3 string enable
    log_archive_dest_state_4 string enable
    log_archive_dest_state_5 string enable
    log_archive_dest_state_6 string enable
    NAME TYPE VALUE
    log_archive_dest_state_7 string enable
    log_archive_dest_state_8 string enable
    log_archive_dest_state_9 string enable
    log_archive_duplex_dest string
    log_archive_format string ARC%S_%R.%T
    log_archive_local_first boolean TRUE
    log_archive_max_processes integer 2
    log_archive_min_succeed_dest integer 1
    log_archive_start boolean FALSE
    log_archive_trace integer 0
    remote_archive_enable string true
    NAME TYPE VALUE
    standby_archive_dest string %ORACLE_HOME%\RDBMS
    Thanks for your response.
    Vincent

  • Archived log not applying

    archived logs are not apllying in DR
    SQL> select max(sequence#) from v$archived_log where applied='YES';
    MAX(SEQUENCE#)
    SQL>
    SQL> select process,status,sequence# from v$managed_standby;
    PROCESS   STATUS        SEQUENCE#
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    RFS       IDLE                  0
    RFS       IDLE                  0
    RFS       IDLE              34473
    MRP0      WAIT_FOR_GAP      31753
    RFS       IDLE                  0
    RFS       IDLE                  0
    RFS       IDLE                  0
    11 rows selected.

    Hello,
    As previously posted, you have a GAP. So, you may have to investigate why the Archived logs cannot be applied.
    To do so, you may check for the File System if any Archive destination is FULL. Also, you may have to verify the parameters on the PRIMARY and STANDBY site:
    sqlplus "/ as sysdba"
    show parameter fal
    show parameter dest
    show parameter config
    show parameter broker
    Check also the Alert log on both Primary and Standby.
    If the DATAGUARD BROKER is enabled you may also execute this
    dgmgrl
    show configuration verbose
    Also try to "tnsping" the services used in the Archive Destination from both site.
    When you are sure that everything is correct then, verify if any Archived logs is missing and try to restore it if possible.
    Else if everything is broken, and there's no way to get back the missing Archived logs, you 'll have to "rebuild" the Standby from a new Backup of the PRIMARY.
    Hope this help.
    Best regards,
    Jean-Valentin Lubiez

  • Archived log missed in standby database

    Hi,
    OS; Windows 2003 server
    Oracle: 10.2.0.4
    Data Guard: Max Performance
    Dataguard missed some of the archivelog files and but latest log files are applying. standby database is not in sync with primary.
    SELECT LOCAL.THREAD#, LOCAL.SEQUENCE# FROM (SELECT THREAD#, SEQUENCE# FROM V$ARCHIVED_LOG WHERE DEST_ID=1) LOCAL WHERE LOCAL.SEQUENCE# NOT IN (SELECT SEQUENCE# FROM V$ARCHIVED_LOG WHERE DEST_ID=2 AND THREAD# = LOCAL.THREAD#);
    I queried above command and I found some files are missed in standby.
    select status, type, database_mode, recovery_mode,protection_mode, srl, synchronization_status,synchronized from V$ARCHIVE_DEST_STATUS where dest_id=2;
    STATUS TYPE DATABASE_MODE RECOVERY_MODE PROTECTION_MODE SRL SYNCHRONIZATION_STATUS SYN
    VALID PHYSICAL MOUNTED-STANDBY MANAGED MAXIMUM PERFORMANCE NO CHECK CONFIGURATION NO
    Anyone can tell me how to apply those missed archive log files.
    Thanks in advacne

    Deccan Charger wrote:
    I got below error.
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION
    ERROR at line 1:
    ORA-01153: an incompatible media recovery is activeYou need to essentially do the following.
    1) Stop managed recovery on the standby.
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;2) Resolve the archive log gap - if you have configured FAL_SERVER and FAL_CLIENT Oracle should do this when you follow step 3 below, as you've manually copied the missed logs you should be OK
    3) restart managed recovery using the command shown above.
    You can monitor archive log catchup using the alert.log or your original query.
    Niall Litchfield
    http://www.orawin.info/
    Edited by: Niall Litchfield on May 4, 2010 2:29 PM
    missed tag

  • Archive logs are not transferred to STDBY database

    Hi,
    I have create a STDBY database (I am running the release 9.2.0.7.0).
    I see that the archivelogs are not correctly transferred into STDBY server.
    From Primary alert log, I see the following error:
    ARC1: Evaluating archive log 1 thread 1 sequence 16734
    ARC1: LGWR is actively archiving destination LOG_ARCHIVE_DEST_2
    ARC1: Destination LOG_ARCHIVE_DEST_2 archival not expedited
    ARC1: Beginning to archive log 1 thread 1 sequence 16734
    Creating archive destination LOG_ARCHIVE_DEST_1: '/backup/archivelogs/log1_16734.arc'
    ARC1: LGWR is actively archiving destination LOG_ARCHIVE_DEST_2
    Invoking non-expedited destination LOG_ARCHIVE_DEST_2 thread 1 sequence 16734 host STDBY_PROD
    ARC1: Completed archiving log 1 thread 1 sequence 16734
    Thu Nov 17 14:54:42 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-03114: not connected to ORACLE
    Thu Nov 17 14:54:42 2011
    ARC0: FAL archive failed, see trace file.
    ARCH: FAL archive failed. Archiver continuing
    Thu Nov 17 14:54:42 2011
    ORACLE Instance PROD - Archival Error. Archiver continuing.
    ARCH: Connecting to console port...
    Thu Nov 17 14:54:42 2011
    ORA-16055: FAL request rejected
    ARCH: Connecting to console port...
    Thu Nov 17 14:54:42 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-16055: FAL request rejected
    ARC0: Begin FAL archive (thread 1 sequence 16483 destination STDBY_PROD)
    Creating archive destination LOG_ARCHIVE_DEST_2: 'STDBY_PROD'
    Thu Nov 17 15:05:44 2011
    LGWR: I/O error 3114 archiving log 2 to 'STDBY_PROD'
    Thu Nov 17 15:05:44 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_lgwr_5265.trc:
    ORA-03114: not connected to ORACLE
    Thu Nov 17 15:10:08 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-03114: not connected to ORACLE
    Thu Nov 17 15:10:08 2011
    ARC0: FAL archive failed, see trace file.
    ARCH: FAL archive failed. Archiver continuing
    Thu Nov 17 15:10:08 2011
    ORACLE Instance PROD - Archival Error. Archiver continuing.
    ARCH: Connecting to console port...
    Thu Nov 17 15:10:08 2011
    ORA-16055: FAL request rejected
    ARCH: Connecting to console port...
    Thu Nov 17 15:10:08 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-16055: FAL request rejected
    I see that both archiving destinations are VALID:
    SQL> select dest_id, dest_name, status from v$archive_dest_status where status != 'INACTIVE';
    DEST_ID ---------- DEST_NAME -------------------------------------------------------------------------------- STATUS
    1 LOG_ARCHIVE_DEST_1 VALID
    2 LOG_ARCHIVE_DEST_2 VALID
    SQL>
    TNSPING works properly in both directions, e.g. from PROD -> STDBY:
    $tnsping STDBY_PROD
    TNS Ping Utility for Linux: Version 9.2.0.7.0 - Production on 17-NOV-2011 16:56:40
    Copyright (c) 1997 Oracle Corporation. All rights reserved.
    Used parameter files:
    /mnt/orclEBS/oracle/proddb/9.2.0/network/admin/PROD_ebslive/sqlnet_ifile.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION= (ADDRESS=(PROTOCOL=tcp) (HOST=EBSSTDBY) (PORT=1521)) (CONNECT_DATA=(SID=PROD)))
    OK (10 msec)
    Please help me to solve this issue.
    Thank you.

    A few notes :
    Troubleshooting Tips For Dataguard Switchover (9i and 10gR1) [ID 298986.1]
    Although it is about Switchover, it does mandate a password file exclusive. If the switchover requires the password file, shouldn't ARCH shipping require it ?
    Step By Step Guide To Create Physical Standby Database Using RMAN [ID 469493.1]
    also covers 9i with specific mentions of 9i differences from 10g. Passwordfile is not mentioned as different in 9i from 10g.
    Data Guard 9i - Net8 Configuration for a 2-node database environment. [ID 175122.1
    has log_archive_dest_2='SERVICE=NODE2STDBY.ACME.COM'    so it connects to the service
    *However* the document "Introduction to Oracle 9i Data Guard Manager" from
    Data Guard 9i Introduction to Data Guard Manager GUI [ID 150217.1]
    page 59 has this : "The creation process will attempt to create a remote login password file for the standby database to
    enable remote connections to the database. (Note: A remote login password file is not necessary for
    Data Guard operation; it is only needed to allow remote connections to the database from clients,
    such as other Enterprise Manager tools.)"
    Similarly the section "3.2.6 Set Initialization Parameters on a Physical Standby Database"
    in the 9i DataGuard manual at http://download.oracle.com/docs/cd/B10501_01/server.920/a96653/create_ps.htm#62941
    only has
    remote_archive_enable=TRUE
    (this is an instance parameter)
    Similarly the section "5.8.2.1 Primary Database Initialization Parameters"
    at http://download.oracle.com/docs/cd/B10501_01/server.920/a96653/log_transport.htm#1067609
    explains
    The last parameter, REMOTE_ARCHIVE_ENABLE=SEND, allows the primary database to send redo data to the standby database, but prevents the primary database from receiving redo data from another system.
    with REMOTE_ARCHIVE_ENABLE=RECEIVE on the Standby
    to Receive and archive the incoming redo data from the primary database, but only while the database is running in the standby role
    So, I guess that REMOTE_ARCHIVE_ENABLE is the important one. Not REMOTE_LOGIN_PASSWORDFILE.
    In 9i : http://download.oracle.com/docs/cd/B10501_01/server.920/a96536/ch1174.htm#1023087
    In 10g it is deprecated : http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/initparams176.htm#sthref541 so that LOG_ARCHIVE_CONFIG is used.
    (and no longer appears in the 11.2 Reference)
    Hemant K Chitale

  • Oracle recommended location for archive logs in  oracle 10g rac

    Hello All,
    We would like to know the oracle recommended location for the archive logs in oracle10g RAC .we are using ASM.
    Thanks...

    user4487322 wrote:
    thanks. Is it the recommended setting ,if we go for a DR setup?I mean archive logs in ASM.If you can use dataguard, the archivelog copy to the standby system would be handled by Oracle and it supports ASM.
    Just remember, what ever your strategy, the archivelogs must be in a SHARED location (where all nodes can read/write to this location.)

  • How to determine which archive logs are needed in flashback.

    Hi,
    Let's assume I have archive logs 1,2,3,4, then a "backup database plus archivelogs" in RMAN, and then archive logs 5+6. If I want to flashback my database to a point immediately after the backup, how do I determine which archive logs are needed?
    I would assume I'd only need archive logs 5 and/or 6 since I did a full backup plus archivelogs and the database would have been checkpointed at that time. I'd also assume archive logs 1,2,3,4 would be obsolete as they would have been flushed to the datafiles in the checkpoint.
    Are my assumptions correct? If not what queries can I run to determine what files are needed for a flashback using the latest checkpointed datafiles?
    Thanks.

    Thanks for the explanation, let me be more specific with my problem.
    I am trying to do a flashback on a failed primary database, the only reason why I want to do a flashback is because dataguard uses the flashback command to try and synchronize the failed database. Specifically dataguard is trying to run:
    FLASHBACK DATABASE TO SCN 865984
    But it fails, if I run it manually then I get:
    SQL> FLASHBACK DATABASE TO SCN 865984;
    FLASHBACK DATABASE TO SCN 865984
    ERROR at line 1:
    ORA-38754: FLASHBACK DATABASE not started; required redo log is not available
    ORA-38761: redo log sequence 5 in thread 1, incarnation 3 could not be accessed
    Looking at the last checkpoint I see:
    CHECKPOINT_CHANGE#
    865857
    Also looking at the archive logs:
    RECID STAMP THREAD# SEQUENCE# FIRST_CHANGE# FIRST_TIM NEXT_CHANGE# RESETLOGS_CHANGE# RESETLOGS
    25 766838550 1 1 863888 10-NOV-11 863892 863888 10-NOV-11
    26 766838867 1 2 863892 10-NOV-11 864133 863888 10-NOV-11
    27 766839225 1 3 864133 10-NOV-11 864289 863888 10-NOV-11
    28 766839340 1 4 864289 10-NOV-11 864336 863888 10-NOV-11
    29 766840698 1 5 864336 10-NOV-11 865640 863888 10-NOV-11
    30 766841128 1 6 865640 10-NOV-11 865833 863888 10-NOV-11
    31 766841168 1 7 865833 10-NOV-11 865857 863888 10-NOV-11
    How can I determine what archive logs are needed by a flashback command? I deleted any archive logs with a SCN less than the checkpoint #, I can restore them from backup but I am trying to figure out how to query what is required for a flashback. Maybe this coincides with the point that flashbacks have nothing to do with the backups of datafiles or the checkpoints?

  • Archive log can not ship to GAP logfiles to standby DB automatically

    we have a non-real time standby database, which will receive the archive file from the primary database server most of the time, and will apply the logfiles only at one point of time daily.
    Some times, we need to shutdown the Standby DB server for a while ( 3-4 hours).
    The missed logfiles will catch up during the standby down time later.
    But since last week we had an storage incident, the primary DB server stops to catch up the missed logfiles, and saw this message at the archvie trace file:
    ABC: tkrsf_al_read: No mirror copies to re-read data
    Current, we found the archive log gaps on the standby server, and have to manually copy those logfiles over and register them.
    Saw some tips on the internet to change the parameter "log_archive_max_processes", but no help for us at all.
    Here is the parameter on the Primary DB server:
    log_archive_dest_2 = SERVICE=Standby_server reopen=300

    which will receive the archive file from the primary database server most of the timeMost times from primary. Then remaining times? so you Copy manually and register?
    Then it's dataguard not an manual standby.
    Error 1034 received logging on to the standby
    Errors in file /******/***arc210536.trc:
    ORA-01034: ORACLE not available
    FAL[server, ARC2]: FAL archive failed, see trace file.these errors in primary when , stanby is down and when primary tries to connect to standby, so tese errors not considerable to investigate ,
    When you don't want to apply archives on standby no need to shutdown. Just put this value log_archive_dest_state_2='defer'
    Once you enable check what are the errors in primary alert log file.
    How is your network band width speed? Is it capable to hold that much archive data?
    May be it will take some time when you pause & start.
    Also use LGWR in log_archive_dest_2 for real time apply after creatin standby redo logs.
    So post the alert log information once you enable standby database.

  • How to delete archive log on standby site

    hi
    i configured dataguard on 10g. could anyone suggest how to delete archived log on standby site..

    Hi,
    There is no automatic feature out-of-the-box but you have the following views:
    1) Shows archive logs transfered:
    select sequence#, first_time, next_time from v$archived_log order by sequence#;
    2) Shows archive logs applied:
    select sequence#, archived, applied from v$archived_log order by sequence#;
    If you see your archivelogs applied you can safely remove archive log files from your disk. It is your decision if you keep or delete them completely.

  • Archive log not shipped to standby.

    Dear Experts,
    We have are using oracle 10.2.0.4.0 on AIX.
    And we have configured physical standby database which was working fine till date.
    suddenly i found that the MRP process is waiting for gap.
    i have all the archive logs on the primary system and all the logs are getting shipped to standby database expect one file.
    /oracle/PRD/saparch/PRDarch1_46811_755493268.dbf only this file is not getting shipped to standby database....due to which the standby is waiting for this log.
    log applied till 46810
    other logs are getting shipped.
    However i tried to manually copy this log to standby but it didnt start applying this log.
    and i heard that if i copy the log file manually then i need to register it...please let me know how to register it if it is the correct method.
    Please help to overcome this issue....
    Thanks,
    Shankar.
    Edited by: Shankar Viji on Jun 19, 2012 10:34 PM

    Shankar Viji wrote:
    Dear Mesberg,
    Is there any way to over come this bug.
    Bug 6851669 - ORA-308 / ORA-27037 from Dataguard GAP resolution [ID 6851669.8]
    if not i am planing to take offline backup of my primary and restore it on standby and start the DR again.
    Please suggest is that the correct way to do it r not.
    Thanks,
    Shankar.
    Edited by: Shankar Viji on Jun 21, 2012 10:51 PMHi,
    It looks for me that archive unable to read, it may be because of any header verification or corrupted blocks.
    Now do you have backup of archives ?
    If so restore it and retry again.
    If not, I suggest you to go with incremental scn for standby instead of recreation whole DR.
    Check this link http://www.oracle-ckpt.com/rman-incremental-backups-to-roll-forward-a-physical-standby-database-2/

  • Standby database Archive log destination confusion

    Hi All,
    I need your help here..
    This is the first time that this situation is arising. We had sync issues in the oracle 10g standby database prior to this archive log destination confusion.So we rebuilt the standby to overcome this sync issue. But ever since then the archive logs in the standby database are moving to two different locations.
    The spfile entries are provided below:
    *.log_archive_dest_1='LOCATION=/m99/oradata/MARDB/archive/'
    *.standby_archive_dest='/m99/oradata/MARDB/standby'
    Prior to rebuilding the standby databases the archive logs were moving to /m99/oradata/MARDB/archive/ location which is the correct location. But now the archive logs are moving to both /m99/oradata/MARDB/archive/ and /m99/oradata/MARDB/standby location, with the majority of them moving to /m99/oradata/MARDB/standby location. This is pretty unusual.
    The archives in the production are moving to /m99/oradata/MARDB/archive/ location itself.
    Could you kindly help me overcome this issue.
    Regards,
    Dan

    Hi Anurag,
    Thank you for update.
    Prior to rebuilding the standby database the standby_archive_dest was set as it is. No modifications were made to the archive destination locations.
    The primary and standby databases are on different servers and dataguard is used to transfer the files.
    I wanted to highlight one more point here, The archive locations are similar to the ones i mentioned for the other stndby databases. But the archive logs are moving only to /archive location and not to the /standby location.

Maybe you are looking for