Archived Log is not shifting from Primary to DR

we have increase the table size without taking care of DR, so
sapdata1 becomes 100% at DR site after that applying archived had
stopped at DR, after that i manually added datafile using following
commands,
select * from v$recover_file where error like '%FILE%';
Primary-:
1)select file#,name from v$datafile where file#=33;
/oracle/BP1/sapdata1/sr3_21/sr3.data21
2)select file#,name from v$datafile where file#=34;
/oracle/BP1/sapdata1/sr3_22/sr3.data22
alter system set standby_file_management =Manual;
DR-:
1)select file#,name from v$datafile where file#=33;
/oracle/BP1/102_64/dbs/UNNAMED00033
2)select file#,name from v$datafile where file#=34;
/oracle/BP1/102_64/dbs/UNNAMED00034
alter database create datafile '/oracle/BP1/102_64/dbs/UNNAMED00033'
as '/oracle/BP1/sapdata1/sr3_21/sr3.data21';
alter database create datafile '/oracle/BP1/102_64/dbs/UNNAMED00034'
as '/oracle/BP1/sapdata1/sr3_22/sr3.data22';
alter system set standby_file_management =Auto;
alter database recover managed standby database disconnect from
session;
after that all logfile which is available at DR is applied but now
automatic shifting from primary to DR is not happen, please check and
provide the solution as early as possible because if log gap increase
we have to go for backup and restore and our DR site is so far.

Dear Orkun Gedik ,
Please find the Aler Log-:
Errors in file /oracle/BP1/saptrace/background/bp1_mrp0_20361.trc:
ORA-00313: open failed for members of log group 1 of thread 1
Thu Aug 11 18:39:53 2011
Errors in file /oracle/BP1/saptrace/background/bp1_mrp0_20361.trc:
ORA-19527: physical standby redo log must be renamed
ORA-00312: online log 1 thread 1: '/oracle/BP1/origlogA/log_g11m1.dbf'
Clearing online redo logfile 1 complete
Please find the log for  bp1_mrp0_20361.trc
Start recovery at thread 1 ckpt scn 232442669 logseq 4297 block 2
2011-08-11 18:39:53.881
Media Recovery add redo thread 1
2011-08-11 18:39:53.881 1180 krsm.c
Managed Recovery: Active posted.
ORA-00313: open failed for members of log group 1 of thread 1
2011-08-11 18:39:53.891 64311 /se_autofs/patch_archive/102044/197/10107454/k
crr.c
Clearing online redo logfile 1 /oracle/BP1/origlogA/log_g11m1.dbf
ORA-00313: open failed for members of log group 1 of thread 1
ORA-19527: physical standby redo log must be renamed
ORA-00312: online log 1 thread 1: '/oracle/BP1/origlogA/log_g11m1.dbf'
Error 19527 creating/clearing online redo logfile 1
2011-08-11 18:39:53.950 64311 /se_autofs/patch_archive/102044/197/10107454/k
crr.c
Clearing online redo logfile 1 complete
2011-08-11 18:39:53.960 64311 /se_autofs/patch_archive/102044/197/10107454/k
crr.c
Media Recovery Waiting for thread 1 sequence 4297

Similar Messages

  • Archive Logs Will Not Delete from OS

    Hi Everyone,
    Here's my issue. Somehow i have a few archivelogs that are corruptued, plus a backup of the database that is corrupted. Since then, I have made a new backup of the database and everything is running smoothly. WHen i try to delete the files from DISK (which i cannot delete), i get this error: CANNOT DELETE <file_name>: The file or directory is corrupted and unreadable. I try to search for the corrupted files in rman, but i get this message "specification does not match any archived log in the recovery catalog". It seems that rman doesn't recognize that the files exist.
    I go to the alert log and I have this error:
    Corrupt block seq: 10511 blocknum=1.
    Bad header found during deleting archived log
    Data in bad block - seq:0. bno:0. time:0
    beg:0 cks:0
    calculated check value: 0
    Reread of seq=10511, blocknum=1, file=APP\ORACLE\FLASH_RECOVERY_AREA\PPROD\ARCHIVELOG\2010_11_04\O1_MF_1_10511_6F5SOFC7_.ARC, found same corrupt data
    Reread of seq=10511, blocknum=1, file=APP\ORACLE\FLASH_RECOVERY_AREA\PPROD\ARCHIVELOG\2010_11_04\O1_MF_1_10511_6F5SOFC7_.ARC, found same corrupt data
    Reread of seq=10511, blocknum=1, file=APP\ORACLE\FLASH_RECOVERY_AREA\PPROD\ARCHIVELOG\2010_11_04\O1_MF_1_10511_6F5SOFC7_.ARC, found same corrupt data
    Reread of seq=10511, blocknum=1, file=APP\ORACLE\FLASH_RECOVERY_AREA\PPROD\ARCHIVELOG\2010_11_04\O1_MF_1_10511_6F5SOFC7_.ARC, found same corrupt data
    Reread of seq=10511, blocknum=1, file=APP\ORACLE\FLASH_RECOVERY_AREA\PPROD\ARCHIVELOG\2010_11_04\O1_MF_1_10511_6F5SOFC7_.ARC, found same corrupt data
    I would really appreciate some help to delete these files from the OS. Thanks in advance to anyone that can help.

    RMAN> Change Archivelog All Validate;
    released channel: ORA_DISK_1
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=363 device type=DISK
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_04\O1_MF_1_10514_6F6SJ9D6_.ARC RECID=11389 STAMP=734220234
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_07\O1_MF_1_10546_6FDRCHQX_.ARC RECID=11421 STAMP=734412064
    validation succeeded for archived log
    archived log file name=\APP\ORACLE\FLASH_RECOVERY_AREAPPROD\ARCHIVELOG\2010_11_07\O1_MF_1_10556_6FGTGZNJ_.ARC RECID=11431 STAMP=734479760
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_07\O1_MF_1_10557_6FGX6CG5_.ARC RECID=11432 STAMP=734482556
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_07\O1_MF_1_10558_6FGZ9SLK_.ARC RECID=11433 STAMP=734484714
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10559_6FH12R00_.ARC RECID=11434 STAMP=734486537
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10560_6FH6OJRW_.ARC RECID=11435 STAMP=734492259
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10561_6FHL3J98_.ARC RECID=11436 STAMP=734503969
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10562_6FHTF2QX_.ARC RECID=11437 STAMP=734512468
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10563_6FJ32ZSD_.ARC RECID=11438 STAMP=734521361
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10564_6FJB1OMK_.ARC RECID=11439 STAMP=734528486
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10565_6FJJH3TV_.ARC RECID=11440 STAMP=734535062
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10566_6FJKJ6PY_.ARC RECID=11441 STAMP=734536118
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10567_6FJV244W_.ARC RECID=11442 STAMP=734545911
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10568_6FK5BMYR_.ARC RECID=11443 STAMP=734556421
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10569_6FK68QRW_.ARC RECID=11444 STAMP=734557384
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10570_6FKDM74D_.ARC RECID=11445 STAMP=734563863
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10571_6FKGK4LK_.ARC RECID=11446 STAMP=734565848
    validation failed for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_08\O1_MF_1_10572_6FKH42MK_.ARC RECID=11447 STAMP=734566452
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10573_6FKOQX6C_.ARC RECID=11448 STAMP=734573238
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10574_6FKVHC8S_.ARC RECID=11449 STAMP=734579100
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10575_6FL6T300_.ARC RECID=11450 STAMP=734590708
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10576_6FLJ3NMK_.ARC RECID=11451 STAMP=734600229
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10577_6FLO72K2_.ARC RECID=11452 STAMP=734605459
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10578_6FLTF31G_.ARC RECID=11453 STAMP=734610772
    validation succeeded for archived log
    archived log file name=APP\ORACLE\FLASH_RECOVERY_AREA\PROD\ARCHIVELOG\2010_11_09\O1_MF_1_10579_6FM1N500_.ARC RECID=11454 STAMP=734618166
    Crosschecked 26 objects

  • ORA-00339: archived log does not contain any redo

    Hi All,
    recently we faced 'ORA-00339: archived log does not contain any redo' issue at standby side,
    after searching on Google and from Metalink (note 30866.1 and 7197445.8 ) I find out that this is the known issue for 10g and below versions, our's is 11.2.0.3,
    Error in Alert Log :
    Errors in file /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/trace/DWPRD_pr0a_48412.trc:
    ORA-00339: archived log does not contain any redo
    ORA-00334: archived log: '/redolog2/redo/redolog3a.log'
    Errors in file /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/trace/DWPRD_pr0a_48412.trc (incident=190009):
    ORA-00600: internal error code, arguments: [kdBlkCheckError], [1], [56702], [6114], [], [], [], [], [], [], [], []
    Incident details in: /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/incident/incdir_190009/DWPRD_pr0a_48412_i190009.trc
    Use ADRCI or Support Workbench to package the incident.
    See Note 411.1 at My Oracle Support for error and packaging details.
    Slave exiting with ORA-10562 exception
    Errors in file /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/trace/DWPRD_pr0a_48412.trc:
    ORA-10562: Error occurred while applying redo to data block (file# 1, block# 56702)
    ORA-10564: tablespace SYSTEM
    ORA-01110: data file 1: '/oradata1/database/DATAFILES/system01.dbf'
    ORA-10561: block type 'TRANSACTION MANAGED DATA BLOCK', data object# 2
    ORA-00600: internal error code, arguments: [kdBlkCheckError], [1], [56702], [6114], [], [], [], [], [], [], [], []
    Mon Apr 15 11:34:12 2013
    Dumping diagnostic data in directory=[cdmp_20130415113412], requested by (instance=1, osid=48412 (PR0A)), summary=[incident=190009].
    Thanks

    Hi,
    "The archived log is not the correct log.
    It is a copy of a log file that has never been used for redo generation, or was an online log being prepared to be the current log."
    "Restore the correct log file."
    Can you say, what is last changes on your database, On log files?
    Did you copies your '/redolog2/redo/redolog3a.log' log file from other ?
    Regards
    Mahir M. Quluzade

  • Archived log files not registered in the Database

    I have Widows Server 2008 R2
    I have Oracle 11g R2
    I configured primary and standby database in 2 physical servers , please find below the verification:
    I am using DG Broker
    Renetly I did failover from primary to standby database
    Then I did REINSTATE DATABASE to returen the old primary to standby mode
    Then I did Switchover again
    I have problem that archive logs not registered and not imeplemented.
    SQL> select max(sequence#) from v$archived_log; 
    MAX(SEQUENCE#)
             16234
    I did alter system switch logfile then I ssue the following statment to check and I found same number in primary and stanbyd has not been changed
    SQL> select max(sequence#) from v$archived_log;
    MAX(SEQUENCE#)
             16234
    Any body can help please?
    Regards

    Thanks for reply
    What I mean after I do alter system switch log file, I can see the archived log files is generated in the physical Disk but when
    select MAX(SEQUENCE#) FROM V$ARCHIVED_LOG;
    the sequence number not changed it should increase by 1 when ever I do switch logfile.
    however I did as you asked please find the result below:
    SQL> alter system switch logfile;
    System altered.
    SQL> /
    System altered.
    SQL> /
    System altered.
    SQL> /
    System altered.
    SQL> SELECT DB_NAME,HOSTNAME,LOG_ARCHIVED,LOG_APPLIED_02,LOG_APPLIED_03,APPLIED_TIME,LOG_ARCHIVED - LOG_APPLIED_02 LOG_GAP_02,
      2  LOG_ARCHIVED - LOG_APPLIED_03 LOG_GAP_03
      3  FROM (SELECT NAME DB_NAME FROM V$DATABASE),
      4  (SELECT UPPER(SUBSTR(HOST_NAME, 1, (DECODE(INSTR(HOST_NAME, '.'),0, LENGTH(HOST_NAME),(INSTR(HOST_NAME, '.') - 1))))) HOSTNAME FROM V$INSTANCE),
      5  (SELECT MAX(SEQUENCE#) LOG_ARCHIVED FROM V$ARCHIVED_LOG WHERE DEST_ID = 1 AND ARCHIVED = 'YES'),
      6  (SELECT MAX(SEQUENCE#) LOG_APPLIED_02 FROM V$ARCHIVED_LOG WHERE DEST_ID = 2 AND APPLIED = 'YES'),
      7  (SELECT MAX(SEQUENCE#) LOG_APPLIED_03 FROM V$ARCHIVED_LOG WHERE DEST_ID = 3 AND APPLIED = 'YES'),
      8  (SELECT TO_CHAR(MAX(COMPLETION_TIME), 'DD-MON/HH24:MI') APPLIED_TIME FROM V$ARCHIVED_LOG WHERE DEST_ID = 2 AND APPLIED = 'YES');
    DB_NAME HOSTNAME           LOG_ARCHIVED   LOG_APPLIED_02    LOG_APPLIED_03     APPLIED_TIME     LOG_GAP_02      LOG_GAP_03
    EPPROD  CORSKMBBOR01     16252                  16253                        (null)                      15-JAN/12:04                  -1                   (       null)

  • Archive logs are not transferred to STDBY database

    Hi,
    I have create a STDBY database (I am running the release 9.2.0.7.0).
    I see that the archivelogs are not correctly transferred into STDBY server.
    From Primary alert log, I see the following error:
    ARC1: Evaluating archive log 1 thread 1 sequence 16734
    ARC1: LGWR is actively archiving destination LOG_ARCHIVE_DEST_2
    ARC1: Destination LOG_ARCHIVE_DEST_2 archival not expedited
    ARC1: Beginning to archive log 1 thread 1 sequence 16734
    Creating archive destination LOG_ARCHIVE_DEST_1: '/backup/archivelogs/log1_16734.arc'
    ARC1: LGWR is actively archiving destination LOG_ARCHIVE_DEST_2
    Invoking non-expedited destination LOG_ARCHIVE_DEST_2 thread 1 sequence 16734 host STDBY_PROD
    ARC1: Completed archiving log 1 thread 1 sequence 16734
    Thu Nov 17 14:54:42 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-03114: not connected to ORACLE
    Thu Nov 17 14:54:42 2011
    ARC0: FAL archive failed, see trace file.
    ARCH: FAL archive failed. Archiver continuing
    Thu Nov 17 14:54:42 2011
    ORACLE Instance PROD - Archival Error. Archiver continuing.
    ARCH: Connecting to console port...
    Thu Nov 17 14:54:42 2011
    ORA-16055: FAL request rejected
    ARCH: Connecting to console port...
    Thu Nov 17 14:54:42 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-16055: FAL request rejected
    ARC0: Begin FAL archive (thread 1 sequence 16483 destination STDBY_PROD)
    Creating archive destination LOG_ARCHIVE_DEST_2: 'STDBY_PROD'
    Thu Nov 17 15:05:44 2011
    LGWR: I/O error 3114 archiving log 2 to 'STDBY_PROD'
    Thu Nov 17 15:05:44 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_lgwr_5265.trc:
    ORA-03114: not connected to ORACLE
    Thu Nov 17 15:10:08 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-03114: not connected to ORACLE
    Thu Nov 17 15:10:08 2011
    ARC0: FAL archive failed, see trace file.
    ARCH: FAL archive failed. Archiver continuing
    Thu Nov 17 15:10:08 2011
    ORACLE Instance PROD - Archival Error. Archiver continuing.
    ARCH: Connecting to console port...
    Thu Nov 17 15:10:08 2011
    ORA-16055: FAL request rejected
    ARCH: Connecting to console port...
    Thu Nov 17 15:10:08 2011
    Errors in file /mnt/orclEBS/oracle/proddb/9.2.0/admin/PROD_ebslive/bdump/prod_arc0_5277.trc:
    ORA-16055: FAL request rejected
    I see that both archiving destinations are VALID:
    SQL> select dest_id, dest_name, status from v$archive_dest_status where status != 'INACTIVE';
    DEST_ID ---------- DEST_NAME -------------------------------------------------------------------------------- STATUS
    1 LOG_ARCHIVE_DEST_1 VALID
    2 LOG_ARCHIVE_DEST_2 VALID
    SQL>
    TNSPING works properly in both directions, e.g. from PROD -> STDBY:
    $tnsping STDBY_PROD
    TNS Ping Utility for Linux: Version 9.2.0.7.0 - Production on 17-NOV-2011 16:56:40
    Copyright (c) 1997 Oracle Corporation. All rights reserved.
    Used parameter files:
    /mnt/orclEBS/oracle/proddb/9.2.0/network/admin/PROD_ebslive/sqlnet_ifile.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION= (ADDRESS=(PROTOCOL=tcp) (HOST=EBSSTDBY) (PORT=1521)) (CONNECT_DATA=(SID=PROD)))
    OK (10 msec)
    Please help me to solve this issue.
    Thank you.

    A few notes :
    Troubleshooting Tips For Dataguard Switchover (9i and 10gR1) [ID 298986.1]
    Although it is about Switchover, it does mandate a password file exclusive. If the switchover requires the password file, shouldn't ARCH shipping require it ?
    Step By Step Guide To Create Physical Standby Database Using RMAN [ID 469493.1]
    also covers 9i with specific mentions of 9i differences from 10g. Passwordfile is not mentioned as different in 9i from 10g.
    Data Guard 9i - Net8 Configuration for a 2-node database environment. [ID 175122.1
    has log_archive_dest_2='SERVICE=NODE2STDBY.ACME.COM'    so it connects to the service
    *However* the document "Introduction to Oracle 9i Data Guard Manager" from
    Data Guard 9i Introduction to Data Guard Manager GUI [ID 150217.1]
    page 59 has this : "The creation process will attempt to create a remote login password file for the standby database to
    enable remote connections to the database. (Note: A remote login password file is not necessary for
    Data Guard operation; it is only needed to allow remote connections to the database from clients,
    such as other Enterprise Manager tools.)"
    Similarly the section "3.2.6 Set Initialization Parameters on a Physical Standby Database"
    in the 9i DataGuard manual at http://download.oracle.com/docs/cd/B10501_01/server.920/a96653/create_ps.htm#62941
    only has
    remote_archive_enable=TRUE
    (this is an instance parameter)
    Similarly the section "5.8.2.1 Primary Database Initialization Parameters"
    at http://download.oracle.com/docs/cd/B10501_01/server.920/a96653/log_transport.htm#1067609
    explains
    The last parameter, REMOTE_ARCHIVE_ENABLE=SEND, allows the primary database to send redo data to the standby database, but prevents the primary database from receiving redo data from another system.
    with REMOTE_ARCHIVE_ENABLE=RECEIVE on the Standby
    to Receive and archive the incoming redo data from the primary database, but only while the database is running in the standby role
    So, I guess that REMOTE_ARCHIVE_ENABLE is the important one. Not REMOTE_LOGIN_PASSWORDFILE.
    In 9i : http://download.oracle.com/docs/cd/B10501_01/server.920/a96536/ch1174.htm#1023087
    In 10g it is deprecated : http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/initparams176.htm#sthref541 so that LOG_ARCHIVE_CONFIG is used.
    (and no longer appears in the 11.2 Reference)
    Hemant K Chitale

  • Archive log can not ship to GAP logfiles to standby DB automatically

    we have a non-real time standby database, which will receive the archive file from the primary database server most of the time, and will apply the logfiles only at one point of time daily.
    Some times, we need to shutdown the Standby DB server for a while ( 3-4 hours).
    The missed logfiles will catch up during the standby down time later.
    But since last week we had an storage incident, the primary DB server stops to catch up the missed logfiles, and saw this message at the archvie trace file:
    ABC: tkrsf_al_read: No mirror copies to re-read data
    Current, we found the archive log gaps on the standby server, and have to manually copy those logfiles over and register them.
    Saw some tips on the internet to change the parameter "log_archive_max_processes", but no help for us at all.
    Here is the parameter on the Primary DB server:
    log_archive_dest_2 = SERVICE=Standby_server reopen=300

    which will receive the archive file from the primary database server most of the timeMost times from primary. Then remaining times? so you Copy manually and register?
    Then it's dataguard not an manual standby.
    Error 1034 received logging on to the standby
    Errors in file /******/***arc210536.trc:
    ORA-01034: ORACLE not available
    FAL[server, ARC2]: FAL archive failed, see trace file.these errors in primary when , stanby is down and when primary tries to connect to standby, so tese errors not considerable to investigate ,
    When you don't want to apply archives on standby no need to shutdown. Just put this value log_archive_dest_state_2='defer'
    Once you enable check what are the errors in primary alert log file.
    How is your network band width speed? Is it capable to hold that much archive data?
    May be it will take some time when you pause & start.
    Also use LGWR in log_archive_dest_2 for real time apply after creatin standby redo logs.
    So post the alert log information once you enable standby database.

  • Archive logs are not transferring to DR server in DataGuard

    Dear Gurus,
    We have implemented data guard on 10g and it was working properly till last week.All of a sudden it is stopped working and i am getting the error in primary server alert log as
    "ORA-12154: TNS:could not resolve the connect identifier specified
    PING[ARCr]: Heartbeat failed to connect to standby 'STANDRR1'"
    and the trace file entries are as follows
    " Error 12154 received logging on to the standby,Error 12154 connecting to destination LOG_ARCHIVE_DEST_2 standby host 'STANDRR1',Error 12154 attaching to destination LOG_ARCHIVE_DEST_2 standby host 'STANDRR1',ORA-12154: TNS:could not resolve the connect identifier specified".
    We have changed nothing on either side of the servers.
    Please suggest how to resolve the issue and Thanks in advance.
    Thanks & Regards,
    venkat

    Hi,
    The output from standby,
    lsnrctl servicesLSNRCTL for HPUX: Version 10.2.0.4.0 - Production on 04-FEB-2012 15:50:20
    Copyright (c) 1991, 2007, Oracle. All rights reserved.
    Connecting to (ADDRESS=(PROTOCOL=IPC)(KEY=PP1.WORLD))
    Services Summary...
    Service "RR1" has 1 instance(s).
    Instance "RR1", status UNKNOWN, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:0 refused:0
    LOCAL SERVER
    The command completed successfully.
    The output of tnsping from standby:
    tnsping STANDPP1
    TNS Ping Utility for HPUX: Version 10.2.0.4.0 - Production on 04-FEB-2012 15:50:
    56
    Copyright (c) 1997, 2007, Oracle. All rights reserved.
    Used parameter files:
    /oracle/RR1/102_64/network/admin/sqlnet.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (COMMUNITY = SAP
    .WORLD) (PROTOCOL = TCP) (HOST = pildrprd) (PORT = 1527))) (CONNECT_DATA = (SID
    = RR1) (GLOBAL_NAME = RR1.WORLD)))
    OK (10 msec)
    The output of tnsping from primary:
    tnsping STANDPP1
    TNS Ping Utility for HPUX: Version 10.2.0.4.0 - Production on 04-FEB-2012 15:55:
    59
    Copyright (c) 1997, 2007, Oracle. All rights reserved.
    Used parameter files:
    /oracle/PP1/102_64/network/admin/sqlnet.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (COMMUNITY = SAP
    .WORLD) (PROTOCOL = TCP) (HOST = pildrprd) (PORT = 1527))) (CONNECT_DATA = (SID
    = RR1) (GLOBAL_NAME = RR1.WORLD)))
    OK (30 msec).
    Still i am getting the error.Please suggest.

  • A Group of Archive Logs Will Not Backup

    I restored a database to a different server, and while that was happening I copied some logarchives out of the production logarchive directory. After the restore completed, but before I put those logarchives back into that directory, I ran CHANGE ARCHIVELOG ALL CROSSCHECK on the production server to sync the catalog with the database. That was a mistake to do that yet. I then put the logarchives back into the production logarchive directory and tried to run a logarchive backup. I should have put these files back before running crosscheck. Now I receive this error:
    RMAN-06004: ORACLE error from recovery catalog database: RMAN-20242: specification does not match any archive log in the recovery catalog
    How can I backup these production logarchives that I copied back into the logarchive directory?
    Thank you.

    Database version? To know that makes it easier to advice, because since RMAN was introduced in 8.0 there were some changes.
    In 10g and higher use 'catalog archivelog...'.
    http://download.oracle.com/docs/cd/B19306_01/backup.102/b14194/rcmsynta011.htm#sthref230
    Werner

  • To Archive Log or Not to Archive Log

    I was recently in a discussion w/ a senior oracle dba who said that you shouldn't run your database in archive log mode if it is possible to take a complete cold backup each night. His recommendation was to have large enough redo logs where, in a days time, they wouldn't be overwritten by log switches. Are there any opinions out there on this?
    Thanks.

    This option depends on requirements on your DB.
    If your DB is in production state and must be running 24x7 and also data loss is not acceptable then best backup strategy is hot backups with RMAN (this option requires Archivelog mode).
    If your DB is not in production and data loss is acceptable then cold backup is enough but this backup strategy has some disadvantages.

  • How to find thead 2 archive log when i recover from 2-node RAC to single

    I backup 2-ndoe RAC and restore to single - node
    Control file created.
    SQL> recover database iuckup controlfile
    SQL> recover database using backup controlfIle;
    ORA-00279: change 12100176131169 generated at 07/06/2013 16:36:57 needed for
    thread 1
    ORA-00289: suggestion : /arch/hop1_566085708_1_212692.dat                 -- Oracle suggest
    ORA-00280: change 12100176131169 for thread 1 is in sequence #212692
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    arch/sch1_566085708_1_212692.dat                                            --- I input  
    ORA-00279: change 12100176131169 generated at  needed for thread 2
    ==> Oracle didn't suggest thread 2 archive log file/
    How can I find ?

    You have to query V$ARCHIVED_LOG for THREAD# 2 on the source database to find the first archivelog where 12100176131169 is less than the NEXT_CHANGE#
    select SEQEUNCE# from V$ARCHIVED_LOG where THREAD#=2 and 12100176131169 between FIRST_CHANGE# and NEXT_CHANGE#
    Once you provide the Sequence# and file name for the first Thread 2 file, the RECOVER command will automatically generate the expected file name for the subsequent archivelogs (of both threads).
    Hemant K Chitale

  • Generate archive logs are not in sequence number?

    On last friday... the latest archive log number was ARC00024.ARC. Tomorrow when I come backup, the archive logs ARC00001.ARC and ARC00002.ARC were being generated by oracle itself. I wondering the archive log sequence should be in sequence. What is happening?
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination C:\oracle\ora92\RDBMS
    Oldest online log sequence 1
    Next log sequence to archive 3
    Current log sequence 3
    SQL>
    FAN
    Edited by: user623471 on Jun 7, 2009 7:35 PM

    khurram,
    Its our production instance and havent issued resetlogs option but when listing the arvchives it shows in different sequence number...
    and also while copying the archives by RMAN it doesnt copy in sequence
    -rw-r----- 1 xxx dba 69363859 May 28 19:16 2_10373.arc.gz
    -rw-r----- 1 xxx dba 43446622 May 28 19:16 1_10553.arc.gz
    -rw-r----- 1 xxx dba 52587365 May 28 19:16 1_10578.arc.gz
    -rw-r----- 1 xxx dba 45251820 May 28 19:16 1_10543.arc.gz
    -rw-r----- 1 xxx dba 60890256 May 28 19:17 1_10579.arc.gz
    -rw-r----- 1 xxx dba 46659008 May 28 19:17 1_10548.arc.gz
    -rw-r----- 1 xxx dba 116899466 May 28 19:17 2_10353.arc.gz
    -rw-r----- 1 xxx dba 77769517 May 28 19:17 1_10531.arc.gz
    -rw-r----- 1 xxx dba 66401923 May 28 19:18 1_10530.arc.gz
    -rw-r----- 1 xxx dba 45972697 May 28 19:18 1_10605.arc.gz
    -rw-r----- 1 xxx dba 55082543 May 28 19:18 1_10600.arc.gz
    -rw-r----- 1 xxxq dba 42682207 May 28 19:19 1_10547.arc.gz
    thanks,
    baskar.l

  • Archived log file not displaying

    While navigating around the "home" page for OCS as an administrator...I was trying to run a report under Reports>Conferences>Diagnostics.
    The links says:
    Click the link below to view comprehensive conference diagnostics. To see the log file correctly, use Internet Explorer 6.0 or higher.
    I am using IE 6 and the page shows up as being done...but it is blank. Any idea what is wrong? The URL reads:
    https://mywebserver/imtapp/logs/imtLogs.jsp?fileName=D:/ocs_onebox/mtier/imeeting/logs/sessions/12.20.2004/10000-clbsvr_OCS_home_mid.mywebserver.imt-collab.0-06_34_01.xml
    The file is there on the filesystem.
    TIA.

    Stages means Transformations in Data flow...
    Transformation names are not displaying correctly in log file.
    for example if i given name as  "TC_table_name" for Table compare Transformation then its displaying only "Table Comparison"  in Log file

  • How: Script archive log transfer to standby db

    Hi,
    I’m implementing disaster recovery right now. For some special reason, the only option for me is to implement non-managed standby (manual recovery) database.
    The following is what I’m trying to do using shell script:
    1.     Compress archive logs and copy them from Primary site to Standby site every hour. ( I have a very low network )
    2.     Decompress archive logs at standby site
    3.     Check if there are missed archive logs. If no, then do the manual recovery
    Did I miss something above? And I’m not skill in to build shell scripts, is there any sample scripts I can follow? Thanks.
    Nabil
    Message was edited by:
    11iuser

    Hi,
    Take a look at data guard packages. There is a package just for this purpose: Bipul Kumar notes:
    http://www.dba-oracle.com/t_oracledataguard_174_unskip_table_.htm
    "the time lag between the log transfer and the log apply service can be built using the DELAY attribute of the log_archive_dest_n initialization parameter on the primary database. This delay timer starts when the archived log is completely transferred to the standby site. The default value of the DELAY attribute is 30 minutes, but this value can be overridden as shown in the following example:
    LOG_ARCHIVE_DEST_3=’SERVICE=logdbstdby DELAY=60’;"
    1. Compress archive logs and copy them from Primary site to Standby site every hour.Me, I use tar (or compress) and rcp, but I don't know the details of your environment. Jon Emmons has some good notes:
    http://www.lifeaftercoffee.com/2006/12/05/archiving-directories-and-files-with-tar/
    2. Decompress archive logs at standby siteSee the man pages for uncompress. I do it through a named pipe to simplify the process:
    http://www.dba-oracle.com/linux/conditional_statements.htm
    3. Check if there are missed archive logs.I keep my standby data in recovery mode, and as soon as the incoming logs are uncompressed, they are applied automatically.
    Again, if you don't feel comfortable writing your own, consider using the data guard packages.
    Hope this helps. . .
    Donald K. Burleson
    Oracle Press author

  • Physycal Standby archive log gap....

    Archive log gap caused... The reason being before the logs can be shipped to standby location where deleted by rman backup... So restored the archives on primary database site back again... These old logs from the gap are not getting shipped to the standby site, but the new ones generated currently are getting shipped.
    Can some one help what action do I have to take to resolve the gap? And how to know what's causing and not letting this shipping happen?
    Or shall I manually ship these gap archive logs to the standby site?

    1) Yep running 9i.. But still its not shipping...Are FAL_CLIENT & FAL_SERVER parameters are defined at standby level?
    If not, define them at standby level. Those parameter will help to get missing (gap) archives from primary database.
    2) If so shipped manually do have to register the archive logs? Just copy from primary to standby and don't need to register any gap, that was in 8i and when there was no background process MRP (media recovery process). If the standby database is in auto media recovery, then, it will automatically applies all the archived logs.
    Jaffar

  • Archived log not applying

    archived logs are not apllying in DR
    SQL> select max(sequence#) from v$archived_log where applied='YES';
    MAX(SEQUENCE#)
    SQL>
    SQL> select process,status,sequence# from v$managed_standby;
    PROCESS   STATUS        SEQUENCE#
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    RFS       IDLE                  0
    RFS       IDLE                  0
    RFS       IDLE              34473
    MRP0      WAIT_FOR_GAP      31753
    RFS       IDLE                  0
    RFS       IDLE                  0
    RFS       IDLE                  0
    11 rows selected.

    Hello,
    As previously posted, you have a GAP. So, you may have to investigate why the Archived logs cannot be applied.
    To do so, you may check for the File System if any Archive destination is FULL. Also, you may have to verify the parameters on the PRIMARY and STANDBY site:
    sqlplus "/ as sysdba"
    show parameter fal
    show parameter dest
    show parameter config
    show parameter broker
    Check also the Alert log on both Primary and Standby.
    If the DATAGUARD BROKER is enabled you may also execute this
    dgmgrl
    show configuration verbose
    Also try to "tnsping" the services used in the Archive Destination from both site.
    When you are sure that everything is correct then, verify if any Archived logs is missing and try to restore it if possible.
    Else if everything is broken, and there's no way to get back the missing Archived logs, you 'll have to "rebuild" the Standby from a new Backup of the PRIMARY.
    Hope this help.
    Best regards,
    Jean-Valentin Lubiez

Maybe you are looking for

  • HP 6210 printer problems with upgrade to Snow Leopard

    I have recently bought a new Macbook Pro with Snow Leapard and my HP printer 6210 gets errors when I try to print.  (It worked fine with the old 10.5.x. OSX)  I have uninstalled the old HP software, then installed the most current from the HP website

  • Prob in using rank in pl/sql ,need logic same of rank function in any way

    I have a query as of the following <br> <br> SELECT sr_no,cod_acct_no,dat_arrears_due,amt_arrear_due<br> FROM ( select cod_acct_no,dat_arrears_due,sum(amt_arrears_due) <br>amt_arrear_due,rank() over (partition by cod_acct_no order by <br>dat_arrears_

  • ADF 11g: Desktop Integrator- Export to Excel issue

    Hi, We are using ADF 11g. In our application we have one feature where end users can download data in excel format from our web application. To export data to excel file, users need to click Export to Excel button. But after deployment of code to sta

  • Can we do multiple transaction testing in different browsers?

    Helllo,  I am trying to do transaction testing instead of URL based i.e. suppose take net banking and wanted test login , post login and credit card section. Then how can i do that ?

  • List of SFTP/PGP Adapters?

    Hi All, Can any one give me list of vendors who provides "SFTP-SSh with /PGP encryption" Adapters? I Googled & search on SDN and found only 4 vendors below, but want to search more vendors. 1) www.advantco.com/?q=node/9 2) www.seeburger.com/fileadmin