Automatic log shipping error

HI
I have a 10gR2 standby database.Normally when logs are created in primary they automatically get shipped to standby but sometimes I get FAL request rejected or connection lost contact error and the logs are not automatically shipped .Then when I manually switch a logfile the RFS process copies the old logs into DR site(around 4 or 5 logs).
What is the problem...How to resolve this????
How to make FAL check for missing gaps after particular time...

eRROR IN pr
>
Completed checkpoint up to RBA [0x3668.2.10], SCN: 258642313474
Tue Apr 5 16:28:08 2011
Errors in file ******************i_arc1_11891.trc:
ORA-03135: connection lost contact
Tue Apr 5 16:28:10 2011
FAL[server, ARC1]: FAL archive failed, see trace file.
Tue Apr 5 16:28:10 2011
Errors in file -*********************arc111891.trc:
ORA-16055: FAL request rejected
ARCH: FAL archive failed. Archiver continuing
Tue Apr 5 16:28:10 2011
ORACLE Instance ******* - Archival Error. Archiver continuing.
Tue Apr 5 16:42:49 2011
ALTER SYSTEM SET log_archive_dest_state_3='ENABLE' SCOPE=BOTH;
Tue Apr 5 16:59:06 2011
Beginning log switch checkpoint up to RBA [0x3669.2.10], SCN: 258642979476
Thread 1 advanced to log sequence 13929
Current log# 2 seq# 13929 mem# 0:**************02_01.log
Tue Apr 5 17:04:11 2011
Completed checkpoint up to RBA [0x3669.2.10], SCN: 258642979476
Tue Apr 5 17:09:43 2011
Errors in file ****************************i_arc1_11891.trc:
ORA-03135: connection lost contact
Tue Apr 5 17:09:46 2011
FAL[server, ARC1]: FAL archive failed, see trace file.
Tue Apr 5 17:09:46 2011
Errors in file ********************************arc111891.trc:
ORA-16055: FAL request rejected
ARCH: FAL archive failed. Archiver continuing
Tue Apr 5 17:09:46 2011
ORACLE Instance ******** - Archival Error. Archiver continuing.
Tue Apr 5 17:12:15 2011
ALTER SYSTEM SET log_archive_dest_state_3='ENABLE' SCOPE=BOTH;
Tue Apr 5 17:48:29 2011
Errors in file *******************************arc111891.trc:
ORA-03135: connection lost contact
Tue Apr 5 17:48:31 2011
FAL[server, ARC1]: FAL archive failed, see trace file.
Tue Apr 5 17:48:31 2011
Errors in file *********************************arc111891.trc:
ORA-16055: FAL request rejected
ARCH: FAL archive failed. Archiver continuing
Tue Apr 5 17:48:31 2011
ORACLE Instance *********i - Archival Error. Archiver continuing.
Tue Apr 5 17:48:34 2011
Beginning log switch checkpoint up to RBA [0x366a.2.10], SCN: 258643727276
Thread 1 advanced to log sequence 13930

Similar Messages

  • SQL 2012 Log Shipping error 14261

    I have two servers and I had log shipping working but could not add any additional databases to the process.  I deleted all the log shipping and I am still unable to add the databases back to the log shipping configuration.  The process fails while
    trying to all the Secondary Log shipping configuration.  I get the following error: The specified @name ('Log shipping Primary Server Alert.') already exists. (.Net SqlClient Data Provider) and it specifically refers to my secondary server by name in
    the error also referencing the stored procedure sp_verify_alert.  I have not been able to find a solution.  Please help.

    You ran this on secondary?
    2. On the secondary server, execute sp_delete_log_shipping_secondary_database to delete the secondary database.
    sp_delete_log_shipping_secondary_database N'LogShipAdventureWorks'

  • Archive log shipping error in Logical standby database

    Hi,
    We have Primay and logical database configuration both in RAC in Production. both database are in 10.2.0.4.
    Everyday we having issue of some archivelog not transfered to logical server around 2AM. I am having good exposure in Logical/Log minor/Goldgendate and can say fal server/client server configuration is correct.
    Can you guys please give some input of what can cause this.
    Yes we take RMAN backup of both database around 1:30 using control file instead of recovery catalog and I doubt this put lock on control file and cause this issue.
    I would be glad of you could help me out in this.
    Thanks a lot!
    Pradeep
    Below is the configuration of archive/fal on both database.
    PIP1REP2 is logical database(instance 2) where apply process is running/configured.
    Primary server configuration on (PIP11) (same are on other instances).
    SQL> sho parameter fal
    NAME TYPE VALUE
    fal_client string PIP1REP2
    fal_server string PIP11
    SQL> sho parameter archive
    NAME TYPE VALUE
    log_archive_dest_1 string LOCATION=+ARCH01
    log_archive_dest_2 string SERVICE=pip1rep2 lgwr async
    log_archive_dest_state_1 string enable
    log_archive_dest_state_2 string ENABLE
    llog_archive_local_first boolean TRUE
    log_archive_max_processes integer 10
    log_archive_min_succeed_dest integer 1
    log_archive_start boolean FALSE
    log_archive_trace integer 0
    remote_archive_enable string true
    Logical server configuration
    SQL> sho parameter fal
    NAME TYPE VALUE
    fal_client string PIP1REP
    fal_server string pip11, pip12, pip13, pip14, pi
    p15
    SQL> sho parameter archive
    NAME TYPE VALUE
    alog_archive_dest_1 string LOCATION=+ARCH01/PIP1REP valid
    for=(onlinelogfiles, all_rol
    es) db_unique_name=pip1rep
    log_archive_dest_2 string LOCATION=+ARCH01/PIP1REP/ARCH_
    DEST2_STANDBY_LOGS VALID_FOR=(
    STANDBY_LOGFILES,STANDBY_ROLE)
    db_unique_name=pip1rep
    log_archive_dest_state_1 string enable
    log_archive_dest_state_2 string enable
    log_archive_local_first boolean TRUE
    log_archive_max_processes integer 20
    log_archive_min_succeed_dest integer 1
    log_archive_trace integer 0
    remote_archive_enable string true
    standby_archive_dest string +ARCH01                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

    Do you think creating recovery catalog will resolve it?
    Yes, I do. (But I do not know for sure)
    If it helps when I got to this part I decided to backup either the Primary or the Standby, but not both.
    I decided Primary only and documented how to recreate my standby without downtime.
    I hope this helps.
    regards.

  • MAXDB Log Shipping

    Hi folks,
    we had working log shipping scripts which automatically copy log backups to the standby server. We had to interrupt the log backups (overwrite mode set during EHP installation) and now after a full backup and restore the log restores do not work, please see the log below (the last two lines).
    The log file =42105
    FirstLogPage=5233911
    LastLogPage=6087243
    Database UsedLogPage=5752904
    I would expect the restore to work since Database log page is within the range of the first and last log pages of the log backup file.
    Is this not the case? How should it work when re-establishing the log shipping?
    Mark.
    Find first log file to apply
    FindLowestLogNo()
    File: log_backup.42105
       File extension is numeric '42105'
       LogNo=42105
    FindLowestLogNo()
    LogMin=42105 LogMax=42105
    Execute Command: C:\sapdb\programs\pgm\dbmcli.exe -n localhost -d ERP -u SUPERDBA,Guer0l1t0 db_admin & C:\sapdb\programs\pgm\dbmcli.exe -n localhost -d ERP -u SUPERDBA,Guer0l1t0 medium_label LOG1 42105
    [START]
    OK
    OK
    Returncode               0
    Date                     20110308
    Time                     00111416
    Server                   saperpdb.rh.renold.com
    Database                 ERP
    Kernel Version           Kernel    7.7.06   Build 010-123-204-327
    Pages Transferred        0
    Pages Left               0
    Volumes                  
    Medianame                
    Location                 F:\log_shipping\log_backup.42105
    Errortext                
    Label                    LOG_000042105
    Is Consistent            
    First LOG Page           5233911
    Last LOG Page            6087243
    DB Stamp 1 Date          20110209
    DB Stamp 1 Time          00190733
    DB Stamp 2 Date          20110308
    DB Stamp 2 Time          00111415
    Page Count               853333
    Devices Used             1
    Database ID              saperpdb.rh.renold.com:ERP_20110209_210432
    Max Used Data Page       
    Converter Page Count     
    [END]
    LogNo=42105 FirstLogPage=5233911 LastLogPage=6087243 (UsedLogPage=5752904)
    WARNING: Log file not yet applied but NOT the first log file. Either sequence error or first log file is missing/yet to arrive

    Hello Birch Mark,
    the recovery with intialization is the correct step to recreate the shadow database.
    What has to be done before:
    Source database:
    SA)Activate the database logging.
    SB) Create the complete databackup
    SC)  Set the aoutolog on or create the logbackup,
        the first logbackup after completedatabackup created => check the backup history of the source database.
    Shadow database:
    ShA) Run the recovery with initialization or use db_activate dbm command, see more details in the MAXDB library,
        I gave you references in the my reply before.
    ShB) After the restore of the complete databackup created in step SB) don't restart the database in the online.
             Keep the shadow database in admin or offline < when you run execute recover_cancel >.
            Please post output of the db_restartinfo command.
       ShC) You could restart the logbackups recovery created in SC), check the backup history of the source database which logbackup will be first to recover.
    Did you follow those steps?
    There is helpful documentation/hints at Wiki:
    http://wiki.sdn.sap.com/wiki/display/MaxDB/SAPMaxDBHowTo
    ->"HowTo - Standby DB log shipping"
    Regards, Natalia Khlopina

  • Log Shipping - encrypted? Resilience?

    Hi,
    A couple of questions about log shipping that I did not see covered in documentation; if they are covered somewhere I would be glad to be pointed to the existing relevant docs.  This would be (at least at first) using SQL Server 2008.
    1) Is there any built-in mechanism for encryption of data transferred through the Log Shipping process, or would we have to use IPSec or similar transport-level strategies?
    2) Does Log Shipping provide any resilience features to automatically recover from failures such as: network errors preventing transfer of data; database unavailability at the scheduled times for log application; temporary lack of disk space, etc.?
    thanks!
    Martin

     similarly, during database unavailablity - while backing up or restore, the job might fail, but will restore\backups all of them,when it comes back ."
    is exactly the sort of thing I mean by resilience features; are you aware of any documentation or articles that explain how this works - such as, under what circumstances the jobs will catch up automatically versus requiring manual intervention?
    thanks!
    Martin
    Yes even if primary database is unavailable lets say due to server restart or SQL services down, when it comes online backup and restore will come backup.
    Only time log shipping will fail if log chain gets broken due to missing tlog backup or corrupt transaction log file due to network or media issue and that when need manual intervention.
    Please see below link:
    How log shipping works
    http://www.sqlshack.com/sql-server-log-shipping/
    setup log shipping:
    http://omaralzabir.com/how_to_setup_sql_server_2005_transaction_log_ship_on_large_database_that_really_works/
    Log shipping FAQ videos:
    http://www.brentozar.com/archive/2013/03/log-shipping-faq/
    Frequently asked questions - Log shipping
    https://support.microsoft.com/en-us/kb/314515
    Some more links for log shipping:
    http://www.tech-archive.net/Archive/SQL-Server/microsoft.public.sqlserver.setup/2004-11/0310.html

  • How to suppress Log Shipping alert for a specific database?

    I want to disable log shipping for a database temporarily.   I have disabled the backup job, copy job, and restore job created automatically when log shipping is configured for the database.  However, I cannot disable the Log Shipping alert
    job since there are other databases configured for log shipping.   How can I suppress Log Shipping alert for specific database?  I don't want to disable the log shipping for the primary database since it will delete all jobs and history
    related to the log shipping configuration at the primary, secondary and monitor server instances.

    Too late but this is possible. We just need to set the value threshold_alert_enabled to 0 in the system table msdb.dbo.log_shipping_monitor_primary on the primary server and in msdb.dbo.log_shipping_monitor_secondary on the secondary server.
    I just tested it out and it should work.
    To test this out, perform the below:-
    1. Run exec master.sys.sp_check_log_shipping_monitor_alert will show the same error message that thresholds have been crossed. This is the same script used in LSAlert job.
    Use something like the below command to edit:-
    2. Run the below to change the values:
    update msdb.dbo.log_shipping_monitor_primary
    set threshold_alert_enabled = 0
    where primary_database = 'XYZ'
    Run this for all databases that needs to be excluded from monitoring.
    3. Run the script in step 1 again and it should run now. 
    Please mark the answer as helpful if i have answered your query. Thanks and Regards, Kartar Rana

  • ORA-16191: Primary log shipping client not logged on standby.

    Hi,
    Please help me in the following scenario. I have two nodes ASM1 & ASM2 with RHEL4 U5 OS. On node ASM1 there is database ORCL using ASM diskgroups DATA & RECOVER and archive location is on '+RECOVER/orcl/'. On ASM2 node, I have to configure STDBYORCL (standby) database using ASM. I have taken the copy of database ORCL via RMAN, as per maximum availability architecture.
    Then I have ftp'd all to ASM2 and put them on FS /u01/oradata. Have made all necessary changes in primary and standby database pfile and then perform the duplicate database for standby using RMAN in order to put the db files in desired diskgroups. I have mounted the standby database but unfortunately, log transport service is not working and archives are not getting shipped to standby host.
    Here are all configuration details.
    Primary database ORCL pfile:
    [oracle@asm dbs]$ more initorcl.ora
    stdbyorcl.__db_cache_size=251658240
    orcl.__db_cache_size=226492416
    stdbyorcl.__java_pool_size=4194304
    orcl.__java_pool_size=4194304
    stdbyorcl.__large_pool_size=4194304
    orcl.__large_pool_size=4194304
    stdbyorcl.__shared_pool_size=100663296
    orcl.__shared_pool_size=125829120
    stdbyorcl.__streams_pool_size=0
    orcl.__streams_pool_size=0
    *.audit_file_dest='/opt/oracle/admin/orcl/adump'
    *.background_dump_dest='/opt/oracle/admin/orcl/bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='+DATA/orcl/controlfile/current.270.665007729','+RECOVER/orcl/controlfile/current.262.665007731'
    *.core_dump_dest='/opt/oracle/admin/orcl/cdump'
    *.db_block_size=8192
    *.db_create_file_dest='+DATA'
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='orcl'
    *.db_recovery_file_dest='+RECOVER'
    *.db_recovery_file_dest_size=3163553792
    *.db_unique_name=orcl
    *.fal_client=orcl
    *.fal_server=stdbyorcl
    *.instance_name='orcl'
    *.job_queue_processes=10
    *.log_archive_config='dg_config=(orcl,stdbyorcl)'
    *.log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST'
    *.log_archive_dest_2='SERVICE=stdbyorcl'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_format='%t_%s_%r.dbf'
    *.open_cursors=300
    *.pga_aggregate_target=121634816
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=364904448
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS'
    *.user_dump_dest='/opt/oracle/admin/orcl/udump'
    Standby database STDBYORCL pfile:
    [oracle@asm2 dbs]$ more initstdbyorcl.ora
    stdbyorcl.__db_cache_size=251658240
    stdbyorcl.__java_pool_size=4194304
    stdbyorcl.__large_pool_size=4194304
    stdbyorcl.__shared_pool_size=100663296
    stdbyorcl.__streams_pool_size=0
    *.audit_file_dest='/opt/oracle/admin/stdbyorcl/adump'
    *.background_dump_dest='/opt/oracle/admin/stdbyorcl/bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='u01/oradata/stdbyorcl_control01.ctl'#Restore Controlfile
    *.core_dump_dest='/opt/oracle/admin/stdbyorcl/cdump'
    *.db_block_size=8192
    *.db_create_file_dest='/u01/oradata'
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='orcl'
    *.db_recovery_file_dest='+RECOVER'
    *.db_recovery_file_dest_size=3163553792
    *.db_unique_name=stdbyorcl
    *.fal_client=stdbyorcl
    *.fal_server=orcl
    *.instance_name='stdbyorcl'
    *.job_queue_processes=10
    *.log_archive_config='dg_config=(orcl,stdbyorcl)'
    *.log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST'
    *.log_archive_dest_2='SERVICE=orcl'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_format='%t_%s_%r.dbf'
    *.log_archive_start=TRUE
    *.open_cursors=300
    *.pga_aggregate_target=121634816
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=364904448
    *.standby_archive_dest='LOCATION=USE_DB_RECOVERY_FILE_DEST'
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS'
    *.user_dump_dest='/opt/oracle/admin/stdbyorcl/udump'
    db_file_name_convert=('+DATA/ORCL/DATAFILE','/u01/oradata','+RECOVER/ORCL/DATAFILE','/u01/oradata')
    log_file_name_convert=('+DATA/ORCL/ONLINELOG','/u01/oradata','+RECOVER/ORCL/ONLINELOG','/u01/oradata')
    Have configured the tns service on both the hosts and its working absolutely fine.
    <p>
    ASM1
    =====
    [oracle@asm dbs]$ tnsping stdbyorcl
    </p>
    <p>
    TNS Ping Utility for Linux: Version 10.2.0.1.0 - Production on 19-SEP-2008 18:49:00
    </p>
    <p>
    Copyright (c) 1997, 2005, Oracle. All rights reserved.
    </p>
    <p>
    Used parameter files:
    </p>
    <p>
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.20.20)(PORT = 1521))) (CONNECT_DATA = (SID = stdbyorcl) (SERVER = DEDICATED)))
    OK (30 msec)
    ASM2
    =====
    </p>
    <p>
    [oracle@asm2 archive]$ tnsping orcl
    </p>
    <p>
    TNS Ping Utility for Linux: Version 10.2.0.1.0 - Production on 19-SEP-2008 18:48:39
    </p>
    <p>
    Copyright (c) 1997, 2005, Oracle. All rights reserved.
    </p>
    <p>
    Used parameter files:
    </p>
    <p>
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.20.10)(PORT = 1521))) (CONNECT_DATA = (SID = orcl) (SERVER = DEDICATED)))
    OK (30 msec)
    Please guide where I am missing. Thanking you in anticipation.
    Regards,
    Ravish Garg

    Following are the errors I am receiving as per alert log.
    ORCL alert log:
    Thu Sep 25 17:49:14 2008
    ARCH: Possible network disconnect with primary database
    Thu Sep 25 17:49:14 2008
    Error 1031 received logging on to the standby
    Thu Sep 25 17:49:14 2008
    Errors in file /opt/oracle/admin/orcl/bdump/orcl_arc1_4825.trc:
    ORA-01031: insufficient privileges
    FAL[server, ARC1]: Error 1031 creating remote archivelog file 'STDBYORCL'
    FAL[server, ARC1]: FAL archive failed, see trace file.
    Thu Sep 25 17:49:14 2008
    Errors in file /opt/oracle/admin/orcl/bdump/orcl_arc1_4825.trc:
    ORA-16055: FAL request rejected
    ARCH: FAL archive failed. Archiver continuing
    Thu Sep 25 17:49:14 2008
    ORACLE Instance orcl - Archival Error. Archiver continuing.
    Thu Sep 25 17:49:44 2008
    FAL[server]: Fail to queue the whole FAL gap
    GAP - thread 1 sequence 40-40
    DBID 1192788465 branch 665007733
    Thu Sep 25 17:49:46 2008
    Thread 1 advanced to log sequence 48
    Current log# 2 seq# 48 mem# 0: +DATA/orcl/onlinelog/group_2.272.665007735
    Current log# 2 seq# 48 mem# 1: +RECOVER/orcl/onlinelog/group_2.264.665007737
    Thu Sep 25 17:55:43 2008
    Shutting down archive processes
    Thu Sep 25 17:55:48 2008
    ARCH shutting down
    ARC2: Archival stopped
    STDBYORCL alert log:
    ==============
    Thu Sep 25 17:49:27 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-01017: invalid username/password; logon denied
    Thu Sep 25 17:49:27 2008
    Error 1017 received logging on to the standby
    Check that the primary and standby are using a password file
    and remote_login_passwordfile is set to SHARED or EXCLUSIVE,
    and that the SYS password is same in the password files.
    returning error ORA-16191
    It may be necessary to define the DB_ALLOWED_LOGON_VERSION
    initialization parameter to the value "10". Check the
    manual for information on this initialization parameter.
    Thu Sep 25 17:49:27 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-16191: Primary log shipping client not logged on standby
    PING[ARC0]: Heartbeat failed to connect to standby 'orcl'. Error is 16191.
    Thu Sep 25 17:51:38 2008
    FAL[client]: Failed to request gap sequence
    GAP - thread 1 sequence 40-40
    DBID 1192788465 branch 665007733
    FAL[client]: All defined FAL servers have been attempted.
    Check that the CONTROL_FILE_RECORD_KEEP_TIME initialization
    parameter is defined to a value that is sufficiently large
    enough to maintain adequate log switch information to resolve
    archivelog gaps.
    Thu Sep 25 17:55:16 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-01017: invalid username/password; logon denied
    Thu Sep 25 17:55:16 2008
    Error 1017 received logging on to the standby
    Check that the primary and standby are using a password file
    and remote_login_passwordfile is set to SHARED or EXCLUSIVE,
    and that the SYS password is same in the password files.
    returning error ORA-16191
    It may be necessary to define the DB_ALLOWED_LOGON_VERSION
    initialization parameter to the value "10". Check the
    manual for information on this initialization parameter.
    Thu Sep 25 17:55:16 2008
    Errors in file /opt/oracle/admin/stdbyorcl/bdump/stdbyorcl_arc0_4813.trc:
    ORA-16191: Primary log shipping client not logged on standby
    PING[ARC0]: Heartbeat failed to connect to standby 'orcl'. Error is 16191.
    Please suggest where I am missing.
    Regards,
    Ravish Garg

  • Transaction log shipping restore with standby failed: log file corrupted

    Restore transaction log failed and I get this error: for only 04 no of database in same SQL server, renaming are working fine.
    Date                     
    9/10/2014 6:09:27 AM
    Log                        
    Job History (LSRestore_DATA_TPSSYS)
    Step ID                
    1
    Server                  
    DATADR
    Job Name                           
    LSRestore_DATA_TPSSYS
    Step Name                        
    Log shipping restore log job step.
    Duration                             
    00:00:03
    Sql Severity        0
    Sql Message ID 0
    Operator Emailed           
    Operator Net sent          
    Operator Paged               
    Retries Attempted         
    0
    Message
    2014-09-10 06:09:30.37  *** Error: Could not apply log backup file '\\10.227.32.27\LsSecondery\TPSSYS\TPSSYS_20140910003724.trn' to secondary database 'TPSSYS'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred while processing the log for database 'TPSSYS'. 
    If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log.
    An error occurred during recovery, preventing the database 'TPSSYS' (13:0) from restarting. Diagnose the recovery errors and fix them, or restore from a known good backup. If errors are not corrected or expected, contact Technical Support.
    RESTORE LOG is terminating abnormally.
    Processed 0 pages for database 'TPSSYS', file 'TPSSYS' on file 1.
    Processed 1 pages for database 'TPSSYS', file 'TPSSYS_log' on file 1.(.Net SqlClient Data Provider) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  Skipping log backup file '\\10.227.32.27\LsSecondery\TPSSYS\TPSSYS_20140910003724.trn' for secondary database 'TPSSYS' because the file could not be verified.
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred restoring the database access mode.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteScalar requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred restoring the database access mode.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteScalar requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  Deleting old log backup files. Primary Database: 'TPSSYS'
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  The restore operation completed with errors. Secondary ID: 'dd25135a-24dd-4642-83d2-424f29e9e04c'
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not cleanup history.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.38  ----- END OF TRANSACTION LOG RESTORE    
    Exit Status: 1 (Error)

    I Have restore the database to new server and check with new log shipping but its give this same error again, If it is network issue i believe issue need to occur on every database in  that server with log shipping configuration
    error :
    Message
    2014-09-12 10:50:03.18    *** Error: Could not apply log backup file 'E:\LsSecondery\EAPDAT\EAPDAT_20140912051511.trn' to secondary database 'EAPDAT'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-12 10:50:03.18    *** Error: An error occurred while processing the log for database 'EAPDAT'.  If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log.
    An error occurred during recovery, preventing the database 'EAPDAT' (8:0) from restarting. Diagnose the recovery errors and fix them, or restore from a known good backup. If errors are not corrected or expected, contact Technical Support.
    RESTORE LOG is terminating abnormally.
    can this happened due to data base or log file corruption, if so how can i check on that to verify the issue
    Its not necessary if the issue is with network it would happen every day IMO it basically happens when load on network is high and you transfer log file which is big in size.
    As per message database engine was not able to restore log backup and said that you must rebuild log because it did not find out log to be consistent. From here it seems log corruption.
    Is it the same log file you restored ? if that is the case since log file was corrupt it would ofcourse give error on wehatever server you restore.
    Can you try creating logshipping on new server by taking fresh full and log backup and see if you get issue there as well. I would also say you to raise case with Microsoft and let them tell what is root cause to this problem
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Articles

  • Log shipping job failing

    HI guys,
    We are using SQL SERVER 2005.
    I am having a LSAlert_Serv job and this job runs the system stored procedure sys.sp_check_log_shipping_monitor_alert.
    when this job is run I am getting the following error message: 
    Here is the error message I am getting :
    The log shipping primary database SHARP has backup threshold of 60 minutes and has not performed
    a backup log operation for 7368 minutes. Check agent log and logshipping monitor information. [SQLSTATE 42000] (Error 14420). The step failed.
    The database named SHARP that is mentioned in the above error message is now moved to another
    server. 
    When I looked into the stored procedure and when i ran the below query from the Stored procedure
    select primary_server
    ,primary_database
    ,isnull(threshold_alert, 14420)
    ,backup_threshold
    ,cast(0 as int)
    from msdb.dbo.log_shipping_monitor_primary
    where threshold_alert_enabled = 1
    I can still see the database SHARP in the table msdb.dbo.log_shipping_monitor_primary. So,
    is it the reason for failure? If so, what to do to update the table msdb.dbo.log_shipping_monitor_primary and to fix the issue.
    Thanks

    The database named SHARP that is mentioned in the above error message is now moved to another server. 
    When I looked into the stored procedure and when i ran the below query from the Stored procedure :
    Since you said you moved database to different server can you please check that SQL Server service account( in new server where you moved database)  has full permissions on folder where you have configured log backup job to back the transaction logs.
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • Log shipping is not restoring log files ata particular time

    Hi,
    I have configured log shipping and it restores all the log files upto a particular time after which it throws error and not in consistent sate. I tried deleting and again configuring log shipping couple of times but no success. Can any one tell me 
    how to prevent it as I have already configured log shipping from another server to same destination server and it is working fine for more than 1 year? The new configuration is only throwing errors in restoration job.
    Thanks,
    Preetha

    Message
    2014-07-21 14:00:21.62    *** Error: The log backup file 'E:\Program Files\MSSQL10_50.MSSQLSERVER\MSSQL\Backup\Qcforecasting_log\Qcforecasting_20140721034526.trn' was verified but could not be applied to secondary database 'Qcforecasting'.(Microsoft.SqlServer.Management.LogShipping)
    2014-07-21 14:00:21.62    Deleting old log backup files. Primary Database: 'Qcforecasting'
    2014-07-21 14:00:21.62    The restore operation completed with errors. Secondary ID: '46b20de0-0ccf-4411-b810-2bd82200ead8'
    2014-07-21 14:00:21.63    ----- END OF TRANSACTION LOG RESTORE     -----
    The same file was tried thrice and it threw error all the 3 times.
    But when I manually restored the Qcforecasting_20140721034526 transaction log it worked. Not sure why this is happening. After the manual restoration it worked fine for one run.ow waiting for the other to complete.
    This seems strange to me error points to fact that backup was consistent but could not be applied because may be restore process found out that log backup was not it correct sequence or another log backup which was accidentally taken can be applied.
    But then you said you can apply this manually then this must be related to permission. Because if it can restore manually it can do it with job unless agent account has some permission issue.
    Ofcourse more logs would help
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • Shrink Log file in log shipping and change the database state from Standby to No recovery mode

    Hello all,
    I have configured sql server 2008 R2 log shipping for some databases and I have two issues:
    can I shrink the log file for these databases: If I change the primary database from full to simple and shrink the log file then change it back to full recovery mode the log shipping will fail, I've seen some answers talked about using "No
    Truncate" option, but as I know this option will not affect the log file and it will shrink the data file only.
          I also can't create maintenance to reconfigure the log shipping every time I want to shrink the log file because the database size is huge and it will take time to restore in the DR site, so the reconfiguration
    is not an option :( 
    how can I change the secondary database state from Standby to No recovery mode? I tried to change it from the wizard and wait until the next restore for the transaction log backup, but the job failed and the error was: "the step failed". I need
    to do this to change the mdf and ldf file location for the secondary databases.
    can any one help?
    Thanks in advance,
    Faris ALMasri
    Database Administrator

    1. can I shrink the log file for these databases: If I change the primary database from full to simple and shrink the log file then change it back to full recovery mode the log shipping will fail, I've seen some answers talked about using "No Truncate"
    option, but as I know this option will not affect the log file and it will shrink the data file only.
          I also can't create maintenance to reconfigure the log shipping every time I want to shrink the log file because the database size is huge
    and it will take time to restore in the DR site, so the reconfiguration is not an option :( 
    2. how can I change the secondary database state from Standby to No recovery mode? I tried to change it from the wizard and wait until the next restore for the transaction log backup, but the job failed and the error was: "the step failed". I need to do
    this to change the mdf and ldf file location for the secondary databases.
    can any one help?
    Thanks in advance,
    Faris ALMasri
    Database Administrator
    1. If you change recovery model of database in logshipping to simple and back to full Logshipping will break and logs wont be resored on Secondary server as log chain will be broken.You can shrink log file of primary database but why would you need that
    what is schedule of log backup. Frequent log backup is already taking care of log files why to shrink it and create performance load on system when log file will ultimately grow and since because instant file initilaization is not for Log files it takes time
    to grow and thus slows performace.
    You said you want to shrink as Database size is huge is it huge or does it have lots of free space. dont worry about data file free space it will eventually be utilized by SQL server when more data comes
    2. You are following wrong method changing state to no recovery would not even allow you to run select queries which you can run in Standby mode. Please refer below link to move Secondary data and log files
    http://www.mssqltips.com/sqlservertip/2836/steps-to-move-sql-server-log-shipping-secondary-database-files/
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • 2012 SQL server log shipping problem

    Scenario:
    Site 1: 2012 Enterprise version 64-bit (11.0.3128.0) - Log shipping is enabled for four databases.  Shipping logs over a WAN every 15 minutes to a DR site.
    Site 2: 2012 Enterprise version 64-bit (11.0.3128.0) - Secondary server for log shipping.
    Site two functions for a period of time with the databases in Standby/Read-Only mode, then the databases will mysteriously switch to "restoring..." mode and error 14420 Severity 16 State 1 will appear in the logs.  I have located kb2987585
    and this did apply to the situtation as the sector sizes were different, so as a work around I moved the log shipping shared folder on the secondary server to a drive with the same sector settings as the primary server but this did not resolve the issue.
    Has anyone had this issue and can you offer potential solutions?
    Thank you -

    Hi,
    I still consider the bug raised in
    https://support.microsoft.com/kb/2987585?wa=wsignin1.0
    is not fixed. I have seen lot of other users complaining. But at this point I need to have look at SQL Server errorlog
    sp_readerrorlog
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • Failed to send an environment change notification to a log shipping partner node while revalidating the login???

    Hi All,
    Recently I received the following error message in my SQL instance (Microsoft SQL Server 2012 - 11.0.5532.0 (X64)) :-
    23/02/2015 5:09:05 PM
    Error: 18456, Severity: 14, State: 51.
    Login failed for user 'webapp'. Reason: Failed to send an environment change notification to a log shipping partner node while revalidating the login. [CLIENT: 192.168.82.18]
     23/02/2015 5:09:05 PM
    Error: 18056, Severity: 20, State: 51.
    The client was unable to reuse a session with SPID 98, which had been reset for connection pooling. The failure ID is 51. This error may have been caused by an earlier operation failing. Check the error logs for failed operations immediately before this
    error message.
    I am using database mirroring instead of log shipping actually but why there's some log shipping related error message coming out? Is it some kind of bug error message?  I am just wondering and hope can get some little help about this.
    Thanks and highly appreciated.
    Best Regards,
                  Han

    Hi Harn,
    According to the error message, as my understanding, the issue could be due to that the login 'webapp' has no access to the default database which is configured in log shipping. The database is offline or removed. You could run the command (select * from
    msdb.dbo.log_shipping_primary_secondaries) to verify if that database is configured in log shipping. Another cause is that parallel execution occupies multiple worker threads. When you have many parallel executions, the number of available worker threads might
    run out and you fail to login. Please help to post complete SQL Server error log for analysis.
    To work around the issue, you could follow the two solutions below.
    1.Change the default database for the login 'webapp' according to the following article:
    https://msdn.microsoft.com/en-us/library/ms189828.aspx
    2.Reduce the 'max degree of parallelism' as the following example.
    exec sp_configure 'max degree of parallelism', 4;
    GO
    RECONFIGURE WITH OVERRIDE;
    GO
    Regards,
    Michelle Li

  • Log shipping Could not retrieve backup settings for primary ID

    hello,
    I implemented log shipping in our server the process of implementation went fine but when I view job history I found
    Message
    2014-06-11 12:00:01.53    *** Error: Could not retrieve backup settings for primary ID '99817903-626e-4380-bcf1-c09ca6f48b6d'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-06-11 12:00:01.53    *** Error: The specified agent_id 99817903-626E-4380-BCF1-C09CA6F48B6D or agent_type 0 do not form a valid pair for log shipping monitoring processing.(.Net SqlClient Data Provider) ***
    2014-06-11 12:00:01.53    *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-06-11 12:00:01.53    *** Error: The specified agent_id 99817903-626E-4380-BCF1-C09CA6F48B6D or agent_type 0 do not form a valid pair for log shipping monitoring processing.(.Net SqlClient Data Provider) ***
    2014-06-11 12:00:01.53    *** Error: Could not cleanup history.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-06-11 12:00:01.53    *** Error: The specified agent_id 99817903-626E-4380-BCF1-C09CA6F48B6D or agent_type 0 do not form a valid pair for log shipping monitoring processing.(.Net SqlClient Data Provider) ***
    2014-06-11 12:00:01.53    ----- END OF TRANSACTION LOG BACKUP   -----
    Exit Status: 1 (Error)
    also I check for database ID using select * from msdb..log_shipping_primary_databases
    your help is appreciated.
    Please Mark it as Answered if it answered your question
    OR mark it as Helpful if it help you to solve your problem
    Elmozamil Elamir Hamid
    MCSE Data Platform
    MCITP: SQL Server 2008 Administration/Development
    MCSA SQL Server 2012
    MCTS: SQL Server Administration/Development
    MyBlog

    Thank you all for your contribution.
    After testing and debugging I found that when they move their database from one server (server2) to another server (server1) the database still using the old name (server2) which you can findusing this script:
    SELECT @@SERVERNAME
    so the replication can't retrieve information, unfortunately for me it is difficult to rename the server because it sis development server so they need to make analysis about side effects if we renamed it. 
    Please Mark it as Answered if it answered your question
    OR mark it as Helpful if it help you to solve your problem
    Elmozamil Elamir Hamid
    MCSE Data Platform
    MCITP: SQL Server 2008 Administration/Development
    MCSA SQL Server 2012
    MCTS: SQL Server Administration/Development
    MyBlog

  • SQL Server 2012 - DB Mirroring + Log Shipping + Multi-Subnet

    Hi again,
    So AG would be nice to have, but considering the costs of getting it, well, it is a no go for me at the moment (This is like a 1:20 ratio considering Std comes with server licencing and enterprise is only per 2 cores).
    So, having 2 sites, with different subnet. Is it possible to configure automatic failover for DBM ? We are running Windows 2012 R2 which supports multi subnet failover. But what about DBM in this case ?
    Can the witness be configured on a third site ? (I was thinking of having the following weights: Site 1:2, Site 2:1, Quorum:2 just to have no reconfig to do in case of Datacenter failure)
    Also, in order to offload some of the backup load as the DB are somewhat big (1TB), I was thinking of having one additional node per site to which the logs are shipped and that are used to run the backups. The question being, is it possible to log ship from
    each "primary" on each site rather than have the primary on site 1 do it for all the servers across the subnets ?
    It would be a setup like this:
    Site 1 Primary <---DBM---> Site 2 Primary (Hot failover)
    Site 1 Primary <---LogShips---> Site 1 Secondary
    Site 2 Primary <---LogShips---> Site 2 Secondary
    And thus reduces the load on the link between the 2 sites.
    I do hope to see AG in the future for Std as they are removing DBM in the next version.
    Thanks,

    The idea behind the cluster is to have a single IP on which the applications can reach the DB in case of a failover to the secondary site.
    This requires a failover cluster instance - mirroring does not work this way.
    Because as far as I understand it, either the application has to be aware and able to switch to the new DB primary, or we can use a multisubnet cluster or we can manually reconfigure the applications (so painful)
    Mirroring has the failover partner attribute in the connection string that can be utilized. A multi-subnet cluster would require, again, a FCI.
    AFAIK, DBM is not recommended for more than 10 DB on an instance due to the worker threads.
    "It depends". There can be worker thread issues, but this will entirely depend on your hardware and architecture of processor/SQL Server. In 32-bit world it'll eat up all of your worker threads. In 64-bit world on a decent server it shouldn't be
    a problem to have 10-20 out of the box.
    I don't mind having to failover multiple database residing on the same instance and having to setup them per DB.
    If any of your databases have a dependency on one another, this isn't going to work well as one database could be failed over and another might not be. Just food for thought.
    What it sound like you really want is a failover cluster instance and use log shipping to have a "readable copy" somewhere even if it is behind on time. The problem comes in is that a cluster all has to be in the same domain and part of the same
    cluster. Your thoughts involve a separate cluster - so I'm not sure why two different ones unless you have multiple environments you're trying to add some HA to.
    -Sean
    The views, opinions, and posts do not reflect those of my company and are solely my own. No warranty, service, or results are expressed or implied.

Maybe you are looking for

  • Production Variances in COPA   CO-PA

    Hi Gurus Production variances are genrated in COPA at the time of Prod Order Settlement.How can we display these in COPA reports apportioned according to Products If not what is the best practice to present Production Variances in COPA reports

  • Is it possible to edit a MDT Task Sequence after starting a deployment, between Hyper-V snapshots?

    Using a Hyper-V VM, I take snapshots of my base image build between major sections (automatic updates, application installation, scripts and extras, etc.) with the intention of being able to go back to that area and look at it for troubleshooting's s

  • Add a view on custom view container

    Hello to everybody, first of all i'm not an expert on java wd. I'm triing to modify a stardard view of ESS package whit JDI. I need add a custom view . First of all i created my web dynpro component whit her view. Later i add a view container in stan

  • How to start creating multidimantional database

    hi; I am new user for oracle 10g. and I want to creat a multidimantional database using Oracle Enterprise Manager GUI to create my dimension?? I do not know how?? do I have first to creat the tables in sql* and then I continu in Oracle Enterprise Man

  • File management - changing "managed" to "referenced"

    I've imported several thousand images as managed files and would like to change them to referenced to conserve disk space. What is the easiest way to do this? And can it be done without disrupting the order of my folder/project hierarchy? Thanks, M