Data guard real time apply vs archived log apply on physical standby

Dear DBA's,
last week i configuared DR , now the phyiscal stanby database is archive apply mode,
i want to confirm is it better to apply the archived log or should i cahnge it to real time apply .
give me sugesstions.
Thanks and Regards
Raja...

One question are you using ARCH transport to move the redo? or have you configured standby redo logs and logwr transport (either async or syncronous), if you are using the archiver to transport the logs then you can not use real time apply.
If you are using log writer to transpor the redo the realtime apply reduces the recovery time required if you need to failover as trher should be less redo to apply to bring the standby up to date, which mode you use to transport redo will depend on what is acceptable in terms of data loss and the impact on performance.

Similar Messages

  • Could anybody explain me what is data guard real time sync ??

    This was asked as an interview question to me . am a fresher please help me out..
    thanks in advance.

    Hi,
    I suppose you mean Real Time Apply ? That's very popular mode used when the shipping method of standby databases is synchronous (as you probably know Dataguard support two redolog shipping modes - asynchronous and synchronous). In your case of asynchronous redolog shipping there is one step further - use of Real Time Apply instead of shipping archived logs. So instead of waiting for redologs to become archived (either by database log switch or setting ARCHIVE_LAG_TARGET) and then shipped you could set Real Time Apply where blocks from the log buffer of the primary database are send directly to the standby logs of the standby database. The benefits are obvious - the lag is almost few seconds i.e. changes got applied on the standby few seconds after the primary, you don't have to worry whether the archivelogs are sent or not.
    Here is link for more information about Real Time Apply at Oracle documentation Apply Services
    Regards,
    Sve

  • Data guard synchronization after link down b/w primary and physical standby

    Hi All,
    I have configured data guard on oracle 11gr2 db. Normally switchover between my primary and physical standby happens smoothly and the Apply lag would be zero. Recently We had to test a scenario when the network link between Primary and Physical standby is completely down and Physical standby is isolated completely for more than half an hour.
    When we brought up the link every thing worked smoothly but apply lag started increasing from 0 to around 3 hrs. And then it started reducing to 0. Currently Apply lag and transport lag shows 0.
    But is this normal behaviour of oracle data guard that when the link between primary and physical standby is completely down, It requires 3-4 hrs for resynchronization ??? Even when during isolation, there were very few transactions happend on primary database ??
    Are there any documents available for this scenario??
    Thanks

    Hi, after the link is up, if there were some transactions and produced archive logs it's normal to take some time for resync. To check if 3-4 hours is normal or not, you can repeat the scenario and this time check
    - how many archivelogs does primary produce in this period.
    - after the link is up, does archivelog transfer immediately starts from primary to standby? Is primary able to send these archivelogs parallel?
    - Is there anything wrong with the apply process?
    check primary & standby alert log files, and run this query on standby to check the transport and apply processes:
    SELECT PROCESS, STATUS, THREAD#, SEQUENCE#, BLOCK#, BLOCKS FROM V$MANAGED_STANDBY;
    regards

  • Delete archive log files on physical standby

    I want to delete the archive log files on the phsical standby database when they were applied. The archive log files are in the flash recovery area. So i used
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON STANDBY;
    in rman.
    My question:
    Does rman delete the archive log files immediately when they are applied or when the disk space becomes scarce?

    Hi,
    They're marked as reclaimable when applied, so they're still on the disk but when the flah recovery area will become full they'll be deleted to give space to the new one, and so...
    Loïc

  • How to switch from data guard really time apply (redo) to archive log

    Hi
    Standby database configured with broker and applying the redo in really time; however, I want to change this to archive log apply mode without losing the broker configuration. Is it possible? If it is not possible to use broker to do archive log apply, can I remove the broker and use data guard to set up the standby to use archive log apply?
    Regards

    user3076922 wrote:
    Hi
    Standby database configured with broker and applying the redo in really time; however, I want to change this to archive log apply mode without losing the broker configuration. Is it possible? If it is not possible to use broker to do archive log apply, can I remove the broker and use data guard to set up the standby to use archive log apply?
    RegardsHi
    I think mseberg is answered correct, you can use enable/disable apply log with change of state on standby database with DGMGRL, as writen mseberg.
    or you can disable recover standby database with following script from SQL*Plus.
    SQL> alter database recover managed standby database cancel;Regards
    Mahir M. Quluzade
    www.mahir-quluzade.com

  • Oracle Data Integrator - Real Time Integration

    Hi,
    I want to know that, is there any possibility of integrating data in real-time using Oracle Data Integrator?
    If yes, does it affect the OLTP system performance? (Could it read from db logs,etc..)
    Thanks..

    Using ODI with Logminer-based CDC will affect performance on the source system more than using Oracle Goldengate, let me explain why:
    When using the ODI Logminer-based Journalisation the Logminer functionality will move the primary-key of the affected row into the journal table. When you are then ready to move the changed data, running an interface to move the data reads the Journal-view, which joins the journal table (primary keys) to the source data (and dealing with deleted rows), optimising out duplicate rows in order to bring across the then-current state of the data, which can then be loaded into the target system, on completion, the moved rows can be removed from the journal table. The data will appear in the journal table as soon as Logminer puts it there, which may be a lag of up to two minutes using asynchronous setting, whereas Synchronous Logminer applies "system triggers to the table, with the consequent overhead.
    With Oracle Goldengate, the comitted transactions are read from the log as soon as the log-writer puts the commit into the log. All the data is picked up from the log, at that point. It is then written to the trail-file system of Oracle Goldengate which can be propogated to multiple other systems potentially with sub-second latency, and minimal impact on the source system due to the efficient reading and writing mechanisms. One other consequence of using Oracle Goldengate is that you get every change of the data, not the optimised then-current state of the data when moved.
    Hope this explanation helps.

  • Read data in real time and save as an excel file

    Hi,
    I want to write a LabVIEW progarmme which able to read data in real time and save it as an excel file from Varian Vacuum muli-gauge.
    It is using RS232 port.
    Can anyone give me some examples or point me in the right direction?
    I am a beginner of LabVIEW. Hope anyone can help me.
    Thank you very much!!
    Joanne

    Thanks for your reply.
    I just use MAX to verify that the rs232 port is operational.
    However, there is an error (please refer to the attachment).
    One possible reason is in MAX I am trying to do the default command *idn? ...but it doesn't work.
    I read the vacuum multi-gauge manual but I don't know which command should I use...
    I attached the manual and can you tell me which command should I use?
    Or can you tell me other possible reason for this error code?
    Thank you very much. 
    Joan
    Attachments:
    Varian Multi-Gauge Controller.pdf ‏2747 KB
    error1.JPG ‏111 KB

  • Viewing data in real-time and saving on-demand - USB4000 Spectrometer

    Hey,
    I am trying to create a program with a USB4000 Fiber Optic Spectrometer to allow the user to view the data in real-time on the front-panel, yet to be implemented, and also be able to save the data to an excel datasheet, or if another format if more appropriate. I am reasonably new to LabVIEW and have limited experience with saving data to excel.
    I am currently trying to build upon one of the example VIs given with the drivers. Currently the VI produces data like the following:
    Wavelength (nm)    Spectrum Data
    xyz                               xyz                      
    xyz                               xyz
    xyz                               xyz
    xyz                               xyz
    I want my VI to continue to append to a new column every iteration but I think due to me inputting an array of dimension size 2 it can't but am unsure how to get around this?
    The current VI starts off by creating the Headers and then appending to insert the spectrometer Data.
    Any help would be much appreciated, apologies if I am unclear about anything.
    Attachments:
    USB4000 15-07-14.vi ‏18 KB

    Ok, a couple comments. First, it is easier to build a 2D array by appending rows and then transposing the array before saving it. Also it is much more efficient to preallocate the array if you know how many columns the data is going to have.
    Second, adding a new column or row isn't a problem because an additional column (or row) doesn't effect the number of dimensions.
    Mike...
    Certified Professional Instructor
    Certified LabVIEW Architect
    LabVIEW Champion
    "... after all, He's not a tame lion..."
    Be thinking ahead and mark your dance card for NI Week 2015 now: TS 6139 - Object Oriented First Steps

  • Confirmation date in real time

    Hi, Is it possible to set the confirmation date as real time date meaning, the system should accept prsent date confirmation only, not yesterday confirmation in CO11N screen. Pls advise how to set the same.

    HI,
    If you know the Exit name, go to transaction CMOD.
    Choose menu Utillities->SAP Enhancements. Enter the exit name and press enter.
    You will now come to a screen that shows the function module exits for the exit.
    Using Project management of SAP Enhancements, we want to create a project to enahance trasnaction CO11N
    Go to transaction CMOD
    Create a project called ZCO11N
    Choose the Enhancement assign radio button and press the Change button.
    Please check this links which will help you to know more detail on user exits.
    http://sap-img.com/abap/a-short-tutorial-on-user-exits.htm
    http://sap-img.com/abap/what-is-user-exits.htm
    http://sap-img.com/abap/what-is-the-difference-between-smod-and-cmod.htm
    http://sap-img.com/ab038.htm
    http://help.sap.com/saphelp_46c/helpdata/en/4a/5b75387be80518e10000009b38f889/frameset.htm
    Hope this will help.
    Regards,
    R.Brahmankar

  • Standby database archive log apply in production server.

    Dear All,
    How I apply standby database archive log apply in production server.
    Please help me.
    Thanks,
    Manas

    How can i use standby database as primary for that 48 hour ?Perform a switchover (role transitions).
    First check if the standby is in sync with the primary database.
    Primary database:
    sql>select max(sequence#) from v$archived_log; ---> Value AStandby database:
    sql>select max(sequence#) from v$archived_log where applied='YES'; -----> Value BCheck if Value B is same as Value A.
    If the standby is in sycn with the primary database, then perform the switchover operation (refer the below link)
    http://www.articles.freemegazone.com/oracle-switchover-physical-standby-database.php
    http://docs.oracle.com/cd/B19306_01/server.102/b14230/sofo.htm
    http://www.oracle-base.com/articles/9i/DataGuard.php#DatabaseSwitchover
    manas
    Handle: manas
    Email: [email protected]
    Status Level: Newbie
    Registered: Jan 24, 2012
    Total Posts: 10
    Total Questions: 3 (3 unresolved)
    Name Manas
    Location kolkata Mark your questions as answered if you feel that you have got better answers rather than building up a heap of unanswered questions.

  • Hangup transport archive log in primary and archive log apply

    Hi
    I am building Dataguard from 3-node primary cluster to 3-node standby cluster
    Oracle Version:10.2.0.4
    Operating system : LInux 64 bit
    After I restored standby database, I configured dataguard broker with wrong unique_name parameter in standby cluster using grid control
    after I corrected mistake, disabled dataguard broker parameters, delete dataguard broker files and reboot standby cluster but don't reboot primary cluster because is production enviroment.
    I have problem with following symptoms:
    -Hangup transport archive log while recovery database in standby then gap archivelog is produced.
    -Copy and register all archivelog gap in standby but don't apply archive log.
    - Don't register like applied in v$archived_log in primary the archives applied in standby manually.
    -RMAN command: "backup as COMPRESSED BACKUPSET tag 'Backup Full Disk' archivelog all not backed up delete all input;"
    don't delete in primary archive log applied because message " archive log is necessary"
    I think that is necessary reboot primary cluster.
    Please helpme

    Post the results of queries. It is difficult to understand.
    post from primary
    SQL> select thread#,max(sequence#) from v$archived_log group by thread#;select     ds.dest_id id
    ,     ad.status
    ,     ds.database_mode db_mode
    ,     ad.archiver type
    ,     ds.recovery_mode
    ,     ds.protection_mode
    ,     ds.standby_logfile_count "SRLs"
    ,     ds.standby_logfile_active active
    ,     ds.archived_seq#
    from     v$archive_dest_status     ds
    ,     v$archive_dest          ad
    where     ds.dest_id = ad.dest_id
    and     ad.status != 'INACTIVE'
    order by
         ds.dest_id
    Post from standby.
    SQL> select thread#,max(sequence#) from v$archived_log where applied='YES' group by thread#;
    select * from v$managed_standby;

  • Export journalized data in real time

    Hello,
    I must pull journalized data in real time out of Oracle DB source.
    my package has ODIWaiForData operator and interface which triggers only in case of Journalized Data in my source table. Interface loads data to target Oracle table in real time and it works.
    Is it possible to send Journalized Data out of Oracle database in real time? Maybe some insert/update statements within sql file, csv file or through ODIOSCommand operator somehow?
    Regards

    Hi,
    did you already see about Logminer?
    Cezar Santos
    www.odiexperts.com

  • Will RMAN delete archive log files on a Standby server?

    Environment:
    Oracle 11.2.0.3 EE on Solaris 10.5
    I am currently NOT using an RMAN repository (coming soon).
    I have a Primary database sending log files to a Standby.
    My Retention Policy is set to 'RECOVERY WINDOW OF 8 DAYS'.
    Question: Will RMAN delete the archive log files on the Standby server after they become obsolete based on the Retention Policy or do I need to remove them manually via O/S command?
    Does the fact that I'm NOT using an RMAN Repository at the moment make a difference?
    Couldn't find the answer in the docs.
    Thanks very much!!
    -gary

    Hello again Gary;
    Sorry for the delay.
    Why is what you suggested better?
    No, its not better, but I prefer to manage the archive. This method works, period.
    Does that fact (running a backup every 4 hours) make my archivelog deletion policy irrelevant?
    No. The policy is important.
    Having the Primary set to :
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON STANDBYBut set to "NONE" on the Standby.
    Means the worst thing that can happen is RMAN will bark when you try to delete something. ( this is a good thing )
    How do I prevent the archive backup process from backing up an archive log file before it gets shipped to the standby?
    Should be a non-issue, the archive does not move, the REDO is transported and applied. There's SQL to monitor both ( Transport and Apply )
    For Data Guard I would consider getting a copy of
    "Oracle Data Guard 11g Handbook" - Larry Carpenter (AKA Dr. Paranoid ) ISBN 978-0-07-162111-2
    Best Oracle book I've read in 10 years. Covers a ton of ground clearly.
    Also Data Guard forum here :
    Data Guard
    Best Regards
    mseberg
    Edited by: mseberg on Apr 10, 2012 4:39 PM

  • 11.2 DB & Data Guard : ORA-16014 how to archive a sequence?

    Hi,
    I've installed 11.2 Oracle Database in my laptop with Oracle Enterprise Linux 5.3 and I have created two standby databases, orcl (primary) and orclstby (physical standby). I performed a switchover to orclstby and consequently, orclstby was the new primary and orcl the physical standby, I checked those values with SQLPlus executing the select database_role from v$database statement, so, there were no problems during the switchover.I also shutteddown and started both primary and stanby databases to check all was fine.
    Today, I've tried to start the environment again and I've encountered the following problem during the startup of the primary database:
    [oracle@mredon-es ~]$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.1.0 Production on Wed Oct 21 17:34:21 2009
    Copyright (c) 1982, 2009, Oracle. All rights reserved.
    Connected to an idle instance.
    SQL> startup
    ORACLE instance started.
    Total System Global Area 839282688 bytes
    Fixed Size 2217992 bytes
    Variable Size 507512824 bytes
    Database Buffers 322961408 bytes
    Redo Buffers 6590464 bytes
    Database mounted.
    ORA-03113: end-of-file on communication channel
    Process ID: 5562
    Session ID: 9 Serial number: 3
    I've checked the newest log created and this is its content:
    Dump file /u01/app/oracle/product/11.2.0/dbhome_1/rdbms/log/orclstby_ora_5382.trc
    *** 2009-10-21 17:34:30.467
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORACLE_HOME = /u01/app/oracle/product/11.2.0/dbhome_1
    System name: Linux
    Node name: mredon-es.us.oracle.com
    Release: 2.6.18-128.el5
    Version: #1 SMP Wed Jan 21 08:45:05 EST 2009
    Machine: x86_64
    Instance name: orclstby
    Redo thread mounted by this instance: 0 <none>
    Oracle process number: 0
    Unix process pid: 5382, image: [email protected]
    *** 2009-10-21 17:34:30.526
    2009-10-21 17:34:30.457: [ default]utgdv:2:ocr loc file /etc/oracle/olr.loc cann
    ot be opened. errno 2
    2009-10-21 17:34:30.527: [ default]utgdv:2:ocr loc file /etc/oracle/ocr.loc cann
    ot be opened. errno 2
    I've searched in Metalink information about ocr.loc and olr.loc and they seem to be part of an Oracle RAC installation, I don't know why the database needs these files to start if I am using single instance...
    Any idea would be kindly appreciated because I'm a bit confused and I don't really know what steps I am suppposed to take in order to solve this problem.
    Thanks in advance.
    Edited by: mredon on Oct 26, 2009 5:22 AM
    Reason: title of the post changed

    sorry for the delay, here is the alert log:
    <msg time='2009-10-21T23:03:09.576+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>Data Guard: version check completed
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:09.642+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6486'>
    <txt>LGWR: STARTING ARCH PROCESSES
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:09.709+02:00' org_id='oracle' comp_id='rdbms'
    msg_id='ksbrdp:3833:3697353022' type='NOTIFICATION' group='process start'
    level='16' host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130'
    pid='6588'>
    <txt>RSM0 started with pid=23, OS id=6588
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:09.831+02:00' org_id='oracle' comp_id='rdbms'
    msg_id='ksbrdp:3833:3697353022' type='NOTIFICATION' group='process start'
    level='16' host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130'
    pid='6590'>
    <txt>ARC0 started with pid=24, OS id=6590
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:10.832+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6486'>
    <txt>ARC0: Archival started
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:10.832+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6486'>
    <txt>LGWR: STARTING ARCH PROCESSES COMPLETE
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:10.833+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module=''
    pid='6590'>
    <txt>ARC0: STARTING ARCH PROCESSES
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.180+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>ARCH: LGWR is scheduled to archive destination LOG_ARCHIVE_DEST_2 after log switch
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.183+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>ARCH: LGWR is scheduled to archive destination LOG_ARCHIVE_DEST_1 after log switch
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.258+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>Errors in file /u01/app/oracle/product/diag/rdbms/orclstby/orclstby/trace/orclstby_ora_6582.trc:
    ORA-16014: log 1 sequence# 27 not archived, no available destinationsORA-00312: online log 1 thread 1: &apos;/u01/app/oracle/product/oradata/orclstby/redo01.log&apos;
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.263+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>USER (ospid: 6582): terminating the instance due to error 16014 </txt>
    </msg>
    <msg time='2009-10-21T23:03:11.429+02:00' org_id='oracle' comp_id='rdbms'
    msg_id='ksbrdp:3833:3697353022' type='NOTIFICATION' group='process start'
    level='16' host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130'
    pid='6592'>
    <txt>ARC1 started with pid=25, OS id=6592
    </txt>
    </msg>
    <msg time='2009-10-21T23:03:12.728+02:00' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='mredon-es.us.oracle.com' host_addr='192.168.237.130' module='[email protected] (TNS V1-V3)'
    pid='6582'>
    <txt>Instance terminated by USER, pid = 6582
    </txt>
    </msg>
    I've marked in bold the errors which may be causing the problem, ¿what do you think?

  • HOW CAN I ACQUIRE AND DEPLOY DATA IN REAL TIME WITH DAQ 6024E?

    HI ENGINEER,
    I'M PROGRAMING AN APLICATION WHICH I NEED ACUIRING AND DEPLOYING AN ANALOGOUS SIGNAL IN REAL TIME.
    I HAVE:
    LabVIEW 7.0
    WINDOWS 2000
    DAQ CARD 6024E
    THE SIGNAL I WANT TO ACQUIRE IS 0-10 Vcd FROM A SENSOR, THE PROBLEM I HAVE, IS THAT I AM ACQUIERING THE DATA AND IT SEEM THAT THEY ARE UPDATING EACH SECOND, BUT I HAVE BEEN MOVING THE PARAMETERS SUCH AS TASK TIMING IN THE DAQ ASSISTANT VI AND NOTHING HAPPENS.
    CAN YOU TELL ME WHICH PARAMETERS I HAVE TO MOVE TO AQUIRE AND DEPLOY DATA IN REAL TIME

    Hi barron,
    If you want to see the data update quicker, you can decrease the value in the "Samples to Read" field of your DAQ Assistant. This number controls when data is transferred from the DAQ buffer to the PC. If the sampling frequency remains the same, but this field decreases, it means the buffered data will be returned more quickly.
    Hope this helps.
    Thanks,
    Lesley Y.

Maybe you are looking for