Data Guard archive log remove

Hi,
I am using 9i Data Guard now. I try to set up automatic procedure to remove the archive log on the standby site once it got applied. But except the manual remove/delete, there is no option to set the automatic procedure in Oracle Data Guard setting.
Do anyone has solution for it?
Thanks

user3076922 wrote:
Hi
Standby database configured with broker and applying the redo in really time; however, I want to change this to archive log apply mode without losing the broker configuration. Is it possible? If it is not possible to use broker to do archive log apply, can I remove the broker and use data guard to set up the standby to use archive log apply?
RegardsHi
I think mseberg is answered correct, you can use enable/disable apply log with change of state on standby database with DGMGRL, as writen mseberg.
or you can disable recover standby database with following script from SQL*Plus.
SQL> alter database recover managed standby database cancel;Regards
Mahir M. Quluzade
www.mahir-quluzade.com

Similar Messages

  • Data Guard  Archive log size

    Hi Experts,
    I would like to know do we have any views where we can see the size of the archive log file transfered and applied to the physical standby database. I wanted to see how much space it takes in a day.
    Thanks
    Shaan
    Message was edited by:
    Shaan_dmp

    SQL> desc v$archived_log
    Name Null? Type
    RECID NUMBER
    STAMP NUMBER
    NAME VARCHAR2(257)
    DEST_ID NUMBER
    THREAD# NUMBER
    SEQUENCE# NUMBER
    RESETLOGS_CHANGE# NUMBER
    RESETLOGS_TIME DATE
    FIRST_CHANGE# NUMBER
    FIRST_TIME DATE
    NEXT_CHANGE# NUMBER
    NEXT_TIME DATE
    BLOCKS NUMBER
    BLOCK_SIZE NUMBER
    CREATOR VARCHAR2(7)
    REGISTRAR VARCHAR2(7)
    STANDBY_DEST VARCHAR2(3)
    ARCHIVED VARCHAR2(3)
    APPLIED VARCHAR2(3)
    DELETED VARCHAR2(3)
    STATUS VARCHAR2(1)
    COMPLETION_TIME DATE
    DICTIONARY_BEGIN VARCHAR2(3)
    DICTIONARY_END VARCHAR2(3)
    END_OF_REDO VARCHAR2(3)
    BACKUP_COUNT NUMBER
    ARCHIVAL_THREAD# NUMBER
    ACTIVATION# NUMBER
    Refer to blocks and block_size
    Other than that, you can look this up in the documentation on v$archived_log.
    Why do you act as a spoiled 3 year old, who wants everything on a golden plate, and can't be bothered to do anything himself?
    Blocks and block_size: That is really obvious, isn't it?
    It is just an issue of using your brains!!!
    Sybrand Bakker
    Senior Oracle DBA

  • Data Guard Archive Log Latency Between Primary and Physical Standby

    How can I get the time it takes (latency) for the primary instance to get an archive log over to the physical standby instance and get it "archived" and "applied". I have been looking at the V$ARCHIVED_LOG view on each side but the COLUMN "COMPLETION_TIME" always shows a date "MM/DD/YY" and no timestamp. I thought the DATE datatype include data and time. Any ideas on how I can get the latency info I'm looking for?
    Thanks
    Steve

    the COLUMN "COMPLETION_TIME" always shows a date "MM/DD/YY"
    and no timestamp. Did you try using TO_CHAR ? e.g.
    to_char(completion_time,'dd/mm/yyyy hh24:mi:ss')

  • Data Guard - Redo log files

    Hi....
    I have done data guard .......every thing is fine.......archives are bring transferred to standby..........
    Also, during configuration, I had created standby redolog groups 4,5,6 and copied to standby.....
    But in real time apply.......the standby is not using standby redolog groups 4,5,6........when i am query v$log it is showing group 1,2,3.
    Actually, its should use standby redo log for maximum availability.
    Please help....
    Thanks in advance.

    There was a similar question here just a few days ago:
    Data Guard - redo log files

  • Read data from Archive logs

    Does anyone have any recommandations on how to read data from archive logs. When i use log minor, i am getting only bind variables for DML operations. But i need actual data from the archive logs..
    Any thoughts
    Thanks
    -Prasad

    Log miner is the closest to command issued as possible. Depending on the Oracle version you will be able to see DML or DML and DDL. From 9i and on Oracle was able to translate the DML against data dictionary as the actual DDL command. On its first 8i release only DML was visible.
    ~ Madrid

  • How to delete the data in archived log files

    hi
    how can i delete the enteries in archived log files. and what is the disadvantage of deleting archived log enteries.

    There is no documented way to delete data stored in archived log files: you can only remove the archived log files if needed.

  • Data Warehouse Archive logging questions

    Hi all,
    I'd like some opinions/advice on archive logging and OWB 10.2 with a 10.2 database.
    Do you use archive logging on your non-production OWB instances? I have a development system that only has "on demand" backups done and the archive logs fill frequently. In this scenario, should I disable archive logging? I realize that this limits my recovery options to cold backups but on a development environment, this seems sufficient for me. Would I be messing up any OWB features by turning off archive logging?
    For production instances, how large do you make your archive log (as a percentage of your total DW size perhaps)?
    How do you manage them? With Flash recovery areas? Manually? RMAN or other tools?
    Thanks in Advance,
    Mike

    Usualy, I don't set any DW tables to log. Since it's a data warehouse, I believe it's better to make cold backups. In some cases, ETL Mappings may work like backup procedures themselves.
    In OWB, select the object you need (table or index) to create. Right-click it, select Configuration -> Performance Parameters -> Logging Mode -> NOLOGGING
    Flash RecoveryDon't think it's going to help you, since most of your data manipulation is based on batch jobs.
    RMANIf you want to make hot backups, this is something that can really help you manage backup procedures.
    ManuallyMaybe... Why not?
    I don't take hot backups from DW databases. I prefer to take cold backups. In a recovery scenario, you restore the cold backup and if it's 3 days late, I execute the ETL mappings for the last 3 days.
    Regards,
    Marcos

  • Oracle8i Data Guard with log shipping

    Is it true that :
    in Oralce8i, with the data guard, there will be zero data loss if online redologs have been mirrored in DR site and in the event of DR, the last un-finished redolog can be used to recover the database.
    What product is used to apply the redolog ?
    I know Oracle9i claim this is possible, but when will Oracle9i available to Sun platform ?

    >
    Thomas Schulz wrote:
    > Here are my questions:
    >
    > 1. Is it correct, that I have to restore the last successful restored log (if not the first) from the previous session with "recover_start", before I can restore the next log with "recover_replace" in a new session?
    Yes, that's correct. As soon as you leave the restore session, you have to provide a 'overlap' of log information so that the recovery can continue.
    > 2. Can't I mix the restoring of incremental and log backups in this way: log001, incremental data, log002, ...? In my tests I was able to restore the incremental data direct after the complete data, but not between the log backups.
    No, that's not possible. After you've recovered some log-information the incremental backup cannot be appliead as a "delta" to the data are anymore as the data area has already changed.
    > 3. Can I avoid the state change to OFFLINE after a log restore?
    Of course - don't use recover_cancel
    As soon as you stop the recovery, the database is stopped - no way around this.
    There are some 3rd party tools available for this, like LIBELLE.
    KR Lars

  • AIX Data Guard Isolating Log Traffic

    Anyone have experience in configuration of Data Guard on AIX? We would like to isolate log traffic to a separate circuit and are struggling with configuration to accomplish this. Would appreciate some suggestions.
    Allan Vath
    [email protected]

    We have successfully configured data guard between our two sites, but the network traffic for transporting the logs was significant. We purchased a separate data circuit for log transport but are struggling with a configuration to force the log file shipment to use the dedicated circuit. This is more of a network/aix issue, but hoped someone has implemented the same type of traffic isolation.
    Allan

  • Data Guard archive deleted from Primary datbase before transfaring Standby

    Dear all
    I have a physical standby database. And Primary database on RAC node 1 and node 2.
    Accidently archive log deleted from Primary db before shipped standby db .
    I saw this error which is showing the GAP.
    FAL[client]: Failed to request gap sequence
    GAP - thread 1 sequence 1_7090 1_7090
    ir
    Edited by: user8244545 on Jun 7, 2010 11:33 AM
    Edited by: user8244545 on Jun 7, 2010 6:47 PM

    Well, I feel sorry for you after all these guys gave you such a hard time :^)
    You should be using at least an RMAN Archive Log deletion policy to manage your archive log files.
    See "Deletion Policy for Archived Redo Log Files In Flash Recovery Areas" at http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/rman.htm#SBYDB00750 and http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/rman.htm#i1031870 for 10gR2. 10g does require you use a Flash Recovery Area (FRA, renamed to Fast Recovery Area) . Read the paper at http://www.oracle.com/technology/deploy/availability/pdf/RMAN_DataGuard_10g_wp.pdf to sort out not having to use MANDATORY (which we do not recommend).
    Now, if you are at 11g (R1 or R2) you don't even have to be using the FRA nor mandatory to have it work. See http://www.oracle.com/pls/db111/to_URL?remark=ranked&urlname=http:%2F%2Fdownload.oracle.com%2Fdocs%2Fcd%2FB28359_01%2Fbackup.111%2Fb28270%2Frcmconfb.htm%23BRADV89439.
    If you do not use RMAN to backup the archive logs at least these deletion policy will prevent this from happening again since you will no longer be deleting archive logs manually.
    Now, it may be too late for this, I am behind in my mail, but you could get the standby past this gap by using an incremental backup of the primary from the last SCN that was applied at the standby. See the manual at http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/scenarios.htm#CIHIAADC for more information.
    Larry

  • Data warehouse Archive log

    Are there any performance implications for archive log in DWH?
    Is it recommended to enable archive log in DWH?

    General speaking, it's still recommend to enable archive logging in DWH environment. With non-archivelog your only viable backup strategy is cold backup which require database shutdown.
    Turn on archivelog does have some performance impact on massive loading. However you can use nologging with direct patch/parallele loading to counter that, remember to take an immediate backup before/after loading.

  • Data Guard Archive Latency with Physical Standby

    I'm trying to get a latency number that tells me how long it took ARCH to move an archive file from the Primary instance to the physical standby instance and "apply it". I see lots of numbers but I can't make heads or tails of them. Does a metric such as this exist in a log file, a V$ view, or can it possibly be calculated or extrapolated?
    Thanks
    Steve ([email protected])

    This is how calculated the same in a test env.
    1. stop standby listener
    2. shutdown standby
    3. perform a data load (import)
    4. start the standby listener. This will cause arch files to be shipped to the standby. Monitor the files being shipped and time it
    5. start the standby db in recovery mode
    6. repeatedly execute the following query while timing the log apply services:
    SELECT SEQUENCE#,APPLIED FROM V$ARCHIVED_LOG ORDER BY SEQUENCE#;

  • 11g Data guard 建physical standby database配置相关问题

    General information
    OS:red hat Linux 2.6.32-200.13.1.el5uek x86_64(primary,standby)
    Home version:11.2.0.3(primary,standby)
    Situation:
    实验用库,想学习data guard,在另一个服务器上克隆了一个HOME并按照官方文档http://docs.oracle.com/cd/E11882_01/server.112/e25608/create_ps.htm#i1225703
    中去实验,步骤均按照官方文档来,最后在mount standby库时校验select sequence# from v$archived_log没有记录,出错日志如下(从mount开如截取)
    standby alert_dbacoe.log
    Error logs
    alter database mount
    Completed: alter database mount
    Error 604 received logging on to the standby
    +FAL[client, ARC2]: Error 604 connecting to PRIMARYSV for fetching gap sequence+
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo01_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 1 thread 1: '/u03/app/oracle/oradata/dbacoe/redo01.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo01_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 1 thread 1: '/u03/app/oracle/oradata/dbacoe/redo01.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 2 of thread 1
    ORA-00312: online log 2 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo02_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 2 thread 1: '/u03/app/oracle/oradata/dbacoe/redo02.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 2 of thread 1
    ORA-00312: online log 2 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo02_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 2 thread 1: '/u03/app/oracle/oradata/dbacoe/redo02.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 3 of thread 1
    ORA-00312: online log 3 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo03_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 3 thread 1: '/u03/app/oracle/oradata/dbacoe/redo03.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u05/oracle/app/oracle/diag/rdbms/standby/dbacoe/trace/dbacoe_lgwr_1912.trc:
    ORA-00313: open failed for members of log group 3 of thread 1
    ORA-00312: online log 3 thread 1: '/u04/app/oracle/redundancy/dbacoe/redo03_s.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ORA-00312: online log 3 thread 1: '/u03/app/oracle/oradata/dbacoe/redo03.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    ARC3: Archival started
    ARC0: STARTING ARCH PROCESSES COMPLETE
    Addition:
    因为是实验用库,所以结构有点怪,大体如下:
    Primary
    数据文件在BASE所在的oradata下 即/u03/app/oracle下,
    redo文件每组两份,一份在oradata下,另一份在同机的/u04/app/oracle/redundancy中(以_s结尾命名)
    archive文件在/u04/app/oracle/fast_recovery_area中
    standby
    数据文件在/u05/oracle/app/oracle/oradata下
    redundancy中的redo文件随着该目录放到oradata下的red中了
    archive文件目录则是/u05/oracle/app/oracle/fast_recovery_area中
    接下来是对应的init文件与TNS信息
    Primary端的init主要信息
    db_unique_name='PRIMARY'
    fal_client='PRIMARYSV'
    fal_server='STANDBYSV'
    log_archive_config='DG_CONFIG=(PRIMARY,STANDBY)'
    log_archive_dest_1='location=/u04/app/oracle/fast_recovery_area/DBACOE/archivelog VALID_FOR=(ALL_LOGFILES,ALL_ROLES) db_unique_name=PRIMARY'
    log_archive_dest_2='service=STANDBY ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) db_unique_name=STANDBY'
    log_archive_dest_state_1='enable'
    log_archive_dest_state_2='enable'
    db_file_name_convert='/u05/oracle/app/oracle/oradata/dbacoe','/u03/app/oracle/oradata/dbacoe',/u05/oracle/app/oracle/oradata/red','/u04/app/oracle/redundancy'
    log_file_name_convert='/u05/oracle/app/oracle/fast_recovery_area/DBACOE/archivelog','/u04/app/oracle/fast_recovery_area/DBACOE/archivelog'
    Standby端的init信息
    log_archive_config='DG_CONFIG=(PRIMARY,STANDBY)'
    db_unique_name='STANDBY'
    db_file_name_convert='/u03/app/oracle/oradata/dbacoe','/u05/oracle/app/oracle/oradata/dbacoe','/u04/app/oracle/redundancy','/u05/oracle/app/oracle/oradata/red'
    log_file_name_convert='/u04/app/oracle/fast_recovery_area/DBACOE/archivelog','/u05/oracle/app/oracle/fast_recovery_area/DBACOE/archivelog'
    log_archive_dest_1='location=/u05/oracle/app/oracle/fast_recovery_area/DBACOE/archivelog VALID_FOR=(ALL_LOGFILES,ALL_ROLES) db_unique_name=STANDBY'
    log_archive_dest_state_1=enable
    log_archive_format=log%t_%s_%r.arc
    log_archive_dest_2='SERVICE=PRIMARYSV ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PRIMARY'
    log_archive_dest_state_2=enable
    TNS信息如下(均能tnsping通)
    PRIMARYSV =
    +(DESCRIPTION =+
    +(ADDRESS = (PROTOCOL = TCP)(HOST = db1.ad.xxxxxx.com)(PORT = 1521))+
    +(CONNECT_DATA =+
    +(SERVER = DEDICATED)+
    +(SERVICE_NAME = dbacoe)+
    +)+
    +)+
    STANDBYSV =
    +(DESCRIPTION =+
    +(ADDRESS = (PROTOCOL = TCP)(HOST = db2.ad.xxxxx.com)(PORT = 1529))+
    +(CONNECT_DATA =+
    +(SERVER = DEDICATED)+
    +(SERVICE_NAME = dbacoe)+
    +)+
    +)+
    PS:请大家花费一点点时间阅读一下这冗长的段子,刚开学data guard,参数上设置都很生疏,(光照着文档做了,是否有redo log相关的操作遗漏?)敬请指教
    Edited by: 961394 on Dec 10, 2012 1:52 AM
    Edited by: 961394 on 2012-12-10 上午4:58

    结贴了,感谢maclean一群的Jesse Lui帮我解决问题

  • Goldengate extracting from Archive log

    Dear All,
    We are conducting a replication of Oracle Database. Due to high load on Production server, the client is providing us a near DR database server for extraction. This database server is using Oracle Data Guard to be in Sync with Production server. Hence we will be using Archive logs shipped from Production to DR to Extract changed data. I have searched on the Net and read manuals, but I am unable to understand the how to configure OGG for this scenario. I understand that we can used Tranlogoptions parameter to extract data from Archive logs, but since the near DR database is in Mount state, how can we Log in the database using userid /password in extract process. Also we need to have Select any dictionary privileges on the database, but database is in mount state, so how to work on it??
    Awaiting eagerly for the reply.l

    For a physical standby, you can configure GG to read the archived standby logs in archive log only (ALO) mode.
    ARCHIVEDLOGONLY causes Extract to read from the archived logs exclusively, without querying or validating the logs from system views such as v$log and v$archived_log. This parameter puts Extract into Archived Log Only mode (ALO). For more information, see the Oracle GoldenGate Oracle Installation and Setup Guide.
    You can configure the Extract process to read exclusively from the archived logs. This is
    known as Archived Log Only (ALO) mode. In this mode, Extract only reads from archived
    logs that are stored in a specified location. ALO mode allows Oracle GoldenGate to use
    production logs that are shipped over to a secondary database (such as a standby) as the
    data source for Oracle GoldenGate. The online logs will not be used. Oracle GoldenGate
    will connect to the secondary database to get metadata and other required data as needed.
    As an alternative, ALO mode is supported on the production system.
    Extract automatically operates in ALO mode if it detects that the database is a physical standby.
    Lots of other info in the Oracle install guide.

  • Oracle Data Guard 9i

    This is a quote from an Oracle publication on Oracle Data Guard 9i.
    "Oracle Data Guard's new SQL Apply feature creates a "logical standby" that enables continuous access to a working copy of your primary database. With SQL Apply, the standby database is continuously available for reporting and other operations, while always at-the-ready for unplanned failovers. Data Guard provides log shipping, remote mirroring, zero-data-loss disaster protection, and a single point of control and automation. "
    My question is: Is the standby database open in read-write or read-only mode?
    If, its only read, is there ANY other way read-write version of a database could be created at a remote site besides:
    a) Complete copying - duplicate database at remote site open in read-write.
    b) Distributed database - replication - remote users access their own version and the two keep themselves in sync.
    c) Parallel server - one database - two instances.

    Data Guard is available for Linux (within Oracle9i).
    It's a tool to make it much more easier to create a standby
    database, but I don't think it's required!
    Regards,
    Thomas.
    Is Oracle 9i data guard for Linux available??
    Well, my real question is, I need to setup 9i standby database
    on Linux. Do I need data guard installed in order to setup
    standby database??

Maybe you are looking for