RAC online and archive logs question

Hello All,
I setup a RAC database instances prod1 and prod2 (10.2.0.4). Datafiles and onlinelogs are on ASM.
Does these results look good queried from two instances. I am kind of concerned about the Group3 that has the same name for both the members.
Also archived logs are going to the ASM, is this a good practice. I was reading Oracle RMAN book and it mentioned archived logs go to local disk.
Is it possible to archive to local disk for online that are on ASM? Please advice. Early reply appreciated.. Thanks San~
PROD1 Instance
SQL> select member from v$logfile;
MEMBER
+DATA/prod/onlinelog/group_2.264.706892209
+FLASH/prod/onlinelog/group_2.259.706892211
+DATA/prod/onlinelog/group_1.261.706892209
+FLASH/prod/onlinelog/group_1.260.706892209
+DATA/prod/onlinelog/group_3.258.706892235
+FLASH/prod/onlinelog/group_3.258.706892235
+DATA/prod/onlinelog/group_4.256.706892237
+FLASH/prod/onlinelog/group_4.257.706892237
8 rows selected.
PROD2 Instance
SQL> select member from v$logfile;
MEMBER
+DATA/prod/onlinelog/group_2.264.706892209
+FLASH/prod/onlinelog/group_2.259.706892211
+DATA/prod/onlinelog/group_1.261.706892209
+FLASH/prod/onlinelog/group_1.260.706892209
+DATA/prod/onlinelog/group_3.258.706892235
+FLASH/prod/onlinelog/group_3.258.706892235
+DATA/prod/onlinelog/group_4.256.706892237
+FLASH/prod/onlinelog/group_4.257.706892237
8 rows selected.
===
SQL> archive log list
Database log mode Archive Mode
Automatic archival Enabled
Archive destination USE_DB_RECOVERY_FILE_DEST
Oldest online log sequence 3
Next log sequence to archive 4
Current log sequence 4
====
Thanks
San

Hi San,
sannidhi wrote:
Also archived logs are going to the ASM, is this a good practice. I was reading Oracle RMAN book and it mentioned archived logs go to local disk.
Is it possible to archive to local disk for online that are on ASM? Please advice. Early reply appreciated.. Thanks San~
It is recommend to store archived log files on ASM and on Shared disk, check your archive log format which suppose to represent uniqueness across all instances.
Yes, technically it is possible to archive to local disk, but not recommended as if you loose local disk there will be gaps in the archived log files and also it increases the administration.
Regards,
Thota

Similar Messages

  • Urgent: Huge diff in total redo log size and archive log size

    Dear DBAs
    I have a concern regarding size of redo log and archive log generated.
    Is the equation below is correct?
    total size of redo generated by all sessions = total size of archive log files generated
    I am experiencing a situation where when I look at the total size of redo generated by all the sessions and the size of archive logs generated, there is huge difference.
    My total all session redo log size is 780MB where my archive log directory size has consumed 23GB.
    Before i start measuring i cleared up archive directory and started to monitor from a specific time.
    Environment: Oracle 9i Release 2
    How I tracked the sizing information is below
    logon as SYS user and run the following statements
    DROP TABLE REDOSTAT CASCADE CONSTRAINTS;
    CREATE TABLE REDOSTAT
    AUDSID NUMBER,
    SID NUMBER,
    SERIAL# NUMBER,
    SESSION_ID CHAR(27 BYTE),
    STATUS VARCHAR2(8 BYTE),
    DB_USERNAME VARCHAR2(30 BYTE),
    SCHEMANAME VARCHAR2(30 BYTE),
    OSUSER VARCHAR2(30 BYTE),
    PROCESS VARCHAR2(12 BYTE),
    MACHINE VARCHAR2(64 BYTE),
    TERMINAL VARCHAR2(16 BYTE),
    PROGRAM VARCHAR2(64 BYTE),
    DBCONN_TYPE VARCHAR2(10 BYTE),
    LOGON_TIME DATE,
    LOGOUT_TIME DATE,
    REDO_SIZE NUMBER
    TABLESPACE SYSTEM
    NOLOGGING
    NOCOMPRESS
    NOCACHE
    NOPARALLEL
    MONITORING;
    GRANT SELECT ON REDOSTAT TO PUBLIC;
    CREATE OR REPLACE TRIGGER TR_SESS_LOGOFF
    BEFORE LOGOFF
    ON DATABASE
    DECLARE
    PRAGMA AUTONOMOUS_TRANSACTION;
    BEGIN
    INSERT INTO SYS.REDOSTAT
    (AUDSID, SID, SERIAL#, SESSION_ID, STATUS, DB_USERNAME, SCHEMANAME, OSUSER, PROCESS, MACHINE, TERMINAL, PROGRAM, DBCONN_TYPE, LOGON_TIME, LOGOUT_TIME, REDO_SIZE)
    SELECT A.AUDSID, A.SID, A.SERIAL#, SYS_CONTEXT ('USERENV', 'SESSIONID'), A.STATUS, USERNAME DB_USERNAME, SCHEMANAME, OSUSER, PROCESS, MACHINE, TERMINAL, PROGRAM, TYPE DBCONN_TYPE,
    LOGON_TIME, SYSDATE LOGOUT_TIME, B.VALUE REDO_SIZE
    FROM V$SESSION A, V$MYSTAT B, V$STATNAME C
    WHERE
    A.SID = B.SID
    AND
    B.STATISTIC# = C.STATISTIC#
    AND
    C.NAME = 'redo size'
    AND
    A.AUDSID = sys_context ('USERENV', 'SESSIONID');
    COMMIT;
    END TR_SESS_LOGOFF;
    Now, total sum of REDO_SIZE (B.VALUE) this is far less than archive log size. This at time when no other user is logged in except myself.
    Is there anything wrong with query for collecting redo information or there are some hidden process which doesnt provide redo information on session basis.
    I have seen the similar implementation as above at many sites.
    Kindly provide a mechanism where I can trace which user is generated how much redo (or archive log) on a session basis. I want to track which all user/process are causing high redo to generate.
    If I didnt find a solution I would raise a SR with Oracle.
    Thanks
    [V]

    You can query v$sess_io, column block_changes to find out which session generating how much redo.
    The following query gives you the session redo statistics:
    select a.sid,b.name,sum(a.value) from v$sesstat a,v$statname b
    where a.statistic# = b.statistic#
    and b.name like '%redo%'
    and a.value > 0
    group by a.sid,b.name
    If you want, you can only look for redo size for all the current sessions.
    Jaffar

  • Redo and archive log association

    Hi,
    I am curious about the redo file and archive log file association, is it one to one, or one to many? That is, does one arc log file hold data from just one redo file or many redo files?
    Or , is the association redo group to arc file as opposed to redo file to arc file?
    The size of the arc log files on my machine, sometimes far exceeds the size of a single redo file and sometimes goes well under the size.
    Thanks.

    I am curious about the redo file and archive log file association, is it one to one, or one to many?One archive log file represents one redo log file contents in a group.You can have multiple logfile members in a group. All the members have the same data.
    The size of the arc log files on my machine, sometimes far exceeds the size of a single redo file and sometimes goes well under the size.The size of the archive log can be smaller from redo logfile in following scenarios:--
    1. Manual log switch.
    2. Setting archive_lag_target parameter.
    sometimes far exceeds the size of a single redo file I am not very sure about it.I haven't seen size greater than the redo logfile.
    Anand
    Edited by: Anand... on Sep 2, 2009 7:45 PM

  • RMAN BACKUPS AND ARCHIVED LOG ISSUES

    제품 : RMAN
    작성날짜 : 2004-02-17
    RMAN BACKUPS AND ARCHIVED LOG ISSUES
    =====================================
    Scenario #1:
    1)RMAN이 모든 archived log들을 삭제할 때 실패하는 경우.
    database는 두 개의 archive destination에 archive file을 생성한다.
    다음과 같은 스크립트를 수행하여 백업후에 archived redo logfile을 삭제한다.
    run {
    allocate channel c1 type 'sbt_tape';
    backup database;
    backup archivelog all delete input;
    Archived redo logfile 삭제 유무를 확인하기 위해 CROSSCHECK 수행시 다음과
    같은 메시지가 발생함.
    RMAN> change archivelog all crosscheck;
    RMAN-03022: compiling command: change
    RMAN-06158: validation succeeded for archived log
    RMAN-08514: archivelog filename=
    /oracle/arch/dest2/arcr_1_964.arc recid=19 stamp=368726072
    2) 원인분석
    이 문제는 에러가 아니다. RMAN은 여러 개의 arhive directory중 하나의
    directoy안에 있는 archived file들만 삭제한다. 그래서 나머지 directory안의
    archived log file들은 삭제되지 않고 남게 되는 것이다.
    3) 해결책
    RMAN이 강제로 모든 directory안의 archived log file들을 삭제하게 하기 위해서는
    여러 개의 채널을 할당하여 각 채널이 각 archive destination안의 archived file을
    백업하고 삭제하도록 해야 한다.
    이것은 아래와 같이 구현될 수 있다.
    run {
    allocate channel t1 type 'sbt_tape';
    allocate channel t2 type 'sbt_tape';
    backup
    archivelog like '/oracle/arch/dest1/%' channel t1 delete input
    archivelog like '/oracle/arch/dest2/%' channel t2 delete input;
    Scenario #2:
    1)RMAN이 archived log를 찾을 수 없어 백업이 실패하는 경우.
    이 시나리오에서 database를 incremental backup한다고 가정한다.
    이 경우 RMAN은 recover시 archived redo log대신에 incremental backup을 사용할
    수 있기 때문에 백업 후 모든 archived redo log를 삭제하기 위해 OS utility를 사용한다.
    그러나 다음 번 backup시 다음과 같은 Error를 만나게 된다.
    RMAN-6089: archive log NAME not found or out of sync with catalog
    2) 원인분석
    이 문제는 OS 명령을 사용하여 archived log를 삭제하였을 경우 발생한다. 이때 RMAN은
    archived log가 삭제되었다는 것을 알지 못한다. RMAN-6089는 RMAN이 OS 명령에 의해
    삭제된 archived log가 여전히 존재하다고 생각하고 백업하려고 시도하였을 때 발생하게 된다.
    3) 해결책
    가장 쉬운 해결책은 archived log를 백업할 때 DELETE INPUT option을 사용하는 것이다.
    예를 들면
    run {
    allocate channel c1 type 'sbt_tape';
    backup archivelog all delete input;
    두 번째로 가장 쉬운 해결책은 OS utility를 사용하여 archived log를 삭제한 후에
    다음과 같은 명령어를 RMAN prompt상에서 수행하는 것이다.
    RMAN>allocate channel for maintenance type disk;
    RMAN>change archivelog all crosscheck;
    Oracle 8.0:
         RMAN> change archivelog '/disk/path/archivelog_name' validate;
    Oracle 8i:
    RMAN> change archivelog all crosscheck ;
    Oracle 9i:
    RMAN> crosscheck archivelog all ;
    catalog의 COMPATIBLE 파라미터가 8.1.5이하로 설정되어 있으면 RMAN은 찾을 수 없는
    모든 archived log의 status를 "DELETED" 로 셋팅한다. 만약에 COMPATIBLE이 8.1.6이상으로
    설정되어 있으면 RMAN은 Repository에서 record를 삭제한다.

    Very strange, I issue following command in RMAN on both primary and standby machine, but it they don't delete the 1_55_758646076.dbf, I find in v$archived_log, this "/home/oracle/app/oracle/dataguard/1_55_758646076.dbf" had already been applied.
    RMAN> connect target /
    RMAN> CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON ALL STANDBY;
    old RMAN configuration parameters:
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON ALL STANDBY;
    new RMAN configuration parameters:
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON ALL STANDBY;
    new RMAN configuration parameters are successfully stored
    RMAN>
    ----------------------------------------------------------------------------------

  • Putting a Oracle RAC in  to archive log mode??

    Our RAC is a 10gR2 Three(3) node RAC with a SAN as the shared storage. cluster is installed on linux. at the moment the RAC is in noarchieve log mode.
    SQL> archive log list;
    Database log mode No Archive Mode
    Automatic archival Disabled
    Archive destination /u01/app/oracle/product/10.2.
    0/db_1/dbs/arch
    Oldest online log sequence 665
    Current log sequence 666
    we created raw partitions in SAN and use ASM to manage shared storage.
    when running DBCA at "specify locations for the database files to be created" step we selected "use oracle managed files" and gave database area = +DATA/
    then at choose the recovery options for the database" step we ticked specify flash recovery area and gave Flash Recovery AREA = +BACKUP/ and flash recovery area = 10240 MB. we didnt click "enable archiving" option. hence the RAC DB is in NOARCHIEVELOG mode.
    i found some guide to put RAC in archive log mode which i have given below....
    ALTER SYSTEM SET log_archive_start=TRUE SCOPE=spfile;
    ALTER SYSTEM SET log_archive_dest_1='location=/u01/oradata/MYSID/archive/' SCOPE=spfile;
    ALTER SYSTEM SET log_archive_format='arch_%t_%s.arc' SCOPE=spfile;
    Since we need to mount the database in exclusive mode we must also alter the following parameter:
    ALTER SYSTEM SET cluster_database=FALSE SCOPE=spfile;
    From the command line we can stop the entire cluster using:
    srvctl stop database -d MYSID
    With the cluster down we can connect to a single node and issue the following commands:
    STARTUP MOUNT;
    ARCHIVE LOG START;
    ALTER DATABASE ARCHIVELOG;
    ALTER SYSTEM SET cluster_database=TRUE SCOPE=spfile;
    SHUTDOWN IMMEDIATE;
    Notice that the CLUSTER_DATABASE parameter has been reset to it's original value. Since the datafiles and spfile are shared between all instances this operation only has to be done from a single node.
    From the command line we can now start the cluster again using:
    srvctl start database -d MYSID
    my main question is about the setting of
    ALTER SYSTEM SET log_archive_dest_1='location= ????? ' SCOPE=spfile;
    since i have set +BACKUP/ as the FRA during installation how should i define the location of log_archive_dest_1???
    many thanks in advance.
    regards
    Dilan

    many thanks for the replies
    will the command you posted send the archive logs to the "+BACKUP/" location in my SAN.
    if yes how does oracle know that USE_DB_RECOVERY_FILE_DEST = BACKUP/ in my SAN. or how do i set oracle to use BACKUP/ as the USE_DB_RECOVERY_FILE_DEST??
    im a newbie to RAC so pls bear with me. also at the moment i get the following outputs from my database.... pls note the bold part below, is the archive destination pointing to a location in my local server?
    SQL> archive log list;
    Database log mode No Archive Mode
    Automatic archival Disabled
    Archive destination            /u01/app/oracle/product/10.2.0/db_1/dbs/arch
    Oldest online log sequence 665
    Current log sequence 666
    SQL> show parameter db_recovery;
    NAME TYPE VALUE
    db_recovery_file_dest string
    db_recovery_file_dest_size big integer 0
    SQL> show parameter archive
    NAME TYPE VALUE
    archive_lag_target integer 0
    log_archive_config string
    log_archive_dest string
    log_archive_dest_1 string
    log_archive_dest_10 string
    log_archive_dest_2 string
    log_archive_dest_3 string
    log_archive_dest_4 string
    log_archive_dest_5 string
    log_archive_dest_6 string
    log_archive_dest_7 string
    NAME TYPE VALUE
    log_archive_dest_8 string
    log_archive_dest_9 string
    log_archive_dest_state_1 string enable
    log_archive_dest_state_10 string enable
    log_archive_dest_state_2 string enable
    log_archive_dest_state_3 string enable
    log_archive_dest_state_4 string enable
    log_archive_dest_state_5 string enable
    log_archive_dest_state_6 string enable
    log_archive_dest_state_7 string enable
    log_archive_dest_state_8 string enable
    NAME TYPE VALUE
    log_archive_dest_state_9 string enable
    log_archive_duplex_dest string
    log_archive_format string %t_%s_%r.dbf
    log_archive_local_first boolean TRUE
    log_archive_max_processes integer 2
    log_archive_min_succeed_dest integer 1
    log_archive_start boolean FALSE
    log_archive_trace integer 0
    remote_archive_enable string true
    standby_archive_dest string ?/dbs/arch

  • Dataguard physical standby archive log question

    Hi all,
    I will try to keep this simple..
    I have a 4 node RAC primary shipping logs to a 2 node physical standby.
    On the primary when I run 'alter system archive log current' on an instance I only see 1 log being applied on the standby, that is by querying v$archived_log.
    If I run the following on the standby:
    select thread#,sequence#,substr(name,43,70)"NAME",registrar,applied,status,first_time from v$archived_log where first_time
    in
    (select max(first_time) from v$archived_log group by thread#)
    order by thread#
    I get:
    THREAD# SEQUENCE# NAME REGISTR APPLIED S FIRST_TIME
    1 602 thread_1_seq_602.2603.721918617 RFS YES A 17-jun-2010 12:56:58
    2 314 thread_2_seq_314.2609.721918627 RFS NO A 17-jun-2010 12:56:59
    3 311 thread_3_seq_311.2604.721918621 RFS NO A 17-jun-2010 12:57:00
    4 319 thread_4_seq_319.2606.721918625 RFS NO A 17-jun-2010 12:57:00
    Why do we only see the max(sequence#) having been applied and not all of them?
    This is the same no matter how many times I archive the current log files on any of the instances on the primary and also the standby does not have any gaps.
    Hope this is clear..
    any ideas?
    jd

    ok output from gv$archived_log on standby BEFORE 'alter system archive log current' on primary
    THREAD# SEQUENCE# NAME REGISTR APPLIED S FIRST_TIME
    1 679 thread_1_seq_679.1267.722001505 RFS NO A 18-jun-2010 11:58:22
    1 679 thread_1_seq_679.1267.722001505 RFS NO A 18-jun-2010 11:58:22
    2 390 thread_2_seq_390.1314.722001507 RFS NO A 18-jun-2010 11:58:23
    2 390 thread_2_seq_390.1314.722001507 RFS NO A 18-jun-2010 11:58:23
    3 386 thread_3_seq_386.1266.722001505 RFS YES A 18-jun-2010 11:58:22
    3 386 thread_3_seq_386.1266.722001505 RFS YES A 18-jun-2010 11:58:22
    4 393 thread_4_seq_393.1269.722001507 RFS NO A 18-jun-2010 11:58:23
    4 393 thread_4_seq_393.1269.722001507 RFS NO A 18-jun-2010 11:58:23
    Output from v$archived_log on standby AFTER 'alter system archive log current' on primary
    THREAD# SEQUENCE# NAME REGISTR APPLIED S FIRST_TIME
    1 680 thread_1_seq_680.1333.722004227 RFS NO A 18-jun-2010 11:58:29
    1 680 thread_1_seq_680.1333.722004227 RFS NO A 18-jun-2010 11:58:29
    2 391 thread_2_seq_391.1332.722004227 RFS NO A 18-jun-2010 11:58:30
    2 391 thread_2_seq_391.1332.722004227 RFS NO A 18-jun-2010 11:58:30
    3 387 thread_3_seq_387.1271.722004225 RFS NO A 18-jun-2010 11:58:28
    3 387 thread_3_seq_387.1271.722004225 RFS NO A 18-jun-2010 11:58:28
    4 394 thread_4_seq_394.1270.722004225 RFS YES A 18-jun-2010 11:58:29
    4 394 thread_4_seq_394.1270.722004225 RFS YES A 18-jun-2010 11:58:29
    as a reminder we have a 4 node RAC system shipping logs to a 2 node RAC standby. There are no gaps but only one log is ever registered as being applied.
    Why is that? Why arnt all logs registered as being applied?

  • Best option for hosting forms online and archiving the completed forms?

    Forgive the length of this post, but I feel it necessary to give a little background information into our problem.
    We were looking for a simple solution to a seemingly simple problem: Host ~15 safety forms online for field technicians (electricians) to fill out via their netbooks and submit to some sort of repository for archiving with the ability for the technicians to reopen the forms and make additions over the course of the job / week as well as firing an email off to our H&S rep per form submitted.
    Acrobat + Remote Desktop Services (SBS 2011 network) + Reader seemed like a good solution and does work, but it's somewhat cumbersome for the less computer literate techs and proving to be more time in support then we are potentially saving going with paperless forms. This is the system we have in place now.
    We initially purchased Acrobat Standard and subsequently upgraded to Professional when we found some limitations we weren't expecting in Acrobat Standard. An Adobe Sales rep also suggested that we check out Livecycle Designer 9, which was included with Acrobat Pro which included even more nice stuff. I started re-authoring the forms in LCD9.
    The issue now is that I have Dynamic XML forms with flowed content that cannot be submitted by Reader to sharepoint. Why? You can't reader extend livecycle forms with acrobat, which it is bundled with.....
    Why exactly adobe has two different and somewhat incompatible forms authoring tools bundled in the same product is beyond me, especially when one of them requires what I would imagine is an expensive server product to enable it's full use?
    So it seems I have a choice between having my employer purchase Reader Extensions ES2 to maintain the nice form design with flowed content, purchasing 12 copies of acrobat to put on the netbooks when Reader SHOULD be all that we need or converting all the dynamic XML forms authored in LCD9 into static AcroForms so they can be reader extended and deployed to sharepoint.
    We envision our technicians browsing to and logging in to our hosted sharepoint server, selecting the appropriate form template from a drop down list in a doc library or list, which opens the form for completion in the browser and on submit posts the form to a list on sharepoint for completed forms, preferably with some key fields mapped to custom columns in said list AND with the ability for the submitting technician to re-open the submitted form from sharepoint and make additions as the week / job progresses.
    What would be the best option given our current situation and needs? With all due respect, I am looking for suggestions from technically oriented or experienced persons, not sales reps that don't fully know the product.    

    Ok, so I have done some additional reading and I've found out that Livecycle forms can be submitted via reader without reader extensions so long as the submit is via HTTP Post or as XML data only.
    This effectively means you can collect the data from the submitted forms for processing, but you can not re-render the form from the form template and the XML without the presumably expensive server component, Reader Extensions, which for all I know, requires other components to run.
    This all seems awfully expensive for a solution to host a dozen or so forms and archive the submissions in one place where they can be re-opened and added to.
    I don't undestand why it's so prohibitively expensive to submit dynamic PDF forms as PDF vs submitting static PDF forms as PDF, which can be done with Acrobat Pro all by itself at ~$500.

  • FRA and archive logs

    Hello all,
    i have db from 10.2.0.4 to 11.2.0.3, my question is with regards to FRA. We do not use the full blow FRA, but only use it to store our Archive logs ONLY. This is still in testing phase for us...
    Our Current method it to backup the archive logs and blow them away (script ran from cron and RMAN)...
    and now from what i understand about FRA, if my archive logs are sitting in FRA then once its get 80% or whatever that value is, it will delete it for me as long as i have backed up my archive logs.... So my question is, is there a acutal threahold or no ?? Reading from below doc looks like this is unpredictable.... but is there a hidden parameter or some setting i can set to say....once the FRA is lets say 50% full, delete the files(as long as they are backed up first)
    http://docs.oracle.com/cd/B19306_01/backup.102/b14192/setup005.htm
    section -- 3.5.6.1 When Files are Eligible for Deletion from the Flash Recovery Area
    if there is no such setting, how can we setup managing archive logs (besides using rman to backup and delete)...

    Hello;
    It is not unpredictable. Oracle decides. The document is a bit vague to say the least.
    Just another reason to check your alert logs often.
    As a general rule I set db_recovery_file_dest_size on the high side.
    We kicked this around the Data Guard forum a few weeks back :
    FRA - Flashback logs usage varies on Primary against Standby Database
    Best Regards
    mseberg
    Edited by: mseberg on Jun 28, 2012 10:22 AM

  • RMAN,Data Guard and Archive log deletion

    Our DG environment is running Oracle 11g R2
    we have a 3 node DG environment with
    A being the Primary
    B and C being Active Data Guard Standbys
    Backups are taken off of B and go directly to tape.
    Standby Redo Logs and Fast Recovery Area are being used
    Taking recommendation from "Using Recovery Manager with Oracle Data Guard in Oracle Database 10g"
    RMAN Setting on Primary ("A")
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON STANDBY
    RMAN Setting on Standby ("B") where Backup is done
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE
    RMAN Setting on other Standby ("C")
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON STANDBY
    How can we know what archive logs are eligible to be deleted from "A" and "C" ?
    When does the delete take place?
    How can we tell when the archive logs are being deleted from "A" and "C" ?

    Dear user10260925,
    The documentation that you have read is reliable but insufficient.
    The Oracle can manage the archivelog directory and knows which one is eligible for the deletion. Those stuff you have posted in here has been taken from the online documentation and is supported and can be used when the Oracle knows and manages the archivelogs. That is simply called the flash recovery area. Please read about the FRA in this very moment.
    Under normal circumstances people in the industry uses some scripts to achive the archivelog deletion on the standby system.
    Here is a useful example to you;
    # Remove old archivelogs
    00,30 * * * * /home/oracle/scripts/delete_applied_redo_logs_OPTSTBY.sh
    vals3:/home/oracle#cat /home/oracle/scripts/delete_applied_redo_logs_OPTSTBY.sh
    export ORACLE_SID=optstby
    export ORACLE_HOME=/oracle/product/10.2.0/db_1
    cd /db/optima/archive/OPTPROD/archivelog
    /oracle/product/10.2.0/db_1/bin/sqlplus "/ as sysdba" @delete_applied_redo_logs.sql
    grep arc delete_applied_redo_logs.lst > delete_applied_redo_logs_1.sh
    chmod 755 delete_applied_redo_logs_1.sh
    sh delete_applied_redo_logs_1.sh
    rm delete_applied_redo_logs_1.sh
    rm delete_applied_redo_logs.lst
    vals3:/home/oracle#cd /db/optima/archive/OPTPROD/archivelog
    vals3:/db/optima/archive/OPTPROD/archivelog#cat delete_applied_redo_logs.sql
    set echo off
    set heading off
    spool /db/optima/archive/OPTPROD/archivelog/delete_applied_redo_logs.lst
    select 'rm -f ' || name from v$archived_log where applied = 'YES';
    spool off
    exit
    vals3:/db/optima/archive/OPTPROD/archivelog#Hope That Helps.
    Ogan

  • Full Backups, Level 0 Backups, and Archived Logs

    We have an active Oracle server and a standby Oracle server. We keep the standby database up to date with a cron script. The script tells the active database to do 'alter system switch logfile;'. We then rsync the archived logs to our standby server and have rman apply them.
    This works everyday except Monday (of course!) and it only recently started failing on Mondays. The only change was that our Sunday backups used to be 'Full' backups but are now 'level 0' backups. Ever since that change, the first attempt to apply the archived logs to the standby server after the level 0 is taken on the active server gives us something like this:
    ORA-00308: cannot open archived log
    '/opt/oracle/flash_recovery_area/ORCL/archivelog/2012_04_16/o1_mf_1_60519_%u_.arc'
    ORA-27037: unable to obtain file status
    Of course, the file is not there and doesn't exist on the active server either. And of course, the nightly level1 backups fo not give us problems applying archived logs to the standby database the rest of the week.
    The only way I know to recover from this is to apply the level 0 backup or take a new level 0 and apply it. After that, all subsequent archive logs just work. Any idea why changing from Full to Level 0 would break this? The Oracle docs insist that a Level 0 is identical to a Full except that level 1s can reference them as parents. This simply cannot be true based on what I'm seeing! I really want to keep the level 0 backups in play if possible. Level 1 cumulatives wont be useful without them.

    Here are the RMAN settings:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 7 DAYS;
    CONFIGURE BACKUP OPTIMIZATION OFF; # default
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP OFF; # default
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '%F'; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '/opt/oracle/102/dbs/snapcf_ORCL.f'; # default
    I'm not sure how changing ARCHIVELOG BACKUP COPIES would help. Can you give me a little more information about how that setting comes into play in this situation?
    I actually don't want an archive deletion policy here. We have this done in script three days after the needed archive logs have been applied. Is it possible that the we're deleting archivelogs too soon? Would we ever need to reach back in time to previously applied archive logs to apply new ones?
    The %u does resolve, but this message isn't showing it. Here is that same log entry plus a few previous entries that show it does resolve.
    ORA-00279: change 1284618956 generated at 04/13/2012 15:30:05 needed for thread
    1
    ORA-00289: suggestion :
    /opt/oracle/flash_recovery_area/ORCL/archivelog/2012_04_16/o1_mf_1_60518_%u_.arc
    ORA-00280: change 1284618956 for thread 1 is in sequence #60518
    ORA-00278: log file
    '/opt/oracle/flash_recovery_area/ORCL/archivelog/2012_04_13/o1_mf_1_60517_7rjzox
    0l_.arc' no longer needed for this recovery
    ORA-00279: change 1284618958 generated at 04/13/2012 15:30:05 needed for thread
    1
    ORA-00289: suggestion :
    /opt/oracle/flash_recovery_area/ORCL/archivelog/2012_04_16/o1_mf_1_60519_%u_.arc
    ORA-00280: change 1284618958 for thread 1 is in sequence #60519
    ORA-00278: log file
    '/opt/oracle/flash_recovery_area/ORCL/archivelog/2012_04_13/o1_mf_1_60518_7rjzox
    0x_.arc' no longer needed for this recovery
    ORA-00308: cannot open archived log
    '/opt/oracle/flash_recovery_area/ORCL/archivelog/2012_04_16/o1_mf_1_60519_%u_.ar
    c'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3

  • Database refresh using dump and archive logs

    Hi all,
    I have a full datapump dump taken from oracle 11g R2 database (PROD) running in HP-UX server.The dump was taken on March,1,2012.
    Also i have all the archive logs till today (March 13,2012).
    I want to clone it to a new database (TEST) in windows machine with this dump, and i want to refresh(restore) the database (TEST) using this archive logs to make the database sync with the (PROD) till today.
    I need your suggestions.

    raoofdba wrote:
    Hi all,
    I have a full datapump dump taken from oracle 11g R2 database (PROD) running in HP-UX server.The dump was taken on March,1,2012.
    Also i have all the archive logs till today (March 13,2012).
    I want to clone it to a new database (TEST) in windows machine with this dump, and i want to refresh(restore) the database (TEST) using this archive logs to make the database sync with the (PROD) till today.
    I need your suggestions.I suggest you to perform transport tablespace method.below is the link: -
    http://neeraj-dba.blogspot.in/2012/01/cross-platform-transportable.html

  • Data Warehouse Archive logging questions

    Hi all,
    I'd like some opinions/advice on archive logging and OWB 10.2 with a 10.2 database.
    Do you use archive logging on your non-production OWB instances? I have a development system that only has "on demand" backups done and the archive logs fill frequently. In this scenario, should I disable archive logging? I realize that this limits my recovery options to cold backups but on a development environment, this seems sufficient for me. Would I be messing up any OWB features by turning off archive logging?
    For production instances, how large do you make your archive log (as a percentage of your total DW size perhaps)?
    How do you manage them? With Flash recovery areas? Manually? RMAN or other tools?
    Thanks in Advance,
    Mike

    Usualy, I don't set any DW tables to log. Since it's a data warehouse, I believe it's better to make cold backups. In some cases, ETL Mappings may work like backup procedures themselves.
    In OWB, select the object you need (table or index) to create. Right-click it, select Configuration -> Performance Parameters -> Logging Mode -> NOLOGGING
    Flash RecoveryDon't think it's going to help you, since most of your data manipulation is based on batch jobs.
    RMANIf you want to make hot backups, this is something that can really help you manage backup procedures.
    ManuallyMaybe... Why not?
    I don't take hot backups from DW databases. I prefer to take cold backups. In a recovery scenario, you restore the cold backup and if it's 3 days late, I execute the ETL mappings for the last 3 days.
    Regards,
    Marcos

  • Hangup transport archive log in primary and archive log apply

    Hi
    I am building Dataguard from 3-node primary cluster to 3-node standby cluster
    Oracle Version:10.2.0.4
    Operating system : LInux 64 bit
    After I restored standby database, I configured dataguard broker with wrong unique_name parameter in standby cluster using grid control
    after I corrected mistake, disabled dataguard broker parameters, delete dataguard broker files and reboot standby cluster but don't reboot primary cluster because is production enviroment.
    I have problem with following symptoms:
    -Hangup transport archive log while recovery database in standby then gap archivelog is produced.
    -Copy and register all archivelog gap in standby but don't apply archive log.
    - Don't register like applied in v$archived_log in primary the archives applied in standby manually.
    -RMAN command: "backup as COMPRESSED BACKUPSET tag 'Backup Full Disk' archivelog all not backed up delete all input;"
    don't delete in primary archive log applied because message " archive log is necessary"
    I think that is necessary reboot primary cluster.
    Please helpme

    Post the results of queries. It is difficult to understand.
    post from primary
    SQL> select thread#,max(sequence#) from v$archived_log group by thread#;select     ds.dest_id id
    ,     ad.status
    ,     ds.database_mode db_mode
    ,     ad.archiver type
    ,     ds.recovery_mode
    ,     ds.protection_mode
    ,     ds.standby_logfile_count "SRLs"
    ,     ds.standby_logfile_active active
    ,     ds.archived_seq#
    from     v$archive_dest_status     ds
    ,     v$archive_dest          ad
    where     ds.dest_id = ad.dest_id
    and     ad.status != 'INACTIVE'
    order by
         ds.dest_id
    Post from standby.
    SQL> select thread#,max(sequence#) from v$archived_log where applied='YES' group by thread#;
    select * from v$managed_standby;

  • Archive log question

    i have a db in archive log mode
    i found harddisk always full with archive log
    and db can not work then
    so i decide to comment this 3 parameter
    # log_archive_start = true
    #log_archive_dest_1= "location=/u01/app/oracle/admin/dbname/arch"
    # log_archive_format = arch_%t_%s.arc
    in init.ora
    then i enter into svrmgrl
    shutdown abort
    startup
    db open and i found db continued to archive redolog
    when redolog are full
    db stop work and alertlog report all redolog need archive
    why ?

    A better solution would be to create a job at the OS level that
    would run automatically at set times to backup the archive logs
    to another location, then remove the backed up logs from your
    drive. Depending on the available space, and the amount of log
    that your database creates, you can set the timing of this job
    to keep some space free all the time. On most of our servers,
    the timing ranges from every 2 hours to once a week.
    When you stop archiving, you can only recover the database to
    the time of your last backup. Also, the backup needs to be
    done "cold", that is, with the database shutdown. Worst case;
    You backup tonight, work all day tomorrow and seconds before the
    backup finishes, your disk explodes. Can you afford to lose 24
    hours work?

  • RMAN and archived log

    I have 2 identical Database from day 1 (1 Prod and 1 Dev). Prod db runs in archive mode
    which produces a number of archived logs after a few hours.
    Can I use rman to apply these archived log to the Dev instance to get a copy of the Prod DB? I've copied all *.arc files rom
    Prod back to Oracle\oradata\devdb directory
    1. rman target internal/<passwd> nocatalog
    2. shutdown normal;
    3. startup mount pfile=init.ora
    4. run {
    allocate channel ch1 type disk;
    restore archivelog from logseq 11 until logseq 13;
    recover database;
    I am getting an error saying
    RMAN 06050: archivelog thread 1 sequence 11 is already on disk ....
    RMAN 06177: restore not done, all files readonly, offline or already restored?
    Thanks
    null

    Hi!
    The quick answer is no. But there are several things to know.
    1. Production Database in archive mode.
    2. Develpment database in no archive mode?
    You cannot apply the redo log of (1) to (2) because the SCN (sequence change number) and won't be the same in the both databases. Means more, except if you clone the databases, the control files could be different so remember that redo-logs are binaries, not just a clear-text script.
    Other think to know is that in development you can modify, delete apppend data/tables/datafiles/tablespaces.. how can synchronize this with the production? (this situation can be the same production vs development).
    If you have a standby database you can do something like you wish, but not identical (tell me and I will explain my little knowledge about it ;)
    So you must first know what do you wanna keep in development database, if you just want to have a daily-mirror and can lost production data the best option is to backup production database with RMAN (full, incremental...) in development and use it to replicate another instance every day. Well I have some troubles replicating with RMAN, but people here say it's possible ;)
    If you want to keep the same data, you can use an incremental data export (exp utility) and import modified data to development.
    A10!
    P.S:Sorry for this loooooong text ;)

Maybe you are looking for