Archive log in WAS J2EE

Hello,
I install a WAS J2EE 6.40 SP13 for EP6.0 and I want to know how can I delete log file after archiving them with the Visual Administrator -> Cluster -> <SAPSID> -> Server -> Services -> Log Configurator -> Archive ?
Thanks,
Regards,
Chris

Hi Chris,
> So my log file size up every day !
No, only up to the size (limit * count) for each defined destination. For "limit" and "count" see http://help.sap.com/saphelp_nw04/helpdata/en/04/4795bf75082a4088b9093e4722fa86/frameset.htm
For log rotating politics, see http://help.sap.com/saphelp_nw04/helpdata/en/c7/d80f13d70e49c58e0d305ef2b62a05/frameset.htm and/or http://help.sap.com/saphelp_nw04/helpdata/en/f6/7d814282cfc153e10000000a1550b0/frameset.htm
Hope it helps
Detlev

Similar Messages

  • Generate archive logs are not in sequence number?

    On last friday... the latest archive log number was ARC00024.ARC. Tomorrow when I come backup, the archive logs ARC00001.ARC and ARC00002.ARC were being generated by oracle itself. I wondering the archive log sequence should be in sequence. What is happening?
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination C:\oracle\ora92\RDBMS
    Oldest online log sequence 1
    Next log sequence to archive 3
    Current log sequence 3
    SQL>
    FAN
    Edited by: user623471 on Jun 7, 2009 7:35 PM

    khurram,
    Its our production instance and havent issued resetlogs option but when listing the arvchives it shows in different sequence number...
    and also while copying the archives by RMAN it doesnt copy in sequence
    -rw-r----- 1 xxx dba 69363859 May 28 19:16 2_10373.arc.gz
    -rw-r----- 1 xxx dba 43446622 May 28 19:16 1_10553.arc.gz
    -rw-r----- 1 xxx dba 52587365 May 28 19:16 1_10578.arc.gz
    -rw-r----- 1 xxx dba 45251820 May 28 19:16 1_10543.arc.gz
    -rw-r----- 1 xxx dba 60890256 May 28 19:17 1_10579.arc.gz
    -rw-r----- 1 xxx dba 46659008 May 28 19:17 1_10548.arc.gz
    -rw-r----- 1 xxx dba 116899466 May 28 19:17 2_10353.arc.gz
    -rw-r----- 1 xxx dba 77769517 May 28 19:17 1_10531.arc.gz
    -rw-r----- 1 xxx dba 66401923 May 28 19:18 1_10530.arc.gz
    -rw-r----- 1 xxx dba 45972697 May 28 19:18 1_10605.arc.gz
    -rw-r----- 1 xxx dba 55082543 May 28 19:18 1_10600.arc.gz
    -rw-r----- 1 xxxq dba 42682207 May 28 19:19 1_10547.arc.gz
    thanks,
    baskar.l

  • Data guard - missed archive log

    Hi:
    I am on 10.2.0.3, using physical standby. One of my check-up sqls shows missed (not applied) arch log:
    SELECT THREAD#, SEQUENCE#, APPLIED FROM V$ARCHIVED_LOG
    where applied='NO'
    THREAD# SEQUENCE# APP
    1 16595 NO
    It's relatively old one, my max is 18211i and the log application is going fine.
    How can I apply this missed one? Do I have to stop managed recovery and do manual recovery for it?
    TIA.

    GregX,
    Please check if this file actually needs to be applied.
    You can find out the when the archived log file was created – FIRST_TIME/NEXT_TIME columns in V$ARCHIVED_LOG.
    If the archived log file was created before you created the stand-by DB, you do not need ( nor you can) to apply that file.
    Please note that the physical standby would stop applying archivelog if you fail to provide the next required archive log file. Since this has not happened (as per your statement) there is a good chance that you archive redo in question is not needed.
    Hope that helps,
    Edited by: IordanIotzov on Jan 30, 2009 1:34 PM

  • Why archive log is full so fast?

    Hi All,
    I have installed 10g RAC on RHEL4 using MSA1000 as a shared storage. and using ASM for database files and archive log files (external reduandancy) with RAID5. I have allocated 100GB for archive log files. database also created successfully with archive log enabled. After that i import 3 schemas in the RAC db each of 2GB size. The import was successful and when i try to connect the schema using Sqlplus it will "archive error". archive log space was full. i checked in asmcmd also, the 100GB fully occupied. My doubt is..How come the 100GB space is filled so fast?
    Can anybody help me?
    Thanks,
    Praveen.

    Hi Don,
    No error, just code from an older release! Right ?
    Not found any reference of the TIME column from V$LOG_HISTORY into the 8i and 9i doc. Maybe from Oracle 7 ?
    Connected to:
    Oracle8i Enterprise Edition Release 8.1.7.4.1 - Production
    With the Partitioning option
    JServer Release 8.1.7.4.1 - Production
    SQL> select substr(time,1,5) day,
      2  to_char(sum(decode(substr(time,10,2),'00',1,0)),'99') "00",
      3  to_char(sum(decode(substr(time,10,2),'01',1,0)),'99') "01",
      4  to_char(sum(decode(substr(time,10,2),'02',1,0)),'99') "02",
      5  to_char(sum(decode(substr(time,10,2),'03',1,0)),'99') "03",
      6  to_char(sum(decode(substr(time,10,2),'04',1,0)),'99') "04",
      7  to_char(sum(decode(substr(time,10,2),'05',1,0)),'99') "05",
      8  to_char(sum(decode(substr(time,10,2),'06',1,0)),'99') "06",
      9  to_char(sum(decode(substr(time,10,2),'07',1,0)),'99') "07",
    10  to_char(sum(decode(substr(time,10,2),'08',1,0)),'99') "08",
    11  to_char(sum(decode(substr(time,10,2),'09',1,0)),'99') "09",
    12  to_char(sum(decode(substr(time,10,2),'10',1,0)),'99') "10",
    13  to_char(sum(decode(substr(time,10,2),'11',1,0)),'99') "11",
    14  to_char(sum(decode(substr(time,10,2),'12',1,0)),'99') "12",
    15  to_char(sum(decode(substr(time,10,2),'13',1,0)),'99') "13",
    16  to_char(sum(decode(substr(time,10,2),'14',1,0)),'99') "14",
    17  to_char(sum(decode(substr(time,10,2),'15',1,0)),'99') "15",
    18  to_char(sum(decode(substr(time,10,2),'16',1,0)),'99') "16",
    19  to_char(sum(decode(substr(time,10,2),'17',1,0)),'99') "17",
    20  to_char(sum(decode(substr(time,10,2),'18',1,0)),'99') "18",
    21  to_char(sum(decode(substr(time,10,2),'19',1,0)),'99') "19",
    22  to_char(sum(decode(substr(time,10,2),'20',1,0)),'99') "20",
    23  to_char(sum(decode(substr(time,10,2),'21',1,0)),'99') "21",
    24  to_char(sum(decode(substr(time,10,2),'22',1,0)),'99') "22",
    25  to_char(sum(decode(substr(time,10,2),'23',1,0)),'99') "23"
    26  from v$log_history
    27  group by substr(time,1,5);
    group by substr(time,1,5)
    ERROR at line 27:
    ORA-00904: invalid column name
    SQL> desc v$log_history
    Name                                      Null?    Type
    RECID                                              NUMBER
    STAMP                                              NUMBER
    THREAD#                                            NUMBER
    SEQUENCE#                                          NUMBER
    FIRST_CHANGE#                                      NUMBER
    FIRST_TIME                                         DATE
    NEXT_CHANGE#                                       NUMBER
    SQL>
    Connected to:
    Oracle9i Enterprise Edition Release 9.2.0.8.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.8.0 - Production
    SQL> select substr(time,1,5) day,
      2  to_char(sum(decode(substr(time,10,2),'00',1,0)),'99') "00",
      3  to_char(sum(decode(substr(time,10,2),'01',1,0)),'99') "01",
      4  to_char(sum(decode(substr(time,10,2),'02',1,0)),'99') "02",
      5  to_char(sum(decode(substr(time,10,2),'03',1,0)),'99') "03",
      6  to_char(sum(decode(substr(time,10,2),'04',1,0)),'99') "04",
      7  to_char(sum(decode(substr(time,10,2),'05',1,0)),'99') "05",
      8  to_char(sum(decode(substr(time,10,2),'06',1,0)),'99') "06",
      9  to_char(sum(decode(substr(time,10,2),'07',1,0)),'99') "07",
    10  to_char(sum(decode(substr(time,10,2),'08',1,0)),'99') "08",
    11  to_char(sum(decode(substr(time,10,2),'09',1,0)),'99') "09",
    12  to_char(sum(decode(substr(time,10,2),'10',1,0)),'99') "10",
    13  to_char(sum(decode(substr(time,10,2),'11',1,0)),'99') "11",
    14  to_char(sum(decode(substr(time,10,2),'12',1,0)),'99') "12",
    15  to_char(sum(decode(substr(time,10,2),'13',1,0)),'99') "13",
    16  to_char(sum(decode(substr(time,10,2),'14',1,0)),'99') "14",
    17  to_char(sum(decode(substr(time,10,2),'15',1,0)),'99') "15",
    18  to_char(sum(decode(substr(time,10,2),'16',1,0)),'99') "16",
    19  to_char(sum(decode(substr(time,10,2),'17',1,0)),'99') "17",
    20  to_char(sum(decode(substr(time,10,2),'18',1,0)),'99') "18",
    21  to_char(sum(decode(substr(time,10,2),'19',1,0)),'99') "19",
    22  to_char(sum(decode(substr(time,10,2),'20',1,0)),'99') "20",
    23  to_char(sum(decode(substr(time,10,2),'21',1,0)),'99') "21",
    24  to_char(sum(decode(substr(time,10,2),'22',1,0)),'99') "22",
    25  to_char(sum(decode(substr(time,10,2),'23',1,0)),'99') "23"
    26  from v$log_history
    27  group by substr(time,1,5);
    group by substr(time,1,5)
    ERROR at line 27:
    ORA-00904: "TIME": invalid identifier
    SQL> desc v$log_history
    Name                                      Null?    Type
    RECID                                              NUMBER
    STAMP                                              NUMBER
    THREAD#                                            NUMBER
    SEQUENCE#                                          NUMBER
    FIRST_CHANGE#                                      NUMBER
    FIRST_TIME                                         DATE
    NEXT_CHANGE#                                       NUMBER
    SQL> Well, Gints has already check the doc.
    To be fair, I would agree that could be a nightmare to maintain a bunch of scripts through different release, but at least see the evidence, on that case, on the age of the scritp if it was tested a day.
    Nicolas.
    Message was edited by:
    N. Gasparotto

  • Create procedure is generating too many archive logs

    Hi
    The following procedure was run on one of our databases and it hung since there were too many archive logs being generated.
    What would be the answer? The db must remain in archivelog mode.
    I understand the nologging concept, but as I know this applies to creating tables, views, indexes and tablespaces. This script is creating procedure.
    CREATE OR REPLACE PROCEDURE APPS.Dfc_Payroll_Dw_Prc(Errbuf OUT VARCHAR2, Retcode OUT NUMBER
    ,P_GRE NUMBER
    ,P_SDATE VARCHAR2
    ,P_EDATE VARCHAR2
    ,P_ssn VARCHAR2
    ) IS
    CURSOR MainCsr IS
    SELECT DISTINCT
    PPF.NATIONAL_IDENTIFIER SSN
    ,ppf.full_name FULL_NAME
    ,ppa.effective_date Pay_date
    ,ppa.DATE_EARNED period_end
    ,pet.ELEMENT_NAME
    ,SUM(TO_NUMBER(prv.result_value)) VALOR
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAf.ASSIGNMENT_ID ASSG_ID
    ,paf.ORGANIZATION_ID
    FROM
    pay_element_classifications pec
    , pay_element_types_f pet
    , pay_input_values_f piv
    , pay_run_result_values prv
    , pay_run_results prr
    , pay_assignment_actions paa
    , pay_payroll_actions ppa
    , APPS.pay_all_payrolls_f pap
    ,Per_Assignments_f paf
    ,per_people_f ppf
    WHERE
    ppa.effective_date BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND ppa.payroll_id = pap.payroll_id
    AND paa.tax_unit_id = NVL(p_GRE, paa.tax_unit_id)
    AND ppa.payroll_action_id = paa.payroll_action_id
    AND paa.action_status = 'C'
    AND ppa.action_type IN ('Q', 'R', 'V', 'B', 'I')
    AND ppa.action_status = 'C'
    --AND PEC.CLASSIFICATION_NAME IN ('Earnings','Alien/Expat Earnings','Supplemental Earnings','Imputed Earnings','Non-payroll Payments')
    AND paa.assignment_action_id = prr.assignment_action_id
    AND prr.run_result_id = prv.run_result_id
    AND prv.input_value_id = piv.input_value_id
    AND piv.name = 'Pay Value'
    AND piv.element_type_id = pet.element_type_id
    AND pet.element_type_id = prr.element_type_id
    AND pet.classification_id = pec.classification_id
    AND pec.non_payments_flag = 'N'
    AND prv.result_value &lt;&gt; '0'
    --AND( PET.ELEMENT_INFORMATION_CATEGORY LIKE '%EARNINGS'
    -- OR PET.element_type_id IN (1425, 1428, 1438, 1441, 1444, 1443) )
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PET.EFFECTIVE_START_DATE AND PET.EFFECTIVE_END_DATE
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PIV.EFFECTIVE_START_DATE AND PIV.EFFECTIVE_END_DATE --dcc
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN Pap.EFFECTIVE_START_DATE AND Pap.EFFECTIVE_END_DATE --dcc
    AND paf.ASSIGNMENT_ID = paa.ASSIGNMENT_ID
    AND ppf.NATIONAL_IDENTIFIER = NVL(p_ssn, ppf.NATIONAL_IDENTIFIER)
    ------------------------------------------------------------------TO get emp.
    AND ppf.person_id = paf.person_id
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN ppf.EFFECTIVE_START_DATE AND ppf.EFFECTIVE_END_DATE
    ------------------------------------------------------------------TO get emp. ASSIGNMENT
    --AND paf.assignment_status_type_id NOT IN (7,3)
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN paf.effective_start_date AND paf.effective_end_date
    GROUP BY PPF.NATIONAL_IDENTIFIER
    ,ppf.full_name
    ,ppa.effective_date
    ,ppa.DATE_EARNED
    ,pet.ELEMENT_NAME
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAF.ASSIGNMENT_ID
    ,paf.ORGANIZATION_ID
    BEGIN
    DELETE cust.DFC_PAYROLL_DW
    WHERE PAY_DATE BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND tax_unit_id = NVL(p_GRE, tax_unit_id)
    AND ssn = NVL(p_ssn, ssn)
    COMMIT;
    FOR V_REC IN MainCsr LOOP
    INSERT INTO cust.DFC_PAYROLL_DW(SSN, FULL_NAME, PAY_DATE, PERIOD_END, ELEMENT_NAME, ELEMENT_INFORMATION_CATEGORY, CLASSIFICATION_ID, ELEMENT_INFORMATION1, VALOR, TAX_UNIT_ID, ASSG_ID,ELEMENT_TYPE_ID,ORGANIZATION_ID)
    VALUES(V_REC.SSN,V_REC.FULL_NAME,v_rec.PAY_DATE,V_REC.PERIOD_END,V_REC.ELEMENT_NAME,V_REC.ELEMENT_INFORMATION_CATEGORY, V_REC.CLASSIFICATION_ID, V_REC.ELEMENT_INFORMATION1, V_REC.VALOR,V_REC.TAX_UNIT_ID,V_REC.ASSG_ID, v_rec.ELEMENT_TYPE_ID, v_rec.ORGANIZATION_ID);
    COMMIT;
    END LOOP;
    END ;
    So, how could I assist our developer with this, so that she can run it again without it generating a ton of logs ? ?
    Thanks
    Oracle 9.2.0.5
    AIX 5.2

    The amount of redo generated is a direct function of how much data is changing. If you insert 'x' number of rows, you are going to generate 'y' mbytes of redo. If your procedure is destined to insert 1000 rows, then it is destined to create a certain amount of redo. Period.
    I would question the <i>performance</i> of the procedure shown ... using a cursor loop with a commit after every row is going to be a slug on performance but that doesn't change the fact 'x' inserts will always generate 'y' redo.

  • I have one problem with Data Guard. My archive log files are not applied.

    I have one problem with Data Guard. My archive log files are not applied. However I have received all archive log files to my physical Standby db
    I have created a Physical Standby database on Oracle 10gR2 (Windows XP professional). Primary database is on another computer.
    In Enterprise Manager on Primary database it looks ok. I get the following message “Data Guard status Normal”
    But as I wrote above ”the archive log files are not applied”
    After I created the Physical Standby database, I have also done:
    1. I connected to the Physical Standby database instance.
    CONNECT SYS/SYS@luda AS SYSDBA
    2. I started the Oracle instance at the Physical Standby database without mounting the database.
    STARTUP NOMOUNT PFILE=C:\oracle\product\10.2.0\db_1\database\initluda.ora
    3. I mounted the Physical Standby database:
    ALTER DATABASE MOUNT STANDBY DATABASE
    4. I started redo apply on Physical Standby database
    alter database recover managed standby database disconnect from session
    5. I switched the log files on Physical Standby database
    alter system switch logfile
    6. I verified the redo data was received and archived on Physical Standby database
    select sequence#, first_time, next_time from v$archived_log order by sequence#
    SEQUENCE# FIRST_TIME NEXT_TIME
    3 2006-06-27 2006-06-27
    4 2006-06-27 2006-06-27
    5 2006-06-27 2006-06-27
    6 2006-06-27 2006-06-27
    7 2006-06-27 2006-06-27
    8 2006-06-27 2006-06-27
    7. I verified the archived redo log files were applied on Physical Standby database
    select sequence#,applied from v$archived_log;
    SEQUENCE# APP
    4 NO
    3 NO
    5 NO
    6 NO
    7 NO
    8 NO
    8. on Physical Standby database
    select * from v$archive_gap;
    No rows
    9. on Physical Standby database
    SELECT MESSAGE FROM V$DATAGUARD_STATUS;
    MESSAGE
    ARC0: Archival started
    ARC1: Archival started
    ARC2: Archival started
    ARC3: Archival started
    ARC4: Archival started
    ARC5: Archival started
    ARC6: Archival started
    ARC7: Archival started
    ARC8: Archival started
    ARC9: Archival started
    ARCa: Archival started
    ARCb: Archival started
    ARCc: Archival started
    ARCd: Archival started
    ARCe: Archival started
    ARCf: Archival started
    ARCg: Archival started
    ARCh: Archival started
    ARCi: Archival started
    ARCj: Archival started
    ARCk: Archival started
    ARCl: Archival started
    ARCm: Archival started
    ARCn: Archival started
    ARCo: Archival started
    ARCp: Archival started
    ARCq: Archival started
    ARCr: Archival started
    ARCs: Archival started
    ARCt: Archival started
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    ARC1: Becoming the heartbeat ARCH
    Attempt to start background Managed Standby Recovery process
    MRP0: Background Managed Standby Recovery process started
    Managed Standby Recovery not using Real Time Apply
    MRP0: Background Media Recovery terminated with error 1110
    MRP0: Background Media Recovery process shutdown
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[1]: Assigned to RFS process 2148
    RFS[1]: Identified database type as 'physical standby'
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[2]: Assigned to RFS process 2384
    RFS[2]: Identified database type as 'physical standby'
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[3]: Assigned to RFS process 3188
    RFS[3]: Identified database type as 'physical standby'
    Primary database is in MAXIMUM PERFORMANCE mode
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: No standby redo logfiles created
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    RFS[4]: Assigned to RFS process 3168
    RFS[4]: Identified database type as 'physical standby'
    RFS[4]: No standby redo logfiles created
    Primary database is in MAXIMUM PERFORMANCE mode
    RFS[3]: No standby redo logfiles created
    10. on Physical Standby database
    SELECT PROCESS, STATUS, THREAD#, SEQUENCE#, BLOCK#, BLOCKS FROM V$MANAGED_STANDBY;
    PROCESS STATUS THREAD# SEQUENCE# BLOCK# BLOCKS
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    ARCH CONNECTED 0 0 0 0
    RFS IDLE 0 0 0 0
    RFS IDLE 0 0 0 0
    RFS IDLE 1 9 13664 2
    RFS IDLE 0 0 0 0
    10) on Primary database:
    select message from v$dataguard_status;
    MESSAGE
    ARC0: Archival started
    ARC1: Archival started
    ARC2: Archival started
    ARC3: Archival started
    ARC4: Archival started
    ARC5: Archival started
    ARC6: Archival started
    ARC7: Archival started
    ARC8: Archival started
    ARC9: Archival started
    ARCa: Archival started
    ARCb: Archival started
    ARCc: Archival started
    ARCd: Archival started
    ARCe: Archival started
    ARCf: Archival started
    ARCg: Archival started
    ARCh: Archival started
    ARCi: Archival started
    ARCj: Archival started
    ARCk: Archival started
    ARCl: Archival started
    ARCm: Archival started
    ARCn: Archival started
    ARCo: Archival started
    ARCp: Archival started
    ARCq: Archival started
    ARCr: Archival started
    ARCs: Archival started
    ARCt: Archival started
    ARCm: Becoming the 'no FAL' ARCH
    ARCm: Becoming the 'no SRL' ARCH
    ARCd: Becoming the heartbeat ARCH
    Error 1034 received logging on to the standby
    Error 1034 received logging on to the standby
    LGWR: Error 1034 creating archivelog file 'luda'
    LNS: Failed to archive log 3 thread 1 sequence 7 (1034)
    FAL[server, ARCh]: Error 1034 creating remote archivelog file 'luda'
    11)on primary db
    select name,sequence#,applied from v$archived_log;
    NAME SEQUENCE# APP
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00003_0594204176.001 3 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00004_0594204176.001 4 NO
    Luda 4 NO
    Luda 3 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00005_0594204176.001 5 NO
    Luda 5 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00006_0594204176.001 6 NO
    Luda 6 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00007_0594204176.001 7 NO
    Luda 7 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\IRINA\ARC00008_0594204176.001 8 NO
    Luda 8 NO
    12) on standby db
    select name,sequence#,applied from v$archived_log;
    NAME SEQUENCE# APP
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00004_0594204176.001 4 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00003_0594204176.001 3 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00005_0594204176.001 5 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00006_0594204176.001 6 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00007_0594204176.001 7 NO
    C:\ORACLE\PRODUCT\10.2.0\ORADATA\LUDA\ARC00008_0594204176.001 8 NO
    13) my init.ora files
    On standby db
    irina.__db_cache_size=79691776
    irina.__java_pool_size=4194304
    irina.__large_pool_size=4194304
    irina.__shared_pool_size=75497472
    irina.__streams_pool_size=0
    *.audit_file_dest='C:\oracle\product\10.2.0\admin\luda\adump'
    *.background_dump_dest='C:\oracle\product\10.2.0\admin\luda\bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='C:\oracle\product\10.2.0\oradata\luda\luda.ctl'
    *.core_dump_dest='C:\oracle\product\10.2.0\admin\luda\cdump'
    *.db_block_size=8192
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_file_name_convert='luda','irina'
    *.db_name='irina'
    *.db_unique_name='luda'
    *.db_recovery_file_dest='C:\oracle\product\10.2.0\flash_recovery_area'
    *.db_recovery_file_dest_size=2147483648
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=irinaXDB)'
    *.fal_client='luda'
    *.fal_server='irina'
    *.job_queue_processes=10
    *.log_archive_config='DG_CONFIG=(irina,luda)'
    *.log_archive_dest_1='LOCATION=C:/oracle/product/10.2.0/oradata/luda/ VALID_FOR=(ALL_LOGFILES, ALL_ROLES) DB_UNIQUE_NAME=luda'
    *.log_archive_dest_2='SERVICE=irina LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=irina'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_max_processes=30
    *.log_file_name_convert='C:/oracle/product/10.2.0/oradata/irina/','C:/oracle/product/10.2.0/oradata/luda/'
    *.open_cursors=300
    *.pga_aggregate_target=16777216
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=167772160
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='C:\oracle\product\10.2.0\admin\luda\udump'
    On primary db
    irina.__db_cache_size=79691776
    irina.__java_pool_size=4194304
    irina.__large_pool_size=4194304
    irina.__shared_pool_size=75497472
    irina.__streams_pool_size=0
    *.audit_file_dest='C:\oracle\product\10.2.0/admin/irina/adump'
    *.background_dump_dest='C:\oracle\product\10.2.0/admin/irina/bdump'
    *.compatible='10.2.0.1.0'
    *.control_files='C:\oracle\product\10.2.0\oradata\irina\control01.ctl','C:\oracle\product\10.2.0\oradata\irina\control02.ctl','C:\oracle\product\10.2.0\oradata\irina\control03.ctl'
    *.core_dump_dest='C:\oracle\product\10.2.0/admin/irina/cdump'
    *.db_block_size=8192
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_file_name_convert='luda','irina'
    *.db_name='irina'
    *.db_recovery_file_dest='C:\oracle\product\10.2.0/flash_recovery_area'
    *.db_recovery_file_dest_size=2147483648
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=irinaXDB)'
    *.fal_client='irina'
    *.fal_server='luda'
    *.job_queue_processes=10
    *.log_archive_config='DG_CONFIG=(irina,luda)'
    *.log_archive_dest_1='LOCATION=C:/oracle/product/10.2.0/oradata/irina/ VALID_FOR=(ALL_LOGFILES, ALL_ROLES) DB_UNIQUE_NAME=irina'
    *.log_archive_dest_2='SERVICE=luda LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=luda'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_max_processes=30
    *.log_file_name_convert='C:/oracle/product/10.2.0/oradata/luda/','C:/oracle/product/10.2.0/oradata/irina/'
    *.open_cursors=300
    *.pga_aggregate_target=16777216
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=167772160
    *.standby_file_management='AUTO'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='C:\oracle\product\10.2.0/admin/irina/udump'
    Please help me!!!!

    Hi,
    After several tries my redo logs are applied now. I think in my case it had to do with the tnsnames.ora. At this moment I have both database in both tnsnames.ora files using the SID and not the SERVICE_NAME.
    Now I want to use DGMGRL. Adding a configuration and a stand-by database is working fine, but when I try to enable the configuration DGMGRL gives no feedback and it looks like it is hanging. The log, although says that it succeeded.
    In another session 'show configuration' results in the following, confirming that the enable succeeded.
    DGMGRL> show configuration
    Configuration
    Name: avhtest
    Enabled: YES
    Protection Mode: MaxPerformance
    Fast-Start Failover: DISABLED
    Databases:
    avhtest - Primary database
    avhtestls53 - Physical standby database
    Current status for "avhtest":
    Warning: ORA-16610: command 'ENABLE CONFIGURATION' in progress
    It there anybody that experienced the same problem and/or knows the solution to this?
    With kind regards,
    Martin Schaap

  • How to find out which archived logs needed to recover a hot backup?

    I'm using Oracle 11gR2 (11.2.0.1.0).
    I have backed up a database when it is online using the following backup script through RMAN
    connect target /
    run {
    allocate channel d1 type disk;
    backup
    incremental level=0 cumulative
    filesperset 4
    format '/san/u01/app/backup/DB_%d_%T_%u_%c.rman'
    database
    }The backup set contains the backup of datafiles and control file. I have copied all the backup pieces to another server where I will restore/recover the database but I don't know which archived logs are needed in order to restore/recover the database to a consistent state.
    I have not deleted any archived log.
    How can I find out which archived logs are needed to recover the hot backup to a consistent state? Can this be done by querying V$BACKUP_DATAFILE and V$ARCHIVED_LOG? If yes, which columns should I query?
    Thanks for any help.

    A few ways :
    1a. Get the timestamps when the BACKUP ... DATABASE began and ended.
    1b. Review the alert.log of the database that was backed up.
    1c. From the alert.log identify the first Archivelog that was generated after the begin of the BACKUP ... DATABASE and the first Archivelog that was generated after the end of the BACKUP .. DATABASE.
    1d. These (from 1c) are the minimal Archivelogs that you need to RECOVER with. You can choose to apply additional Archivelogs that were generated at the source database to contininue to "roll-forward"
    2a. Do a RESTORE DATABASE alone.
    2b. Query V$DATAFILE on the restored database for the lowest CHECKPOINT_CHANGE# and CHECKPOINT_TIME. Also query for the highest CHECKPOINT_CHANGE# and CHECKPOINT_TIME.
    2c. Go back to the source database and query V$ARCHIVED_LOG (FIRST_CHANGE#) to identify the first Archivelog that has a higher SCN (FIRST_CHANGE#) than the lowest CHECKPOINT_CHANGE# from 2b above. Also query for the first Archivelog that has a higher SCN (FIRST_CHANGE#) than the highest CHECKPOINT_CHANGE# from 2b above.
    2d. These (from 2c) are the minimal Archivelogs that you need to RECOVER with.
    (why do you need to query V$ARCHIVED_LOG at the source ? If RESTORE a controlfile backup that was generated after the first Archivelog switch after the end of the BACKUP ... DATABASE, you would be able to query V$ARCHIVED_LOG at the restored database as well. That is why it is important to force an archivelog (log switch) after a BACKUP ... DATABASE and then backup the controlfile after this -- i.e. last. That way, the controlfile that you have restored to the new server has all the information needed).
    3. RESTORE DATABASE PREVIEW in RMAN if you have the archivelogs and subsequent controlfile in the backup itself !
    Hemant K Chitale

  • How to find a session with high archive logs

    Any query, to see which active session is generating high archive logs in oracle 8i &9i and high rbs usage

    Though, there is no direct option or view where you can get this information.
    However, you may can find out the session which are generating lot of redo and undo. I guess, when the session has lot of redo & undo, definately, it contribution towards archive would be more.
    You can query v$sess_io and v$session to findout the the session which is generating lot of redo. i.e. lot of block changes occuring.
    SELECT s.sid, s.serial#, s.username, s.program, i.block_changes
    FROM v$session s, v$sess_io i
    WHERE s.sid = i.sid
    ORDER BY 5 desc
    Also query v$transaction and v$session to find out the session that is generating lot of undo information.
    Jaffar
    Message was edited by:
    Syed Jaffar

  • Archive logs not getting shipped when we do a swtich

    Hi Team.
    I am facing an strange issue on one of our standby.
    After the setup of the standby when we enabled the MrM mode, the archive got shipped smoothly.After after sometime when we tried to switch log file and check if it was working or not.
    From the alert log of the DR , we can only see these msg " Media Recovery Waiting for thread 1 sequence ***". But the moment we cancel the MrM mode and re enable it , its ships the same smooth.
    So need some guidance to debug the same .
    Please advise .
    Thanks

    Hello;
    On something like this I would check BOTH the Primary and Standby alert logs.
    Here's a SWITCH I forced yesterday:
    Thu Sep 12 16:01:11 2013
    ALTER SYSTEM ARCHIVE LOG
    Thu Sep 12 16:01:11 2013
    Thread 1 advanced to log sequence 811 (LGWR switch)
      Current log# 1 seq# 811 mem# 0: /u01/app/oracle/oradata/PRIMARY/redo01.log
    Thu Sep 12 16:01:11 2013
    LNS: Standby redo logfile selected for thread 1 sequence 811 for destination LOG_ARCHIVE_DEST_2
    Notice the last line is logged on the Primary almost the very moment I do the switch. Does your Database in  Primary mode show that?
    Best Regards
    mseberg

  • Recovery question using archive log

    HPUX 11X
    Oracle 10.2.0.2
    I have a standby database that the customers want to test the ability to fail over. They also want to test the data integrity of the standby once it's brought on line as the primary.
    I've already told the customer that any writes that they do on the standby during the test time WILL be lost. I plan on doing a incomplete recovery back to a specific sequence number / archive log when the tests are complete.
    My plan is to take note of the current sequence number just before I bring the standby live. When all is said and done, I was going to do a recover back to the sequence number one PRIOR to that one. Then when I place the db back in standby mode, it will sync back up to the primary.
    What I need to know is if this is a good or a bad idea. Should I plan on doing it differently and for what reason.
    Thanks! :)

    Use flashback database to recover the former primary database to a new standby.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/scenarios.htm#i1049997
    Therefore the changes made on standby won't be lost (unless you want to lose them).
    If your databases do not have Flashback Database feature enabled, you can use incomplete recovery instead.
    I think Flashback Database should be your first option.

  • ORA-00339: archived log does not contain any redo

    Hi All,
    recently we faced 'ORA-00339: archived log does not contain any redo' issue at standby side,
    after searching on Google and from Metalink (note 30866.1 and 7197445.8 ) I find out that this is the known issue for 10g and below versions, our's is 11.2.0.3,
    Error in Alert Log :
    Errors in file /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/trace/DWPRD_pr0a_48412.trc:
    ORA-00339: archived log does not contain any redo
    ORA-00334: archived log: '/redolog2/redo/redolog3a.log'
    Errors in file /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/trace/DWPRD_pr0a_48412.trc (incident=190009):
    ORA-00600: internal error code, arguments: [kdBlkCheckError], [1], [56702], [6114], [], [], [], [], [], [], [], []
    Incident details in: /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/incident/incdir_190009/DWPRD_pr0a_48412_i190009.trc
    Use ADRCI or Support Workbench to package the incident.
    See Note 411.1 at My Oracle Support for error and packaging details.
    Slave exiting with ORA-10562 exception
    Errors in file /oracle/ora_home/diag/diag/rdbms/dwprd/DWPRD/trace/DWPRD_pr0a_48412.trc:
    ORA-10562: Error occurred while applying redo to data block (file# 1, block# 56702)
    ORA-10564: tablespace SYSTEM
    ORA-01110: data file 1: '/oradata1/database/DATAFILES/system01.dbf'
    ORA-10561: block type 'TRANSACTION MANAGED DATA BLOCK', data object# 2
    ORA-00600: internal error code, arguments: [kdBlkCheckError], [1], [56702], [6114], [], [], [], [], [], [], [], []
    Mon Apr 15 11:34:12 2013
    Dumping diagnostic data in directory=[cdmp_20130415113412], requested by (instance=1, osid=48412 (PR0A)), summary=[incident=190009].
    Thanks

    Hi,
    "The archived log is not the correct log.
    It is a copy of a log file that has never been used for redo generation, or was an online log being prepared to be the current log."
    "Restore the correct log file."
    Can you say, what is last changes on your database, On log files?
    Did you copies your '/redolog2/redo/redolog3a.log' log file from other ?
    Regards
    Mahir M. Quluzade

  • How to recover the database when some of the archive log file get deleted.

    I am facing a problem with Oracle database, which is related to archivelogs.
    Our development database is running in archivelog mode, but we don't have backups scheduled and have no recovery catalog.
    When the database was in running condition, disk got full, so some archivelogs were deleted manually.
    After this they restarted the DB, and now DB is not coming up. Errors are as follows:
    SQL> startup
    ORACLE instance started.
    Total System Global Area 1444383504 bytes
    Fixed Size 731920 bytes
    Variable Size 486539264 bytes
    Database Buffers 956301312 bytes
    Redo Buffers 811008 bytes
    Database mounted.
    ORA-01589: must use RESETLOGS or NORESETLOGS option for database open
    SQL> alter database open resetlogs;
    alter database open resetlogs
    ERROR at line 1:
    ORA-01113: file 1 needs media recovery
    ORA-01110: data file 1: '/export/home/oracle/dev/ADVFRW/ADVFRW.system'
    SQL> recover datafile '/export/home/oracle/dev/ADVFRW/ADVFRW.system'
    ORA-00283: recovery session canceled due to errors
    ORA-01610: recovery using the BACKUP CONTROLFILE option must be done
    SQL> recover database using backup controlfile;
    ORA-00279: change 215548705 generated at 09/02/2008 17:06:10 needed for thread
    1
    ORA-00289: suggestion :
    /export/home/oracle/dev/ADVFRW/ADVFRW.archivelog1/LOG_ADVFRW_1107_1.ARC
    ORA-00280: change 215548705 for thread 1 is in sequence #1107
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    /export/home/oracle/dev/ADVFRW/ADVFRW.archivelog1/LOG_ADVFRW_1107_1.ARC
    ORA-00308: cannot open archived log
    '/export/home/oracle/dev/ADVFRW/ADVFRW.archivelog1/LOG_ADVFRW_1107_1.ARC'
    ORA-27037: unable to obtain file status
    SVR4 Error: 2: No such file or directory
    Additional information: 3
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    CANCEL
    Media recovery cancelled.
    SQL>
    1. How to recover the database and bring it online
    Any help will be highly appreciated.
    With Regards
    Hemant Joshi
    Edited by: hem_Kec on Sep 7, 2008 9:07 AM

    Hi,
    Archive log files are the copies of redolog files.As redo log files are circularly overwritten,oracle generates archive log file of the corresponding redo logfiles being overwritten.So if you have a backup that dates back to 10 am in the morning and if your database creashed at 3 pm,you cannot use the redo log files alone as they have incomplete information.To completely recover the database upto 3 pm,you need archive log files generated between 10 am to 3 pm. In your case since you are missing one archive log file,you cannot perform complete recovery and hence would suffer data loss.

  • Question about Archive Log Deletion policy

    I've a problem to understand the Archive Log Deletion policy, and I I'd like to this problem explain with the following example.
    Messages of the database are in German, but I guess you'll understand them.
    SQL> startup
    ORACLE-Instance hochgefahren.
    Total System Global Area 5344731136 bytes
    Fixed Size                  2129240 bytes
    Variable Size            2684355240 bytes
    Database Buffers         2617245696 bytes
    Redo Buffers               41000960 bytes
    Datenbank mounted.
    Datenbank geöffnet.
    SQL> archive log list
    Datenbank-Log-Modus              Archive-Modus
    Automatische Archivierung             Aktiviert
    Archivierungsziel            E:\oracle\thetis_iv\arch
    Älteste Online-Log-Sequenz     17917
    Nächste zu archivierende Log-Sequenz   17919
    Aktuelle Log-Sequenz           17919
    SQL> alter system switch logfile;
    System wurde geändert.I created a brand new archive log.
    SQL> exit
    Verbindung zu Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options beendet
    D:\OracleDB\product\11.1.0\db_1\BIN>dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:04    <DIR>          .
    06.04.2011  15:04    <DIR>          ..
    06.04.2011  15:04        17.137.152 ARC17919_0721667907.001
                   1 Datei(en),     17.137.152 Bytes
                   2 Verzeichnis(se), 41.073.258.496 Bytes freiand this is the only archive log in the directory. Now I start rman:
    D:\OracleDB\product\11.1.0\db_1\BIN>rman target / catalog rmanrepo@rmanrepo
    Recovery Manager: Release 11.1.0.7.0 - Production on Mi Apr 6 15:05:35 2011
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Mit Ziel-Datenbank verbunden: ENTWIV (DBID=21045568)
    Kennwort für Recovery-Katalog-Datenbank:
    Verbindung mit Datenbank des Recovery-Katalogs
    RMAN> show all;
    RMAN-Konfigurationsparameter für Datenbank mit db_unique_name ENTWIV sind:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO 'SBT_TAPE';
    CONFIGURE CONTROLFILE AUTOBACKUP OFF;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO 'E:\oracle\thetis_iv\backup\CF_%F_ENTWIV.ORA';
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE SBT_TAPE TO '%F'; # default
    CONFIGURE DEVICE TYPE SBT_TAPE PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE 'SBT_TAPE' PARMS  'ENV=(TPDO_OPTFILE=D:\Services\Tivoli\TSM\AgentOBA64\tpdo.opt)';
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE COMPRESSION ALGORITHM 'BZIP2'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO 'SBT_TAPE';
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO 'D:\ORACLEDB\PRODUCT\11.1.0\DB_1\DATABASE\SNCFENTWIV.ORA'; # defaultThe archive log deletion policy says the the logfiles have to be backed up for two times before they get deleted.
    Now I backup all archive logs, that havn't been backed up for at least two times.
    RMAN> run { backup archivelog all not backed up 2 times
    2>       format '%d_AR_%Y%M%D_%s_%t'
    3>       tag 'ARCHIVE LOGS'
    4>       DELETE ALL INPUT;
    5>     }
    Starten backup um 06.04.2011 15:08:01
    Aktuelles Log archiviert
    Zugewiesener Kanal: ORA_SBT_TAPE_1
    Kanal ORA_SBT_TAPE_1: SID=253 Device-Typ=SBT_TAPE
    Kanal ORA_SBT_TAPE_1: Data Protection for Oracle: version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set für Archive Log wird begonnen
    Kanal ORA_SBT_TAPE_1: Archive Logs in Backup Set werden angegeben
    Eingabe-Archive-Log-Thread=1 Sequence=17919 RECID=147 STAMP=747759899
    Eingabe-Archive-Log-Thread=1 Sequence=17920 RECID=148 STAMP=747760081
    Kanal ORA_SBT_TAPE_1: Piece 1 wird auf 06.04.2011 15:08:02 begonnen
    Kanal ORA_SBT_TAPE_1: Piece 1 auf 06.04.2011 15:08:09 beendet
    Piece Handle=ENTWIV_AR_20110406_23_747760082 Tag=ARCHIVE LOGS Kommentar=API Version 2.0,MMS Version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set vollstõndig, abgelaufene Zeit: 00:00:08
    Kanal ORA_SBT_TAPE_1: Archive Logs werden gel÷scht
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17919_0721667907.001 RECID=147 STAMP=747759899
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17920_0721667907.001 RECID=148 STAMP=747760081
    Beendet backup um 06.04.2011 15:08:10
    RMAN> exit
    Recovery Manager abgeschlossen.
    D:\OracleDB\product\11.1.0\db_1\BIN> dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:08    <DIR>          .
    06.04.2011  15:08    <DIR>          ..
                   0 Datei(en),              0 Bytes
                   2 Verzeichnis(se), 41.090.396.160 Bytes freirman deleted all archive logs, even I they are on tape only once by now.
    Thats not what I expected. Where is my mistake?

    Hi,
    I do new tests it's very strange.
    BACKUP ARCHIVELOG command is not obeying the policy of archivelog.
    You can open a SR on MOS. (to check bugs)
    I reproduce the same test and the result was the same, it seems that this is a bug.
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO DISK;
    RMAN> backup archivelog all not backed up 2 times delete all input;
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=1 sequence=15 RECID=16 STAMP=747753711
    input archived log thread=2 sequence=20 RECID=17 STAMP=747753714
    input archived log thread=1 sequence=16 RECID=19 STAMP=747753729
    input archived log thread=2 sequence=21 RECID=18 STAMP=747753729
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t132210_0.304.747753731 tag=TAG20110406T132210 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    channel ORA_DISK_1: deleting archived log(s)
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_15.293.747753711 RECID=16 STAMP=747753711
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_20.295.747753715 RECID=17 STAMP=747753714
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_16.294.747753729 RECID=19 STAMP=747753729
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_21.298.747753729 RECID=18 STAMP=747753729
    Finished backup at 06-APR-11
    RMAN> list archivelog all;
    specification does not match any archived log in the repositoryOracle Docs Says:
    The BACKUP ARCHIVELOG ... DELETE INPUT command deletes archived log files after they are backed up.
    This command eliminates the separate step of manually deleting archived redo logs.
    With DELETE INPUT, RMAN deletes only the specific copy of the archived log chosen for the backup set.
    With DELETE ALL INPUT, RMAN deletes each backed-up archived redo log file from all log archiving destinations.
    As explained in "Configuring an Archived Redo Log Deletion Policy",
    the BACKUP ... DELETE INPUT and DELETE ARCHIVELOG commands obey the archived redo log deletion policy
    for logs in all archiving locations. For example, if you specify that logs should only be deleted when backed
    up at least twice to tape, then BACKUP ... DELETE honors this policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10642/rcmbckba.htm#BRADV89524
    But in ours case it's not honors this policy.
    Only with the FORCE command should this happen. But it is not our case.
    Oracle Docs:
    If FORCE is not specified on the deletion commands,
    then these deletion commands obey the archived log deletion policy.
    If FORCE is specified, then the deletion commands ignore the archived log deletion policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10643/rcmsynta010.htm#RCMRF113
    Alternatively you can do the following:
    Set the commands separately.
    Check this:
    RMAN>  run {
    2> backup archivelog all not backed up 2 times ;
    3> delete archivelog all backed up 2 times to disk;
    4> }
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=2 sequence=22 RECID=21 STAMP=747755128
    input archived log thread=1 sequence=17 RECID=20 STAMP=747755127
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t134528_0.295.747755129 tag=TAG20110406T134528 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    Finished backup at 06-APR-11
    released channel: ORA_DISK_1
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=78 instance=orcl1 device type=DISK
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_17.298.747755127 thread=1 sequence=17
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_22.294.747755129 thread=2 sequence=22
    RMAN>Edited by: Levi Pereira on Apr 6, 2011 1:35 PM

  • Duplexing the Archive Log - is there a potential performance hit

    Good Afternoon Oracle People -
    I apologize if this question is silly or out of place.
    Basically, we are looking at options for implementing a (cheap) DR solution for our Oracle Database.
    Bottom line objective is to have a second copy of our production system (not running, offline) with a usable archive log to recover from at a remote site with a similar set of disk technologies etc. 
    The sites are linked by 100MBPs link - so any access to secondary site is relatively fast.
    What I was thinking of doing is creating an iSCSI target on the destination SAN and adding this as a disk into the production database.  Then, I was going to go in and define a duplex archive log destination to this iSCSI target.
    My fear is that if the archiver waits for the write to the destination, I would imagine that this would slow down database access.  Is this a valid concern?  Or does Oracle treat the Duplex as a replicated (slower) copy?
    Again, sorry if this question is stupid - i have tried the google machine but couldn't find anything, and it isn't all that clear from the documentation. 
    Kind Regards,
    Aleksei

    Or does Oracle treat the Duplex as a replicated (slower) copy?
    Oracle treats it in a way you tell Oracle to treat it – please take a look at LOG_ARCHIVE_DEST_n parameters such as MANDATORY and OPTIONAL, MAX_FAILURE, etc
    There are two major things to consider:
    ->Is the throughput of the destination enough to handle max archive log generation?
    You need to take a look at the size of archive logs and how frequently they are
    generated during peak load.
    ->What happens when the destination is not available?
    HTH,
    Iordan Iotzov

  • Archived log to standby server stopped

    Hi,
    I configured dataguard standby server few weeks ago,
    I set log_archive_dest_2='SERVICE=JHRD0107_OFS1 OPTIONAL'
    and log_archive_dest_state=ENABLE in primary server,
    and very go fine until yesterday, I shutdown my standby server and set the log_archive_dest_state=DEFER,
    when I tried to startup again the standby server, and set log_archive_dest_state=ENABLE again,
    the archive stop sending the archived log to log_archive_dest_2.
    No error was found in alert log of primary and standby server.
    It seem that the primary server don't know there is another archive destination..
    The database version is 10g 10.2.0.4
    Anyone what could be the cause of this?
    Thanks

    Hi, Below is the parameter related to archive in my primary server
    SQL> alter system set log_archive_dest_state_2=enable;
    System altered.
    SQL> show parameter archive
    NAME TYPE VALUE
    archive_lag_target integer 0
    log_archive_config string
    log_archive_dest string
    log_archive_dest_1 string LOCATION=I:\OFSDB\OFS1\archive
    MANDATORY REOPEN=300
    log_archive_dest_10 string LOCATION=USE_DB_RECOVERY_FILE_
    DEST OPTIONAL REOPEN=300
    log_archive_dest_2 string SERVICE=JHRD0107_OFS1 OPTIONAL
    log_archive_dest_3 string
    log_archive_dest_4 string
    log_archive_dest_5 string
    NAME TYPE VALUE
    log_archive_dest_6 string
    log_archive_dest_7 string
    log_archive_dest_8 string
    log_archive_dest_9 string
    log_archive_dest_state_1 string enable
    log_archive_dest_state_10 string enable
    log_archive_dest_state_2 string ENABLE
    log_archive_dest_state_3 string enable
    log_archive_dest_state_4 string enable
    log_archive_dest_state_5 string enable
    log_archive_dest_state_6 string enable
    NAME TYPE VALUE
    log_archive_dest_state_7 string enable
    log_archive_dest_state_8 string enable
    log_archive_dest_state_9 string enable
    log_archive_duplex_dest string
    log_archive_format string ARC%S_%R.%T
    log_archive_local_first boolean TRUE
    log_archive_max_processes integer 2
    log_archive_min_succeed_dest integer 1
    log_archive_start boolean FALSE
    log_archive_trace integer 0
    remote_archive_enable string true
    NAME TYPE VALUE
    standby_archive_dest string %ORACLE_HOME%\RDBMS
    Thanks for your response.
    Vincent

Maybe you are looking for

  • Problem when download IR to CSV format in Apex 4.1

    Hi, I've an interactive report that using date variables in its query. In order this report to produce data, you must submit those variables otherwise it will show text 'no data found'. the report can be produced after those variable is submitted, an

  • Over-protected external HDs-now can't see them :(

    Folks I "protected" my External HDs in 'Ownership and Permissions" whilst my grandkids were wanting to "play" on my G5, but I must have overdone it because now I can't see them on my Desktop to UN-protect them! Can anybody help get 'em back, please?

  • Usage of - BAPI_POSRVAPS_SAVEMULTI3

    Hello Frnds, Im using this above mentioned BAPI to create Stock Transfer Requisition in APO and send it to R3. Presently I could successfully create an Order in APO but it is not ciffing across to R3 automatically. I have to manually reconcile the or

  • How to create a new database

    hi I am not able to create a new database other than the one created during the installation. i am using the oracle 9i client. i am able to create new database from the enterprise manager console,add a database manually but not able to connect to the

  • SQL 2005 cluster rejects SQL logins when in failed over state

    When SQL 2005 SP4 on Windows 2003 server cluster is failed over from Server_A to Server_B, it rejects all SQL Server logins. domain logins are OK. The message is "user is not associated with a trusted server connection", then the IP of the client. Th