Redo log API - Logminer

In my never ending quest (no pun intended) to get data off of Oracle into
another database, I am now looking towards scanning the Redo logs (using
Logminer?) to pull out the updates and transmit them to the target DBMS in
close to real time.
The question is, is this capability in Oracle 8i and/or 9i?
Does anyone have any sample code? Has anyone ever done this? Pitfalls?
Gotchas?
Andre
null

we are all wondering why you want it to switch at 1800 seconds?
If you set LOG_CHECKPOINT_TIMEOUT=1800 and LOG_CHECKPOINT_INTERVAL=0 it should work. ALTER SYSTEM SWTITCH LOGFILE will work too, but if you're using a standby you could run into an ORA-00332: archived log is too small - may be incompletely archived. If 1800 seconds go by and only 1Mb of redologs are generated, you may hit this error on your sdb.
If you're in a situation where you can't afford to lose more than 1800 seconds worth of redo being archived, you could create a job to run during your busiest hours, and then not run during your off-time (if that exists) or slow time.
Maybe some more info about why you want to do this would help us provide better solutions?

Similar Messages

  • Standby Database fails to read dictionary from redo log

    hi,
    I am attempting to create a Logical standby database on same machine as the primary database. I have executed the steps outlined in Oracle Documentation several times, but end up with the same error. Detailes of setup and error are provided below. Please help. Thanks.
    ==========
    OS: REdhat 8 (2.4.18-14)
    RDBMS: Oracle EE Server 9.2.0.3.0
    primary db init details:
    *.log_archive_dest_1='LOCATION=/usr3/oracle/admin/lbsp/archive/ MANDATORY'
    *.log_archive_dest_2='SERVICE=STDBY'
    standby db init details:
    log_archive_dest_1='LOCATION=/usr3/oracle/admin/stdby/archive/'
    standby_archive_dest='/usr3/oracle/admin/lbsp/archive_pdb/'
    Standby alert log file (tail)
    LOGSTDBY event: ORA-01332: internal Logminer Dictionary error
    Sun Jul 13 11:37:20 2003
    Errors in file /usr3/oracle/admin/stdby/bdump/stdby_lsp0_13691.trc:
    ORA-01332: internal Logminer Dictionary error
    LSP process trace file:
    Instance name: stdby
    Redo thread mounted by this instance: 1
    Oracle process number: 18
    Unix process pid: 13691, image: oracle@prabhu (LSP0)
    *** 2003-07-13 11:37:19.972
    *** SESSION ID:(12.165) 2003-07-13 11:37:19.970
    <krvrd.c:krvrdfdm>: DDL or Dict mine error exit. 384<krvrd.c:krvrdids>: Failed to mine dictionary. flgs 180
    knahcapplymain: encountered error=1332
    *** 2003-07-13 11:37:20.217
    ksedmp: internal or fatal error
    . (memory dump)
    KNACDMP: Unassigned txns = { }
    KNACDMP: *******************************************************
    error 1332 detected in background process
    OPIRIP: Uncaught error 447. Error stack:
    ORA-00447: fatal error in background process
    ORA-01332: internal Logminer Dictionary error
    another trace file created by error is: stdby_p001_13695.trc
    Instance name: stdby
    Redo thread mounted by this instance: 1
    Oracle process number: 20
    Unix process pid: 13695, image: oracle@prabhu (P001)
    *** 2003-07-13 11:37:19.961
    *** SESSION ID:(22.8) 2003-07-13 11:37:19.908
    krvxmrs: Leaving by exception: 604
    ORA-00604: error occurred at recursive SQL level 1
    ORA-01031: insufficient privileges
    ORA-06512: at "SYS.LOGMNR_KRVRDREPDICT3", line 68
    ORA-06512: at line 1
    there are no errors anywhere during the creation, mounting or opening of standby database. After the initial log register, any log switch on primary is communicated to standby and visible in DBA_LOGSTDBY_LOG. Also, archived logs from primary are successfuly copied by system to directory pointed by standby db's standby_archive_dest parameter.
    I noticed, somehow everytime I issue "ALTER DATABASE START LOGICAL STANDBY APPLY" command the procedures and packages related to logmnr get invalid. I compile them and again after "APPLY" they become invalid.
    Invalid object list:
    OBJECT_TYPE OBJECT_NAME
    VIEW DBA_LOGSTDBY_PROGRESS
    PACKAGE BODY DBMS_INTERNAL_LOGSTDBY
    PACKAGE BODY DBMS_STREAMS_ADM_UTL
    VIEW LOGMNR_DICT
    PACKAGE BODY LOGMNR_DICT_CACHE
    PROCEDURE LOGMNR_GTLO3
    PROCEDURE LOGMNR_KRVRDA_TEST_APPLY
    Anybody point out what I am doing wrong. Thanks for the help

    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mountedhave you mentioned parameter LOG_FILE_NAME_CONVERT in standby when online redo log locations are different?
    post from standby:-
    SQL> select name, state From v$asm_diskgroup;
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privilegesPost from primary & standby
    SQL> select * from v$pwfile_users;
    User Profile for 919131
    919131     
    Handle:     919131  
    Status Level:     Newbie
    Registered:     Mar 6, 2012
    Total Posts:     16
    Total Questions:     8 (8 unresolved)
    OTN failed 100% to help you, then why you posted another question?
    First close all your old answered threads and then better continue your updates in your thread.
    Edited by: CKPT on Jul 9, 2012 11:45 AM

  • Redo log mining.

    Hello,
    We've a database on which auditing having been activated. Some fraudulous acctivities have been operated on that database. The database was not running in archivelog mode. Now management wants information about those activities to be retrieved. The database is 9iR2.
    1- Is it possible to get the trace of those fraudulous activities given the situation I described?
    2- Some people suggest we exploit the content of the redos. Can anyone explain/tell the techniques that we can use to mine the redo log?
    Thanks in advance.

    Damorga,Erka, thanks for your replies. The team that was managing the databases wasn't a team of DBAs. They were rather System/network Administrators and we cannot blame them.
    New structural chages are being made and the new team of DBAs was set up about 2 months ago. It's only at that moment that we realized the sinning I described above. We turned auditing and archiving on. The problem which brought me to post here happened long before we were hired. That means the content of the redos were flushed everytime they were full, but management is putting pressure on us to try and get a trace of those activities. I wanted to confirm that we cannot get those trace in the situation I presented.
    Now I would like to know, how far in the past can logminer go? I mean, I know it works best in archivelog mode but if we apply logminer on archives that are like 4 months or more old, can we have significant information about everything that happended during that period?
    Edited by: [email protected] on 7 févr. 2009 02:41

  • Logmnr/capure error b'coz of corruption of redo log block

    Hi,
    We all know that capture process reads the REDO entries from redo log files or archived log files. Therefore we need to ahev db in ARCHIVELOG mode.
    In alert log file, I found error saying :
    Creating archive destination LOG_ARCHIVE_DEST_1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\LOCATION01\1_36.ARC'
    ARC0: Log corruption near block 66922 change 0 time ?
    ARC0: All Archive destinations made inactive due to error 354
    Fri Apr 04 12:57:44 2003
    Errors in file e:\oracle\admin\repf\bdump\trishul_arc0_1724.trc:
    ORA-00354: corrupt redo log block header
    ORA-00353: log corruption near block 66922 change 0 time 04/04/2003 11:05:40
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    As a normal practice, we do have multiplexing of redo log files at diff location, but even that second copy of redo log is of no use to recover the redo log. This explains redo log could not be archived, since it can't be read. Same is true even for Logmnr process, it could not read the redo log file and it failed. Now, we have wae to recover from this situation (as far as DB is concern, not Stream Replication), since the shutdown after this error was IMMEDIATE causing checkpoing, and rollback/rollforward is not required during system startup. (No instance recovery) We can make db NOARCHIVELOG mode, drop that particular group, and create new one, and turn db to ARCHIVELOG mode This will certainly serve the purpose as far as consistency of DB is concern.
    Here is a catch for Stream Replication. The redo log that got corrupted must be having few transaction which are not being archived, and each will be having corresponding SCN. Now, Capture Process read the info sequentially in order of SCN. Few transaction are now missed, and Capture process can't jump to next SCN skipping few SCN in between. So, we have to re-instantiate the objects on the another system which has no erros, and start working on it. My botheration is what will happen to those missed transaction on the another database. It's absolete loss of the data. In development I can manage that. But in real time Production stage, this is a critical situation. How to recover from this situation to get back the corrupted info from redo log ?
    I have not dropped any of the log group yet. B'coz I would like to recover from this situation without LOSS of data.
    Thanx, & regards,
    Kamlesh Chaudhary
    Content of trace files :
    Dump file e:\oracle\admin\repf\bdump\trishul_arc0_1724.trc
    Fri Apr 04 12:57:31 2003
    ORACLE V9.2.0.2.1 - Production vsnsta=0
    vsnsql=12 vsnxtr=3
    Windows 2000 Version 5.0 Service Pack 2, CPU type 586
    Oracle9i Enterprise Edition Release 9.2.0.2.1 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.2.0 - Production
    Windows 2000 Version 5.0 Service Pack 2, CPU type 586
    Instance name: trishul
    Redo thread mounted by this instance: 1
    Oracle process number: 16
    Windows thread id: 1724, image: ORACLE.EXE
    *** SESSION ID:(13.1) 2003-04-04 12:57:31.000
    - Created archivelog as 'E:\ORACLE\ORADATA\REPF\ARCHIVE\LOCATION02\1_36.ARC'
    - Created archivelog as 'E:\ORACLE\ORADATA\REPF\ARCHIVE\LOCATION01\1_36.ARC'
    *** 2003-04-04 12:57:44.000
    ARC0: All Archive destinations made inactive due to error 354
    *** 2003-04-04 12:57:44.000
    kcrrfail: dest:2 err:354 force:0
    *** 2003-04-04 12:57:44.000
    kcrrfail: dest:1 err:354 force:0
    ORA-00354: corrupt redo log block header
    ORA-00353: log corruption near block 66922 change 0 time 04/04/2003 11:05:40
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    *** 2003-04-04 12:57:44.000
    ARC0: Archiving not possible: error count exceeded
    ORA-16038: log 2 sequence# 36 cannot be archived
    ORA-00354: corrupt redo log block header
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-16014: log 2 sequence# 36 not archived, no available destinations
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-16014: log 2 sequence# 36 not archived, no available destinations
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-16014: log 2 sequence# 36 not archived, no available destinations
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-16014: log 2 sequence# 36 not archived, no available destinations
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG'
    ORA-16014: log 2 sequence# 36 not archived, no available destinations
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\REDO02.LOG'
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\ORADATA\REPF\ARCHIVE\REDO02.LOG
    Dump file e:\oracle\admin\repf\udump\trishul_cp01_2048.trc
    Fri Apr 04 12:57:27 2003
    ORACLE V9.2.0.2.1 - Production vsnsta=0
    vsnsql=12 vsnxtr=3
    Windows 2000 Version 5.0 Service Pack 2, CPU type 586
    Oracle9i Enterprise Edition Release 9.2.0.2.1 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.2.0 - Production
    Windows 2000 Version 5.0 Service Pack 2, CPU type 586
    Instance name: trishul
    Redo thread mounted by this instance: 1
    Oracle process number: 30
    Windows thread id: 2048, image: ORACLE.EXE (CP01)
    *** 2003-04-04 12:57:28.000
    *** SESSION ID:(27.42) 2003-04-04 12:57:27.000
    TLCR process death detected. Shutting down TLCR
    error 1280 in STREAMS process
    ORA-01280: Fatal LogMiner Error.
    OPIRIP: Uncaught error 447. Error stack:
    ORA-00447: fatal error in background process
    ORA-01280: Fatal LogMiner Error
    **********************

    I have the similar problem - I am using Steams environment, and have got this
    "ORA-00353: log corruption near block" errors in the alert.log file
    during capture the changes on the primary database, and Capture
    process became aborted after that.
    Was that transactions lost, or after i've started the Capture process
    again the were captured and send to the target database?
    Have anyone solved that problem?
    Can you help me with it?

  • Streams capture waiting for dictionary redo log

    Hi ,
    The stream capture is waiting for the dictionary redo log
    as per the alert logs ,the logminer is able to register the logfile
    RFS LogMiner: Registered logfile [TEMP_114383_1_628824420.arc] to LogMiner session id [142]
    Fri Feb 13 00:00:39 2009
    Capture Session Redo Total
    Capture Process Session Serial Entries LCRs
    Name Number ID Number State Scanned Enqueued
    C_REF C001 675 2707 WAITING FOR DICTIONARY REDO 0 0
    Capture Capture Capture
    Process Process Positive Negative Process
    Name Queue Rule Set Rule Set Status
    C_REF CA_REF RULESET$_80 ENABLED
    Capture Capture
    Capture Process Process
    Name Queue START_SCN Status STATUS_CHAN CAPTURED_SCN APPLIED_SCN USE FIRST_SCN
    C_REF CA_REF 8586133398117 ENABLED 12-Feb-2009 8586133398117 8586133398117 YES 8586133398117
    CONSUMER_NAM SEQUENCE# FIRST_SCN NEXT_SCN TO_DATE(FIR TO_DATE(NEX NAME
    C_REF 114378 8586133399062 8586162685837 12-Feb-2009 12-Feb-2009 /TEMP_114378_1_628824420.arc
    C_REF 114379 8586162685837 8586163112496 12-Feb-2009 12-Feb-2009 /TEMP_114379_1_628824420.arc
    C_REF 114380 8586163112496 8586163984886 12-Feb-2009 12-Feb-2009 /TEMP_114380_1_628824420.arc
    C_REF 114381 8586163984886 8586163986301 12-Feb-2009 12-Feb-2009 /TEMP_114381_1_628824420.arc
    C_REF 114382 8586163986301 8586163987651 12-Feb-2009 12-Feb-2009 /TEMP_114382_1_628824420.arc
    C_REF 114383 8586163987651 8586163989497 12-Feb-2009 13-Feb-2009 /TEMP_114383_1_628824420.arc
    C_REF 114384 8586163989497 8586163989674 13-Feb-2009 13-Feb-2009 /TEMP_114384_1_628824420.arc
    Capture Time of
    Name LogMiner ID Last Redo SCN Last Redo SCN
    C_REF 142 8586166339742 00:10:13 02/13/09
    i am not still able to make out even after the archivelogs are registered they are not done logmining by logminer.
    Can you please help ,i am stuck up this situation.i have rebuild streams by completely removing the stream configuration and also dropped and recreated the strmadmin.

    Perhaps I missed it in your post but I didn't see a version number or any information as to what form of Streams was implemented or how.
    There are step-by-step instructions for debugging Streams applications at metalink. I would suggest you find the directions for your version and follow them.

  • Third party tools for redo log

    Dear,
    Any third party tools can read redo log for oracle9i?
    ManyThanks

    Most 3rd parties gave up when they realized that
    a) Oracle now owns, and includes for free, the logminer, and
    b) Oracle is free - and willing to - change the contents of the log files at any time, including patch releases.

  • Redo Log Files - more than 12 per hour

    Hello @all,
    I have a problem with my redo log files. I get more than 12 switches per Hour. I have 3 Files with 5oM size. I increased the sitz to 15o M, but
    I still have 12 switches per hour.
    Do anyone know, what I did wrong?
    Database:
    Oracle 9i
    Thanks
    Martin

    user9528362 wrote:
    Hello @all,
    yes I know, that 3 switches per hour are perfekt, but I had increased the Size from 5o M to 15oM already and the amount from switches, are not reduced.
    So there must be something else, that causes the log switches.Martin,
    As I said somewhere above too, 150meg is a tiny size if you are managing a production db. I have already mentioned that make your log file size to at least 500meg and than check. As for the high redo activity, only you can confirm that whether this has been started from now only or was happening before too? In anycase, for an active oltp, 500 to 1gb of redo log file size should be okay.
    For the extra redo generation, you have been given link for mining log files using logminer. Try using it to see what is causing extra redo.
    HTH
    Aman....

  • Redo log

    Can I read/exploit the redo log files (outside recovery activities)?
    My purpose would be to use the redo log file in case of trouble/complain in order to see who did what and when.
    Are the redo log files encrypted or could they be encrypted?
    Thanks in advance

    Can I read/exploit the redo log files (outside recovery activities)?
    My purpose would be to use the redo log file in case of trouble/complain in order to see who did what and when.
    Are the redo log files encrypted or could they be encrypted?You can read a redo log files using logminer. It is not opening the file like in textpad or notepad, but is a tool provided by oracle. Depending upon you setting if you use advanced securty and encrypted data in while storing on to file then it will be encrypted otherwise it will be in normal binary format.
    Regards.

  • Reading Orcle current online redo logs

    Hi,
    We are able to read the SQL Server Transactional Log. We are trying to do that in case Oracle. Can anyone put some light on this?
    Is there any third party tools doing this? Is this a feasible idea or not.
    Regards,
    Kalyan

    You can use Logminer to look into the archived redo logs.
    Why do you want to look at this though?

  • Redo Log and Supplemental Logging related doubts

    Hi Friends,
    I am studying Supplemental logging in detail. Have read lots of articles and oracle documentation about it and redo logs. But couldnot found answers of some doubts..
    Please help me clear it.
    Scenario: we have one table with primary key. And we execute an update query on that table which is not using the primary key column in any clause..
    Question: In this case, does the redo log entry generated for the changes done by update query contain the primary columns values..?
    Question: If we have any table with primary key, do we need to enable the supplemental logging on primary columns of that table? If yes, in which circumstances, do we need to enable it?
    Question: If we have to configure stream replication on that table(having primary key), why do we actually need to enable its supplemental logging ( I have read the documentation saying that stream requires some more information so.., but actually what information does it need. Again this question is highly related to the first question.)
    Also please suggest any good article/site which provide inside details of redo log and supplemental logging, if you know.
    Regards,
    Dipali..

    1) Assuming you are not updating the primary key column and supplemental logging is not enabled, Oracle doesn't need to log the primary key column to the redo log, just the ROWID.
    2) Is rather hard to answer without being tautological. You need to enable supplemental logging if and only if you have some downstream use for additional columns in the redo logs. Streams, and those technologies built on top of Streams, are the most common reason for enabling supplemental logging.
    3) If you execute an update statement like
    UPDATE some_table
      SET some_column = new_value
    WHERE primary_key = some_key_value
       AND <<other conditions as well>>and look at an update statement that LogMiner builds from the redo logs in the absence of supplemental logging, it would basically be something like
    UPDATE some_table
      SET some_column = new_value
    WHERE rowid = rowid_of_the_row_you_updatedOracle doesn't need to replay the exact SQL statement you issued, (and thus it doesn't have to write the SQL statement to the redo log, it doesn't have to worry if the UPDATE takes a long time to run (otherwise, it would take as long to apply an archived log as it did to generate the log, which would be disasterous in a recovery situation), etc). It just needs to reconstruct the SQL statement from the information in redo, which is just the ROWID and the column(s) that changed.
    If you try to run this statement on a different database (via Streams, for example) the ROWIDs on the destination database are likely totally different (since a ROWID is just a physical address of a row on disk). So adding supplemental logging tells Oracle to log the primary key column to redo and allows LogMiner/ Streams/ etc. to reconstruct the statement using the primary key values for the changed rows, which would be the same on both the source and destination databases.
    Justin

  • Redo Log Generation Rate is more as compaired to Single Instacne

    Hi Experts,
    i need views regarding the experience that i m facing in rac environment. actually we have two node rac 10gr2 running on AIX 6.1
    on application side we have a process of 4 hrs and during this process we observed average 110 Gb of logs. but if we run same process on application and single instance DB is being used at back end, then it generates only 40Gb of log. now my question is why same process have different redo log generation values against rac db and single instacne.

    Amiabu,
    Your question calls for crystal balls, as you have provided way too few details.
    Also you didn't use Logminer to find out what is going on. This would have been a sensible approach, even before posting a question which can be paraphrased as 'It doesn't work as expected, why?'.
    Two general observations:
    if your code doesn't scale, using RAC will make things worse.
    Oracle 10g is a heavily instrumented product, assuming you use AWR or ADDM or statspack.
    The Cache Fusion feature of RAC can easily result in extra wait events, which in turn are being tracked (INSERTed that is) in the AWR or statspack repository.
    Only you can answer the question what was going on by using Logminer.
    Sybrand Bakker
    Senior Oracle DBA
    Edited by: sybrand_b on 16-jun-2011 14:33

  • What will happen when redo log file or archive log file, which is yet to be

    What will happen when redo log file or archive log file, which is yet to be read by logminer is corrupted? It seems that the captures process hangs between “Paused for flow control” and “Enqueuing Messages”. How to come out of this condition without recreating the captures process?
    Any clue is helpful
    Thanks in advance for your help.

    Basically you can't skip SCN since it will result in a data integrity issues (say you had skipped some inserts and later there will be some updates to a not replicated data).
    Streams maintain their own checkpoint tables with transaction related stuff. So there is no way you can jump over a range of SCN's without recreating capture.
    The only thing you can try - temporary give capture process a rule without any objects. But it will need to mine through the redo anyway.

  • Will Analyze create too many redo logs ??

    DB : 10gR2
    Hello Folks,
    Is analyze process creates too many redo log files ?? Is their any other valid option other than no log ?

    I guess that's also going to create too many log switches as it's part of DDLHUH?
    REDO log switches result from DML.
    If you are obsessed with LOG switches, then you need to use LOGMINER to observe the the relative per centage of changed data is from DBMS_STATS
    when compared to normal DML activity.
    I seriously doubt that DBMS_STATS has any significant impact to redo log switches.

  • How do I manually archive 1 redo log at a time?

    The database is configured in archive mode, but automatic archiving is turned off.
    For both Oracle 901 and 920 on Windows, when I try to manually archive a single redo log, the database
    archives as many logs as it can up to the log just before the current log:
    For example:
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 NO INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 NO INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 NO INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 NO INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    6 rows selected.
    SQL> alter system archive log next;
    System altered.
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 YES INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 YES INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 YES INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 YES INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    See - instead of only 1 log being archive, 4 of them were. Oracle behaves the same way if I use the "sequence" option:
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 NO INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 NO INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 NO INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 NO INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    6 rows selected.
    SQL> alter system archive log next;
    System altered.
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 YES INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 YES INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 YES INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 YES INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    Is there some default system configuration property telling Oracle to archive as many logs as it can?
    Thanks,
    DGR

    Thanks Yoann (and Syed Jaffar Jaffar Hussain too),
    but I don't have a problem finding the group to archive or executing the alter system archive log command.
    My problem is that Oracle doesn't work as I expect it.
    This comes from the Oracle 9.2 online doc:
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96540/statements_23a.htm#2053642
    "Specify SEQUENCE to manually archive the online redo log file group identified by the log sequence number integer in the specified thread."
    This implies that Oracle will only archive the log group identified by the log sequence number I specify in the alter system archive log sequence statement. However, Oracle is archiving almost all of the log groups (see my first post for an example).
    This appears to be a bug, unless there is some other system parameter that is configured (by default) to allow Oracle to archive as many log groups as possible.
    As to the reason why - it is an application requirement. The Oracle db must be in archive mode, automatic archiving must be disabled and the application must control online redo log archiving.
    DGR

  • Select from .. as of - using archived redo logs - 10g

    Hi,
    I was under the impression I could issue a "Select from .. as of" statement back in time if I have the archived redo logs.
    I've been searching for a while and cant find an answer.
    My undo_management=AUTO, database is 10.2.0.1, the retention is the default of 900 seconds as I've never changed it.
    I want to query a table as of 24 hours ago so I have all the archived redo logs from the last 48 hours in the correct directory
    When is issue the following query
    select * from supplier_codes AS OF TIMESTAMP
    TO_TIMESTAMP('2009-08-11 10:01:00', 'YYYY-MM-DD HH24:MI:SS')
    I get a snapshot to old ORA-01555 error. I guess that is because my retention is only 900 seconds but I thought the database should query the archive redo logs or have I got that totally wrong?!
    My undo tablespace is set to AUTOEXTEND ON and MAXSIZE UNLIMITED so there should be no space issues
    Any help would be greatly appreciated!
    Thanks
    Robert

    If you want to go back 24 hours, you need to undo the changes...
    See e.g. the app dev guide - fundamentals, chapter on Flashback features: [doc search|http://www.oracle.com/pls/db102/ranked?word=flashback&remark=federated_search].

Maybe you are looking for

  • Possible to count only hand hits

    hey guys, This exercise has been in my back pocket for a month. Worked with one of our coders on this and she's empty of ideas. I've been trying to create a teach game to make correct change. So what I have is change, a hand and a table. I want a liv

  • Help for a newbie

    I am to iphone, been an An$%*#d user for years and now am loving life on the iphone. However, my son also has an ipod. When he downloads his games to is ipod, they show up on my iphone. He had the ipod before I had an iphone, so I am assuming our acc

  • How to use the Function module BDM_DISPUTE_NOTES_GET

    i have to use the FM BDM_DISPUTE_NOTES_GET to ET_NOTES CALL FUNCTION 'BDM_DISPUTE_NOTES_GET'   EXPORTING     im_guid        = IMPORTING ES_RETURN       =    ET_NOTES        = V_STRING. V_STRING type ???????(i want this) my output will be a paragraph

  • Drawing a String at an angle

    Hi, I have use the Graphics class to develope a map of streets etc and I wish to place the name of the streets running parallel to the streets themselves. The drawString method allows me to specifiy the string and start co-ordinates but can anyone te

  • How can I run C programs under a Java GUI?

    Hi I'm looking for some info on how to run a command line program written in another language behind a Java GUI, any advice greatly appreciated, particularly in how to transfer data between the two. Thanks