Redo log sequence number

hi everyone,
I am looking at our SB system, and try to do only redo log backup. however it comes back with error message: The current redo log sequence number is not greater than the sequence number of the last saved offline redo log file.
So, we suspect that since last refresh, we may reset the sequence back from 1.  Any one know how to solve this problem?  Thanks in advance.
Amy

Hi,
Is your probh solved?
Can u pl tell if u executed a recovery of database earlier?
Pl check if there was a case of reset logs.
To do this run BRTOOLS-->Instance Mgt->Show Inst Status.
This will show if there was a case of reset logs.
If it is so try the following:
Shut down sap and the database.
Look for a folder named oraarch
Oracle/sid/oraarch
Copy all the archive files to another folder and delete all the arc files from the oraarch folder.
The name of the arch files will be like <SID>ARCHARC123456.001
After copying to another location and subsequently deleting them from oraarch folders start up oracle in mount mode.
Then issue the following sql
alter database open resetlogs;
then start sap.
NB:Dont forget to take a complete backup before and after diong this.
Hope this will help.
Most probably your database contains two icarnations of redo log files so this error is comming.
regards

Similar Messages

  • Doubt about CURRENT LOG SEQUENCE NUMBER

    I just want to know that why the "current log sequence number" is added to the header of each data file ? why oracle needs to do that ? as it is also added to the control file and whenever it needs to know the current log sequence number , it can get it from control file , then why to add this in the header of data files ?

    For Recovery.
    When you are doing a Recovery of a database where the datafiles have been restored from backups but the controlfile has been recreated (thus, the controlfile has no information about log sequence numbers and checkpoint scns), the Recovery must read this information from the datafile headers.
    Say you have 5 datafiles and they have been backed up 6 months ago using ALTER TABLESPACE BEGIN BACKUP commands and, therefore, have different checkpoint SCNs. The checkpoint SCNs could be 1021, 1059, 1085, 1745, 2045.
    How would Oracle know which Log Sequence Numbers are needed ? It knows that a Recovery must start with SCN 1021. But if the information about Log Sequence Numbers is not available in the controlfile (either because it is as of today and the controlfile doesn't store information about 6 months of redo logs OR because the controlfile has been recreated), what is the Log Sequence Number corresponding to that SCN 1021 ?
    That information is available in the datafile header. You will NOT see it in V$DATAFILE or V$DATAFILE_HEADER. You will not see it in a level 1 dump of datafile headers.
    You will see an example here :
    http://web.singnet.com.sg/~hkchital/Incomplete_Recovery_with_BackupControlfile.doc
    Hemant K Chitale
    http://hemantoracledba.blogspot.com

  • Current log sequence number

    i have a small doubt :
    when there is log switch , the current log sequence number is stored in the control file . Is the same current log sequence number stored in the headers of the datafiles??????
    waiting for your answers eargerly.....
    kumaresan.M

    nops i don think so.
    there is no need for doing this, only controlfile needs to know about the current sequence number
    Regards

  • Log sequence number

    who updates the log sequence number in control file and datafiles , I know chkpoint is responsible to update checkpoint number in datafile and control file number.

    it is LGWR.

  • Who writes LOG SEQUENCE NUMBER to CONTROL FILE,whether CKPT or LGWR process

    I am in confusion that,
    When log switch occurs that time Log sequence no. is generated by LGWR. then after who writes that LSN to CONTROL FILE whether Checkpoint or LGWR.
    Please provide me proper answer with example.....

    I notice by using fuser on a controlfile on hp-ux, a number of processes keep the controlfile open.  So it would make sense that the one that writes logs would write log sequence numbers to the controlfile too.
    $ for z in  `/usr/sbin/fuser  control03.ctl|awk ' {x=split($0,args); for (i=2;i<=x;i++){print args[i]}}'`
            do
                ps -efx|grep $z|grep -v grep
            done
    control03.ctl: oooooooooo
      oracle 27129     1  0  Sep  5  ?        12:15 ora_arc1_TPRD
      oracle 27073     1  0  Sep  5  ?        52:45 ora_dbw0_TPRD
      oracle 27127     1  0  Sep  5  ?        10:52 ora_arc0_TPRD
      oracle  9485     1  0  Nov  3  ?         9:14 oracleTPRD (LOCAL=NO)
      oracle 22395     1  0 08:36:22 ?         0:24 oracleTPRD (LOCAL=NO)
      oracle 27075     1  0  Sep  5  ?        68:53 ora_lgwr_TPRD
      oracle 22116     1  0 08:33:24 ?         1:06 oracleTPRD (LOCAL=NO)
      oracle 27077     1  0  Sep  5  ?        41:05 ora_ckpt_TPRD
      oracle 26510     1  0  Nov 30  ?        21:26 oracleTPRD (LOCAL=NO)

  • Command that give me the log sequence number of Specifc date

    hi,
    I need a command that gives me the LSN number of previous date. e.g if current date is 20-02-2013. & i want to get the LSN number that have date last month.28-01-2013.
    thanks.
    Regards,

    No Replies

  • Redo log files are not applying to standby database

    Hi everyone!!
    I have created standby database on same server ( windows XP) and using oracle 11g . I want to synchronize my standby database with primary database . So I tried to apply redo logs from primary to standby database as follow .
    My standby database is open and Primary database is not started (instance not started) because only one database can run in Exclusive Mode as DB_NAME is same for both database.  I run the following command on the standby database.
                SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;
    It returns "Database altered" . But when I checked the last archive log on primary database, its sequence is 189 while on standby database it is 177. That mean archived redo logs are not applied on standby database.
    The tnsnames.ora file contains entry for both service primary & standby database and same service has been used to transmit and receive redo logs.
    1. How to resolve this issue ?
    2.Is it compulsory to have Primary database open ?
    3. I have created standby  control file by using  command
              SQL> ALTER DATABASE CREATE STANDBY CONTROLFILE AS ‘D:\APP\ORACLE\ORADATA\TESTCAT\CONTROLFILE\CONTROL_STAND1.CTL‘;
    So database name in the standby control file is same as primary database name (PRIM). And hence init.ora file of standby database also contains DB_NAME = 'PRIM' parameter. I can't change it because it returns error of mismatch database name on startup. Should  I have different database name for both or existing one is correct ?
    Can anybody help me to come out from this stuck ?
    Thanks & Regards
    Tushar Lapani

    Thank you Girish. It solved  my redo apply problem. I set log_archive_dest parameter again and then I checked archive redo log sequence number. It was same for both primary and standby database. But still table on standby database is not being refresh.
    I did following scenario.
    1.  Inserted 200000 rows in emp table of Scott user on Primary database and commit changes.
    2. Then I synchronized standby database by using Alter database command. And I also verify that archive log sequence number is same for both database. It mean archived logs from primary database has been applied to standby database.
    3. But when I count number of rows in emp table of scott user on standby database, it returns only 14 rows even of redo log has been applied.
    So my question is why changes made to primary database is not reflected on standby database although redo logs has been applied ?
    Thanks

  • Thread 1 advanced to log sequence (redo log)

    Hi All,
    db:oracle 10gR2(10.2.0.4)
    os:solaris10
    My alert log file shows following information all the time....can you please explain me whats going on??
    Mon Jan 30 01:40:30 2012
    Thread 1 advanced to log sequence 253305
    Current log# 2 seq# 253305 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO02.LOG
    Mon Jan 30 01:45:40 2012
    Thread 1 advanced to log sequence 253306
    Current log# 3 seq# 253306 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO03.LOG
    Mon Jan 30 01:51:21 2012
    Thread 1 advanced to log sequence 253307
    Current log# 1 seq# 253307 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO01.LOG
    Mon Jan 30 01:56:53 2012
    Thread 1 advanced to log sequence 253308
    Current log# 2 seq# 253308 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO02.LOG
    Mon Jan 30 02:01:57 2012
    Thread 1 advanced to log sequence 253309
    Current log# 3 seq# 253309 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO03.LOG
    Mon Jan 30 02:06:31 2012
    Thread 1 advanced to log sequence 253310
    Current log# 1 seq# 253310 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO01.LOG
    Mon Jan 30 02:11:37 2012
    Thread 1 advanced to log sequence 253311
    Current log# 2 seq# 253311 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO02.LOG
    Mon Jan 30 02:16:44 2012
    Thread 1 advanced to log sequence 253312
    Current log# 3 seq# 253312 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO03.LOG
    Mon Jan 30 02:21:51 2012
    Thread 1 advanced to log sequence 253313
    Current log# 1 seq# 253313 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO01.LOG
    Mon Jan 30 02:27:11 2012
    Thread 1 advanced to log sequence 253314
    Current log# 2 seq# 253314 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO02.LOG
    Mon Jan 30 02:32:31 2012
    Thread 1 advanced to log sequence 253315
    Current log# 3 seq# 253315 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO03.LOG
    Mon Jan 30 02:37:25 2012
    Thread 1 advanced to log sequence 253316
    Current log# 1 seq# 253316 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO01.LOG
    Mon Jan 30 02:42:15 2012
    Thread 1 advanced to log sequence 253317
    Current log# 2 seq# 253317 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO02.LOG
    Mon Jan 30 02:47:05 2012
    Thread 1 advanced to log sequence 253318
    Current log# 3 seq# 253318 mem# 0: C:\ORACLE\PRODUCT\10.2.0\ORADATA\JNPC\REDO03.LOG
    I appreciate your response......
    thanks.

    This information shows that the log switch has occured.
    The LGWR writes the redo information from the redo log buffer to the online redo log file. When the current online redo log file fills, LGWR begins writing to the next available online redo log file. In the mean time, the ARC process archives the contents of the filled online redo log file (If the database is in archivelog mode. If its in noarchive mode, then the filled online redo log file would not be archived). These archives are given a unique log sequence number.
    In your case 253305, 253306..are the log sequence numbers.
    You can refer http://docs.oracle.com/cd/B10501_01/server.920/a96521/onlineredo.htm for a detailed information

  • In standby db, can't find sequence number of Last Applied Log.

    Hello,
    Standby database is behind the primary database for over 200 hours, to repaire this, we are using a incremental backup from primary database and a restored control file.
    after starting up standby database, in Grid Control (OEM), can't find "last applied log" sequence number,
    go to that standby, do
    standby> select max(sequence#) from v$archived_log where applied='YES';
    MAX(SEQUENCE#)
    go to primary,
    do
    SQL> select max(sequence#) from v$archived_log where applied='YES';
    MAX(SEQUENCE#)
    83833
    then using OEM grid control, to Verify checks various standby database settings.
    Initializing
    Connected to instance standby_server:standby
    Starting alert log monitor...
    Updating Data Guard link on database homepage...
    Skipping verification of fast-start failover static services check.
    Data Protection Settings:
    Protection mode : Maximum Performance
    Redo Transport Mode settings:
    primary.com: ASYNC
    standby.com: ASYNC
    Checking standby redo log files.....OK
    Checking Data Guard status
    primary.com : Normal
    standby.com : Normal
    Checking inconsistent properties
    Checking agent status
    Checking applied log on standby........WARNING:
    Timed out after 60 seconds waiting for log to be applied.
    Processing completed.
    so how to fix this?
    thanks you very much.
    Edited by: 951932 on Oct 18, 2012 7:44 AM

    Hello;
    Probably nothing to fix. This is a common warning message.
    It even occurs in this example :
    http://www.databasejournal.com/features/oracle/article.php/10893_3826706_2/Oracle-11g-Data-Guard-Grid-Control-Management.htm
    Best Regards
    mseberg

  • How do I manually archive 1 redo log at a time?

    The database is configured in archive mode, but automatic archiving is turned off.
    For both Oracle 901 and 920 on Windows, when I try to manually archive a single redo log, the database
    archives as many logs as it can up to the log just before the current log:
    For example:
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 NO INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 NO INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 NO INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 NO INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    6 rows selected.
    SQL> alter system archive log next;
    System altered.
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 YES INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 YES INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 YES INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 YES INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    See - instead of only 1 log being archive, 4 of them were. Oracle behaves the same way if I use the "sequence" option:
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 NO INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 NO INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 NO INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 NO INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    6 rows selected.
    SQL> alter system archive log next;
    System altered.
    SQL> select * from v$log order by sequence#;
    GROUP# THREAD# SEQUENCE# BYTES MEMBERS ARC STATUS FIRST_CHANGE# FIRST_TIM
    1 1 14 104857600 1 YES INACTIVE 424246 19-JAN-05
    2 1 15 104857600 1 YES INACTIVE 425087 28-MAR-05
    3 1 16 104857600 1 YES INACTIVE 425088 28-MAR-05
    4 1 17 512000 1 YES INACTIVE 425092 28-MAR-05
    5 1 18 512000 1 NO INACTIVE 425100 28-MAR-05
    6 1 19 512000 1 NO CURRENT 425102 28-MAR-05
    Is there some default system configuration property telling Oracle to archive as many logs as it can?
    Thanks,
    DGR

    Thanks Yoann (and Syed Jaffar Jaffar Hussain too),
    but I don't have a problem finding the group to archive or executing the alter system archive log command.
    My problem is that Oracle doesn't work as I expect it.
    This comes from the Oracle 9.2 online doc:
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96540/statements_23a.htm#2053642
    "Specify SEQUENCE to manually archive the online redo log file group identified by the log sequence number integer in the specified thread."
    This implies that Oracle will only archive the log group identified by the log sequence number I specify in the alter system archive log sequence statement. However, Oracle is archiving almost all of the log groups (see my first post for an example).
    This appears to be a bug, unless there is some other system parameter that is configured (by default) to allow Oracle to archive as many log groups as possible.
    As to the reason why - it is an application requirement. The Oracle db must be in archive mode, automatic archiving must be disabled and the application must control online redo log archiving.
    DGR

  • BRARCHIVE fails with BR0017E error message: "off line redo log not found"

    After upgrade, archive format was changed to introduce resetlogs id in archive log file name
    Archive destination            /oracle/GDD/oraarch/GDDarch
    Archive format                 %t_%s_%r.dbf
    NOTE: %s log sequence number, %t thread number, %r resetlogs ID that ensures unique names are constructed for the archived log files across multiple incarnations of the database
    We did have problems with this upgrade so database was recovered once and started with resetlogs (new incarnation created).
    Now when we run brarchive it is trying to archive archive_log_file
    BR0017E Offline redo log file '/oracle/GDD/oraarch/GDDarch1_853_544272161.dbf' not found
    while on archive_log_dest we have files like "/oracle/GDD/oraarch/GDDarch1_853_618656359.dbf"
    Can you please help me to find out from where brarchive is picking parameter for log_archive_format and why it is not picking up current one causing missmatch between archvie_redo_log_file create and one BRARCHIVE is trying to back up.

    I already tried it (renamed archGDD.log to andrija.archGDD.log) and rerun
    brarchive -c -u / -r initGDD.utl.nbkp.bdhp4320 -cs
    it gave me output:
    brarchive -c -u / -r initGDD.utl.nbkp.bdhp4320 -cs
    BR0002I BRARCHIVE 7.00 (14)
    BR0006I Start of offline redo log processing: advfexyn.cps 2007-05-08 08.33.13
    BR0477I Oracle pfile /oracle/GDD/102_64/dbs/initGDD.ora created from spfile /oracle/GDD/102_64/dbs/spfileGDD.ora
    BR0013W No offline redo log files found for processing
    BR0007I End of offline redo log processing: advfexyn.cps 2007-05-08 08.33.15
    BR0280I BRARCHIVE time stamp: 2007-05-08 08.33.15
    BR0004I BRARCHIVE completed successfully with warnings
    and in archGDD.log that is created there is againreference to 0544272161 which seems to be still reference to old resetlog ID:
    bdhp4320:ac8558 41> vi archGDD.log
    "archGDD.log" 4 lines, 322 characters
    GDD  util_file  advfexyn cps  2007-05-08 08.33.13  2007-05-08 08.33.15  9  ...........         0        0        0
         0  -
    7.00 (14)  @0544272161
    GDD  util_file  advfexyn cps  2007-05-08 08.33.13  2007-05-08 08.33.15  1  ...........         0        0        0
         0  -
    7.00 (14)  @0544272161
    and in /oracle/GDD/saparch/advfexyn.cps file there is output like:
    Name                           Value
    oracle_sid                     GDD
    oracle_home                    /oracle/GDD/102_64
    oracle_profile                 /oracle/GDD/102_64/dbs/initGDD.ora
    sapdata_home                   /oracle/GDD
    sap_profile                    /oracle/GDD/102_64/dbs/initGDD.sap
    backup_dev_type                util_file
    util_par_file                  /oracle/GDD/102_64/dbs/initGDD.utl.nbkp.bdhp4320
    system_info                    oragdd/oragdd bdhp4320 HP-UX B.11.11 U 9000/800
    oracle_info                    GDD 10.2.0.2.0 8192 862 2969171885
    sap_info                       640 SAPGDD GDD W1381514715 R3_ORA 0020195637
    make_info                      hp_64 OCI_101 May 27 2006
    command_line                   brarchive -c -u / -r initGDD.utl.nbkp.bdhp4320 -cs
    BR0013W No offline redo log files found for processing
    BR0007I End of offline redo log processing: advfexyn.cps 2007-05-08 08.33.15
    BR0280I BRARCHIVE time stamp: 2007-05-08 08.33.15
    BR0004I BRARCHIVE completed successfully with warnings

  • Doubts in Redo log

    Hi,
    DB: 10.2.0.4
    OS: UNIX/Windows
    I have some doubts in archive log generation.
    *1)* Once current redo log is filed , LGWR will move to next redo log which is available with status INACTIVE , and current one will become ACTIVE.If more archive logs are generating , based on business , ACTIVE will become INACTIVE after sometime and can use for CURRENT redo log.
    If no archive logs are generated , my redo log is not coming out from status ACTIVE to INACTIVE .Even i waited for 5 minutes , but i had the same status.
    What will be the time to make ACTIVE to INACTIVE?. What causing to make this happen ?.
    Normally , when checkpoint is happened DBWR will write the CURRENT log data to data files , right?.In ACTIVE state , has archive log generated?. Many times i seen archive log is generated even it is in ACTIVE state.
    *2)* My Archive logs size (25MB) is not equal to the redo log size. Normally , when we switch the logs or any rman backup , logs will be switched automatically and how much redo is filled , same will be generated with size.
    But, in my case , more archive logs are generating and many have the different size out of generated and no manual switching or rman script are running during this period.
    Any idea on this?.
    Thanks,
    Sunand

    sunand wrote:
    Hi,
    DB: 10.2.0.4
    OS: UNIX/Windows
    I have some doubts in archive log generation.
    *1)* Once current redo log is filed , LGWR will move to next redo log which is available with status INACTIVE , and current one will become ACTIVE.If more archive logs are generating , based on business , ACTIVE will become INACTIVE after sometime and can use for CURRENT redo log.
    If no archive logs are generated , my redo log is not coming out from status ACTIVE to INACTIVE .Even i waited for 5 minutes , but i had the same status.
    What will be the time to make ACTIVE to INACTIVE?. What causing to make this happen ?.
    Normally , when checkpoint is happened DBWR will write the CURRENT log data to data files , right?.In ACTIVE state , has archive log generated?. Many times i seen archive log is generated even it is in ACTIVE state.
    That is wrong.DBWR process write dirty blocks from database buffer cache to datafiles but not "CURRENT log data to data files".So if your current online log group is full then ARCH process will try archiving this group to available destination and without archiving there will not happen LOG SWITCH.After archiving log switch will happen and LOG SEQUENCE NUMBER will increase then you will get new current log group.IF old group status is ACTIVE it means this group still need instance recovery.SO when you execute ALTER SYSTEM CHECKPOINT in this case status of this group will INACTIVE.
    *2)* My Archive logs size (25MB) is not equal to the redo log size. Normally , when we switch the logs or any rman backup , logs will be switched automatically and how much redo is filled , same will be generated with size.
    But, in my case , more archive logs are generating and many have the different size out of generated and no manual switching or rman script are running during this period.
    Any idea on this?.
    As you know archivelogs is a copy of online redologs,but there can be several reasons and result these size can be different.For example manual log switch or if you set ARCHIVE_LAG_TARGET != 0.Finally archive logs contains information for media recover,but online redo logs contain instance and media recovery it means these size can be different.

  • GoldenGate Extract Process will not read from redo log with manual help

    Here is my issue.
    I have GoldenGate replication successfully setup one-way from 1 Source to Many Targets. There is 1 source extract on the DB and many pumps that push the trail file data to the Targets. Replication does work but after manual help with starting up the Source Extract process.
    If I execute the command:
    GGSCI> alter extract <source extract name> begin now
    GGSCI> view report <source extract name>
    The extract starts and reads the source trail file but will not process data, I continually see in the ggserr.log file "OGG-01515: Positioning to begin time MMM DD, YYYY, HH:MM:SS" The date and time are irrevelant for this problem.
    When I see this, I SQ*Plus into the database and look in the v$log table for the current log and sequence #.
    I return to GGSCI and issue the following command:
    GGSCI> alter extract <source extract name> thread 1 extseqno <sequence # from v$log query>
    GGSCI> start <source extract name>
    It then works as expected. Why is this so? I thought the alter extract <source extract name> begin now would do the same output.
    We do use ASM but like I said when I issue the:
    GGSCI> alter extract <source extract name> thread 1 extseqno <sequence # from v$log query>
    It works like it should.
    Very weird.
    - Jason

    Yes, supplemental logging is enabled on both the source and the targets, but why would supplemental logging on the targets have any affect on why the Source Extract on the source can't read from the source redo log?
    This is not a RAC database, rather single-instance with one thread. Also, we are using DBLOGREADER functionality as it is an 11.2.0.3 database.
    My issue is simply, when I start the source extract from being down, meaning it isn't running, I issue this command:
    alter <source extract> begin now
    start <source extract>
    view report <source extract>
    OGG-01515 Positioning to begin time <today's date and time> ie Mar 4, 2013, 3:26:39 PM. (this is repeated over and over and over)
    If I perform a
    info <source extract> detail---> I see the following:
    Log Read Checkpoint Oracle Redo Logs 2013-03-04 15:26:39 Thread 1, Seqno 0, RBA 0 (why is it showing 0, becuase it can't read the redo, WHY NOT?)
    Extract Source BEGIN END
    Not Available <today's date> <today's date> (repeat....)
    However, if I retreive the Redo Log number and I issue:
    alter spe thread 1 extseqno (redo log sequence #)
    start spe.
    Then it works okay. I have to manually tell it what redo log to begin reading from. Why?
    - Jason
    Edited by: 924317 on Mar 4, 2013 9:03 AM

  • Redo Log Switch 결과...

    환경 : 8.1.7.3.0 (no archive log 모드)
    log_checkpoint_timeout = 0
    log_checkpoint_interval = 999999999
    redo log size = 200M
    현재 check point는 log switch 상태에서만 가능하도록 설정된 것 같습니다.
    거래량이 적어서 그런지 log switch는 30시간 주기입니다.
    제가 실행한 것은 아래 4번째 로그가 current일때
    alter system checkpoint를 하고 조금 있다가..
    alter system switch logfile를 하여 1번째가 current가 된 상황입니다.
    3/16일 14시 까지도 계속 active상태입니다...
    1. 문제가 생긴건지요??? 도움부탁합니다...
    2. no archive log mode에서도 switch 주기를 줄이는 것이 복구에 도움이 되나요...?
    ===========================================
    STATUS , FIRST_CHANGE#,FIRST_TIME
    CURRENT , 8846777646687,2007-03-15 16:57:55
    INACTIVE, 8846777587798,2007-03-14 10:34:40
    INACTIVE, 8846777609448,2007-03-14 17:17:38
    ACTIVE , 8846777643690,2007-03-15 16:01:22

    no archivemode에서 정상복구를 바라는 것인지요?
    잘못된 정책이란 생각이 듭니다.
    no archive mode에서 v$log의 first_change# 중에 가장 작은 것
    이 v$recover_file의 change# 보다 크거나 같다면 복구불능입니다.
    배치작업이라도 있어서 log switch가 한번의 cycle을 돌게 되면 이전
    백업으로는 복구불능입니다. archive mode로 지금 바로 바꾸시지요..
    log_checkpoint_timeout은 checkpoint에 대한 timeout 시간값을
    지정하는 것입니다.
    LOG_CHECKPOINT_INTERVAL specifies the frequency of checkpoints in terms of the number of redo log file blocks that can exist between an incremental checkpoint and the last block written to the redo log. This number refers to physical operating system blocks, not database blocks.
    checkpoint는 아시는 바와 같이 데이터파일과
    리두로그 컨트롤파일의 SCN을 일치시키는 것입니다. 주요한 것은
    DBWR프로세스가 데이터파일에 write를 하겠구요.
    물론 checkpoint와 인스턴스 복구는 관련이 있습니다. checkpoint timeout을
    적당히주면 instance recovery에서 좀 더 빠르게 instance 복구후 DB가
    open되겠습니다. 만약 설정하신대로 하신다면 DB를 abort로
    내리고 open하게 되면
    instance recovery시에 더 많은 시간이 필요하겠습니다.
    게다가 트랜잭션으로 인한 log switch하는 시간이 30시간보다
    작다면 timeout을 준들 영향을 주지 않겠지요. redo log가 꽉
    차게 되면 log switch를 자동으로 하게 되는데 log switch를
    하기전에 checkpoint를 주게 되어 있으니까요.
    그런데 checkpoint와 물리적/논리적 복구와는 다른 개념입니다.
    checkpoint는 위에서 말씀드린 instance recovery와 관련이 있고
    물리적/논리적 복구에서는 archive file이 떨어져 있는가 current redo log가
    존재하는가에 따라서 복구가능여부를 결정되는 것이지요..
    그리고 ACTIVE 상태라는 것은 문서상의 정의에서는 archive mode일 경우
    archiving이 되는 중일 경우, 그리고 이 상태는 complete recovery시에 redo log
    적용시 필요한 정보가 있다는 것입니다.
    no archive mode에서 복구정책을 적용 하겠다는 것은 위험한 발상인 것 같습니다.
    물론 DSS시스템의 경우에는 이미 정책을 no archive mode로
    만들어두고 주말마다 offline backup을 하기도 합니다.
    하지만 DSS에서는 하루에 300번 이상의 log switch가 일어나는
    경우가 있을 정도이니 아무리 백업이 되어 있다 한들 완전복구는
    불능이겠지요. offline backup을 했을 때까지만 복구가 됩니다.
    $LOG
    This view contains log file information from the control files.
    Column Datatype Description
    GROUP#
    NUMBER
    Log group number
    THREAD#
    NUMBER
    Log thread number
    SEQUENCE#
    NUMBER
    Log sequence number
    BYTES
    NUMBER
    Size of the log (in bytes)
    MEMBERS
    NUMBER
    Number of members in the log group
    ARCHIVED
    VARCHAR2(3)
    Archive status (YES |NO)
    STATUS
    VARCHAR2(16)
    Log status:
    UNUSED - Online redo log has never been written to. This is the state of a redo log that was just added, or just after a RESETLOGS, when it is not the current redo log.
    CURRENT - Current redo log. This implies that the redo log is active. The redo log could be open or closed.
    ACTIVE - Log is active but is not the current log. It is needed for crash recovery. It may be in use for block recovery. It might or might not be archived.
    CLEARING - Log is being re-created as an empty log after an ALTER DATABASE CLEAR LOGFILE statement. After the log is cleared, the status changes to UNUSED.
    CLEARING_CURRENT - Current log is being cleared of a closed thread. The log can stay in this status if there is some failure in the switch such as an I/O error writing the new log header.
    INACTIVE - Log is no longer needed for instance recovery. It may be in use for media recovery. It might or might not be archived.
    no archive mode에서도 복구하는 여러가지 방법들이 있기는 합니다.
    예를들어 current redo log가 깨졌을 때에 recovery 방법이라던지
    등등이 문서에 있긴하지요. 하지만 no archivemode에서 백업을 붓고
    복구하는 방법은 찾아보기 힘드실 것입니다. 위에서도 말씀드렸듯이
    no archive mode에서
    v$recover_file의 CHANGE# > v$logl의 minimum FIRST_CHANGE# 이면 데이터파일은 복구가능합니다.
    그러나 CHANGE# <= minimum FIRST_CHANGE# 이면 복구 불가능합니
    다. 그러니 백업을 붓고 복구를 하는 방법에 대한 문서는 거의
    찾기 힘듭니다. advance 방법에 대한 문서에서만 adjust_scn을 쓴다던지 하는 등이 나와있을 뿐입니다.
    글 수정:
    민천사(민연홍)
    아무래도 졸면서 썼나봅니다.;;
    interval과 timeout은 엄연히 다른데요. 왜 timeout과 interval을
    혼동했는지..;;
    LOG_CHECKPOINT_INTERVAL specifies the frequency of checkpoints in terms of the number of redo log file blocks that can exist between an incremental checkpoint and the last block written to the redo log. This number refers to physical operating system blocks, not database blocks.
    LOG_CHECKPOINT_TIMEOUT specifies (in seconds) the amount of time that has passed since the incremental checkpoint at the position where the last write to the redo log (sometimes called the tail of the log) occurred. This parameter also signifies that no buffer will remain dirty (in the cache) for more than integer seconds.

  • How to remove a redo log file

    Hi Experts,
    I want to remove a wrong redo log file from 10G R2 database in window.
    how to do that without loss data?
    my steps as
    1. alter system switch logfile;
    2. select * from v$log;
    which ARC and sataus do I can drop redo log file based on above SQL
    no archive and active status?
    ALso Which account should I do above action?
    fExamp, system account added redo log file, i only
    can drop by system? how about sys?
    Thanks help in detail steps
    Jim
    Edited by: user589812 on Dec 23, 2008 4:35 PM

    Jim,
    Check this link out for how to drop a redo log file
    Make sure a redo log group is archived (if archiving is enabled) before dropping it. To see whether this has happened, use the V$LOG view.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14231/onlineredo.htm#i1006489
    SELECT GROUP#, ARCHIVED, STATUS FROM V$LOG;
       GROUP# ARC STATUS
            1 YES ACTIVE
            2 NO  CURRENT
            3 YES INACTIVE
            4 YES INACTIVE
    Drop a redo log group with the SQL statement ALTER DATABASE with the DROP LOGFILE clause.
    The following statement drops redo log group number 3:
    ALTER DATABASE DROP LOGFILE GROUP 3;

Maybe you are looking for

  • Transporting user defined selection screen from one server to another

    How to transport a user defined selection screen (0100) from one server to another. PBO and PAI modules written in PBO and PAI events for the selection screen are not transported. How to transport the modules in the events for the selection screen.

  • Target file

    Hi, My scenario is a simple file-2-file scenario but want to create target file name based on element of the payload without any udf? thanks sai

  • How to use the readLine()  method when reading data from a URL?

    Hello, I have a URL which contains text input. The only way to get the data from this URL is by opening a URL connection to it. I would like to get the data from this URL but at the same time I would like to be able to use readLine() method of Buffer

  • Custom assembly & Report Caching

    Our reports are using a custom assembly which encapsulates some common functionality. The problem is that report caching stopped working since we've implemented this cahnge. We can still activate caching via "Manage processing options" --> "Use cache

  • Search function with Java script

    Hi, i'm using ADF 11g . In one of the application page , i need to give a search function , in which user can enter any text , the entered text needs to be highlighted . Since ADF is very new to me , could some one point me how to use the javascript