Goldengate extracting from Archive log

Dear All,
We are conducting a replication of Oracle Database. Due to high load on Production server, the client is providing us a near DR database server for extraction. This database server is using Oracle Data Guard to be in Sync with Production server. Hence we will be using Archive logs shipped from Production to DR to Extract changed data. I have searched on the Net and read manuals, but I am unable to understand the how to configure OGG for this scenario. I understand that we can used Tranlogoptions parameter to extract data from Archive logs, but since the near DR database is in Mount state, how can we Log in the database using userid /password in extract process. Also we need to have Select any dictionary privileges on the database, but database is in mount state, so how to work on it??
Awaiting eagerly for the reply.l

For a physical standby, you can configure GG to read the archived standby logs in archive log only (ALO) mode.
ARCHIVEDLOGONLY causes Extract to read from the archived logs exclusively, without querying or validating the logs from system views such as v$log and v$archived_log. This parameter puts Extract into Archived Log Only mode (ALO). For more information, see the Oracle GoldenGate Oracle Installation and Setup Guide.
You can configure the Extract process to read exclusively from the archived logs. This is
known as Archived Log Only (ALO) mode. In this mode, Extract only reads from archived
logs that are stored in a specified location. ALO mode allows Oracle GoldenGate to use
production logs that are shipped over to a secondary database (such as a standby) as the
data source for Oracle GoldenGate. The online logs will not be used. Oracle GoldenGate
will connect to the secondary database to get metadata and other required data as needed.
As an alternative, ALO mode is supported on the production system.
Extract automatically operates in ALO mode if it detects that the database is a physical standby.
Lots of other info in the Oracle install guide.

Similar Messages

  • Read data from Archive logs

    Does anyone have any recommandations on how to read data from archive logs. When i use log minor, i am getting only bind variables for DML operations. But i need actual data from the archive logs..
    Any thoughts
    Thanks
    -Prasad

    Log miner is the closest to command issued as possible. Depending on the Oracle version you will be able to see DML or DML and DDL. From 9i and on Oracle was able to translate the DML against data dictionary as the actual DDL command. On its first 8i release only DML was visible.
    ~ Madrid

  • GoldenGate extract from DB2 in rollforward pending mode

    Hi,
    Could anyone tell me whether it would be possible to use GoldenGate extract to extract transactions from a DB2 database in "rollforward pending mode"?
    Thanks!

    (answering this just to learn something about DB2)
    "By default, if archive logging was enabled for a database, when you issue the restore command, the database is left in a rollforward pending state. The database or tablespace in a rollforward pending state cannot be used until it is brought out of this state by applying the rollforward command."
    I would say no.

  • Recover database from archive log: Time based recovery

    Hi,
    Could you pelase help regarding the following:
    I have power outage in my machine running oracle 9i on Solaris OS 9
    Oracle is mounted but failed to open
    Once it is started showing the error message
    SQL> startup;
    ORACLE instance started.
    Total System Global Area 9457089744 bytes
    Fixed Size 744656 bytes
    Variable Size 3154116608 bytes
    Database Buffers 6291456000 bytes
    Redo Buffers 10772480 bytes
    Database mounted.
    ORA-01113: file 4 needs media recovery
    ORA-01110: data file 4: '/opt/oracle/oradata/sysdb/indx/indx01.dbf'
    So I tried to recover the database
    SQL> recover database;
    ORA-00279: change 1652948240 generated at 12/03/2007 13:09:08 needed for thread
    1
    ORA-00289: suggestion : /opt/oracle/oradata/nobilldb/archive/1_183942.dbf
    ORA-00280: change 1652948240 for thread 1 is in sequence #183942
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    ORA-00279: change 1652948816 generated at 12/03/2007 13:09:19 needed for thread
    1
    ORA-00289: suggestion : /opt/oracle/oradata/nobilldb/archive/1_183943.dbf
    ORA-00280: change 1652948816 for thread 1 is in sequence #183943
    ORA-00278: log file '/opt/oracle/oradata/nobilldb/archive/1_183942.dbf' no
    longer needed for this recovery
    The power outage at 16:00pm and the recovery archive log file '/opt/oracle/oradata/nobilldb/archive/1_183942.dbf' at 11 am
    Always I am applying the next sequence it is giving the same message and asking the next sequence. I have more than 900 archive log from 11am to 4pm and each of them having the size of 100mb and it take 1 minute to get back from each to provide this error message.
    How I can start my recovery from say 15:45 onwards until 16:15?
    I have all archive logs in the proper destination.
    Still my database is not opened and it is starts applying archive log since 5 hours back, please help me regarding this
    Thanks in advance

    Wrong forum. Post your question in the following forum:
    General Database Discussions

  • Update turbotax 2013 error can't extract from archive  on install and launch

    I installed my turbotax 2013. Then it said to update. I downloaded the update. Then it said install and relaunch. Up comes this message, "An error occurred while extracting the archive. Please try again later. No matter what I try. It comes up with same error message. I was online with intuit techs and they haven't solved it yet. Any ideas?

    First off, you are not seeing a kernel panic but an inability to load the kernel file "mach_kernel," which is damaged or missing.
    This is apparently because of a file system error (Invalid Key Length), a serious error Disk Utility is not designed to repair, since the repair usually involves loss of data & that app is very conservative about such things.
    It is almost impossible that the update itself is responsible for creating this error but very common for one to exacerbate it if it already exists. This is because file system errors corrupt files when the OS writes files to the disk, & updates write a lot of files to the disk.
    If you have implemented a backup strategy, the fix is straightforward, if a little time consuming: reformat the drive, do a new install of the OS from the installer disks, & restore your data from your backups.
    If you have not implemented a backup strategy, you now know why one is essential: no data you cannot afford to lose should reside only in one place like your internal hard drive. Data is lost every day because of human error, theft, hardware failure, or a host of other reasons. Only a good backup strategy can prevent this.
    Assuming the second situation applies to you, your best bet is to purchase a more powerful disk repair utility like DiskWarrior to see if it can repair the file system damage with minimal data loss. Be sure you read & heed the app's manual before using it! Failure to do so, or attempting to use the computer before the errors are repaired, will almost certainly result in more data loss.

  • Backup and Recovery from Archive Log

    Dear all,
    I implemented hot backup with oracle9i and
    Can anybody suggest me how to backup and recovery archive log ?
    Thanks for advance.
    Chara

    assuming you are using RMAN for backup -
    backup ( archivelog all );
    backup ( archivelog all delete input ); - to delete from file system.
    If you are not using RMAN but backup using a script which puts the tablespaces in hot backup mode then you just need to backup archive log directory.
    The archive logs are required when you need to do a database recovery. You don't recover archive log files.
    rgds

  • Restore from archive log

    Hi,
    I want to restore a record or a data that I have deleted accidentally from a table located in a database.
    Can you help me to do it?
    Thanks

    Querying Data at a Point in Time (Flashback Query)
    By default, operations on the database use the most recent committed data available. If you want to query the database as it was at some time in the past, you can do so with the flashback query feature. It lets you specify either a time or a system change number (SCN) and query using the committed data from the corresponding time.
    Some potential applications of flashback query are:
    Recovering lost data or undoing incorrect changes, even after the changes are committed. For example, a user who deletes or updates rows and then commits can immediately repair a mistake.
    Comparing current data against the data at some time in the past. For example, you might run a daily report that shows the change from yesterday, rather than just the current data.
    Checking the state of transactional data at a particular time. For example, you might want to verify an account balance on a certain day.
    Simplifying application design by removing the need to store some kinds of temporal data.
    Enabling packaged applications, such as report generation tools, to work on past versions of data.
    The flashback query mechanism is most effective when you use automatic undo management. The DBA requests that undo data be kept for a specified period of time. Depending on the available storage capacity, the database might not always be able to keep all the requested undo data. If you use flashback queries, you might need to familiarize yourself with automatic undo management to understand its capabilities and limitations.
    Other features are available to recover lost data. The unique feature of flashback query is that you can see the data as it was in the past, then choose exactly how to process the information; you might do an analysis, undoing the changes, or capture changed data for further processing.
    See Also:
    Oracle9i Database Concepts for background information about flashback query and automatic undo management.
    Oracle9i Database Administrator's Guide for the associated DBA responsibilities, such as setting up automatic undo management and granting privileges.
    Oracle9i Supplied PL/SQL Packages and Types Reference for details about the DBMS_FLASHBACK package that you can use in combination with flashback queries.
    Oracle9i Database Reference and Oracle9i Recovery Manager User's Guide and Reference for information about system change numbers, which are used in backup and recovery and are present in some V$ data dictionary views.
    http://download-west.oracle.com/docs/cd/B10501_01/appdev.920/a96590/adg08sql.htm#10851
    Joel Pérez
    http://otn.oracle.com/experts

  • Goldengate Extracts reads slow during Table Data Archiving and Index Rebuilding Operations.

    We have configured OGG on a  near-DR server. The extracts are configured to work in ALO Mode.
    During the day, extracts work as expected and are in sync. But during any dialy maintenance task, the extracts starts lagging, and read the same archives very slow.
    This usually happens during Table Data Archiving (DELETE from prod tables, INSERT into history tables) and during Index Rebuilding on those tables.
    Points to be noted:
    1) The Tables on which Archiving is done and whose Indexes are rebuilt are not captured by GoldenGate Extract.
    2) The extracts are configured to capture DML opeartions. Only INSERT and UPDATE operations are captured, DELETES are ignored by the extracts. Also DDL extraction is not configured.
    3) There is no connection to PROD or DR Database
    4) System functions normally all the time, but just during table data archiving and index rebuild it starts lagging.
    Q 1. As mentioned above, even though the tables are not a part of capture, the extracts lags ? What are the possible reasons for the lag ?
    Q 2. I understand that Index Rebuild is a DDL operation, then too it induces a lag into the system. how ?
    Q 3. We have been trying to find a way to overcome the lag, which ideally shouldn't have arised. Is there any extract parameter or some work around for this situation ?

    Hi Nick.W,
    The amount of redo logs generated is huge. Approximately 200-250 GB in 45-60 minutes.
    I agree that the extract has to parse the extra object-id's. During the day, there is a redo switch every 2-3 minutes. The source is a 3-Node RAC. So approximately, 80-90 archives generated in an hour.
    The reason to mention this was, that while reading these archives also, the extract would be parsing extra Object ID's, as we are capturing data only for 3 tables. The effect of parsing extract object id's should have been seen during the day also. The reason being archive size is same, amount of data is same, the number of records to be scanned is same.
    The extract slows down and read at half the speed. If normally it would take 45-50 secs to read an archive log of normal day functioning, then it would take approx 90-100 secs to read the archives of the mentioned activities.
    Regarding the 3rd point,
    a. The extract is a classic extract, the archived logs are on local file system. No ASM, NO SAN/NAS.
    b. We have added  "TRANLOGOPTIONS BUFSIZE" parameter in our extract. We'll update as soon as we see any kind of improvements.

  • Disabling Log_Archive_Start in Archive Log Mode

    Guys,
    I have a Database running in Archive Log mode.
    I have a situation where in I have to do bunch of bulk updates for which I dont want any logs generated.
    sO I was thinking to set the intialization parameter log_archive_start to false before I run the DML updates
    ALTER SYSTEM SET log_archive_start = false
    after the DML is run I will turn it back on
    ALTER SYSTEM SET log_archive_start = True
    Do you think if this is the right to go about or would I have to switch from Archive Log Mode to No Archive Log Mode
    run the DML update and then switch back to Archive Log Mode
    Any comments/suggestions/inputs woo

    Hello,
    I would use some other way to get a full nologing behaviour.
    First, ensure your tablespace is in force logging mode.
    ALTER TABLESPACE data FORCE LOGGING NOLOGGING;
    ALTER TABLESPACE index FORCE LOGGING NOLOGGING;
    This will ensure that you're usually logging.
    Set all tables/indexes to be nologging
    BEGIN
    FOR vRes IN (SELECT OWNER,TABLE_NAME FROM DBA_TABLES WHERE TABLESPACE_NAME='data')
    LOOP
    EXECUTE IMMEDIATE 'ALTER TABLE ' || vRes.OWNER || '.' || vRes.TABLE_NAME ' NOLOGGING';
    END LOOP;
    END;
    Do even for indexes.
    From now on, when you want to do your huge amount of updates/... targetting objects in tablespaces data and index, jou just have to:
    ALTER TABLESPACE data NO FORCE LOGGING;
    ALTER TABLESPACE indexes NO FORCE LOGGING;
    - Your DML
    ALTER TABLESPACE data FORCE LOGGING;
    ALTER TABLESPACE indexes FORCE LOGGING.
    4 commands to disable logging for every object in the tablespace.
    Beware, when in NOLOGGING mode, you won't be able to "RECOVER" any of the tables. Thiis is not crash-friendly, but you'll have no redo info generated, which is what you require.
    Regards,
    Yoann.

  • Powershell script to search specific event ID's from Archive Event files

    Hi,
    I have to search some Windows Event ID's from Archive event (evtx) files. for that i am trying powershell code mentioned on below link but its not working. Kindly suggest how to search specific event id from archive logs.
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/83cbf185-bd96-4efc-b36b-59439024960a/parse-message-property-from-archived-eventlog
    Regards,
    Dhiraj

    Hi,
    Thanks for your response. I am using below code & getting below message. I changed the event ID & got some results. Can we remove those lines form results where record not found. like how can we get only those records which match given
    event ID. also it's not showing all field values. Is there any formating issue.
    $SearchID='4740' 
    get-childitem "C:\Windows\System32\Winevt\Logs\Archive-Security*.evtx" | select FullName | forEach{
     %{get-winevent  -filterHashTable @{path=$_.Fullname;ID=$SearchID} |                
        ? {$_.Properties[5].Value -match $UserName} |   
        Select-Object -Property TimeCreated, `
                                @{Name='SecurityId';Expression={$_.Properties[4].Value}}, `
                                @{Name='AccountName';Expression={$_.Properties[5].Value}}, `
                                @{Name='AccountDomain';Expression={$_.Properties[6].Value}}, `
                                @{Name='LogonId';Expression={$_.Properties[7].Value}}, `
                                @{Name='LogonType';Expression={$_.Properties[8].Value}}, `
                                @{Name='Workstation';Expression={$_.Properties[11].Value}}, `
                                @{Name='LogonGuid';Expression={$_.Properties[12].Value}} |
        Format-Table
    Get-WinEvent : No events were found that match the specified selection criteria.
    At C:\Users\sidadmin1\Documents\EventIDCheck.ps1:3 char:16
    +  %{get-winevent <<<<   -filterHashTable @{path=$_.Fullname;ID=$SearchID} |
        + CategoryInfo          : ObjectNotFound: (:) [Get-WinEvent], Exception
        + FullyQualifiedErrorId : NoMatchingEventsFound,Microsoft.PowerShell.Commands.GetWinEventCommand
    TimeCreated    SecurityId     AccountName     AccountDomain LogonId        LogonType      Workstation    LogonGuid
    3/11/2014 4... DC04$          MyDomain                 999
    Get-WinEvent : No events were found that match the specified selection criteria.
    At C:\Users\sidadmin1\Documents\EventIDCheck.ps1:3 char:16
    +  %{get-winevent <<<<   -filterHashTable @{path=$_.Fullname;ID=$SearchID} |
        + CategoryInfo          : ObjectNotFound: (:) [Get-WinEvent], Exception
        + FullyQualifiedErrorId : NoMatchingEventsFound,Microsoft.PowerShell.Commands.GetWinEventCommand
    Dhiraj

  • GoldenGate Extract Process will not read from redo log with manual help

    Here is my issue.
    I have GoldenGate replication successfully setup one-way from 1 Source to Many Targets. There is 1 source extract on the DB and many pumps that push the trail file data to the Targets. Replication does work but after manual help with starting up the Source Extract process.
    If I execute the command:
    GGSCI> alter extract <source extract name> begin now
    GGSCI> view report <source extract name>
    The extract starts and reads the source trail file but will not process data, I continually see in the ggserr.log file "OGG-01515: Positioning to begin time MMM DD, YYYY, HH:MM:SS" The date and time are irrevelant for this problem.
    When I see this, I SQ*Plus into the database and look in the v$log table for the current log and sequence #.
    I return to GGSCI and issue the following command:
    GGSCI> alter extract <source extract name> thread 1 extseqno <sequence # from v$log query>
    GGSCI> start <source extract name>
    It then works as expected. Why is this so? I thought the alter extract <source extract name> begin now would do the same output.
    We do use ASM but like I said when I issue the:
    GGSCI> alter extract <source extract name> thread 1 extseqno <sequence # from v$log query>
    It works like it should.
    Very weird.
    - Jason

    Yes, supplemental logging is enabled on both the source and the targets, but why would supplemental logging on the targets have any affect on why the Source Extract on the source can't read from the source redo log?
    This is not a RAC database, rather single-instance with one thread. Also, we are using DBLOGREADER functionality as it is an 11.2.0.3 database.
    My issue is simply, when I start the source extract from being down, meaning it isn't running, I issue this command:
    alter <source extract> begin now
    start <source extract>
    view report <source extract>
    OGG-01515 Positioning to begin time <today's date and time> ie Mar 4, 2013, 3:26:39 PM. (this is repeated over and over and over)
    If I perform a
    info <source extract> detail---> I see the following:
    Log Read Checkpoint Oracle Redo Logs 2013-03-04 15:26:39 Thread 1, Seqno 0, RBA 0 (why is it showing 0, becuase it can't read the redo, WHY NOT?)
    Extract Source BEGIN END
    Not Available <today's date> <today's date> (repeat....)
    However, if I retreive the Redo Log number and I issue:
    alter spe thread 1 extseqno (redo log sequence #)
    start spe.
    Then it works okay. I have to manually tell it what redo log to begin reading from. Why?
    - Jason
    Edited by: 924317 on Mar 4, 2013 9:03 AM

  • How do I setup RMAN not to delete archive log files on the source database so GoldenGate can process DDL/DML changes?

    I want to setup RMAN not to delete any archive log files that will be used by GoldenGate.   Once GoldenGate is completed with the archive log file, the archive log file can be backup and deleted by RMAN.   It's my understanding that I can issue the following command "REGISTER EXTRACT <ext_name>, LOGRETENTION" to enable to functionally.   Is this the only thing I need to do to execute to enable this functionally?

    Hello,
    Yes this is the rigth way  using clasic capture.
    Using the command : REGISTER EXTRACT Extract_name LOGRETENTION.
    Create a Oracle Streams Group Capture (Artificial)  that prevent RMAN archive deletion if these are pending to process for Golden Gate capture process.
    You can see this integration doing a SELECT * FROM DBA_CAPTURE; after execute the register command.
    Then, when RMAN try to delete a archive file pending to process for GG this warning appear AT RMAN logs:
    Error:     RMAN 8317 (RMAN-08317 RMAN-8317)
    Text:     WARNING: archived log not deleted, needed for standby or upstream capture process.
    Then , this is a good manageability feature. I think is a 11.1 GG new feature.
    Tip. To avoid RMAN backup multiples times a archive pending to process, there is a option called BACKUP archivelog not backed 1 times.
    If you remove a Capture process that is registered with the database you need to use this comand to remove the streams capture group:
    unREGISTER EXTRACT extract_name LOGRETENTION;
    Then if you query dba_capture, the artificial Streams group is deleted.
    I hope help.
    Regards
    Arturo

  • Extract is abended and looking 8 days old archive log.

    I was doing some maintenance and I stop all the extracts and replicats. Couple of my extracts are stopped and looking for 8 days old archive logs. Please tell me how can I recover from that situation. I dont have that archive any more we only keep 5days arch.
    GGSCI (ggastdb01.infra.mms) 5> info e_it860t
    EXTRACT E_IT860T Initialized 2011-02-08 16:04 Status STOPPED
    Checkpoint Lag 00:10:05 (updated 04:34:30 ago)
    Log Read Checkpoint Oracle Redo Logs
    2011-02-24 09:53:04 Seqno 15721, RBA 103192
    Report file Error:
    2011-02-24 14:34:04 ERROR OGG-00446 Could not find archived log for sequence 15494 thread 1 under alternative or default destinations. SQL <SELECT name FROM v$archived_log WHERE sequence# = :ora_seq_no AND thread# = :ora_thread AND resetlogs_id = :ora_resetlog_id AND archived = 'YES' AND deleted = 'NO>. Last alternative log tried /u10/oradata/PPMTST10/arch/1_15494_721133148.arc, error retrieving redo file name for sequence 15494, archived = 1, use_alternate = 0Not able to establish initial position for sequence 15494, rba 68728848.
    2011-02-24 14:34:04 ERROR OGG-01668 PROCESS ABENDING.

    If you don't have the log file(s) in question, then you need to:
    re-sync the databases/schema/tables, or
    advance the CSN to a point in time where you do have log files and deal with missing data separately
    Is an archived transaction log missing?
    ● Extract reads the archived logs when an online log is not available. If you started
    Extract after it was stopped for a long time, the records it needs to process might
    already be archived away.
    ● Restore the archive log that is specified in the error message in the report file, and also
    restore all subsequent logs.
    NOTE Extract should not be stopped for a long time. If you must stop Extract, make
    certain that archives are available all the way back to the time that you stopped
    Extract. If Extract remains stopped for an exceptionally long time, it might be more
    practical to resynchronize the table(s) rather than wait for Extract to go through the
    archives and catch up.

  • Changing the location of archive log from flash recovery area PLZ HELP!!!

    Hi All,
    My archive log is being stored in flash memory area which got full and the production server went down.
    alert log file details.....
    ORA-19809: limit exceeded for recovery files
    ORA-19804: cannot reclaim 43432960 bytes disk space from 2147483648 limit
    *** 2010-04-25 14:22:49.777 62692 kcrr.c
    ARCH: Error 19809 Creating archive log file to
    '/oracle/product/10.2.0/flash_rec
    overy_area/EDWREP/archivelog/2010_04_25/o1_mf_1_232_%u_.arc'
    *** 2010-04-25 14:22:49.777 60970 kcrr.c
    kcrrfail: dest:10 err:19809 force:0 blast:1I removed the files and started the database,
    Can someone kindly tell me as to how to avoid this problem in future by keeping archive log destination in flash recovery area.
    I want to change the location of archive log files, can someone please guide me as to hiow to do that
    I changed the size of flash recovery area for the time being, but i am afraid it will be full again!!
    SQL> select * from v$flash_recovery_area_usage;
    FILE_TYPE    PERCENT_SPACE_USED PERCENT_SPACE_RECLAIMABLE NUMBER_OF_FILES
    CONTROLFILE                   0                         0               0
    ONLINELOG                     0                         0               0
    ARCHIVELOG                99.44                         0              57
    BACKUPPIECE                   0                         0               0
    IMAGECOPY                     0                         0               0
    FLASHBACKLOG                  0                         0               0
    6 rows selected.
    SQL> alter system set DB_RECOVERY_FILE_DEST_SIZE = 4G ;
    System altered.
    SQL> select * from v$flash_recovery_area_usage;
    FILE_TYPE    PERCENT_SPACE_USED PERCENT_SPACE_RECLAIMABLE NUMBER_OF_FILES
    CONTROLFILE                   0                         0               0
    ONLINELOG                     0                         0               0
    ARCHIVELOG                49.72                         0              57
    BACKUPPIECE                   0                         0               0
    IMAGECOPY                     0                         0               0
    FLASHBACKLOG                  0                         0               0
    6 rows selected.regards,
    Edited by: user10243788 on Apr 25, 2010 6:12 AM

    user10243788 wrote:
    Hi All,
    My archive log is being stored in flash memory area which got full and the production server went down.
    alert log file details.....
    ORA-19809: limit exceeded for recovery files
    ORA-19804: cannot reclaim 43432960 bytes disk space from 2147483648 limit
    *** 2010-04-25 14:22:49.777 62692 kcrr.c
    ARCH: Error 19809 Creating archive log file to
    '/oracle/product/10.2.0/flash_rec
    overy_area/EDWREP/archivelog/2010_04_25/o1_mf_1_232_%u_.arc'
    *** 2010-04-25 14:22:49.777 60970 kcrr.c
    kcrrfail: dest:10 err:19809 force:0 blast:1I removed the files and started the database,
    Can someone kindly tell me as to how to avoid this problem in future by keeping archive log destination in flash recovery area.
    I want to change the location of archive log files, can someone please guide me as to hiow to do that
    I changed the size of flash recovery area for the time being, but i am afraid it will be full again!!
    SQL> select * from v$flash_recovery_area_usage;
    FILE_TYPE    PERCENT_SPACE_USED PERCENT_SPACE_RECLAIMABLE NUMBER_OF_FILES
    CONTROLFILE                   0                         0               0
    ONLINELOG                     0                         0               0
    ARCHIVELOG                99.44                         0              57
    BACKUPPIECE                   0                         0               0
    IMAGECOPY                     0                         0               0
    FLASHBACKLOG                  0                         0               0
    6 rows selected.
    SQL> alter system set DB_RECOVERY_FILE_DEST_SIZE = 4G ;
    System altered.
    SQL> select * from v$flash_recovery_area_usage;
    FILE_TYPE    PERCENT_SPACE_USED PERCENT_SPACE_RECLAIMABLE NUMBER_OF_FILES
    CONTROLFILE                   0                         0               0
    ONLINELOG                     0                         0               0
    ARCHIVELOG                49.72                         0              57
    BACKUPPIECE                   0                         0               0
    IMAGECOPY                     0                         0               0
    FLASHBACKLOG                  0                         0               0
    6 rows selected.regards,
    Edited by: user10243788 on Apr 25, 2010 6:12 AMPointing the archive log dest (and/or the FRA) to a new location, or enlarging them, will do no good if you are not performing regular housekeeping on the archivelogs. You will just keep knocking down the same problem over and over.
    If you simply delete the archivelogs at the OS level, the database will never know about it and it will continue to think the destination is full, based on records kept in the control file.
    For regular housekeeping, you need to be doing something similar to this in rman:
    run {
      backup archivelog all not backed up 1 times tag='bkup_vlnxora1_arch';
      delete noprompt archivelog all backed up 1 times to device type disk;
    run {
    delete noprompt obsolete;
    crosscheck archivelog all;
    delete noprompt expired archivelog all;

  • Recover from currnet log file corruption in archive mode

    Hi,
    I am using oracle 10.2.0 on windows xp. my database is in archivelog mode. I am learning recovery. I take backup of database using
    RMAN> backup database;
    Starting backup at 03-SEP-09
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting full datafile backupset
    channel ORA_DISK_1: specifying datafile(s) in backupset
    input datafile fno=00004 name=D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
    input datafile fno=00001 name=D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSTEM01.DBF
    input datafile fno=00003 name=D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSAUX01.DBF
    input datafile fno=00002 name=D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\UNDOTBS01.DBF
    channel ORA_DISK_1: starting piece 1 at 03-SEP-09which completed.
    my current logfile and current sequence no is
    SQL> select * from v$log;
        GROUP#    THREAD#  SEQUENCE#      BYTES    MEMBERS ARC STATUS
    FIRST_CHANGE# FIRST_TIM
             1          1         88   52428800          1 NO  CURRENT
          4475270 01-SEP-09
             2          1         86   52428800          1 YES INACTIVE
          4426127 31-AUG-09
             3          1         87   52428800          1 YES INACTIVEThen i delete my only one redolog file in my current group. my instance crashes i want to recover now first i try
    SQL> startup
    ORACLE instance started.
    Total System Global Area  289406976 bytes
    Fixed Size                  1248576 bytes
    Variable Size              92275392 bytes
    Database Buffers          188743680 bytes
    Redo Buffers                7139328 bytes
    Database mounted.
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1:
    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO01.LOG'
    ORA-27041: unable to open file
    OSD-04002: unable to open file
    O/S-Error: (OS 2) The system cannot find the file specified.
    SQL> select open_mode from v$database;
    OPEN_MODE
    MOUNTED
    SQL> recover database until cancel;
    ORA-00279: change 4500421 generated at 09/03/2009 10:49:48 needed for thread 1
    ORA-00289: suggestion : D:\1\ARC00088_0693601900.001
    ORA-00280: change 4500421 for thread 1 is in sequence #88
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    cancel
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01194: file 1 needs more recovery to be consistent
    ORA-01110: data file 1: 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSTEM01.DBF'
    ORA-01112: media recovery not startedit doesn't work then i do
    RMAN> restore database
    2> ;
    Starting restore at 03-SEP-09
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: sid=156 devtype=DISK
    channel ORA_DISK_1: starting datafile backupset restore
    channel ORA_DISK_1: specifying datafile(s) to restore from backup set
    restoring datafile 00001 to D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSTEM01.DBF
    restoring datafile 00002 to D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\UNDOTBS01.DBF
    restoring datafile 00003 to D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSAUX01.DBF
    restoring datafile 00004 to D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
    channel ORA_DISK_1: reading from backup piece D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03
    KP
    channel ORA_DISK_1: restored backup piece 1
    piece handle=D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03\O1_MF_NNNDF_TAG20090903T104947_5
    channel ORA_DISK_1: restore complete, elapsed time: 00:02:16
    Finished restore at 03-SEP-09
    RMAN>
    RMAN> recover database until sequence 87;
    Starting recover at 03-SEP-09
    using channel ORA_DISK_1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of recover command at 09/03/2009 11:06:06
    RMAN-06556: datafile 1 must be restored from backup older than scn 4444743It means i need an older backup but i have new backup how to restore or get my database up.
    Thanks
    umesh

    Thanks for reply
    i have this in my hand
    RMAN> list backupset;
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    1       Full    1.38G      DISK        00:02:21     03-SEP-09
            BP Key: 1   Status: AVAILABLE  Compressed: NO  Tag: TAG20090903T104947
            Piece Name: D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03\O1_MF_NNNDF_TAG20090903T104947_59YNHNSK_.BKP
      List of Datafiles in backup set 1
      File LV Type Ckp SCN    Ckp Time  Name
      1       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSTEM01.DBF
      2       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\UNDOTBS01.DBF
      3       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSAUX01.DBF
      4       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    2       Full    6.80M      DISK        00:00:03     03-SEP-09
            BP Key: 2   Status: AVAILABLE  Compressed: NO  Tag: TAG20090903T104947
            Piece Name: D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03\O1_MF_NCSNF_TAG20090903T104947_59YNN8J9_.BKP
      Control File Included: Ckp SCN: 4500486      Ckp time: 03-SEP-09
      SPFILE Included: Modification time: 03-SEP-09
    RMAN> restore database preview;
    Starting restore at 03-SEP-09
    using channel ORA_DISK_1
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    1       Full    1.38G      DISK        00:02:21     03-SEP-09
            BP Key: 1   Status: AVAILABLE  Compressed: NO  Tag: TAG20090903T104947
            Piece Name: D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03\O1_MF_NNNDF_TAG20090903T104947_59YNHNSK_.BKP
      List of Datafiles in backup set 1
      File LV Type Ckp SCN    Ckp Time  Name
      1       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSTEM01.DBF
      2       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\UNDOTBS01.DBF
      3       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSAUX01.DBF
      4       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
    archive logs generated after SCN 4500421 not found in repository
    Media recovery start SCN is 4500421
    Recovery must be done beyond SCN 4500421 to clear data files fuzziness
    Finished restore at 03-SEP-09
    RMAN> backup archivelog all ;
    Starting backup at 03-SEP-09
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archive log backupset
    channel ORA_DISK_1: specifying archive log(s) in backup set
    input archive log thread=1 sequence=60 recid=3 stamp=694522772
    input archive log thread=1 sequence=61 recid=4 stamp=694543115
    input archive log thread=1 sequence=62 recid=5 stamp=694693173
    input archive log thread=1 sequence=63 recid=6 stamp=694707783
    input archive log thread=1 sequence=64 recid=7 stamp=694785054
    input archive log thread=1 sequence=65 recid=8 stamp=694787797
    input archive log thread=1 sequence=66 recid=9 stamp=694794198
    input archive log thread=1 sequence=67 recid=10 stamp=694799765
    input archive log thread=1 sequence=68 recid=11 stamp=694806827
    input archive log thread=1 sequence=69 recid=12 stamp=694809124
    input archive log thread=1 sequence=70 recid=13 stamp=694868744
    input archive log thread=1 sequence=71 recid=14 stamp=695132728
    input archive log thread=1 sequence=72 recid=15 stamp=695210729
    input archive log thread=1 sequence=73 recid=16 stamp=695330401
    input archive log thread=1 sequence=74 recid=17 stamp=695399695
    input archive log thread=1 sequence=75 recid=18 stamp=695412690
    input archive log thread=1 sequence=76 recid=19 stamp=695511043
    input archive log thread=1 sequence=77 recid=20 stamp=695557945
    input archive log thread=1 sequence=78 recid=21 stamp=695597432
    input archive log thread=1 sequence=79 recid=22 stamp=695728451
    input archive log thread=1 sequence=80 recid=23 stamp=695729334
    input archive log thread=1 sequence=81 recid=24 stamp=695729976
    input archive log thread=1 sequence=82 recid=25 stamp=695835471
    input archive log thread=1 sequence=83 recid=26 stamp=695988878
    input archive log thread=1 sequence=84 recid=27 stamp=696160938
    input archive log thread=1 sequence=85 recid=28 stamp=696333853
    input archive log thread=1 sequence=86 recid=29 stamp=696360658
    input archive log thread=1 sequence=87 recid=30 stamp=696454218
    channel ORA_DISK_1: starting piece 1 at 03-SEP-09
    channel ORA_DISK_1: finished piece 1 at 03-SEP-09
    piece handle=D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03\O1_MF_ANNNN_TAG20090903T115246_59YR5VRV_.BKP tag=TAG20090903T1
    =NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:01:08
    Finished backup at 03-SEP-09
    RMAN> restore database preview;
    Starting restore at 03-SEP-09
    using channel ORA_DISK_1
    List of Backup Sets
    ===================
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    1       Full    1.38G      DISK        00:02:21     03-SEP-09
            BP Key: 1   Status: AVAILABLE  Compressed: NO  Tag: TAG20090903T104947
            Piece Name: D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_09_03\O1_MF_NNNDF_TAG20090903T104947_59YNHNSK_.BKP
      List of Datafiles in backup set 1
      File LV Type Ckp SCN    Ckp Time  Name
      1       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSTEM01.DBF
      2       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\UNDOTBS01.DBF
      3       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\SYSAUX01.DBF
      4       Full 4500421    03-SEP-09 D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
    archive logs generated after SCN 4500421 not found in repository
    Media recovery start SCN is 4500421
    Recovery must be done beyond SCN 4500421 to clear data files fuzziness
    Finished restore at 03-SEP-09now i have two backup one is full database backup which will not work another is backup of all archivelog upto sequence 87 i think it should work.
    but rman is choosing full database backup for restore.
    so i need to force rman to choose archivelog backups for restore but don't know how to do it. please suggest.
    Thanks
    umesh

Maybe you are looking for

  • CS5 64-Bit applications crashing on Basic File Operations

    Hi all, I am having a big problem running photoshop 64-bit (32-bit works ok), it opens ok but when you go to file and click either Open or Save the application just closes with no error report. I have been on quite a few forums about this, common ans

  • Webpage displays correctly only in safe mode

    A pop up window on ebay does not show correctly in Firefox. It shows correctly (fully displaying all information) in Safari and Chrome. I reset Firefox, still same issue. The weird thing is, it displays properly in Safe Mode. I noticed that once the

  • Maximum length of field in OPEN DATASET

    Hello, I try to import a file from server (for test purposes downloaded on the server directory with transaction CG3Z in ASC) with an unknown length of lines. Here is my coding for the import into ABAP: OPEN DATASET p_pa_snam FOR INPUT IN TEXT MODE .

  • Upgrading from Premiere Pro CS5.5 to CS6

    I have Premiere Pro CS5.5 as part of Adobe Creative Suite 5.5 Production Premium.  I just purchased Premiere Pro CS6 upgrade but when I try to register I select Premiere Pro CS5.5 as my qualifying product upgrade and enter the SN from Creative Suite

  • Grand Total a Percentage

    So I am calculating the % change between two Quarters with the following formula: ((Current Qtr / Prior Qtr) -1) *100 my rows are calculating correctly but my column's grand total does a sum and it isn't correct. How can I get my grand total to calcu