Refresh Materilaized Views without generating Archive Logs

Hello Gurus,
I'm facing an emebarassing situation,
I've a job every night to execute a dbms_refresh on several materialiled Views, but there are 2 Mviews that take almost 30 minutes each but another issue thay generates almost 10G of data and my File system becomes full in a minute.
I've done an alter Mview Nologging but it doesn't change anything.
Does soemon ahve an idea how to do a refresh without generating Archive logs - or
without logging?
thxs in advance
C.

Since this is a TRUNCATE and INSERT, any user / application querying the MV while the Refresh is running will see ZERO rows. Furthermore if the INSERT fails (e.g. insufficent space to add an extent), the MV remains with ZERO rows.
A DELETE and INSERT avoids such situations (even if the refresh fails, both INSERT and DELETE are rolled-back and the MV is reverted to the state as it was before the Refresh began -- thus showing data from the previous refresh)
Hemant K Chitale
http://hemantoracledba.blogspot.com
Just to clarify : My paragraphs above are not about data "missing" but about precautions to be taken. A COMPLETE Refresh is done during defined outages -- users/applications are made aware that data is not available while the Refresh is running.
Edited by: Hemant K Chitale on Sep 1, 2009 10:53 AM

Similar Messages

  • Generate archive logs are not in sequence number?

    On last friday... the latest archive log number was ARC00024.ARC. Tomorrow when I come backup, the archive logs ARC00001.ARC and ARC00002.ARC were being generated by oracle itself. I wondering the archive log sequence should be in sequence. What is happening?
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination C:\oracle\ora92\RDBMS
    Oldest online log sequence 1
    Next log sequence to archive 3
    Current log sequence 3
    SQL>
    FAN
    Edited by: user623471 on Jun 7, 2009 7:35 PM

    khurram,
    Its our production instance and havent issued resetlogs option but when listing the arvchives it shows in different sequence number...
    and also while copying the archives by RMAN it doesnt copy in sequence
    -rw-r----- 1 xxx dba 69363859 May 28 19:16 2_10373.arc.gz
    -rw-r----- 1 xxx dba 43446622 May 28 19:16 1_10553.arc.gz
    -rw-r----- 1 xxx dba 52587365 May 28 19:16 1_10578.arc.gz
    -rw-r----- 1 xxx dba 45251820 May 28 19:16 1_10543.arc.gz
    -rw-r----- 1 xxx dba 60890256 May 28 19:17 1_10579.arc.gz
    -rw-r----- 1 xxx dba 46659008 May 28 19:17 1_10548.arc.gz
    -rw-r----- 1 xxx dba 116899466 May 28 19:17 2_10353.arc.gz
    -rw-r----- 1 xxx dba 77769517 May 28 19:17 1_10531.arc.gz
    -rw-r----- 1 xxx dba 66401923 May 28 19:18 1_10530.arc.gz
    -rw-r----- 1 xxx dba 45972697 May 28 19:18 1_10605.arc.gz
    -rw-r----- 1 xxx dba 55082543 May 28 19:18 1_10600.arc.gz
    -rw-r----- 1 xxxq dba 42682207 May 28 19:19 1_10547.arc.gz
    thanks,
    baskar.l

  • Refresh a view without losing changes?

    Hi,
    Is there a way to refresh a view to get the new rows added (and commited) using another transaction but without losing the rows added in the current transaction? (update keeping the dirty stuff)
    Thank you!

    I think it would be possible with polling.
    You will have to poll the server (with a poll component), attach it as a partial trigger to your table so that it refresh its model on the server.
    Also, you will have to check some settings on the pageDef definition of your iterator so that data gets always queried (avoiding cache).
    I would be interested by what this would do to your current data however...
    Good luck!

  • Delete large partition generates archive log?

    Hi!
    I have a partition 3 partitions with 100 milj rows in each partition. If I drop a partition, does it generats 300milj logs in archive or just 3?
    /Andreas

    Deletes perform normal DML: take locks on rows, generate redo (lots of it), If a mistake is made a rollback can be issued to restore the records prior to a commit. A delete does not relinquish segment space thus a table in which all records have been deleted retains all of its original blocks.
    Truncate is DDL, it moves the High Water Mark of the table back to zero. No row-level locks are taken, no redo or rollback is generated. All extents bar the initial are de-allocated from the table As truncate is a DDL command, you can't roll it back.
    in your case, if You need to permanenty delete all data in one partiotion, then use truncate or drop partition. It will not generate redo.

  • Oracle Restore from Online backup without any archive logs

    Hi,
    May be a dumb question. If I have a good online backup (say it took 2 hours to do that. And there is database activity while backup is going on), and lost all the archive logs happened after the online backup, is it possible at all to do the restore using that backup? complete or incomplete? and bring back the database to normal operation mode? Some details on this.
    Thanks.

    Let us see the reasoning behind this.
    Database:WORLDDB
    WORLDDB configuration:-
    TBSP_WDB
    ->TBSP_WDB_01.dbf
    ->TBSP_WDB_02.dbf
    SYSTEM
    ->SYSTEM.dbf
    USERS
    and so on for the tablespaces.
    CASE 1:-
    User managed backup.
    You issue
    ALTER TABLESPACE TBSP_WDB BEGIN BACKUP;
    DB keeps working all transactions are recorded in the redo stream, the scn information is not updated in the file header.
    Also note that the other tablespaces were been continuously worked on hence their scn numbers are a moving target.
    http://download-east.oracle.com/docs/cd/B10501_01/server.920/a96572/osbackups.htm#10012
    So if you begin tablespace TBSP_WDB backup at time t1 and by the time you are done with your backups you are at time t2, you will need all of the archive log's between time t1 and t2.
    CASE 2:-
    If we do with rman, rman takes care of the file header update information and goes on its merry way, those details are hidden from the end user. If one of the tablespace is fractured, updated while backups are going on , rman knows about it and will re-read the affected blocks in question.
    My thinking would be that you might need archive logs, unless it is a cold backup, with switch logfiles in between.
    If you do hot backups you need archivelogs is my thought. The number of archivelogs required will be decreased considerably through rman.
    http://download-east.oracle.com/docs/cd/B10501_01/server.920/a96566/rcmconc1.htm#458821
    Summary:-
    Either way archivelogs are needed.

  • How to online datafile of rollback segment of NO archive log available

    I set offline datafile of rool back segement and rename it
    but when I try to online , get error to recover
    I try to recover but unfortunately all archive log was deleted ( kind of cron job in unix aotu delete these file)
    Pls advice how can I set this data file online without the archive log ( while databse still up and running) Oracle version is 8i not possible to drop datafile by sql.
    in roll back segment I have 3 data files as below and the rb04.dbf is offline
    usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs01.dbf
    /usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs03.dbf
    /usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf
    sequence of what I did is below:
    SQL> alter database datafile '/usr/Systems/1354RM_1/databases/dbsnml/data/rbs04.dbf' offline;
    Database altered.
    SQL> alter tablespace RBS rename datafile '/usr/Systems/1354RM_1/databases/dbsnml/data/rbs04.dbf' to '/usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf';
    Tablespace altered.
    SQL> alter database datafile '/usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf' online;
    alter database datafile '/usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf' online
    ERROR at line 1:
    ORA-01113: file 13 needs media recovery
    ORA-01110: data file 13:
    '/usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf'
    SQL> recover datafile '/usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf';
    ORA-00279: change 71449081 generated at 12/23/2008 20:18:05 needed for thread 1
    ORA-00289: suggestion :
    /usr/Systems/1354RM_1_7.1_Master/OSRES/data/warm_repl/WarmArchive/arch_1_455922.
    arc
    ORA-00280: change 71449081 for thread 1 is in sequence #455922
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    ORA-00308: cannot open archived log
    '/usr/Systems/1354RM_1_7.1_Master/OSRES/data/warm_repl/WarmArchive/arch_1_455922
    .arc'
    ORA-27037: unable to obtain file status
    HP-UX Error: 2: No such file or directory
    Additional information: 3

    SQL> alter tablespace RBS rename datafile '/usr/Systems/1354RM_1/databases/dbsnml/data/rbs04.dbf' to '/usr/Systems/1354RM_1_7.1_Master/databases/dbsnml/data/rbs04.dbf';Did you rename the file at OS level after you issued above command? Oracle will not rename the file at OS level.
    Other than that, if you plan to change rollback segment tablespace, better create a new one and create new rollback segment in new tablespace and drop the old one.

  • Archive Logs in every 15 minutes(Oracle 11g 64 bit  EE on Linux RHEL 4

    In our production database we have very few transactions may be of few MB's in whole day but it is generating archive logs constantly in every 15 minutes(may be sometime in 14 minutes also) of file size 50 MB each and this way consume 4 GB of space in a day for archive log which is way above than expected.
    I have checked archive_lag_target and value of this is 0.
    Any clue why it is creating 50 MB archive log file after every 14-15 minutes?

    It's easy enough to reduce redo log file size without downtime; just add new smaller redo log files, switch logfile a couple of times and drop the old redo log files.
    However, if the redo logs are filling up before they switch, then this will probably only make matters worse.
    If the redo logs are switching before they are full then maybe you also need to consider log_checkpoint_interval and log_checkpoint_timeout settings.
    If the redo logs are filling up before they switch then use the techniques suggested by a couple of the other posters to track down the guilty SQL.

  • How to recover the database when some of the archive log file get deleted.

    I am facing a problem with Oracle database, which is related to archivelogs.
    Our development database is running in archivelog mode, but we don't have backups scheduled and have no recovery catalog.
    When the database was in running condition, disk got full, so some archivelogs were deleted manually.
    After this they restarted the DB, and now DB is not coming up. Errors are as follows:
    SQL> startup
    ORACLE instance started.
    Total System Global Area 1444383504 bytes
    Fixed Size 731920 bytes
    Variable Size 486539264 bytes
    Database Buffers 956301312 bytes
    Redo Buffers 811008 bytes
    Database mounted.
    ORA-01589: must use RESETLOGS or NORESETLOGS option for database open
    SQL> alter database open resetlogs;
    alter database open resetlogs
    ERROR at line 1:
    ORA-01113: file 1 needs media recovery
    ORA-01110: data file 1: '/export/home/oracle/dev/ADVFRW/ADVFRW.system'
    SQL> recover datafile '/export/home/oracle/dev/ADVFRW/ADVFRW.system'
    ORA-00283: recovery session canceled due to errors
    ORA-01610: recovery using the BACKUP CONTROLFILE option must be done
    SQL> recover database using backup controlfile;
    ORA-00279: change 215548705 generated at 09/02/2008 17:06:10 needed for thread
    1
    ORA-00289: suggestion :
    /export/home/oracle/dev/ADVFRW/ADVFRW.archivelog1/LOG_ADVFRW_1107_1.ARC
    ORA-00280: change 215548705 for thread 1 is in sequence #1107
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    /export/home/oracle/dev/ADVFRW/ADVFRW.archivelog1/LOG_ADVFRW_1107_1.ARC
    ORA-00308: cannot open archived log
    '/export/home/oracle/dev/ADVFRW/ADVFRW.archivelog1/LOG_ADVFRW_1107_1.ARC'
    ORA-27037: unable to obtain file status
    SVR4 Error: 2: No such file or directory
    Additional information: 3
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    CANCEL
    Media recovery cancelled.
    SQL>
    1. How to recover the database and bring it online
    Any help will be highly appreciated.
    With Regards
    Hemant Joshi
    Edited by: hem_Kec on Sep 7, 2008 9:07 AM

    Hi,
    Archive log files are the copies of redolog files.As redo log files are circularly overwritten,oracle generates archive log file of the corresponding redo logfiles being overwritten.So if you have a backup that dates back to 10 am in the morning and if your database creashed at 3 pm,you cannot use the redo log files alone as they have incomplete information.To completely recover the database upto 3 pm,you need archive log files generated between 10 am to 3 pm. In your case since you are missing one archive log file,you cannot perform complete recovery and hence would suffer data loss.

  • How to delete archive log on standby site

    hi
    i configured dataguard on 10g. could anyone suggest how to delete archived log on standby site..

    Hi,
    There is no automatic feature out-of-the-box but you have the following views:
    1) Shows archive logs transfered:
    select sequence#, first_time, next_time from v$archived_log order by sequence#;
    2) Shows archive logs applied:
    select sequence#, archived, applied from v$archived_log order by sequence#;
    If you see your archivelogs applied you can safely remove archive log files from your disk. It is your decision if you keep or delete them completely.

  • OVM 3.0 Database Creating Lots of Archive Logs

    Greetings - ever since we initially installed OVM 3.0 earlier this fall (~October), the OVM database has generated archive logs at a very rapid rate. It continually threatens to fill up our 16 GB filesystem dedicated to archive logs, even after daily backup and purging.
    Our OVM database itself is about 4-6 GB large, and we would need to increase the archive log filesystem to about 20-25 GB in size, which we see as unreasonable for such a small database.
    What is causing OVM to generate so many redo logs? Our best guess is that OVM is continuously gathering guest VM CPU usage on each physical server.
    Is there a way to configure the OVM application in order to reduce the amount of redo/archive logs being created?
    We are currently running 3.0.3, having upgraded each time a 3.0.* patch was released. OVMM running on OEL 6.1, database running on latest HP-UX.

    majedian21 wrote:
    Greetings - ever since we initially installed OVM 3.0 earlier this fall (~October), the OVM database has generated archive logs at a very rapid rate. It continually threatens to fill up our 16 GB filesystem dedicated to archive logs, even after daily backup and purging.I would log an SR with Oracle Support for this, so that Development can look at it. Sounds like your environment has lots of VMs running and yes, collecting usage stats for all of those environments. However, there may be some old data from the previous versions that's causing more stats to be collected than necessary.

  • Sync database using archive log files

    Hi,
    I am having two databases A nd B. A is the live Database. I need to replicate the A database and its name as B. I had done the replication using hotbackup.
    1. create the B database
    2. then take the hotbackup and copy the files to the B database
    3. Then delete the controlfile, logfiles and datafiles then paste the files from A database.
    4. Recreate the control files in B database
    5. Now the B database is up and running.
    But my question is in A database is still live and its generating archive log files. I would like to know how to sync the B Database similar to A database using archive log files.
    Kindly share me an example..
    Rgds..

    As you have renamed and opened the clone database, it is no longer the same DBID / Database Incarnation as the source. As you point out, it is not a standby database.
    You cannot apply archivelogs from the source database to this clone.
    You have to consider [Oracle Streams|http://download.oracle.com/docs/cd/B19306_01/server.102/b14229/toc.htm] to implement data replication.
    Hemant K Chitale
    Edited by: Hemant K Chitale on Jan 5, 2010 3:34 PM

  • Populate archive logs from production (Solaris) to new production(Linux) db

    I want to migrate my production database(Solaris) to different environment(Linux)
    After production database migration my production database(Solaris) will be up and running for 2 days.
    We are going to check applications for 2 days on temporary production(Linux) database.
    For 2 days my primary database(Solaris) will be up and running and generating archive logs. After that i want to populate those 2 days archives to temporary database(Linux).
    Question 1: How would i populate 2 days archive logs from primary database(Solaris) to new production environment(Linux)
    Is there any way/utility do apply generated archives to different database?\
    I would be thankful for any expert suggestion.
    Thanks,

    Is there any way/utility do apply generated archives to different database?\They have to be applied to the same database (identified by DBID).
    As soon as you OPEN the Linux database (and do any transactions therein), it diverges from your Solaris database. You cannot apply archivelogs of the Solaris database to the Linux database.
    Hemant K Chitale

  • Archive Log vs Full Backup Concept

    Hi,
    I just need some clarification on how backups and archive logs work. Lets say starting at 1PM I have archive logs 1,2,3,4,5 and then I perform a full backup at 6PM.
    Then I resume generating archive logs at 6PM to get logs 6,7,8,9,10. I then stop at 11PM.
    If my understanding is correct, the archive logs should allow me to restore oracle to a point in time anywhere between 1PM and 11PM. But if I only have the full backup then I can only restore to a single point, which is 6PM. Is my understanding correct?
    Do the archive logs only get applied to the datafiles when the backup occurs or only when a restore occurs? It doesn't seem like the archive logs get applied on the fly.
    Thanks in advance.

    thelok wrote:
    Thanks for the great explanation! So I can do a point in time restore from any time since the datafiles have last been written (or from when I have the last set of backed up datafiles plus the archive logs). From what you are saying, I can force the datafiles to be written from the redo logs (by doing a checkpoint with "alter set archive log current" or "backup database plus archivelog"), and then I can delete all the archive logs that have a SCN less than the checkpoint SCN on the datafiles. Is this true? This would be for the purposes of preserving disk space.Hi,
    See this example. I hope this explain your doubt.
    # My current date is 06-11-2011 17:15
    # I not have backup of this database
    # My retention policy is to have 1 backup
    # I start listing  archive logs.
    RMAN> list archivelog all;
    using target database control file instead of recovery catalog
    List of Archived Log Copies
    Key     Thrd Seq     S Low Time            Name
    29      1    8       A 29-10-2011 12:01:58 +HR/dbhr/archivelog/2011_10_31/thread_1_seq_8.399.766018837
    30      1    9       A 31-10-2011 23:00:30 +HR/dbhr/archivelog/2011_11_03/thread_1_seq_9.409.766278025
    31      1    10      A 03-11-2011 23:00:23 +HR/dbhr/archivelog/2011_11_04/thread_1_seq_10.391.766366105
    32      1    11      A 04-11-2011 23:28:23 +HR/dbhr/archivelog/2011_11_06/thread_1_seq_11.411.766516065
    33      1    12      A 05-11-2011 23:28:49 +HR/dbhr/archivelog/2011_11_06/thread_1_seq_12.413.766516349
    ## See I have archive logs from time "29-10-2011 12:01:58" until "05-11-2011 23:28:49" but I dont have any backup of database.
    # So I perfom backup of database including archive logs.
    RMAN> backup database plus archivelog delete input;
    Starting backup at 06-11-2011 17:15:21
    ## Note above RMAN forcing archive current log, this archivelog generated will be usable only for previous backup.
    ## Is not my case... I don't have backup of database.
    current log archived
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: sid=159 devtype=DISK
    channel ORA_DISK_1: starting archive log backupset
    channel ORA_DISK_1: specifying archive log(s) in backup set
    input archive log thread=1 sequence=8 recid=29 stamp=766018840
    input archive log thread=1 sequence=9 recid=30 stamp=766278027
    input archive log thread=1 sequence=10 recid=31 stamp=766366111
    input archive log thread=1 sequence=11 recid=32 stamp=766516067
    input archive log thread=1 sequence=12 recid=33 stamp=766516350
    input archive log thread=1 sequence=13 recid=34 stamp=766516521
    channel ORA_DISK_1: starting piece 1 at 06-11-2011 17:15:23
    channel ORA_DISK_1: finished piece 1 at 06-11-2011 17:15:38
    piece handle=+FRA/dbhr/backupset/2011_11_06/annnf0_tag20111106t171521_0.268.766516525 tag=TAG20111106T171521 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:16
    channel ORA_DISK_1: deleting archive log(s)
    archive log filename=+HR/dbhr/archivelog/2011_10_31/thread_1_seq_8.399.766018837 recid=29 stamp=766018840
    archive log filename=+HR/dbhr/archivelog/2011_11_03/thread_1_seq_9.409.766278025 recid=30 stamp=766278027
    archive log filename=+HR/dbhr/archivelog/2011_11_04/thread_1_seq_10.391.766366105 recid=31 stamp=766366111
    archive log filename=+HR/dbhr/archivelog/2011_11_06/thread_1_seq_11.411.766516065 recid=32 stamp=766516067
    archive log filename=+HR/dbhr/archivelog/2011_11_06/thread_1_seq_12.413.766516349 recid=33 stamp=766516350
    archive log filename=+HR/dbhr/archivelog/2011_11_06/thread_1_seq_13.414.766516521 recid=34 stamp=766516521
    Finished backup at 06-11-2011 17:15:38
    ## RMAN finish backup of Archivelog and Start Backup of Database
    ## My backup start at "06-11-2011 17:15:38"
    Starting backup at 06-11-2011 17:15:38
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting full datafile backupset
    channel ORA_DISK_1: specifying datafile(s) in backupset
    input datafile fno=00001 name=+HR/dbhr/datafile/system.386.765556627
    input datafile fno=00003 name=+HR/dbhr/datafile/sysaux.396.765556627
    input datafile fno=00002 name=+HR/dbhr/datafile/undotbs1.393.765556627
    input datafile fno=00004 name=+HR/dbhr/datafile/users.397.765557979
    input datafile fno=00005 name=+BFILES/dbhr/datafile/bfiles.257.765542997
    channel ORA_DISK_1: starting piece 1 at 06-11-2011 17:15:39
    channel ORA_DISK_1: finished piece 1 at 06-11-2011 17:16:03
    piece handle=+FRA/dbhr/backupset/2011_11_06/nnndf0_tag20111106t171539_0.269.766516539 tag=TAG20111106T171539 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:24
    Finished backup at 06-11-2011 17:16:03
    ## And finish at "06-11-2011 17:16:03", so I can recovery my database from this time.
    ## I will need archivelogs (transactions) which was generated during backup of database.
    ## Note during backup some blocks are copied others not. The SCN is inconsistent state.
    ## To make it consistent I need apply archivelog which have all transactions recorded.
    ## Starting another backup of archived log generated during backup.
    Starting backup at 06-11-2011 17:16:04
    ## So automatically RMAN force another "checkpoint" after backup finished,
    ## forcing archive current log, because this archivelog have all transactions to bring database in a consistent state.
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archive log backupset
    channel ORA_DISK_1: specifying archive log(s) in backup set
    input archive log thread=1 sequence=14 recid=35 stamp=766516564
    channel ORA_DISK_1: starting piece 1 at 06-11-2011 17:16:05
    channel ORA_DISK_1: finished piece 1 at 06-11-2011 17:16:06
    piece handle=+FRA/dbhr/backupset/2011_11_06/annnf0_tag20111106t171604_0.272.766516565 tag=TAG20111106T171604 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:02
    channel ORA_DISK_1: deleting archive log(s)
    archive log filename=+HR/dbhr/archivelog/2011_11_06/thread_1_seq_14.414.766516565 recid=35 stamp=766516564
    Finished backup at 06-11-2011 17:16:06
    ## Note: I can recover my database from time "06-11-2011 17:16:03" (finished backup full)
    ##  until "06-11-2011 17:16:04" (last archivelog generated) that is my recover window in this scenary.
    ## Listing Backup I have:
    ## Archive Logs in backupset before backup full start - *BP Key: 40*
    ## Backup Full database in backupset - *BP Key: 41*
    ##  Archive Logs in backupset after backup full stop - *BP Key: 42*
    RMAN> list backup;
    List of Backup Sets
    ===================
    BS Key  Size       Device Type Elapsed Time Completion Time
    40      196.73M    DISK        00:00:15     06-11-2011 17:15:37
            *BP Key: 40*   Status: AVAILABLE  Compressed: NO  Tag: TAG20111106T171521
            Piece Name: +FRA/dbhr/backupset/2011_11_06/annnf0_tag20111106t171521_0.268.766516525
      List of Archived Logs in backup set 40
      Thrd Seq     Low SCN    Low Time            Next SCN   Next Time
      1    8       766216     29-10-2011 12:01:58 855033     31-10-2011 23:00:30
      1    9       855033     31-10-2011 23:00:30 896458     03-11-2011 23:00:23
      1    10      896458     03-11-2011 23:00:23 937172     04-11-2011 23:28:23
      1    11      937172     04-11-2011 23:28:23 976938     05-11-2011 23:28:49
      1    12      976938     05-11-2011 23:28:49 1023057    06-11-2011 17:12:28
      1    13      1023057    06-11-2011 17:12:28 1023411    06-11-2011 17:15:21
    BS Key  Type LV Size       Device Type Elapsed Time Completion Time
    41      Full    565.66M    DISK        00:00:18     06-11-2011 17:15:57
            *BP Key: 41*   Status: AVAILABLE  Compressed: NO  Tag: TAG20111106T171539
            Piece Name: +FRA/dbhr/backupset/2011_11_06/nnndf0_tag20111106t171539_0.269.766516539
      List of Datafiles in backup set 41
      File LV Type Ckp SCN    Ckp Time            Name
      1       Full 1023422    06-11-2011 17:15:39 +HR/dbhr/datafile/system.386.765556627
      2       Full 1023422    06-11-2011 17:15:39 +HR/dbhr/datafile/undotbs1.393.765556627
      3       Full 1023422    06-11-2011 17:15:39 +HR/dbhr/datafile/sysaux.396.765556627
      4       Full 1023422    06-11-2011 17:15:39 +HR/dbhr/datafile/users.397.765557979
      5       Full 1023422    06-11-2011 17:15:39 +BFILES/dbhr/datafile/bfiles.257.765542997
    BS Key  Size       Device Type Elapsed Time Completion Time
    42      3.00K      DISK        00:00:02     06-11-2011 17:16:06
            *BP Key: 42*   Status: AVAILABLE  Compressed: NO  Tag: TAG20111106T171604
            Piece Name: +FRA/dbhr/backupset/2011_11_06/annnf0_tag20111106t171604_0.272.766516565
      List of Archived Logs in backup set 42
      Thrd Seq     Low SCN    Low Time            Next SCN   Next Time
      1    14      1023411    06-11-2011 17:15:21 1023433    06-11-2011 17:16:04
    ## Here make sense what I trying explain
    ## As I don't have backup of database before of my Last backup, all archivelogs generated before of my backup full is useless.
    ## Deleting what are obsolete in my env, RMAN choose backupset 40 (i.e all archived logs generated before my backup full)
    RMAN> delete obsolete;
    RMAN retention policy will be applied to the command
    RMAN retention policy is set to redundancy 1
    using channel ORA_DISK_1
    Deleting the following obsolete backups and copies:
    Type                 Key    Completion Time    Filename/Handle
    *Backup Set           40*     06-11-2011 17:15:37
      Backup Piece       40     06-11-2011 17:15:37 +FRA/dbhr/backupset/2011_11_06/annnf0_tag20111106t171521_0.268.766516525
    Do you really want to delete the above objects (enter YES or NO)? yes
    deleted backup piece
    backup piece handle=+FRA/dbhr/backupset/2011_11_06/annnf0_tag20111106t171521_0.268.766516525 recid=40 stamp=766516523
    Deleted 1 objectsIn the above example, I could before starting the backup run "delete archivelog all" because they would not be needed, but to show the example I follow this unnecessary way. (backup archivelog and delete after)
    Regards,
    Levi Pereira
    Edited by: Levi Pereira on Nov 7, 2011 1:02 AM

  • Archive log generation

    Hai all,
    In My production environment, some times archives log are generating 5-6 logs a min.. very less no of users are connected to the database now..
    rw-r----- 1 oraprod dba 10484224 Jan 12 14:10 prod_arch_4810.arc
    -rw-r----- 1 oraprod dba 10484224 Jan 12 14:10 prod_arch_4811.arc
    -rw-r----- 1 oraprod dba 10483712 Jan 12 14:10 prod_arch_4812.arc
    -rw-r----- 1 oraprod dba 10484224 Jan 12 14:10 prod_arch_4813.arc
    -rw-r----- 1 oraprod dba 10484224 Jan 12 14:10 prod_arch_4814.arc
    -rw-r----- 1 oraprod dba 10484224 Jan 12 14:10 prod_arch_4815.arc
    Y is this happenin ?
    Any comments or ideas to resolve this?
    Yusuf

    Whenever you create a thread, it is always advisable to specify your current OS and DB versions.
    You could be generating this redo information by means of your scheduled tasks or by the current users activity, the few concurrent number of users doesn't mean they won't be generating a lot of transactions. Check your v$undostat and v$rollstat, v$transaction, and v$session to monitor users and transaction activity.
    10M for a redo log size IMO is very little for the current transaction requirements for most databases. Your database currently generates transaction information at a rate of about 50M/min. With 100M redolog files you would be generating one archivelog file around each two minutes, instead of the current 5 archivelog per minute.
    Since your database is highly transactional, make sure you have enough free space to store your generated archive log files, you will be generating about 3G/hr.
    ~ Madrid

  • Heavy archive log generation

    Hi,
    Database Version: Oracle 11.1.0.6
    Platform: Enterprise Linux 5
    Can someone please tell me the troubleshooting steps_ in a situation where there is heavy inflow of archive log generation i mean, say around 3 files of around 50MB every minute and eating away the space on disk.
    1) How to find out what activity is causing such heavy archive log generation? I can run the below query to find out the currently running sql queries with status;
    select a.username, b.sql_text,a.status from v$session a inner join v$sqlarea b on a.sql_id=b.sql_id;
    But is there any other query or a better way to find out current db activity in this situation.
    Tried using DBMS_LGMNR Log Miner but failed because (i) utl_file_dir is not set in init param file (So, mining the archive log file on production is presently ruled out as I cannot take an outage)
    (ii) alter database add supplemental log data (all) columns query takes for ever because of the locks. (So, cannot mine the generated archive log file on another machine due to DBID mismatch.)
    2) How to deal with this situation? I read here in OTN discussion board that increasing the number of redo log groups or redo log members will help to manage this situation when there is lot of DML activity going on application side....But I didn't understand how it is going to help in controlling the rigorous archive logs generation
    Edited by: user10313587 on Feb 11, 2011 8:43 AM
    Edited by: user10313587 on Feb 11, 2011 8:44 AM

    Hi,
    Other than logminer, which will tell you exactly what the redo is by definition, you can run something like the following:
    select value/((sysdate-logon_time) * 1440) redo_per_minute,s.sid,serial#,logon_time,value
      from v$statname sn,
           v$sesstat ss,
           v$session s
      where s.sid = ss.sid
        and sn.statistic# = ss.statistic#
        and name = 'redo size'
    and value>0Then trace the "high" sessions above and it should jump out at you. If not, then run logmnr with something like...
    set serveroutput on size unlimited
    begin
      dbms_logmnr.add_logfile(logfilename => '&log_file_name');
      dbms_logmnr.start_logmnr(options => dbms_logmnr.dict_from_online_catalog + dbms_logmnr.no_rowid_in_stmt);
      FOR cur IN (SELECT *
                    FROM v$logmnr_contents) loop
        dbms_output.put_line(cur.sql_redo);
      end loop;
    end;
    /Note you don't need utl_file_dir for log miner if you use the online catalog.
    HTH,
    Steve

Maybe you are looking for

  • Remote laptop won't print to new HP4622 wireless printer

    My new HP 4622 won't print from wife's laptop (Compaq Presario CQ-60-224NR w/ Windows Vista) in another part of the house on my home Centurylink DSL network with very good connectivity.  I have the printer installed at my desktop (Gateway Microsoft X

  • Dynamic table in interactive form & WD ABAP

    Hi,       I have created a adobe interactive form adding the rows dynamically when pressing add button. The form works fine in Life cycle designer preview. But it doesnot work in Webdynpro ABAP runtime. Please let me know if any of you have a solutio

  • ADF: Calling concurrent program in Oracle Apps from ADF

    Hi All, I am using Jdev 11G... I have a requirement to call a concurrent program which is in Oracle E-Business from ADF page. i will have ADF page with one button, On click of that ADF button we need to call the concurrent program and show the output

  • Need some help getting started with Mobility Services Engine

    Someone has ordered several 3602 access points, a 5508 controller, and a MSE 3310 for one of our remote locations. The access points and controller are in place and are working fine (we use lightweight APs at many other locations), but I'm not sure h

  • Database refresh on Tablespace Level

    Hi Guys, Please can you advise if for refreshing database on tablespace level we require to move the source tablespaces in read only mode? Is there any workaround? Cheers, Shaz