Errors reported in alert log file regarding redo log files

Hi,
In my database i see the following entries regarding the redo log files frequently.
Thread 1 advanced to log sequence 88
Current log# 3 seq# 88 mem# 0: D:\ORACLE\PRODUCT\10.2.0\ORADATA\PELICAN\REDO03.LOG
Thread 1 advanced to log sequence 89
Current log# 1 seq# 89 mem# 0: D:\ORACLE\PRODUCT\10.2.0\ORADATA\PELICAN\REDO01.LOG
Thread 1 cannot allocate new log, sequence 90
Checkpoint not complete
I have 3 redo log files of 50MB each. My database is 10g.
What is the reason?
What do i need to do ?
Thanks.

Hi,
It is clearly stated that checkpoint is not complete. Thats why it could not allocate new sequence number to group 1.
We can understand that your redo log group size is insufficient. or increase the number of redo log groups.
Just test adding one more redo log group or test by increasing the size of redo log groups.

Similar Messages

  • The file structure online redo log, archived redo log and standby redo log

    I have read some Oracle documentation for file structure and settings in Data Guard environment. But I still have some doubts. What is the best file structure or settings in Oracle 10.2.0.4 on UNIX for a data guard environment with 4 primary databases and 4 physical standby databases. Based on Oracle documents, there are 3 redo logs. They are: online redo logs, archived redo logs and standby redo logs. The basic settings are:
    1. Online redo logs --- This redo log must be on Primary database and logical standby database. But it is not necessary to be on physical standby database because physical standby is not open. It doesn't generate redo log. However, if don't set up online redo log on physical standby, when primary failover and switch standby as primary. How can standby perform without online redo logs? In my standby databases, online redo logs have been set up.
    2. Archived redo logs --- It is obviously that primary database, logical and physical standby database all need to have this log file being set up. Primary use it to archive log files and ship to standby. Standby use it to receive data from archived log and apply to database.
    3. Standby redo logs --- In the document, it says A standby redo log is similar to an online redo log, except that a standby redo log is used to store redo data received from another database. A standby redo log is required if you want to implement: The maximum protection and maximum availability levels of data protection and Real-time apply as well as Cascaded destinations. So it seems that this standby redo log only should be set up on standby database, not on primary database. Am my understanding correct? Because I review current redo log settings on my environment, I have found that Standby redo log directory and files have been set up on both primary and standby databases. I would like to get more information and education from experts. What is the best setting or structure on primary and standby database?

    FZheng:
    Thanks for your input. It is clear that we need 3 type of redo logs on both databases. You answer my question.
    But I have another one. In oracle ducument, it says If you have configured a standby redo log on one or more standby databases in the configuration, ensure the size of the current standby redo log file on each standby database exactly matches the size of the current online redo log file on the primary database. It says: At log switch time, if there are no available standby redo log files that match the size of the new current online redo log file on the primary database. The primary database will shut down
    My current one data gurard envirnment setting is: On primary DB, online redo log group size is 512M and standby redo log group size is 500M. On the standby DB, online redo log group size is 500M and standby redo log group size is 750M.
    This was setup by someone I don't know. Is this setting OK? or I should change Standby Redo Log on standby DB to 512M to exactly meatch with redo log size on primary?
    Edited by: 853153 on Jun 22, 2011 9:42 AM

  • Online redo logs vs standby redo logs

    Hi All,
    Oracle version : 10.2.0.4
    OS : SunOS 5.10 Generic_118855-36 i86pc i386 i86pc
    This might be a silly question for the experts :(
    We are having 2 standby databases and 1 logical database. Everything is running fine , I have one confusion : My primary server having 3 online redo log files and 3 standby redo log files (on primary local), but as per the architecture standby redo logs should be on standby server.My question is whether standby redo logs on standby are online redo logs of standby server or standby redo logs of primary server. e.g If i perform a log switch on primary then what is the role of sync and affirm, will lgwr write parallel on online redo logs ,standby redo logs (of primary) or online redo logs (of primary) and online redo logs (which is known as standby redo logs on standby ). Please help me to understand the scenerio.
    Thanks in advance!

    RFS receives redo information from the primary database.
    RFS can write the redo into standby redo logs or directly to archived redo logs. Each LNSn and ARCn process from the primary database has its own RFS process.
    A standby redo log is used only when the database is in the standby role to store redo data received from the primary database. Standby redo logs form a separate pool of log file groups.
    Configuring standby redo log files is highly recommended on all standby databases in a Data Guard configuration, including the primary database to aid in role reversal.
    A standby redo log is required to implement:
    • The maximum protection and maximum availability levels of data protection
    • Real-time apply
    • Cascaded redo log destinations
    Standby redo logs are recommended for maximum performance data protection mode. Unless you are using the real-time apply feature, standby redo logs must be archived before the data can be applied to the standby database. The standby archival operation occurs automatically.

  • Dataguard lost both Primary redo log and standby redo log files

    Hi,
    I am new to data guard, i came acorss a scenario where we loose both primary redo log file and standby redo log files.
    Can someone please help me understand how to recover from this situation.
    Thanks!

    >loose both primary redo log file and standby redo log files
    We have to be very clear.
    There are (set A) online redo log files  and (set B) standby redo log files at (location 1) Primary and (location 2) Standby.
    The standby redo log files, depending on the configuration, aren't strictly mandatory.  The standby can be applying redo without online redo log files present as well, depending on how it was setup.
    So, the question is  : Did you lose online redo log files at the primary ?  Didn't the primary shutdown itself then ? If so, you have to do an incomplete recovery at the primary OR switch over to the standby (which may or may not have received the last transaction, depending on how it was configured and operating)   OR restore from the standby (again, with possible loss of transactions) to the primary.
    Hemant K Chitale

  • Behaviuor in reading logs from DB13 - redo log backup for DB

      Hi all,
    I have a question ,
    We have scheduled redo log backup using HP DP Tool.
    When I am reading a backup log from DB13 , I am facing the weird behavior.
    copy, delete, save
    0000
    aeomzyok.cds
    copy, delete, save
    0001
    aeomzyoh.cds
    Where ever the RC is 000 I am able to read the log files available at OS level . But not able to read the log for other RC.
    Please help me to understand this
    Thanks,
    Gnana

    Hi Gnana.
    Use the latest BR*Tools as there were problems with permission issues while creating logfiles. In the latest BR*Tools version, the different logfiles are created with predefined permission levels that suits to SAP needs.
    Also it worth to check if the BR*Tools executables hav the correct ownership.
    113747 Owners and authorizations for BR*Tools
    Regards,
    János

  • Where RFS exactly write redo data ?  ( archived redo log or standby redo log ) ?

    Good Morning to all ;
    I am getting bit confused from oracle official link . REF_LINK : Log Apply Services
    Redo data transmitted from the primary database is received by the RFS on the standby system ,
    where the RFS process writes the redo data to either archived redo log files  or  standby redo log files.
    In standby site , does rfs write redo data in any one file or both ?
    Thanks in advance ..

    Hi GTS,
    GTS (DBA) wrote:
    Primary & standby log file size should be same - this is okay.
    1) what are trying to disclose about  largest & smallest here ? -  You are confusing.
    Read: http://docs.oracle.com/cd/E11882_01/server.112/e25608/log_transport.htm#SBYDB4752
    "Each standby redo log file must be at least as large as the largest redo log file in the redo log of the redo source database. For administrative ease, Oracle recommends that all redo log files in the redo log at the redo source database and the standby redo log at a redo transport destination be of the same size."
    GTS (DBA) wrote:
    2) what abt group members ? should be same as primary or need  to add some members additionally. ?
    Data Guard best practice for performance, is to create one member per each group in standby DB. on standby DB, one member per group is reasonable enough. why? to avoid write penalty; writing to more than one log files at the standby DB.
    SCENARIO 1: if in your source primary DB you have 2 log member per group, in standby DB you can have 1 member  per group, additionally create an extra group.
    primary
    standby
    Member per group
    2
    1
    Number of log group
    4
    5
    SCENARIO 2: you can also have this scenario 2 but i will not encourage it
    primary
    standby
    Member per group
    2
    2
    Number of log group
    4
    5
    GTS (DBA) wrote:
    All standby redo logs of the correct size have not yet been archived.
      - at this situation , can we force on standby site ? any possibilities ? 
    you can not force it , just size your standby redo files correctly and make sure you don not have network failure that will cause redo gap.
    hope there is clarity now
    Tobi

  • How could I find specified date of REDO log and archive REDO log ?

    we use Oracle11gr2 on win2008R2.
    1
    How could I find specified date of REDO log(2013/10/17,etc) and archive REDO log ?
    2
    What is the format of archive REDO log.? (zipped file ?)

    user12075536123 wrote:
    1)
    select * from v$archived_log;
    select * from v$log_history;
    but there is a possibility there is no old data
    below contains no filename column
    SQL> desc v$log_history
    Name                                      Null?    Type
    RECID                                              NUMBER
    STAMP                                              NUMBER
    THREAD#                                            NUMBER
    SEQUENCE#                                          NUMBER
    FIRST_CHANGE#                                      NUMBER
    FIRST_TIME                                         DATE
    NEXT_CHANGE#                                       NUMBER
    RESETLOGS_CHANGE#                                  NUMBER
    RESETLOGS_TIME                                     DATE
    there is NO data when archive mode is disabled

  • How to start a database by copying the files(tablespace, redo, ctl files)

    Hi,
    I have two databases called Q1 and Q2. I am planning to copy from a cold backup files from Q1 to Q2 database.
    could anybody will provide the steps involved in how to copy from one database to another and how to rename the target database after copying and renaming the files in the target database to up and run?
    Please help.
    Thank you.

    For Oracle version from 9i the steps are :
    Perform Cold Backup of the Production database
    1. Note down the datafile, controlfile and redologfile locations on proddb
    2. Shutdown the production database Ensure clean shutdown.
    3. Copy the database files (datafiles, logfiles, controlfiles and parameter file) into the clone destination.
    4. Modify the parameter CONTROL_FILES in the pfile for clonedb to reflect
    the new path of the control files.
    1. Mount the database and rename the datafiles and redolog files.
    export ORACLE_SID=clonedb
    DB02> sqlplus / as sysdba
    DB02> startup mount
    DB02> alter database rename file '/u01/<path_to_proddb_files>/<filename>' to '/u01/<path_to_clonedb_files>/<filename>'
    2. Rename the database with the dbnewid (nid) utility
    3. Edit the pfile and update the new database name
    Modify parameter DB_NAME to reflect the new database name.
    also ensure you make necessary changes to the environment for ORACLE_SID, ORACLE_HOME
    as needed.
    4. Open the clone database.

  • How to clean redo log file?

    Hi, guys:
    I need to migrate data from external tables to normal tables. But the connection is always frozen half the way of executing script. The error message is ORA-00257. it looks redo log file is full. I tried to delete redo log file with RMAN, but I got this error message:
    delete archivelog until time 'trunc(sysdate)';
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=43 device type=DISK
    specification does not match any archived log in the repository
    Any suggestion would be appreciated.
    Sam

    lxiscas wrote:
    Hi, guys:
    I need to migrate data from external tables to normal tables. But the connection is always frozen half the way of executing script. The error message is ORA-00257. it looks redo log file is full. I tried to delete redo log file with RMAN, but I got this error message:
    Sounds like you are confusing (online) redo logs with "archivelogs" (archived redo logs).
    You don't delete redo logs at all. You can delete archivelogs with rman, which is what your command is trying to do. Do you ever backup the archivelogs? A proper backup/recovery policy would backup the archivelogs on at least a daily (if not more often) basis, and delete them after they are backed up. That will keep the archivelog destination from filling up in all but the most extreme circumstances.
    delete archivelog until time 'trunc(sysdate)';
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=43 device type=DISK
    specification does not match any archived log in the repository
    Any suggestion would be appreciated.
    Sam

  • Redo log files deleted

    Hello,
    One of my friend deleted redo log and control files of one machine in cluster environment. Oracle parallel server(8.1.7) is installed on this cluster.
    Redo log and control files of the second disk is still there.
    Can i copy the control files and redo log files of the second disk to the first one...??? Will it work ???
    Please let me know how should I go by it...??
    Thanking in anticipation.
    Regards,
    Rajeev

    Hello,
    I don't have any experiences with cluster/parallel server. But in general:
    1. Controlfile: shutdown database and cp exisiting controlfile to "old" destination. Filename must be similar to deleted controlfile.
    2. Redo Logs: Drop and recreate groups step by step (no user activity recommended). Database won't crash, because members are still "alive" on second disk.
    HTH,
    juergen_klinsmann

  • Deleted redo log files

    hi everbody,
    i use oracle 9.2 and 3 redo log files was created when installation. Now, i could not open database because two of redo log files was deleted (that i haven't understood how deleted.) So what can i do ? recovery is possible ? if poosibel how ?
    NOTES :
    my database does not work Archive mode and give below error messages...
    Total System Global Area 126950220 bytes
    Fixed Size 453452 bytes
    Variable Size 96468992 bytes
    Database Buffers 29360128 bytes
    Redo Buffers 667648 bytes
    Database mounted.
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: 'C:\ORACLE\ORADATA\ABUZER\REDO01.LOG'

    thanks everbody. i couldn't give information because some problems was internet connections. i solved problems. i found command and info in the Oracle 8i Admin Guide documentation. first i dropped two log groups whose files are missing then recreate these these log groups with redo log files..before that i start up an new instance, but do not mount the database. after that i use commands like below :
    ALTER DATABASE DROP LOGFILE GROUP 1 ;
    ALTER DATABSE DROP LOGFILE GROUP 2 ;
    ALTER DATABASE ADD LOGFILE GROUP 1 ('C:\...\REDO01.LOG') SIZE 1020400K
    ALTER DATABASE ADD LOGFILE GROUP 2 ('C:\...\REDO02.LOG') SIZE 1020400K
    thanks everbody...

  • "Could not redo log record..." error

    The following error killed the entire database:
    "Could not redo log record (21816:100853:13), for transaction ID (0:100066184), on page (1:621028), database 'MyDatabase' (database ID 5). Page: LSN = (21816:99572:9), type = 2. Log: OpCode = 2, context 4, PrevPageLSN: (21816:99584:9). Restore from a backup of the database, or repair the database."
    I have no idea what caused this.  Thank goodness, my backup worked.  Could anyone offer any clue?  Does anyone know how to repair a database under this condition?
    Thanks,
    hz

    The following error killed the entire database:
    "Could not redo log record (21816:100853:13), for transaction ID (0:100066184), on page (1:621028), database 'MyDatabase' (database ID 5). Page: LSN = (21816:99572:9),
    type = 2. Log: OpCode = 2, context 4, PrevPageLSN: (21816:99584:9). Restore from a backup of the database, or repair the database."
    I have no idea what caused this.  Thank goodness, my backup worked.  Could anyone offer any clue?  Does anyone know how to repair a database under this condition?
    Thanks,
    hz
    Hello,
    Hong this seems to me as if your log file is damaged or corrupted ,DBCC CHECKDB does not checks consistency for log file.YOu can run checkdb for database see if it comess clean.
    Please see errormessages in SQL Server errorlog and eventviewer and please post it here,its important you need to fins out why this message occured.
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • Deleting all users WER (Windows error reporting) files

    Hi,
    I am using windows server 2008 R2 (Standard), I have created around 120 users in this server, yesterday i found that my C drive is full, after checking all i found around 30GB is used for Windows error reporting files, Please help me to delete This files
    Please, i cant delete all files one by one...

    For what it's worth, this is the script I use to find Windows Error Reports that have not been sent, and delete archived Windows Error Reports:
    It's not currently doing per-user error reports, as we don't have people logging into the servers, typically.
    Karl
    <#
    .SYNOPSIS
    Find-WER - Finds Windows Error Reports that have not been sent to Microsoft.
    .DESCRIPTION
    This script finds Windows Error Reports that have not been sent to Microsoft.
    This script also deletes archived Windows Error Reports.
    .NOTES
    File Name: Find-WER.ps1
    Author: Karl Mitschke
    Requires: Powershell V2
    Created: 12/16/2011
    .EXAMPLE
    C:\PS>.\Find-Wer.ps1
    Description
    Finds Windows Error Reports on the local computer.
    .EXAMPLE
    C:\PS>.\Find-WER -Computer Exch2010
    Description
    Finds Windows Error Reports on the computer Exch2010.
    .EXAMPLE
    C:\PS> Get-ExchangeServer | Find-WER
    Description
    Finds Windows Error Reports on all Exchange servers.
    .EXAMPLE
    C:\PS> $cred = Get-Credential -Credential mitschke\karlm
    C:\PS>.\Find-WER.ps1 -ComputerName (Get-Content -Path ..\Servers.txt) -Credential $cred
    Description
    Finds Windows Error Reports on all servers in the file Servers.txt.
    .PARAMETER ComputerName
    The Computer(s) to search for Windows Error Reports. If not specified, defaults to the local computer.
    .PARAMETER Credential
    The Credential to use. If not specified, runs under the current security context.
    #>
    [CmdletBinding(SupportsShouldProcess=$true, ConfirmImpact='Medium')]
    param (
    [parameter(
    Position = 1,
    Mandatory=$false,
    ValueFromPipeline=$true)
    [String[]]$ComputerName=$Env:ComputerName,
    [Parameter(
    Position = 2,
    Mandatory = $false,
    ValueFromPipeline=$false)
    $Credential
    BEGIN{
    #region PSBoundParameters modification
    if ($Credential -ne $null -and $Credential.GetType().Name -eq "String"){
    $PSBoundParameters.Remove("Credential") | Out-Null
    $PSBoundParameters.Add("Credential", (Get-Credential -Credential $Credential))
    $PSBoundParameters.Remove("WhatIf") | Out-Null
    #endregion
    $ReportPath = "C:\\ProgramData\\Microsoft\\Windows\\WER\\ReportQueue"
    $ArchivePath = "C:\\ProgramData\\Microsoft\\Windows\\WER\\ReportArchive"
    $FoundReport = $false
    $DeleteFailure = $false
    PROCESS {
    $PSBoundParameters.Remove("ComputerName") | Out-Null
    foreach ($StrComputer in $ComputerName){
    $Count = @((Get-WmiObject Win32_Directory -Filter "Name= '$ReportPath'" -ComputerName $StrComputer @PSBoundParameters).GetRelated('Win32_directory') |? {$_.name -match $ReportPath}).Count
    if ($Count -gt 0){
    Write-Warning "There are $Count Windows Error Reports on $StrComputer."
    $FoundReport = $true
    Else{
    Write-Output "There are no Windows Error Reports on $StrComputer."
    $Archives = @((Get-WmiObject Win32_Directory -Filter "Name= '$ArchivePath'" -ComputerName $StrComputer @PSBoundParameters).GetRelated('Win32_directory') |? {$_.name -match $ArchivePath})
    if ($Archives.Count -gt 0){
    Write-Output "Deleting Archived Windows Error Reports on $StrComputer."
    foreach ($Folder in $Archives)
    if ($pscmdlet.ShouldProcess($($StrComputer), "Deleting '$($folder.name.split("\")[-1])'")){
    $Return = $Folder.Delete()
    if ($Return.ReturnValue -ne 0){
    $DeleteFailure = $true
    if ($DeleteFailure -eq $true){
    Write-Warning "Could not delete all Archived Windows Error Reports on $StrComputer."
    END {
    If ($FoundReport){
    Write-Warning "Remember to run the Action items on the servers."
    else{
    Write-Output "There are no Windows Error Reports."
    When you see answers and helpful posts, please click Vote As Helpful, Propose As Answer, and/or Mark As Answer.
    My Blog: Unlock PowerShell
    My Book:
    Windows PowerShell 2.0 Bible
    My E-mail: -join ('6F6C646B61726C406F75746C6F6F6B2E636F6D'-split'(?<=\G.{2})'|%{if($_){[char][int]"0x$_"}})

  • Multiplexing Online redo logs, archive logs, and control files.

    Currently I am only multiplexing my control files and online redo logs, My archive logs are only going to the FRA and then being backed up to tape.
    We have to replace disks that hold the FRA data. HP says there is a chance we will have to rebuild the FRA.
    As my archive logs are going to the FRA now, can I multiplex them to another disk group? And if all of the control files, online redo logs and archive logs are multiplexed to another disk group, when ASM dismounts the FRA disk group due to insufficient number of disks, will the database remain open and on line.
    If so then I will just need to rebuild the ASM volumes, and the FRA disk group and bring it to the mount state, correct?
    Thanks!

    You can save your online redo logs and archive logs anywhere you want by making use of of init params create_online_log_dest and log_archive_dest_n. You will have to create new redo log groups in the new location and drop the ones in the FRA. The archive logs will simply land wherever you designate with log_archive_dest_n parameters. Moving the control files off FRA is a little trickier because you will need to restore your controlfile to a non-FRA destination and then shutdown your instance, edit the control file param to reflect changes and restart.
    I think you will be happier if you move everything off the FRA diskgroup before dismounting it, and not expecting the db to automagically recover from the loss of files on the FRA.

  • Backup vs redo log files

    Hi,
    Can anyone tell me what's the real difference between a backup file and a redo log file/archived redo log file and the scenarios (examples) when each of them can be used (for example, "......") ? Both are used for database/instance recovery. I have read the concepts of there 2 terms, but I need some additional information about the difference.
    Thanks!

    Roger25 wrote:
    What i still don't understand is how redo mechanism works; I know, redo logs records changes made to the database and it's used to re-do informations in case of a system failure/crash, for example. Ok, let's say I have a huge update, but when the system crash has occured, the transaction is neither comitted, nor rolled back. Then, how this redo is useful? If a system crash occur, all those updates aren't rolled back automatically? So the database state is as before executing that huge update? So in this case, 'what to redo'?No, with the system's crash, the transaction never gets a chance to get committed (and even rolled back) because a commit is only supposed to happen if there is an explicit or implicit commit statement is issued. Now, for the redo , with the statement, a commit marker is entered at the end of the redo stream of that transaction denoting that it's finally over now and for the corresponding transaction, the status is cleared from the transaction table that's stored in the Undo segment in which the Undo information of that transaction is stored. If you have given a huge update, the redo would be very highly required as you should be knowing that it's not all the time that the dirty buffers that are written to the data files unlikely the redo log files which are written every 3 econds. This means, there is a very fair chance that many changes have yet not even propagated to the data files for which the change vectors are now already there in the redo log files. Now, for the sake of clarity, assume that this transaction was actually committed! So now, without those changes be applied to the data files how Oracle can actually show the committed reults when the database next time would be opened? For this purpose, the redo change vectors are required. And for the uncommitted transactions , the applcation of the redo change vectors would keep on updating the Checkpoint numbers of the data files and would eventually get them synched with the control file and redo log files, without which the database would never be opened.
    HTH
    Aman....

Maybe you are looking for

  • Unable to create user in PKGBUILD / fakeroot

    tl;tr Package creation witk makepkg crashes if a certain user does not exist I'm maintaining package Icinga [1] and I've had three comments about the way the pakage wants to install. Following the install from the official documentation [2] a user ne

  • Case and if statements in Ehancment points

    Hi Friends I have a standard code "Standard code IF <condition>. "In between i have an enhacment point Enhancment point. enhancment. ELSEIF <condition1> endenhacment. ENDIF. I get an error no if condition or no endif within the enhacment point. Is th

  • Is it Possible to upgrade CQ62-219wm processors and ram?

    I have a compaq presario CQ62-219wm laptop and would like to know what processors are compatable with it, and can it be upgraded?

  • Mapping problem of JnlpDownloadServlet from .Jnlp file

    Hello All, 1. I am trying to run a java web start application, i am using tomcat 5.5.20 server. I am using the JnlpDownloadServlet to use the $$codebase and $$name to dynimically code the path to the jar's and the jnlp file. Here is what i have done:

  • Configuring Apache HTTP as an RPS

    Longback i configured Apache HTTP as an Reverse proxy to weblogic for singlesignon with OAM. that time i downloaded the Apache HTTP server Plugin from the below link http://e-docs.bea.com/wls/docs92/plugins/apache.html but after oracle aquiring the o