Capture process issue...archive log missing!!!!!

Hi,
Oracle Streams capture process is alternating between INITIALIZING and DICTIONARY INITIALIZATION state and not proceeding after this state to capture updates made on table.
we have accidentally missing archivelogs and no backup archive logs.
Now I am going to recreate the capture process again.
How I can start the the capture process from new SCN ?
And Waht is the batter way to remove the archive log files from central server, because
SCN used by capture processes?
Thanks,
Faziarain
Edited by: [email protected] on Aug 12, 2009 12:27 AM

Using dbms_Streams_Adm to add a capture, perform also a dbms_capture_adm.build. You will see in v$archived_log at the column dictionary_begin a 'yes', which means that the first_change# of this archivelog is first suitable SCN for starting capture.
'rman' is the prefered way in 10g+ to remove the archives as it is aware of streams constraints. If you can't use rman to purge the archives, then you need to check the min required SCN in your system by script and act accordingly.
Since 10g, I recommend to use rman, but nevertheless, here is the script I made in 9i in the old time were rman was eating the archives needed by Streams with appetite.
#!/usr/bin/ksh
# program : watch_arc.sh
# purpose : check your archive directory and if actual percentage is > MAX_PERC
#           then undertake the action coded by -a param
# Author : Bernard Polarski
# Date   :  01-08-2000
#           12-09-2005      : added option -s MAX_SIZE
#           20-11-2005      : added option -f to check if an archive is applied on data guard site before deleting it
#           20-12-2005      : added option -z to check if an archive is still needed by logminer in a streams operation
# set -xv
#--------------------------- default values if not defined --------------
# put here default values if you don't want to code then at run time
MAX_PERC=85
ARC_DIR=
ACTION=
LOG=/tmp/watch_arch.log
EXT_ARC=
PART=2
#------------------------- Function section -----------------------------
get_perc_occup()
  cd $ARC_DIR
  if [ $MAX_SIZE -gt 0 ];then
       # size is given in mb, we calculate all in K
       TOTAL_DISK=`expr $MAX_SIZE \* 1024`
       USED=`du -ks . | tail -1| awk '{print $1}'`    # in Kb!
  else
    USED=`df -k . | tail -1| awk '{print $3}'`    # in Kb!
    if [ `uname -a | awk '{print $1}'` = HP-UX ] ;then
           TOTAL_DISK=`df -b . | cut -f2 -d: | awk '{print $1}'`
    elif [ `uname -s` = AIX ] ;then
           TOTAL_DISK=`df -k . | tail -1| awk '{print $2}'`
    elif [ `uname -s` = ReliantUNIX-N ] ;then
           TOTAL_DISK=`df -k . | tail -1| awk '{print $2}'`
    else
             # works on Sun
             TOTAL_DISK=`df -b . | sed  '/avail/d' | awk '{print $2}'`
    fi
  fi
  USED100=`expr $USED \* 100`
  USG_PERC=`expr $USED100 / $TOTAL_DISK`
  echo $USG_PERC
#------------------------ Main process ------------------------------------------
usage()
    cat <<EOF
              Usage : watch_arc.sh -h
                      watch_arc.sh  -p <MAX_PERC> -e <EXTENTION> -l -d -m <TARGET_DIR> -r <PART>
                                    -t <ARCHIVE_DIR> -c <gzip|compress> -v <LOGFILE>
                                    -s <MAX_SIZE (meg)> -i <SID> -g -f
              Note :
                       -c compress file after move using either compress or gzip (if available)
                          if -c is given without -m then file will be compressed in ARCHIVE DIR
                       -d Delete selected files
                       -e Extention of files to be processed
                       -f Check if log has been applied, required -i <sid> and -g if v8
                       -g Version 8 (use svrmgrl instead of sqlplus /
                       -i Oracle SID
                       -l List file that will be processing using -d or -m
                       -h help
                       -m move file to TARGET_DIR
                       -p Max percentage above wich action is triggered.
                          Actions are of type -l, -d  or -m
                       -t ARCHIVE_DIR
                       -s Perform action if size of target dir is bigger than MAX_SIZE (meg)
                       -v report action performed in LOGFILE
                       -r Part of files that will be affected by action :
                           2=half, 3=a third, 4=a quater .... [ default=2 ]
                       -z Check if log is still needed by logminer (used in streams),
                                it requires -i <sid> and also -g for Oracle 8i
              This program list, delete or move half of all file whose extention is given [ or default 'arc']
              It check the size of the archive directory and if the percentage occupancy is above the given limit
              then it performs the action on the half older files.
        How to use this prg :
                run this file from the crontab, say, each hour.
     example
     1) Delete archive that is sharing common arch disk, when you are at 85% of 2500 mega perform delete half of the files
     whose extention is 'arc' using default affected file (default is -r 2)
     0,30 * * * * /usr/local/bin/watch_arc.sh -e arc -t /arc/POLDEV -s 2500 -p 85 -d -v /var/tmp/watch_arc.POLDEV.log
     2) Delete archive that is sharing common disk with oother DB in /archive, act when 90% of 140G, affect by deleting
     a quater of all files (-r 4) whose extention is 'dbf' but connect before as sysdba in POLDEV db (-i) if they are
     applied (-f is a dataguard option)
     watch_arc.sh -e dbf -t /archive/standby/CITSPRD -s 140000 -p 90 -d -f -i POLDEV -r 4 -v /tmp/watch_arc.POLDEV.log
     3) Delete archive of DB POLDEV when it reaches 75% affect 1/3 third of files, but connect in DB to check if
     logminer do not need this archive (-z). this is usefull in 9iR2 when using Rman as rman do not support delete input
     in connection to Logminer.
     watch_arc.sh -e arc -t /archive/standby/CITSPRD  -p 75 -d -z -i POLDEV -r 3 -v /tmp/watch_arc.POLDEV.log
EOF
#------------------------- Function section -----------------------------
if [ "x-$1" = "x-" ];then
      usage
      exit
fi
MAX_SIZE=-1  # disable this feature if it is not specificaly selected
while getopts  c:e:p:m:r:s:i:t:v:dhlfgz ARG
  do
    case $ARG in
       e ) EXT_ARC=$OPTARG ;;
       f ) CHECK_APPLIED=YES ;;
       g ) VERSION8=TRUE;;
       i ) ORACLE_SID=$OPTARG;;
       h ) usage
           exit ;;
       c ) COMPRESS_PRG=$OPTARG ;;
       p ) MAX_PERC=$OPTARG ;;
       d ) ACTION=delete ;;
       l ) ACTION=list ;;
       m ) ACTION=move
           TARGET_DIR=$OPTARG
           if [ ! -d $TARGET_DIR ] ;then
               echo "Dir $TARGET_DIR does not exits"
               exit
           fi;;
       r)  PART=$OPTARG ;;
       s)  MAX_SIZE=$OPTARG ;;
       t)  ARC_DIR=$OPTARG ;;
       v)  VERBOSE=TRUE
           LOG=$OPTARG
           if [ ! -f $LOG ];then
               > $LOG
           fi ;;
       z)  LOGMINER=TRUE;;
    esac
done
if [ "x-$ARC_DIR" = "x-" ];then
     echo "NO ARC_DIR : aborting"
     exit
fi
if [ "x-$EXT_ARC" = "x-" ];then
     echo "NO EXT_ARC : aborting"
     exit
fi
if [ "x-$ACTION" = "x-" ];then
     echo "NO ACTION : aborting"
     exit
fi
if [ ! "x-$COMPRESS_PRG" = "x-" ];then
   if [ ! "x-$ACTION" =  "x-move" ];then
         ACTION=compress
   fi
fi
if [ "$CHECK_APPLIED" = "YES" ];then
   if [ -n "$ORACLE_SID" ];then
         export PATH=$PATH:/usr/local/bin
         export ORAENV_ASK=NO
         export ORACLE_SID=$ORACLE_SID
         . /usr/local/bin/oraenv
   fi
   if [ "$VERSION8" = "TRUE" ];then
      ret=`svrmgrl <<EOF
connect internal
select max(sequence#) from v\\$log_history ;
EOF`
LAST_APPLIED=`echo $ret | sed 's/.*------ \([^ ][^ ]* \).*/\1/' | awk '{print $1}'`
   else
    ret=`sqlplus -s '/ as sysdba' <<EOF
set pagesize 0 head off pause off
select max(SEQUENCE#) FROM V\\$ARCHIVED_LOG where applied = 'YES';
EOF`
   LAST_APPLIED=`echo $ret | awk '{print $1}'`
   fi
elif [ "$LOGMINER" = "TRUE" ];then
   if [ -n "$ORACLE_SID" ];then
         export PATH=$PATH:/usr/local/bin
         export ORAENV_ASK=NO
         export ORACLE_SID=$ORACLE_SID
         . /usr/local/bin/oraenv
   fi
    var=`sqlplus -s '/ as sysdba' <<EOF
set pagesize 0 head off pause off serveroutput on
DECLARE
hScn number := 0;
lScn number := 0;
sScn number;
ascn number;
alog varchar2(1000);
begin
  select min(start_scn), min(applied_scn) into sScn, ascn from dba_capture ;
  DBMS_OUTPUT.ENABLE(2000);
  for cr in (select distinct(a.ckpt_scn)
             from system.logmnr_restart_ckpt\\$ a
             where a.ckpt_scn <= ascn and a.valid = 1
               and exists (select * from system.logmnr_log\\$ l
                   where a.ckpt_scn between l.first_change# and l.next_change#)
              order by a.ckpt_scn desc)
  loop
    if (hScn = 0) then
       hScn := cr.ckpt_scn;
    else
       lScn := cr.ckpt_scn;
       exit;
    end if;
  end loop;
  if lScn = 0 then
    lScn := sScn;
  end if;
   select min(sequence#) into alog from v\\$archived_log where lScn between first_change# and next_change#;
  dbms_output.put_line(alog);
end;
EOF`
  # if there are no mandatory keep archive, instead of a number we just get the "PLS/SQL successfull"
  ret=`echo $var | awk '{print $1}'`
  if [ ! "$ret" = "PL/SQL" ];then
     LAST_APPLIED=$ret
  else
     unset LOGMINER
  fi
fi
PERC_NOW=`get_perc_occup`
if [ $PERC_NOW -gt $MAX_PERC ];then
     cd $ARC_DIR
     cpt=`ls -tr *.$EXT_ARC | wc -w`
     if [ ! "x-$cpt" = "x-" ];then
          MID=`expr $cpt / $PART`
          cpt=0
          ls -tr *.$EXT_ARC |while read ARC
              do
                 cpt=`expr $cpt + 1`
                 if [ $cpt -gt $MID ];then
                      break
                 fi
                 if [ "$CHECK_APPLIED" = "YES" -o "$LOGMINER" = "TRUE" ];then
                    VAR=`echo $ARC | sed 's/.*_\([0-9][0-9]*\)\..*/\1/' | sed 's/[^0-9][^0-9].*//'`
                    if [ $VAR -gt $LAST_APPLIED ];then
                         continue
                    fi
                 fi
                 case $ACTION in
                      'compress' ) $COMPRESS_PRG $ARC_DIR/$ARC
                                 if [ "x-$VERBOSE" = "x-TRUE" ];then
                                       echo " `date +%d-%m-%Y' '%H:%M` : $ARC compressed using $COMPRESS_PRG" >> $LOG
                                 fi ;;
                      'delete' ) rm $ARC_DIR/$ARC
                                 if [ "x-$VERBOSE" = "x-TRUE" ];then
                                       echo " `date +%d-%m-%Y' '%H:%M` : $ARC deleted" >> $LOG
                                 fi ;;
                      'list'   )   ls -l $ARC_DIR/$ARC ;;
                      'move'   ) mv  $ARC_DIR/$ARC $TARGET_DIR
                                 if [ ! "x-$COMPRESS_PRG" = "x-" ];then
                                       $COMPRESS_PRG $TARGET_DIR/$ARC
                                       if [ "x-$VERBOSE" = "x-TRUE" ];then
                                             echo " `date +%d-%m-%Y' '%H:%M` : $ARC moved to $TARGET_DIR and compressed" >> $LOG
                                       fi
                                 else
                                       if [ "x-$VERBOSE" = "x-TRUE" ];then
                                             echo " `date +%d-%m-%Y' '%H:%M` : $ARC moved to $TARGET_DIR" >> $LOG
                                       fi
                                 fi ;;
                  esac
          done
      else
          echo "Warning : The filesystem is not full due to archive logs !"
          exit
      fi
elif [ "x-$VERBOSE" = "x-TRUE" ];then
     echo "Nothing to do at `date +%d-%m-%Y' '%H:%M`" >> $LOG
fi

Similar Messages

  • What is the meaning of stream capture waiting for archive log ?

    today I fing the oracle alert that stream capture waiting for archive log in other class,and use a lot of database time.
    What is it? And Can I prevent it happening ?

    today I fing the oracle alert that stream capture waiting for archive log in other class,and use a lot of database time.
    What is it? And Can I prevent it happening ?

  • Archive log miss : How to restart capture

    Hi Gurus,
    I configured hotlog CDC distributed on 10.2.0.4 Databases.
    I make a mistake: in my source Db I have deleted an Archive log,
    and now the state of Capture process in V_$STREAM_CAPTURE is "WAITING FOR REDO: LAST SCN MINED 930696".
    Now I'd like to restart the capture process from the next archive (just after the missed archive)
    How is it possible?
    tnk Fabio

    I'm sorry to tell you that, but it's not possible. (Just as it's not possible to recover database with missing logs...)
    You will have to recreate the capture process and to re-instantiate the replicated tables.
    Regards,

  • Archive log missing on standby: FAL[client]: Failed to request gap sequence

    My current environment is Oracle 10.2.0.4 with ASM 10.2.0.4 on a 2 node RAC in production and a standby that is the same setup. I'm also running on Oracle Linux 5. Almost daily now an archivelog doesnt make it to the standby and oracle doesnt seem to resolve the gap sequence from the primary. If I stop and restart recovery it gets the logfile and continues recovery just fine. I have checked my fal_client and fal_server settings and they look good. The logs after this error do continue to get written to the standby but the standby wont continue recovery until I stop and restart recovery and it fetches this missing log.
    The only thing I know thats happening is that the firewall people are disconnecting any connections that are inactive for 60 minutes and recently did an upgrade that they are claiming didnt change anything:)  I dont know if this is causing this problem or not. Any thoughts on what might be happening?
    Error in standby alert.log:
    Tue Jun 29 23:15:35 2010
    RFS[258]: Possible network disconnect with primary database
    Tue Jun 29 23:15:36 2010
    Fetching gap sequence in thread 2, gap sequence 9206-9206
    Tue Jun 29 23:16:46 2010
    FAL[client]: Failed to request gap sequence
    GAP - thread 2 sequence 9206-9206
    DBID 661398854 branch 714087609
    FAL[client]: All defined FAL servers have been attempted.
    Error on primary alert.log:
    Tue Jun 29 23:00:07 2010
    ARC0: Creating remote archive destination LOG_ARCHIVE_DEST_2: 'WSSPRDB' (thread 1 sequence 9265)
    (WSSPRD1)
    ARC0: Transmitting activation ID 0x29c37469
    Tue Jun 29 23:00:07 2010
    Errors in file /u01/app/oracle/admin/WSSPRD/bdump/wssprd1_arc0_14024.trc:
    ORA-03135: connection lost contact
    FAL[server, ARC0]: FAL archive failed, see trace file.
    Tue Jun 29 23:00:07 2010
    Errors in file /u01/app/oracle/admin/WSSPRD/bdump/wssprd1_arc0_14024.trc:
    ORA-16055: FAL request rejected
    ARCH: FAL archive failed. Archiver continuing
    Tue Jun 29 23:00:07 2010
    ORACLE Instance WSSPRD1 - Archival Error. Archiver continuing.
    Tue Jun 29 23:00:41 2010
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    Tue Jun 29 23:00:41 2010
    FAL[server, ARC2]: Begin FAL archive (dbid 0 branch 714087609 thread 2 sequence 9206 dest WSSPRDB)
    FAL[server, ARC2]: FAL archive failed, see trace file.
    Tue Jun 29 23:00:43 2010
    Errors in file /u01/app/oracle/admin/WSSPRD/bdump/wssprd1_arc2_14028.trc:
    ORA-16055: FAL request rejected
    ARCH: FAL archive failed. Archiver continuing
    Tue Jun 29 23:00:43 2010
    ORACLE Instance WSSPRD1 - Archival Error. Archiver continuing.
    Tue Jun 29 23:01:16 2010
    Redo Shipping Client Connected as PUBLIC
    -- Connected User is Valid
    Tue Jun 29 23:15:01 2010
    Thread 1 advanced to log sequence 9267 (LGWR switch)
    I have checked the trace files that get spit out but they arent anything meaningful to me as to whats really happening. Snipit of the trace file:
    tkcrrwkx: Starting to process work request
    tkcrfgli: SRL header: 0
    tkcrfgli: SRL tail: 0
    tkcrfgli: ORL to arch: 4
    tkcrfgli: le# seq thr for bck tba flags
    tkcrfgli: 1 359 1 2 0 3 0x0008 ORL active cur
    tkcrfgli: 2 358 1 0 1 1 0x0000 ORL active
    tkcrfgli: 3 361 2 4 0 0 0x0008 ORL active cur
    tkcrfgli: 4 360 2 0 3 2 0x0000 ORL active
    tkcrfgli: 5 -- entry deleted --
    tkcrfgli: 6 -- entry deleted --
    tkcrfgli: 7 -- entry deleted --
    tkcrfgli: 8 -- entry deleted --
    tkcrfgli: 9 -- entry deleted --
    tkcrfgli: 191 -- entry deleted --
    tkcrfgli: 192 -- entry deleted --
    *** 2010-03-27 01:30:32.603 20998 kcrr.c
    tkcrrwkx: Request from LGWR to perform: <startup>
    tkcrrcrlc: Starting CRL ARCH check
    *** 2010-03-27 01:30:32.603 66085 kcrr.c
    Beginning controlfile transaction 0x0x7fffd0b53198 [kcrr.c:20395 (14011)]
    *** 2010-03-27 01:30:32.645 66173 kcrr.c
    Acquired controlfile transaction 0x0x7fffd0b53198 [kcrr.c:20395 (14024)]
    *** 2010-03-27 01:30:32.649 66394 kcrr.c
    Ending controlfile transaction 0x0x7fffd0b53198 [kcrr.c:20397]
    tkcrrasgn: Checking for 'no FAL', 'no SRL', and 'HB' ARCH process
    # HB NoF NoS CRL Name
    29 NO NO NO NO ARC0
    28 NO YES YES NO ARC1
    27 NO NO NO NO ARC2
    26 NO NO NO NO ARC3
    25 YES NO NO NO ARC4
    24 NO NO NO NO ARC5
    23 NO NO NO NO ARC6
    22 NO NO NO NO ARC7
    21 NO NO NO NO ARC8
    20 NO NO NO NO ARC9
    Thanks.
    Kristi

    It's the network that's messing up; unlikely due to firewall timeout as it waits for 60 minutes and you are switching every 15 minutes. There may be some other network glitch that needs rectified.
    In any case - arch file missing/ corrupt / halfway through - FAL setting should have refetched the problematic archive log automatically.
    As many had suggested already, the best way to resolve RFS issues I believe is to use real-time apply by configuring standby redo logs. It's very easy to configure it and you can opt for real-time apply even in max-performance mode that you are using right now.
    Even though you are maintaining (I guess) 1-1 between primary & standby instances, you can provide both primary instances in fal_server (like fal_server=string1,string2). See if that helps.
    lastly, check if you are having simiar issue at other times as well that might be getting rectified automatically as expected.
    col message for a80
    col time for a20
    select message, to_char(timestamp,'dd-mon-rr hh24:mi:ss') time
    from v$dataguard_status
    where severity in ('Error','Fatal')
    order by timestamp;
    Cheers.

  • Redo Archive Logs Missing

    Hi Gurus
    While Configuring Data Guard for ORacle 10g (10.2.0.4) 64 bits on Windows 2007 Server 64 bits.
    I got few questions
    1. What is the Default mode of Standby Database?
    2. Should we Always Start Physical Standby Database to Recover Missing Redo Archive Log?
    SQL> startup mount;
    ORACLE instance started.
    Total System Global Area 591396864 bytes
    Fixed Size 2067496 bytes
    Variable Size 163578840 bytes
    Database Buffers 419430400 bytes
    Redo Buffers 6320128 bytes
    Database mounted.
    SQL> alter databse recover managed standby database disconnect from session;
                   Database altered.
    3. When there are missing Redo Log Archives e.g.
    ----On Standby Database--------
    SQL> SELECT RESETLOGS_ID,SEQUENCE#,STATUS,ARCHIVED FROM V$ARCHIVED_LOG
    2 ORDER BY RESETLOGS_ID,SEQUENCE#;
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 15 A YES
    812980008 16 A YES
    812980008 17 A YES
    812980008 18 A YES
    812980008 19 A YES
    812980008 20 A YES
    812980008 21 A YES
    812980008 22 A YES
    812980008 23 A YES
    812980008 24 A YES
    812980008 25 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 26 A YES
    812980008 27 A YES
    812980008 28 A YES
    812980008 29 A YES
    812980008 30 A YES
    812980008 31 A YES
    812980008 32 A YES
    812980008 33 A YES
    812980008 34 A YES
    812980008 35 A YES
    812980008 36 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 37 A YES
    812980008 38 A YES
    812980008 39 A YES
    812980008 40 A YES
    812980008 41 A YES
    812980008 42 A YES
    812980008 43 A YES
    29 rows selected.
    ---------------On Primary Database---------------------
    SQL> SELECT RESETLOGS_ID,SEQUENCE#,STATUS,ARCHIVED FROM V$ARCHIVED_LOG
    2 ORDER BY RESETLOGS_ID,SEQUENCE# ;
    RESETLOGS_ID SEQUENCE# S ARC
    *812980008 8 A YES*
    *812980008 9 A YES*
    *812980008 10 A YES*
    *812980008 11 A YES*
    *812980008 12 A YES*
    *812980008 13 A YES*
    *812980008 14 A YES*
    812980008 15 A YES
    812980008 15 A YES
    812980008 16 A YES
    812980008 16 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 17 A YES
    812980008 17 A YES
    812980008 18 A YES
    812980008 18 A YES
    812980008 19 A YES
    812980008 19 A YES
    812980008 20 A YES
    812980008 20 A YES
    812980008 21 A YES
    812980008 21 A YES
    812980008 22 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 22 A YES
    812980008 23 A YES
    812980008 23 A YES
    812980008 24 A YES
    812980008 24 A YES
    812980008 25 A YES
    812980008 25 A YES
    812980008 26 A YES
    812980008 26 A YES
    812980008 27 A YES
    812980008 27 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 28 A YES
    812980008 28 A YES
    812980008 29 A YES
    812980008 29 A YES
    812980008 30 A YES
    812980008 30 A YES
    812980008 31 A YES
    812980008 31 A YES
    812980008 32 A YES
    812980008 32 A YES
    812980008 33 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 33 A YES
    812980008 34 A YES
    812980008 34 A YES
    812980008 35 A YES
    812980008 35 A YES
    812980008 36 A YES
    812980008 36 A YES
    812980008 37 A YES
    812980008 37 A YES
    812980008 38 A YES
    812980008 38 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 39 A YES
    812980008 39 A YES
    812980008 40 A YES
    812980008 40 A YES
    812980008 41 A YES
    812980008 41 A YES
    812980008 42 A YES
    812980008 42 A YES
    812980008 43 A YES
    812980008 43 A YES
    65 rows selected.
    Log 8, 9, 10, 11, 12, 13, 14, 15 are missing.
    How to Apply / Recover These Logs on Standby Database?
    Regards
    Thunder2777

    Hi
    Thunder2777 wrote:
    Hi Gurus
    While Configuring Data Guard for ORacle 10g (10.2.0.4) 64 bits on Windows 2007 Server 64 bits.
    I got few questions
    1. What is the Default mode of Standby Database?
    What is default mode? I think you want ask in what mode standby database is apply redo logs.
    Standby database can apply received redo only MOUNT mode, (your version is 10g, after 11g can apply open mode with READ ONLY WITH APPLY)
    2. Should we Always Start Physical Standby Database to Recover Missing Redo Archive Log?
    If Standby database opened mount mode, then database can receive redo.
    If you are start Redo Apply then MPR can request from primary for missing redo logs.
    SQL> startup mount;
    ORACLE instance started.
    Total System Global Area 591396864 bytes
    Fixed Size 2067496 bytes
    Variable Size 163578840 bytes
    Database Buffers 419430400 bytes
    Redo Buffers 6320128 bytes
    Database mounted.
    SQL> alter databse recover managed standby database disconnect from session;
                   Database altered.
    It is started recovery, in other words Redo Apply (MRP0 process)
    >
    3. When there are missing Redo Log Archives e.g.
    ----On Standby Database-------->
    SQL> SELECT RESETLOGS_ID,SEQUENCE#,STATUS,ARCHIVED FROM V$ARCHIVED_LOG
    2 ORDER BY RESETLOGS_ID,SEQUENCE#;
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 15 A YES
    812980008 16 A YES
    812980008 17 A YES
    812980008 18 A YES
    812980008 19 A YES
    812980008 20 A YES
    812980008 21 A YES
    812980008 22 A YES
    812980008 23 A YES
    812980008 24 A YES
    812980008 25 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 26 A YES
    812980008 27 A YES
    812980008 28 A YES
    812980008 29 A YES
    812980008 30 A YES
    812980008 31 A YES
    812980008 32 A YES
    812980008 33 A YES
    812980008 34 A YES
    812980008 35 A YES
    812980008 36 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 37 A YES
    812980008 38 A YES
    812980008 39 A YES
    812980008 40 A YES
    812980008 41 A YES
    812980008 42 A YES
    812980008 43 A YES
    29 rows selected.
    ---------------On Primary Database---------------------
    SQL> SELECT RESETLOGS_ID,SEQUENCE#,STATUS,ARCHIVED FROM V$ARCHIVED_LOG
    2 ORDER BY RESETLOGS_ID,SEQUENCE# ;
    RESETLOGS_ID SEQUENCE# S ARC
    *812980008 8 A YES*
    *812980008 9 A YES*
    *812980008 10 A YES*
    *812980008 11 A YES*
    *812980008 12 A YES*
    *812980008 13 A YES*
    *812980008 14 A YES*
    812980008 15 A YES
    812980008 15 A YES
    812980008 16 A YES
    812980008 16 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 17 A YES
    812980008 17 A YES
    812980008 18 A YES
    812980008 18 A YES
    812980008 19 A YES
    812980008 19 A YES
    812980008 20 A YES
    812980008 20 A YES
    812980008 21 A YES
    812980008 21 A YES
    812980008 22 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 22 A YES
    812980008 23 A YES
    812980008 23 A YES
    812980008 24 A YES
    812980008 24 A YES
    812980008 25 A YES
    812980008 25 A YES
    812980008 26 A YES
    812980008 26 A YES
    812980008 27 A YES
    812980008 27 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 28 A YES
    812980008 28 A YES
    812980008 29 A YES
    812980008 29 A YES
    812980008 30 A YES
    812980008 30 A YES
    812980008 31 A YES
    812980008 31 A YES
    812980008 32 A YES
    812980008 32 A YES
    812980008 33 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 33 A YES
    812980008 34 A YES
    812980008 34 A YES
    812980008 35 A YES
    812980008 35 A YES
    812980008 36 A YES
    812980008 36 A YES
    812980008 37 A YES
    812980008 37 A YES
    812980008 38 A YES
    812980008 38 A YES
    RESETLOGS_ID SEQUENCE# S ARC
    812980008 39 A YES
    812980008 39 A YES
    812980008 40 A YES
    812980008 40 A YES
    812980008 41 A YES
    812980008 41 A YES
    812980008 42 A YES
    812980008 42 A YES
    812980008 43 A YES
    812980008 43 A YES
    65 rows selected.
    Log 8, 9, 10, 11, 12, 13, 14, 15 are missing.It is no missing, you are created standby database, after sequence 15.
    As you know , if a sequence redo is not applied, then after is sequenced redo log is cannot apply to standby database.
    It means GAP.
    There have 43 archived redo log, and your last sequenced archive log received by standby database
    and applied.
    You can check with following scripts, too
    select max(Sequence#) from v$archived_log; -- on primary
    select max(Sequence#) from v$archived_log where applied = 'YES';  - on standby  Regards
    Mahir M. Quluzade

  • Archived log missed in standby database

    Hi,
    OS; Windows 2003 server
    Oracle: 10.2.0.4
    Data Guard: Max Performance
    Dataguard missed some of the archivelog files and but latest log files are applying. standby database is not in sync with primary.
    SELECT LOCAL.THREAD#, LOCAL.SEQUENCE# FROM (SELECT THREAD#, SEQUENCE# FROM V$ARCHIVED_LOG WHERE DEST_ID=1) LOCAL WHERE LOCAL.SEQUENCE# NOT IN (SELECT SEQUENCE# FROM V$ARCHIVED_LOG WHERE DEST_ID=2 AND THREAD# = LOCAL.THREAD#);
    I queried above command and I found some files are missed in standby.
    select status, type, database_mode, recovery_mode,protection_mode, srl, synchronization_status,synchronized from V$ARCHIVE_DEST_STATUS where dest_id=2;
    STATUS TYPE DATABASE_MODE RECOVERY_MODE PROTECTION_MODE SRL SYNCHRONIZATION_STATUS SYN
    VALID PHYSICAL MOUNTED-STANDBY MANAGED MAXIMUM PERFORMANCE NO CHECK CONFIGURATION NO
    Anyone can tell me how to apply those missed archive log files.
    Thanks in advacne

    Deccan Charger wrote:
    I got below error.
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION
    ERROR at line 1:
    ORA-01153: an incompatible media recovery is activeYou need to essentially do the following.
    1) Stop managed recovery on the standby.
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;2) Resolve the archive log gap - if you have configured FAL_SERVER and FAL_CLIENT Oracle should do this when you follow step 3 below, as you've manually copied the missed logs you should be OK
    3) restart managed recovery using the command shown above.
    You can monitor archive log catchup using the alert.log or your original query.
    Niall Litchfield
    http://www.orawin.info/
    Edited by: Niall Litchfield on May 4, 2010 2:29 PM
    missed tag

  • Status of db backup taken after 2 logs archived went missing

    Hi ,
    what would be the effect or issues of the full db backup & archivelog backup of 9i db taken after 2 archive logs missing that were never backed up, since backup, or crosscheck did't alerted any error messages. will this be still a valid backup to restore or recovery in case.
    posting the backup script details;
    note:we are not using catalog repository
    run {
    allocate channel dev1 type disk format 'Z:\Rbackup\DMP\db_t%t_s%s_p%p';
    allocate channel dev2 type disk format 'Z:\Rbackup\DMP\db_t%t_s%s_p%p';
    backup database;
    backup archivelog all;
    delete noprompt archivelog until time 'sysdate-1' ;
    crosscheck archivelog all;
    release channel dev1 ;
    release channel dev2 ;
    }

    logs missing date is Jan 14th and no backups were taken on this day. In that case, the "BACKUP ARCHIVELOG ALL" on the 15th would have generated an error flagging the missing archivelogs because the "ALL" specification is a direction to backup all known ArchiveLogs that are available for backup. The only reasons why it wouldn't have flagged the missing ArchiveLogs would be :
    a. You have another job that deletes archivelogs and this one ran before the backup on 15th Jan
    b. The ArchiveLogs are in the FRA and Oracle had automatically purged ArchiveLogs when the FRA was running low on free space.
    The Backup on 15th is good for restore and recovery. Use the "VALIDATE" command in RMAN to validate the database and archivelogs.
    See http://hemantoracledba.blogspot.com/2010/03/misinterpreting-restore-database.html
    However, as two archivelogs are missing and not backed up, you will not be able to do a Point In Time Recovery to any point in time from the time of the first missing archivelog to the end of the backup of the 15th.
    Hemant K Chitale

  • RMAN-6089: archive log not found or out of sync with catalog

    Hi:
    I am using OEM, not RMAN, to do a DB backup and got the above error message. I think the problem is caused by some archive logs missing from my NT machine.
    I tried to manually copy from an existing archive log and rename it to the missing archive log but it didn't work.
    I then tried to backup the archive logs only, not whole DB, and specified "delete input" option, and the process completed successfully.
    However, when I tried to re-do the whole DB backup, it failed with the same error.
    I looked at v$archived_log and I can see all archive logs, including the ones from previous incarnation and the missing ones from the current incarnation. How can I delete them or change their status (from OEM not from RMAN) from not deleted (NO) to deleted (YES) in order to perform a DB backup?

    Hi,
    I'm using controlfile for rman. I've issued "change archivelog all validate" before a full backup Before the full backup, I had a incremental backup process crashed(don't know why), which left some entres in v$archived_log. I have no problem completing the rman backup, but "restore archivelog all validate" command gives me "falure of restore..." and all those seq can be found in v$archived_log with "yes" for deleted column. Does that mean controlfile is still not in sync with the system? I don't need those target seq because those were before the full backup. Will there be potential problem later on during recovery? How do I (or when Oracle server will) get rid of those entries in v$archived_log? Thanks.
    RMAN> restore archivelog all validate;
    Starting restore at 11-AUG-05
    using target database controlfile instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: sid=28 devtype=DISK
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of restore command at 08/11/2005 10:20:24
    RMAN-06026: some targets not found - aborting restore
    RMAN-06025: no backup of log thread 1 seq 1048 scn 122640071 found to restore
    RMAN-06025: no backup of log thread 1 seq 1047 scn 122572176 found to restore
    RMAN-06025: no backup of log thread 1 seq 1046 scn 122431604 found to restore
    SQL> select name, completion_time,deleted from v$archived_log;
    NAME COMPLETIO DEL
    /u04/arch/arch_1_1046.arc 10-AUG-05 YES
    /u04/arch/arch_1_1047.arc 10-AUG-05 YES
    /u04/arch/arch_1_1048.arc 11-AUG-05 YES
    /u04/arch/arch_1_1049.arc 11-AUG-05 YES
    /u04/arch/arch_1_1050.arc 11-AUG-05 YES

  • Resetting SCN from removed Capture Process

    I've come across a problem in Oracle Streams where the Capture Processes seem to get stuck. There are no reported errors in the alert log and no trace files, but the capture process fails to continue capturing changes. It stays enabled, but in an awkward state where the OEM Console reports zeros across the board (0 messages, 0 enqueued), when in fact there had been accurate totals in the past.
    Restarting the Capture process does no good. The Capture process seems to switch its state back and forth from Dictionary Initialization to Initializing and vice versa. The only thing that seems to kickstart Streams again is to remove the Capture process and recreate the same process.
    However my problem is that I want to set the start_scn of the new capture process to the captured_scn of the remove capture process so that the new one can start from where the old one left off? However, I'm getting an error that this cannot be performed (cannot capture from specified SCN).
    Am I understanding this correctly? Or should the new Capture process start from where the removed left off automatically?
    Thanks

    Hi,
    I seem to have the same problem.
    I now have a latency of round about 3 days while nothing happened in the database so I want to be able to set the capture process to a further SCN. Setting the Start_SCN gives me an error (can't remember it now unfortunately). Somethimes it seems that the capture process gets stuck in an archived log. It then takes a long time for it to go further and when it goes further it sprints through a bunch of logs before it gets stuck again. During that time all the statuses look good, no heavy cpu-usage is monitored. We saw that the capture-builder has the highest cpu-load, where I would expect the capture-reader to be busy.
    I am able to set the first_scn. So a rebuild of the logminer dictionary might help a bit. But then again: why would the capture process need such a long time to process the archived-logs where no relevant events are expected.
    In my case the Streams solution is considered as a candidate for a replication solution where Quest's Sharedplex is considered to be expensive and unable to meet the requirements. One main reason it is considered inadaquate is that it is not able to catch up after a database-restart of a heavy batch. Now it seems that our capture-process might suffer from the same problem. I sincerly hope I'm wrong and it proofs to be capable.
    Regards,
    Martien

  • Recovery window retention policy deletes archive logs before a backup?

    Hi All,
    Oracle 11G on Windows 2008 R2
    This afternoon, I changed my RMAN retention policy from 'RETENTION POLICY REDUNDANCY 3' to RETENTION POLICY RECOVERY WINDOW OF 3 DAYS'
    However, i checked tonight and after my daily backup ran, all the archive logs prior to the backup had been deleted! Thus meaning i dont think i can restore to any point in time, prior to my daily backup. All the .arc logs were there after the backup. So i tried another test and kicked off the daily backup again. During the backup process, the archive logs got deleted again! Now i don't have any archive logs..
    Is this proper behaviour of RETENTION POLICY RECOVERY WINDOW?? I thought it would keep all the files needed for me to restore to any point in time for the previous 3 days. When i used REDUNDANCY, with my daily backups, it kept 3 days worth of backups+ archive logs so i could restore point-in-time to any point. How can i use RECOVERY WINDOW so that i can actually restore to any point-in-time for the 3 days??
    I wanted to change to RECOVERY WINDOW since i read that by using REDUNDANCY it only keeps X copies of a backup (so if i ran the backup 3 times in a day, i would only have those 3).
    Thanks in advance.

    Hi All,
    Here is the SHOW ALL output:
    RMAN> show all;
    RMAN configuration parameters for database with db_unique_name MMSPRD7 are:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 3 DAYS;
    CONFIGURE BACKUP OPTIMIZATION OFF; # default
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '%F'; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE COMPRESSION ALGORITHM 'BASIC' AS OF RELEASE 'DEFAULT' OPTIMIZE FOR LOA
    D TRUE ; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO 'D:\ORACLE\DATABASE\ORA11G\DATABASE\SNCFM
    MSPRD7.ORA'; # default
    Here is the RMAN script:
    Recovery Manager: Release 11.2.0.3.0 - Production on Mon Jan 20 23:03:12 2014
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    connected to target database: XXX (DBID=)
    RMAN> CROSSCHECK ARCHIVELOG ALL;
    2> CROSSCHECK BACKUPSET;
    3> CROSSCHECK BACKUP;
    4> CROSSCHECK COPY;
    5> DELETE NOPROMPT EXPIRED ARCHIVELOG ALL;
    6> DELETE NOPROMPT EXPIRED BACKUPSET;
    7> DELETE NOPROMPT OBSOLETE;
    8> BACKUP CURRENT CONTROLFILE;
    9> BACKUP AS COMPRESSED BACKUPSET DATABASE PLUS ARCHIVELOG DELETE INPUT;
    Also, not sure if you needed the whole RMAN output, but here is the deletion part:
    RMAN retention policy will be applied to the command
    RMAN retention policy is set to recovery window of 3 days
    using channel ORA_DISK_1
    Deleting the following obsolete backups and copies:
    Type                 Key    Completion Time    Filename/Handle
    Backup Set           1392   15-JAN-14        
      Backup Piece       1392   15-JAN-14          F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_15\O1_MF_NNNDF_TAG20140115T190054_9FG89R8N_.BKP
    Backup Set           1393   15-JAN-14        
      Backup Piece       1393   15-JAN-14          F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_15\O1_MF_ANNNN_TAG20140115T192204_9FG9KDHX_.BKP
    Backup Set           1397   16-JAN-14        
      Backup Piece       1397   16-JAN-14          F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_16\O1_MF_ANNNN_TAG20140116T190027_9FJWNW6L_.BKP
    Backup Set           1400   17-JAN-14        
      Backup Piece       1400   17-JAN-14          F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_17\O1_MF_ANNNN_TAG20140117T190138_9FMK349M_.BKP
    deleted backup piece
    backup piece handle=F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_15\O1_MF_NNNDF_TAG20140115T190054_9FG89R8N_.BKP RECID=1392 STAMP=836938856
    deleted backup piece
    backup piece handle=F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_15\O1_MF_ANNNN_TAG20140115T192204_9FG9KDHX_.BKP RECID=1393 STAMP=836940124
    deleted backup piece
    backup piece handle=F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_16\O1_MF_ANNNN_TAG20140116T190027_9FJWNW6L_.BKP RECID=1397 STAMP=837025228
    deleted backup piece
    backup piece handle=F:\ORAFRA\MMSPRD7\BACKUPSET\2014_01_17\O1_MF_ANNNN_TAG20140117T190138_9FMK349M_.BKP RECID=1400 STAMP=837111700
    Deleted 4 objects

  • How to monitor the successful archive log shipment to the standby database

    Hello,
    From the primary database, which data dictionary do I view to get information whether the standby database has received and processed the archive logs from the primary database. Both databases are 11202 on linux. Another question, which views on the primary database contain a flag sent by the standby database to the primary database to let the primary database knows that the standby database is up and functional? My purpose is to query the primary database for infomation that tell me the standby database is alive and functioning. I want to only query the primary database for these information about the standby database. Thank you in advance.

    watson2011 wrote:
    Hello,
    From the primary database, which data dictionary do I view to get information whether the standby database has received and processed the archive logs from the primary database. Both databases are 11202 on linux. Another question, which views on the primary database contain a flag sent by the standby database to the primary database to let the primary database knows that the standby database is up and functional? My purpose is to query the primary database for infomation that tell me the standby database is alive and functioning. I want to only query the primary database for these information about the standby database. Thank you in advance.You can troubleshoot standby by views as
    v$managed_standby (standby)
    v$dataguard_status
    v$dataguard_stats
    Some of the new views are introduced from 11g, check below link too
    http://docs.oracle.com/cd/B28359_01/server.111/b28294/views.htm#i79129

  • Archive Log Gap between Disaster recovery server & Production Server

    Hi
        Can anybody provide me the script  to find out the archive log missing on Production Server means that archive log backed up & deleted from the production server & rest of the Archive logs has been shipped on to the DR server. Means it will shot the gap between production server & DR server until that log which has been deleted from the production server , but next logs has been shipped to  the DR server. But rest of the logs only be applied when delete log be restored on the production server & shipped on the DR server.
    thanks.
    Regards
    Ravi Kant Arya

    Ravi,
    The question is how are you getting the archive logs shipped over to your standby database?  If you are doing it via FTP or scripts, then you may want to look at configuring DataGuard to do it for you.  Look at "Setting Up Oracle Dataguard for SAP" at:
    http://www.oracle.com/us/solutions/sap/wp-ora4sap-dataguard11g-303811.pdf
    Good Luck.
    Mike Kennedy

  • CAPTURE process error - missing Archive log

    Hi -
    I am getting cannot open archived log 'xxxx.arc' message when I try to start a newly created capture process. The archive files have been moved by the DBAs.
    Is there a way to set the capture process to start from a new archive ?
    I tried      
    exec DBMS_CAPTURE_ADM.ALTER_CAPTURE ( capture_name => 'STRMADMIN_SCH_CAPTURE', start_scn =>9668840362577);
    I got the new scn from DBMS_FLASHBACK.GET_SYSTEM_CHANGE_NUMBER();.
    But I still get the same error.
    Any ideas ?
    Thanks,
    Sadeepa

    If you are on 9i, I know that trying to reset the scn that way won't work. You have to drop and recreate the capture process. You can leave all the rules and rulesets in place, but I think you have to prepare all of the tables again.

  • Local capture only seems to process archived logs - not online logs

    In Oracle 10g R2 I have set up a local capture and apply on a logical standby database. I am configured to capture table level changes with LOGFILE_ASSIGNMENT = IMPLICIT, but the capture process only seems to find changes when the local redo logs are archived. I read that local capture will try to read the online redo if it can and then resort to the archive log. I don't see anything on why the online redo couldn't be used (except if LOGFILE_ASSIGNMENT = EXPLICIT). I want the capture to read the online redo rather than waiting for an archive. Can you help me? The status of the capture shows "WAITING FOR REDO" until archive time.
    -Nick

    Anybody else ever experience this?

  • Error running Archived-Log Downstream Capture Process

    I have created a Archived-Log Downstream Capture Process with ref. to following link
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28321/strms_ccap.htm#i1011654
    After executing the capture process get following error in trace
    ============================================================================
    Trace file /home/oracle/app/oracle/diag/rdbms/orcl/orcl/trace/orcl_cp01_13572.trc
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORACLE_HOME = /home/oracle/app/oracle/product/11.2.0/dbhome_1
    System name: Linux
    Node name: localhost.localdomain
    Release: 2.6.18-194.el5
    Version: #1 SMP Fri Apr 2 14:58:14 EDT 2010
    Machine: x86_64
    Instance name: orcl
    Redo thread mounted by this instance: 1
    Oracle process number: 37
    Unix process pid: 13572, image: [email protected] (CP01)
    *** 2011-08-20 14:21:38.899
    *** SESSION ID:(146.2274) 2011-08-20 14:21:38.899
    *** CLIENT ID:() 2011-08-20 14:21:38.899
    *** SERVICE NAME:(SYS$USERS) 2011-08-20 14:21:38.899
    *** MODULE NAME:(STREAMS) 2011-08-20 14:21:38.899
    *** ACTION NAME:(STREAMS Capture) 2011-08-20 14:21:38.899
    knlcCopyPartialCapCtx(), setting default poll freq to 0
    knlcUpdateMetaData(), before copy IgnoreUnsuperrTable:
    source:
    Ignore Unsupported Error Table: 0 entries
    target:
    Ignore Unsupported Error Table: 0 entries
    knlcUpdateMetaData(), after copy IgnoreUnsuperrTable:
    source:
    Ignore Unsupported Error Table: 0 entries
    target:
    Ignore Unsupported Error Table: 0 entries
    knlcfrectx_Init: rs=STRMADMIN.RULESET$_66, nrs=., cuid=0, cuid_prv=0, flags=0x0
    knlcObtainRuleSetNullLock: rule set name "STRMADMIN"."RULESET$_66"
    knlcObtainRuleSetNullLock: rule set name
    knlcmaInitCapPrc+
    knlcmaGetSubsInfo+
    knlqgetsubinfo
    subscriber name EMP_DEQ
    subscriber dblinke name
    subscriber name APPLY_EMP
    subscriber dblinke name
    knlcmaTerm+
    knlcmaTermSrvs+
    knlcmaTermSrvs-
    knlcmaTerm-
    knlcCCAInit()+, err = 26802
    knlcnShouldAbort: examining error stack
    ORA-26802: Queue "STRMADMIN"."STREAMS_QUEUE" has messages.
    knlcnShouldAbort: examing error 26802
    knlcnShouldAbort: returning FALSE
    knlcCCAInit: no combined capture and apply optimization err = 26802
    knlzglr_GetLogonRoles: usr = 91,
    knlqqicbk - AQ access privilege checks:
    userid=91, username=STRMADMIN
    agent=STRM05_CAPTURE
    knlqeqi()
    knlcRecInit:
    Combined Capture and Apply Optimization is OFF
    Apply-state checkpoint mode is OFF
    last_enqueued, last_acked
    0x0000.00000000 [0] 0x0000.00000000 [0]
    captured_scn, applied_scn, logminer_start, enqueue_filter
    0x0000.0004688c [288908] 0x0000.0004688c [288908] 0x0000.0004688c [288908] 0x0000.0004688c [288908]
    flags=0
    Starting persistent Logminer Session : 13
    krvxats retval : 0
    CKPT_FREE event=FALSE CCA=FALSE Checkptfreq=1000 AV/CDC flags=0
    krvxssp retval : 0
    krvxsda retval : 0
    krvxcfi retval : 0
    #1: krvxcfi retval : 0
    #2: krvxcfi retval : 0
    About to call krvxpsr : startscn: 0x0000.0004688c
    state before krvxpsr: 0
    dbms_logrep_util.get_checkpoint_scns(): logminer sid = 13 applied_scn = 288908
    dbms_logrep_util.get_checkpoint_scns(): prev_ckpt_scn = 0 curr_ckpt_scn = 0
    *** 2011-08-20 14:21:41.810
    Begin knlcDumpCapCtx:*******************************************
    Error 1304 : ORA-01304: subordinate process error. Check alert and trace logs
    Capture Name: STRM05_CAPTURE : Instantiation#: 65
    *** 2011-08-20 14:21:41.810
    ++++ Begin KNST dump for Sid: 146 Serial#: 2274
    Init Time: 08/20/2011 14:21:38
    ++++Begin KNSTCAP dump for : STRM05_CAPTURE
    Capture#: 1 Logminer_Id: 13 State: DICTIONARY INITIALIZATION [ 08/20/2011 14:21:38]
    Capture_Message_Number: 0x0000.00000000 [0]
    Capture_Message_Create_Time: 01/01/1988 00:00:00
    Enqueue_Message_Number: 0x0000.00000000 [0]
    Enqueue_Message_Create_Time: 01/01/1988 00:00:00
    Total_Messages_Captured: 0
    Total_Messages_Created: 0 [ 01/01/1988 00:00:00]
    Total_Messages_Enqueued: 0 [ 01/01/1988 00:00:00]
    Total_Full_Evaluations: 0
    Elapsed_Capture_Time: 0 Elapsed_Rule_Time: 0
    Elapsed_Enqueue_Time: 0 Elapsed_Lcr_Time: 0
    Elapsed_Redo_Wait_Time: 0 Elapsed_Pause_Time: 0
    Apply_Name :
    Apply_DBLink :
    Apply_Messages_Sent: 0
    ++++End KNSTCAP dump
    ++++ End KNST DUMP
    +++ Begin DBA_CAPTURE dump for: STRM05_CAPTURE
    Capture_Type: DOWNSTREAM
    Version:
    Source_Database: ORCL2.LOCALDOMAIN
    Use_Database_Link: NO
    Logminer_Id: 13 Logfile_Assignment: EXPLICIT
    Status: ENABLED
    First_Scn: 0x0000.0004688c [288908]
    Start_Scn: 0x0000.0004688c [288908]
    Captured_Scn: 0x0000.0004688c [288908]
    Applied_Scn: 0x0000.0004688c [288908]
    Last_Enqueued_Scn: 0x0000.00000000 [0]
    Capture_User: STRMADMIN
    Queue: STRMADMIN.STREAMS_QUEUE
    Rule_Set_Name[+]: "STRMADMIN"."RULESET$_66"
    Checkpoint_Retention_Time: 60
    +++ End DBA_CAPTURE dump
    +++ Begin DBA_CAPTURE_PARAMETERS dump for: STRM05_CAPTURE
    PARALLELISM = 1 Set_by_User: NO
    STARTUP_SECONDS = 0 Set_by_User: NO
    TRACE_LEVEL = 7 Set_by_User: YES
    TIME_LIMIT = -1 Set_by_User: NO
    MESSAGE_LIMIT = -1 Set_by_User: NO
    MAXIMUM_SCN = 0xffff.ffffffff [281474976710655] Set_by_User: NO
    WRITE_ALERT_LOG = TRUE Set_by_User: NO
    DISABLE_ON_LIMIT = FALSE Set_by_User: NO
    DOWNSTREAM_REAL_TIME_MINE = FALSE Set_by_User: NO
    MESSAGE_TRACKING_FREQUENCY = 2000000 Set_by_User: NO
    SKIP_AUTOFILTERED_TABLE_DDL = TRUE Set_by_User: NO
    SPLIT_THRESHOLD = 1800 Set_by_User: NO
    MERGE_THRESHOLD = 60 Set_by_User: NO
    +++ End DBA_CAPTURE_PARAMETERS dump
    +++ Begin DBA_CAPTURE_EXTRA_ATTRIBUTES dump for: STRM05_CAPTURE
    USERNAME Include:YES Row_Attribute: YES DDL_Attribute: YES
    +++ End DBA_CAPTURE_EXTRA_ATTRIBUTES dump
    ++ LogMiner Session Dump Begin::
    SessionId: 13 SessionName: STRM05_CAPTURE
    Start SCN: 0x0000.00000000 [0]
    End SCN: 0x0000.00046c2d [289837]
    Processed SCN: 0x0000.0004689e [288926]
    Prepared SCN: 0x0000.000468d4 [288980]
    Read SCN: 0x0000.000468e2 [288994]
    Spill SCN: 0x0000.00000000 [0]
    Resume SCN: 0x0000.00000000 [0]
    Branch SCN: 0x0000.00000000 [0]
    Branch Time: 01/01/1988 00:00:00
    ResetLog SCN: 0x0000.00000001 [1]
    ResetLog Time: 08/18/2011 16:46:59
    DB ID: 740348291 Global DB Name: ORCL2.LOCALDOMAIN
    krvxvtm: Enabled threads: 1
    Current Thread Id: 1, Thread State 0x01
    Current Log Seqn: 107, Current Thrd Scn: 0x0000.000468e2 [288994]
    Current Session State: 0x20005, Current LM Compat: 0xb200000
    Flags: 0x3f2802d8, Real Time Apply is Off
    +++ Additional Capture Information:
    Capture Flags: 4425
    Logminer Start SCN: 0x0000.0004688c [288908]
    Enqueue Filter SCN: 0x0000.0004688c [288908]
    Low SCN: 0x0000.00000000 [0]
    Capture From Date: 01/01/1988 00:00:00
    Capture To Date: 01/01/1988 00:00:00
    Restart Capture Flag: NO
    Ping Pending: NO
    Buffered Txn Count: 0
    -- Xid Hash entry --
    -- LOB Hash entry --
    -- No TRIM LCR --
    Unsupported Reason: Unknown
    --- LCR Dump not possible ---
    End knlcDumpCapCtx:*********************************************
    *** 2011-08-20 14:21:41.810
    knluSetStatus()+{
    *** 2011-08-20 14:21:44.917
    knlcapUpdate()+{
    Updated streams$_capture_process
    finished knlcapUpdate()+ }
    finished knluSetStatus()+ }
    knluGetObjNum()+
    knlsmRaiseAlert: keltpost retval is 0
    kadso = 0 0
    KSV 1304 error in slave process
    *** 2011-08-20 14:21:44.923
    ORA-01304: subordinate process error. Check alert and trace logs
    knlz_UsrrolDes()
    knstdso: state object 0xb644b568, action 2
    knstdso: releasing so 0xb644b568 for session 146, type 0
    knldso: state object 0xa6d0dea0, action 2 memory 0x0
    kadso = 0 0
    knldso: releasing so 0xa6d0dea0
    OPIRIP: Uncaught error 447. Error stack:
    ORA-00447: fatal error in background process
    ORA-01304: subordinate process error. Check alert and trace logs
    Any suggestions???

    Output of above query
    ==============================
    CAPTURE_NAME STATUS ERROR_MESSAGE
    STRM05_CAPTURE ABORTED ORA-01304: subordinate process error. Check alert and trace logs
    Alert log.xml
    =======================
    <msg time='2011-08-25T16:58:01.865+05:30' org_id='oracle' comp_id='rdbms'
    client_id='' type='UNKNOWN' level='16'
    host_id='localhost.localdomain' host_addr='127.0.0.1' module='STREAMS'
    pid='30921'>
    <txt>Errors in file /home/oracle/app/oracle/diag/rdbms/orcl/orcl/trace/orcl_cp01_30921.trc:
    ORA-01304: subordinate process error. Check alert and trace logs
    </txt>
    </msg>
    The orcl_cp01_30921.trc has the same thing posted in the first message.

Maybe you are looking for

  • TRCS 2LIS_03_BF_TR - CUBE 0IC_C03

    Hi Gurus TI am unable to solve the routine for RCS 2LIS_03_BF_TR -> CUBE 0IC_C03. It is giving me the following error. E:In PERFORM or CALL FUNCTION "ROUTINE_9998", the actual parameter "SOURCE_PACKAGE" is incompatible with the formal parameter "DATA

  • T code f-58 , issue

    hi gurus while doin f-58 for vendor payment am getting error as " Tax relevent accounts not defined in this area , please correct " could anybody please tell me what it is ????? please help me

  • Everything is read only?

    i accidentally clicked the "check out" option on my site folder while in dreamweaver and now all my files are read only. is there a way to undo this?

  • Cannot delete gmail contact list

    i am having problems syncing my verizon/yahoo account back to outlook on the computer.  I had heard that gmail accounts can cause these types of problems.  However when I go into Contact lists I see 4 listed there and I only want 1 contact list.  How

  • Connection using sql 2000

    i try to connect t o database using sql 2000 this my code and the execptions can anyone help me to tell me what is the right code to connect to database package javaapplication37; import java.sql.Connection; import java.sql.DriverManager; import java