Job number from alert log file to information

Hello!
I have a question about job numbers in alert log file. Today one of our Oracle 10g R2 [10.2.0.4] RAC nodes crashed. After examining alert log file for one of the nodes I saw a lot of messages like:
Tue Jul 26 11:52:43 2011
Errors in file /u01/app/oracle/admin/zeme/bdump/zeme2_j002_28952.trc:
ORA-12012: error on auto execute of job *20627358*
ORA-12705: Cannot access NLS data files or invalid environment specified
Tue Jul 26 11:52:43 2011
Errors in file /u01/app/oracle/admin/zeme/bdump/zeme2_j001_11018.trc:
ORA-12012: error on auto execute of job *20627357*
ORA-12705: Cannot access NLS data files or invalid environment specified
Tue Jul 26 11:52:43 2011
Errors in file /u01/app/oracle/admin/zeme/bdump/zeme2_j000_9684.trc:
ORA-12012: error on auto execute of job *20627342*
ORA-12705: Cannot access NLS data files or invalid environment specified
After examining trc files I have found no further information about error except session ids.
My question is: how to find what job caused these messages to appear in alert log file.
How do I map number in alert log file to some "real" information (owner, statement executed, schedule)?
Marx.

Sorry for the delay
Try this to find the job :
select job, what from dba_jobs ;
How do I find NLS_LANG version?SQL> show parameter NLS_LANG
Do you mean ALTER SESSION inside a job?I meant anywhere, but your question is better.
ORA-12705 - Common Reasons and How to Resolve Them [ID 158654.1]
If OS is Windows lookout for NLS_LANG=NA in the registry
Is it possible you are doing this somewhere ?
ALTER SESSION SET NLS_DATE_FORMAT = 'RRRR-MM-DD\"T\"HH24:MI:SS';NLS database settings are superseded by NLS instance settings
SELECT * from NLS_SESSION_PARAMETERS;
These are the settings used for the current SQL session.
NLS_LANG could be set in a profile for example.
NLS_LANG=_AMERICA.WE8ISO8859P1     ( correct )
NLS_LANG=AMERICA.WE8ISO8859P1 ( Incorrect )
you need to set the "_" as separator.
Windows
set NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P1
Unix
export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P1
mseberg
Edited by: mseberg on Jul 28, 2011 3:51 PM
Edited by: mseberg on Jul 29, 2011 4:05 AM

Similar Messages

  • Shell script for oracle alert.log file

    Hi Gurus,
    I wanted to write shell script to know the last 10 shutdown timings of the database from alert log file. I'm working on oracle 9i.
    Could anyone please advice on this.
    Thanks in advance
    regards,
    Shaan
    Edited by: Shaan_dmp on Jan 5, 2009 1:27 PM
    Edited by: Shaan_dmp on Jan 5, 2009 1:28 PM

    Use awk. I don't have a 9i to hand but here is a very simple version for 10g XE
    My awk file (the line numbers are for the notes below - don't include them:
    01:BEGIN { prevline = "";}
    02:
    03:/Completed: alter database close/ {print prevline,FS,$0;}
    04:
    05:{prevline = $0;}The command line and results (from my 300k alert log)
    $ awk -f alert.awk.txt alert_xe.log
    Fri Apr 11 18:08:40 2008   Completed: alter database close normal
    Fri May 16 18:53:21 2008   Completed: alter database close normal
    Tue May 20 17:28:23 2008   Completed: alter database close normal
    Thu Jul 17 19:08:52 2008   Completed: alter database close normal
    Fri Aug 15 15:12:48 2008   Completed: alter database close normal
    Wed Nov 05 08:52:59 2008   Completed: alter database close normal
    Fri Nov 14 16:36:03 2008   Completed: alter database close normal
    Tue Dec 09 10:46:23 2008   Completed: alter database close normal
    Mon Jan 05 11:12:22 2009   Completed: alter database close normalWhat it means:
    1) the BEGIN section at line 01 defines the variable to hold the previous line
    2) the /search string/ at line 03 finds the marker in the file for a shutdown, then does the required action (print the time which was in the previous line, and then this line; use FS (the awk field separator - normally space) as a separator
    3) at line 05 is an instruction we do on every line - so we remember it in case it is the timestamp for the shutdown.
    Now, you can include more of the corner cases for shutdowns by adding more search patterns etc. For more information, google for awk examples.
    Awk is really good at this sort of thing!
    HTH
    Regards Nigel

  • Will the informations  be recorded in the alert.log file? -----No.168

    will the informations about the loss of a temporary file be recorded in the alert.log file?

    Yes, because whe your database starts needs to "mount" a tablespace with temporary files (case of tablespace "TEMP"). But don't worry with a loss of this tablespace because doesn't contain nothing when database starts.

  • RMAN success, but errors in alert.log file

    My RMAN backup script runs well, but generates errors in alert.log file.
    Here is the trace file contents:
    /usr/lib/oracle/xe/app/oracle/admin/XE/udump/xe_ora_3990.trc
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    ORACLE_HOME = /usr/lib/oracle/xe/app/oracle/product/10.2.0/server
    System name: Linux
    Node name: plockton
    Release: 2.6.18-128.2.1.el5
    Version: #1 SMP Wed Jul 8 11:54:54 EDT 2009
    Machine: i686
    Instance name: XE
    Redo thread mounted by this instance: 1
    Oracle process number: 26
    Unix process pid: 3990, image: oracle@plockton (TNS V1-V3)
    *** 2009-07-23 23:05:01.835
    *** ACTION NAME:(0000025 STARTED111) 2009-07-23 23:05:01.823
    *** MODULE NAME:(backup full datafile) 2009-07-23 23:05:01.823
    *** SERVICE NAME:(SYS$USERS) 2009-07-23 23:05:01.823
    *** SESSION ID:(33.154) 2009-07-23 23:05:01.823
    *** 2009-07-23 23:05:18.689
    *** ACTION NAME:(0000045 STARTED111) 2009-07-23 23:05:18.689
    *** MODULE NAME:(backup archivelog) 2009-07-23 23:05:18.689
    Does anyone know why? Thanks.
    Richard

    I'm not sure if this will answer your question or not, but I believe these messages can likely be ignored.
    I'm currently running 10.2.0.1.0 Enterprise Edition in pre-production (yes, I know I should apply the latest patchset and I plan to do so as soon as I get a development box allocated to me and can test it's impact). I see the same types of messages that you've reported with each of my regularly-scheduled backups:
    a) The alert_<$SID>.log reports that there are errors in trace files:
    Mon Aug 10 04:33:49 2009
    Starting control autobackup
    Mon Aug 10 04:33:50 2009
    Errors in file /opt/oracle/admin/blah/udump/blah_ora_32520.trc:
    Mon Aug 10 04:33:50 2009
    Errors in file /opt/oracle/admin/blah/udump/blah_ora_32520.trc:
    Mon Aug 10 04:33:50 2009
    Errors in file /opt/oracle/admin/blah/udump/blah_ora_32520.trc:
    Control autobackup written to DISK device
    handle '/backup/physical/BLAH/RMAN/cf_c-2740124895-20090810-00'
    b) The .trc files, when you look at them contain no errors - only these "informational" messages:
    *** 2009-08-10 04:33:50.781
    *** ACTION NAME:(0000105 STARTED111) 2009-08-10 04:33:50.754
    *** MODULE NAME:(backup archivelog) 2009-08-10 04:33:50.754
    *** SERVICE NAME:(SYS$USERS) 2009-08-10 04:33:50.754
    *** SESSION ID:(126.28030) 2009-08-10 04:33:50.754
    c) I've verified that LOG_ARCHIVE_TRACE is set to 0:
    SQL*Plus> show parameter log_archive_trace
    NAME TYPE VALUE
    log_archive_trace integer 0
    As best I can discern from my own experience, these should just be ignored and I trust (read: "hope") they will simply go away once the latest patchset is applied. As for you running Oracle XE, a patchset is not an option, unfortunately.
    V/R
    -Eric

  • Ora-07445 reported in the alert log file

    Hi all,
    We are using the following platform:-
    OS: Solaris Operating System (SPARC 32-bit)
    Oracle Server - Enterprise Edition / Product Version: 9.2.0.5.0
    We encountered the following problem:-
    There is Ora-07445 reported in the alert log file.
    "ORA-07445: exception encountered: core dump [kgghash()+308] [SIGSEGV] [Address not mapped to object] [0x3222
    000] [] []reported in the database."
    These errors are signaled in more than 3 independent, unexplained, occurrences every day.
    When this error occurred while User was accessing the application system, by right the case will be shown on the system, but they were no case found.
    alert log are as below:-
    Fri Jul 27 09:12:30 2007
    Errors in file /disc3/oracle9205/RFDB/udump/rfdb_ora_27371.trc:
    ORA-07445: exception encountered: core dump [kgghash()+340] [SIGSEGV] [Address not mapped to object] [0x3184000] [] []
    Fri Jul 27 09:22:10 2007
    Thread 1 advanced to log sequence 10730
    Current log# 2 seq# 10730 mem# 0: /disc3/oracle9205/RFDB/RDO/logRFDB2a.rdo
    Current log# 2 seq# 10730 mem# 1: /disc3/oracle9205/RFDB/RDO/logRFDB2b.rdo
    Fri Jul 27 09:29:26 2007
    Errors in file /disc3/oracle9205/RFDB/udump/rfdb_ora_27372.trc:
    ORA-07445: exception encountered: core dump [kgghash()+296] [SIGSEGV] [Address not mapped to object] [0x3182000] [] []
    The applications have encountered ora-3113 after a short period of time followed by ora-3114.
    application log:-
    RF0120-1 2007-Jul-27 09:46:57] Load m[RF0120-1 2007-Jul-27 09:29:30] SQLCODE: -3113
    [RF0120-1 2007-Jul-27 09:29:30] Error Code -4105 returning from get score pan no.
    [RF0120-1 2007-Jul-27 09:29:30] Message type :120
    [RF0120-1 2007-Jul-27 09:29:30] Primary Account Number(PAN) DE0
    02 :5440640155262702
    [RF0120-1 2007-Jul-27 09:29:30] Processing code DE003 :003000
    [RF0120-1 2007-Jul-27 09:29:30] Transaction amount DE004 :000000000001
    [RF0120-1 2007-Jul-27 09:29:30] Settlement amount DE005 :000000000000
    [RF0120-1 2007-Jul-27 09:29:30] Transmission Date and time
    DE007 :0727092717
    [RF0120-1 2007-Jul-27 09:29:30] Settlement conversion rate DE009 :60263158
    [RF0120-1 2007-Jul-27 09:29:30] System trace audit no. DE011 :754710
    [RF0120-1 2007-Jul-27 09:29:30] Local transaction time DE012 :092717
    [RF0120-1 2007-Jul-27 09:29:30] Local transaction date DE013 :0727
    [RF0120-1 2007-Jul-27 09:29:30] Expiration date D
    E014 :0712
    [RF0120-1 2007-Jul-27 09:29:30] Settlement date DE015 :0727
    [RF0120-1 2007-Jul-27 09:29:30] Merchant type DE018 :5311
    [RF0120-1 2007-Jul-27 09:29:30] Point-of-service(POS) entry code DE022 :051
    [RF0120-1 2007-Jul-27 09:29:30] Acquiring inst. ID code DE032 :001912
    [RF0120-1 2007-Jul-27 09:29:30] Forwarding Inst. ID code DE033 :001912
    [RF0120-1 2007-Jul-27 09:29:30] Retrieval ref. no.
    DE037 :754710356390
    [RF0120-1 2007-Jul-27 09:29:30] Autholization ID response DE038 :356390
    [RF0120-1 2007-Jul-27 09:29:30] Response code DE039 :00
    [RF0120-1 2007-Jul-27 09:29:30] Card acceptor terminal ID DE041 :19306002
    [RF0120-1 2007-Jul-27 09:29:30] Card acceptor ID code DE042 :000001106
    020132
    [RF0120-1 2007-Jul-27 09:29:30] Card acceptor Name/Location
    What could have caused the above mentioned errors i.e Ora-07445 ; ora-3113 / ora-3114? How to resolve the problem.
    Please help .
    Thanks.

    i am also facing the same some time in ora 9.2.0.6 on Sun OS 9 SPARC 64-bits
    Errors in file /oracle/oracle9i/admin/FINPROD/udump/finprod_ora_6076.trc:
    ORA-07445: exception encountered: core dump [0000000100FDE088] [SIGSEGV] [Address not mapped to object] [0x00000013A] [] []
    Thu Aug 30 08:52:39 2007
    Errors in file /oracle/oracle9i/admin/FINPROD/udump/finprod_ora_6078.trc:
    ORA-07445: exception encountered: core dump [0000000100FDE088] [SIGSEGV] [Address not mapped to object] [0x00000013A] [] []
    Thu Aug 30 09:41:49 2007

  • I am geting errors in my database alert log file ora-12012 and ora 44003

    Hi all,
    I am using oracle 10g ...10.2.0.1.0 and OS is solaris 10.......
    i am geting these errors in my database alert log file last two days..i am not geting why it is coming?
    ORA-12012: error on auto execute of job 8887
    ORA-44003: invalid SQL name
    and trace file is i got i am posting here please help any one is aware about this....
    System name: SunOS
    Node name: obrmzone1
    Release: 5.10
    Version: Generic_137111-06
    Machine: sun4v
    Instance name: ------
    Redo thread mounted by this instance: 1
    Oracle process number: 144
    Unix process pid: 26874, image:--------- (J000)
    *** ACTION NAME:(AUTO_SPACE_ADVISOR_JOB) 2011-06-10 22:00:49.372
    *** MODULE NAME:(DBMS_SCHEDULER) 2011-06-10 22:00:49.372
    *** SERVICE NAME:(SYS$USERS) 2011-06-10 22:00:49.372
    *** SESSION ID:(110.7114) 2011-06-10 22:00:49.372
    *** 2011-06-10 22:00:49.372
    ORA-12012: error on auto execute of job 8887
    ORA-44003: invalid SQL name
    thank you in advance,
    Dav

    I just google with "ORA-44003: invalid SQL name" and found below link:
    http://www.oracle-base.com/articles/10g/dbms_assert_10gR2.php
    SIMPLE_SQL_NAME
    The SIMPLE_SQL_NAME function checks the input string conforms to the basic characteristics of a simple SQL name:
    The first character of the name is alphabetic.
    The name only contains alphanumeric characters or the "_", "$", "#"
    Quoted names must be enclosed by double quotes and may contain any characters, including quotes provided they are represented by two quotes in a row ("").
    The function ignores leading and trailing white spaces are ignored
    The length of the input string is not validated.

  • How to check particular error in alert log file

    Hi all,
    How to check particular error in alert log file,for supose if i get error in batabase yesterday 4 pm & today i want to check alert log file to get basic idea..it might be a big file so how to check that particular error..
    Thanks & regards,
    Eswar..

    What's your oracle version?
    If you are in 11g you can use adrci tool
    1- set homes diag/rdbms/orawiss/ORAWISS/ : the rdbms home
    2- show alert -P "MESSAGE_TEXT LIKE '%ORA-%'" -term : to find all the ORA-% errors in the alert file
    3- show alert -P "MESSAGE_TEXT LIKE '%ORA-%' and originating_timestamp > systimestamp-51 " -term : to find all the ORA-% errors in the alert file during the last 51 days,
    4- show alert -P "MESSAGE_TEXT LIKE '%ORA-%' and originating_timestamp > systimestamp-1/24 " -term : to find all the ORA-% errors in the alert file during the last hour,
    5- show alert -P "MESSAGE_TEXT LIKE '%ORA-12012%' and originating_timestamp > systimestamp-1/24 " -term : to find the particular ORA-12012 error in the alert file during the last hour,
    Example:
    [oracle@wissem wissem]$ adrci
    ADRCI: Release 11.2.0.1.0 - Production on Wed May 4 10:24:54 2011
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    ADR base = "/home/oracle/app/oracle"
    adrci> set homes diag/rdbms/orawiss/ORAWISS/
    adrci> show alert -P "MESSAGE_TEXT LIKE '%ORA-'" -term
    ADR Home = /home/oracle/app/oracle/diag/rdbms/orawiss/ORAWISS:
    adrci> show alert -P "MESSAGE_TEXT LIKE '%ORA-%'" -term
    ADR Home = /home/oracle/app/oracle/diag/rdbms/orawiss/ORAWISS:
    2010-12-11 19:45:41.289000 +01:00
    ORA-1109 signalled during: ALTER DATABASE CLOSE NORMAL...
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01152: file 1 was not restored from a sufficiently old backup
    ORA-01110: data file 1: '/home/oracle/app/oracle/oradata/ORAWISS/system01.dbf'
    ORA-1547 signalled during: ALTER DATABASE RECOVER  database until time '2011-01-21:10:48:00'  ...
    Errors in file /home/oracle/app/oracle/diag/rdbms/orawiss/ORAWISS/trace/ORAWISS_j000_5692.trc:
    ORA-12012: error on auto execute of job 29
    ORA-01435: user does not exist
    2011-03-15 11:39:37.571000 +01:00
    opiodr aborting process unknown ospid (31042) as a result of ORA-609
    2011-03-15 12:04:15.111000 +01:00
    opiodr aborting process unknown ospid (3509) as a result of ORA-609
    adrci>
    adrci> show alert -P "MESSAGE_TEXT LIKE '%ORA-%' and originating_timestamp > systimestamp-51 " -term
    ADR Home = /home/oracle/app/oracle/diag/rdbms/orawiss/ORAWISS:
    2011-03-15 10:19:45.316000 +01:00
    Errors in file /home/oracle/app/oracle/diag/rdbms/orawiss/ORAWISS/trace/ORAWISS_j006_5536.trc:
    ORA-12012: error on auto execute of job 26
    ORA-01435: user does not exist
    Errors in file /home/oracle/app/oracle/diag/rdbms/orawiss/ORAWISS/trace/ORAWISS_j000_5692.trc:
    ORA-12012: error on auto execute of job 29
    ORA-01435: user does not exist
    2011-03-15 11:39:37.571000 +01:00
    opiodr aborting process unknown ospid (31042) as a result of ORA-609
    2011-03-15 12:04:15.111000 +01:00
    opiodr aborting process unknown ospid (3509) as a result of ORA-609
    adrci>

  • "Error in the alert log file " - please help

    Hi ALL,
    I am getting following errors in the alert log frequently.
    My database version is:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bi
    PL/SQL Release 10.2.0.1.0 - Production
    CORE 10.2.0.1.0 Production
    TNS for IBM/AIX RISC System/6000: Version 10.2.0.1.0 - Productio
    NLSRTL Version 10.2.0.1.0 - Production
    Thu Aug 30 16:58:19 2007
    Memory Notification: Library Cache Object loaded into SGA
    Heap size 2172K exceeds notification threshold (2048K)
    KGL object name :select object_name from all_objects where object_name='TOAD_PROFILER' and object_type='PACKAGE'
    Thu Aug 30 17:09:36 2007
    ************************************************************************************

    Additionally you can visit here:
    Alert Error from trc files--Heap size 5236K exceeds notification threshold
    geting error in alert log file
    Adith

  • Trace file with different name is alert log file.

    I am strange today
    i found a trace file generated in alert log file with name
    /bdump/stlbas_cjq0_1880.trc:
    but when i am trying to find out in bdump folder i cannot found
    any file with this named. instead i found /bdump/stlbas_cjq0_1853.trc
    is there any possibility of being this scenario ?
    database version:- 10.2.0.4
    Os:- sun solaries spark 64 bits
    Regards
    Halim

    The trace file has been removed/deleted from an OS user. I think you should verify that with OS logs.
    Thanks

  • Errors in Alert log file

    Hi,
    I have encountered following errors in alrt log file, its strange to see FAST_START_MTTR_TARGET is out of valid and aksing to use another value and that value is changing every second, it should be in increasing order or it should be consistent. I think i have a serious problem, so anyone could help me out on this. I am getting this "FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 24381 instead" error regurarly.
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 24381 instead.
    Wed Jan 16 23:40:34 2008
    Completed checkpoint up to RBA [0x2aadb.2.10], SCN: 5473742880
    Wed Jan 16 23:45:51 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 24017 instead.
    Wed Jan 16 23:46:21 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 23647 instead.
    Wed Jan 16 23:46:52 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 23298 instead.
    Wed Jan 16 23:47:22 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 22906 instead.
    Wed Jan 16 23:47:52 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 22507 instead.
    Wed Jan 16 23:48:22 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 22133 instead.
    Wed Jan 16 23:48:52 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 21786 instead.
    Wed Jan 16 23:49:22 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 21443 instead.
    Wed Jan 16 23:49:52 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 21062 instead.
    Wed Jan 16 23:50:22 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 20698 instead.
    Wed Jan 16 23:53:53 2008
    Incremental checkpoint up to RBA [0x2aadb.31b4.0], current log tail at RBA [0x2aadb.3e38.0]
    Thu Jan 17 00:00:54 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 19582 instead.
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 16287 instead.
    Thu Jan 17 00:34:01 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 15902 instead.
    Thu Jan 17 00:34:31 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 14638 instead.
    Thu Jan 17 00:35:01 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 13600 instead.
    Thu Jan 17 00:35:31 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 12773 instead.
    Thu Jan 17 00:36:02 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 12460 instead.
    Thu Jan 17 00:36:32 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 12338 instead.
    Thu Jan 17 00:37:01 2008
    Process P187 died, see its trace file
    Thu Jan 17 00:37:02 2008
    FAST_START_MTTR_TARGET 121 is out of the valid MTTR range, use 12248 instead.
    Thu Jan 17 00:37:02 2008
    Process P188 died, see its trace file
    Thu Jan 17 00:37:03 2008
    Process P189 died, see its trace file
    Thu Jan 17 00:37:04 2008
    Process P190 died, see its trace file
    Thu Jan 17 00:37:16 2008
    Process J000 died, see its trace file
    Thu Jan 17 00:37:16 2008
    kkjcre1p: unable to spawn jobq slave process
    Thu Jan 17 00:37:16 2008
    Errors in file /oracle/BP3/saptrace/background/bp3_cjq0_4788.trc:
    Process J000 died, see its trace file
    Thu Jan 17 00:37:22 2008
    kkjcre1p: unable to spawn jobq slave process
    Thu Jan 17 00:37:22 2008
    Errors in file /oracle/BP3/saptrace/background/bp3_cjq0_4788.trc:
    Thanks
    Jafar

    This parameter, FAST_START_MTTR_TARGET ranges from 0 to 3600 seconds. It is important to specify your oracle version, since a similar issue is reported for the 10gR1 release.
    The solution for it would be to increase the value to the suggested one in the alert.log file, but it could be completely out of range from the valid values, so set it to zero and let the oracle instance calculate the value. You can check the output at the V$INSTANCE_RECOVERY view at the columns TARGET_MTTR, and ESTIMATED_MTTR prior to proceeding with modificaiton.
    On the other hand, I suggest you to read this metalink note
    Kkjcre1p: Unable To Spawn Jobq Slave Process, Error 1089
         Doc ID:      Note:344275.1
    for the Kkjcre1p error message.
    ~ Madrid

  • Oracle 9i (SE) Solaris 9 (Alert.log file)

    Hi Guys,
    How do I clean up my alert.log file. Do I have to delete it completely or only delete the content of the file.
    Thank you.

    Laura Gaigala wrote:
    Why You want to clean up it? Too large or some other reasons?
    If You have some planned db restart, then during restart rename archive.log and upon startup it will be created automaticaly again.
    If no restart is planned, then delete it's content, but before that make a copy of alert.log. Just in order You will need to find something.Actually, I've found that it is quite safe to rename the alert.log right out from under a running database. It will simply start a new one with the next message to be written. However, I agree it is much nicer to do this housekeeping while the db is down, so that the first entry in an alert log is the startup.
    This renameing of the log against a running process does not work for the listener log. I've found that if I rename listener.log to listener.log.1, the listener will continue to log to the original file (now listener.log.1) until such time as it is restarted. THEN it will start a new listener.log.
    This was all on unix. Windows may behave differently.

  • Urgent Database Down - alert log file

    Hi Guru's
    this is the alert log file - today morning production database went down - any one can you please let me know the problem area!
    Thread 1 advanced to log sequence 17563
    Current log# 4 seq# 17563 mem# 0: E:\ORACLE\ORADATA\PNLDB\REDO04.LOG
    Thu Jan 26 08:12:56 2006
    KCF: write/open error block=0x7f0bc online=1
    file=12 F:\ORACLE\ORADATA\PNLDB\USERS02.DBF
    error=27069 txt: 'OSD-04026: Invalid parameter passed. (OS 520380)'
    Thu Jan 26 08:12:57 2006
    Errors in file f:\oracle\admin\pnldb\udump\pnldb_ora_1420.trc:
    ORA-01115: IO error reading block from file 12 (block # 517514)
    ORA-01110: data file 12: 'F:\ORACLE\ORADATA\PNLDB\USERS02.DBF'
    ORA-27091: skgfqio: unable to queue I/O
    ORA-27069: skgfdisp: attempt to do I/O beyond the range of the file
    OSD-04026: Invalid parameter passed. (OS 517513)
    ORA-01115: IO error reading block from file 12 (block # 132287)
    ORA-01110: data file 12: 'F:\ORACLE\ORADATA\PNLDB\USERS02.DBF'
    ORA-27091: skgfqio: unable to queue I/O
    ORA-27069: skgfdisp: attempt to do I/O beyond the range of the file
    OSD-04026: Invalid parameter passed. (OS 132286)
    Thu Jan 26 08:12:57 2006
    Errors in file f:\oracle\admin\pnldb\bdump\pnldb_dbw0_956.trc:
    ORA-01242: data file suffered media failure: database in NOARCHIVELOG mode
    ORA-01114: IO error writing block to file 12 (block # 520380)
    ORA-01110: data file 12: 'F:\ORACLE\ORADATA\PNLDB\USERS02.DBF'
    ORA-27069: skgfdisp: attempt to do I/O beyond the range of the file
    OSD-04026: Invalid parameter passed. (OS 520380)
    DBW0: terminating instance due to error 1242
    Thu Jan 26 08:12:59 2006
    Errors in file f:\oracle\admin\pnldb\udump\pnldb_ora_1420.trc:
    ORA-00603: ORACLE server session terminated by fatal error
    ORA-01115: IO error reading block from file 12 (block # 517514)
    ORA-01110: data file 12: 'F:\ORACLE\ORADATA\PNLDB\USERS02.DBF'
    ORA-27091: skgfqio: unable to queue I/O
    ORA-27069: skgfdisp: attempt to do I/O beyond the range of the file
    OSD-04026: Invalid parameter passed. (OS 517513)
    ORA-01115: IO error reading block from file 12 (block # 132287)
    ORA-01110: data file 12: 'F:\ORACLE\ORADATA\PNLDB\USERS02.DBF'
    ORA-27091: skgfqio: unable to queue I/O
    ORA-27069: skgfdisp: attempt to do I/O beyond the range of the file
    OSD-04026: Invalid parameter passed. (OS 132286)
    Thu Jan 26 08:12:59 2006
    Errors in file f:\oracle\admin\pnldb\bdump\pnldb_pmon_3804.trc:
    ORA-01242: data file suffered media failure: database in NOARCHIVELOG mode
    Instance terminated by DBW0, pid = 956
    Thanks
    Ravi

    You are having problems on disks E\ and F\ - possibly hardware problems.
    (Hopefully these disks are not network drives. If they are, your problem might be network contention. Network mounted drives are never a good idea for databases.)
    If this is truly urgent, then I encourage you to use your Oracle Support contract to open a service request for assistance.

  • Thread 1 cannot allocate new log -- from alert log

    in my alert log file, I have noticed occasionally that this message was showing out during the business hour (8 am ~ 8 pm). Not often , maybe 3 or 4 times, but there are more during the midnight when system run some administration process such as KGL object name :SELECT /*+rule*/ SYS_XMLGEN(VALUE(KU$), XMLFORMAT.createFormat2('VIEW_T', '7')), KU$.OBJ_NUM ,K
    U$.SCHEMA_OBJ.NAME ,KU$.SCHEMA_OBJ.NAME ,'VIEW' ,KU$.SCHEMA_OBJ.OWNER_NAME FROM SYS.KU$_VIEW_VIEW KU$ WHERE KU$
    .OBJ_NUM IN (SELECT * FROM TABLE(DBMS_METADATA.FETCH_OBJNUMS(200001)))
    . My question is : Do I have to add another redo log file to system? and how much is proper?
    I also noticed there where message of "Memory Notification: Library Cache Object loaded into SGA
    Heap size 2829K exceeds notification threshold (2048K)" during that hours. Appreciate any advice. my system is 10g r2 on AIX 5.3 L.

    The "Thread cannot allocate new log" occurs more often off the regular hours due to system run process. It seems more regular occurring at evry 20 minutes. By OEM, I noticed thate there some system processes run at 2 am, 8 am, 2 pm, 8pm. for examples
    1) EMD_MAINTENANCE.EXECUTE_EM_DBMS_JOB_PROCS();
    2) MGMT_ADMIN_DATA.EVALUATE_MGMT_METRICS;
    3) SELECT /*+ USE_HASH(f a) */ SUM(NVL(F.BYTES - NVL(A.BYTES, 0), 0)/1024/1024), SUM(NVL(F.BYTES, 0)/1024/1024) FROM (SELECT DISTINCT TABLESPACE_NAME FROM USER_TABLES) U, (SELECT TABLESPACE_NAME, SUM(BYTES) BYTES FROM DBA_DATA_FILES GROUP BY TABLESPACE_NAME) F, (SELECT TABLESPACE_NAME, SUM(BYTES) BYTES FROM DBA_FREE_SPACE GROUP BY TABLESPACE_NAME) A WHERE A.TABLESPACE_NAME = U.TABLESPACE_NAME AND F.T...
    I wonder why they are necessary? or I have to ad one more redo log.

  • "checkpoint not complete" in alert log file.

    Hi, all.
    I have got a message of "Checkpoint not complete" in alert log file.
    Thread 2 advanced to log sequence 531
    Current log# 7 seq# 531 mem# 0: \\.\REDO231
    Current log# 7 seq# 531 mem# 1: \\.\REDO232
    Thread 2 cannot allocate new log, sequence 532
    Checkpoint not complete
    Current log# 7 seq# 531 mem# 0: \\.\REDO231
    Current log# 7 seq# 531 mem# 1: \\.\REDO232
    I searched "Checkpoint not complete" issue in this forum.
    As solutions,
    1. add more redo log groups
    2. increase the size of redo log
    3. check I/O contention
    4. set LOG_CHECKPOINT_INTERVAL, LOG_CHECKPOINT_TIMEOUT or
    FAST_START_MTTR_TARGET
    I think No.4 is the possible first approach in our environment.
    I think No.1 and No2 are not the ploblems in our environment.
    I ask the above issue oracle support center, but
    I was told that "if you are not getting this message frequently, you do not need to worry about it" from an oracle engineer.
    Is this true?? If I am not getting this message frequently, there is no problem
    in terms of database integrity, consistency, and performance?
    I will be waiting for your advice and experience in real life.
    Thanks and Regards.

    Redo Log Tuning Advisory and Automatic Checkpoint Tuning are new features introduced in Oracle 10G, if you are on 10g you may benefit from these features.
    The size of the redo log files can influence performance, because the behavior of the database writer and archiver processes depend on the redo log sizes. Generally, larger redo log files provide better performance, however it must balanced out with the expected recovery time. Undersized log files increase checkpoint activity and increase CPU usage. As rule of thumb switching logs at most once every fifteen minutes.
    Checkpoint frequency is affected by several factors, including log file size and the setting of the FAST_START_MTTR_TARGET initialization parameter. If the FAST_START_MTTR_TARGET parameter is set to limit the instance recovery time, Oracle automatically tries to checkpoint as frequently as necessary. Under this condition, the size of the log files should be large enough to avoid additional checkpointing due to under sized log files.
    The redo logfile sizing advisory is specified by column optimal_logfile_size of v$instance_recovery. This feature require setting the parameter "fast_start_mttr_target" for the advisory to take effect and populate the column optimal_logfile_size.
    Also you can obtain redo sizing advice on the Redo Log Groups page of Oracle Enterprise Manager Database Control.
    To enable automatic checkpoint tuning, unset FAST_START_MTTR_TARGET or set it to a nonzero value(This is measured in seconds and by default, this feature is not enabled, because FAST_START_MTTR_TARGET has a default value of 0). If you set this parameter to zero this feature will be disabled. When you enable fast-start checkpointing, remove or disable(set to 0) the following initialization parameters:
    - LOG_CHECKPOINT_INTERVAL
    - LOG_CHECKPOINT_TIMEOUT
    - FAST_START_IO_TARGET
    Enabling fast-start checkpointing can be done statically using the initialization files or dynamically using -
    SQL> alter system set FAST_START_MTTR_TARGET=10;
    Best regards.

  • Message in alert log file during the startup of the DB

    Hi,
    i have two instances, i have the same warning in the alert log file during the startup of the DB:
    Oracle instance running on a system with low open file descriptor limit.
    Tune your system to increase this limit to avoid severe performance degradation.
    The db_files parameter was at 1024 for the two instances, i have set it to 300 for the first instance and 550 for second, the message for the first instance was droped(it's ok), but the message was there in the alert log for the second instance.
    The two instance are on the same machine.
    The number of the open file limit on the linux machine is equal to 1024.
    I have shutdown the first instance (with db_files=300) the problem persits for the second instance (with db_files=550).
    My question is how to determinate the right value of db_files parameter to avoid this message.
    Regards.
    Message was edited by:
    learn

    Increase the max number of open file on Linux or decrease db_files.
    f the OS limit is 1024 db_files would have to be >= 472 for the message to be printed. t is printed if an instance is configured to access more datafiles than the OS-specific limit on the number of files that can be open in a single process. In this case the server will re-cycle the file descriptors.
    Vadim Bobrov
    Oracle Database Tools
    http://www.fourthelephant.com

Maybe you are looking for