Log file creation ( Trace)

Hi,
I am trying to do the global trace for the application and
set the following in the mm.cfg file
ErrorReportingEnable = 1
TraceOutputFileEnable =1
MaxWarnings = 100
TraceOutputFileName = "C:\Documents and
Settings\user\myProject.log"
The file is not created when I am using a trace("Test") in
the button click event on the page.
Is there anything important that I am missing?
Thanks in advance.

Hi,
Post your question in the appropriate General Database Discussions.
Thanks,
Hussein

Similar Messages

  • Diff between Log files and Trace files

    hi
    What is the exact differences between Log Files and Trace Files.
    And the relevant informative URL/reference materils would be highly appriciated.
    Thanks
    Sekhar

    Hi,
    Go through these,
    http://help.sap.com/saphelp_nw04/helpdata/en/d1/7b1e40777cdd5fe10000000a155106/frameset.htm
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/0bf8c890-0201-0010-8b8c-fc93fa1f4795
    Hope it clarifies
    Regards
    Srinivasan T

  • Question regarding alert log file and trace files

    What should be the alert log file size ? And when should it be deleted? And for how many days user trace files should be kept?
    Also will anyone please tell me the importance of these files.
    Thanks

    This may help: http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14231/manproc.htm#sthref729
    There are a few discussions on it here:
    Re: Alert Log File
    alert log file contents viewing
    Re: how to read alert log file? is there any tool available?

  • Log File Creation Confusion

    SQL*Plus: Release 10.2.0.3.0 - Production on Mon Mar 11 11:42:45 2013
    Copyright (c) 1982, 2006, Oracle.  All Rights Reserved.There are some initialization parameters that decide the location of the online redo log files in general.These initialization parameters are
    - DB_CREATE_ONLINE_LOG_DEST_n
    - DB_RECOVERY_FILE_DEST
    - DB_CREATE_FILE_DEST
    I could not understand the level of precedence of these parameters if you set each of them for creating online logfile, if i set all these parameter then creating online log file always goes to the path which define in parameter DB_CREATE_ONLINE_LOG_DEST_n and ignores the others parameter (DB_RECOVERY_FILE_DEST,DB_CREATE_FILE_DEST).
    If i just set the last two parameter (DB_RECOVERY_FILE_DEST,DB_CREATE_FILE_DEST) and do not set the DB_CREATE_ONLINE_LOG_DEST_n the logfile created in both location DB_RECOVERY_FILE_DEST,DB_CREATE_FILE_DEST) with mirrored mechanisim.
    SQL> select name,value
      2    from v$parameter
      3   where upper(name) in ('DB_CREATE_ONLINE_LOG_DEST_1','DB_RECOVERY_FILE_DEST','DB_CREATE_FILE_DEST')
      4  /
    NAME                                                                             VALUE
    db_create_file_dest                                                              D:\ORACLE\PRODUCT\10.2.0\DB_1\dbfile
    db_create_online_log_dest_1
    db_recovery_file_dest                                                            D:\oracle\product\10.2.0\db_1\flash_recovery_area
    SQL> select * from v$logfile
      2  /
        GROUP# STATUS  TYPE    MEMBER                                                                              
             3         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO03.LOG                                    
             2         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO02.LOG                                    
             1 STALE   ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO01.LOG                                    
    SQL> alter database add logfile
      2  /
    Database altered.
    SQL> select * from v$logfile
      2  /
        GROUP# STATUS  TYPE    MEMBER                                                                                      
             3         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO03.LOG                                            
             2         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO02.LOG                                            
             1 STALE   ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO01.LOG                                            
             4         ONLINE  D:\ORACLE\PRODUCT\10.2.0\DB_1\DBFILE\ORCL\ONLINELOG\O1_MF_4_8MTHLWTJ_.LOG                   
             4         ONLINE  D:\ORACLE\PRODUCT\10.2.0\DB_1\FLASH_RECOVERY_AREA\ORCL\ONLINELOG\O1_MF_4_8MTHLZB8_.LOGAs you can see above result , creating a logfile adhere defining parameters DB_RECOVERY_FILE_DEST,DB_CREATE_FILE_DEST) , when i define the parameter DB_CREATE_ONLINE_LOG_DEST_1 , logfile creation will goes to only defining within parameter DB_CREATE_ONLINE_LOG_DEST_1 no matter what you define for DB_RECOVERY_FILE_DEST,DB_CREATE_FILE_DEST).Here you go.
    SQL> alter database drop logfile group 4
      2  /
    Database altered.
    SQL> select * from v$logfile
      2  /
        GROUP# STATUS  TYPE    MEMBER                                                                      
             3         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO03.LOG                            
             2         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO02.LOG                            
             1 STALE   ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO01.LOG                            
    SQL> alter system set db_create_online_log_dest_1='D:\oracle' scope=both
      2  /
    System altered.
    SQL> select name,value
      2    from v$parameter
      3   where upper(name) in ('DB_CREATE_ONLINE_LOG_DEST_1','DB_RECOVERY_FILE_DEST','DB_CREATE_FILE_DEST')
      4  /
    NAME                                                                             VALUE
    db_create_file_dest                                                              D:\ORACLE\PRODUCT\10.2.0\DB_1\dbfile
    db_create_online_log_dest_1                                                      D:\oracle
    db_recovery_file_dest                                                            D:\oracle\product\10.2.0\db_1\flash_recovery_area
    SQL> alter database add logfile
      2  /
    Database altered.
    SQL> select * from v$logfile
      2  /
        GROUP# STATUS  TYPE    MEMBER                                                                              
             3         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO03.LOG                                    
             2         ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO02.LOG                                    
             1 STALE   ONLINE  D:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\REDO01.LOG                                    
             4         ONLINE  D:\ORACLE\ORCL\ONLINELOG\O1_MF_4_8MTJ10B8_.LOG                                       My confusion is here why the mechanisim of (DB_RECOVERY_FILE_DEST,DB_CREATE_FILE_DEST) is same while the same with both of them becomes differ when you define
    'DB_CREATE_ONLINE_LOG_DEST_n'?

    DB_CREATE_FILE_DEST is used if DB_CREATE_ONLINE_LOG_DEST_n is not defined.
    DB_RECOVERY_FILE_DEST is used for multiplexed log files.
    Thus, if Oracle uses DB_CREATE_FILE_DEST (because DB_CREATE_ONLINE_LOG_DEST_n is not defined), it multiplexes the log file to DB_RECOVERY_FILE_DEST if DB_RECOVERY_FILE_DEST is also defined.
    If, however, DB_CREATE_ONLINE_LOG_DEST_1 is used, Oracle expects you to define DB_CREATE_ONLINE_LOG_DEST_2 as well for multiplexing the log file; else it assumes that you do not want the log file multiplexed. The fact that the parameter ends with an n means that Oracle uses the n=2 f or the multiplexed location if defined.
    Hemant K Chitale

  • Sqlnet.log file creation permissions

    One of my sites has a sqlnet.log file that has been symlinked to /dev/null and we are trying to determine if this has caused the issue with /dev/null's permissions being reset to 660 instead of 666.
    It looks like the other sqlnet.log files that are not symlinked to /dev/null are 640 so we are not sure this is the issue.
    Does anyone know if oracle changes permissions on the sqlnet.log file at creation or access time? And how it might change those permissions?
    This is on Oracle 10g running on RHEL4
    edit: I typed sqlnet.ora instead of sqlnet.log ... oops.
    Edited by: user12198769 on Nov 10, 2009 7:35 AM

    looks to be just the default in that file:
    NAMES.DIRECTORY_PATH= (TNSNAMES, ONAMES, HOSTNAME)
    There are several sqlnet.log files in other places on the server yet there is only one oracle home on here so is there another location that has sqlnet.ora info?

  • Disable Log file creation for a Berkeley DB database

    Hi,
    I'm using Berkeley DB 6 with Oracle Mobile Server 11.3.  When I sync a lot of data, a lot of logfile are created and I think that this is really slowing down my sync process.  Since I never need to recover those client database, I would like to know if it is possible to disable log creation on a Berkeley Database?
    Thank you

    The version of BDB that is used for DMS is TDS (Transaction Data Store).   In that environment, logging is needed to ensure recoverability.   There isnt a way to disable logging.     If you never need to do recovery, then you can use the BDB utilities and occasionally do checkpoints which will flush the cache, or you can shut down the client application.  After this is done then you can remove the log files, since you are claiming that you will not need them for recovery.   
    thanks
    mike

  • Exchange Server 2010 - Message Tracking Logs - Log file creation

    Hi,
    I would like to find out on the behavior of the exchange server in the way that it logs the message tracking.
    Currently the parameter used is 
    MessageTrackingLogMaxDirectorySize - 10GBMessageTrackingLogMaxAge - 30daysI would like to check when the Max Directory Size has exceeded the value indicated, does Exchange server immediately deletes the oldest log file to make space for the new logs?And in the event that the oldest file is being open or locked, will exchange server delete the next oldest file? or it will reattempt to delete the "locked" file for a period of time?Lastly, when these "oldest" files is not able to be deleted, will exchange server stops logging new tracking events?Thanks!

    Hi Zack,
    Thank you for your question.
    If you have configured the parameter of “MessageTrackingLogMaxDirectorySize” and “MessageTrackingLogMaxAge”, we think you have enable circular logging, it will delete the oldest message tracking log files for new log file when the either of the following
    conditions is true:
    The message tracking log directory reaches its specified maximum size.
    A message tracking log file reaches its specified maximum age.
    In addition, it didn’t exceeded the value indicated.
    If there are any questions regarding this issue, please be free to let me know. 
    Best Regard,
    Jim
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact [email protected]
    Jim Xu
    TechNet Community Support

  • Log file creation using km api

    Hi,
    how to create log file using km api . please provide me if any sample code available.
    Thanks and Regards,
    Nari.

    Thanks for your quick reply but one more requirement is... here i can able create text file in km and adding content to created text file on the same line but i want to update new content in next line(newline).Please see below code and correct it.
         Date dt = new Date(Calendar.getInstance().getTimeInMillis());
                                  com.sapportals.portal.security.usermanagement.IUser iuser = WPUMFactory.getServiceUserFactory().getServiceUser("cmadmin_service");
                                  IResourceContext irCtx = new ResourceContext(iuser);
                                  RID docsResource = RID.getRID(filepath);
                                  IContent initCont = new Content(new ByteArrayInputStream("".getBytes()),"text/plain",-1,null);
                                  if(ResourceFactory.getInstance().getResource(RID.getRID(filepath+"/"+filename), irCtx) == null)
                                       ICollection docsColl = (ICollection)com.sapportals.wcm.repository.ResourceFactory.getInstance().getResource(docsResource,irCtx);
                                       docsColl.createResource(filename,null,initCont);
                              String InputData = Exception;
                              RID sugg_html = RID.getRID(filepath+"/"+filename);
                              IResource resource = com.sapportals.wcm.repository.ResourceFactory.getInstance().getResource(sugg_html,irCtx);
                              String existingComments;
                              IContent cont = resource.getContent();
                              BufferedReader buf_in = new BufferedReader(new InputStreamReader(cont.getInputStream()));
                              existingComments = buf_in.readLine();
                              existingComments = existingComments+"   "+"\n"+dt+InputData;
                              ByteArrayInputStream inputStream = new ByteArrayInputStream(existingComments.getBytes());
                              cont = new Content(inputStream,"text/plain",-1,null);
                              resource.updateContent(cont);
                              cont.close();

  • Log file creation -- BDC

    Hi friends.This is Sudhir . I have a scenario of loading mass vendor data. If in case any error occurs while loading the error record should be created as log file. Do we have any function module to create the log file?
    Hope your answers will be helpful to proceed further.
    With regards ,
    Sudhir S

    Are you looking a way for generating logs as the ones you can see in SLG0?
    You can also store them into a spool.
    To store them into a file, you can simply read the spool output (or you can maybe use a SUBMIT ... EXPORTING LIST TO MEMORY to avoid the spool).
    In that case, please refer to [sap library|http://help.sap.com/saphelp_nw2004s/helpdata/en/d3/1fa03940fab918e10000000a114084/frameset.htm]
    and SBAL* demo programs (use of BAL_* function modules)
    Edited by: Sandra Rossi on Jul 20, 2010 10:39 PM

  • Log file creation

    hi ,
            can anybody tell me how to create a log file to know that which master data has been replicated?

    hi ,
            can anybody tell me how to create a log file to know that which master data has been replicated?

  • Database not starting,REDO log file error?

    My DB was working fine uptill last evening, when i restarted it for some DB task it showed me an error ..
    ORA-00333: redo log read error block 57346 count 8192
    I studied the alert log file & related trace file .... but could not conclude what to do ?? Can any one suggest me what could be the solution...
    ORACLE 10g R2
    Windows 2003 Server
    Database in NOARCHIVELOG mode.
    I don't have the recent backup.The one i have is 20 days old.
    Alert Log file contents : -
    Wed Jun 02 20:22:19 2010
    Aborting crash recovery due to error 333
    Wed Jun 02 20:22:19 2010
    Errors in file e:\oracle\product\10.2.0\admin\ora10g\udump\ora10g_ora_140.trc:
    ORA-00333: redo log read error block 57346 count 8192
    ORA-333 signalled during: ALTER DATABASE OPEN...
    Wed Jun 02 20:37:15 2010
    db_recovery_file_dest_size of 2048 MB is 0.00% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    Wed Jun 02 20:43:53 2010
    Shutting down instance: further logons disabled
    Wed Jun 02 20:43:53 2010
    Stopping background process CJQ0
    Wed Jun 02 20:43:54 2010
    Stopping background process MMNL
    Wed Jun 02 20:43:55 2010
    Stopping background process MMON
    Wed Jun 02 20:43:56 2010
    Shutting down instance (immediate)
    License high water mark = 3
    Wed Jun 02 20:43:56 2010
    Stopping Job queue slave processes
    Wed Jun 02 20:43:56 2010
    Job queue slave processes stopped
    Waiting for dispatcher 'D000' to shutdown
    All dispatchers and shared servers shutdown
    Wed Jun 02 20:43:58 2010
    ALTER DATABASE CLOSE NORMAL
    ORA-1109 signalled during: ALTER DATABASE CLOSE NORMAL...
    Wed Jun 02 20:43:58 2010
    ALTER DATABASE DISMOUNT
    Completed: ALTER DATABASE DISMOUNT
    ARCH: Archival disabled due to shutdown: 1089
    Shutting down archive processes
    Archiving is disabled
    Archive process shutdown avoided: 0 active
    ARCH: Archival disabled due to shutdown: 1089
    Shutting down archive processes
    Archiving is disabled
    Archive process shutdown avoided: 0 active
    Trace file to read ora10g_ora_140.trc:-
    Dump file e:\oracle\product\10.2.0\admin\ora10g\udump\ora10g_ora_140.trc
    Wed Jun 02 20:22:18 2010
    ORACLE V10.2.0.1.0 - Production vsnsta=0
    vsnsql=14 vsnxtr=3
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Windows Server 2003 Version V5.2 Service Pack 1
    CPU                 : 2 - type 586, 1 Physical Cores
    Process Affinity    : 0x00000000
    Memory (Avail/Total): Ph:3527M/4094M, Ph+PgF:4386M/5973M, VA:800M/2047M
    Instance name: ora10g
    Redo thread mounted by this instance: 1
    Oracle process number: 15
    Windows thread id: 140, image: ORACLE.EXE (SHAD)
    *** SERVICE NAME:() 2010-06-02 20:22:18.406
    *** SESSION ID:(159.3) 2010-06-02 20:22:18.406
    Successfully allocated 2 recovery slaves
    Using 545 overflow buffers per recovery slave
    Thread 1 checkpoint: logseq 14261, block 2, scn 86974094
      cache-low rba: logseq 14260, block 4023
        on-disk rba: logseq 14262, block 3758, scn 86978396
      start recovery at logseq 14261, block 2, scn 86974094
    ORA-00333: redo log read error block 57346 count 8192
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\PRODUCT\10.2.0\ORADATA\ORA10G\REDO02.LOG'
    ORA-27070: async read/write failed
    OSD-04016: Error queuing an asynchronous I/O request.
    O/S-Error: (OS 2) The system cannot find the file specified.
    ORA-00333: redo log read error block 57346 count 8192
    ORA-00312: online log 2 thread 1: 'E:\ORACLE\PRODUCT\10.2.0\ORADATA\ORA10G\REDO02.LOG'
    ORA-27091: unable to queue I/O
    ORA-27070: async read/write failed
    OSD-04006: ReadFile() failure, unable to read from file
    O/S-Error: (OS 2) The system cannot find the file specified.
    ORA-00333: redo log read error block 57346 count 8192
    *** 2010-06-02 20:43:58.812
    *** 2010-06-02 20:43:58.812 60680 kcrr.c
    ARCH: Archival disabled due to shutdown: 1089
    *** 2010-06-02 20:43:59.812 60680 kcrr.c
    ARCH: Archival disabled due to shutdown: 1089

    Hi,Aman this(following) may is not good way for opening database,but OP can try and open own database:
    @ DBA4 you will see from following example how you can open your database.In your online log files was corrupted,therefore you have to do following step by step then you database will open.
    SQL> alter database clear unarchived logfile group 2;
    alter database clear unarchived logfile group 2
    ERROR at line 1:
    ORA-01624: log 2 needed for crash recovery of instance p (thread 1)
    ORA-00312: online log 2 thread 1:
    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\REDO02.LOG'
    SQL> alter database backup controlfile to trace as 'D:\oracle\product\10.2.0\ora
    data\P\c.txt';
    Database altered.
    SQL> shutdown abort;
    ORACLE instance shut down.
    SQL> startup nomount;
    ORACLE instance started.
    Total System Global Area  125829120 bytes
    Fixed Size                  1288100 bytes
    Variable Size              75499612 bytes
    Database Buffers           41943040 bytes
    Redo Buffers                7098368 bytes
    SQL> CREATE CONTROLFILE REUSE DATABASE "P" RESETLOGS   ARCHIVELOG
      2      MAXLOGFILES 16
      3      MAXLOGMEMBERS 3
      4      MAXDATAFILES 100
      5      MAXINSTANCES 8
      6      MAXLOGHISTORY 292
      7  LOGFILE
      8    GROUP 1 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\REDO01.LOG'  SIZE 50M,
      9    GROUP 2 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\REDO02.LOG'  SIZE 50M,
    10    GROUP 3 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\REDO03.LOG'  SIZE 50M
    11  DATAFILE
    12    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\SYSTEM01.DBF',
    13    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\UNDOTBS01.DBF',
    14    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\SYSAUX01.DBF',
    15    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\USERS01.DBF',
    16    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\TS_TEST.ORA',
    17    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\TEST.ORA',
    18    'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\TEST1.ORA'
    19  CHARACTER SET WE8MSWIN1252;
    Control file created.
    SQL> alter database open resetlogs;
    alter database open resetlogs
    ERROR at line 1:
    ORA-01194: file 1 needs more recovery to be consistent
    ORA-01110: data file 1: 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\SYSTEM01.DBF'
    SQL> recover database using backup controlfile until cancel;
    ORA-00279: change 2032832 generated at 06/03/2010 12:42:29 needed for thread 1
    ORA-00289: suggestion :
    D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\P\ARCHIVELOG\2010_06_03\O1_MF_1_87_
    %U_.ARC
    ORA-00280: change 2032832 for thread 1 is in sequence #87
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    cancel
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01194: file 1 needs more recovery to be consistent
    ORA-01110: data file 1: 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\SYSTEM01.DBF'
    ORA-01112: media recovery not started
    then add *_allow_resetlogs_corruption=true* in pfile then startup database with pfile as
    SQL> startup mount pfile=D:\oracle\product\10.2.0\admin\P\pfile\init.ora.pr
    ORACLE instance started.
    Total System Global Area  125829120 bytes
    Fixed Size                  1288100 bytes
    Variable Size              62916700 bytes
    Database Buffers           54525952 bytes
    Redo Buffers                7098368 bytes
    Database mounted.
    SQL> recover;
    ORA-00283: recovery session canceled due to errors
    ORA-01610: recovery using the BACKUP CONTROLFILE option must be done
    SQL> recover database using backup controlfile until cancel;
    ORA-00279: change 2032832 generated at 06/03/2010 12:42:29 needed for thread 1
    ORA-00289: suggestion :
    D:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\P\ARCHIVELOG\2010_06_03\O1_MF_1_87_
    %U_.ARC
    ORA-00280: change 2032832 for thread 1 is in sequence #87
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    cancel
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01194: file 1 needs more recovery to be consistent
    ORA-01110: data file 1: 'D:\ORACLE\PRODUCT\10.2.0\ORADATA\P\SYSTEM01.DBF'
    ORA-01112: media recovery not started
    SQL> alter database open;
    alter database open
    ERROR at line 1:
    ORA-01589: must use RESETLOGS or NORESETLOGS option for database open
    SQL> alter database open resetlogs;
    Database altered.
    SQL>finally remove allowresetlogs_corruption=true parameter from pfile.

  • Re: identifying server log file when runningdistributed

    Hi John, I can give you some TOOL code which will get the process
    id, but I do have to stray outside of Framework :-). The following
    code uses classes from the SystemMonitor project which was
    introduced in release 2 of Forte (this code won't work on R1):
    partAgent : SystemAgent;
    pidInst : ConfigValueInst;
    pid : TextData;
    partAgent = SystemAgent(task.Part.Agent);
    pidInst = ConfigValueInst(partAgent.FindInstrument('ProcessId');
    pid = TextData(pidInst.GetData);
    The result is that the variable pid contains the process id in string form.
    This could be converted to numeric form if needed.
    If what you're really after is the partition's log file name, then the following
    code will do the trick (it takes into account the differences in how the log
    files are named for interpreted vs. compiled partitions):
    partAgent : SystemAgent;
    logFileInst : ConfigValueInst;
    logFileName : TextData;
    -- Get our agent and try to get the log file inst
    partAgent = SystemAgent(task.Part.Agent);
    logFileInst = ConfigValueInst(partAgent.FindInstrument('LogFile');
    -- Interpreted partition don't have their own log file, so check
    if (logFileInst = NIL) then
    pidInst : ConfigValueInst;
    pid : TextData;
    -- We must be an interpreted partition get our pid
    pidInst = ConfigValueInst(partAgent.FindInstrument('ProcessId');
    pid = TextData(pidInst.GetData);
    -- Build log file name
    logFileName = 'forte_ex_';
    logFileName.Concat(pid);
    else
    -- Get the name of the log file from the instrument
    logFileName = TextData(logFileInst.GetData);
    end;
    The available agents and their instruments and commands are documented
    in the manual "SystemMonitor Project". I'm at home now, so I don't have
    the page numbers. Some additional agents (which were added after this
    manual went to press) can be found in Tech Note #10475. Also, econsole
    and escript can be handy since any instrument you can see in these tools
    can be accessed from TOOL code. Hope this is of some use.
    Sean
    At 05:24 PM 7/30/96 -0700, John L. Jamison wrote:
    >
    I'd like to solicit some ideas from you folks. As many of you are probably
    aware, when running in distributed mode, log output for server partitions is
    written out to log files on the server partition. However it is sometimes a
    trick trying to identify the process which is running your individual
    partitions, and
    thus knowing which log file to read.
    At one client, we added a 3gl call-out to obtain the process id and return it
    to the client. However this is not a good option at a new client which uses
    Sequent (3gl wrappering difficult in statically linked environments such as
    sequent). I am also aware that Econsole allows you to browse active
    partitions and display log files, but you still have to know which active
    partitions to watch.
    I have not yet seen a way to programmatically obtain the process ID for a
    partition within TOOL and using FrameWork classes.
    What kinds of strategies are folks employing out there?
    Thanks in advance,
    -John
    John Jamison
    Sage Solutions, Inc.
    353 Sacramento Street, Suite 1360
    San Francisco, CA 94111
    415 392 7243 x 508
    [email protected]

    Hi John,
    I think that Sean Fits answered your question about TOOL code to get the PID
    number. I just want to complement on the loging strategy.
    There is one log file for every active partition of an application. I think
    it is useful in some cases that a distributed application gets a centralized
    log file to trace the exact sequential flow of processing among all the
    partitions. This is useful during the initial debuging and tuning. In fact
    something similar to the UNIX syslog file.
    For doing so it is easy to implement a custom central log Mgr in one
    partition and to have all partitions use it when needed (it doesn't prevent
    to continue using the standard LogMgr in addition). This central LogMgr
    automatically adds the date&time plus the node name, partition name, ... to
    the log messages it receives.
    The flags which apply are those of the partition where the central Log Mgr is.
    Because of potential concurency requests from the several partitions
    accessing the central Log Mgr, it is not possible to support the "Put" and
    PutHex" methods. Only complete lines can be logged (Putline and PutHexLine).
    Attached is the TOOL code of my TraceService plan that implements it.
    Remark : the "Phr" in the names relate to the name of the application we
    have here under development.
    To use the central Log Mgr, a partition must create an object of class
    PartitionLog, and then log messages must be sent to it the way you send them
    to the standard LogMgr; it will manage to send them to the central Log Mgr.
    At 17:24 30/07/96 -0700, John Jamison wrote:
    >
    I'd like to solicit some ideas from you folks. As many of you are probably
    aware, when running in distributed mode, log output for server partitions is
    written out to log files on the server partition. However it is sometimes a
    trick trying to identify the process which is running your individual
    partitions, and
    thus knowing which log file to read.
    At one client, we added a 3gl call-out to obtain the process id and return it
    to the client. However this is not a good option at a new client which uses
    Sequent (3gl wrappering difficult in statically linked environments such as
    sequent). I am also aware that Econsole allows you to browse active
    partitions and display log files, but you still have to know which active
    partitions to watch.
    I have not yet seen a way to programmatically obtain the process ID for a
    partition within TOOL and using FrameWork classes.
    What kinds of strategies are folks employing out there?

  • Who writes alert log file?

    Hi guys ,
    I have been searching for answer to this question,.............
    Which process(backgroud process) writes to alert log file?

    Trace files for user processes are normally in the user_dump. They are created only when tracing is requested OR when Oracle encounters an error.
    The alert.log is a file that is used to continously display the status of the database as it changes with important events (e.g. archival of log, alter system commands, ora-1555 errors, indexes unusable, datafile space allocation etc). Most of these are issues which affect the entire instance/database. However, where a sever error is encountered in a user process, Oracle writes a trace file in user_dump_dest for that error and a message indicating the error and the name of the trace file to the alert.log.
    Similarly, the background processes may also be writing to their own trace files to indcate status /tracing.
    The level of detail being logged can vary by version and by setting specific database events (if specified by Oracle Support) in the instance.
    See http://download.oracle.com/docs/cd/B19306_01/server.102/b14231/manproc.htm#sthref729
    Hemant K Chitale

  • Jar file: log file is not created!

    Hi, I create a log file with these instructions:
    //logger
              FileHandler fh = new FileHandler(LOG_FILE,true); //append mode
              fh.setFormatter(new SimpleFormatter());
              logger = Logger.getLogger(this.getClass().getName());
              logger.addHandler(fh);It works well as long I execute it normally, but when I create a .jar file and put this logger class inside logger file is not created..how come?

    Hi, I create a log file with these instructions:The code extract creates a Logger object, sure, but does that mean it creates a log file? The log file creation could be deferred until the first usage of the logger...
    It works well as long I execute it normally,Let's clarify: do you mean, when executing from your IDE, or when executing via a command line?
    but when I create a .jar file and put this logger class inside logger file is not created..how come?Let's clarify: do you mean, when you execute the same code packaged in a jar file?
    The most likely issue is configuration: apparently the various ways in which you run your program use different logging configurations.
    J.

  • How can I extract the data from a Real-time Execution Trace ".log" file?

    I would like to get the data for the traces from the Real-time Execution Trace toolkit ".log" file to read in Excel and generate a report.

    Hi Chuck,
    Have you tried reading it into a text or binary file first and then generating a report using the Report Generation VIs?
    Ipshita C.
    National Instruments
    Applications Engineer

Maybe you are looking for