The analysis of a Apache log file

What will you recommend for the analysis of a Apache log file?

Your favorite text-editor?
Are you looking for some specific analysis? There surely are tools available for that.
Can you be more specific in your question, please.
cu
Andreas

Similar Messages

  • Apache Log File size

    My Apache log files are getting filled and causes Apache to crash.
    How can I increase the size of the Apache logs?
    Thanks.

    Update: apparently the version of Java matters.  I down-rev'ed to java 6 and the app works straightaway.
    So, this version of Java works:
    java version "1.6.0_29"
    Java(TM) SE Runtime Environment (build 1.6.0_29-b11)
    Java HotSpot(TM) Client VM (build 20.4-b02, mixed mode, sharing)
    This version of Java does *not* work:
    java version "1.7.0_01"
    Java(TM) SE Runtime Environment (build 1.7.0_01-b08)
      Java HotSpot(TM) Client VM (build 21.1-b02, mixed mode, sharing)

  • Print a custom Error message in the XML bursting program's log file...

    Hi,
    I having this requirement, where i need to print a custom error message in the xml bursting program's log file.
    Actually i am having a report where i create invoices and then those invoices are emailed to the respective customers, now say if a customer has three contacts and there is only two valid email id's so what happens is bursting will be successful for two contacts whereas the third contact dosen't get any emails. when this happens i need to log a message in the bursting programs log file stating a custom message.
    Two things i want to know..
    1- Whether is it possible to write into the xml bursting programs log file
    2- If yes, then how..
    note: it ll be greatly appreciated if the answer is elaborated.
    thanks,
    Ragul

    Hi,
    I having this requirement, where i need to print a custom error message in the xml bursting program's log file.
    Actually i am having a report where i create invoices and then those invoices are emailed to the respective customers, now say if a customer has three contacts and there is only two valid email id's so what happens is bursting will be successful for two contacts whereas the third contact dosen't get any emails. when this happens i need to log a message in the bursting programs log file stating a custom message.
    Two things i want to know..
    1- Whether is it possible to write into the xml bursting programs log file
    2- If yes, then how..
    note: it ll be greatly appreciated if the answer is elaborated.
    thanks,
    Ragul

  • What is the purpose of standby redo log files

    Hi,
    What is the purpose of the standby redo log files in DR?
    what if the standby redo log files are created? or else not created?
    Please explain
    Thanks

    3.1.3 Configure a Standby Redo LogA standby redo log is required for the maximum protection and maximum availability modes and the LGWR ASYNC transport mode is recommended for all databases. Data Guard can recover and apply more redo data from a standby redo log than from archived redo log files alone.
    You should plan the standby redo log configuration and create all required log groups and group members when you create the standby database. For increased availability, consider multiplexing the standby redo log files, similar to the way that online redo log files are multiplexed.>
    Reference http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/create_ps.htm#i1225703
    HTH
    Anand

  • Can anybody explain the difference REDOLG vs ARCHIVE log files

    can anybody explain the difference REDOLG vs ARCHIVE log files?
    and the relation between redolog files and redolog buffers?
    thanx in advance

    Changes made to the database are recorded in the Redo log buffer. This Buffer contains Redo records or entries, which are a description of changes made to the database.
    The Redo Logs are critical in order to recover the database or to reconstruct data files and undo segments after a system crash or hardware failure.
    Each Oracle instance has only one log writer process (LGWR). The log writer operates in the background and writes all records from the Redo log buffer to the Redo log files. When the database operates in Archivelog mode, the redo logs are are archived before overwritten.

  • What's The Terminal Command To Find Apache Log Files?

    I can't find it anywhere.
    Thank you.

    It looks like either:
    $ grep ErrorLog /private/etc/apache2/httpd.conf
    # ErrorLog: The location of the error log file.
    # If you do not specify an ErrorLog directive within a <VirtualHost>
    ErrorLog "/private/var/log/apache2/error_log"
    or
    /usr/sbin/apachectl -V | egrep 'HTTPD_ROOT|ERRORLOG'
    -D HTTPD_ROOT="/usr"
    -D DEFAULTERRORLOG="logs/errorlog"
    Though red_menace probably knows better that I.
    Message was edited by: Nils C. Anderson

  • Error occurs during the rendering of jsp component log file?

    Hello. I am having troubles on locating the log file for this particular error:
    Portal Runtime Error
    An exception occurred while processing a request for :
    iView : N/A
    Component Name : N/A
    Error occurs during the rendering of jsp component.
    Exception id: 11:59_31/07/07_0004_19351950
    See the details for the exception ID in the log file
    I searched all folders on my drive, using the exception id as keyword to search files that contains part of the id (0004_19351950), but to no avail. I wanted to see the log for me to identify the problem. Any ideas where I can look for it? or, better yet, what can be a solution to this runtime error?
    I'm using the 2004s developer environment version. Thanks in advance.

    hi Arnold,
    Go to Java Developer role -> expand Portal Tools folder -> click on Portal Logviewer. Put the exception id in the text-box corresponding to search -> click on search all files. Open the file that is displayed in the search result and make a search with the id again to refer to the exact line.
    N.B: if you don't have Java developer role, using user admin, assign that role to you first
    Cheers.......

  • What's the command to dump archived log file?

    Could someone share if you remember? I forgot the syntax. Thanks alot.

    Are you referring to extracting data from the archive log files? If so, you should refer to this Using LogMiner.

  • Why did the permissions change on my log files?

    My Messaging Server was running just fine, but then it crashed with an error
    message that it couldn't write the log files. What happened?
    <P>
    Under the INSTALLATION root, there is only one file that should be owned by
    root:
    <P>
    ./bin/msg/store/bin/ressh
    <P>
    ALL other files should be owned by the user that Messaging is to be run as.
    Changing the ownership of the log files will enable the server to start
    immediately, but the ownership of the files should be addressed.

    Check your settings in System Preferencees > Desktop & Screen Saver > Hot Corners

  • How to get the form run time debug log file

    Hi,
    I followed the instruction of Metalink Notes:
    R12: Forms Runtime Diagnostics (FRD), Tracing And Logging For Forms In Oracle Applications [ID 438652.1]
    Option 1: Obtain FRD Trace Using Profile Options
    Method 1: Using profile option 'ICX: Forms Launcher'
    I was not able to get the log file.
    Any suggestions

    This problem is specific to the eBusiness-suite. Try it in an appropiate forum, maybe here OA Framework

  • How to find the configuration status of REDO logs files?

    I am in the process of moving the redo log files.
    Before that I want to find out whether it is set up as duplex or any other configurations.
    How to find the REDO log configurations.
    Oracle 10g R2
    Thank you,
    Smith

    Example:
    Not duplexed redo log - if number of member is 1, then it is not duplexed:
    SQL> select group#, members from v$log;
        GROUP#    MEMBERS
             1          1
             2          1
             3          1
    SQL> select group#, member from v$logfile;
        GROUP# MEMBER
             1 /u01/app/oracle/oradata/db1/redo01.log
             2 /u01/app/oracle/oradata/db1/redo02.log
             3 /u01/app/oracle/oradata/db1/redo03.log
    SQL> alter database add logfile member '/u01/app/oracle/oradata/db1/redo01_2.log' to group 1;
    Database altered.
    SQL> alter database add logfile member '/u01/app/oracle/oradata/db1/redo02_2.log' to group 2;
    Database altered.
    SQL> alter database add logfile member '/u01/app/oracle/oradata/db1/redo03_2.log' to group 3;
    Database altered.Duplexed redolog
    SQL> select group#, members from v$log;
        GROUP#    MEMBERS
             1          2
             2          2
             3          2
    SQL> select group#, member from v$logfile;
        GROUP# MEMBER
             1 /u01/app/oracle/oradata/db1/redo01.log
             2 /u01/app/oracle/oradata/db1/redo02.log
             3 /u01/app/oracle/oradata/db1/redo03.log
             1 /u01/app/oracle/oradata/db1/redo01_2.log
             2 /u01/app/oracle/oradata/db1/redo02_2.log
             3 /u01/app/oracle/oradata/db1/redo03_2.log

  • Catch the error in to a log file

    Hi Team,
    Created the script for testing the server online or offline, Could you please tell me how to catch the error in to a file
    Try
    $Result = Test-Connection -count 1 -computer (Get-Content ServerName.txt)  -ErrorAction "Continue"
    Catch
       $Result = New-Object PSCustomObject -Property @{
                TimeStamp = Get-Date
                __Server = $env:COMPUTERNAME
                Address = $Computer
                Protocol = ""
                ResponseTime = "Failed"
    $Result | Select TimeStamp,__Server,Address,ProtocolAddress,ResponseTime | ft -AutoSize

    As mjolinor pointed, you should change the -ErrorAction to Stop to make the control flow into catch block when error occurs. Just change alone is not sufficient because, the control will go to catch block if any of the computer in servername.txt gives error,
    and it will not go back to next computer in your text file perform the ping.
    So try chaning the code as shown below.
    $Outarr = @()
    foreach($Comp in (Get-Content C:\temp\servers.txt)) {
    Try {
    $Result = Test-Connection -count 1 -computer $Comp -ErrorAction stop
    Catch {
    $Result = New-Object PSCustomObject -Property @{
    TimeStamp = Get-Date
    __Server = $env:COMPUTERNAME
    Address = $Comp
    Protocol = ""
    ResponseTime = "Failed"
    $Outarr +=$Result
    $Outarr| Select TimeStamp,__Server,Address,ProtocolAddress,ResponseTime | ft -AutoSize
    Hope this helps.
    Thanks,
    Sitaram Pamarthi
    Blog : http://techibee.com
    Follow on Twitter
    This posting is provided AS IS with no warranties or gurentees,and confers no rights
    You can simplify that to:
    $Outarr =
    foreach($Comp in (Get-Content C:\temp\servers.txt)) {
    Try {
    Test-Connection -count 1 -computer $Comp -ErrorAction stop
    Catch {
    [PSCustomObject]@{
    TimeStamp = Get-Date
    __Server = $env:COMPUTERNAME
    Address = $Comp
    Protocol = ""
    ResponseTime = "Failed"
    $Outarr | ft -AutoSize
    The [PSCustomObject] type accelerator will take the hash table as a constructor, and coerce it to an [ordered] hash table, so the properties will stay in the same order they were declared.  That eliminates the need to use Select-Object afterward to
    get them back into the right order.
    [string](0..33|%{[char][int](46+("686552495351636652556262185355647068516270555358646562655775 0645570").substring(($_*2),2))})-replace " "

  • Apache log analysis for portal page visits etc

    We are about to go into production with Oracle Portal (9iAS rel 9.0.2.0.1).
    I have been looking into how to gather statistics on web site usage and have concluded that I cannot get the information from within portal logs since many pages are served from the cache and there is no interaction with the database at all.
    I think I need to, therefore, analyse the apache log files. Can anyone recommend a good 3rd party product for log analysis? I am also concerned that the portal specific urls might severely limit the usefullness of the data I can extract.
    Has anyone used a log analysis tool in conjunction with Portal or knows of a product with a specific add-on for Oracle Portal (I have seen these for other portals and CMS products).
    Any pointers appreciated, I am completely new to this area.
    Peter.

    Peter,
    You might want to look at the "ClickStream Collector" that is included in a 9iAS installation.
    I'm not sure if it will do what you want with Portal, but it's starting point.
    Good luck!
    Bill G...

  • .default files and Apache logs in 10g AS(10.1.2.3)

    Hi Friends,
    I am using Oracle 10g AS (10.1.2.3) and i have some .default files in my Middle Tier and Infra home like
    httpd.conf.default , access.conf.default.....
    Is it safe to remove these files?
    Will these files cause any Vulnerability issue on my server?
    Also will Apache Log files cause Vulnerability issue on my server?
    Please clarify
    Regards,
    DB

    Default files are there as a backup for those files. If you happen to modify your files without saving a copy then this would take you back to how a typical httpd.conf (for instance, or plsql.conf or sso.properties, etc) would look like.
    I do not think they are any greater source of vulnerability than any other backup files you have on your server. I do not delete them either and wouldn't from any of the servers that I administer.
    thanks,
    AMN

  • How to Monitor the size of log file (Log4j) During its Generation.

    i have made a program using log4j technology. Now i want to know that is there any method that i can restrict the size of the log file. for example if the size of generating log file exceeds the restricted size then it should be placed in backup and a new file should be automatically generate and remaining contents should be written in that file.
    is there any method that can monitor the size of file during the generation of that (say) log file.
    Waiting for ur Urgent response

    I have wrote that code
    <?xml version="1.0" encoding="UTF-8" ?>
    <!DOCTYPE log4j:configuration SYSTEM "log4j.dtd">
    <log4j:configuration xmlns:log4j="http://jakarta.apache.org/log4j/">
    <appender name="appender" class="org.apache.log4j.FileAppender">
    <param name="File" value="c:\\abc.txt"/>
    <param name="MaxFileSize" value="100B"/>
    <param name="MaxBackupIndex" value="3"/>
    <param name="Append" value="false"/>
    <layout class="org.apache.log4j.SimpleLayout"></layout>
    </appender>
    <root>
    <priority value ="debug"/>
    <appender-ref ref="appender"/>
    </root>
    </log4j:configuration>
    When i run it it gave me error Message
    log4j:WARN No such property [maxFileSize] in org.apache.log4j.FileAppender.
    log4j:WARN No such property [maxBackupIndex] in org.apache.log4j.FileAppender.

Maybe you are looking for