Console generating huge logs

Console is generating multiple logs ranging from 1 to 12 gigabytes on my childrens' user log-ins. This is leading to "startup disk full" errors.
1) Why?
2) Can I stop it?
3) Does this indicate another problem?

Hi Troy, runaway logs are a fairly frequent cause of disappearing disk space but to stop the process you need to identify the problem...
1) Why? Because there is a problem somewhere that is endlessly writing to the log.
2) Can I stop it? Yes, by identifying the problem and sorting it out
3) Does this indicate another problem? Yes, see above
I'm sorry is the above seems facetious but without knowing what the runaway process is there is not much anyone can advise.
I am no expert at reading error logs but you need to read through the most recent entry and see if you can get some kind of a clue from it. There should be a recurring theme that you can identify. You could post the most recent entry here if it isn't too long but please don't post 1-12 GBs
You could try the following:
Starting up in Safe Mode
Safe Boot takes longer than normal startup
What is Safe Boot, Safe Mode.
Hopefully the errant process won't load in Safe Boot so it might be easier to identify, plus the Disk Repair utility of Safe Boot may fix the problem. Ultimately though it would be better to know what the problem is beforehand in case it recurs.
Good luck,
Adrian

Similar Messages

  • Oracle 10g Enterprise Manager console giving Alert log

    Dear all
    I have One HP server running Oracle 10g on Redhat Enterprise Linux 3.0 and two other systems with Oracle 9i and 8i. I have written sql query in my Oracle10g system which updates the database in 9i.
    The sql statement is
    CREATE MATERIALIZED VIEW ADASNAP REFRESH FAST START WITH SYSDATE NEXT SYSDATE+1 AS
    SELECT * FROM ADA@WCMMISLINK;
    CREATE MATERIALIZED VIEW BGTABSNAP REFRESH FAST START WITH SYSDATE NEXT SYSDATE+1 AS
    SELECT * FROM BGTAB@WCMMISLIN;
    The above query is sheduled to run every day .
    and i get the following Alert log in 10g enterprise manager console
    Generic Alert log ORA-12012: error on auto execute of job 54
    ORA-04052: error occured when looking up remote object WCMM.SYS@WCMMSERVERLINK
    ORA-00604: error occured at recursive SQL level 3
    ORA-12514: tns no listener
    ORA-06512: AT "SYS.DBMS_SNAPSHOT" line 1883
    can you guid me why the above error are occuring
    the ORA-12514: tns no listener
    the above error why it is occuring , then listener is running on oracle 10g server and other 8i and 9i server also and i am able to connect from oracle10g to other oracle 8i and 9i from sql plus and viceversa..
    what could be the problem with tns listener
    Regards
    Niranjan

    Hi Ugonic
    Sorry both are same, it's spelling mistake it's " WCMMISLINK". Acutally, my database is getting updated but this alert log is getting generated in enterprise manager console of 10g.
    The sql statement is
    CREATE MATERIALIZED VIEW ADASNAP REFRESH FAST START WITH SYSDATE NEXT SYSDATE+1 AS
    SELECT * FROM ADA@WCMMISLINK;
    CREATE MATERIALIZED VIEW BGTABSNAP REFRESH FAST START WITH SYSDATE NEXT SYSDATE+1 AS
    SELECT * FROM BGTAB@WCMMISLINK;
    The above query is sheduled to run every day .
    and i get the following Alert log in 10g enterprise manager console
    Generic Alert log ORA-12012: error on auto execute of job 54
    ORA-04052: error occured when looking up remote object WCMM.SYS@WCMMSERVERLINK
    ORA-00604: error occured at recursive SQL level 3
    ORA-12514: tns no listener
    ORA-06512: AT "SYS.DBMS_SNAPSHOT" line 1883
    can you guid me why the above error are occuring
    Regards
    Niranjan

  • Huge log.xml are generaing in 11.1.0.7 database.

    Hi
    we have an issue where listner log file is generating hugely and also it is creatin very huge listner logs. we tried bounce the listner and database but still it geerates huge log files

    how big is the listener log file.... ?
    to save on space, you may backup and truncate the existing listener log file with the below steps.
    lsnrctl set log_status off (turn listener logging off)
    gzip <listener.log>listener.log.07052011.gz (backup listener file by zipping)
    next step to truncate by doing this $>listener.log (do this to truncate the existing listener log)
    lsnrctl set log_status on (turn listener logging on)
    Again, I would like to know the listener.log file size..
    Edited by: s_n_i_dba on Jul 5, 2011 11:37 AM

  • How to generate rule log on client side

    I am looking for a solution where I could generate a rule log and make it available on client side. I know generating a log during consolidation and save it on server. However, I want to make it available on client side.
    I have created four scenarios in HFM application, namely -
    1. ACTUAL_REPORTING - To capture TB data and use it for reporting
    2. ACTUAL_ICP - To capture ICP data and run custom business rule as per requirement
    3. ACTUAL_ADDINPUT - To capture additional memorandum inputs
    During consolidation,data is copied from ACTUAL_ICP and ACTUAL_ADDINPUT to ACTUAL_REPORTING. However, problem is that user should consolidate ACTUAL_ICP AND ACTUAL_ADDINPUT scenarios first. If they miss executing the consolidation in these scenarios, then the data copied would be wrong. Just as to ensure this, there's a provision is HFM to check calculation status in rule file and looking at that, calcuation could be stopped and a message could be sent to the user. But delimma is, how to report the same on the user side.
    Possible solution that I could have thought are -
    1. To generate a rule log and post the same to the client, but how to do this? As path of the rule log could be of the server only (that is C:\Hyperion\Logs\Rule)
    2. To generate a rule log and pop a web page terminating consolidation, but how to do this?
    3. To raise an error from the rule file and direct it to the system messages, but how to to this?
    Please advice and any further clarification is required please contact me.
    Thanks & Regards,
    Mustafa

    You can define the output folder as a share, which has already been mentioned. I strongly advise against implementing any solution that in production will regularly generate an external file.
    1) The output file is generated by the DCOM user. If you want this user to generate a file in a specific location, it must be writeable by the DCOM user, and readable by the intended human user. Make sure both NTFS and share level permissions on the target file and its containing folder consider this. Latency for the file write can degrade performance.
    2) Most HFM implementations have two or more HFM application servers. You have no control over which server will execute the file, and also no control over what happens if two or more servers try to write to the same file. You can create a process which causes HFM to become single threaded if every time Sub Calculate for any entity on any server needs to open a single file. This is because the file object is single threaded. It could be multi-threaded if you write out to a database, but certainly far more complex.
    3) Wait state: while as in #2, this can single thread the process, and you have to consider whether your process will error out during moments when another process has access to the file, or if you will ignore that and simply proceed. This can have a significant impact on performance time if you decide to wait until the file becomes available.
    While I have done implementations where the overall solution required combining data across multiple scenarios, I consistently find this a cumbersome, error-prone, and poorly performing approach. For this reason, I always try to keep all of the data required in a single scenario. This is one key reason why the "DataSource" or "DataType" (or whatever you call it) approach is so popular and successful.
    Finally, you cannot write to the HFM system messages. You can, on the other hand, use Calc Manager to write out to a system log.
    --Chris                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Error for Generating a log file

    Hi Cezar sanos,
    i am trying to generate a log file for ODI with details like who logged in and what is is doing kind of things.
    For this i am executing the command like
    lagentscheduler.bat "-PORT=20910" "-NAME=localagent" "-V=2" > C:\OraHome_1\logs\agent1.log.
    But its getting the error like
    A JDK is required to execute Web Services with OracleDI. You are currently using a JRE.
    OracleDI: Starting Scheduler Agent ...
    Starting Oracle Data Integrator Agent...
    Version : 10.1.3.5 - 10/11/2008
    DwgJv.main: Exit. Return code:-1

    Just in case,
    the following message :
    A JDK is required to execute Web Services with OracleDI. You are currently using a JRE.
    is only a warning and not an error message....

  • Generating a log file from oracle db

    I want to generate a log file in which i want to dump some useful messages, when anyone does a dml operation on a table, and also, i want to have a switch like YES or NO (may be an environment variable,,) if i switch it to YES the log file should get generated, if NO then, no log file will be generated..
    can anyone help, how can you do this task ?
    thanks a lot in advance..
    srini

    You can use a Trigger and UTL_FILE to write to a file (on the server).
    Example:
    create or replace trigger test_file
      after insert or delete or update on test_case 
      for each row
    declare
    v_logfile utl_file.file_type;
    begin
      v_logfile := utl_file.fopen('\myfiles','test_file.log','a');
      if inserting then
         utl_file.put_line(v_logfile,'Inserting to table');
      elsif deleting then
         utl_file.put_line(v_logfile,'Deleting to table');
      else
         utl_file.put_line(v_logfile,'Updating to table');
      end if;
      utl_file.fclose(v_logfile);
    end test_file;
    I want to generate a log file in which i want to dump some useful messages, when anyone does a dml operation on a table, and also, i want to have a switch like YES or NO (may be an environment variable,,) if i switch it to YES the log file should get generated, if NO then, no log file will be generated..
    can anyone help, how can you do this task ?
    thanks a lot in advance..
    srini

  • Reader 10.1 update fails, creates huge log files

    Last night I saw the little icon in the system tray saying an update to Adobe Reader was ready to be installed.
    I clicked it to allow the install.
    Things seemed to go OK (on my Windows XP Pro system), although very slowly, and it finally got to copying files.
    It seemed to still be doing something and was showing that it was copying file icudt40.dll.  It still displayed the same thing ten minutes later.
    I went to bed, and this morning it still showed that it was copying icutdt40.dll.
    There is no "Cancel" button, so this morning I had to stop the install through Task Manager.
    Now, in my "Local Settings\TEMP" directory, I have a file called AdobeARM.log that is 2,350,686 KB in size and a file MSI38934.LOG that is 4,194,304 KB in size.
    They are so big I can't even look at them to see what's in them.  (Too big for Notepad.  When I tried to open the smaller log file, AdobeARM.log, with Wordpad it was taking forever and showing only 1% loaded, so after five minutes, I terminated the Wordpad process so I could actually do something useful with my computer.)
    You would think the installer would be smart enough to stop at some point when the log files begin to get enormous.
    There doesn't seem to be much point to creating log files that are too big to be read.
    The update did manage to remove the Adobe Reader X that was working on my machine, so now I can no longer read PDF files.
    Maybe I should go back Adobe Reader 9.
    Reader X never worked very well.
    Sometimes the menu bar showed up, sometimes it didn't.
    PDF files at the physics e-print archive always loaded with page 2 displayed first.  And if you forgot to disable the look-ahead capability, you could get banned from the e-print archive site altogether.
    And I liked the user interface for the search function a lot better in version 9 anyway.  Who wants to have to pop up a little box for your search phrase when you want to search?  Searching is about the most important and routine activity one does, other than going from page to page and setting the zoom.

    Hi Ankit,
    Thank you for your e-mail.
    Yesterday afternoon I deleted the > 2 GB AdobeARM.log file and the > 4.194 GB
    MSI38934.LOG file.
    So I can't upload them.  I expect I would have had a hard time doing so
    anyway.
    It would be nice if the install program checked the size of the log files
    before writing to them and gave up if the size was, say, three times larger
    than some maximum expected size.
    The install program must have some section that permits infinite retries or
    some other way of getting into an endless loop.  So another solution would be
    to count the number of retries and terminate after some reasonable number of
    attempts.
    Something had clearly gone wrong and there was no way to stop it, except by
    going into the Task Manager and terminating the process.
    If the install program can't terminate when the log files get too big, or if
    it can't get out of a loop some other way, there might at least be a "Cancel"
    button so the poor user has an obvious way of stopping the process.
    As it was, the install program kept on writing to the log files all night
    long.
    Immediately after deleting the two huge log files, I downloaded and installed
    Adobe Reader 10.1 manually.
    I was going to turn off Norton 360 during the install and expected there
    would be some user input requested between the download and the install, but
    there wasn't.
    The window showed that the process was going automatically from download to
    install. 
    When I noticed that it was installing, I did temporarily disable Norton 360
    while the install continued.
    The manual install went OK.
    I don't know if temporarily disabling Norton 360 was what made the difference
    or not.
    I was happy to see that Reader 10.1 had kept my previous preference settings.
    By the way, one of the default settings in "Web Browser Options" can be a
    problem.
    I think it is the "Allow speculative downloading in the background" setting.
    When I upgraded from Reader 9 to Reader 10.0.x in April, I ran into a
    problem. 
    I routinely read the physics e-prints at arXiv.org (maintained by the Cornell
    University Library) and I got banned from the site because "speculative
    downloading in the background" was on.
    [One gets an "Access denied" HTTP response after being banned.]
    I think the default value for "speculative downloading" should be unchecked
    and users should be warned that one can lose the ability to access some sites
    by turning it on.
    I had to figure out why I was automatically banned from arXiv.org, change my
    preference setting in Adobe Reader X, go to another machine and find out who
    to contact at arXiv.org [I couldn't find out from my machine, since I was
    banned], and then exchange e-mails with the site administrator to regain
    access to the physics e-print archive.
    The arXiv.org site has followed the standard for robot exclusion since 1994
    (http://arxiv.org/help/robots), and I certainly didn't intend to violate the
    rule against "rapid-fire requests," so it would be nice if the default
    settings for Adobe Reader didn't result in an unintentional violation.
    Richard Thomas

  • MDIS failed to generate the Log file!!!

    Hello All,
    Having a issue where MDIS is not generating the log file..
    The scenario is something like this-
    The files are getting archived and the records not flowing into MDM
    Basis team says-
    2014-06-30T14:11:33.339,47083231971072,24,"[MDS=sapdpm1 Repository=REAL_ESTATE ClientSystem=MDM_REAL_ESTATE Port=Building]: Nigerian Building updates part 2 - SLKDDY.txt is empty, the file will be skipped
    But the source file was having data it was not empty(bit strange!!)
    Also its not generating the LOG to analyze
    Regards,
    Girish

    Hi Shenoy,
    Let me explain the scenario--
    User uploads the file through Portal and through FTP records resides in MDM...the issues is when i tried to import through IM it worked and i tried manually push file through Filezilla FTP it worked.
    But when we upload file through portal, the file resides in Archive and generating the message-
    2014-06-30T14:11:33.339,47083231971072,24,"[MDS=sapdpm1 Repository=REAL_ESTATE ClientSystem=MDM_REAL_ESTATE Port=Building]: Nigerian Building updates part 2 - SLKDDY.txtis empty, the file will be skipped
    But the file is having data.
    Regards,
    Girish

  • Want to Generate custom logs with Log4J within Java Embedding activity

    Hi Gurus,
    i want to Generate custom logs with Log4J within Java Embedding activity. For that i have performed some steps, but the log file has not been created.
    ServerSide Configuration:
    *===============*
    1. I have copied the log4j-1.2.15.jar file to the "oracle.soa.ext_11.1.1" location and recreate the oracle.soa.ext.jar file using an-script, which appends the log4j.jar file to the classpath.( That was created successfully)
    2. Specify a File location in log4j.xml for creating the log file. (e.g. (<middleware_home>/config/customLog.log)
    3. Then create a folder and add log4j.xml and log4j.dtd file under the folder.
    4. Modify the startManagedWeblogicServer by adding JAVA_OPTIONS="-Dlog4j.configuration=<middleware_home>/config/log4j.debug.xml”
    5. Then restart the SOA server.
    log4j.xml :
    I have used fileAppender.
    DevelopmentSide Configuration:
    *===================*
    1. Create synchronous BPEL process and Add "JAVA Embedding Activity" inbetween receiveInput and replyOutput activity.
    2.Import the log4j-1.2.15.jar to the project libs
    3.In the BPEL source code, import the Logger class.
    4.Insert the code in the JAVA Embedding Activity
    Logger logger = Logger.getLogger("CustomLog");
    logger.info("CustomeLog " + getTitle());
    Compile and deploy the project into the SOA_server and test it. Test status is completed but log is not created in the mentioned location.
    Ref: http://blog.andrade.inf.br/search/label/Log4j
    Is there any other way to achieve this requirement. Please suggest.
    Thanks in Advance,
    Sharmistha

    You can try this.
    http://veejai24.blogspot.co.uk/2008/04/simple-way-to-implement-log4j-in-your.html
    Thanks,
    Vijay

  • How To Generate Debug Log Files for ebs jsp?

    hi   How To Generate Debug Log Files for ebs r12 jsp?
    and where i get the log .please help me thanks!

    Please check following MOS Document
    Oracle Application Server Diagnostic Tools and Log Files in Applications Release 12 (Doc ID 454178.1)

  • Idoc types that generate the log files

    Hi All,
    Can anybody provide me the list of IDOC types that generated application log files.
    As I found that material master IDOCs generate log files. Is there IDOC types other
    than material master that generate log files.
    Regards,
    Azra.

    Hi Azra,
    You can tell whether an IDOC does create the application log from the function module it uses.
    Go to SE37, and type in IDOC_INPUT_<message_type> for inbound or IDOC_OUTPUT_<message_type> for outbound, and search for any function module which start with 'BAL_'. If you can find it, then it creates application log files.
    Regards,
    Lim...

  • How a select statement generates redo logs

    Can some one explain how a select statement generates redo logs
    Naveen

    Redo with a select statement happens when dirty blocks get written to the database, and are then "cleaned up" when next read of that data block. This could happen when a large DML statement does not commit before the DB Writer needs to write modified blocks to disk. The next time the blocks are read by a select statement they get modified, hence REDO is generated by the select statement.
    Read the following for more and better understandings.
    http://www.dbspecialists.com/specialists/specialist2003-10.htm
    Jaffar

  • Incosistencies between Analyzer Server Console and stout.log file

    Hi,<BR><BR>In stout.log file of application server file there is a record:<BR>"Setting Current User Count To: 2 Users. Maximum Concurrent Licensed User Count Is: 10 Users." So 2 licences are ured, but checking "Analyer Server Console" there is only one user connected.<BR><BR>After restarting computer "Analyzer Server Console" and stout.log number of users are sinhronized. But I don't know what happens, but after some time this two parameters are not sinhronized anymore.<BR><BR>My problem: I have to show the number of user licences used and I am reading info rom stout.log. But something is not correct - it looks like stout.log doesn't show correct values?<BR><BR>Do I need to specify some setting or is there a bug in code?<BR><BR><BR>My system:<BR>Hyperion Analytic Server version 7.1.0<BR>Hyperion Analyzer Server version 7.0.1.8.01830<BR>IBM DB2 Workgroup Edition version 8 fixpack 9<BR>Tomcat version 4<BR>Windows 2003 Server

    Hi grofaty.<BR><BR>We use 7.0.0.0.01472 and I had experienced the same behaviour, <BR>Analyzer Server Console shows one more session than stdout.log.<BR><BR>If this difference 1 is a static value than you can assume it as an systematic bug...and do your license counting on it...<BR><BR>But again the Analyzer Server Console is not good as it should be for productive usage because all the information is only logged online til the next application restart. E.g. it is not helpful in using it for user tracking purposes. Do you use the stdout.log in such a way or have an idea how to grep measures for session logging analysis:<BR> - Session ID <BR> - User ID<BR> - Client ID <BR> - Total Number of Requests <BR> - Average Response (sec) <BR> - Login Time<BR> - Number of concurrent sessions<BR><BR>?

  • Generating encrypted logs in a untrusted machine

    Hi guys,
    I would like to get some ideas from your guys on how to solve a problem.
    Basically, we have an application that is installed in a unstrusted computer (i.e. the person whom uses this computer cannot be trusted).
    Unfortunately, the guys whom install the application is untrusted as well.
    Basically this application allows the user to perform some operations in the field (generate some log of those operations) and when the user returns
    to the headquarter those logs are read, so we can check if "field user" did any malicious action. The problem is that this log MUST be encrypted, otherwise
    he can edit the log file and cheat the system.
    Do your guys have any idea on how to solve that?
    PS: I was thinking on give a public key that should be used during the product installation (by using keytool) and use a private key on the server to read the logs, however I don't know if that is the best solution.
    Thanks and Regards

    Those cmdlets use DPAPI (
    http://msdn.microsoft.com/en-us/library/ms995355.aspx .)  DPAPI does generate new Master Keys from time to time, but they're all stored in the user's profile, and you should not lose the ability to decrypt old data.  Few questions:
    Is this script running as a domain account, or as a local user?
    Does anyone mess with the account's user profile, deleting all or parts of it and allowing them to be recreated?
    Your DPAPI master keys are stored in <Profile Dir>\AppData\Roaming\Microsoft\Protect\<Account SID>\ .  They're marked as Hidden and System, so you'll need to make sure you're viewing those files to be able to see them.
    The master key files in that directory are encrypted using a hash of the user's password; if anyone resets the password, you lose access to those keys (unless you're either in Active Directory, which has a backup mechanism in place on the domain controllers,
    or you have made a Password Recovery Disk for a local account.)  You mentioned that the account's password hasn't changed, so that's probably not an issue here, but I figured I'd mention it so you know how DPAPI works.
    If the account's password has not changed at all, my best guess is that someone has managed to delete the old master key(s) from the user profile, one of which was used to encrypt the SecureString originally.

  • Refresh Materilaized Views without generating Archive Logs

    Hello Gurus,
    I'm facing an emebarassing situation,
    I've a job every night to execute a dbms_refresh on several materialiled Views, but there are 2 Mviews that take almost 30 minutes each but another issue thay generates almost 10G of data and my File system becomes full in a minute.
    I've done an alter Mview Nologging but it doesn't change anything.
    Does soemon ahve an idea how to do a refresh without generating Archive logs - or
    without logging?
    thxs in advance
    C.

    Since this is a TRUNCATE and INSERT, any user / application querying the MV while the Refresh is running will see ZERO rows. Furthermore if the INSERT fails (e.g. insufficent space to add an extent), the MV remains with ZERO rows.
    A DELETE and INSERT avoids such situations (even if the refresh fails, both INSERT and DELETE are rolled-back and the MV is reverted to the state as it was before the Refresh began -- thus showing data from the previous refresh)
    Hemant K Chitale
    http://hemantoracledba.blogspot.com
    Just to clarify : My paragraphs above are not about data "missing" but about precautions to be taken. A COMPLETE Refresh is done during defined outages -- users/applications are made aware that data is not available while the Refresh is running.
    Edited by: Hemant K Chitale on Sep 1, 2009 10:53 AM

Maybe you are looking for

  • GarageBand running in the background with video app open

    I was wondering if GarageBand will run in the background while I'm using the video app on the iPad. Basically, I'm looking to have my guitar running into my iPad with GarageBand running, while watching/following along with a guitar video lesson I als

  • Annoying question mark in finder window

    There is a very annoying question mark in the finder window of my Mac, When I click on it it directs me to a skype disk image that I keep on an external hard drive with other backups. How can I get rid of it?

  • Could not resolve host (nil); Unknown error

    Trying to install arch linux into virtualbox... I got it to boot and everything, got my internet set up (I think), installed virtualbox guest stuff, and then had to reboot. I try to do sudo pacman -S xorg-twm xorg-xclock xterm so that I can test X be

  • How to implement hyperlink in java using swing components

    hi..... pls help me to find out solution how to implement hyperlink in java using swing components

  • CALL FUNCTION STARTING NEW TASK  DESTINATION

    Hi All, I have tried to use CALL FUNCTION 'ZWHV_TEST' STARTING NEW TASK 'TEST' DESTINATION 'NONE' But it didnt work. Please explain me  to use above statement and detail me more about NEW TASK 'TEST'. How to declare TEST in new task ? Thanks , Vinay.