[Solved] log files

Hi everyone,
I would like to know what log file I must check to find out why my laptop is shutting down unexpectedly.  It is an 8 year old laptop, so it may have a hardware problem. If it is a hardware problem (like a bad power supply or overheating),  I would like to know if these type of things are logged?
Thanks in advance!
Last edited by fwin (2011-09-15 04:30:08)

fwin wrote:
Thank you karol and lifeafter2am,
I checked dmesg and kernel.log, and I could only understand about 10 percent of the output, I will spend a few days trying to learn this (googling) and then I will report back if I still have any question.
I don't think it is overheating, it just feels warm (normal). This laptop is my home server so it is usually not running many applications, it does not even have a GUI. Anyway, I will try the lm_sensor thing just to learn about that, it sounds like fun!
Thanks again!
Feel free to post some of it either here or on pastebin if you're too lost. 
Last edited by lifeafter2am (2011-09-13 04:18:20)

Similar Messages

  • [SOLVED]Log files getting LARGE

    I ran a pacman -Syu for the first time in several months yesterday, and my computer has become almost useless due the fact that everything.log, kernel.log and messages.log gets extremely large (3.8 GB) after a while, causing / to become 100% full.
    I've located the following in kernel.log:
    Mar 14 15:06:44 elvix attempt to access beyond end of device
    Mar 14 15:06:45 elvix attempt to access beyond end of device
    Mar 14 15:06:45 elvix sda5: rw=0, want=1812442544, limit=412115382
    Mar 14 15:06:45 elvix attempt to access beyond end of device
    Mar 14 15:06:45 elvix sda5: rw=0, want=1812442544, limit=412115382
    Mar 14 15:06:45 elvix attempt to access beyond end of device
    Not sure what it means, but the last two lines are repeated XX times and are the reason why log files grow beyond limits. Anyone got ideas to what can be done to fix this?
    Last edited by bistrototal (2008-03-14 16:27:15)

    logrotate works really well:
    http://www.archlinux.org/packages/14754/
    There's quite a few threads about configuration floating around.

  • Allowing special user to read log files [Solved]

    Hi
    I want to add a group (log) that has the privledges to read /var/log files
    and then add my user to that group.
    How do i accomplish this
    This is to run root-tail in .xinitrc
    Or is there an other way?
    //Clanman

    [root@master fredrik]# ls -al /usr/bin/root-tail
    -rwxr-xr-x  1 root root 22600 2006-01-08 04:58 /usr/bin/root-tail
    [root@master fredrik]# chmod 640 /usr/bin/root-tail
    [root@master fredrik]# ls -al /usr/bin/root-tail
    -rw-r-----  1 root root 22600 2006-01-08 04:58 /usr/bin/root-tail
    [root@master fredrik]# chgrp log /usr/bin/root-tail
    [root@master fredrik]# ls -al /usr/bin/root-tail
    -rw-r-----  1 root log 22600 2006-01-08 04:58 /usr/bin/root-tail
    exiting root
    fredrik@master ~]$ root-tail -g 800x250+10+10 -justify -fn '-*-verdana-*-*-*-*-10-*-*-*-*-*-*-*' /var/log/errors.log,orange,'ALERT'
    -bash: /usr/bin/root-tail: Permission denied
    Still no go
    //Clanman

  • [SOLVED] dvd drive not working, log file filled with error.

    hi,
    i dont know if this is something to do with the latest kernel update or something else that may have occured.
    but my dvd writer has stopped working. no light on the front, pressing the button does not open the tray.
    it wtill works fine when i boot into windows (dual boot with XP) so it's not a hardware issue.
    also my everything.log file is filling up with the following every two seconds!
    Mar  1 19:27:56 adax hda: status error: status=0x59 { DriveReady SeekComplete DataRequest Error }
    Mar  1 19:27:56 adax hda: status error: error=0x00 { }
    Mar  1 19:27:56 adax ide: failed opcode was: unknown
    Mar  1 19:27:56 adax hda: drive not ready for command
    Mar  1 19:27:56 adax hda: status error: status=0x59 { DriveReady SeekComplete DataRequest Error }
    Mar  1 19:27:56 adax hda: status error: error=0x00 { }
    Mar  1 19:27:56 adax ide: failed opcode was: unknown
    Mar  1 19:27:56 adax hda: drive not ready for command
    Mar  1 19:27:56 adax hda: status error: status=0x59 { DriveReady SeekComplete DataRequest Error }
    Mar  1 19:27:56 adax hda: status error: error=0x00 { }
    Mar  1 19:27:56 adax ide: failed opcode was: unknown
    Mar  1 19:27:56 adax hda: drive not ready for command
    Mar  1 19:27:56 adax hda: status error: status=0x59 { DriveReady SeekComplete DataRequest Error }
    Mar  1 19:27:56 adax hda: status error: error=0x00 { }
    Mar  1 19:27:56 adax ide: failed opcode was: unknown
    Mar  1 19:27:56 adax hda: drive not ready for command
    can someone please help me before i run out of disk space for my logs!
    thanks,
    ad.
    Last edited by adax (2008-03-01 23:25:37)

    Hello,
    I strongly suspect that you need to change 'ide' to 'pata' in your /etc/mkinitcpio.conf
    #HOOKS="base udev autodetect ide scsi sata usb keymap filesystems"
    HOOKS="base udev autodetect pata scsi sata usb keymap filesystems"
    Then recreate with: 
    mkinitcpio -g /boot/kernel26.img
    Last edited by colnago (2008-03-01 21:21:37)

  • [SOLVED] Writting to log file on Oracle AS in ADF BC

    Hi!
    I'm trying to write some information to any of the log files on Oracle Application Server but can't seem to get it working.
    I have a http servlet class in which I try to log records like this:
    Logger loLogger = Logger.getAnonymousLogger();
    loLogger.info("HttpServlet: doGet start!");This code displays
    1.4.2008 9:44:02 package.HttpServlet doGet
    INFO: HttpServlet: doGet start!
    in JDeveloper output window, but after deployment to AS it is not written to any of the files. I do not use a custom logger or log files in my application so my question is: how can I write information to any of the log files on AS (for example logs in <ORACLE_HOME>\opmn\logs folder or in log.xml file in oc4j folder of j2ee home log)?
    I've been reading about j2ee logging but as I said, can't get it to work. Are there any ADF classes to do this?
    Thanks,
    regards!

    Hi!
    Thank you both for your effort. Both ways work now, I messed something up at the beginning.
      Logger loLogger = Logger.getAnonymousLogger();
        loLogger.info("method start!");writes to 'default_group~home~default_group~1.log' file in opmn/logs directory, while
      ADFLogger loADFLogger = ADFLogger.createADFLogger("oracle");
        loADFLogger.log(ADFLogger.NOTIFICATION, "method start!");writes to 'log.xml' in j2ee\home\log\home_default_group_1\oc4j directory.
    Now I just have to decide which is better for me ;).
    BB

  • Display string from a log file on a Dashboard?

    Hello everyone,
    I need a little guidance on how to solve this problem in SCOM 2012 R2.
    I need SCOM to do this:
    Read a log file on a specific application server (say its hostname is server21)
    Look for the text: Total number of users: 15
    Display the text “Users on Server21 =
     15” on a SCOM dashboard (or something to that effect)
    Sounds simple enough? Well, I am stumped :-(
    I have read and followed tutorials on internet, telling me how to create a rule and alerts. However I don’t want alerts. I want this data to be simply shown on a dashboard.
    How can I do that?
    PS: I hear that I may have to create my own management pack with only server21 in it. Is that right?
    -Rajeev rajdude.com

    Thanks for the tip. The powershell grid widget is pretty powerful. However it will take me quite some time and effort to write a PS script which can extract the exact info I need from that application's log file. I was hoping SCOM's own log parsing
    capabilities would do it.
    By the way, yes, we can create a MP and have only one server in it. We can use the (free) 
    MPAuthor to make a MP with only one server in it. Here is a video showing how to do it...
    http://www.silect.com/static/mpauthor/MP_Author_Creating_a_New_Single_Server_Application_MP.mp4
    I tried doing what I want using MPAuthor, but it again boiled down to me writing a PS script which can extract the exact info I need
    from that application's log file.
    -Rajeev rajdude.com

  • Read data from a log file to flex textarea while deploying

    Hi,
    I am new to flex. I am able to read data from a log file that placed in C drive and able to write that data into flex textarea, but I am unable to read the same file while it is deployed into weblogic sever.
    Could anyone please tell me how to solve this problem.
    Thanks,
    Sri.

    Do you have it trying to read that same file on your C drive or is the file actually deployed somewhere on the server?

  • Empty Log File - log settings will not save

    Description of Problem or Question:
    Cannot get logging to work in folder D:\Program Files\Business Objects\Dashboard and Analytics 12.0\server\log
    (empty log file is created)
    Product\Version\Service Pack\Fixpack (if applicable):
    BO Enterorise 12.0
    Relevant Environment Information (OS & version, java or .net & version, DB & version):
    Server: windows Server 2003 Enterprise SP2.
    Database Oracle 10g
    Client : Vista
    Sporadic or Consistent (if applicable):
    Consistent
    What has already been tried (where have you searched for a solution to your question/problem):
    Searched forum, SMP
    Steps to Reproduce (if applicable):
    From InfoViewApp, logged in as Admin
    Open ->Dashboard and Analytics Setp -> Parameters -> Trace
    Check "Log to folder" and "SQL Queries", Click Apply.
    Now, navigate away and return to this page - the "Log to folder" is unchecked. Empty log file is created.

    Send Apple feedback. They won't answer, but at least will know there is a problem. If enough people send feedback, it may get the problem solved sooner.
    Feedback
    Or you can use your Apple ID to register with this site and go the Apple BugReporter. Supposedly you will get an answer if you submit feedback.
    Feedback via Apple Developer
    Do a backup.
    Quit the application.
    Go to Finder and select your user/home folder. With that Finder window as the front window, either select Finder/View/Show View options or go command - J.  When the View options opens, check ’Show Library Folder’. That should make your user library folder visible in your user/home folder.  Select Library. Then go to Preferences/com.apple.systempreferences.plist. Move the .plist to your desktop.
    Restart, open the application and test. If it works okay, delete the plist from the desktop.
    If the application is the same, return the .plist to where you got it from, overwriting the newer one.
    Thanks to leonie for some information contained in this.

  • Moving the log file of a publisher database SQL Server 2008

    There are many threads on this here. Most of them not at all helpful and some of them wrong. Thus a fresh post.
    This post regards SQL Server 2008 (10.0.5841)
    The PUBLISHER database primary log file which is currently of 3 blocks and not extendable,
    must be moved as the LUN is going away.
    The database has several TB of data and a large number of push transactional replications as well as a couple of bi-directional replications.
    While the primary log file is active, it is almost never (if ever) used due to its small fixed size.
    We are in the 20,000 TPS range at peak (according to perfmon). This is a non-trivial installation.
    This means that
    backup/restore is not even a remotely viable option (it never is in the real world)
    downtime minimization is critical - measured in minutes or less.
    dismantling and recreating the replications is doable, but I have to say, I have zero trust in the script writer to generate accurate scripts. Many of these replications were originally set up in older versions of SQL Server and have come along for the
    ride as upgrades have occurred. I consider scripting everything and dismantling the whole lot pretty high risk. In any case, I do not want to have to reinitialize any replications as this takes, effectively, an eternity.
    Possible solution:
    The only option I can think of is to wind down everything, such that there are zero outstanding uncommitted transactions and detach the database, delete the offending log file and reattach using the CREATE DATABASE xyz ATTACH_REBUILD_LOG option.
    This should, if I have understood things correctly, cause SQL Server to recreate the default log file in the same directory as the .mdf file. I am not sure what will happen to the secondary log file which is not moving anywhere at this point.
    The hard bit is insuring that every transaction in the active log files have been replicated before shutdown. This is probably doable. I do not know how to manually flush any left over transactions to replication. I expect if I shut down all "real"
    activity and wait for a certain amount of time, eventually all the replications will show "No replicated transactions are available" and then I would be good to go.
    Hillary, if you happen to be there, comments appreciated.

    Hi Philip
    you should try this long back suggested way of stopping replication and restore db and rename or detach attach
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/6731803b-3efa-4820-a303-4ffb7edf154a/detaching-a-replicated-database?forum=sqlreplication
    Thanks
    Saurabh Sinha
    http://saurabhsinhainblogs.blogspot.in/
    Please click the Mark as answer button and vote as helpful
    if this reply solves your problem
    I do not wish to be rude, but which part of the OP didn't you understand?
    Specifically the bit about 20,000 transactions a second and database size of several TB. Do you have any concept whatsoever of what this means? I will answer for you, "no, you are clueless" as your answer clearly shows.
    Stop wasting bandwidth by proposing pointless and wrong solutions which indicate that you did not read the OP, or do you just do this to generate points?
    Also, you clearly failed to notice that I was on the thread to which you referred, and I had some pointed comments to make. This thread was an attempt to garner some input for an alternative proposal.

  • Transaction log shipping restore with standby failed: log file corrupted

    Restore transaction log failed and I get this error: for only 04 no of database in same SQL server, renaming are working fine.
    Date                     
    9/10/2014 6:09:27 AM
    Log                        
    Job History (LSRestore_DATA_TPSSYS)
    Step ID                
    1
    Server                  
    DATADR
    Job Name                           
    LSRestore_DATA_TPSSYS
    Step Name                        
    Log shipping restore log job step.
    Duration                             
    00:00:03
    Sql Severity        0
    Sql Message ID 0
    Operator Emailed           
    Operator Net sent          
    Operator Paged               
    Retries Attempted         
    0
    Message
    2014-09-10 06:09:30.37  *** Error: Could not apply log backup file '\\10.227.32.27\LsSecondery\TPSSYS\TPSSYS_20140910003724.trn' to secondary database 'TPSSYS'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred while processing the log for database 'TPSSYS'. 
    If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log.
    An error occurred during recovery, preventing the database 'TPSSYS' (13:0) from restarting. Diagnose the recovery errors and fix them, or restore from a known good backup. If errors are not corrected or expected, contact Technical Support.
    RESTORE LOG is terminating abnormally.
    Processed 0 pages for database 'TPSSYS', file 'TPSSYS' on file 1.
    Processed 1 pages for database 'TPSSYS', file 'TPSSYS_log' on file 1.(.Net SqlClient Data Provider) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  Skipping log backup file '\\10.227.32.27\LsSecondery\TPSSYS\TPSSYS_20140910003724.trn' for secondary database 'TPSSYS' because the file could not be verified.
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred restoring the database access mode.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteScalar requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred restoring the database access mode.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteScalar requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  Deleting old log backup files. Primary Database: 'TPSSYS'
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  The restore operation completed with errors. Secondary ID: 'dd25135a-24dd-4642-83d2-424f29e9e04c'
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not cleanup history.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.38  ----- END OF TRANSACTION LOG RESTORE    
    Exit Status: 1 (Error)

    I Have restore the database to new server and check with new log shipping but its give this same error again, If it is network issue i believe issue need to occur on every database in  that server with log shipping configuration
    error :
    Message
    2014-09-12 10:50:03.18    *** Error: Could not apply log backup file 'E:\LsSecondery\EAPDAT\EAPDAT_20140912051511.trn' to secondary database 'EAPDAT'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-12 10:50:03.18    *** Error: An error occurred while processing the log for database 'EAPDAT'.  If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log.
    An error occurred during recovery, preventing the database 'EAPDAT' (8:0) from restarting. Diagnose the recovery errors and fix them, or restore from a known good backup. If errors are not corrected or expected, contact Technical Support.
    RESTORE LOG is terminating abnormally.
    can this happened due to data base or log file corruption, if so how can i check on that to verify the issue
    Its not necessary if the issue is with network it would happen every day IMO it basically happens when load on network is high and you transfer log file which is big in size.
    As per message database engine was not able to restore log backup and said that you must rebuild log because it did not find out log to be consistent. From here it seems log corruption.
    Is it the same log file you restored ? if that is the case since log file was corrupt it would ofcourse give error on wehatever server you restore.
    Can you try creating logshipping on new server by taking fresh full and log backup and see if you get issue there as well. I would also say you to raise case with Microsoft and let them tell what is root cause to this problem
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Articles

  • Error in sql_engine_core_inst_Cpu64_1.log file for SQL Server 2012

    I was running SQL Server 2012 in Win 7 64 bit OS. Some how I found SQLSERVER server is not running and I was not able to run the service. Then I tried to unisntall and install again. And I found that some of the SQL Server
    features got failed,  like: Reporting Services - Native, Database Engine Services, Data Quality Services, Full-Text and Semantic Extractions for Search and SQL Server Replication. I checked the Summary.txt
    file from location C:\Program Files\Microsoft SQL Server\110\Setup Bootstrap\Log. It thsi logs referes to another log file location like: 
    Feature:                       Database Engine Services
      Status:                        Failed: see logs for details
      Reason for failure:            An error occurred during the setup process of the feature.
      Next Step:                     Use the following information to resolve the error, and then try the setup process again.
      Component name:                SQL Server Database Engine Services Instance Features
      Component error code:          1639
      Component log file:            C:\Program Files\Microsoft SQL Server\110\Setup Bootstrap\Log\20140127_151409\sql_engine_core_inst_Cpu64_1.log
      Error help link:               http://go.microsoft.com/fwlink?LinkId=20476&ProdName=Microsoft+SQL+Server&EvtSrc=setup.rll&EvtID=50000&ProdVer=11.0.3128.0&EvtType=sql_engine_core_inst.msi%400x162A16FE%400x1639
    I did unisntall/install multipel times. But fhw to fix the issue that saying in sql_engine_core_inst_Cpu64_1.log file.: 
    === Verbose logging started: 1/27/2014  15:45:04  Build type: SHIP UNICODE 5.00.7601.00  Calling process: C:\Program Files\Microsoft SQL Server\110\Setup Bootstrap\SQLServer2012\x64\ScenarioEngine.exe ===
    MSI (c) (C4:BC) [15:45:04:321]: Resetting cached policy values
    MSI (c) (C4:BC) [15:45:04:321]: Machine policy value 'Debug' is 0
    MSI (c) (C4:BC) [15:45:04:321]: ******* RunEngine:
               ******* Product: E:\Software\SQL Full Installation\x64\setup\sql_engine_core_inst_msi\sql_engine_core_inst.msi
               ******* Action: 
               ******* CommandLine: **********
    MSI (c) (C4:BC) [15:45:04:322]: Client-side and UI is none or basic: Running entire install on the server.
    MSI (c) (C4:BC) [15:45:04:322]: Grabbed execution mutex.
    MSI (c) (C4:BC) [15:45:04:323]: Cloaking enabled.
    MSI (c) (C4:BC) [15:45:04:323]: Attempting to enable all disabled privileges before calling Install on Server
    MSI (c) (C4:BC) [15:45:04:323]: Incrementing counter to disable shutdown. Counter after increment: 0
    MSI (s) (38:8C) [15:45:04:327]: Running installation inside multi-package transaction E:\Software\SQL Full Installation\x64\setup\sql_engine_core_inst_msi\sql_engine_core_inst.msi
    MSI (s) (38:8C) [15:45:04:327]: Grabbed execution mutex.
    MSI (s) (38:A0) [15:45:04:327]: Resetting cached policy values
    MSI (s) (38:A0) [15:45:04:327]: Machine policy value 'Debug' is 0
    MSI (s) (38:A0) [15:45:04:327]: ******* RunEngine:
               ******* Product: E:\Software\SQL Full Installation\x64\setup\sql_engine_core_inst_msi\sql_engine_core_inst.msi
               ******* Action: 
               ******* CommandLine: **********
    MSI (s) (38:A0) [15:45:04:452]: Machine policy value 'TransformsSecure' is 0
    MSI (s) (38:A0) [15:45:04:452]: User policy value 'TransformsAtSource' is 0
    MSI (s) (38:A0) [15:45:04:475]: Machine policy value 'DisableUserInstalls' is 0
    MSI (s) (38:A0) [15:45:04:475]: Specified instance {18B2A97C-92C3-4AC7-BE72-F823E0BC895B} via transform :InstID01.mst;:InstName02.mst is already installed. MSINEWINSTANCE requires a new instance that is not installed.
    MSI (s) (38:A0) [15:45:04:475]: MainEngineThread is returning 1639
    MSI (s) (38:8C) [15:45:04:477]: User policy value 'DisableRollback' is 0
    MSI (s) (38:8C) [15:45:04:477]: Machine policy value 'DisableRollback' is 0
    MSI (s) (38:8C) [15:45:04:477]: Incrementing counter to disable shutdown. Counter after increment: 0
    MSI (s) (38:8C) [15:45:04:478]: Note: 1: 1402 2: HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\Installer\InProgress 3: 2 
    MSI (s) (38:8C) [15:45:04:478]: Note: 1: 1402 2: HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\Installer\InProgress 3: 2 
    MSI (s) (38:8C) [15:45:04:478]: Decrementing counter to disable shutdown. If counter >= 0, shutdown will be denied.  Counter after decrement: -1
    MSI (s) (38:8C) [15:45:04:478]: Restoring environment variables
    MSI (c) (C4:BC) [15:45:04:481]: Decrementing counter to disable shutdown. If counter >= 0, shutdown will be denied.  Counter after decrement: -1
    MSI (c) (C4:BC) [15:45:04:481]: MainEngineThread is returning 1639
    === Verbose logging stopped: 1/27/2014  15:45:04 ===
    Please advice. I searched many places and did not get actual solution that will resolve my case.
    Nusrat Akhter

    Hi Nusrat Akhter,
    According to your description, when installing the first install failed, and it looks like the instance did not uninstall completely. To solve the issue I recommend you changed the instance name in re-installing SQL Server, and check if it can work fine.
    There is a similar issue that the SQL engine will not start, and there is no method of uninstalling either it, or any of the components etc. You can review the following post.
    https://connect.microsoft.com/SQLServer/feedback/details/710892/sql-server-2012-rc0-install-fails-citing-invalid-drivers
    Thanks,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • Corrupt LOG File MaxDB 7.5 (Netweaver 04)

    At the Installation the LOG went full. LOG-Mode = SINGLE.
    Then I have added a new LOG File. Nothing happens. After rebooting the maschine,
    the database doesnt come up, in krnldiag there is following:
    2004-11-14 22:46:30  7579     11000 vattach  '/sapdb/XI3/sapdata/DISKD0001' devno 1 T58 succeeded
    2004-11-14 22:46:30  7520     11597 IO       Open '/sapdb/XI3/saplog/DISKL001' successfull, fd: 16
    2004-11-14 22:46:30  7633     12821 TASKING  Thread 7633 starting
    2004-11-14 22:46:30  7632     11565 startup  DEVi started
    2004-11-14 22:46:30  7520     11597 IO       Open '/sapdb/XI3/saplog/DISKL001' successfull, fd: 17
    2004-11-14 22:46:30  7634     12821 TASKING  Thread 7634 starting
    2004-11-14 22:46:30  7633     11565 startup  DEVi started
    2004-11-14 22:46:30  7579     11000 vattach  '/sapdb/XI3/saplog/DISKL001' devno 2 T58 succeeded
    2004-11-14 22:46:30  7520     11597 IO       Open '/data2/sapdb/XI3/DISKL002' successfull, fd: 19
    2004-11-14 22:46:30  7635     12821 TASKING  Thread 7635 starting
    2004-11-14 22:46:30  7634     11565 startup  DEVi started
    2004-11-14 22:46:30  7520     11597 IO       Open '/data2/sapdb/XI3/DISKL002' successfull, fd: 20
    2004-11-14 22:46:30  7636     12821 TASKING  Thread 7636 starting
    2004-11-14 22:46:30  7635     11565 startup  DEVi started
    2004-11-14 22:46:30  7579     11000 vattach  '/data2/sapdb/XI3/DISKL002' devno 3 T58 succeeded
    2004-11-14 22:46:30  7579 ERR    30 IOMan    Log volume configuration corrupted: Linkage missmatch between volume 1 and 2
    2004-11-14 22:46:30  7579     11000 vdetach  '/sapdb/XI3/sapdata/DISKD0001' devno 1 T58
    2004-11-14 22:46:30  7520     12822 TASKING  Thread 7631 joining
    2004-11-14 22:46:30  7636     11565 startup  DEVi started
    2004-11-14 22:46:30  7631     11566 stop     DEVi stopped
    2004-11-14 22:46:30  7520     12822 TASKING  Thread 7632 joining
    2004-11-14 22:46:30  7579 ERR    16 Admin    RestartFilesystem failed with 'I/O error'
    2004-11-14 22:46:30  7579 ERR     8 Admin    ERROR 'disk_not_accessibl' CAUSED EMERGENCY SHUTDOWN
    2004-11-14 22:46:30  7579     12696 DBSTATE  Change DbState to 'SHUTDOWN'(25)
    Is there any way to delete the corrupt LOG files in order to add new one? The data file is ok.
    Thanks
    Helmut

    Thread was also opened on the Mailing list:
    Here is the answer
    Hi,
    in most situations an "add log volume" does not solve
    a log full situation, because the added logvolume is
    accessible first at the moment, when the write position
    on the log is at the end of the old logvolume.
    If a LOGFULL occurs, the save log!
    All "parts" of the LogVolume are numbered and each one
    has stored the identifying number of it's predecessor and
    it successor in it's header-page. The message
         "Linkage missmatch between volume"
    means, that these values does not match any more.
    Did you somehow changed the contents of this file/device?
    If so: can you revert the change? Then the restart should
    work again. do a logbackup and your database will run again.
    If this doesn't work for you, you could patch the
    header pages of the logvolume (I assume you do not really
    want this).
    Another way to solve the situation is to do a data backup
    in admin-mode, reinstall the database and recover the
    databackup, because the removal of a logvolume is not
    supported.
    kind regards, Martin

  • Log file full

    Hi,
    My log file is full, steps to reduce the size of the logfile?

    Old question: While drafting question you get similar questions . If you whould have read that , you might have fixed same till now.
    Btw if below solution doesnt work you can check :Sql Server Transaction Log File Is Not Shrinking
    Thanks Saurabh Sinha
    http://saurabhsinhainblogs.blogspot.in/
    Please click the Mark as answer button and vote as helpful
    if this reply solves your problem

  • Help! SQL server database log file increasing enormously

    I have 5 SSIS jobs running in sql server job agent and some of them are pulling transactional data into our database over the interval of 4 hours frequently. The problem is log file of our database is growing rapidly which means in a day, it eats up 160GB of
    disk space. Since our requirement dont need In-point recovery, so I set the recovery model to SIMPLE, eventhough I set it to SIMPLE, the log
    data consumes more than 160GB in a day. Because of disk full, the scheduled jobs getting failed often.Temporarily I am doing DETACH approach
    to cleanup the log.
    FYI: All the SSIS packages in the job is using Transaction on
    some tasks. for eg. Sequence Cointainer
    I want a permanent solution to keep log file in a particular memory limit and as I said earlier I dont want my log data for future In-Point recovery, so no need to take log backup at all.
    And one more problem is that in our database,the transactional table has 10 million records in it and some master tables have over 1000 records on them but our mdf file
    size is about 50 GB now.I dont believe that this 10 million records should make it to 50GB memory consumption.Whats the problem here?
    Help me on these issues. Thanks in advance.

    And one more problem is that in our database,the transactional table has 10 million records in it and some master tables have over 1000 records on them but our mdf file
    size is about 50 GB now.I dont believe that this 10 million records should make it to 50GB memory consumption.Whats the problem here?
    Help me on these issues.
    For SSIS part of question it would be better if you ask in SSIS forum although noting is going to change about logging behavior. You can increase some space on log file and also should batch your transactions as already suggested
    Regarding memory question about SQL Server, once it utilizes memory is not going to release unless there is windows OS faces memory pressure and SQLOS asks SQL Server to trim down its memory consumption. So again if you have set max server memory to some
    where near 50 SQL Server will utilize that much memory eventually. So what you are seeing is totally normal. Remember its costtly task for SQL Server to release and take memory so it avoids it by caching as much as possible plus it caches more so as to avoid
    physical reads which are costly
    When log file is getting full what does below transaction return
    select log_reuse_wait_desc from sys.databases where name='db_name'
    Can you manually introduce chekpoint in ETL query. Try this it might help you
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • Unable to debug the Data Template Error in the Log file

    Hi,
    I am unable to debug the log file error message Please can anybody explain me in detail where the error lies and how to solve the error.The log file shows the following message.
    XDO Data Engine ver 1.0
    Resp: 50554
    Org ID : 204
    Request ID: 2865643
    All Parameters: USER_ID=1318:REPORT_TYPE=Report Only:P_SET_OF_BOOKS_ID=1:TRNS_STATUS=Posted:P_APPROVED=Not Approved:PERIOD=Sep-05
    Data Template Code: ILDVAPDN
    Data Template Application Short Name: CLE
    Debug Flag: Y
    {TRNS_STATUS=Posted, REPORT_TYPE=Report Only, PERIOD=Sep-05, USER_ID=1318, P_SET_OF_BOOKS_ID=1, P_APPROVED=Not Approved}
    Calling XDO Data Engine...
    java.lang.NullPointerException
         at oracle.apps.xdo.dataengine.DataTemplateParser.getObjectVlaue(DataTemplateParser.java:1424)
         at oracle.apps.xdo.dataengine.DataTemplateParser.replaceSubstituteVariables(DataTemplateParser.java:1226)
         at oracle.apps.xdo.dataengine.XMLPGEN.writeData(XMLPGEN.java:398)
         at oracle.apps.xdo.dataengine.XMLPGEN.writeGroupStructure(XMLPGEN.java:281)
         at oracle.apps.xdo.dataengine.XMLPGEN.processData(XMLPGEN.java:251)
         at oracle.apps.xdo.dataengine.XMLPGEN.processXML(XMLPGEN.java:192)
         at oracle.apps.xdo.dataengine.XMLPGEN.writeXML(XMLPGEN.java:222)
         at oracle.apps.xdo.dataengine.DataProcessor.processData(DataProcessor.java:334)
         at oracle.apps.xdo.oa.util.DataTemplate.processData(DataTemplate.java:236)
         at oracle.apps.xdo.oa.cp.JCP4XDODataEngine.runProgram(JCP4XDODataEngine.java:272)
         at oracle.apps.fnd.cp.request.Run.main(Run.java:148)
    Start of log messages from FND_FILE
    Start of After parameter Report Trigger Execution..
    Gl Set of Books.....P
    Organization NameVision Operations
    Entering TRNS STATUS POSTED****** 648Posted
    end of the trns status..687 Posted
    currency_code 20USD
    P_PRECISION 272
    precision 332
    GL NAME 40Vision Operations (USA)
    Executing the procedure get format ..
    ExecutED the procedure get format and the Result..
    End of Before Report Execution..
    End of log messages from FND_FILE
    Executing request completion options...
    ------------- 1) PUBLISH -------------
    Beginning post-processing of request 2865643 on node AP615CMR at 28-SEP-2006 07:58:26.
    Post-processing of request 2865643 failed at 28-SEP-2006 07:58:38 with the error message:
    One or more post-processing actions failed. Consult the OPP service log for details.
    Finished executing request completion options.
    Concurrent request completed
    Current system time is 28-SEP-2006 07:58:38
    Thanks & Regards
    Suresh Singh

    Generally the DBAs are aware of the OPP service log. They can tell you the cause of the problem.
    Anyway, how did you resolve the issue?

Maybe you are looking for