Transactional log grows abnormally to 160GB

Dear All
I am new to SQL database . Im my production server transactional log grow upto 160GB and only 7 GB left in the file system
Like dont want to add the any space in the file system and we have everyday full backup schedule.
Is there any permanent solution. I want to take the proactive action before my database hang.
Is also check the note 421644.

Hi Tarik!
Probably the fact that your TLOG is so high is because you did not back it up at all in the beginning or you just did not back it up for some time. This does not mean that its allocated space is full, and in fact, if you defined a proper backup policy you can shrink it to a more reasonable value as explained in SAP note 363018 (section E.b ).
Now, in order to avoid having it grown so much again is to ensure that you perform TLOG backups regularly (the frequency depends on your business needs, you should take your own conclusions depending on your system activity); check the autogrow settings just in case a single grow means a huge increase (imagine you set 120 GB!).
In case that you have any problems still e.g. because the TLOG grows even though you back it up regularly, you can check if there is a open transaction that prevents it to be backed up, as you already know note 421644.
Good luck! 
-Jesú

Similar Messages

  • SQL Server Database - Transaction logs growing largely with Simple Recovery model

    Hello,
    There is SQL server database on client side in production environment with huge transaction logs.
    Requirement :
    1. Take database backup
    2. Transaction log backup is not required. - so it is set to Simple recovery model.
    I am aware that, Simple Recovery model also increases the transaction logs same as in Full Recovery model as given on below link.
    http://realsqlguy.com/origins-no-simple-mode-doesnt-disable-the-transaction-log/
    Last week, this transaction log became of 1TB size and blocked everything on the database server.
    How to over come with this situation?
    PS :  There are huge bulk uploads to the database tables.
    Current Configuration :
    1. Simple Recovery model
    2. Target Recovery time : 3 Sec
    3. Recovery interval : 0
    4. No SQL Agent job schedule to shrink database.
    5. No other checkpoints created except automatic ones.
    Can anyone please guide me to have correct configuration on SQL server for client's production environment?
    Please let me know if any other details required from server.
    Thank you,
    Mittal.

    @dave_gona,
    Thank you for your response.
    Can you please explain me this in more details -- 
    What do you mean by one batch ?
    1. Number of rows to be inserted at a time ?
    2. or Size of data in one cell does matter here.
    As in my case, I am clubbing together all the data in one xml (on c# side) and inserting it as one record. Data is large in size, but only 1 record is inserted.
    Is it a good idea to shrink transaction log periodically, as it is not happening itself in simple recovery model.
    HI Mittal,
    Shrinking is bad activity yu should not shrink log files regularly, in rare case if you want to recovery space you may do it.
    Have manual chekpoints in Bulk insert operation.
    I cannot tell upfront what should be batch size but you can start with 1/4 th of what you are currently inserting.
    Most important what does below query return for database
    select log_reuse_wait_desc from sys.databases where name='db_name'
    The value it returns is what stopping the log from getting cleared and reused.
    What is version and editon of SQl server we are talking about. What is output of
    select @@version
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • Why is my transaction file growing when db is in simple mode and there are no waiting transactions

    I noticed that my database log file has reached 160GBs. My database is in simple recovery mode and I used the following query to see if there are
    any waiting transactions.
    SELECT d.name,d.log_reuse_wait,d.log_reuse_wait_desc
    FROM sys.databases d
    It returned log_reuse_wait 0 and description 'NOTHING'. Why is my transaction log growing then? Please advice.
    mayooran99

    DBCC opentran will give spid, you can pass that spid to dbcc inputbuffer(spid) to see what transaction is running. Also you can do sp_who2 spid will give other details.
    You can make use of these queries to get MORE details about OPEN and ACTIVE transactions.
    SELECT
    trans.session_id as [Session ID],
    trans.transaction_id as [Transaction ID],
    tas.name as [Transaction Name],
    tds.database_id as [Database ID]
    FROM sys.dm_tran_active_transactions tas
    INNER JOIN sys.dm_tran_database_transactions tds
    ON (tas.transaction_id = tds.transaction_id )
    INNER JOIN sys.dm_tran_session_transactions trans
    ON (trans.transaction_id=tas.transaction_id)
    WHERE trans.is_user_transaction = 1 -- user
    AND tas.transaction_state = 2 -- active
    AND tds.database_transaction_begin_time IS NOT NULL
    SELECT
        [s_tst].[session_id],
        [s_es].[login_name] AS [Login Name],
        DB_NAME (s_tdt.database_id) AS [Database],
        [s_tdt].[database_transaction_begin_time] AS [Begin Time],
        [s_tdt].[database_transaction_log_bytes_used] AS [Log Bytes],
        [s_tdt].[database_transaction_log_bytes_reserved] AS [Log Rsvd],
        [s_est].text AS [Last T-SQL Text],
        [s_eqp].[query_plan] AS [Last Plan]
    FROM
        sys.dm_tran_database_transactions [s_tdt]
    JOIN
        sys.dm_tran_session_transactions [s_tst]
    ON
        [s_tst].[transaction_id] = [s_tdt].[transaction_id]
    JOIN
        sys.[dm_exec_sessions] [s_es]
    ON
        [s_es].[session_id] = [s_tst].[session_id]
    JOIN
        sys.dm_exec_connections [s_ec]
    ON
        [s_ec].[session_id] = [s_tst].[session_id]
    LEFT OUTER JOIN
        sys.dm_exec_requests [s_er]
    ON
        [s_er].[session_id] = [s_tst].[session_id]
    CROSS APPLY
        sys.dm_exec_sql_text ([s_ec].[most_recent_sql_handle]) AS [s_est]
    OUTER APPLY
        sys.dm_exec_query_plan ([s_er].[plan_handle]) AS [s_eqp]
    ORDER BY
        [Begin Time] ASC;
    GO
    http://www.sqlskills.com/blogs/paul/script-open-transactions-with-text-and-plans/

  • Big backup file of transaction log

    we run transaction log backup of database A every 15 minutes.
    backup file of transaction log file is usually less than 1 GB.
    All of sudden, backup of transaction log file is more than 20GB continuously.
    What could happen? Need to fix because no disk space for continuous 20+GB transaction log backup.

    we run transaction log backup of database A every 15 minutes.
    backup file of transaction log file is usually less than 1 GB.
    All of sudden, backup of transaction log file is more than 20GB continuously.
    What could happen? Need to fix because no disk space for continuous 20+GB transaction log backup.
    You need to analyse why your T-log has started growing all of sudden, You probably have a long running transaction running (Index maintenance or Big batch delete or update).
    Also analyse the out put of below query
    SELECT log_reuse_wait_desc FROM SYS.databases WHERE name = 'DB_NAME' -- replacce your database name
    you can also refer below article for deeper analysis.
    A transaction log grows unexpectedly or becomes full in SQL Server
    Please mark solved if I've answered your question, vote for it as helpful to help other users find a solution quicker
    Praveen Dsa | MCITP - Database Administrator 2008 |
    My Blog | My Page

  • Transaction log usage grows due to replication even if I don't use replication at all

    Hi
    The transaction log usage keeps growing a lot on my user database since few days back. the database is in full recovery model and I do transaction log backups every 10 minutes. The DB was part of Database Mirroring but I removed it. The usage was controlled
    for many years by the backups but something happened that is messing up the transaction log
    this is DBCC OPENTRAN
    Transaction information for database 'MyDB'.
    Replicated Transaction Information:
            Oldest distributed LSN     : (0:0:0)
            Oldest non-distributed LSN : (1450911:6823:1)
    DBCC execution completed. If DBCC printed error messages, contact your system administrator.
    log_reuse_wait_desc reports REPLICATION
    the funny thing is that I am not using replication at all. I am using CDC.
    To reduce the transaction log usage I run below statement every day since the problem started
    EXEC sp_repldone @xactid = NULL, @xact_segno = NULL, 
        @numtrans = 0, @time = 0, @reset = 1
    Any idea what should I do to solve this problem and be back on a normal situation?
    BTW, The server is SQL 2012 (11.0.2383)
    Thanks
    Javier Villegas |
    @javier_vill | http://sql-javier-villegas.blogspot.com/
    Please click "Propose As Answer" if a post solves your problem or "Vote As Helpful" if a post has been useful to you

    CDC uses the replication log reader agent and if you manually ran sp_repldone like that you lost information in your CDC capture.  If the capture job can't keep up with the workload or is not running for CDC, you would have the exact problems you describe.
     If you execute sp_repldone like that, you might as well disable CDC.
    http://technet.microsoft.com/en-us/library/dd266396(v=sql.100).aspx
    Jonathan Kehayias | Principal Consultant | MCM: SQL Server 2008
    My Blog |
    Twitter |
    MVP Profile
    Training |
    Consulting |
    Become a SQLskills Insider
    Troubleshooting SQL Server

  • Transactional log file grow

    hello,
     how we will solve the problem in this scenario,
    1) I configured mirroring 
    2)my mirror database is full & not available to take any more transactions from principle
    3) so, that transaction log is growing up in principle.

    Have you ever done a trx log backup on the principle?
    No, I have not done trx log backup on the principle.
    ok I will take log backup
    then  I can see that mirroring is paused or suspended ,I can resume mirroring .
    what would happen if there is no space in mirror database? 
    whether I need to reconfigure my mirroring by dropping mirror database or else can I add any disk space to mirror database? 

  • Transaction log shipping restore with standby failed: log file corrupted

    Restore transaction log failed and I get this error: for only 04 no of database in same SQL server, renaming are working fine.
    Date                     
    9/10/2014 6:09:27 AM
    Log                        
    Job History (LSRestore_DATA_TPSSYS)
    Step ID                
    1
    Server                  
    DATADR
    Job Name                           
    LSRestore_DATA_TPSSYS
    Step Name                        
    Log shipping restore log job step.
    Duration                             
    00:00:03
    Sql Severity        0
    Sql Message ID 0
    Operator Emailed           
    Operator Net sent          
    Operator Paged               
    Retries Attempted         
    0
    Message
    2014-09-10 06:09:30.37  *** Error: Could not apply log backup file '\\10.227.32.27\LsSecondery\TPSSYS\TPSSYS_20140910003724.trn' to secondary database 'TPSSYS'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred while processing the log for database 'TPSSYS'. 
    If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log.
    An error occurred during recovery, preventing the database 'TPSSYS' (13:0) from restarting. Diagnose the recovery errors and fix them, or restore from a known good backup. If errors are not corrected or expected, contact Technical Support.
    RESTORE LOG is terminating abnormally.
    Processed 0 pages for database 'TPSSYS', file 'TPSSYS' on file 1.
    Processed 1 pages for database 'TPSSYS', file 'TPSSYS_log' on file 1.(.Net SqlClient Data Provider) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  Skipping log backup file '\\10.227.32.27\LsSecondery\TPSSYS\TPSSYS_20140910003724.trn' for secondary database 'TPSSYS' because the file could not be verified.
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred restoring the database access mode.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteScalar requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: An error occurred restoring the database access mode.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteScalar requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  Deleting old log backup files. Primary Database: 'TPSSYS'
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  The restore operation completed with errors. Secondary ID: 'dd25135a-24dd-4642-83d2-424f29e9e04c'
    2014-09-10 06:09:30.37  *** Error: Could not log history/error message.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.37  *** Error: Could not cleanup history.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-10 06:09:30.37  *** Error: ExecuteNonQuery requires an open and available Connection. The connection's current state is closed.(System.Data) ***
    2014-09-10 06:09:30.38  ----- END OF TRANSACTION LOG RESTORE    
    Exit Status: 1 (Error)

    I Have restore the database to new server and check with new log shipping but its give this same error again, If it is network issue i believe issue need to occur on every database in  that server with log shipping configuration
    error :
    Message
    2014-09-12 10:50:03.18    *** Error: Could not apply log backup file 'E:\LsSecondery\EAPDAT\EAPDAT_20140912051511.trn' to secondary database 'EAPDAT'.(Microsoft.SqlServer.Management.LogShipping) ***
    2014-09-12 10:50:03.18    *** Error: An error occurred while processing the log for database 'EAPDAT'.  If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log.
    An error occurred during recovery, preventing the database 'EAPDAT' (8:0) from restarting. Diagnose the recovery errors and fix them, or restore from a known good backup. If errors are not corrected or expected, contact Technical Support.
    RESTORE LOG is terminating abnormally.
    can this happened due to data base or log file corruption, if so how can i check on that to verify the issue
    Its not necessary if the issue is with network it would happen every day IMO it basically happens when load on network is high and you transfer log file which is big in size.
    As per message database engine was not able to restore log backup and said that you must rebuild log because it did not find out log to be consistent. From here it seems log corruption.
    Is it the same log file you restored ? if that is the case since log file was corrupt it would ofcourse give error on wehatever server you restore.
    Can you try creating logshipping on new server by taking fresh full and log backup and see if you get issue there as well. I would also say you to raise case with Microsoft and let them tell what is root cause to this problem
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Articles

  • Exchange 2010 SP3, RU5 - Massive Transaction Log File Generation

    Hey All,
    I am trying to figure out why 1 of our databases is generating 30k Log Files a day! The other one is generating 20K log files a day. The database does not grow in size as the log files are generated, the problem is log file generation.
    I've tried running through some of the various solutions out there, reviewed message tracking logs, rpc client access logs, IIS Logs - all of which show important info, but none of which actually provide the answers.
    I Stopped the following services to see if that would affect the log file generation in any way, and it has not!
    MS Exchange Transport
    Mail Submission
    IIS (Site Stopped in IIS)
    Mailbox Assistants
    Content Indexing Service
    With the above services stopped, I still see dozens (or more) log files generated in under 10 minutes, I also checked mailbox size reports (top 10) and found that several users mailboxes were generating item count increases for one user of
    about 300, size increases for one user of about 150Mb (over the whole day).
    I am not sure what else to check here? Any ideas?
    Thanks,
    Robert
    Robert

    Hmm - this sounds like an device is chewing up the logs.
    If you use log parser studio, are there any stand out devices in terms of the # of hits?
    And for the ExMon was that logged over a period of time?  The default 60 second window normally misses a lof of stuff.  Just curious!
    Cheers,
    Rhoderick
    Microsoft Senior Exchange PFE
    Blog:
    http://blogs.technet.com/rmilne 
    Twitter:   LinkedIn:
      Facebook:
      XING:
    Note: Posts are provided “AS IS” without warranty of any kind, either expressed or implied, including but not limited to the implied warranties of merchantability and/or fitness for a particular purpose.
    Rhoerick, 
    Thanks for the response. When checking the logs the highest number of hits were the (Source) Load Balancers, Port 25 VIP. The problem i was experience was the following: 
    1) I kept expecting the log file generation to drop to an acceptable rate of 10~20 MB Per Minute (Max). We have a large environment and use the exchange sevrers as the mail relays for the hated Nagios monitoring environment
    2) We didn't have our enterprise monitoring system watching SMTP traffic, this is  being resolved. 
    3) I needed to look closer at the SMTP transport database counters, logs, log files and focus less on the database log generation, i did do some of that but not enough of that. 
    4) My troubleshooting kept getting thrown off due to the monitoring notifications seeming to be sent out in batches (or something similar) stopping the transport service for 10 ~ 15 minutes several times seemed to finally "stop the transactions logs
    from growing at a psychotic rate". 
    5) I am re-running my data captures now that i have told the "Nagios Team" to quit killing the exchange servers, with their notifications, sometimes as much as 100+ of the same notifications for the same servers, issues. so far at a quick glance
    the log file generation seems to have dropped by about 30%. 
    Question: What would be the best counters to review in order to "Put it all together"? Also note: our Server roles are split, MBX and CAS/HT. 
    Robert 
    Robert

  • Backup and restore full and transaction log in nonrecovery mode failed due to LSN

    In SQL 2012 SP1 enterprise, when taking a full backup and followed up a transaction log backup immediately, the transaction log backup starts with an earlier LSN than the ending LSN of the full backup. As a result, I cannot restore
    the transaction log backup after the full backup both as nonrecovery on another machine. I was trying to make the two machine in sync for mirroring purpose. An example is as follows.
    full backup:       first 1121000022679500037, last 1121000022681200001
    transaction log: first 1121000022679000001, last 1121000022682000001
    --- SQL Scripts used  
    BACKUP DATABASE xxx  TO DISK = xxx WITH FORMAT
    go
    backup log  xxx to disk = xxx
    --- When restore, I tried the
    restore log BarraOneArchive  from disk=xxx  WITH STOPATMARK  = 'lsn:1121000022682000001', NORECOVERY
    Also tried StopBeforeMark, did not work either. Complained about the LSN too early to apply to the database

    I think that what I am saying is correct .I said in sync mirroring ( i was not talking about witness) if network goes for few minutes or some longer time may be 20 mins ( more than that is reare scenario IS team has backup for that) logs on Principal will
    continue to grow as they wont be able to commit because there connection with mirror is gone so commit from mirror is not coming.After network comes online Mirror will replay all logs and will soon try to come up with principal
    Books Online says this: This is achieved by waiting to commit a transaction on the principal database, until the principal server receives a message from the mirror server stating that it has hardened the transaction's log to disk. That is,
    if the remote server would go away in a way so that the primary does not notice, transactions would not commit and the primary would also be stalled.
    In practice it does not work that way. When a timeout expires, the principal will consider the mirror to be gone, and Books Online says about this case
    If the mirror server instance goes down, the principal server instance is unaffected and runs exposed (that is without mirroring the data). In this section, BOL does not discussion transaction logs, but it appear reasonable that the log records are
    retained so that the mirror can resync once it is back.
    In Async Mirroring Transaction log is sent to Mirror but it does not waits for Acknowledgement from mirror and commits the transaction.
    But I would expect that the principal still gets acknowledgement that the log records have been consumed, or else your mirroring could start failing f you backup the log too frequently. That is, I would not expect any major difference between sync and async
    mirroring in this regard. (Where it matters is when you fail over. With async mirroring, you are prepared to accept some data loss in case of a failover.)
    These are theories that could be fairly easily tested if you have a mirroring environment set up in a lab, but I don't.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Transaction Logs in SQL Server

    Hi, the BW system has the following properties:
    BW 3.1C Patch 14
    BASIS/ABA 6.20 Patch 38
    BI_CONT 310 Patch 2
    PI_BASIS Patch 2004_1_620
    Windows 2000 Service Pack 4
    SQL Server 2000 SP3 version 8.00.760
    Database used space: 52 GB
    Database free space: 8.9 GB
    Transaction log space: 8 GB
    I am having the following problem.  The SQL transaction logs on the SQL Server fill up very rapidly while aggregates are rolling up.  Sometimes taking up to 16-20 GB of transaction log space in the SQL Server.  We only have 8 GB of space available for the transaction logs.  When the aggregates are not rolling up, the logs do not fill up at all.  I have tried changing the logs to Simple logging, but all that does is delay the fill, and at that point you cannot backup simple logs to free up DB space.
    What is it about aggregates that fills up the transaction log?  Anybody know a solution to this without adding disk space to the transaction log disk?
    Thanks,

    Hello,
    the log backup on simple mode is not necessary. The full database after switching back to full is a must.
    Please keep in mind, that even running on simple mode the log can be filled up, as all transactions are still written to the log. Commited transaction then can truncated from the log. But when you run a hugh transaction like a client copy, the log might grow as well. The log will be freed once the transaction commits or rolls back. And no, you can't split a client copy in several transactions.
    Best regards
      Clas

  • Big transaction log file

    Hi,
    I found a sql server database with a transaction log file of 65 GB.
    The database is configured with the recovery model option = full.
    Also, I noticed than since the database exist, they only took database backup.
    No transaction log backup were executed.
    Now, the "65 GB transaction log file" use more than 70% of the disk space.
    Which scenario do you recommend?
    1- Backup the database, backup the transaction log to a new disk, shrink the transaction log file, schedule transaction log backup each hour.
    2- Backup the database, put the recovery model option= simple, shrink the transaction log file, Backup the database.
    Does the " 65 GB file shrink" operation would have impact on my database users ?
    The sql server version is 2008 sp2 (10.0.4000)
    regards
    D

    I've read the other posts and I'm at the position of: It really doesn't matter.
    You've not needed point in time restore abilities inclusive of this date and time since inception. Since a full database backup contains all of the log needed to bring the database into a consistent state, doing a full backup and then log backup is redundant
    and just taking up space.
    For the fastest option I would personally do the following:
    1. Take a full database backup
    2. Set the database recovery model to Simple
    3. Manually issue two checkpoints for good measure or check to make sure the current VLF(active) is near the beginning of the log file
    4. Shrink the log using the truncate option to lop off the end of the log
    5. Manually re-size the log based on usage needed
    6. Set the recovery model to full
    7. Take a differential database backup to bridge the log gap
    The total time that will take is really just the full database backup and the expanding of the log file. The shrink should be close to instantaneous since you're just truncating the end and the differential backup should be fairly quick as well. If you don't
    need the full recovery model, leave it in simple and reset the log size (through multiple grows if needed) and take a new full backup for safe keeping.
    Sean Gallardy | Blog |
    Twitter

  • Very high transaction log file growth

    Hello
    Running Exchange 2010 sp2 in a two node DAG configuration.  Just recently i have noticed a very high transaction log file growth for one database. The transaction logs are growing so quickly that i have had to turn on circular logging in order to prevent
    the log lun from filling up and causing the database to dismount. I have tried several things to try find out what is causing this issue. At first i thought this could be happening because of virus, an Active Sync user, a users outlook client, or our salesforce
    integration, howerver when i used exmon, i could not see any unusual high user activity, also when i looked at the itemcount for all mailboxes in the particular database that is experiencing the high transaction log file growth, i could not see any mailboxes
    with unusual high item count, below is the command i ran to determine this, i ran this command sever times. I also looked at the message tracking log files, and again could see no indication of a message loop or unusual high message traffic for a
    particlar day. I also followed this guide hopping that it would allow me to see inside the transaction log files, but it didnt produce anything that would help me understand the cause of this issue. When i ran the below tool againts the transaction log files,
    i saw DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD, or OOOOOOOOOOOOOOOOOOOOOOOOOOOOOOO, or HHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHH.
    I am starting to run out of ideas on how to figure out what is causing the log file build up. Any help is greatly appreciated.
    http://blogs.msdn.com/b/scottos/archive/2007/07/12/rough-and-tough-guide-to-identifying-patterns-in-ese-transaction-log-files.aspx
    Get-Mailbox -database databasethatkeepsgrowing | Get-MailboxStatistics | Sort-Object ItemCount -descending |Select-Object DisplayName,ItemCount,@{name="MailboxSize";exp={$_.totalitemsize}} -first 10 | Convertto-Html | out-File c:\temp\report.htm
    Bulls on Parade

    If you have users with iPhones or Smart Phones using ActiveSync then one of the quickest ways to see if this is the issue is to have users shot those phones off to see if the problem is resolved.  If it is one or more iPhones then perhaps look at
    what IOS they are on and get them to update to the latest version or adjust the ActiveSync connection timeout.  NOTE: There was an issue where iPhones caused runaway transactions logs and I believe it was resolved with IOS 4.0.1
    There was also a problem with the MS CRM client awhile back so if you are using that check out this link.
    http://social.microsoft.com/Forums/en/crm/thread/6fba6c7f-c514-4e4e-8a2d-7e754b647014
    I would also deploy some tracking methods to see if you can hone in on the culprits, i.e. If you want to see if the problem is coming from an internal Device/Machine you can use one of the following
    MS USER MONITOR:
    http://www.microsoft.com/downloads/en/details.aspx?FamilyId=9A49C22E-E0C7-4B7C-ACEF-729D48AF7BC9&displaylang=en and here is a link on how to use it
    http://www.msexchange.org/tutorials/Microsoft-Exchange-Server-User-Monitor.html
    And this is a great article as well
    http://blogs.msdn.com/b/scottos/archive/2007/07/12/rough-and-tough-guide-to-identifying-patterns-in-ese-transaction-log-files.aspx
    Also check out ExMon since you can use it to confirm which mailbox is unusually active , and then take the appropriate action.
     http://www.microsoft.com/downloads/en/details.aspx?FamilyId=9A49C22E-E0C7-4B7C-ACEF-729D48AF7BC9&displaylang=en
    Troy Werelius
    www.Lucid8.com
    Search, Recover, & Extract Mailboxes, Folders, & Email Items from Offline EDB's and Live Exchange Servers with Lucid8's DigiScope

  • Read transaction log

    hello,
    I have seen that a user is using too much ressources on my exchange server 2010 (used Exmon to get the user)
    the point is, i do not know whats going on. I see on each log file on the server that is name is coming a lot of time, but the file is crypted (file like E0000000A.log).
    Do you know any tools or something to help me out to get what is going on on this mailbox ?
    So much traffic and packet like 4000 per minutes...
    thx for your help.

    Jero,
    Parse the transaction logs using log parser.
    http://blogs.msdn.com/b/scottos/archive/2007/11/07/remix-using-powershell-to-parse-ese-transaction-logs.aspx?wa=wsignin1.0
    http://hermannmaurer.blogspot.in/2012/05/exchange-2010-transaction-logfiles-grow.html
    Regards,
    ASP20
    Hello
    I have tryied, log parser 2.2, but no result as i dont know the log type to put. Any more information or tutorials ?
    Also i tried the Get-StoreUsageStatistics -Database DB | sort logrecordbytes -descending | select -first 10 | select displayname, timeinserver, digestcategory,logrecordcount,logrecordbytes,@{n='logrecordMb';e={[math]::round($_.logrecordbytes/1MB,3)}},SampleTime,ServerName,DatabaseName
    | ft -AutoSize, return me the user i have targeted with exmon.exe, but still no info about whats wrong.
    get this return (where User X is the hammering user on the server)
    DisplayName                           TimeInServer DigestCategory LogRecordCount LogRecordBytes logrecordMb SampleTim
    Boîte aux lettres - User X                   32793 LogBytes                58186       58771332      56,049 16/05/20.
    Boîte aux lettres - User X                   33430 LogBytes                55670       56481469      53,865 16/05/20.
    Boîte aux lettres - User X                   30057 LogBytes                54078       54983694      52,437 16/05/20.
    Boîte aux lettres - User X                   33104 LogBytes                54137       54338453      51,821 16/05/20.
    Boîte aux lettres - User X                   28259 LogBytes                49456       50254831      47,927 16/05/20.
    Boîte aux lettres - User X                   23761 LogBytes                38417       39139380      37,326 16/05/20.
    Boîte aux lettres - User HHHHH               20137 LogBytes                19993        7451912       7,107 16/05/20.
    Boîte aux lettres - User X                    3982 TimeInServer             6701        6879487       6,561 16/05/20.
    Boîte aux lettres - User X                    4134 TimeInServer             6760        6846904        6,53 16/05/20.
    Boîte aux lettres - User BB                   7139 LogBytes                 5013        6787055       6,473 16/05/20.
    User X is using an Ipad, Iphone and Outlook 2013 to access to mail / calendar.

  • SQL0964C  The transaction log for the database is full

    Hi,
    i am planning to do QAS refresh from PRD system using Client export\import method. i have done export in PRD and the same has been moved to QAS then started imported.
    DB Size:160gb
    DB:DB2 9.7
    os: windows 2008.
    I facing SQL0964C  The transaction log for the database is full issue while client import and regarding this i have raised incident to SAP then they replied to increase some parameter like(LOGPRIMARY,LOGSECOND,LOGFILSIZ) temporarily and revert them back after the import. Based on that i have increased from as below mentioned calculation.
    the filesystem size of /db2/<SID>/log_dir should be greater than LOGFILSIZ*4*(Sum of LOGPRIMARY+LOGSECONDARY) KB
    From:
    Log file size (4KB)                         (LOGFILSIZ) = 60000
    Number of primary log files                (LOGPRIMARY) = 50
    Number of secondary log files               (LOGSECOND) = 100
    Total drive space required: 33GB
    To:
    Log file size (4KB)                         (LOGFILSIZ) = 70000
    Number of primary log files                (LOGPRIMARY) = 60
    Number of secondary log files               (LOGSECOND) = 120
    Total drive space required: 47GB
    But still facing the same issue. Please help me to resolve the ASAP.
    Last error TP log details:
    3 ETW674Xstart import of "R3TRTABUFAGLFLEX08" ...
    4 ETW000   1 entry from FAGLFLEX08 (210) deleted.
    4 ETW000   1 entry for FAGLFLEX08 inserted (210*).
    4 ETW675 end import of "R3TRTABUFAGLFLEX08".
    3 ETW674Xstart import of "R3TRTABUFAGLFLEXA" ...
    4 ETW000  [     dev trc,00000]  Fri Jun 27 02:20:21 2014                                             -774509399  65811.628079
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] CON = 0 (BEGIN)                    85  65811.628164
    4 ETW000  [     dev trc,00000]  &+     DbSlModifyDB6( SQLExecute ): [IBM][CLI Driver][DB2/NT64] SQL0964C  The transaction log for the database is full. 
    4 ETW000                                                                                                  83  65811.628247
    4 ETW000  [     dev trc,00000]  &+     SQLSTATE=57011 row=1                                                                                             
    4 ETW000                                                                                                  51  65811.628298
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  67  65811.628365
    4 ETW000  [     dev trc,00000]  &+     DELETE FROM "FAGLFLEXA" WHERE "RCLNT" = ?                                                                        
    4 ETW000                                                                                                  62  65811.628427
    4 ETW000  [     dev trc,00000]  &+       cursor type=NO_HOLD, isolation=UR, cc_release=YES, optlevel=5, degree=1, op_type=8, reopt=0                    
    4 ETW000                                                                                                  58  65811.628485
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  53  65811.628538
    4 ETW000  [     dev trc,00000]  &+     Input SQLDA:                                                                                                     
    4 ETW000                                                                                                  52  65811.628590
    4 ETW000  [     dev trc,00000]  &+                        1 CT=WCHAR           T=VARCHAR         L=6     P=9     S=0   
    4 ETW000                                                                                                  49  65811.628639
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  50  65811.628689
    4 ETW000  [     dev trc,00000]  &+     Input data:                                                                                                      
    4 ETW000                                                                                                  49  65811.628738
    4 ETW000  [     dev trc,00000]  &+     row 1:             1 WCHAR           I=6       "210"               34  65811.628772
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  51  65811.628823
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  50  65811.628873
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] (END)                              27  65811.628900
    4 ETW000  [    dbtran  ,00000]  ***LOG BY4=>sql error -964   performing DEL on table FAGLFLEXA                    
    4 ETW000                                                                                                3428  65811.632328
    4 ETW000  [    dbtran  ,00000]  ***LOG BY0=>SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1
    4 ETW000                                                                                                  46  65811.632374
    4 ETW000  [     dev trc,00000]  dbtran ERROR LOG (hdl_dbsl_error): DbSl 'DEL'                             59  65811.632433
    4 ETW000                         RSLT: {dbsl=99, tran=1}
    4 ETW000                         FHDR: {tab='FAGLFLEXA', fcode=194, mode=2, bpb=0, dbcnt=0, crsr=0,
    4 ETW000                                hold=0, keep=0, xfer=0, pkg=0, upto=0, init:b=0,
    4 ETW000                                init:p=0000000000000000, init:#=0, wa:p=0X00000000020290C0, wa:#=10000}
    4 ETW000  [     dev trc,00000]  dbtran ERROR LOG (hdl_dbsl_error): DbSl 'DEL'                            126  65811.632559
    4 ETW000                         STMT: {stmt:#=0, bndfld:#=1, prop=0, distinct=0,
    4 ETW000                                fld:#=0, alias:p=0000000000000000, fupd:#=0, tab:#=1, where:#=2,
    4 ETW000                                groupby:#=0, having:#=0, order:#=0, primary=0, hint:#=0}
    4 ETW000                         CRSR: {tab='', id=0, hold=0, prop=0, max.in@0=1, fae:blk=0,
    4 ETW000                                con:id=0, con:vndr=7, val=2,
    4 ETW000                                key:#=3, xfer=0, xin:#=0, row:#=0, upto=0, wa:p=0X00000001421A3000}
    2EETW125 SQL error "-964" during "-964" access: "SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1"
    4 ETW690 COMMIT "14208" "-1"
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] CON = 0 (BEGIN)                 16208  65811.648767
    4 ETW000  [     dev trc,00000]  &+     DbSlModifyDB6( SQLExecute ): [IBM][CLI Driver][DB2/NT64] SQL0964C  The transaction log for the database is full. 
    4 ETW000                                                                                                  75  65811.648842
    4 ETW000  [     dev trc,00000]  &+     SQLSTATE=57011 row=1                                                                                             
    4 ETW000                                                                                                  52  65811.648894
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  51  65811.648945
    4 ETW000  [     dev trc,00000]  &+     INSERT INTO DDLOG (SYSTEMID, TIMESTAMP, NBLENGTH, NOTEBOOK) VALUES (  ? , CHAR(   CURRENT TIMESTAMP - CURRENT TIME
    4 ETW000                                                                                                  50  65811.648995
    4 ETW000  [     dev trc,00000]  &+     ZONE   ), ?, ? )                                                                                                 
    4 ETW000                                                                                                  49  65811.649044
    4 ETW000  [     dev trc,00000]  &+       cursor type=NO_HOLD, isolation=UR, cc_release=YES, optlevel=5, degree=1, op_type=15, reopt=0                   
    4 ETW000                                                                                                  55  65811.649099
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  49  65811.649148
    4 ETW000  [     dev trc,00000]  &+     Input SQLDA:                                                                                                     
    4 ETW000                                                                                                  50  65811.649198
    4 ETW000  [     dev trc,00000]  &+                        1 CT=WCHAR           T=VARCHAR         L=44    P=66    S=0   
    4 ETW000                                                                                                  47  65811.649245
    4 ETW000  [     dev trc,00000]  &+                        2 CT=SHORT           T=SMALLINT        L=2     P=2     S=0   
    4 ETW000                                                                                                  48  65811.649293
    4 ETW000  [     dev trc,00000]  &+                        3 CT=BINARY          T=VARBINARY       L=32000 P=32000 S=0   
    4 ETW000                                                                                                  47  65811.649340
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  50  65811.649390
    4 ETW000  [     dev trc,00000]  &+     Input data:                                                                                                      
    4 ETW000                                                                                                  49  65811.649439
    4 ETW000  [     dev trc,00000]  &+     row 1:             1 WCHAR           I=14      "R3trans"           32  65811.649471
    4 ETW000  [     dev trc,00000]  &+                        2 SHORT           I=2        12744              32  65811.649503
    4 ETW000  [     dev trc,00000]  &+                        3 BINARY          I=12744   00600306003200300030003900300033003300310031003300320036003400390000...
    4 ETW000                                                                                                  64  65811.649567
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  52  65811.649619
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  51  65811.649670
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] (END)                              28  65811.649698
    4 ETW000  [    dbsyntsp,00000]  ***LOG BY4=>sql error -964   performing SEL on table DDLOG                36  65811.649734
    4 ETW000  [    dbsyntsp,00000]  ***LOG BY0=>SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1
    4 ETW000                                                                                                  46  65811.649780
    4 ETW000  [    dbsync  ,00000]  ***LOG BZY=>unexpected return code 2          calling ins_ddlog           37  65811.649817
    4 ETW000  [     dev trc,00000]  db_syflush (TRUE) failed                                                  26  65811.649843
    4 ETW000  [     dev trc,00000]  db_con_commit received error 1024 in before-commit action, returning 8
    4 ETW000                                                                                                  57  65811.649900
    4 ETW000  [    dbeh.c  ,00000]  *** ERROR => missing return code handler                               1974  65811.651874
    4 ETW000                         caller does not handle code 1024 from dblink#5[321]
    4 ETW000                         ==> calling sap_dext to abort transaction
    2EETW000 sap_dext called with msgnr "900":
    2EETW125 SQL error "-964" during "-964" access: "SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1"
    1 ETP154 MAIN IMPORT
    1 ETP110 end date and time   : "20140627022021"
    1 ETP111 exit code           : "12"
    1 ETP199 ######################################
    Regards,
    Rajesh

    Hi Babu,
    I believe you should have taken a restart of your system if log primary are changed.  If so, then increase log primary to 120 and secondary to 80 provide size and space are enough.
    Note 1293475 - DB6: Transaction Log Full
    Note 1308895 - DB6: File System for Transaction Log is Full
    Note: 495297 - DB6: Monitoring transaction log    
    Regards,
    Divyanshu

  • About transaction logs

    Can you tell me about transaction log space? how does it gets full? How is it related to performance?

    Hi,
    Monitoring the SAP Log Disk
    Use
    The size of the transaction log must be checked regularly to work out how much free space is available on the log disk. There should always be enough free space to allow the next file extension. When the SAP system has been installed the autogrow increment is set. At least the size of this increment should be available on the log disk to permit the next file extension. If less space is available and the transaction log file fills up, the SAP system will come to a standstill.
    Ideally, the transaction log should never be filled to more than 60-70%. If the transaction log regularly exceeds this level between 2 transaction log backups, the transaction log must be saved at more frequent time intervals.
    The size of the log can be assessed on the basis of information given for completed backups in the SAP transaction for Backup and Restore Information.
    Procedure:
           1.      To access the transaction for Backup Restore Information  choose CCMS ® DB Administration ® Backup logs.
    Alternatively, enter the transaction code DB12.
    The initial screen of the monitor CCMS Monitoring Tool u2013 DB12 (Backup Restore Information) appears. 
           2.      Choose Backup history and then Logs Backup.
           3.      A result list appears. Find the largest transaction log backup of the past week. Select a row and then History info to find out the number of pages that were processed during the backup. To work out the amount of space used in the transaction log, multiply the number of dumped pages by 8 KB. You can then work out how much free space is left on the transaction log disk.
    If you use a RAID1 disk system exclusively for the SAP transaction log and create hourly log backups, you will rarely encounter space problems. The SAP log file is initially created with a size of 1 GB. The smallest disk normally has 9 GB space and the log file can therefore grow to 9 GB.
    Hope it Helps
    Srini

Maybe you are looking for

  • How can I set a default for webpages to expand to full screen?

    How can I set a default for webpages to expand to full screen? Every webpage I go to does not expand to a full screen which causes me to have to ctrl+++ each time. How can I set my preference to expand to full screen automatically? Thanks.

  • Video Sync between iTunes and iPad with IOS 7.0.2 Never Completes

    Hi, I travel a lot and load recorded content up to my iPad from iTunes on my MAC on a weekly basis. Everything worked fine until I upgraded my iPad-2 to IOS 7.0.2 and upgraded iTunes to 11.1. Now when I sync the iPad from iTunes, the sync hangs. Afte

  • Importing duplicates question

    When I connect my iPhone and Import photos, somethimes I am asked if I want to import duplicates as it seems some photos are already on my Mac probably via iCloud. But if I say Do not import duplicates, does that mean that when I reach 1,000 photos i

  • Indesign CS4 stalls when exporting to PDF

    Trying to export a file from InDesign CS4 to PDF. It's always worked fine before but one file just won't work. Program seems to stall or take so long I think it's crashed. Is there a progress bar that tells you what the program is actually doing? I r

  • OWB 9.2.0.8 and Oracle Database 10g - Urgent

    hi I have OWB 9.2.0.8 software and my oracle database version is 10.2.0.2.0 hosted on sun solaris 5.9 version, can the OWB support this database, pls let me know your suggestions and where I can get a support document Kishan