Applying transaction log corrupted database?

Hi experts,
Symptoms:
  I built a standby system this Monday and I completed a full DB restore without errors. I used log shipping to apply logs but an error occurred: SQL Server detected a logical consistency-based I/O error: incorrect pageid (expected 20:5746350; actual 0:0). It occurred during a read of page (20:5746350) in database ID 5 at offset 0x00000af5d5c000 in file 'l:\TCPDATA4\TCPDATA19.ndf'.
  I run dbcc page and found page(20:5746350) did corrupt on standby system. However, I ran dbcc page on production and it is ok. I restored this full db to another standby and ran dbcc page and it is ok. Also, Checkalloc, Checkcatalog, Checktable on production are ok.
Questions:
  Does checktable detect secondary indexes corruption? I am re-compressing COSP indexes and running full checkdb on production right now to find out what's going on in our DB. How do I fix this issue? Any information will be really appreciated.
Error message:
2011-08-23 17:26:00.19 spid52      The log shipping secondary database TCCDBM1.TCP has restore threshold of 45 minutes and is out of sync. No restore was performed for 1 minutes. Restored latency is 720 minutes. Check agent log and logshipping monitor information.
2011-08-23 17:26:39.62 Backup      Log was restored. Database: TCP, creation date(time): 2011/07/03(05:36:16), first LSN: 184193:4075900:1, last LSN: 184193:7288500:1, number of dump devices: 1, device information: (FILE=1, TYPE=DISK: {'w:\LogB\TCP_20110822213010.trn'}). This is an informational message. No user action is required.
2011-08-23 17:28:00.88 spid52      Error: 14421, Severity: 16, State: 1.
2011-08-23 17:28:00.88 spid52      The log shipping secondary database TCCDBM1.TCP has restore threshold of 45 minutes and is out of sync. No restore was performed for 2 minutes. Restored latency is 716 minutes. Check agent log and logshipping monitor information.
2011-08-23 17:28:34.03 Backup      Log was restored. Database: TCP, creation date(time): 2011/07/03(05:36:16), first LSN: 184193:7288500:1, last LSN: 184193:10880343:1, number of dump devices: 1, device information: (FILE=1, TYPE=DISK: {'w:\LogB\TCP_20110822213511.trn'}). This is an informational message. No user action is required.
2011-08-23 17:28:53.06 spid65      Error: 824, Severity: 16, State: 2.
2011-08-23 17:28:53.06 spid65      SQL Server detected a logical consistency-based I/O error: incorrect pageid (expected 20:5746350; actual 0:0). It occurred during a read of page (20:5746350) in database ID 5 at offset 0x00000af5d5c000 in file 'l:\TCPDATA4\TCPDATA19.ndf'.  Additional messages in the SQL Server error log or system event log may provide more detail. This is a severe error condition that threatens database integrity and must be corrected immediately. Complete a full database consistency check (DBCC CHECKDB). This error can be caused by many factors; for more information, see SQL Server Books Online.
2011-08-23 17:30:00.54 spid52      Error: 14421, Severity: 16, State: 1.
2011-08-23 17:30:00.54 spid52      The log shipping secondary database TCCDBM1.TCP has restore threshold of 45 minutes and is out of sync. No restore was performed for 2 minutes. Restored latency is 713 minutes. Check agent log and logshipping monitor information.
2011-08-23 17:30:35.94 spid53      Error: 824, Severity: 16, State: 2.

Hi Dennis!
I understand your reasonable doubt, but think that physical corruptions are 99,999...% caused by a hardware failure e.g. a defective RAM module, or a related, low-level failure as e.g. an outdated NIC driver, etc.
Then, in this case it is clear that you have a platform problem only on your standby system, and not in the one running productively right now.
Maybe you already checked your platform, but you should take into account that in most cases just reading the logs is not enough; stress tests on all of the components are usually necessary. Most of the time, when we analyze hardware related problems, the problem is only reproduced when special tests are performed. Furthermore it is possible that more than one component is weak or broken.
My recommendation is that you check the components in this order:
   - disks (complete surface) (while SAP and SQL Server are shut down)
   - disk controller
   - Memory (stress test)
   - CPU
   - all other components
Consider that restoring a backup and running in a damaged platform would imply further problems. You would be waiting until the next problem arises and that would imply a system halt or corruptions, maybe even with data loss.
Once the hardware checked and -if necessary- fixed, you can proceed as explained in SAP notes 142731 and 1420452. Of course you can also restore in a different hardware, but remember checking the platform as it is also exposed to hardware failures.
Cheers!!
--Jesú

Similar Messages

  • The transaction log for database 'BizTalkMsgBoxDb' is full.

    Hi All,
    We are getting the following error continously in the event viewer of our UAT servers. I checked the jobs and all the backup jobs were failing on the step to backup the transaction log file and were giving the same error. Our DBA's cleaned the message box manually and backed up the DB but still after some time the jobs starts failing and this error is logged in the event viewer.
    The transaction log for database 'BizTalkMsgBoxDb' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases".
    Thanks,
    Abdul Rafay
    http://abdulrafaysbiztalk.wordpress.com/
    Please mark this answer if it helps

    Putting the database into simple recovery mode and shrinking the log file isn't going to help: it'll just grow again, it will probably fragment across the disk thereby impacting performance and, eventually, it will fill up again for the same reason
    as before.  Plus you put yourself in a very vulnerable position for disaster recovery if you change the recovery mode of the database: and that's before we've addressed the distributed transaction aspect of the BizTalkDatabases.
    First, make sure you're backing up the log file using the BizTalk job Backup BizTalk Server (BizTalkMgmtDb).  It might be that the log hasn't been backed up and is full of transactions: and, eventually, it will run out of space.  Configuration
    instructions at this link:
    http://msdn.microsoft.com/en-us/library/aa546765(v=bts.70).aspx  Your DBA needs to get the backup job running properly rather than panicking!
    If this is running properly, and backing up (which was the case for me) and the log file is still full, run the following query:
    SELECT Name, log_reuse_wait_desc
    FROM sys.databases
    This will tell you why the log file isn't properly clearing down and why it cannot use the space inside.  When I had this issue, it was due to an active transaction.
    I checked for open transactions on the server using this query:
    SELECT
    s_tst.[session_id],
    s_es
    .[login_name]
    AS [Login Name],
    DB_NAME
    (s_tdt.database_id)
    AS [Database],
    s_tdt
    .[database_transaction_begin_time]
    AS [Begin Time],
    s_tdt
    .[database_transaction_log_record_count]
    AS [Log Records],
    s_tdt
    .[database_transaction_log_bytes_used]
    AS [Log Bytes],
    s_tdt
    .[database_transaction_log_bytes_reserved]
    AS [Log Rsvd],
    s_est
    .[text]
    AS [Last T-SQL Text],
    s_eqp
    .[query_plan]
    AS [Last Plan]
    FROM
    sys.dm_tran_database_transactions
    s_tdt
    JOIN
    sys.dm_tran_session_transactions
    s_tst
    ON s_tst.[transaction_id]
    = s_tdt.[transaction_id]
    JOIN
    sys.[dm_exec_sessions]
    s_es
    ON s_es.[session_id]
    = s_tst.[session_id]
    JOIN
    sys.dm_exec_connections
    s_ec
    ON s_ec.[session_id]
    = s_tst.[session_id]
    LEFT
    OUTER
    JOIN
    sys.dm_exec_requests
    s_er
    ON s_er.[session_id]
    = s_tst.[session_id]
    CROSS
    APPLY
    sys.dm_exec_sql_text
    (s_ec.[most_recent_sql_handle])
    AS s_est
    OUTER
    APPLY
    sys.dm_exec_query_plan
    (s_er.[plan_handle])
    AS s_eqp
    ORDER
    BY [Begin Time]
    ASC;
    GO
    And this told me the spid of the process with an open transaction on BizTalkMsgBoxDB (in my case, this was something that had been open for several days).  I killed the transaction using KILL spid, where spid is an integer.  Then I ran the BizTalk
    Database Backup job again, and the log file backed up and cleared properly.
    Incidentally, just putting the database into simple transaction mode would have emptied the log file: giving it lots of space to fill up again.  But it doesn't deal with the root cause: why the backups were failing in the first place.

  • The transaction log for database 'ECC' is full + ECC6.0 Installation Failur

    Guyz,
    my ecc6 installation failed after 8 hours run with following error log snippet...
    exec sp_bindefault 'numc3_default','SOMG.MSGNO'
    DbSlExecute: rc = 99
      (SQL error 9002)
      error message returned by DbSl:
    The transaction log for database 'ECC' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases
    (DB) ERROR: DDL statement failed
    (ALTER TABLE [SOMG] ADD CONSTRAINT [SOMG~0] PRIMARY KEY CLUSTERED ( [MANDT], [OBJTP], [OBJYR], [OBJNO] ) )
    DbSlExecute: rc = 99
      (SQL error 4902)
      error message returned by DbSl:
    Cannot find the object "SOMG" because it does not exist or you do not have permissions.
    ECCLOG1 data file has got 25GB initial size and growth was restricted to 10% (PROPOSED BY SAPInst)...
    i'm assuming this error was due to lack of growth space for ECCLOG1 datafile...am i right? if so how much should i allocate memory for this log ? or is there any workaround ?
    thanks in advance

    Kasu,
    If SQL is complaining that the log file is full then the phase of the install that creates the SQL data/log files has already occurred (happens early in the install) and the install is importing programs, config and data into the db.
    Look at the windows application event log for "Transaction log full" events to confirm.
    To continue, in SQL Query analyzer try:
    "Backup log [dbname] with truncate_only"
    This will remove only inactive parts of the log and is safe when you don't require point-in-time recovery (which you don't during an install).
    Then, go to the SQL Enterprise manager, choose the db in question and choose the shrink database function, choose to shrink only the transaction log file and the space made empty by the truncate will be removed from the file.
    Change the recovery mode in SQL Server to "simple" so that the log file does not grow for the remainder of the install.
    Make sure you change the recovery mode back to "full" after the install is complete.
    Your transaction log appears to have filled the disk partition you have assigned to it.
    25GB is huge for a transaction log and you would normally not see them grow this large if you are doing regular scheduled tlog backups (say every 30-60 minutes) because the log will truncate every time, but its not unusual to see one get big during an install, upgrade or when applying hotpacks.
    Tim

  • The transaction log for database 'Test_db' is full due to 'LOG_BACKUP'

    My dear All,
    Came up with another issue:
    App team is pushing the data from one Prod1 server 'test_1db' to another Prod2 server 'User_db' through a job, here while pushing the data after some duration job is failing and throwing the following error
    'Error: 9002, Severity: 17, State: 2.'The transaction log for database 'User_db' is full due to 'LOG_BACKUP'''.
    On Prod2 server 'User_db' log is having enough space 400gb on drive and growth is 250mb. I really confused that why job is failing as there is lot of space available. Kindly guide me to troubleshoot the issue as this issue is occuring from more than
    1 week. Kindly refer the screenshot for the same.
    Environment: SQL Server 2012 with sp1 Ent-edition. and log backup duration is every 15 mints and there is no High availability between the servers.
    Note: Changing to simple recovery model may resolve but App team is required to run in Full recovery model as they need of log backups.
    Thanks in advance,
    Nagesh
    Nagesh

    Dear V,
    Thanks for the susggestions.
    I have followed some steps to resolve the issue, as of now my jobs are working without issue.
    Steps:
    Generating log backup for every 5 minutes
    Increased the growth 500mb to unrestricted. 
    Once whole job completed we are shrinking the log file.
    Nagesh

  • The transaction log for database 'SharePoint_Config' is full

    Hi all ,
    I am very new to sharepoint. when i tried to remove a wsp file from central administration i got the message like
    The transaction log for database 'SharePoint_Config' is full. To find out why space in the log cannot
    be reused, see the log_reuse_wait_desc column in sys.databases.
    Can anybody help me to solve this please. I saw one solution in net like this but i don't know how to do this. Can anybody help me how to do these steps please.
    1.       
    Take the configuration database offline and detach it
    2.       
    Copy the current MDF to a new location (to be used as a way of recovering the database if needed)
    3.       
    Put the database back on-line, reattach it and then put it within simple mode (from full), with an aim of this stopping the database from
    increasing in size
    4.       
    Shrink the database and recover log space
    5.       
    Should the shrinking fail, we'd look at detaching the database, making a sideways copy of the log file to another database
    6.       
    We would then reattach the database, which should generate a new log file
    Thank you 

    Hi Soumya,
    I don't have any dba resource. And it is not a lab environment. It's Corporate environment.
    Thanks alot for ur quick reply.
    I normally don't just come onto the threads and tell people that they need a consultant, but you might want to look at getting a consultant to help make sure that your SharePoint environment is health and can be restored in the event of a server failure. 
    Based on the problem that you have described it might not be recoverable, or at least it might not be as recoverable as you want it to be.
    At the very least you'll want to watch this video of my session at TechEd 2014 which talks about backups and how to set them up.
    http://channel9.msdn.com/events/TechEd/NorthAmerica/2014/DBI-B214#fbid=
    Thank You,
    Denny Cherry

  • The transaction log for database 'mydatabase' is full. To find out why space in the log cannot be reused, see the log_reuse_wait

    Every time I get this error, at different points of testing inserts and deletions on my table:
    The transaction log for database 'mydatabase' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases
    Why do I keep getting this?  All I'm doing is deleting several hundred thousand records and inserting them into a couple of tables.  i shouldn't have to truncate my log every time or my application bombs out!
    sys.databases only gives me this info for log_reuse_wait_desc which does nothing for me:
    LOG_BACKUP

    sp_helpdb BizTalkDTADb
    ALTER DATABASE BiztalkDTADb
    SET RECOVERY SIMPLE;
    GO
    DBCC SHRINKFILE (BiztalkDTADb_log, 1);
    GO
    sp_helpdb BizTalkDTADb
    GO
    ALTER DATABASE BiztalkDTADb
    SET RECOVERY FULL
    GO

  • "The transaction log for database 'speakasiaonline' is full." I'm getting this message when I m trying to login to speakasia website and am unable to open it. Pl help.

    The transaction log for database 'speakasiaonline' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases

    What does it return?
    SELECT log_reuse_wait_desc FROM  sys.databases WHERE database_id=2
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • Any body able to help me : The transaction log for database 'KDS' is full

    Hi Experts,
                      I am facing the follwoing problem when i entered into the portal.
    [NWMss][SQLServer JDBC Driver][SQLServer]The transaction log for database 'KDS' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases.
    Exception id: 10:51_28/09/07_0006_3439451
    See the details for the exception ID in the log file
    Please give me a solution urgent.
    Message was edited by:
            Ramanan Panchabakesan

    Ameya,
    The log file seems to be....[ i have bolded the warning number and bold+underline the error number].kindly see this
    #1.5 #001320E973F2004900000281000010D400043B2BF18A5CFD#1190959829445#com.sap.engine.services.deploy##com.sap.engine.services.deploy######01eb74206d8811dc8114001320e973f2#SAPEngine_System_Thread[impl:5]_5##0#0#Warning#1#/System/Server#Plain###
    Warning occurred on server <b>3439450</b> during startApp sap.com/cafruntimemonitoringear : Application sap.com/cafruntimemonitoringear has weak reference to application sap.com/com.sap.jdo and is starting it!#
    #1.5 #001320E973F2004900000283000010D400043B2BF18A5DCF#1190959829445#com.sap.engine.services.deploy##com.sap.engine.services.deploy######01eb74206d8811dc8114001320e973f2#SAPEngine_System_Thread[impl:5]_5##0#0#Warning#1#/System/Server#Plain###
    Warning occurred on server <b>3439450</b> during startApp sap.com/cafruntimemonitoringear : Application sap.com/cafruntimemonitoringear has weak reference to application sap.com/caf~km.proxies and is starting it!#
    #1.5 #001320E973F2004900000285000010D400043B2BF18A5E85#1190959829445#com.sap.engine.services.deploy##com.sap.engine.services.deploy######01eb74206d8811dc8114001320e973f2#SAPEngine_System_Thread[impl:5]_5##0#0#Warning#1#/System/Server#Plain###
    Warning occurred on server <b>3439450</b> during startApp sap.com/cafruntimemonitoringear : Application sap.com/cafruntimemonitoringear has weak reference to application sap.com/cafruntimeear and is starting it!#
    #1.5 #001320E973F2004900000287000010D400043B2BF18A5F72#1190959829461#com.sap.engine.services.deploy##com.sap.engine.services.deploy######01eb74206d8811dc8114001320e973f2#SAPEngine_System_Thread[impl:5]_5##0#0#Warning#1#/System/Server#Plain###
    Warning occurred on server <b>3439450</b> during startApp sap.com/cafruntimemonitoringear : Application sap.com/cafruntimemonitoringear has a weak reference to resource jmsfactory/TopicConnectionFactory with type javax.jms.TopicConnectionFactory but the resource is not available and the application may not work correctly!#
    #1.5 #001320E973F2001100000002000010D400043B2BF18FC93A#1190959829805#com.sap.jms##com.sap.jms.LOCK_EXCEPTION######837a4dd06d8911dcbe35001320e973f2#SAPEngine_System_Thread[impl:5]_34##0#0#Warning#1#/System/Server#Java###Couldn't acquire Lock $service.jms_provider. Current JMS lock owner is server node with id = .#1#<i><b><u>3439451</u></b></i>#
    #1.5 #001320E973F200490000028D000010D400043B2BF1A40473#1190959831134#com.sap.engine.services.jndi##com.sap.engine.services.jndi######01eb74206d8811dc8114001320e973f2#SAPEngine_System_Thread[impl:5]_5##0#0#Path##Java###Caught #1#com.sap.engine.services.jndi.persistent.exceptions.JNDIException: Error during s object serialization.
    at com.sap.engine.services.jndi.implclient.ClientContext.serializeObject(ClientContext.java:3335)
    at com.sap.engine.services.jndi.implclient.ClientContext.serializeDirObject(ClientContext.java:3224)
    at com.sap.engine.services.jndi.implclient.ClientContext.rebind(ClientContext.java:1032)
    at com.sap.engine.services.jndi.implclient.ClientContext.rebind(ClientContext.java:957)
    at com.sap.engine.services.servlets_jsp.server.runtime.context.WebApplicationConfig.bind(WebApplicationConfig.java:455)
    at com.sap.engine.services.servlets_jsp.server.runtime.context.WebApplicationConfig.parse(WebApplicationConfig.java:116)
    at com.sap.engine.services.servlets_jsp.server.runtime.context.ApplicationContext.init(ApplicationContext.java:617)
    at com.sap.engine.services.servlets_jsp.server.container.WebContainerHelper.createContext(WebContainerHelper.java:540)
    at com.sap.engine.services.servlets_jsp.server.container.StartAction.prepareStart(StartAction.java:51)
    at com.sap.engine.services.servlets_jsp.server.container.WebContainer.prepareStart(WebContainer.java:475)
    at com.sap.engine.services.deploy.server.application.StartTransaction.prepareCommon(StartTransaction.java:223)
    at com.sap.engine.services.deploy.server.application.StartTransaction.prepareLocal(StartTransaction.java:176)
    at com.sap.engine.services.deploy.server.application.ApplicationTransaction.makeAllPhasesLocal(ApplicationTransaction.java:365)
    at com.sap.engine.services.deploy.server.application.ParallelAdapter.runInTheSameThread(ParallelAdapter.java:132)
    at com.sap.engine.services.deploy.server.application.ParallelAdapter.makeAllPhasesLocalAndWait(ParallelAdapter.java:250)
    at com.sap.engine.services.deploy.server.DeployServiceImpl.startApplicationLocalAndWait(DeployServiceImpl.java:4450)
    at com.sap.engine.services.deploy.server.DeployServiceImpl.startApplicationsInitially(DeployServiceImpl.java:2610)
    at com.sap.engine.services.deploy.server.DeployServiceImpl.clusterElementReady(DeployServiceImpl.java:2464)
    at com.sap.engine.services.deploy.server.ClusterServicesAdapter.containerStarted(ClusterServicesAdapter.java:42)
    at com.sap.engine.core.service630.container.ContainerEventListenerWrapper.processEvent(ContainerEventListenerWrapper.java:144)
    at com.sap.engine.core.service630.container.AdminContainerEventListenerWrapper.processEvent(AdminContainerEventListenerWrapper.java:19)
    at com.sap.engine.core.service630.container.ContainerEventListenerWrapper.run(ContainerEventListenerWrapper.java:102)
    at com.sap.engine.frame.core.thread.Task.run(Task.java:64)
    at com.sap.engine.core.thread.impl5.SingleThread.execute(SingleThread.java:79)
    at com.sap.engine.core.thread.impl5.SingleThread.run(SingleThread.java:150)
    Caused by: com.sap.engine.services.jndi.persistent.exceptions.JNDIException: Error during s object serialization.
    at com.sap.engine.services.jndi.persistent.RemoteSerializator.toByteArray(RemoteSerializator.java:55)
    at com.sap.engine.services.jndi.implclient.ClientContext.serializeObject(ClientContext.java:3332)
    ... 24 more
    Caused by: java.io.NotSerializableException: com.sap.engine.system.ORBProxy
    at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1054)
    at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:278)
    at com.sap.engine.services.jndi.persistent.RemoteSerializator.toByteArray(RemoteSerializator.java:48)
    ... 25 more

  • The transaction log for database 'EMP_SP_SearchApp_CrawlStoreDB_32fdb1522c5249088db8b09c1917dbec' is full due to 'ACTIVE_TRANSACTION'

    Hi
    we have a sharepoint farm for uploading documents in huge number daily and we conifgured RBS for sql server content db.
    we are facing issue in sql server instance:
    Suddenly the SQL server instance restarts and sometimes failover to node1 and sometimes just restart and keep running on the current server
    and did some invistagations we found that 
    -Analysis of data shows that the slowness of IO for tempdb.ldf is cause of the issue
    -tried to check tempdb using db properties but I couldn’t see the properties it gives me error,
    -There might be a big transaction locking the tempdb.
    Also noted  that there are lot of errors:
     The transaction log for database 'MOJSP_SearchApp_CrawlStoreDB_32fdb1522c5249088db8b09c1917dbec' is full due to 'ACTIVE_TRANSACTION'
    This might also be part of the problem causing locking.
    we found the errors like
    The transaction log for database 'EMP_SP_SearchApp_CrawlStoreDB_32fdb1522c5249088db8b09c1917dbec' is full due to 'ACTIVE_TRANSACTION'
    I see that log_reuse_wait_desc (reuse of transaction log space is currently waiting on as of the last checkpoint) is 'ACTIVE_TRANSACTION'
    adil

    Hi Adil,
    Method1:
    On your SharePoint SQL Server database, open the SQL Server Management Studio.
    Connect to the local SQL Server.
    Right click Your Database Name.
    Go to -> Properties -> Options.
    Change Recover Model from Full to Simple -> Click OK.
    Right click Your Database Name, go to Tasks -> Shrink -> Shrink Files.
    Select a File type of Log, then Shrink.
    Method2:
    Run SQL Management Studio and login to your SharePoint instance.
    Click on "New Query".
    Type in the following commands:
    USE [database name];
    BACKUP LOG [Your Database Name] WITH TRUNCATE_ONLY;
    DBCC
    SHRINKFILE ([Your Log File Name], 1);
    Replace [Your Database Name] and [Your Log File Name] with your log file.
    This will truncate your log file to 1mb.
    Method3:
    Issue CHECKPOINT on the tempdb
    USE tempdb
    GO
    CHECKPOINT
    GO
    Take Transactional log backup on the user database 
        BACKUP LOG CustomerXS TO DISK = N'M:\MSSQL\Backup\MSSQLSERVER\XS and RT\XS_Movement.trn'     WITH COMPRESSION     GO

  • The transaction log for database 'tempdb' is full

    Hi,
    When I try to login into SQL SERVER 2008, suddenly below error message is popped out.
    ADDITIONAL INFORMATION:
    The transaction log for database 'tempdb' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases (Microsoft SQL Server, Error: 9002)
    For help, click:
    http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=10.00.4000&EvtSrc=MSSQLServer&EvtID=9002&LinkId=20476
    BUTTONS:
    OK
    I could restart SQL SERVER, but I need to identify cause of it. Any pointer on cause of this issue may help me. Thanks

    What does it return?
    SELECT log_reuse_wait_desc FROM  sys.databases WHERE database_id=2
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • The transaction log for database 'WSS_Content' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases

    Hi all,
    Observing the below error in event logs once in every day at 3:30 AM on the on SQL Server Machine, it is being used for TFS 2010.
    The transaction log for database 'WSS_Content' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases
    Kindly advice me here on what to do?
    Thanks in advance

    First thing to check is, what is happening at 3:30am to cause the log to fill up?  Is there any maintenance tasks being run against the DB, such as index re-organizations or rebuilds?  If so, take a look at what the job is doing, maybe it could
    be tuned (ex:  A Maintenance Plan with a rebuild indexes task will rebuild *every* index whether it needs it or not.  Look at using something that checks how fragmented an index is, and then intelligently determines what, if anything, to do to that
    index.)
    Basically, the error is telling you that SQL is trying to do something in the DB, and has filled up the log (either because it's not getting backed up often enough, or the transaction in question is running *very* long, or your log is set to not auto-grow)
    So, a couple possible solutions:
    Increase the frequency of Transaction Log backups of WSS_Content (only works if the DB is in Full Recovery)
    If you've got the disk space, turn on auto-growth for the TLog (or just manually grow it larger) {This will work for both Full and Simple Recovery}
    It would be nice to know what's causing the space to not be re-used, but with this occuring at 3:30am, you'll either need to set up an Agent job to run around that time and record its results, or manually run something like "select log_reuse_wait_desc from
    sys.databases where name = 'WSS_Content';"  Me, I prefer to be soundly sleeping at 3:30am and would set up an Agent job to run from say about 3:00am to 3:45am once per minute (maybe even more frequently.)
    DON'T shrink the log, though.  SQL will idealy grow the log as large as it needs it (if autogrowth is enabled) and no more.  If you shrink the file, SQL will just have to re-grow it later again, which can affect performance.
    Jason A.

  • 24985,ERR_MENOTFOUND: medium not found error while applying transaction log

    Hi,
    We are in process of creating a standby database for our maxdb system.
    We have backed up & restored the database (OS level) to the standby system and the standby database is up & running in admin state. Now I want to apply the transaction logs to the standby system.  I have copied all the logs from the OS level & transfered to the standby system. Now, with the recover_start command I get the error
    ERR
    -24985,ERR_MENOTFOUND: medium not found
    Will it not work if I copy the logs from OS level???
    Please let me know, how to proceed further
    Thanks,
    Salman Qayyum

    Hi,
    Yes, they are of same version...I have copied the binaries from primary to standby
    Primary
    dbmcli on LCP>db_enum
    OK
    LCP     /sapdb/LCP/db                           7.7.04.38       fast    running
    LCP     /sapdb/LCP/db                           7.7.04.38       quick   offline
    LCP     /sapdb/LCP/db                           7.7.04.38       slow    offline
    LCP     /sapdb/LCP/db                           7.7.04.38       test    offline
    Standby
    dbmcli on LCP>db_enum
    OK
    LCP     /sapdb/LCP/db                           7.7.04.38       fast    running
    LCP     /sapdb/LCP/db                           7.7.04.38       quick   offline
    LCP     /sapdb/LCP/db                           7.7.04.38       slow    offline
    LCP     /sapdb/LCP/db                           7.7.04.38       test    offline
    dbmcli on LCP>
    Thanks,
    Salman

  • Apply archive logs to database

    I have a production database that is online. I'm planning to setup a pc with the same configurations as that of the db server and use that server as the reporting generation. The clone pc's database will be one day delay compared to the production database. I'm planning to use export/import but someone suggested to use the archive log files of the production database and apply it to the clone pc, is this doable?
    I have a full export backup and all the archive log of the production database, is this enough to restore the clone pc to the state( one day back) of the production database?
    Need your suggestions. Thanks

    You can do this while configuring physical standby database.
    go through the link below to know more about physical standby
    http://www.stanford.edu/dept/itss/docs/oracle/10g/server.101/b10823/create_ps.htm
    regards

  • How to apply transaction logs to a standby DB?

    We already moved the logs to the standby server.
    Please advise us how to use the SQL server GUI to apply them?  Thanks!

    In SQL Server Management studio right click the <database>->tasks->restore->database. You need to "Add" full database and then related translog backup. You can also mention a specific time to restore.
    RT

  • Database Transaction log suspected pages

    We migrated our Production Databases to New SQL Cluster and when I run query to find any suspected pages entries in MSDB Database .I found there are 5 entries in msdb.dbo.suspected_pages tables .These enries for Production Database Transaction file (File_id=2)
    , Pages_id =1,2,3,6,7 and the event _type was updated to 4 for all pages after I did DB restore and error_count is 1 for each page_id.
    As my understanding , before I did the DB restore ,there were transaction log pages were corrupted ,but the restored repaired those corrupted pages .Since pages are repaired then there is no need to concern for now .I have now Database consistency check
    job scheduled to check the Database corruption on Report server each night .I restore Database on report server using the a copy of Production Database Backup .Someone please help me to understand what caused the log file pages to get corrupted .Page_id 1,2,3,6,7 
    are called boot pages for the log file  ? What shold I do if I will find the Log file supected Pages ?
    Thank so your help in advance .
    Daizy

    Hi Andreas , Thanks for your reply .
    FYI- You have the event_type 1and 3 for your Database , but the event_type was updated to 4 on my system after I did restore and the date/time shows the exact date/time when the event_type was updated .
    Please help me understand usually Database Data file is organized in pages ,not the log file ??
    Thanks
    Daizy
    Hello Daizy
    yes, the event types 1-3 were the error-state before the "repair".
    After I did a Full backup + Restore I now have type 4 just as you do.
    Yes, the Log files is organized in so called "Virtual Log Files"/VLFs, which have nothing in common with the 8-KB data-pages of the data-files. Therefore a page_id does not make sense there.
    You can read more on the architecture of the Transaction Log here:
    SQL Server Transaction Log Architecture and Management
    This article by Paul Randal might also be of interest to you for:
    Transaction log corruption and backups
    Hope that helps.
    Andreas Wolter (Blog |
    Twitter)
    MCSM: Microsoft Certified Solutions Master Data Platform, MCM, MVP
    www.SarpedonQualityLab.com |
    www.SQL-Server-Master-Class.com

Maybe you are looking for