Truncating Log file - SQL

Hi All,
We need some help in reducing the size of our log files.
Our database has grown to 17gig and our log file is just over 12gig. Our database recovery model is "full".
We are running complete backups once a night, and also doing an hourly transaction log backup. I have tried right clicking the database in EM->all tasks->shrink database and specifically selected the log file and clicked ok, as well as just clicking ok, but there seems to be no effect.
Can anyone outline how to do this? From what I can tell, we are only using around 50Mb of the logfile and the rest is just empty space!
Thanks
Rajiv

Ok, lets split this up into 2 issues to keep things simple:
1. In our present situation (using the FULL recovery model), we are performing a full backup every night. We are also backing up the transaction log every hour during working hours. In this scenario, from what I understand the transaction log file should be truncated to the point of backup. This is not happening on my server, so I am asking for guidance as to the best way to investigate why this is happening, and then of course the best way of truncating the file manually (T-SQL?).
2. In terms of the recovery model, in our business people phone in with orders, payments etc which are all input directly into B1. We physically wouldn't be able to recover back to the previous day/last full backup, as alot of the input data isn't on paper/email. I have done a test restore with backup + logs etc previously - it's not difficult per se, just lo-ong =).
The change in hardware works 2 ways... yes, you're right in that hardware may be more reliable, but the basic business need for recovery in the event of failure remains unchanged (in our specific case). In contrast however, as the hardware gets faster and faster, our 17gig DB + logfile (note, this should be nowhere near 12gig) will not be as incumbent as perhaps one might expect on our file server. Now if we can do full backups semi-transparently using the simple model every hour on newer hardware... that's one less thing to worry about! =)

Similar Messages

  • !!!Urgent !!!!!Cannot recovery database without log files (SQL 2008)

    I only have single mdf file, the transaction log file was corrupted. When I was trying to attatch the mdf file, I received the following error message. Can someone give me a help..... it's really urgent!
    TITLE: Microsoft SQL Server Management Studio
    Attach database failed for Server 'DEV8\sqlexpress2008'.  (Microsoft.SqlServer.Smo)
    For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=10.0.1600.22+((SQL_PreRelease).080709-1414+)&EvtSrc=Microsoft.SqlServer.Management.Smo.ExceptionTemplates.FailedOperationExceptionText&EvtID=Attach+database+Server&LinkId=20476
    ADDITIONAL INFORMATION:
    An exception occurred while executing a Transact-SQL statement or batch. (Microsoft.SqlServer.ConnectionInfo)
    Could not open new database 'Smartadmin'. CREATE DATABASE is aborted.
    File activation failure. The physical file name "C:\MSSQL\Smartadmin_log.ldf" may be incorrect.
    The log cannot be rebuilt because there were open transactions/users when the database was shutdown, no checkpoint occurred to the database, or the database was read-only. This error could occur if the transaction log file was manually deleted or lost due to a hardware or environment failure. (Microsoft SQL Server, Error: 1813)
    For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=10.00.1600&EvtSrc=MSSQLServer&EvtID=1813&LinkId=20476
    BUTTONS:
    OK
    Javafun

    I killed a dev db by accident (moved it between folders while it had an open transaction) and out of curiosity attempted to fix it as though it were a super-urgent client db. The above link fixed it, but I'll cut to the chase since it's a lot simpler than
    most of the article:
    1) Set aside the database mdf and ldf you're trying to recover.
    2) Create a database in the location you wanted, of the same name as the problem DB.
    3) Stop the SQL Server service.
    4) Copy the mdf and ldf files in, overwriting the dummy db you created.
    5) Start the SQL Server service.
    Done. Fixed.
    Surely you're doing something SQL Server does not like, but it is unfortunate you have to shut down the entire service to perform this trick. All of the other methods in that linked article and sp_attach_single_file_db failed, but this worked like a charm.
    In my dev database I was left with a db that was good as new - no missing data no problems. It seems odd this works so well and yet all these other approaches get errors about an open transaction. To be fair though, my dev database is essentially a single-user
    file that isn't heavily used, I assume a production DB on a busy server would be more problematic.

  • Truncating log files

    Are there any server settings that will truncate logs on startup?
    Thanks in advance,
    Josh Cimino
    Software Developer
    Mxi Technologies Ltd.
    1430 Blair Place, Suite 800
    Ottawa, ON, Canada
    K1J 9N2
    E: [email protected]
    T: 613.747.4698 ext 479
    F: 613.747.1909
    www.mxi.com
    "From the flight line to the bottom line"
    [att1.html]

    Exchange 2013 SP1.  I have about 30 databases in a 4 server DAG, with half the DBs on servers 1 and 2 and the other half on 3 and 4.  All copies show healthy.  When I run Get-MailboxDatabase
    –Status | ft name,last* -auto it shows the last full and incremental backups as about a month ago, despite many successful backups since then, according to Arcserve anyway. And, log files are not truncating.  Any ideas?
    Did you apply the Nov 2014 Updates? If so, that may explain the issue as a VSS bug was introduced. Make sure you have installed the latest updates:
    https://social.technet.microsoft.com/Forums/windowsserver/en-US/c78faf9a-e126-4f51-b8fe-69d547a9d365/november-2014-update-rollup-kb3000853-vss-backups-broken?forum=winserver8gen
    Twitter!: Please Note: My Posts are provided “AS IS” without warranty of any kind, either expressed or implied.

  • CPO SQL Log Files

    Hey guys
    We have started to see an exponential growth in our SQL log files that is in direct correlation with running more processes more consistently.
    Has anyone had any issues with running a log truncating script to both the TEO_processlog and TEO_reportinglog files?
    something along the lines of 
    ALTER DATABASE ExampleDB SET RECOVERY SIMPLE
    DBCC SHRINKFILE('ExampleDB_log', 0, TRUNCATEONLY)
    Thanks 
    Matt

    Matt,
     Yes, many people (if they do not need the t-logs) will reduce them. In 3.0 (during install) you can actually set it to SIMPLE mode instead of the more full recovery mode. I do that on most all of my boxes. Prior to CPO 3.0 you would have to do it like you show above.
    On most of my pre 3.0 boxes I would do something like
    USE <TEOProcess_DB_Name>;GO
    -- Truncate the log by changing the database recovery model to SIMPLE.
    ALTER DATABASE <TEOProces_DB_Name> SET RECOVERY SIMPLE;GO
    -- Shrink the truncated log file to 1000MB.
    DBCC SHRINKFILE ("<TEOProcess_log_file_name>", 1000); GO
    Of course this is on SQL only. You can find more information on DBCC Shrinkfile at Microsoft's help site.
    If you need to reset the database to full mode it's:
    -- Reset the database recovery model.
    ALTER DATABASE <TEOProcess_DB_Name> SET RECOVERY FULL;GO
    --Shaun

  • Crystal Report Server Database Log File Growth Out Of Control?

    We are hosting Crystal Report Server 11.5 on Microsoft SQL Server 2005 Enterprise.  Our Crystal Report Server SQL 2005 database file size = 6,272 KB, and the log file that goes with the database has a size = 23,839,552.
    I have been reviewing the Application Logs and this log file size is auto-increasing about 3-times a week.
    We backup the database each night, and run maintenance routines to Check Database Integrity, re-organize index, rebuild index, update statistics, and backup the database.
    Is it "Normal" to have such a large LOG file compared to the DATABASE file?
    Can you tell me if there is a recommended way to SHRINK the log file?
    Some Technical Documents suggest frist truncating the log, and the using the DBCC SHRINKFILE command:
    USE CRS
    GO
    --Truncate the log by changing the database recovery model to SIMPLE
    ALTER DATABASE CRS
    SET RECOVERY SIMPLE;
    --Shrink the truncated log file to 1 gigabyte
    DBCC SHRINKFILE (CRS_log, 1000);
    GO
    --Reset the database recovery model.
    ALTER DATABASE CRS
    SET RECOVERY FULL;
    GO
    Do you think this approach would help?
    Do you think this approach would cause any problems?

    my bad you didn't put the K on the 2nd number.
    Looking at my SQL server that's crazy big my logs are in the k's like 4-8.
    I think someone enabled some type of debugging on your SQL server, it's more of a Microsoft issue, as our product doesn't require it from looking at my SQL DB's
    Regards,
    Tim

  • MessageBox log file size

    Hi, 
    In our prod environment, the MessageBox data file is withing the recommended limits - 2GB, but the log file is 32GB. Is this a reason to worry, or it is normal?  I couldn't find any recommendations on this. 
    Thank you very much!

    This is not normal.
    IMO your BizTalk database Jobs are not running , Make sure your BizTalk SQL servers jobs have been enabled and SQL server agent is running. 
    Please have a look of
    How to Configure the Backup BizTalk Server Job article to enable the jobs. 
    The BizTalk backup job is responsible for keeping the log file size in the limit. 
    you can try shrinking the log file using following SQL command
    USE BiztalkMsgBoxDb;
    GO
    -- Truncate the log by changing the database recovery model to SIMPLE.
    ALTER DATABASE BiztalkMsgBoxDb
    SET RECOVERY SIMPLE;
    GO
    -- Shrink the truncated log file to 1 MB.
    DBCC SHRINKFILE (BiztalkMsgBoxDb_Log, 2);
    GO
    I would recommend you to have a read of following articles
    BizTalk Environment Maintenance from a DBA perspective 
    BizTalk Databases: Survival Guide
    hope this helps. 
    Greetings,HTH
    Naushad Alam
    When you see answers and helpful posts, please click Vote As Helpful, Propose As Answer, and/or
    Mark As Answer
    alamnaushad.wordpress.com

  • How do look if i have Standby Redo Log files

    How do look if i have Standby Redo Log files
    example of creating them..

    To check existence of Standby Redo Log Files:
    SQL> v$standby_log
    Name                                      Null?    Type
    ----------------------------------------- -------- ---------------------------- GROUP#                                             NUMBER
    DBID                                               VARCHAR2(40)
    THREAD#                                            NUMBER
    SEQUENCE#                                          NUMBER
    BYTES                                              NUMBER
    USED                                               NUMBER
    ARCHIVED                                           VARCHAR2(3)
    STATUS                                             VARCHAR2(10)
    FIRST_CHANGE#                                      NUMBER
    FIRST_TIME                                         DATE
    LAST_CHANGE#                                       NUMBER
    LAST_TIME                                          DATE
    SQL> select * from v$standby_log;
    no rows selected
    SQL>To Create a Standby Redo Log file:
    SQL> alter database add standby logfile group 11 ('/u01/app/test.log') size 5m;
    Database altered.
    SQL> set line 10000
    SQL> select * from v$standby_log;
        GROUP# DBID                                        THREAD#  SEQUENCE#      BYTES       USED ARC STATUS     FIRST_CHANGE# FIRST_TIM LAST_CHANGE# LAST_TIME
            11 UNASSIGNED                                        0          0    5242880        512 YES UNASSIGNED              0                      0
    SQL>and this is how you drop:
    SQL> alter database drop standby logfile group 11;
    Database altered.
    SQL> ! rm /u01/app/test.log
    SQL>Asif Momen
    http://momendba.blogspot.com
    Edited by: Asif Momen on Mar 16, 2010 1:32 PM
    Included DROP example

  • Truncate SQL 2008 R2 log files

    How do I truncate a log file.  Running out of space.
    Help is appreciated.
    Thanks
    Dave Kozlowski

    If your db is in Full Recovery:
    ALTER DATABASE
    [DatabaseName] SET
    RECOVERY SIMPLE WITH
    NO_WAIT
    DBCC SHRINKFILE
    ([LogFileName, 1)
    ALTER DATABASE [DatabaseName]
    SET RECOVERY
    FULL WITH NO_WAIT
    GO
    Ayad

  • Microsoft sql server extended event log file

    Dears
    Sorry for my below questions if it is very beginner level.
    In my implementation I have cluster SQL 2012 on Windows 2012; I am using MountPoints since I have many Clustered Disks.
    My MountPoint Size is only 3 GB; My Extended event log are growing fast and it is storing in the MountPoint Drive directly (Path: F:\MSSQL11.MSSQLSERVER\MSSQL\Log).
    What is the best practice to work with it? (is it to keep all Extended events? or recirculate? or to shrink? or to store in DB?)
    Is there any relation between SQL truncate and limiting the size of Extended event logs?
    How can I recirculate this Extended Events?
    How can I change the default path?
    How can I stop it?
    and in case I stop it, does this means to stop storing SQL event in Windows event Viewer?
    Thank you

    After a lot of checking, I have found below:
    My Case:
    I am having SQL Failover Cluster Instances "FCI" and I am using Mount-Points to store my Instances.
    I am having 2 Passive Copies for each FCI.
    In my configuration I choose to store the Root Instance which include the logs on Mount-Point.
    My Mount Point is 2 GB Only, which became full after few days of deployment.
    Light Technical Information:
    The Extended Event Logs files are generated Coz I have FCI, in single SQL Installation you will not find this files.
    The File Maximum size will be 100 MB.
    The Files start circulating after it become 10 Full Files.
    If you have the FCI installed as 1 Active 2 Passive, and you are doing failover between the nodes, then you will expect to see around 14 - 30 copy of this file.
    Based on above information you will need to have around 100 MB * 10 Files Per Instance copy * 3 Since in my case I have 1 Active and 2 passive instances which will = 3000 MB
    So in my case My Mount-Point was 2 GB, which become full coz of this SQLDIAG Logs.
    Solution:
    I extended my mount point by 3 GB coz I am storing this logs on it.
    In case you will need to change SQLDIAG Extended Logs Size to 50 MB for example and place to F:\Logs, then you will need below commands:
    ALTER SERVER CONFIGURATION SET DIAGNOSTICS LOG OFF;
    ALTER SERVER CONFIGURATION
    SET DIAGNOSTICS LOG MAX_SIZE = 50 MB;
    ALTER SERVER CONFIGURATION
    SET DIAGNOSTICS LOG PATH = 'F:\logs';
    ALTER SERVER CONFIGURATION SET DIAGNOSTICS LOG ON;
    After that you will need to restart the FCI from SQL Server Configuration Manager or Failover Cluster Manager.
    I wish you will find this information helpful if it is your case.
    Regards

  • Sql server 2005 log file shrinking issue.

    hi,
    one of my production server database  log file size is increasing continuously,we can take log backup but no use,
    we can change the database recovery model from  full to simple we shrink the log file, log file shrink successfully but space is not released,and we are getting the alerts continuously your physical disk is 95% like that.
    here what we can do how reduced the log file size and how to control the log file size.
    please give me the answer  ASAP.
    Thanks,

    Hello,
    In SQL Server, log truncation occurs automatically as follows:
    • Under the simple recovery model, after a checkpoint.
    • Under the full recovery model or bulk-logged recovery model, after a log backup, if a checkpoint has occurred since the previous backup.
    As Dan posted earlier, when we run the above code, we will get an integer value. The log_reuse_wait_desc value tells us why the transaction log space is not reused. You can refer to the article below to check the root cause. 
    http://msdn.microsoft.com/en-us/library/ms345414.aspx
    Regards,
    Alisa Tang
    Alisa Tang
    TechNet Community Support

  • SQL Loader Inserting Log File Statistics to a table

    Hello.
    I'm contemplating how to approach gathering the statistics from the SQL Loader log file to insert them into a table. I've approached this from a Korn Shell Script perspective previously, but now that I'm working in a Windows environment and my peers aren't keen about batch files and scripting I thought I'd attempt to use SQL Loader itself to read the log file and insert one or more records into a table that tracks data uploads. Has anyone created a control file that accomplishes this?
    My current environment:
    Windows 2003 Server
    SQL*Loader: Release 10.2.0.1.0
    Thanks,
    Luke

    Hello.
    Learned a little about inserting into multiple tables with delimited records. Here is my current tested control file:
    LOAD DATA
    APPEND
    INTO TABLE upload_log
    WHEN (1:12) = 'SQL*Loader: '
    FIELDS TERMINATED BY WHITESPACE
    TRAILING NULLCOLS
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , filler_field_3  FILLER
    , filler_field_4  FILLER
    , filler_field_5  FILLER
    , day_of_week
    , month
    , day_of_month
    , time_of_day
    , year
    , log_started_on          "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
    INTO TABLE upload_log
    WHEN (1:11) = 'Data File: '
    FIELDS TERMINATED BY ':'
    (  upload_log_id    RECNUM
    , filler_field_0   FILLER  POSITION(1)
    , input_file_name          "TRIM(:input_file_name)"
    INTO TABLE upload_log
    WHEN (1:6) = 'Table '
    FIELDS TERMINATED BY WHITESPACE
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , table_name              "RTRIM(:table_name, ',')"
    INTO TABLE upload_rejects
    WHEN (1:7) = 'Record '
    FIELDS TERMINATED BY ':'
    (  upload_rejects_id  RECNUM
    , record_number      POSITION(1)  "TO_NUMBER(SUBSTR(:record_number,8,20))"
    , reason
    INTO TABLE upload_rejects
    WHEN (1:4) = 'ORA-'
    FIELDS TERMINATED BY ':'
    (  upload_rejects_id  RECNUM
    , error_code         POSITION(1)
    , error_desc
    INTO TABLE upload_log
    WHEN (1:22) = 'Total logical records '
    FIELDS TERMINATED BY WHITESPACE
    (  upload_log_id      RECNUM
    , filler_field_0     FILLER  POSITION(1)
    , filler_field_1     FILLER
    , filler_field_2     FILLER
    , action                     "RTRIM(:action, ':')"
    , number_of_records
    INTO TABLE upload_log
    WHEN (1:13) = 'Run began on '
    FIELDS TERMINATED BY WHITESPACE
    TRAILING NULLCOLS
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , day_of_week
    , month
    , day_of_month
    , time_of_day
    , year
    , run_began_on            "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
    INTO TABLE upload_log
    WHEN (1:13) = 'Run ended on '
    FIELDS TERMINATED BY WHITESPACE
    TRAILING NULLCOLS
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , day_of_week
    , month
    , day_of_month
    , time_of_day
    , year
    , run_ended_on            "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
    INTO TABLE upload_log
    WHEN (1:18) = 'Elapsed time was: '
    FIELDS TERMINATED BY ':'
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , elapsed_time
    INTO TABLE upload_log
    WHEN (1:14) = 'CPU time was: '
    FIELDS TERMINATED BY ':'
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , cpu_time
    )Here are the basic table create scripts:
    TRUNCATE TABLE upload_log;
    DROP TABLE upload_log;
    CREATE TABLE upload_log
    (  upload_log_id      INTEGER
    , day_of_week        VARCHAR2(  3)
    , month              VARCHAR2(  3)
    , day_of_month       INTEGER
    , time_of_day        VARCHAR2(  8)
    , year               INTEGER
    , log_started_on     DATE
    , input_file_name    VARCHAR2(255)
    , table_name         VARCHAR2( 30)
    , action             VARCHAR2( 10)
    , number_of_records  INTEGER
    , run_began_on       DATE
    , run_ended_on       DATE
    , elapsed_time       VARCHAR2(  8)
    , cpu_time           VARCHAR2(  8)
    TRUNCATE TABLE upload_rejects;
    DROP TABLE upload_rejects;
    CREATE TABLE upload_rejects
    (  upload_rejects_id  INTEGER
    , record_number      INTEGER
    , reason             VARCHAR2(255)
    , error_code         VARCHAR2(  9)
    , error_desc         VARCHAR2(255)
    );Now, if I could only insert a single record to the upload_log table (per table logged); adding separate columns for skipped, read, rejected, discarded quantities. Any advice on how to use SQL Loader to do this (writing a procedure would be fairly simple, but I'd like to perform all of the work in one place if at all possible)?
    Thanks,
    Luke
    Edited by: Luke Mackey on Nov 12, 2009 4:28 PM

  • Logical sql in log file.

    Can someone please tell me how to see the complete sql query in the log file. If I run the same query the sql is not being produced I looked in the server log file and also manage sessions log file. It just says all columns from 'Subject Area'. I want to see all the joins and filters as well. Even for repeated queries how can I see complete sql. I set my logging level to 2.

    http://lmgtfy.com/?q=obiee+disable+query+caching
    http://catb.org/esr/faqs/smart-questions.html#homework

  • How to configure log files in SQL 2012?

    I'm installing a SharePoint Solution and using Microsoft SQL 2012 and I have limited knowledge in installing SQL.  I simply run the wizard and create the DB for SharePoint. 
    Is there any links/materials available demonstrating how to correctly configure the log files step by step in a specific partition during the SQL installation for SharePoint 2013?
    thanks, 

    Hello,
    SQL Server database log files benefit from RAID 1 and RAID 10 configurations. RAID 10 is recommended for both data files and log files.
    To control de grow of transaction log files, please backup them regularly. The following article may explain you in detail why:
    http://technet.microsoft.com/en-us/library/ms175495.aspx
    Monitor the growth of log files using Performance Monitor and the SQLServer:Databases --> Log Growths counter. Adjust the size of log files until Log Growths
    is constantly zero.
    Hope this helps.
    Regards,
    Alberto Morillo
    SQLCoffee.com

  • Moving the log file of a publisher database SQL Server 2008

    There are many threads on this here. Most of them not at all helpful and some of them wrong. Thus a fresh post.
    This post regards SQL Server 2008 (10.0.5841)
    The PUBLISHER database primary log file which is currently of 3 blocks and not extendable,
    must be moved as the LUN is going away.
    The database has several TB of data and a large number of push transactional replications as well as a couple of bi-directional replications.
    While the primary log file is active, it is almost never (if ever) used due to its small fixed size.
    We are in the 20,000 TPS range at peak (according to perfmon). This is a non-trivial installation.
    This means that
    backup/restore is not even a remotely viable option (it never is in the real world)
    downtime minimization is critical - measured in minutes or less.
    dismantling and recreating the replications is doable, but I have to say, I have zero trust in the script writer to generate accurate scripts. Many of these replications were originally set up in older versions of SQL Server and have come along for the
    ride as upgrades have occurred. I consider scripting everything and dismantling the whole lot pretty high risk. In any case, I do not want to have to reinitialize any replications as this takes, effectively, an eternity.
    Possible solution:
    The only option I can think of is to wind down everything, such that there are zero outstanding uncommitted transactions and detach the database, delete the offending log file and reattach using the CREATE DATABASE xyz ATTACH_REBUILD_LOG option.
    This should, if I have understood things correctly, cause SQL Server to recreate the default log file in the same directory as the .mdf file. I am not sure what will happen to the secondary log file which is not moving anywhere at this point.
    The hard bit is insuring that every transaction in the active log files have been replicated before shutdown. This is probably doable. I do not know how to manually flush any left over transactions to replication. I expect if I shut down all "real"
    activity and wait for a certain amount of time, eventually all the replications will show "No replicated transactions are available" and then I would be good to go.
    Hillary, if you happen to be there, comments appreciated.

    Hi Philip
    you should try this long back suggested way of stopping replication and restore db and rename or detach attach
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/6731803b-3efa-4820-a303-4ffb7edf154a/detaching-a-replicated-database?forum=sqlreplication
    Thanks
    Saurabh Sinha
    http://saurabhsinhainblogs.blogspot.in/
    Please click the Mark as answer button and vote as helpful
    if this reply solves your problem
    I do not wish to be rude, but which part of the OP didn't you understand?
    Specifically the bit about 20,000 transactions a second and database size of several TB. Do you have any concept whatsoever of what this means? I will answer for you, "no, you are clueless" as your answer clearly shows.
    Stop wasting bandwidth by proposing pointless and wrong solutions which indicate that you did not read the OP, or do you just do this to generate points?
    Also, you clearly failed to notice that I was on the thread to which you referred, and I had some pointed comments to make. This thread was an attempt to garner some input for an alternative proposal.

  • Unable to view SQL Request in Log files

    Hi Folks,
    I am facing an issue which I am unable to find out the solution to view the physical query generated in log files in Presentation Services.
    Below is the SQL Request generated but I want to view the exact physical query i.e SQL Request which is hitting DB.
    So please guiude me to resolve this issue, I guess it is because of Initialization blocks created which is blocking to view the SQL request.
    -------------------- SQL Request:
    set variable LOGLEVEL = 7;SELECT "- Policy Effective-Start Date"."Start Quarter" saw_0, "- Insurance Policy Facts".Revenue saw_1, "- Insurance Policy Facts"."# Insurance Policies" saw_2, "Insurance Policy".Status saw_3, "Insurance Policy".Type saw_4 FROM "Insurance Policies" WHERE ("Insurance Policy".Type = 'Policy') AND ("- Policy Effective-Start Date"."Start Julian Day Number" BETWEEN VALUEOF(CURRENT_JULIAN_DAY)-365 AND VALUEOF("CURRENT_JULIAN_DAY")) ORDER BY saw_0, saw_3, saw_4
    /* QUERY_SRC_CD='rawSQL' */
    Regards
    Dj

    There is no Enterprise Edition of SSMS. There is SSMS Basic and SSMS Complete. Prior to 2012 sp1, only SSMS Basic were available with Express Edition - but as of 2012 sp1 Expredd also offers SSMS Complete. SSMS Complete is selected bu default when you install
    SSMS (unless you are prior to 2012 sp1 and are using Express, of course).
    However, even SSMS Basic *should* show Agent assuming you have permissions for that. This is hearsay, but from trusted sources. Here is what to do:
    Check what is installed for the machine from where you are running SSMS. You can do that using SQL Server Installation Center - see this blog post: http://sqlblog.com/blogs/tibor_karaszi/archive/2011/02/10/what-does-this-express-edition-look-like-anyhow.aspx
     (towards the end).
    On that machine try both this problematic account as well as an account which is sysadmin. Does the sysadmin account see Agent? If so, you know permissions aren't granted properly. If not, then you know the tool is the problme.
    Also try the problematic account from a machine where you know you see Agent normally. Again, this will help you assess whether the problem is the tool (SSMS) or permissions for the account.
    Tibor Karaszi, SQL Server MVP |
    web | blog

Maybe you are looking for

  • Sharing an external HD on a Powermac Network

    Hi ! I'm working in a graphic & motion design agency in Paris, and we would like to share on our network an external hard drive containing our picture, sound and video libraries, so that anyone can access it. Can anyone help us do that ? Do we need i

  • Use Japaness words as search condition, but can't retrive data

    Hi,   I encountered this problem. In report, I use client name as a condition. User can choose a client to retrieve data.    but when user choose a Japanese client name(or other non-English language), can't get data.    I use the sql genereated by bo

  • Automate install...

    Hi all, I am trying to build an Automator workflow to install (actually to copy) an application to the Application folder of a computer, and copy a link to a website onto the desktop folder of the remote computer also. For some reason I can only get

  • RfcAFWException: Service: BS_ECC not found in CPACache

    Hi everybody, I have the follow error: - Error: 2007-07-31 08:53:39 GMT-04:00: com.sap.aii.rfc.afcommunication.RfcAFWException: Service: BS_ECC not found in CPACache Does anybody know what is happening? Thanks, Renato.

  • Availability of the 1 gb N200 in the

    Creative, Alternati'vely to the Zen Micro I could be interested in the gb version of the N200, but it is listed as 'not available yet' on every retailer but is available for backorder. Any idea when this device (any color) will be available? And can