EWA via solman for SQL Server

Hi,
I'm trying to configure EWA via solman for an ECC 6.0 system on MS SQL 2005.
I've completed all the necessary steps i.e. all the RFC's are working fine, RTCCTOLL is ok, the report is getting generated in SDCCN on the satellite system...however it doesnt seem to get transfered to solman.
RSCOLL00 gives a weird "database system not supported " message although it shows finished successfully.
Is there some other program for SQL Server or maybe something that I'm missing...
Pls. help.
Thanks a lot,
Saba.

Hi Saba,
you can't schedule EWA for Solution manager sessions in SDCCN of the satellite. The only kind you can schedule in the satellite is the kind you send directly to SAP, that's why SDCC_OSS appears as the RFC connetion. You have to fetch EWA for Solution Manager sessions from the solution manager using a refresh sessions task. Does anything happen when you perform a refresh sessions task in SDCCN on the satellite (when you pick the RFC to Solution Manager as the destination)?
You should also check that the service definitions are current in both satellite and solution manager systems. Note 727998 explains how to delete and replace them. (Often the only way of being totally sure that the service defintions are correct.)
best regards,
-David.
Edited by: David Murphy  on Nov 20, 2008 4:26 PM

Similar Messages

  • After installed SP1 for SQL Server 2012, can no longer export to csv

    After installing SP1 today via Windows Update, I am no longer able to export data to csv using the SQL Server Import and Export wizard. I get the following error message:
    "Column information for the source and the destination data could not be retrieved, or the data types of source columns were not mapped correctly to those available on the destination provider."
    "Column "col1": Source data type "200" was not found in the data type mapping file."...
    (The above line repeats for each column)
    The work-around I have to do is to manually map each column in the "Edit Mappings..." option from the "Configure Flat File Destination" page of the wizard. This is an extreme inconvenience to have to have to edit the mappings and change
    each column to "string [DT_STR]" type from "byte stream [DT_BYTES]" type each time I want to export to csv. I did not have to do this before installing SP1; it worked perfectly for months with hundreds of exports prior to this update and
    no need to modify mapping.

    I am running Windows 7 64-bit, SQL Server 2012 Express edition. Again, just yesterday from Windows Update, I installed SQL Server 2012 Service Pack 1 (KB2674319), followed by Update Rollup for SQL Server 2012 Service Pack 1 (KB2793634). This situation was
    not occurring before these updates were installed, and I noticed it immediately after they were installed (and of course I restarted my computer after the updates).
    In SSMS I just now created a test DB and table to provide a step-by-step with screenshots.
    Here is the code I ran to create the test DB and table:
    CREATE DATABASE testDB;
    GO
    USE testDB;
    GO
    CREATE TABLE testTable
    id int,
    lname varchar(50),
    fname varchar(50),
    address varchar(50),
    city varchar(50),
    state char(2),
    dob date
    GO
    INSERT INTO testTable VALUES
    (1,'Smith','Bob','123 Main St.','Los Angeles','CA','20080212'),
    (2,'Doe','John','555 Rainbow Ln.','Chicago','IL','19580530'),
    (3,'Jones','Jane','999 Somewhere Pl.','Washington','DC','19651201'),
    (4,'Jackson','George','111 Hello Cir.','Dallas','TX','20010718');
    GO
    SELECT * FROM testTable;
    Results look good:
    id    lname    fname    address    city    state    dob
    1    Smith    Bob    123 Main St.    Los Angeles    CA    2008-02-12
    2    Doe    John    555 Rainbow Ln.    Chicago    IL    1958-05-30
    3    Jones    Jane    999 Somewhere Pl.    Washington    DC    1965-12-01
    4    Jackson    George    111 Hello Cir.    Dallas    TX    2001-07-18
    In Object Explorer, I right-click on the [testDB] database, choose "Tasks", then "Export Data..." and the SQL Server Import and Export Wizard appears. I click Next to leave all settings as-is on the "Choose a Data Source" page, then on the "Choose a Destination"
    page, under the "Destination" drop-down I choose "Flat File Destination" then browse to the desktop and name the file "table_export.csv" then click Next. On the "Specify Table Copy or Query" page I choose "Write a query to specify the data to transfer" then
    click Next. I type the following SQL statement:
    SELECT * FROM testTable;
    When clicking the "Parse" button I get the message "This SQL statement is valid."
    On to the next page, "Configure Flat File Destination" I try leaving the defaults then click Next. This is where I am getting the error message (see screenshot below):
    Then going to the "Edit Mappings..." option on the "Configure Flat File Destination" page, I see that all columns which were defined as varchar in the table are showing as type "byte stream [DT_BYTES]", size "0", the state column which is defined as char(2)
    shows correctly however with type "string [DT_STR]", size "2" (see screenshow below):
    So what I have to do is change the type for the lname, fname, address and city columns to "string [DT_STR]", then I am able to proceed with the export successfully. Again, this just started happening after installing these updates. As you can imagine, this
    is very frustrating, as I do a lot of exports from many tables, with a lot more columns than this test table.
    Thanks for your help.

  • Will Cumulative Update 12 for SQL Server 2008R2 SP2 fix: Warning: Unable to Verify TimeStamp for Path\ProgramName?

    Hello,
    I had a server crash recently and our outsource hosting tech support suggested applying Cumulative Upade 12 for SQL server 2008R2 SP2 to fix the issue.  The exception from our dump file "Warning: Unable to Verify TimeStamp for Path\ProgramName"
    is not in the list of hotfixes for this CU.  Do you know if this will fix this issue?  The CU warns not to apply the CU if your issue id not addressed by the CU.  Here is a portion of the dump file with the relevant error:
    This dump file has an exception of interest stored in it.
    The stored exception information can be accessed via .ecxr.
    (7e8.2ab4): Unknown exception - code 000042ac (first/second chance not available)
    ntdll!NtWaitForSingleObject+0xa:
    00000000`777412fa c3              ret
    0:240> .sympath srv*c:\Websymbols*http://msdl.microsoft.com/download/symbols;
    Symbol search path is: srv*c:\Websymbols*http://msdl.microsoft.com/download/symbols
    Expanded Symbol search path is: srv*c:\websymbols*http://msdl.microsoft.com/download/symbols
    0:240> .reload /f
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\sqlservr.exe, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlservr.exe
    ..........Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\sqlos.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlos.dll
    ...............Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\opends60.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for opends60.dll
    .......Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\BatchParser.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for BatchParser.dll
    ....Unable to load image C:\Program Files\Microsoft SQL Server\100\Shared\instapi10.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for instapi10.dll
    *** ERROR: Module load completed but symbols could not be loaded for instapi10.dll
    ..Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\Resources\1033\sqlevn70.rll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlevn70.rll
    *** ERROR: Module load completed but symbols could not be loaded for sqlevn70.rll
    Press ctrl-c (cdb, kd, ntsd) or ctrl-break (windbg) to abort symbol loads that take too long.
    Run !sym noisy before .reload to track down problems loading symbols.
    .................Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\ftimport.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for ftimport.dll
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\msfte.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for msfte.dll
    *** ERROR: Module load completed but symbols could not be loaded for msfte.dll
    ...........Unable to load image C:\Windows\System32\sqlncli10.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlncli10.dll
    ...Unable to load image C:\Windows\System32\1033\SQLNCLIR10.RLL, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for SQLNCLIR10.RLL
    *** ERROR: Module load completed but symbols could not be loaded for SQLNCLIR10.RLL
    ..Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\xpsqlbot.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for xpsqlbot.dll
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\XPStar.DLL, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for XPStar.DLL
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\sqlscm.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlscm.dll
    ...*** ERROR: Module load completed but symbols could not be loaded for odbcint.dll
    ...Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\Resources\1033\XPStar.RLL, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for XPStar.RLL
    Thank you,
    Steve -

    The error message appears to be related to a debug session and not come from the actual crash.
    So all I know is that you hade a server crash. I don't even know exactly what that means. Did Windows bite the dust? Or was it only SQL Server?
    Assuming the latter, I would expect the SQL Server errorlog to have some information (and that would be ERRORLOG.1 or earlier, since the server have been restarted), but if SQL Server died the output may be incomplete.
    There may be also dump files, but as I rarely look into these, I am not sure how to interpret them. But I am quite confident that "Unable to verify TimeStamp..." is not the reason SQL Server went down.
    I would suggest the following course:
    *  If the server is not critical, do nothing. As long as it has onlyl happened once, it has only happened once.
    *  If the server is critical, open a case with Microsoft if you are not able to figure out the reason yourself. The key here is "Unknown exception - code 000042ac".
    *  If it happens again, you should absolutely open a case. An important thing here is whether the stack dump is identical or something different. If the stack dumps are identical, you may have hit a bug in SQL Server or the OS, and applying CUs or OS
    fixes could help if it is a known issue. If the stack dump is something else, you have ghosts in the machine - that is, bad hardware.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • OGG for SQL Server - Extract stops capturing - Bug?

    Hi, all,
    I've found a problem with OGG for SQL Server where the Extract stops capturing data after the transaction log is backed up. I've looked for ways to reconfigure OGG to avoid the problem but couldn't find any reference to options to workaround this problem. It seems to be a bug to me.
    My Extract configuration is as follows:
    EXTRACT ext1
    SOURCEDB mssql1
    TRANLOGOPTIONS NOMANAGESECONDARYTRUNCATIONPOINT
    EOFDELAY 60
    EXTTRAIL dirdat/e1
    TABLE dbo.TestTable;
    I'm using the EOFDELAY parameter for testing purposes only, since it's easy to reproduce the scenario that causes the issue when the extract polling is configured with longer intervals.
    When the Transaction Log backup runs, SQL Server marks all the virtual logs that are older than the primary and secondary truncation points as inactive (status = 0). These virtual logs can then be reused if required. They still contain change records, though, and OGG can read from then if required, before they are overwritten. This situation will never occur if we are not using SQL Replication and have the Extract configured with the parameter MANAGESECONDARYTRUNCATIONPOINT.
    However, I'm trying to simulate a scenario where OGG is used along SQL Replication and the extract is configured with the NOMANAGESECONDARYTRUNCATIONPOINT option. The situation that I've reproduced and caused the Extract to stop capturing is the follow sequence of events:
    1. Extract reads transaction log and capture change up to LSN X
    2. More change are made to the database and the LSN is incremented
    3. Log Reader reads Transaction Log, captures changes up to LSN X+Y and advances the secondary truncation point to that LSN
    4. A transaction log occurs, backs up all the active virtual logs, advances the primary truncation point to a LSN greater than LSN X+Y, and marks all the virtual logs with LSNs <= X+Y as inactive (status = 0)
    5. Changes continue to happen in the database consuming all the available inactive virtual logs and overwriting them.
    6. The extract wakes up again to capture more changes.
    At this point, the changes between LSNs X and X+Y are not in the Transaction Log anymore, but are available in the backups. From what I understood in the documentation the Extract should detect that situation and retrieve the changes from the Transaction Log backups. This, however, is not happening and the Extract becomes stuck. It still pools the transaction log at the configured interval query the log state with DBCC LOGINFO, but doesn't move forward anymore.
    If I stop and restart the Extract I can see from the trace that it does the right thing upon startup. It realises that it requires information that's missing from the logs, query MSDB for the available backups, and mine the backups to get the required LSNs.
    I would've thought the Extract should do the same during normal operation, without the need for a restart.
    Is this a bug or the normal operation of the Extract? Is there a way to configure it to avoid this situation without using NOMANAGESECONDARYTRUNCATIONPOINT?
    The following is the state of the Extract once it gets stuck. The last replicated change occurred at 2012-07-09 12:46:50.370000. All the changes after that, and there are many, were not captured until I restarted the Extract.
    GGSCI> info extract ext1, showch
    EXTRACT EXT1 Last Started 2012-07-09 12:32 Status RUNNING
    Checkpoint Lag 00:00:00 (updated 00:00:54 ago)
    VAM Read Checkpoint 2012-07-09 12:46:50.370000
    LSN: 0x0000073d:00000aff:0001, Tran: 0000:000bd922
    Current Checkpoint Detail:
    Read Checkpoint #1
    VAM External Interface
    Startup Checkpoint (starting position in the data source):
    Timestamp: 2012-07-09 11:41:06.036666
    LSN: 0x00000460:00000198:0004, Tran: 0000:00089b02
    Recovery Checkpoint (position of oldest unprocessed transaction in the data so
    urce):
    Timestamp: 2012-07-09 12:46:50.370000
    LSN: 0x0000073d:00000afd:0004, Tran: 0000:000bd921
    Current Checkpoint (position of last record read in the data source):
    Timestamp: 2012-07-09 12:46:50.370000
    LSN: 0x0000073d:00000aff:0001, Tran: 0000:000bd922
    Write Checkpoint #1
    GGS Log Trail
    Current Checkpoint (current write position):
    Sequence #: 14
    RBA: 28531192
    Timestamp: 2012-07-09 12:50:02.409000
    Extract Trail: dirdat/e1
    CSN state information:
    CRC: D2-B6-9F-B0
    CSN: Not available
    Header:
    Version = 2
    Record Source = A
    Type = 8
    # Input Checkpoints = 1
    # Output Checkpoints = 1
    File Information:
    Block Size = 2048
    Max Blocks = 100
    Record Length = 20480
    Current Offset = 0
    Configuration:
    Data Source = 5
    Transaction Integrity = 1
    Task Type = 0
    Status:
    Start Time = 2012-07-09 12:32:29
    Last Update Time = 2012-07-09 12:50:02
    Stop Status = A
    Last Result = 400
    Thanks!
    Andre

    It might be something simple (or maybe not); but the best/fastest way to troubleshoot this would be to have Oracle (GoldenGate) support review your configuration. There are a number of critical steps required to allow GG to interoperate with MS's capture API. (I doubt this is it, but is your TranLogOptions on one line? It looks like you have it on two , the way it's formatted here.)
    Anyway, GG support has seen it all, and can probably wrap this up quickly. (And if it was something simple -- or even a bug -- do post back here & maybe someone else can benefit from the solution.)
    Perhaps someone else will be able to provide a better answer, but for the most part troubleshooting this (ie, sql server) via forum tends to be a bit like doing brain surgery blindfolded.

  • Microsoft OLE DB Provider for SQL Server 2012 Analysis Services Error Code = 0x80040E05

    I've been running a package for 6 months without any issues that pulls data from a cube via MDX and moves it to a database.
    We've recently moved to a new server x64 and the connections all succeed but at runtime (in 64 = T&F) I get the following error:
    [SRC AdExchange [87]] Error: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80040E05.
    An OLE DB record is available.  Source: "Microsoft OLE DB Provider for SQL Server 2012 Analysis Services."  Hresult: 0x00000001  Description: "Error Code = 0x80040E05, External Code = 0x00000000:.".
    [SSIS.Pipeline] Error: SRC AdExchange failed the pre-execute phase and returned error code 0xC0202009.
    I've installed SQL_AS_OLEDB, SQL_AS_ADOMD from Microsoft® SQL Server® 2012 SP1 Feature Pack
    Why is my connection failing?

    The server has a 10Gbps network connection.
    I've installed CU6 for SQL Server 2012 SP1
    http://support.microsoft.com/kb/2874879/en-us#hfi
    I'm still getting the same error at runtime in BIDS:
    [SRC AdExchange [87]] Error: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80040E05.
    An OLE DB record is available.  Source: "Microsoft OLE DB Provider for SQL Server 2012 Analysis Services."  Hresult: 0x00000001  Description: "Error Code = 0x80040E05, External Code = 0x00000000:.".
    [SSIS.Pipeline] Error: SRC AdExchange failed the pre-execute phase and returned error code 0xC0202009.
    Looking at the errorlog, everything looks normal until:
    Launched startup procedure 'sp_ssis_startup'.
    Attempting to load library 'xpsqlbot.dll' into memory. This is an informational message only. No user action is required.
    Using 'xpsqlbot.dll' version '2011.110.2100' to execute extended stored procedure 'xp_qv'. This is an informational message only; no user action is required.
    Attempting to load library 'xpstar.dll' into memory. This is an informational message only. No user action is required.
    Using 'xpstar.dll' version '2011.110.3000' to execute extended stored procedure 'xp_instance_regread'. This is an informational message only; no user action is required.
    Attempting to load library 'xplog70.dll' into memory. This is an informational message only. No user action is required.
    Using 'xplog70.dll' version '2011.110.2100' to execute extended stored procedure 'xp_msver'. This is an informational message only; no user action is required.
    AppDomain 2 (SSISDB.dbo[runtime].1) created.
    Unsafe assembly 'microsoft.sqlserver.integrationservices.server, version=11.0.0.0, culture=neutral, publickeytoken=89845dcd8080cc91, processorarchitecture=msil' loaded into appdomain 2 (SSISDB.dbo[runtime].1).
    Unsafe assembly 'microsoft.sqlserver.integrationservices.server, version=11.0.0.0, culture=neutral, publickeytoken=89845dcd8080cc91, processorarchitecture=msil' loaded into appdomain 2 (SSISDB.dbo[runtime].1).
    AppDomain 3 (master.sys[runtime].2) created.
    Software Usage Metrics is enabled.
    SQL Server is terminating because of a system shutdown. This is an informational message only. No user action is required.
    Service Broker manager has shut down.
    Error: 17054, Severity: 16, State: 1.
    The current event was not reported to the Windows Events log. Operating system error = (null). You may need to clear the Windows Events log if it is full.
    .NET Framework runtime has been stopped.
    SQL Trace was stopped due to server shutdown. Trace ID = '1'. This is an informational message only; no user action is required.
    The SQL Server Network Interface library could not deregister the Service Principal Name (SPN) [ MSSQLSvc/XXXXXXXXXX.com ] for the SQL Server service. Error: 0xffffffff, state: 63. Administrator should deregister this SPN manually to avoid client authentication
    errors.
    The SQL Server Network Interface library could not deregister the Service Principal Name (SPN) [ MSSQLSvc/XXXXXXXXXX.com:1433 ] for the SQL Server service. Error: 0xffffffff, state: 63. Administrator should deregister this SPN manually to avoid client authentication
    errors.
    Wondering about the unsafe assembly in SSIS and deregistering the SPN..
    Anyone have any ideas on how to resolve this?
    Thanks!

  • HotFix4 for SQL Server 2005 doesn't work

    I am having problems getting a machine running SQL Server 2008 R2 (Express) to connect to a SQL  Server 2005 (Full Standard Edition) DB on another machine via the LAN. It times out when connecting from a C# application. SSMS on the SQL 2008 machine
    connects fine, but SSMS on the 2005 machine (connecting to a copy of the same DB on the 2008 machine) says it isn't supported and I need a hotfix.
    I downloaded the hotfix link  from Microsoft, but when I paste it into the Browser I get a DNS name failure for the Hotfix4 server. I don't want to paste the full link here because it is not made public by MS and you need to request it. However, the
    first part of the URL is: http://hotfix4.microsoft.com/SQL%20Server%202005...etc.
    I have checked out all of the obvious things in both instances (2005 AND 2008)... Both instances are configured to accept remote connections, both have TCP/IP and named pipes protocols enabled, both are running fine for local connections, both have the SQL
    Browser enabled and the Computer Browser Service running on the network.
    As the hot fix doesn't install, I am kind of out of ideas.
    Any suggestions?
    Thanks,
    Pete
    I used to write COBOL... now I can do anything.

    SQL 2005 is out of
    mainstream support:
    SQL Server
    2005
    SP4
    04/12/2011
    04/12/2016
    Technical support continues till 04/12/2016, yet mainstream (hotfix) support ends as of 04/12/2011; options
    for hotfix support after 04/12/2011:
    ð  Continue
    with self-help
    ð  Upgrade
    to the latest supported service pack for SQL Server 2005 or SQL Server 2008 or SQL Server 2008 R2
    ð  Extended
    hotfix support agreement
    SP4 and CU3 are still available for download, if this helps...
    http://support.microsoft.com/kb/2507769
    http://www.microsoft.com/en-US/download/details.aspx?id=7218
    But you should update your SQL-Server..... SQL 2014 is the latest version

  • SSIS 2008 Logging - SSIS Log Provider for SQL Server

    Simple SSIS package - copies a table from Server A to Server B.
    Everything works except logging using SSIS Log Provider for SQL Server. Selecting the similar events to using
    SSIS Log Provider for Text Files does generates a file on the server.
    I specified server B on the configuration for SSIS Log Provider for SQL Server.
    Logged on with a username with sysadmin rights on Server B. Executed the package using that login and nothing is written to msdb..sysssislog
    any ideas what's going on or what am I doing wrong?

    Just to be sure that the basics are right :
    You sure that you configured the log provider correctly?
    Selected the required 'components' on the left bar?
    Selected events on the Details Tab?
    selected the right table? I mean it should be --> select * from msdb..sysssislog
    The default schema in your DB setup is dbo right? Coz, if not the logging will be a real problem. You might wanna recreate the table then .. 
    Btw, if you still experience issues, I'd advise you to enable all events and check for logging just to make sure that there isnt any prob with the DB setup .. 
    Do try and let us know :) 
    Edited:
    Also a few more:
    Check what database you chose in the connection setting page of the logging configuration wizard and check for entries in sysssislog of tht DB as, the sysssislog will be created only in that db.
    Ensure that you have execute permission on sp_ssis_addlogentry in that particular db (mostly msdb). Can be checked via:
    select name,
    has_perms_by_name(name, 'OBJECT', 'EXECUTE') as has_execute
    from msdb.sys.procedures
    where name like 'sp_ssis_addlogentry' --for 2008 edition and up
    or name like 'sp_dts_addlogentry' --for 2005 edition
    Thanks to @Sathyanarrayanans for pointing this out :)
    Hope this should help ..
    Thanks, Jay <If the post was helpful mark as 'Helpful and if the post answered your query, mark as 'Answered'>

  • Type 3/4 driver required for SQL Server (Weblogic 6.1)

    Hi All,
    I would like to know the site where I can download the driver for SQL Server on Weblogic 6.1 Application server or if someone can send me the driver via email to my email id [email protected]
    Thanks in advance,
    Seetesh Hindlekar

    Hi Sanjay,
    Thanks a lot. I did download a driver for 1 month evaluating period called JSQLConnect but was not able to connect to my local SQL Server.
    There is one driver which gets connected to the SQL Server but I am not able to connect to the SQL Server installed on my comp but it does connect to SQL servers available on the network.
    Any reasons for this not connecting to the local server.
    Seetesh

  • Specifying Partner Failover in odbc.ini for Linux ODBC driver for SQL Server

    Hi,
    We are trying to specify the partner failover when making connections via Linux ODBC driver for SQL Server. Do you know a way to specify partner failover in the DSN entry in odbc.ini? 
    We are using Linux ODBC driver for SQL Server and unixODBC DriverManager.
    Our DSN is:
    [MyDSN]
    Driver=ODBC Driver 11 for SQL Server
    Description=MyDSN
    Server=MySERVER,<port>
    I also tried to use the following connection string in Python (via pyodbc) bypassing the DSN and connecting using the server name.
    cnxn = pyodbc.connect("Server=server1,<port>;UID=sa;PWD=<pwd>;Driver=ODBC Driver 11 for SQL Server;Failover_Partner=Server2,<port2>;Database=MyDB")
    If the standby DB is on server1 connection fails. It doesn't try server2. If the DB is principal on server1 connection succeeds.
    Thank you,
     Liviu
    LLL

    Hello,
    Did you use ODBC Driver for SQL Server to connect to database that uses database mirroring?
    According to
    the BOL,it does not list "Failover_Partner" as ODBC Driver for Linux connection string keywords.
    But ODBC driver for Linux supports
    AlwaysOn Availability Groups or Failover Cluster Instance by specify MultiSubnetFailover=True when connecting to a SQL Server 2012 availability group listener or SQL Server 2012 Failover Cluster Instance.
    The following case is about similar issue and it seems that ODBC Driver on Linux does not recognize Failover_Partner in connection string. (Not test myself).You can refer to:
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/6a8b5223-066d-4493-84a3-b5374a32e433/native-client-on-linux-does-not-recognize-failoverpartner-in-connection-string?forum=sqldataaccess
    Regards,
    Fanny Liu
    Fanny Liu
    TechNet Community Support

  • WSUS service pack updates for SQL Server 2008 R2

    Hello, how do we configure WSUS for SQL Server Service Packs? One problem we have is that we do not see SQL server 2008 R2 SP2 available as a download from our WSUS server and this is when SQL SP1 is installed. Now I can apply these service packs manually
    but when we have 100's of virtual machines with different versions and editions of SQL this task can be quite daunting. Suggestions please?
    RFDZ12

    from memory, SQL service packs and CUs are not released via WSUS.
    This article is a good reference for (products other than Windows) for what has been released via WSUS:
    http://technet.microsoft.com/en-us/windowsserver/bb456965
    The equivalent article for Windows releases, is
    http://support.microsoft.com/kb/894199
    Or, you can check the MU catalog (which doesn't necessarily mean the update is available via WSUS, but if the update you seek in is the MU catalog, you can usually import it into your WSUS)
    http://catalog.update.microsoft.com/
    some updates for products are not able to be released via WSUS, for various reasons, and so in those cases, you need a different deployment solution.
    Don
    (Please take a moment to "Vote as Helpful" and/or "Mark as Answer", where applicable.
    This helps the community, keeps the forums tidy, and recognises useful contributions. Thanks!)

  • Increase Performance and ROI for SQL Server Environments

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

  • How to access the datasource window in SSRS for sql server 2008 R2 for writing my query without having to go through the wizard?

    I have used SSRS a lot years ago with Sql Server 2000 and Sql Server 2005. I have written external assemblies, ... But now I have to do this with Sql Server 2008 (R2 -- which I realize I am way behind the times already but ...)  in sql server 2000 and
    2005 there was a tab for datasource to the left of the tab for design which was to the left of the preview tab. How do I get to the datasource window in sql server 2008 (r2) ?
    I see that datasource explorer.  But where can I get to the datasource window to edit my queries and so forth for sql server 2008 (R2)?
    Thanks
    Rich P

    I think I found the answer to my question --- just right-click on the the Data Sources or Datasets for editing connections and dataset queries.  I'm guessing it gets even fancier with Sql Svr 2012 - 2014.    Man, that's the one thing
    about coding platforms -- you let it go for a few years and come back, and everything has changed (well, a lot of things).  Now I need to figure out how to add an external assembly to SSRS 2008 (R2).
    Rich P

  • Need to connect SQL Server 2008 and CR 2008 via OLE DB SQL Server Provider

    I am relatively new to Crystal but have done some minor design/layout work in the past. I just purchased CR 2008 and dowloaded a Eval Copy of SQL Server 2008 to build test reports that will then be uploaded to a hosted web app we use for use with live data.
    I am having some trouble, however, getting SQL server and CR 08 to talk the way I need them to. We have to use the OLE DB for SQL Server Provider connection for the reports to work in our hosted live environment but I cannot get this to work on SQL 2008. The server never appears in the Server dropdown and if I manually type it get the following error:
    failed to open connection.
    Detail ADO Error Code: 0x80004005
    Source: Microsoft OLE DB Provider for SQL Server
    Description: [DBNETLIB][ConnectionOpen (Connect()).]SQL Server does not exist or access is denied.
    SQL State: 08001
    Native Error: 17 [Database Vendor Code: 17 ]
    I am setting up the connection through the Database Expert choosing the OLE DB (ADO) folder and then selecting 'Microsoft OLE DB Provider for SQL Server' then entering my server name, user id, password, and Database name.
    I have uninstalled and reinstalled everything twice. I have installed the Microsoft SQL Server 2008 Analysis Service 10.0 OLEDB Provider pack.
    My provider/host does use SQL Server 2005 but since I can't get a free Evaluation version of it I have to use 2008 for my local copy. I have searched the web over (including here) and cannot find an answer to my issue.
    I cannot use SQL Server Express version because our copy of the DB is too big. I cannot use the Native 10 provider as the reports will not work once I have uploaded them to the live data.
    any help anyone can provide for this would be GREATLY appreciated. Thank you and sorry if this is a dupe post. I have searched everywhere and cannot find the answer.

    Hello,
    Go to Microsoft's site and download their OLE DB test tool and test the connection. Did you install the MS client tools to test the connection also?
    As a test try creating an ODBC System DSN just to verify you can connect and create a report also.
    And don't use the SA account, MS 2008 disabled it, sort of, so you'll have to create a new account and grant permissions to any table you need to use.
    SQL 2008 changed security model big time so it's not the same as 2005 once was....
    Good luck
    Don

  • Is it possible to have different authentication mode for SQL Server Database Engine and corresponding SQL Server instance?

    Hi,
    I have installed the x64 SQL Server 2008 R2 Express with default settings and run MBSA 2.3 (using default settings too). It shows three SQL Server instances: MSSQL10_50.SQLEXPRESS, SQLEXPRESS and SQLEXPRESS (32-bit). For the first, authentication
    mode is Windows, for the rest two - mixed. Here https://social.msdn.microsoft.com/Forums/sqlserver/en-US/03e470dc-874d-476d-849b-c805acf5b24d/sql-mbsa-question-on-folder-permission?forum=sqlsecurity question
    about such multiple instances was asked and the answer is that "MSSQL10.TEST_DB
    is the instance ID for the SQL Server Database Engine of the instance, TEST_DB", so in my case, it seems that MSSQL10_50.SQLEXPRESS is the instance ID for SQL Server Database Engine  of the SQLEXPRESS instance.
    I have two questions:
    1) How can it be that SQL Server DB Engine instance has different authentication mode than corresponding SQL Server Instance?
    2) Why 32-bit instance reported although I installed only 64-bit version?
    Also, this https://social.technet.microsoft.com/Forums/security/en-US/6b12c019-eaf0-402c-ab40-51d31dce968f/mbsa-23-reporting-sql-32bt-instance-is-running-in-mixed-mode-when-it-is-set-to-integrated?forum=MBSA question seems to be related to this
    issue, but there is no answer :(.
    Upd: Tried on clean Windows 8 installation and Windows 7 with the same result.

      Because I DO NOT want the three people who will be having access to the production SQL Server to also have access to the primary host ProductionA.  Since I have to allow them to RDC into the box to manage the SQL Server, I figure why not create
    a separate VM for each one of them and they can RDC into those instead.
    Does this make any sense?
    Any tips are greatly appreciated.  The main reason for doing this is because the three people who will be accessing the box, I need to isolate each one of them and at the same time keep them off of the primary ProductionA.
    Thanks for your help.
    M
    Hello M,
    Since you dont want the 3 guys to have access to Production machine A.You can install SQL Server client .By client i mean SQL server management studio(SSMS) on there local desktop and then create login for them in SQL Server.Open port on which your SQL server
    is running for three of the machines so that they can connct.Now with SSMS installed on each machine each can connect to SQL server from there own machine.
    I would also like you to be cautious with giving Sysadmin privilege to all three of them ,first please note down what task they would do and then decide what rights to be provided.
    Your option will also work but you need to create 3 VM for that .Which is more tedious task.
    Hope this helps
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • Do I need to install Security Hotfix (KB2977319) after Cumulative Update 12 for SQL Server 2008 R2 SP2

    HI,
    I have installed Cumulative Update 12 for SQL Server 2008 R2 SP2 on my SharePoint instances. This was to resolve a known  issue faced with the instance. CU12 helped resolve the issue. My company is rather strict regarding security hotfixes. But I am
    not sure if this particular hotfix [Security Hotfix (KB2977319)] is required if the instance has CU12 applied.
    Tested this on a Lab server, the installation did run fine, the summary log also stated that the KB is applied. But the Build Number did not change. Hence the doubt.
    Overall summary:
      Final result:                  Passed
      Exit code (Decimal):           0
      Exit message:                  Passed
      Start time:                    2014-09-06 10:31:21
      End time:                      2014-09-06 10:55:49
      Requested action:              Patch
    Instance SPNTSQLTRN overall summary:
      Final result:                  Passed
      Exit code (Decimal):           0
      Exit message:                  Passed
      Start time:                    2014-09-06 10:48:08
      End time:                      2014-09-06 10:55:45
      Requested action:              Patch
    Package properties:
      Description:                   SQL Server Database Services 2008 R2
      ProductName:                   SQL2008
      Type:                          RTM
      Version:                       10
      SPLevel:                       2
      KBArticle:                     KB2977319
      KBArticleHyperlink:            http://support.microsoft.com/?kbid=2977319
      PatchType:                     QFE
      AssociatedHotfixBuild:         0
      Platform:                      x64
      PatchLevel:                    10.52.4321.0
      ProductVersion:                10.52.4000.0
      GDRReservedRange:              10.50.4001.0:10.50.4199.0;10.50.4200.0:10.50.4250.0
      PackageName:                   SQLServer2008-KB2977319-x64.exe
      Installation location:         e:\ac2af22d88ee645b5b32b5c178\x64\setup\
    Please inform if I need to apply the hotfix on CU12. Thanks in advance.
    John S

    Yes you must install Security update mentioned in KB 2977319 it is important for SQL Server to be patches with this security update. Without this it could allow an attacker to compromise your system and gain control over it.
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Articles

Maybe you are looking for

  • Substitution for Functional area

    We are on ECC 5.0 In SAP Note 740519 for deriving the functional area, it mentions a substitution and a callup point can be used for application area FI. Can someone shed some light on what this means and how to determine the substitution? Thanks! Li

  • Making a gun fire.

    Okay, this is a little out of left field, but I wanted to share it with the forum, just for fun. I recently bought some filter plugins from Virtix that promised a simulated "machine gun" effect. It works fine, but you can't get a "single shot" effect

  • H:Form onValueChange

    How can we control value change actions on a form instead of using every form elements onValueChange action? Is it possible to use a generic onValueChange action for h:form tag?

  • Playing AVI files on a Mac

    I'm not sure if this is the correct forum for this question. I have a bunch of AVI files that I want to play. They open in Quicktime which says necessary software is missing. Is there some plug in I need to view AVI in Quicktime, or are there other p

  • About Date Range Selection in Contributor Recognition Program- List

    Hello SCN, Happy that we have a classified date range selection like Year wise, Contest period wise and also with date range in the Top Contributors list. However earlier we had a luxury to see the point classified as  "All Time Contribution".  I per