Raid levels for sql server files

Hi,
I have a doubt on raid levels.
What RAID level should be used for SQL Data files and why?
What RAID level should be used for SQL LOG files and why?
What RAID level should be used for Tempdb files and why?
Thank you.

SQL Log File - The I/O is sequential in nature but only if you have a single database on the raid group of disks. If you have 2 or more database each doing its sequential pattern, you result having random I/O on the raid group.
SQL Data files - yes they are random in nature and could be write intensive but consider that SQL Server writes synchronously only in the log file, the writes on the data files are made in background by lazy writer and check point so it is not mandatory
to have RAID10, I will always choose RAID 5. The write penality of the RAID5 does not affect performances.
One last consideration: modern storage do "wide striping" i.e. they spread I/O operation on a pool of raid group. Every read/write operation first hit the cache of the storage (battery protected RAM) if the cache has not the blocks needed by the
rad operation or has not the amount of space to satisfy the write operation, the I/O is demanded to the disks...
Work with your storage admin if you have one and ask him to do cache partitioning and reserve some only for SQL Server log, you will be amazed by the gain in performances...

Similar Messages

  • Do we need to format data and log files with 64k cluster size for sql server 2012?

    Do we need to format data and log files with 64k cluster size for sql server 2012?
    Does this best practice still applies to sql server 2012 & 2014?

    Yes.  The extent size of SQL Server data files, and the max log block size have not changed with the new versions, so the guidance should remain the same.
    Microsoft SQL Server Storage Engine PM

  • Which .exe file I should choose to apply SP2 for SQL Server 2008r2 Enterprise Edition 64Bit

    Hi All,
    Which .exe file I should choose to apply SP2 for SQL Server 2008r2 Enterprise Edition 64Bit.
    And what is the difference between below .exe files. 
    SQLServer2008R2SP2-KB2630458-x64-ENU.exe
    SQLServer2008R2SP2-KB2630458-IA64-ENU.exe
    Grateful to your time and support. Regards, Shiva

    Good day
    Shiva Shakthi,
    I have change the type of the thread to a question, as this is a question :-)
    please chose the right type next time. If you have a question then use a question thread (this is the default), and that will let you mark the answers that you get.
    Have a nice day :-)
    [Personal Site] [Blog] [Facebook]

  • What would be the reason for sql server agent [MSSQLSERVER] service terminated unexpectedly?

    What would be the reason for sql server agent [MSSQLSERVER] service terminated unexpectedly?
    only below details found in error log
    07:26:44.170 spid60      
    FILESTREAM: effective level = 3, configured level = 3, file system access share name = 'MSSQLSERVER'.
    07:26:45.720 spid51      
    Configuration option 'Agent XPs' changed from 0 to 1. Run the RECONFIGURE statement to install.
    07:26:45.730 spid51      
    FILESTREAM: effective level = 3, configured level = 3, file system access share name = 'MSSQLSERVER'.
    07:28:57.840 spid64      
    Configuration option 'Agent XPs' changed from 1 to 0. Run the RECONFIGURE statement to install.
    07:28:57.840 spid64      
    FILESTREAM: effective level = 3, configured level = 3, file system access share name = 'MSSQLSERVER'.
    and below error on agent log
    2014-01-01 07:26:28 - ? [131] SQLSERVERAGENT service stopping due to a stop request from a user, process, or the OS...
    2014-01-01 07:26:43 - + [188] Scheduler engine timed out (after 15 seconds) waiting for 1 jobs(s) to stop
    2014-01-01 07:26:44 - + [098] SQLServerAgent terminated (forcefully)
    Rahul

    What would be the reason for sql server agent [MSSQLSERVER] service terminated unexpectedly?
    only below details found in error log
    07:26:44.170 spid60      
    FILESTREAM: effective level = 3, configured level = 3, file system access share name = 'MSSQLSERVER'.
    07:26:45.720 spid51      
    Configuration option 'Agent XPs' changed from 0 to 1. Run the RECONFIGURE statement to install.
    07:26:45.730 spid51      
    FILESTREAM: effective level = 3, configured level = 3, file system access share name = 'MSSQLSERVER'.
    07:28:57.840 spid64      
    Configuration option 'Agent XPs' changed from 1 to 0. Run the RECONFIGURE statement to install.
    07:28:57.840 spid64      
    FILESTREAM: effective level = 3, configured level = 3, file system access share name = 'MSSQLSERVER'.
    and below error on agent log
    2014-01-01 07:26:28 - ? [131] SQLSERVERAGENT service stopping due to a stop request from a user, process, or the OS...
    2014-01-01 07:26:43 - + [188] Scheduler engine timed out (after 15 seconds) waiting for 1 jobs(s) to stop
    2014-01-01 07:26:44 - + [098] SQLServerAgent terminated (forcefully)
    Rahul
    Can you post the output of below query
    select @@version.
    If you are using SQl server express edition ,SQL server agent feature is not available.It seems from errorlog that you can start it but actually you cannot.But first confirm what is version and edition of your SQL server
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • Security Update for SQL Server 2005 SP3 (KB2494113) failed

    Here is the Error, Can't get the update install, please help.
    KB Number: KB2494113
    Machine: MJA01
    OS Version: Server 4.0 Service Pack 1 (Build 7601)
    Package Language: 1033 (ENU)
    Package Platform: x86
    Package SP Level: 3
    Package Version: 4060
    Command-line parameters specified:
    Cluster Installation: No
    Prerequisites Check & Status
    SQLSupport: Passed
    Products Detected                         Language  Level  Patch Level       Platform  Edition
    SQL Server Database Services 2005 (BKUPEXEC)  ENU       SP3    2005.090.4035.00  x86       EXPRESS
    SQL Server Tools and Workstation Components 2005  ENU       SP2           9.2.3042  x86       EXPRESS
    Products Disqualified & Reason
    Product                                   Reason
    SQL Server Tools and Workstation Components 2005  The product instance SQL Tools does not have prerequisite update 4035 installed. Update 4060 is dependent on prerequisite update 4035. Exit setup and refer to the Knowledge Base article to find the prerequisite
    patch. Install the prerequisite and rerun the installation.
    Processes Locking Files
    Process Name          Feature               Type          User Name                  PID
    Product Installation Status
    Product                   : SQL Server Database Services 2005 (BKUPEXEC)
    Product Version (Previous): 4035
    Product Version (Final)   : 
    Status                    : Failure
    Log File                  : C:\Program Files (x86)\Microsoft SQL Server\90\Setup Bootstrap\LOG\Hotfix\SQL9_Hotfix_KB2494113_sqlrun_sql.msp.log
    SQL Express Features      : 
    Error Number              : 29528
    Error Description         : MSP Error: 29528  The setup has encountered an unexpected error while Setting Internal Properties. The error is: Fatal error during installation.
    Product                   : SQL Server Tools and Workstation Components 2005
    Product Version (Previous): 3042
    Product Version (Final)   : 
    Status                    : NA
    Log File                  : 
    SQL Express Features      : 
    Error Description         : The product instance SQL Tools does not have prerequisite update 4035 installed. Update 4060 is dependent on prerequisite update 4035. Exit setup and refer to the Knowledge Base article to find the prerequisite
    patch. Install the prerequisite and rerun the installation.
    Summary
         One or more products failed to install, see above for details
         Exit Code Returned: 29528

    Hi Bhanu,
    I have uninstall the SQL Server Tools and Workstation Components 2005.
    But still unable to upg the SQL Server with the KB249411. Any more ideas?
    This is the Summary log.
    Time: 09/10/2014 11:09:48.218
    KB Number: KB2494113
    Machine: MJA01
    OS Version: Server 4.0 Service Pack 1 (Build 7601)
    Package Language: 1033 (ENU)
    Package Platform: x86
    Package SP Level: 3
    Package Version: 4060
    Command-line parameters specified:
    Cluster Installation: No
    Prerequisites Check & Status
    SQLSupport: Passed
    Products Detected                         Language  Level  Patch Level       Platform  Edition
    SQL Server Database Services 2005 (BKUPEXEC)  ENU       SP3    2005.090.4035.00  x86       EXPRESS
    Products Disqualified & Reason
    Product                                   Reason
    Processes Locking Files
    Process Name          Feature               Type          User Name                  PID
    Product Installation Status
    Product                   : SQL Server Database Services 2005 (BKUPEXEC)
    Product Version (Previous): 4035
    Product Version (Final)   : 
    Status                    : Failure
    Log File                  : C:\Program Files (x86)\Microsoft SQL Server\90\Setup Bootstrap\LOG\Hotfix\SQL9_Hotfix_KB2494113_sqlrun_sql.msp.log
    SQL Express Features      : 
    Error Number              : 29528
    Error Description         : MSP Error: 29528  The setup has encountered an unexpected error while Setting Internal Properties. The error is: Fatal error during installation.
    Summary
         One or more products failed to install, see above for details
         Exit Code Returned: 29528
    Here is what is log in the SQL9_Hotfix_KB2494113_sqlrun_sql.msp.log
    Property(S): CommonFilesFolder.D9BC9C10_2DCD_44D3_AACC_9C58CAF76128 = C:\Program Files (x86)\Common Files\
    MSI (s) (B8:F0) [11:09:39:826]: Product: Microsoft SQL Server 2005 Express Edition - Update 'GDR 4060 for SQL Server Database Services 2005 ENU (KB2494113)' could not be installed. Error code 1603. Additional information is available in the log file C:\Program
    Files (x86)\Microsoft SQL Server\90\Setup Bootstrap\LOG\Hotfix\SQL9_Hotfix_KB2494113_sqlrun_sql.msp.log.
    MSI (s) (B8:F0) [11:09:39:827]: Windows Installer installed an update. Product Name: Microsoft SQL Server 2005 Express Edition. Product Version: 9.3.4035.00. Product Language: 1033. Manufacturer: Microsoft Corporation. Update Name: GDR 4060 for SQL Server Database
    Services 2005 ENU (KB2494113). Installation success or error status: 1603.
    MSI (s) (B8:F0) [11:09:39:828]: Note: 1: 1729 
    MSI (s) (B8:F0) [11:09:39:828]: Product: Microsoft SQL Server 2005 Express Edition -- Configuration failed.
    MSI (s) (B8:F0) [11:09:39:829]: Windows Installer reconfigured the product. Product Name: Microsoft SQL Server 2005 Express Edition. Product Version: 9.3.4035.00. Product Language: 1033. Manufacturer: Microsoft Corporation. Reconfiguration success or error
    status: 1603.
    MSI (s) (B8:F0) [11:09:39:829]: Attempting to delete file C:\Windows\Installer\63906e.msp
    MSI (s) (B8:F0) [11:09:39:829]: Unable to delete the file. LastError = 32
    MSI (s) (B8:F0) [11:09:40:092]: Deferring clean up of packages/files, if any exist
    MSI (s) (B8:F0) [11:09:40:092]: Attempting to delete file C:\Windows\Installer\63906e.msp
    MSI (s) (B8:F0) [11:09:40:094]: MainEngineThread is returning 1603
    MSI (s) (B8:6C) [11:09:40:097]: RESTART MANAGER: Session closed.
    MSI (s) (B8:6C) [11:09:40:097]: No System Restore sequence number for this installation.
    === Logging stopped: 9/10/2014  11:09:39 ===
    MSI (s) (B8:6C) [11:09:40:098]: User policy value 'DisableRollback' is 0
    MSI (s) (B8:6C) [11:09:40:098]: Machine policy value 'DisableRollback' is 0
    MSI (s) (B8:6C) [11:09:40:098]: Incrementing counter to disable shutdown. Counter after increment: 0
    MSI (s) (B8:6C) [11:09:40:098]: Note: 1: 1402 2: HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\Installer\Rollback\Scripts 3: 2 
    MSI (s) (B8:6C) [11:09:40:098]: Note: 1: 1402 2: HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\Installer\Rollback\Scripts 3: 2 
    MSI (s) (B8:6C) [11:09:40:098]: Note: 1: 1402 2: HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\Installer\InProgress 3: 2 
    MSI (s) (B8:6C) [11:09:40:098]: Note: 1: 1402 2: HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\Installer\InProgress 3: 2 
    MSI (s) (B8:6C) [11:09:40:098]: Decrementing counter to disable shutdown. If counter >= 0, shutdown will be denied.  Counter after decrement: -1
    MSI (s) (B8:6C) [11:09:40:099]: Restoring environment variables
    MSI (s) (B8:6C) [11:09:40:099]: Destroying RemoteAPI object.
    MSI (s) (B8:30) [11:09:40:099]: Custom Action Manager thread ending.
    MSI (c) (FC:60) [11:09:40:100]: Decrementing counter to disable shutdown. If counter >= 0, shutdown will be denied.  Counter after decrement: -1
    MSI (c) (FC:60) [11:09:40:101]: MainEngineThread is returning 1603
    Thanks,
    Alice

  • Increase Performance and ROI for SQL Server Environments

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

  • After installed SP1 for SQL Server 2012, can no longer export to csv

    After installing SP1 today via Windows Update, I am no longer able to export data to csv using the SQL Server Import and Export wizard. I get the following error message:
    "Column information for the source and the destination data could not be retrieved, or the data types of source columns were not mapped correctly to those available on the destination provider."
    "Column "col1": Source data type "200" was not found in the data type mapping file."...
    (The above line repeats for each column)
    The work-around I have to do is to manually map each column in the "Edit Mappings..." option from the "Configure Flat File Destination" page of the wizard. This is an extreme inconvenience to have to have to edit the mappings and change
    each column to "string [DT_STR]" type from "byte stream [DT_BYTES]" type each time I want to export to csv. I did not have to do this before installing SP1; it worked perfectly for months with hundreds of exports prior to this update and
    no need to modify mapping.

    I am running Windows 7 64-bit, SQL Server 2012 Express edition. Again, just yesterday from Windows Update, I installed SQL Server 2012 Service Pack 1 (KB2674319), followed by Update Rollup for SQL Server 2012 Service Pack 1 (KB2793634). This situation was
    not occurring before these updates were installed, and I noticed it immediately after they were installed (and of course I restarted my computer after the updates).
    In SSMS I just now created a test DB and table to provide a step-by-step with screenshots.
    Here is the code I ran to create the test DB and table:
    CREATE DATABASE testDB;
    GO
    USE testDB;
    GO
    CREATE TABLE testTable
    id int,
    lname varchar(50),
    fname varchar(50),
    address varchar(50),
    city varchar(50),
    state char(2),
    dob date
    GO
    INSERT INTO testTable VALUES
    (1,'Smith','Bob','123 Main St.','Los Angeles','CA','20080212'),
    (2,'Doe','John','555 Rainbow Ln.','Chicago','IL','19580530'),
    (3,'Jones','Jane','999 Somewhere Pl.','Washington','DC','19651201'),
    (4,'Jackson','George','111 Hello Cir.','Dallas','TX','20010718');
    GO
    SELECT * FROM testTable;
    Results look good:
    id    lname    fname    address    city    state    dob
    1    Smith    Bob    123 Main St.    Los Angeles    CA    2008-02-12
    2    Doe    John    555 Rainbow Ln.    Chicago    IL    1958-05-30
    3    Jones    Jane    999 Somewhere Pl.    Washington    DC    1965-12-01
    4    Jackson    George    111 Hello Cir.    Dallas    TX    2001-07-18
    In Object Explorer, I right-click on the [testDB] database, choose "Tasks", then "Export Data..." and the SQL Server Import and Export Wizard appears. I click Next to leave all settings as-is on the "Choose a Data Source" page, then on the "Choose a Destination"
    page, under the "Destination" drop-down I choose "Flat File Destination" then browse to the desktop and name the file "table_export.csv" then click Next. On the "Specify Table Copy or Query" page I choose "Write a query to specify the data to transfer" then
    click Next. I type the following SQL statement:
    SELECT * FROM testTable;
    When clicking the "Parse" button I get the message "This SQL statement is valid."
    On to the next page, "Configure Flat File Destination" I try leaving the defaults then click Next. This is where I am getting the error message (see screenshot below):
    Then going to the "Edit Mappings..." option on the "Configure Flat File Destination" page, I see that all columns which were defined as varchar in the table are showing as type "byte stream [DT_BYTES]", size "0", the state column which is defined as char(2)
    shows correctly however with type "string [DT_STR]", size "2" (see screenshow below):
    So what I have to do is change the type for the lname, fname, address and city columns to "string [DT_STR]", then I am able to proceed with the export successfully. Again, this just started happening after installing these updates. As you can imagine, this
    is very frustrating, as I do a lot of exports from many tables, with a lot more columns than this test table.
    Thanks for your help.

  • Will Cumulative Update 12 for SQL Server 2008R2 SP2 fix: Warning: Unable to Verify TimeStamp for Path\ProgramName?

    Hello,
    I had a server crash recently and our outsource hosting tech support suggested applying Cumulative Upade 12 for SQL server 2008R2 SP2 to fix the issue.  The exception from our dump file "Warning: Unable to Verify TimeStamp for Path\ProgramName"
    is not in the list of hotfixes for this CU.  Do you know if this will fix this issue?  The CU warns not to apply the CU if your issue id not addressed by the CU.  Here is a portion of the dump file with the relevant error:
    This dump file has an exception of interest stored in it.
    The stored exception information can be accessed via .ecxr.
    (7e8.2ab4): Unknown exception - code 000042ac (first/second chance not available)
    ntdll!NtWaitForSingleObject+0xa:
    00000000`777412fa c3              ret
    0:240> .sympath srv*c:\Websymbols*http://msdl.microsoft.com/download/symbols;
    Symbol search path is: srv*c:\Websymbols*http://msdl.microsoft.com/download/symbols
    Expanded Symbol search path is: srv*c:\websymbols*http://msdl.microsoft.com/download/symbols
    0:240> .reload /f
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\sqlservr.exe, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlservr.exe
    ..........Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\sqlos.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlos.dll
    ...............Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\opends60.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for opends60.dll
    .......Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\BatchParser.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for BatchParser.dll
    ....Unable to load image C:\Program Files\Microsoft SQL Server\100\Shared\instapi10.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for instapi10.dll
    *** ERROR: Module load completed but symbols could not be loaded for instapi10.dll
    ..Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\Resources\1033\sqlevn70.rll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlevn70.rll
    *** ERROR: Module load completed but symbols could not be loaded for sqlevn70.rll
    Press ctrl-c (cdb, kd, ntsd) or ctrl-break (windbg) to abort symbol loads that take too long.
    Run !sym noisy before .reload to track down problems loading symbols.
    .................Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\ftimport.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for ftimport.dll
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\msfte.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for msfte.dll
    *** ERROR: Module load completed but symbols could not be loaded for msfte.dll
    ...........Unable to load image C:\Windows\System32\sqlncli10.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlncli10.dll
    ...Unable to load image C:\Windows\System32\1033\SQLNCLIR10.RLL, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for SQLNCLIR10.RLL
    *** ERROR: Module load completed but symbols could not be loaded for SQLNCLIR10.RLL
    ..Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\xpsqlbot.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for xpsqlbot.dll
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\XPStar.DLL, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for XPStar.DLL
    .Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\sqlscm.dll, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for sqlscm.dll
    ...*** ERROR: Module load completed but symbols could not be loaded for odbcint.dll
    ...Unable to load image C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Binn\Resources\1033\XPStar.RLL, Win32 error 0n2
    *** WARNING: Unable to verify timestamp for XPStar.RLL
    Thank you,
    Steve -

    The error message appears to be related to a debug session and not come from the actual crash.
    So all I know is that you hade a server crash. I don't even know exactly what that means. Did Windows bite the dust? Or was it only SQL Server?
    Assuming the latter, I would expect the SQL Server errorlog to have some information (and that would be ERRORLOG.1 or earlier, since the server have been restarted), but if SQL Server died the output may be incomplete.
    There may be also dump files, but as I rarely look into these, I am not sure how to interpret them. But I am quite confident that "Unable to verify TimeStamp..." is not the reason SQL Server went down.
    I would suggest the following course:
    *  If the server is not critical, do nothing. As long as it has onlyl happened once, it has only happened once.
    *  If the server is critical, open a case with Microsoft if you are not able to figure out the reason yourself. The key here is "Unknown exception - code 000042ac".
    *  If it happens again, you should absolutely open a case. An important thing here is whether the stack dump is identical or something different. If the stack dumps are identical, you may have hit a bug in SQL Server or the OS, and applying CUs or OS
    fixes could help if it is a known issue. If the stack dump is something else, you have ghosts in the machine - that is, bad hardware.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Deploy JDBC driver for SQL server 2005 on PI 7.1

    How to deploy JDBC driver for SQL server 2005 on PI 7.1
    We are in SAP NetWeaver 7.1 Oracle 10G
    Third party system is  SQL server 2005
    There are different JDBC versions are available to download for SQL server 2005.
    I am not sure about the applicable version for the PI 7.1 SP level. Again JMS Adapter needs to be deploy along with this.
    Please help

    Hi,
    Hope this How to Guide help you.
    How To Install and Configure External Drivers for the JDBC & JMS Adapters from
    www.sdn.sap.com/irj/sdn/howtoguides
    Regards,
    Karthick.
    Edited by: Karthick Srinivasan on Apr 13, 2009 4:07 PM

  • OGG for SQL Server - Extract stops capturing - Bug?

    Hi, all,
    I've found a problem with OGG for SQL Server where the Extract stops capturing data after the transaction log is backed up. I've looked for ways to reconfigure OGG to avoid the problem but couldn't find any reference to options to workaround this problem. It seems to be a bug to me.
    My Extract configuration is as follows:
    EXTRACT ext1
    SOURCEDB mssql1
    TRANLOGOPTIONS NOMANAGESECONDARYTRUNCATIONPOINT
    EOFDELAY 60
    EXTTRAIL dirdat/e1
    TABLE dbo.TestTable;
    I'm using the EOFDELAY parameter for testing purposes only, since it's easy to reproduce the scenario that causes the issue when the extract polling is configured with longer intervals.
    When the Transaction Log backup runs, SQL Server marks all the virtual logs that are older than the primary and secondary truncation points as inactive (status = 0). These virtual logs can then be reused if required. They still contain change records, though, and OGG can read from then if required, before they are overwritten. This situation will never occur if we are not using SQL Replication and have the Extract configured with the parameter MANAGESECONDARYTRUNCATIONPOINT.
    However, I'm trying to simulate a scenario where OGG is used along SQL Replication and the extract is configured with the NOMANAGESECONDARYTRUNCATIONPOINT option. The situation that I've reproduced and caused the Extract to stop capturing is the follow sequence of events:
    1. Extract reads transaction log and capture change up to LSN X
    2. More change are made to the database and the LSN is incremented
    3. Log Reader reads Transaction Log, captures changes up to LSN X+Y and advances the secondary truncation point to that LSN
    4. A transaction log occurs, backs up all the active virtual logs, advances the primary truncation point to a LSN greater than LSN X+Y, and marks all the virtual logs with LSNs <= X+Y as inactive (status = 0)
    5. Changes continue to happen in the database consuming all the available inactive virtual logs and overwriting them.
    6. The extract wakes up again to capture more changes.
    At this point, the changes between LSNs X and X+Y are not in the Transaction Log anymore, but are available in the backups. From what I understood in the documentation the Extract should detect that situation and retrieve the changes from the Transaction Log backups. This, however, is not happening and the Extract becomes stuck. It still pools the transaction log at the configured interval query the log state with DBCC LOGINFO, but doesn't move forward anymore.
    If I stop and restart the Extract I can see from the trace that it does the right thing upon startup. It realises that it requires information that's missing from the logs, query MSDB for the available backups, and mine the backups to get the required LSNs.
    I would've thought the Extract should do the same during normal operation, without the need for a restart.
    Is this a bug or the normal operation of the Extract? Is there a way to configure it to avoid this situation without using NOMANAGESECONDARYTRUNCATIONPOINT?
    The following is the state of the Extract once it gets stuck. The last replicated change occurred at 2012-07-09 12:46:50.370000. All the changes after that, and there are many, were not captured until I restarted the Extract.
    GGSCI> info extract ext1, showch
    EXTRACT EXT1 Last Started 2012-07-09 12:32 Status RUNNING
    Checkpoint Lag 00:00:00 (updated 00:00:54 ago)
    VAM Read Checkpoint 2012-07-09 12:46:50.370000
    LSN: 0x0000073d:00000aff:0001, Tran: 0000:000bd922
    Current Checkpoint Detail:
    Read Checkpoint #1
    VAM External Interface
    Startup Checkpoint (starting position in the data source):
    Timestamp: 2012-07-09 11:41:06.036666
    LSN: 0x00000460:00000198:0004, Tran: 0000:00089b02
    Recovery Checkpoint (position of oldest unprocessed transaction in the data so
    urce):
    Timestamp: 2012-07-09 12:46:50.370000
    LSN: 0x0000073d:00000afd:0004, Tran: 0000:000bd921
    Current Checkpoint (position of last record read in the data source):
    Timestamp: 2012-07-09 12:46:50.370000
    LSN: 0x0000073d:00000aff:0001, Tran: 0000:000bd922
    Write Checkpoint #1
    GGS Log Trail
    Current Checkpoint (current write position):
    Sequence #: 14
    RBA: 28531192
    Timestamp: 2012-07-09 12:50:02.409000
    Extract Trail: dirdat/e1
    CSN state information:
    CRC: D2-B6-9F-B0
    CSN: Not available
    Header:
    Version = 2
    Record Source = A
    Type = 8
    # Input Checkpoints = 1
    # Output Checkpoints = 1
    File Information:
    Block Size = 2048
    Max Blocks = 100
    Record Length = 20480
    Current Offset = 0
    Configuration:
    Data Source = 5
    Transaction Integrity = 1
    Task Type = 0
    Status:
    Start Time = 2012-07-09 12:32:29
    Last Update Time = 2012-07-09 12:50:02
    Stop Status = A
    Last Result = 400
    Thanks!
    Andre

    It might be something simple (or maybe not); but the best/fastest way to troubleshoot this would be to have Oracle (GoldenGate) support review your configuration. There are a number of critical steps required to allow GG to interoperate with MS's capture API. (I doubt this is it, but is your TranLogOptions on one line? It looks like you have it on two , the way it's formatted here.)
    Anyway, GG support has seen it all, and can probably wrap this up quickly. (And if it was something simple -- or even a bug -- do post back here & maybe someone else can benefit from the solution.)
    Perhaps someone else will be able to provide a better answer, but for the most part troubleshooting this (ie, sql server) via forum tends to be a bit like doing brain surgery blindfolded.

  • Arc is not implemented for SQL Server 2005 physical model

    I added the Arc in my relational model, which is supposed to provide that references are mutually exclusive (so only one refering column can be not null). However there is not any implementation for this in the SQL Server 2005 physical model. The SQL code generated simply ignores the Arc and allows any values for the columns covered by Arc. If we check Oracle physical model. then it generates some trigger, which controls how many columns have values assigned - this is actually an implementation of Arc. But there is nothing of this kind for SQL Server model...
    By the way, implementing this Arc by trigger looks like to heavy approach. The very simple table level check constraint will do the same job. But this is a different subject though.
    Edited by: Otbl on Nov 3, 2011 7:34 PM

    Hi Dmitry,
    it's fixed in Data Modeler 3.1 EA2.
    Philip

  • How to establish a trusted connection with JDBC for SQL SERVER 2000

    Hi!I am using jdk 1.4 and eclipse 3.3.
    I create a servlet in eclipse with in-build tomcat.
    When I run it ,it was working perfectlly has it was suppose to work.
    In this servlet I connect to a sql 2000 database using jdbc-odbc bridge driver.
    But when I tried to deploy the servlet on tomcat 5.5 manully on the same machine ,it gave me error saying
    [Microsoft][SQLServer JDBC Driver][SQLServer]Login failed
    for user 'sa'
    I searched around some post and found that ok ,I need trusted connection
    But I have 2 Questions
    1). Why was in eclipse I was able to connect to the SQL server and why not in the servlet which I deployed manully on tomcat.
    2). How do I create a trusted connection with JDBC for SQL server 2000
    Thnaks for your help in advance.

    Hi! duffymo ,QussayNajjar ,dvohra09 .
    Thank for help.
    The ideas are really great.
    I am trying generate reports for my company.
    When I used eclipse the code worked perfectly.
    below is code which I used
    out.println("Calling For Class Name<br>");
    Class.forName("sun.jdbc.odbc.JdbcOdbcDriver");
    out.println("Calling For Class Name success Now calling database <br>");
    1). jdbcConnection = DriverManager.getConnection("jdbc:odbc:SQLJasper");
    2). jdbcConnection = DriverManager.getConnection("jdbc:odbc:Driver={SQL Server};Server=ServerName;Database=tempdb");
    3). jdbcConnection = DriverManager.getConnection("jdbc:odbc:Driver={SQL Server};Server=ServerName;Database=tempdb","UID=UserName","Password=Password");
    out.println("connecting to database success<br>");
    I had tried to connect the database using this three way.
    In 1st I tried using DSN name .
    Next 2 self explainer for expert like you.
    I used to 2nd variant to connect in eclipse and it worked fine.
    I not an expert in java ,I just doing some research on jasperReport.
    My best guest is that eclipse is using some library files of which I have no clue.
    Thank's for your help,I appretiate it.
    Once again thank a billion.
    Sorry for the messy righting.

  • Has anybody used the microsoft JDBC 2.0 driver for sql server 2000?

    Hi,
    Has anybody used the JDBC 2.0 driver for sql server 2000 downloadable from the
    microsoft website?When I try using it with WL 6.1 sp1 it says it can't load the
    driver.I try viewing the class file from the jar file using the jar utility it
    gives an unknown Zip format error.Anybody has any solution for this ?If anybody
    has managed to work with this microsoft driver i will be grateful if they provide
    me with a solution.
    Thanks
    Thomas

    Hello Thomas,
    You may want to download the driver again and install it again.
    heres a sample xml tag in the config.xml:
    <JDBCConnectionPool
    DriverName="com.microsoft.jdbc.sqlserver.SQLServerDriver"
    InitialCapacity="3" MaxCapacity="12" Name="MSpool"
    Password="{3DES}fUz1bxR0zDg=" Properties="user=uid"
    Targets="myserver"
    URL="jdbc:microsoft:sqlserver://mydbserver:1433"/>
    ensure that you follow the instructions from Microsoft. For using 2000
    driver you will need to have
    Install_dir/lib/msbase.jar and Install_dir/lib/msutil.jar in addition to
    Install_dir/lib/mssqlserver.jar in the CLASSPATH.
    hth
    sree
    "Thomas" <[email protected]> wrote in message
    news:3c91ec0e$[email protected]..
    Hi,
    Has anybody used the JDBC 2.0 driver for sql server 2000 downloadable from
    the
    microsoft website?When I try using it with WL 6.1 sp1 it says it can't load
    the
    driver.I try viewing the class file from the jar file using the jar utility
    it
    gives an unknown Zip format error.Anybody has any solution for this ?If
    anybody
    has managed to work with this microsoft driver i will be grateful if they
    provide
    me with a solution.
    Thanks
    Thomas

  • JDBC Driver for SQL Server 7

    Is there a free JDBC Driver for SQL Server 7 available? If so, where? I created a server program that listens to a port and inserts records based on information it receives. It uses sun's JDBC:ODBC Bridge but it crashes randomly (sometimes weeks, sometimes days, sometimes hours after starting program). Each time it crashes it generates a error log file like this:
    An unexpected exception has been detected in native code outside the VM.
    Unexpected Signal : EXCEPTION_ACCESS_VIOLATION occurred at PC=0x77f821e1
    Function name=RtlEnterCriticalSection
    Library=C:\WINNT\System32\ntdll.dll
    Current Java thread:
         at sun.jdbc.odbc.JdbcOdbc.driverConnect(Native Method)
         at sun.jdbc.odbc.JdbcOdbc.SQLDriverConnect(Unknown Source)
         at sun.jdbc.odbc.JdbcOdbcConnection.initialize(Unknown Source)
         at sun.jdbc.odbc.JdbcOdbcDriver.connect(Unknown Source)
         at java.sql.DriverManager.getConnection(Unknown Source)
         at java.sql.DriverManager.getConnection(Unknown Source)
         at com.xxxxx.data.DataAccess.connect(DataAccess.java:20)
         at com.xxxxx.xxx.server.Broadcaster.insertOrder(POSServer.java:71)
         at com.xxxxx.xxx.server.ClientListener.run(POSServer.java:298)
    Dynamic libraries:
    0x00400000 - 0x00405000      C:\WINNT\system32\java.exe
    0x77F80000 - 0x77FFB000      C:\WINNT\System32\ntdll.dll
    0x77DB0000 - 0x77E0C000      C:\WINNT\system32\ADVAPI32.dll
    0x77E80000 - 0x77F35000      C:\WINNT\system32\KERNEL32.DLL
    0x77D40000 - 0x77DB0000      C:\WINNT\system32\RPCRT4.DLL
    0x78000000 - 0x78046000      C:\WINNT\system32\MSVCRT.dll
    0x6D420000 - 0x6D4F0000      C:\Program Files\JavaSoft\JRE\1.3.1_02\bin\hotspot\jvm.dll
    Local Time = Mon May 06 10:36:33 2002
    Elapsed Time = 124
    # The exception above was detected in native code outside the VM
    # Java VM: Java HotSpot(TM) Client VM (1.3.1_02-b02 mixed mode)
    Or this:
    An unexpected exception has been detected in native code outside the VM.
    Unexpected Signal : EXCEPTION_ACCESS_VIOLATION occurred at PC=0x77f821e1
    Function name=RtlEnterCriticalSection
    Library=C:\WINNT\System32\ntdll.dll
    Current Java thread:
         at sun.jdbc.odbc.JdbcOdbc.freeConnect(Native Method)
         at sun.jdbc.odbc.JdbcOdbc.SQLFreeConnect(Unknown Source)
         at sun.jdbc.odbc.JdbcOdbcDriver.closeConnection(Unknown Source)
         at sun.jdbc.odbc.JdbcOdbcConnection.close(Unknown Source)
         at sun.jdbc.odbc.JdbcOdbcConnection.finalize(Unknown Source)
         at java.lang.ref.Finalizer.invokeFinalizeMethod(Native Method)
         at java.lang.ref.Finalizer.runFinalizer(Unknown Source)
         at java.lang.ref.Finalizer.access$100(Unknown Source)
         at java.lang.ref.Finalizer$FinalizerThread.run(Unknown Source)
    Dynamic libraries:
    0x00400000 - 0x00405000      C:\Program Files\JavaSoft\JRE\1.3.1_02\bin\javaw.exe
    0x77F80000 - 0x77FFB000      C:\WINNT\System32\ntdll.dll
    0x77DB0000 - 0x77E0C000      C:\WINNT\system32\ADVAPI32.dll
    0x77E80000 - 0x77F35000      C:\WINNT\system32\KERNEL32.DLL
    0x77D40000 - 0x77DB0000      C:\WINNT\system32\RPCRT4.DLL
    Local Time = Mon May 06 10:13:59 2002
    Elapsed Time = 137
    # The exception above was detected in native code outside the VM
    # Java VM: Java HotSpot(TM) Client VM (1.3.1_02-b02 mixed mode)
    I think this error occurs because I'm using sun's JDBC:ODBC Bridge. I'm hoping that using a different driver will solve the problem.
    Thanks,
    Dan

    I used Sun's jdbc odbc driver for a while, with no problems, with Sql Server 7. But anyways, if you want a free type 4 driver, you can get the freetds driver at
    https://sourceforge.net/projects/jtds/

  • Using JSP & JDBC driver for SQL Server 2000 on Red Hat

    I successfully have a .jsp app running on windows server 2000 using JDBC
    driver for SQL Server 2000. Which I installed in order to the following
    Red Hat:
    http://msdn.microsoft.com/MSDN-FILES/027/001/779/install.htm
    I moved the .jsp app over to the Red Hat 9 server running Tomcat, while
    keeping the MS SQL 2000 on windows. The issue I have is setting up the
    JDBC driver for SQL Server 2000 on the Red Hat server.
    I created a folder called /usr/java/MSSQLdriver and unzipped the tar file with
    the driver for SQL Server 2000. And ran the install.ksh script.
    The /usr/java/MSSQLdriver/lib has the following files within it:
    msbase.jar, msutil.jar, & mssqlserver.jar
    I chmod 0777 each of the *.jar files.
    I then went into /etc/profile.d/tomcat.sh and adding the following:
    CLASSPATH=.;/opt/msSQLjdbc/lib/msbase.jar;/opt/msSQLjdbc/lib/msutil.jar;/opt/msSQLjdbc/lib/mssqlserver.jar
    Each time I login and pull up the termial I get the following error:
    bash: /opt/msSQLjdbc/lib/msbase.jar: cannot execute binary file
    bash: /opt/msSQLjdbc/lib/msutil.jar: cannot execute binary file
    bash: /opt/msSQLjdbc/lib/mssqlserver.jar: cannot execute binary file
    And can't connnect to the database within the .jsp app.
    Is there anyone out there using DBC driver for SQL Server 2000 on the Red Hat server?
    Michael

    Sorry, I needed to correct some information of where the drivers were installed.
    I created a folder called /usr/java/MSSQLdriver/new and untar the Microsoft file with the driver for SQL Server 2000. I ran the install.ksh script "sh install.ksh"
    installed the driver into the default directory "/opt/msSQLjdbc".
    The /opt/msSQLjdbc/lib has the following files within it:
    msbase.jar, msutil.jar, & mssqlserver.jar
    Michael

Maybe you are looking for