Source System creation in BI 7.0 for SQL Server using DB Connect

Hi
My requirement  is to extract data from SQL Server 2005 to BI 7.0 using DB Connect.
I need to create a SOURCE SYSTEM for that purpose,As i am new to BW i don't know how to start to create a  SOURCE SYSTEM.
What are the parameters i require either from Basis team to install or for DB team User ID in SQL (What type of user).
Questions:
1)  create a db user with authorization to read this views.......as i created views on tables...do i need authorisation to read table or views
"On BW Server side (If BW server isn't based on SQL Server):
(1) Install SQL Server client and test connection to SQL Server DB.
(2) Install DBSL of SQL server (download lib_dbsl_<num>.sar file from service market and sapcar to BW kernel)"
How can i know on which server our BW System is Based and HOW Install SQL Server client and test connection to SQL Server DB and
Install DBSL of SQL server (download lib_dbsl_<num>.sar file from service market and sapcar to BW kernel
and
how to create Source System to SQL Server on BW
Please update me in details i this is my first Source System Creation in BW
Thanks in advance

go through the link this may be helpful                  http://msdn.microsoft.com/en-us/library/dd299430.aspx

Similar Messages

  • Setup SSL for SQL Server 2014

    I'm trying to get SSL setup for SQL Server. I'm assuming version doesn't matter. It's all about the same but I'm using 2014.
    I have an Enterprise CA so I've setup certs through it.  I've tried using several templates but none seem to work.  I thought it might be needing to use a SAN cert but that doesn't seem to work either.
    I create the cert, add it to the computer account. 
    Go to the properties of "Protocols for MSSQLSERVER" and change the settings on the cert tab to my cert. 
    Attempting to restart SQL Server fails.  Once I clear the setting the service will start.
    What am I missing?
    David Jenkins

    Hello,
    Please read the following resources:
    http://support.microsoft.com/kb/316898/en-us
    http://thesqldude.com/2012/04/21/setting-up-ssl-encryption-for-sql-server-using-certificates-issues-tips-tricks/
    Hope this helps.
    Regards,
    Alberto Morillo
    SQLCoffee.com

  • Source system creation

    hi experts,
    this is the first time i am creating source system
    i am have requirement to load data one bw system to another bi system,
    so, to do that i defined and created logical system in bw for bi system and
    i created same for bw in bi system also and everything is ok.
    i guess my next step is to create source system(my self) for bw in bi .
    when i am trying to create it is showing that it is already exists.and here what type of source system i need to choose. i am not founding any my self bw data mart
    status symbol under  source system creation node and what type of the source system i need to choose and whether it is manually or automatically.if i choose automatically it is asking create new destination ,create back ground user and connection of source system to bw (it is asking to fill all fields),for manual system creation only logical system and rfc destination. is it required to creat source system or not because my self bw data mart symbol is present in bw system and bi system also.please correct me if anything wrong
    thanks
    venu

    hi chandra shekhar &chitrat,
    thanks for your quick reply,when i choose sap business information symbol
    it is asking to fill all the entries like crate new destination under this i provided
    target computer in this i provided ip address of bi system( i am creating in bw system),system id & system number,and  back groung user in source system and
    password ,connection to the source system ,target system for back ground user
    and connection to the source system what i need to provide
    thanks
    venu

  • Error while source system creation/Source system transport

    Hi,
         In our BW implementation, the system landscape is only 2 servers. One is BW Dev, another is BW Prd. All configurations are done in BW Dev and I extracted the data from QAS for this BW Dev system.
         Since we can’t create the source systems in the BW Production system, I tried to create the source system for the SAP R3 Production system in BW Dev and trying to transport this to BW Production Instance.
         I have two queries.
    1.  Is this way is correct for creating the Source System for BW Production system?
    2. When I tried to activate the source system I am getting the following error messages. (I tested the RFC Destinations. There is no problem, I also done the configurations for logical system creation and client assignment in both the systems).
    Error in source system GWPCLNT900
         Message no. RSAR502
    Diagnosis
         With Remote Function Call to system GWPCLNT900 error connection closed
         (no data)  ccurred.
    Procedure
         Remove the cause of the error.
    When tried again I got the following error message
    Basic type ZSBB030 doesn’t exist. EA 447
    Result of the destination check: Timeout RSAR 375.
    Could anyone throw some light on these errors?
    Thanks in advance.
    Krishna

    Hi,
    you need to maintain the logical systems on both R/3 and BW. You also need to set up the RFC connection with users that can log in to the other system. The R/3 system must be open for customizing in scc4.
    In transaction se03 the system change options for 'Business Information Warehouse: SAP Namespace' and 'Business Information Warehouse: Customer Namespace' may need to be modifiable. I am not completely sure about this.
    The last thing I will mention is that I have seen source system creation fail when the RFC-users are system users, and the source system creation worked if the RFC-users was display user while creating the source system. The RFC-users must be changed back to system user later...
    Please look at best practice documents on a more step-by-step guide...
    Best regards,
    Christian Frier

  • With out source system creation we have transported all the flow

    Hi all,
    With out source system creation in Quality we have transported all the data flow
    now is there any option to map the source system to transported flow.
    Regards
    Kiran Kumar

    Hi,
    you can reimport the same TR again once the mapping is done in BW from source system.
    If you have left the TR's in the import queue of BW for later import then you can reimport it again.
    This will go correctly.
    In qualtiy to do the mappings you will need development authorization.
    Thanks
    Ajeet

  • R/3 Source system creation

    Hello,
    I wish to add an R/3 system to my BI 7.0 system. What exactly do I need to enter in the "Create SAP Source System" window? Should I leave backgrounduser as ALEREMOTE? I dont see any option of automatic source system creation. Could u pls help me with a step by step solution.
    Thanks. POINTS WILL BE GIVEN.
    SD

    Hi
    Defining the source system
    As of release 640, you can use upper and lower case characters for R/3 user passwords. This is only uppercase characters for previous releases. Remember that your passwords are automatically converted to uppercase in previous Basis releases.
    As BI 7.0 System has Basis 700 below, upper and lower case passwords are important.
    So, for defining source system is almost same as ordinary BW Source System definitions.
    One more important thing is that you have to take into account the EDI Ports. This may cancel your system definition if the <spsettings are not correct. This settings must be controlled on both BI and Source System.</p>
    Here are the steps to configure Source System with Basis Relase lower than 640 in BI 70:
    1. Take a look at table EDIPORT of your BW system and note the next free number for the field "Port". This is the adjacent number of the highest entry like 'A0000000123' for example. In this case e.g. take '124'.
    2. Choose transaction "snum" and object "ediport".
    Select "number ranges" from the menu "Goto" and here the button "status".
    Change the CURRENT NUMBER of the ranges to the next free number you noted from table EDIPORT.
    If the number range of your BW system is correct please check the same in your OLTP system.
    3. Call TCODE RSA1 on BI System in BI Client.
    4. Select Source Systems on Left Menu
    5. Left Click SAP System and Choose Create
    6. Important think is, as source system has Basis Release Lower than 640, You have to give All Passwords (For BI ALEREMOTE User and for Source System ALEREMOTE User) in UPPERCASE.
    My Favourite is giving RFCCPIC for ALEREMOTE Passwords.
    7. It Must Be OK.
    Assign points if it helps
    Regards,
    Sumit

  • A question about item "type and release" of  source system creation

    Hello expert,
    I have a question about item "type and release" of  source system creation.
    As we know,when we create a web servie source system,there will display a pop-up which includes three items as "logical system","source system"and "type and release".
    About the item "type and release",when we push "F4" button,there will be three default selections as below:
    "ORA 115     Oracle Applications 11i
    TLF 205     Tealeaf 2.05B
    XPD 020     SAP xPD".
    Who can tell me when and how should I use the three selections.
    And also I attempted to input the item by some optional letters except the default three selections and it seems that I can input it freely.
    Thank you and Best Regards,
    Maggie

    Hello DMK,
    Thank you very much for your answer.It is very helpful for me.
    Can I ask you further about it?
    I got that it is a semantic description item.
    You said the default selections are set by our basis people.Would you like to tell me how should we creat a new value except the default ones for item "type and release"?Only by inputing the value in the item directly?But you see we canot see the new value item we created by ourself when we push "F4" button next time ,is that ok?Or do we have to ask basis people to define one new value item just like the default seletions before we use it.
    Also if possible would you like to describe detail about "This becomes important when you are troubleshooting certain issues especially when RFC connection problems."
    Thank you and Best Regards,
    Maggie
    Message was edited by: Maggie

  • Kerberos Configuration Manager for SQL Server: Access of system information failed!

    I'm trying to use the new Kerberos Configuration Manager for SQL Server tool that was released recently to verify SPN on several SQL Servers, but any time I attempt to connect to a server I get the following error in the log:
    6/24/2013 3:48:22 PM Info: Connect to WMI, \\<HOSTNAME>\root\cimv2
    6/24/2013 3:48:25 PM Error: Access of system information failed System.DirectoryServices.AccountManagement.PrincipalOperationException: An error (1332) occurred while enumerating the group membership.  The member's SID could not be resolved.
       at System.DirectoryServices.AccountManagement.SAMMembersSet.IsLocalMember(Byte[] sid)
       at System.DirectoryServices.AccountManagement.SAMMembersSet.MoveNextLocal()
       at System.DirectoryServices.AccountManagement.SAMMembersSet.MoveNext()
       at System.DirectoryServices.AccountManagement.FindResultEnumerator`1.MoveNext()
       at System.Linq.Enumerable.Contains[TSource](IEnumerable`1 source, TSource value, IEqualityComparer`1 comparer)
       at KerberosCM.WMIHelper.isUserLocalAdmin(SystemInfo si, UserPrincipal user)
       at KerberosCM.WMIHelper.getUserInfo(SystemInfo mi)
    6/24/2013 3:48:25 PM Error: Error System.Exception: Access of system information failed!
       at KerberosCM.WMIHelper.getUserInfo(SystemInfo mi)
       at KerberosCM.SystemInfo.GetInfo()
       at KerberosConfigMgr.Utility.Login(String serverName, String login, String password, Boolean isCmdLine, Form uiForm)
    Things I have tried to resolve this:
    1. Verified that my account is a Domain Admin.
    2. Attempt to connect locally without inputting any information for server/user/pw
    3. Attempt to connect remotely using server/user/pw
    I always receive the same error message and log: Access of system information failed!
    Has anyone else run into this issue?

    Update to this:
    There were some invalid user accounts added to the Local Administrators group only showing up as a GUID. This caused the enumeration of the group to fail and generate the error in my original post. Removing those user accounts from the Administrators group
    got past the enumeration error.
    However, now when attempting to connect to the servers (locally or remotely) I get this error:
    6/27/2013 10:24:24 AM Info: Connect to WMI, \root\cimv2
    6/27/2013 10:24:38 AM Error: Access of system information failed System.Runtime.InteropServices.COMException (0x80070035): The network path was not found.
       at System.DirectoryServices.DirectoryEntry.Bind(Boolean throwIfFail)
       at System.DirectoryServices.DirectoryEntry.Bind()
       at System.DirectoryServices.DirectoryEntry.get_AdsObject()
       at System.DirectoryServices.PropertyValueCollection.PopulateList()
       at System.DirectoryServices.PropertyValueCollection..ctor(DirectoryEntry entry, String propertyName)
       at System.DirectoryServices.PropertyCollection.get_Item(String propertyName)
       at System.DirectoryServices.AccountManagement.SAMStoreCtx.ResolveCrossStoreRefToPrincipal(Object o)
       at System.DirectoryServices.AccountManagement.SAMMembersSet.MoveNextForeign()
       at System.DirectoryServices.AccountManagement.SAMMembersSet.MoveNext()
       at System.DirectoryServices.AccountManagement.FindResultEnumerator`1.MoveNext()
       at System.Linq.Enumerable.Contains[TSource](IEnumerable`1 source, TSource value, IEqualityComparer`1 comparer)
       at KerberosCM.WMIHelper.isUserLocalAdmin(SystemInfo si, UserPrincipal user)
       at KerberosCM.WMIHelper.getUserInfo(SystemInfo mi)
    6/27/2013 10:24:38 AM Error: Error System.Exception: Access of system information failed!
       at KerberosCM.WMIHelper.getUserInfo(SystemInfo mi)
       at KerberosCM.SystemInfo.GetInfo()
       at KerberosConfigMgr.Utility.Login(String serverName, String login, String password, Boolean isCmdLine, Form uiForm)

  • SSIS Error Code DTS_E_OLEDBERROR. An OLE DB error has occurred. Error code: 0x80004005. An OLE DB record is available. Source: "Microsoft OLE DB Provider for SQL Server 2012 Analysis Services." Hresult: 0x80004005 Description: "Internal error: An une

    SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80004005.  An OLE DB record is available.  Source: "Microsoft OLE DB Provider for SQL Server 2012 Analysis Services."  Hresult: 0x80004005
     Description: "Internal error: An unexpected error occurred (file 'pcxmlacommon.cpp', line 43, function 'PCFault::RaiseError').
    I'm getting above error in the pre- execute phase of a DFT when I'm trying to fetch data from a SSAS cube using mdx query.
    I'm using OLE DB provider for connecting to cube.
    I got one resolution for Error Code 0X8004005, which asked me to add 'Format= Tabular' in cube's connection string. It does not seem to work either. Can any one help me out on this???

    You are probably missing an update.
    I saw a MS Connect post https://connect.microsoft.com/SQLServer/feedback/details/250920/error-using-oledb-or-datareader-to-get-analysis-services-data where that suggestion was proposed as a fix, but is for an older SQL Server version.
    So do there is a question if you pull data from SQL Server 2012 SSAS using SSIS 2012 (so no other build is involved).
    Arthur My Blog

  • Microsoft Access Text Driver missing! and ...Cannot initialize the data source object of OLE DB provider "MSDASQL" for linked server "(null)".

    In order to use OpenRowSet, I installded Microsoft Access Database Engine 2010. However, I could not find Microsoft Access Text Driver in Drivers of ODBC Data Source Administrator.  Could I get some help with that?
    Thank you very much!

    I am local admin and try to run the following script, but I got an error. Could anyone help me look at it?
    EXEC sp_configure 'show advanced options', 1
    go
    RECONFIGURE
    GO
    EXEC sp_configure 'ad hoc distributed queries', 1
    go
    RECONFIGURE
    GO
    SELECT * FROM OPENROWSET('MSDASQL',
    'Driver={Microsoft Access Text Driver (*.txt, *.csv)};
    DefaultDir=D:\;','SELECT * FROM Test.csv')
    Configuration option 'show advanced options' changed from 1 to 1. Run the RECONFIGURE statement to install.
    Configuration option 'Ad Hoc Distributed Queries' changed from 1 to 1. Run the RECONFIGURE statement to install.
    OLE DB provider "MSDASQL" for linked server "(null)" returned message "[Microsoft][ODBC Driver Manager] Data source name not found and no default driver specified".
    Msg 7303, Level 16, State 1, Line 1
    Cannot initialize the data source object of OLE DB provider "MSDASQL" for linked server "(null)".

  • System Center Orchestrator 2012 Integration Pack for Lync Server 2013 release Date?

    Hi,
    I want to know about System Center Orchestrator 2012 Integration Pack for Lync Server 2013 release Date?
    or Microsoft has already released it?
    Amit Sharma

    Hi,
    I believe it will be around Q4. At the moment there isn't any updates on the release dates yet. So I suggest you to wait patiently.
    Guowen Su
    Cisco Certified Network Associate
    Cisco Certified Internetwork professional - MPLS
    Certified Information Systems Security Professional
    Microsoft Partner Network 2011
    Microsoft Certified Professional
    Microsoft Certified Systems Administrator:Security
    Microsoft Certified Systems Engineer: Security
    Microsoft Certified Technology Specialist: Windows Server 2008 Active Directory, Configuration
    Microsoft Certified Technology Specialist: Windows Server 2008 Network Infrastructure, Configuration
    Microsoft Certified Technology Specialist: Windows Server 2008 Applications Infrastructure, Configuration
    Microsoft Certified Technology Specialist: Windows 7, Configuring
    Microsoft Certified IT Professional: Enterprise Administrator
    Microsoft Certified IT Professional: Server Administrator
    Certified Ethical Hacker
    Computer Hacking Forensics Investigator
    Certified Sonicwall Security Administrator
    Microsoft Geeks

  • Increase Performance and ROI for SQL Server Environments

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

    May 2015
    Explore
    The Buzz from Microsoft Ignite 2015
    NetApp was in full force at the recent Microsoft Ignite show in Chicago, talking about solutions for hybrid cloud, and our proven solutions for Microsoft SQL Server and other Microsoft applications.
    Hot topics at the NetApp booth included:
    OnCommand® Shift. A revolutionary technology that lets you move virtual machines back and forth between VMware and Hyper-V environments in minutes.
    Azure Site Recovery to NetApp Private Storage. Replicate on-premises SAN-based applications to NPS for disaster recovery in the Azure cloud.
    These tools give you greater flexibility for managing and protecting important business applications.
    Chris Lemmons
    Director, EIS Technical Marketing, NetApp
    If your organization runs databases such as Microsoft SQL Server and Oracle DB, you probably know that these vendors primarily license their products on a "per-core" basis. Microsoft recently switched to "per-core" rather than "per-socket" licensing for SQL Server 2012 and 2014. This change can have a big impact on the total cost of operating a database, especially as core counts on new servers continue to climb. It turns out that the right storage infrastructure can drive down database costs, increase productivity, and put your infrastructure back in balance.
    In many customer environments, NetApp has noticed that server CPU utilization is low—often on the order of just 20%. This is usually the result of I/O bottlenecks. Server cores have to sit and wait for I/O from hard disk drives (HDDs). We've been closely studying the impact of all-flash storage on SQL Server environments that use HDD-based storage systems. NetApp® All Flash FAS platform delivers world-class performance for SQL Server plus the storage efficiency, application integration, nondisruptive operations, and data protection of clustered Data ONTAP®, making it ideal for SQL Server environments.
    Tests show that All Flash FAS can drive up IOPS and database server CPU utilization by as much as 4x. And with a 95% reduction in latency, you can achieve this level of performance with half as many servers. This reduces the number of servers you need and the number of cores you have to license, driving down costs by 50% or more and paying back your investment in flash in as little as six months.
    Figure 1) NetApp All Flash FAS increases CPU utilization on your SQL Server database servers, lowering costs.
    Source: NetApp, 2015
    Whether you're running one of the newer versions of SQL Server or facing an upgrade of an earlier version, you can't afford not to take a second look at your storage environment.
    End of Support for Microsoft SQL Server 2005 is Rapidly Approaching
    Microsoft has set the end of extended support for SQL Server 2005 for April 2016—less than a year away. With support for Microsoft Windows 2003 ending in July 2015, time may already be running short.
    If you're running Windows Server 2003, new server hardware is almost certainly needed when you upgrade SQL Server. Evaluate your server and storage options now to get costs under control.
    Test Methodology
    To test the impact of flash on SQL Server performance, we replaced a legacy HDD-based storage system with an All Flash FAS AFF8080 EX. The legacy system was configured with almost 150 HDDs, a typical configuration for HDD storage supporting SQL Server. The AFF8080 EX used just 48 SSDs.
    Table 1) Components used in testing.
    Test Configuration Components
    Details
    SQL Server 2014 servers
    Fujitsu RX300
    Server operating system
    Microsoft Windows 2012 R2 Standard Edition
    SQL Server database version
    Microsoft SQL Server 2014 Enterprise Edition
    Processors per server
    2 6-core Xeon E5-2630 at 2.30 GHz
    Fibre channel network
    8Gb FC with multipathing
    Storage controller
    AFF8080 EX
    Data ONTAP version
    Clustered Data ONTAP® 8.3.1
    Drive number and type
    48 SSD
    Source: NetApp, 2015
    The test configuration consisted of 10 database servers connected through fibre channel to both the legacy storage system and the AFF8080 EX. Each of the 10 servers ran SQL Server 2014 Enterprise Edition.
    The publicly available HammerDB workload generator was used to drive an OLTP-like workload simultaneously from each of the 10 database servers to storage. We first directed the workload to the legacy storage array to establish a baseline, increasing the load to the point where read latency consistently exceeded 20ms.
    That workload was then directed at the AFF8080 EX. The change in storage resulted in an overall 20x reduction in read latency, a greater than 4x improvement in IOPS, and a greater than 4x improvement in database server CPU utilization.
    Figure 2) NetApp All Flash FAS increases IOPS and server CPU utilization and lowers latency.
    Source: NetApp, 2015
    In other words, the database servers are able to process four times as many IOPS with dramatically lower latency. CPU utilization goes up accordingly because the servers are processing 4x the work per unit time.
    The All Flash FAS system still had additional headroom under this load.
    Calculating the Savings
    Let's look at what this performance improvement means for the total cost of running SQL Server 2014 over a 3-year period. To do the analysis we used NetApp Realize, a storage modeling and financial analysis tool designed to help quantify the value of NetApp solutions and products. NetApp sales teams and partners use this tool to assist with return on investment (ROI) calculations.
    The calculation includes the cost of the AFF8080 EX, eliminates the costs associated with the existing storage system, and cuts the total number of database servers from 10 to five. This reduces SQL Server licensing costs by 50%. The same workload was run with five servers and achieved the same results. ROI analysis is summarized in Table 2.
    Table 2) ROI from replacing an HDD-based storage system with All Flash FAS, thereby cutting server and licensing costs in half.
    Value
    Analysis Results
    ROI
    65%
    Net present value (NPV)
    $950,000
    Payback period
    six months
    Total cost reduction
    More than $1 million saved over a 3-year analysis period compared to the legacy storage system
    Savings on power, space, and administration
    $40,000
    Additional savings due to nondisruptive operations benefits (not included in ROI)
    $90,000
    Source: NetApp, 2015
    The takeaway here is that you can replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs, with the majority of the savings derived from the reduction in SQL Server licensing costs.
    Replace your existing storage with All Flash FAS and get a big performance bump while substantially reducing your costs.
    Maximum SQL Server 2014 Performance
    In addition to the ROI analysis, we also measured the maximum performance of the AFF8080 EX with SQL Server 2014. A load-generation tool was used to simulate an industry-standard TPC-E OLTP workload against an SQL Server 2014 test configuration.
    A two-node AFF8080 EX achieved a maximum throughput of 322K IOPS at just over 1ms latency. For all points other than the maximum load point, latency was consistently under 1ms and remained under 0.8ms up to 180K IOPS.
    Data Reduction and Storage Efficiency
    In addition to performance testing, we looked at the overall storage efficiency savings of our SQL Server database implementation. The degree of compression that can be achieved is dependent on the actual data that is written and stored in the database. For this environment, inline compression was effective. Deduplication, as is often the case in database environments, provided little additional storage savings and was not enabled.
    For the test data used in the maximum performance test, we measured a compression ratio of 1.5:1. We also tested inline compression on a production SQL Server 2014 data set to further validate these results and saw a 1.8:1 compression ratio.
    Space-efficient NetApp Snapshot® copies provide additional storage efficiency benefits for database environments. Unlike snapshot methods that use copy-on-write, there is no performance penalty; unlike full mirror copies, NetApp Snapshot copies use storage space sparingly. Snapshot copies only consume a small amount of storage space for metadata and additional incremental space is consumed as block-level changes occur. In a typical real-world SQL Server deployment on NetApp storage, database volume Snapshot copies are made every two hours.
    First introduced more than 10 years ago, NetApp FlexClone® technology also plays an important role in SQL Server environments. Clones are fully writable, and, similar to Snapshot copies, only consume incremental storage capacity. With FlexClone, you can create as many copies of production data as you need for development and test, reporting, and so on. Cloning is a great way to support the development and test work needed when upgrading from an earlier version of SQL Server. You'll sometimes see these types of capabilities referred to as "copy data management."
    A Better Way to Run Enterprise Applications
    The performance benefits that all-flash storage can deliver for database environments are significant: more IOPS, lower latency, and an end to near-constant performance tuning.
    If you think the performance acceleration that comes with all-flash storage is cost prohibitive, think again. All Flash FAS doesn't just deliver a performance boost, it changes the economics of your operations, paying for itself with thousands in savings on licensing and server costs. In terms of dollars per IOPS, All Flash FAS is extremely economical relative to HDD.
    And, because All Flash FAS runs NetApp clustered Data ONTAP, it delivers the most complete environment to support SQL Server and all your enterprise applications with capabilities that include comprehensive storage efficiency, integrated data protection, and deep integration for your applications.
    For complete details on this testing look for NetApp TR-4303, which will be available in a few weeks. Stay tuned to Tech OnTap for more information as NetApp continues to run benchmarks with important server workloads including Oracle DB and server virtualization.
    Learn more about NetApp solutions for SQL Server and NetApp All-flash solutions.
    Quick Links
    Tech OnTap Community
    Archive
    PDF

  • Microsoft OLE DB Provider for SQL Server 2012 Analysis Services Error Code = 0x80040E05

    I've been running a package for 6 months without any issues that pulls data from a cube via MDX and moves it to a database.
    We've recently moved to a new server x64 and the connections all succeed but at runtime (in 64 = T&F) I get the following error:
    [SRC AdExchange [87]] Error: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80040E05.
    An OLE DB record is available.  Source: "Microsoft OLE DB Provider for SQL Server 2012 Analysis Services."  Hresult: 0x00000001  Description: "Error Code = 0x80040E05, External Code = 0x00000000:.".
    [SSIS.Pipeline] Error: SRC AdExchange failed the pre-execute phase and returned error code 0xC0202009.
    I've installed SQL_AS_OLEDB, SQL_AS_ADOMD from Microsoft® SQL Server® 2012 SP1 Feature Pack
    Why is my connection failing?

    The server has a 10Gbps network connection.
    I've installed CU6 for SQL Server 2012 SP1
    http://support.microsoft.com/kb/2874879/en-us#hfi
    I'm still getting the same error at runtime in BIDS:
    [SRC AdExchange [87]] Error: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80040E05.
    An OLE DB record is available.  Source: "Microsoft OLE DB Provider for SQL Server 2012 Analysis Services."  Hresult: 0x00000001  Description: "Error Code = 0x80040E05, External Code = 0x00000000:.".
    [SSIS.Pipeline] Error: SRC AdExchange failed the pre-execute phase and returned error code 0xC0202009.
    Looking at the errorlog, everything looks normal until:
    Launched startup procedure 'sp_ssis_startup'.
    Attempting to load library 'xpsqlbot.dll' into memory. This is an informational message only. No user action is required.
    Using 'xpsqlbot.dll' version '2011.110.2100' to execute extended stored procedure 'xp_qv'. This is an informational message only; no user action is required.
    Attempting to load library 'xpstar.dll' into memory. This is an informational message only. No user action is required.
    Using 'xpstar.dll' version '2011.110.3000' to execute extended stored procedure 'xp_instance_regread'. This is an informational message only; no user action is required.
    Attempting to load library 'xplog70.dll' into memory. This is an informational message only. No user action is required.
    Using 'xplog70.dll' version '2011.110.2100' to execute extended stored procedure 'xp_msver'. This is an informational message only; no user action is required.
    AppDomain 2 (SSISDB.dbo[runtime].1) created.
    Unsafe assembly 'microsoft.sqlserver.integrationservices.server, version=11.0.0.0, culture=neutral, publickeytoken=89845dcd8080cc91, processorarchitecture=msil' loaded into appdomain 2 (SSISDB.dbo[runtime].1).
    Unsafe assembly 'microsoft.sqlserver.integrationservices.server, version=11.0.0.0, culture=neutral, publickeytoken=89845dcd8080cc91, processorarchitecture=msil' loaded into appdomain 2 (SSISDB.dbo[runtime].1).
    AppDomain 3 (master.sys[runtime].2) created.
    Software Usage Metrics is enabled.
    SQL Server is terminating because of a system shutdown. This is an informational message only. No user action is required.
    Service Broker manager has shut down.
    Error: 17054, Severity: 16, State: 1.
    The current event was not reported to the Windows Events log. Operating system error = (null). You may need to clear the Windows Events log if it is full.
    .NET Framework runtime has been stopped.
    SQL Trace was stopped due to server shutdown. Trace ID = '1'. This is an informational message only; no user action is required.
    The SQL Server Network Interface library could not deregister the Service Principal Name (SPN) [ MSSQLSvc/XXXXXXXXXX.com ] for the SQL Server service. Error: 0xffffffff, state: 63. Administrator should deregister this SPN manually to avoid client authentication
    errors.
    The SQL Server Network Interface library could not deregister the Service Principal Name (SPN) [ MSSQLSvc/XXXXXXXXXX.com:1433 ] for the SQL Server service. Error: 0xffffffff, state: 63. Administrator should deregister this SPN manually to avoid client authentication
    errors.
    Wondering about the unsafe assembly in SSIS and deregistering the SPN..
    Anyone have any ideas on how to resolve this?
    Thanks!

  • Group managed service accounts for SQL Server

    Hey guys,
    Unfortunately I missed that (g/s)MSAs aren't supported yet for SQL Servers but I'm using them without any worries since ages.
    As i digged a bit deeper I could find different informations due to the related TechNet entrys. So it seems Microsofts Informations about (s)MSAs and gMSAs aren't consistent.
    I'm not a SQL Server guy and use SQL only for System Center testing stuff so i would like to get a real world exps of SQL Server guys.
    Should I continue using gMSAs or are there any worries I should know?
    some sources I found so far:
    Not supported:
    "Hi Adam,
    Thank you for your feedback. Windows Server 2012 Group Managed Service Account is not currently supported as SQL 2012 released earlier than Windows Server 2012. We will consider to support gMSA in future SQL Server release.
    Regards,
    Min He, Program Manager, SQL Server"
    11.2012 -
    https://connect.microsoft.com/SQLServer/feedback/details/767211/gmsa-for-sql-server-failover-Clusters
    gMSA are not yet available, are not yet supported for SQL Server.  gMSA exist and are available and supported in Windows Server 2012 and higher.  SQL does not support them , but
    from an OS perspective, they exist and are supported.    
    http://blogs.msdn.com/b/sqlosteam/archive/2014/02/19/msa-accounts-used-with-sql.aspx
    Within the FAQ Task Scheduler isn't supported as well ...
    http://technet.microsoft.com/en-us/library/ff641729%28WS.10%29.aspx
    ... but also PFEs using them for Tasks... this is confusin... 0o
    http://blogs.msdn.com/b/arvindsh/archive/2014/02/03/managed-service-accounts-msa-and-sql-2012-practical-tips.aspx
    supported?:
    Configure Windows Service Accounts and Permissions
    ... New Account Types Available with Windows 7 and Windows Server 2008 R2
    http://technet.microsoft.com/en-us/library/ms143504(v=sql.110).aspx#Default_Accts
    The MSA must be created in the Active Directory by the domain administrator before SQL Server setup can use it for SQL Server services.
    others sources won't mentioning s/gMSAs...
    I couldn't find clear informations about using gMSA for SQL Server 2014. 
    only the same page which also Looks like the page for 2008 R2 and SQL 2012.
    Configure Windows Service Accounts and Permissions
                SQL Server 2014        
    http://msdn.microsoft.com/en-us/library/ms143504.aspx
    annoying topic so far... ;) 

    Hi Enrico
    aside from what Dan says about the risk for support, on which I agree, the following thread may clear it up a bit:
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/acb2048c-ffce-4d44-b882-6aafc7eb689d/managed-service-accounts-to-run-sql-server-service?forum=sqlsecurity
    Andreas Wolter (Blog |
    Twitter)
    MCM - Microsoft Certified Master SQL Server 2008
    MCSM - Microsoft Certified Solutions Master Data Platform, SQL Server 2012
    www.andreas-wolter.com |
    www.SarpedonQualityLab.com

  • Missing Stored Procedure for SQL Server

    Hello,
    I'm trying to create the standard TestStand 4 schema for SQL Server. I get the following error when I run multiple numeric test cases:
    An error occurred calling 'LogResults' in 'ITSDBLog' of 'zNI TestStand Database Logging'
    An error occurred executing a statement.
    Schema: SQL Server Stored Proc (NI).
    Statement: MEAS_NUMERICLIMIT.
    Could not find stored procedure 'InsertStepMeasNumericLimit'.
    Description: Could not find stored procedure 'InsertStepMeasNumericLimit'.
    Number: -2147217900
    NativeError: 2812
    SQLState: 42000
    Reported by: Microsoft OLE DB Provider for SQL Server
    Source: TSDBLog
    -2147217900; User-defined error code.
    Step 'Log Results to Database' of sequence 'Log To Database' in 'Database.seq'
    My process for creating the database is to create a .sql file from the TestStand database configuration window. I have selected SQL Server Stored Procedure. I then build the database by running the script in the Execute SQL window of the Database viewer. I don't get any errors when creating the database or the stored procedures. I have noticed that a stored procedure, called InsertMeasNumericLimit does exist. When I rename this procedure to InsertStepMeasNumericLimit, the sequence with the multiple numeric executes without error, and my data is in the database.
    Now to my questions>
    1. Am I doing something wrong in my procedure to create the database schema, that would cause this?
    2. Have the NI schemas or DB calls in TS been changed since 3.5 and validated?
    3. Has anyone else run across this problem, and what others can I expect to encounter with the SQL server stored procedure schema?
    Thanks in advance for the help,
    BCE

    bce,
    1. What you are doing is correct. There is a bug with the SQL Server Stored Procedure. Apparently
    2. The schemas hasn't changed since 3.5? Were you able to go through the same process in TS 3.5 without running into this error?
    3. It looks like you have found the way to solve this problem. There are a few other step types that have similar error. Specifically the IVI steps. Before the bug is fixed, you can use validate to determines whether the statement and column information for the selected schema matches the tables and columns in the database. Alternatively, you may look at the file below to find out the correct spelling of the stored procedures.
    <TestStand 4.0>\Components\NI\Models\TestStandModels\Database\SQL Server Create Stored Proc Result Tables.sql
    Regards,
    Song Du
    Systems Software
    National Instruments R&D

Maybe you are looking for

  • How can i set up mail rules from an iPad?

    I am familiar with using rules to filter email. From my Mac at home I acces iCloud.com and have no problems. I went away for the long weekend, and brought only my iPad. I started receiving spam from a new source and wanted to filter it out. From safa

  • Using acrobat pro in word 2007

    I just upgraded to word 2007 and have acrobat pro 7.0.0. I can't figure out how to save a document as a pdf! Adobe Acrobat Pro shows up as an add-in in "word options" - Add-ins, but there is no add-in tab, and I can't find any way to actually use it!

  • How do I connect my iphone to my itunes?

    I need help! I just boughtmy first iphone and I have no idea how to "sync" it to my itunes account so I can start using it. Help!

  • Two-phase commit

    Hi, all. Does WeblogicServer 5.1sp10 support two-phase commit? I'm going to use Informix and Oracle. But I'm not sure that WLS 5.1 support it. For example,when new record is inserted in Informix, it must be inserted in oracle. and update,delete proce

  • Help on Certified Documents from the Superintendent of Documents

    I often review regulations and legislation that have been certified by the Superintendent of Documents.  I've used Acrobat Pro in the past and was able to run OCR to at least allow for commenting/notes, highlighting text and creating my own bookmarks