2012: Execute Package Task (external reference) Very Slow in SSDT

Hi All,
I have 4 packages which I've deployed using the package deployment model to the MSDB package store.
Configuration setup is System Env Var pointing at SSIS Configurations table that populates the connection string of my SSISConfig data source. Source/Dest configurations are in the SSIS Configuration table.
I have a  'master' package that executes the 4 packages in parallel using the Execute Package control flow task (external reference, SSIS Package Store)
When I execute this master package in SSDT it takes 1minute 56 seconds to finish. All the execute package tasks seem to hang until the last few sec then finish one after the other. It's the same if execute one of the Execute Package Task items individually.
It just hangs there for nearly 2 min then rushes through its work (i've verified by looking at the destination tables; nothing happens until the end).
But the packages executed directly in ssdt take 2-3 sec to execute.
I've deployed the master package to the server and executed it via SSMS. It takes <3 sec.
I've RDPed into the ssis box and it's the same. Executing the packages individually or the master package in dtexecui takes less than 3 sec.
Something's going wrong inside SSDT (some kind of timeout? but there are no error msgs or warning) that's making it take nearly 2 minutes.
Have others experienced this? Have I misconfigured something? The env var has been working on my local machine since last week and i've restarted many times since then, so I'm positive SSDT is using the 'current' version of the env var.
Jakub @ Adelaide, Australia

Very strange. The sysssislogs are the same for any method of execution, but there's a 5 sec delay for each step when executing via SSDT
endtime,message
2014-02-06 09:37:15.000,Execute phase is beginning.
2014-02-06 09:37:20.000,PrimeOutput will be called on a component. : 104 : Merge Join : Merge Join
2014-02-06 09:37:20.000,PrimeOutput will be called on a component. : 213 : Unchanged : Unchanged
2014-02-06 09:37:25.000,A component has returned from its PrimeOutput call. : 213 : Unchanged : Unchanged
2014-02-06 09:37:25.000,A component has returned from its PrimeOutput call. : 104 : Merge Join : Merge Join
2014-02-06 09:37:30.000,PrimeOutput will be called on a component. : 143 : ODS - Src - State : ODS - Src - State
2014-02-06 09:37:30.000,PrimeOutput will be called on a component. : 17 : Src - dbo - State : Src - dbo - State
2014-02-06 09:37:35.000,A component has returned from its PrimeOutput call. : 143 : ODS - Src - State : ODS - Src - State
2014-02-06 09:37:35.000,A component has returned from its PrimeOutput call. : 17 : Src - dbo - State : Src - dbo - State
2014-02-06 09:37:35.000,Rows were provided to a data flow component as input. : : 264 : OLE DB Source Output : 182 : Source Checksum : 185 : ChecksumInput : 7 : Source Checksum : Paths[Src - dbo - State.OLE DB Source Output] : Source Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:35.000,Rows were provided to a data flow component as input. : : 263 : OLE DB Source Output : 53 : Destination Checksum : 56 : ChecksumInput : 7 : Destination Checksum : Paths[ODS - Src - State.OLE DB Source Output] : Destination Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:40.000,A component will be given the end of rowset signal. : 53 : Destination Checksum : 56 : ChecksumInput : Destination Checksum : Destination Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:40.000,Rows were provided to a data flow component as input. : : 258 : ChecksumOutput : 104 : Merge Join : 109 : Merge Join Left Input : 7 : Merge Join : Paths[Source Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Left Input]
2014-02-06 09:37:40.000,A component will be given the end of rowset signal. : 182 : Source Checksum : 185 : ChecksumInput : Source Checksum : Source Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:45.000,Rows were provided to a data flow component as input. : : 259 : ChecksumOutput : 104 : Merge Join : 118 : Merge Join Right Input : 7 : Merge Join : Paths[Destination Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Right Input]
2014-02-06 09:37:45.000,Rows were provided to a data flow component as input. : : 264 : OLE DB Source Output : 182 : Source Checksum : 185 : ChecksumInput : 0 : Source Checksum : Paths[Src - dbo - State.OLE DB Source Output] : Source Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:45.000,Rows were provided to a data flow component as input. : : 263 : OLE DB Source Output : 53 : Destination Checksum : 56 : ChecksumInput : 0 : Destination Checksum : Paths[ODS - Src - State.OLE DB Source Output] : Destination Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:50.000,A component has finished processing all of its rows. : 53 : Destination Checksum : 56 : ChecksumInput : Destination Checksum : Destination Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:50.000,A component has finished processing all of its rows. : 182 : Source Checksum : 185 : ChecksumInput : Source Checksum : Source Checksum.Inputs[ChecksumInput]
2014-02-06 09:37:55.000,A component will be given the end of rowset signal. : 104 : Merge Join : 118 : Merge Join Right Input : Merge Join : Merge Join.Inputs[Merge Join Right Input]
2014-02-06 09:38:00.000,Rows were provided to a data flow component as input. : : 259 : ChecksumOutput : 104 : Merge Join : 118 : Merge Join Right Input : 0 : Merge Join : Paths[Destination Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Right Input]
2014-02-06 09:38:05.000,A component has finished processing all of its rows. : 104 : Merge Join : 118 : Merge Join Right Input : Merge Join : Merge Join.Inputs[Merge Join Right Input]
2014-02-06 09:38:10.000,A component will be given the end of rowset signal. : 104 : Merge Join : 109 : Merge Join Left Input : Merge Join : Merge Join.Inputs[Merge Join Left Input]
2014-02-06 09:38:15.000,Rows were provided to a data flow component as input. : : 258 : ChecksumOutput : 104 : Merge Join : 109 : Merge Join Left Input : 0 : Merge Join : Paths[Source Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Left Input]
2014-02-06 09:38:20.000,A component has finished processing all of its rows. : 104 : Merge Join : 109 : Merge Join Left Input : Merge Join : Merge Join.Inputs[Merge Join Left Input]
2014-02-06 09:38:20.000,Rows were provided to a data flow component as input. : : 261 : Merge Join Output : 2 : Add audit columns : 3 : Derived Column Input : 7 : Add audit columns : Paths[Merge Join.Merge Join Output] : Add audit columns.Inputs[Derived Column Input]
2014-02-06 09:38:25.000,A component will be given the end of rowset signal. : 2 : Add audit columns : 3 : Derived Column Input : Add audit columns : Add audit columns.Inputs[Derived Column Input]
2014-02-06 09:38:25.000,Rows were provided to a data flow component as input. : : 260 : Derived Column Output : 195 : Type 1 SCD : 196 : Conditional Split Input : 7 : Type 1 SCD : Paths[Add audit columns.Derived Column Output] : Type 1 SCD.Inputs[Conditional Split Input]
2014-02-06 09:38:30.000,Rows were provided to a data flow component as input. : : 265 : Unchanged : 213 : Unchanged : 214 : Union All Input 1 : 7 : Unchanged : Paths[Type 1 SCD.Unchanged] : Unchanged.Inputs[Union All Input 1]
2014-02-06 09:38:30.000,Rows were provided to a data flow component as input. : : 261 : Merge Join Output : 2 : Add audit columns : 3 : Derived Column Input : 0 : Add audit columns : Paths[Merge Join.Merge Join Output] : Add audit columns.Inputs[Derived Column Input]
2014-02-06 09:38:35.000,A component has finished processing all of its rows. : 2 : Add audit columns : 3 : Derived Column Input : Add audit columns : Add audit columns.Inputs[Derived Column Input]
2014-02-06 09:38:40.000,A component will be given the end of rowset signal. : 195 : Type 1 SCD : 196 : Conditional Split Input : Type 1 SCD : Type 1 SCD.Inputs[Conditional Split Input]
2014-02-06 09:38:45.000,Rows were provided to a data flow component as input. : : 260 : Derived Column Output : 195 : Type 1 SCD : 196 : Conditional Split Input : 0 : Type 1 SCD : Paths[Add audit columns.Derived Column Output] : Type 1 SCD.Inputs[Conditional Split Input]
2014-02-06 09:38:50.000,A component has finished processing all of its rows. : 195 : Type 1 SCD : 196 : Conditional Split Input : Type 1 SCD : Type 1 SCD.Inputs[Conditional Split Input]
2014-02-06 09:38:55.000,A component will be given the end of rowset signal. : 66 : Insert ODS - Src - State : 79 : OLE DB Destination Input : Insert ODS - Src - State : Insert ODS - Src - State.Inputs[OLE DB Destination Input]
2014-02-06 09:38:55.000,A component will be given the end of rowset signal. : 220 : Update ODS - Src - State : 225 : OLE DB Command Input : Update ODS - Src - State : Update ODS - Src - State.Inputs[OLE DB Command Input]
2014-02-06 09:38:55.000,A component will be given the end of rowset signal. : 213 : Unchanged : 214 : Union All Input 1 : Unchanged : Unchanged.Inputs[Union All Input 1]
2014-02-06 09:39:00.000,Rows were provided to a data flow component as input. : : 262 : New : 66 : Insert ODS - Src - State : 79 : OLE DB Destination Input : 0 : Insert ODS - Src - State : Paths[Type 1 SCD.New] : Insert ODS - Src - State.Inputs[OLE DB Destination Input]
2014-02-06 09:39:00.000,Rows were provided to a data flow component as input. : : 257 : Changed : 220 : Update ODS - Src - State : 225 : OLE DB Command Input : 0 : Update ODS - Src - State : Paths[Type 1 SCD.Changed] : Update ODS - Src - State.Inputs[OLE DB Command Input]
2014-02-06 09:39:00.000,"The final commit for the data insertion in ""Insert ODS - Src - State"" has started."
2014-02-06 09:39:00.000,Rows were provided to a data flow component as input. : : 265 : Unchanged : 213 : Unchanged : 214 : Union All Input 1 : 0 : Unchanged : Paths[Type 1 SCD.Unchanged] : Unchanged.Inputs[Union All Input 1]
2014-02-06 09:39:00.000,"The final commit for the data insertion in ""Insert ODS - Src - State"" has ended."
2014-02-06 09:39:05.000,A component has finished processing all of its rows. : 213 : Unchanged : 214 : Union All Input 1 : Unchanged : Unchanged.Inputs[Union All Input 1]
2014-02-06 09:39:05.000,A component has finished processing all of its rows. : 220 : Update ODS - Src - State : 225 : OLE DB Command Input : Update ODS - Src - State : Update ODS - Src - State.Inputs[OLE DB Command Input]
2014-02-06 09:39:05.000,A component has finished processing all of its rows. : 66 : Insert ODS - Src - State : 79 : OLE DB Destination Input : Insert ODS - Src - State : Insert ODS - Src - State.Inputs[OLE DB Destination Input]
2014-02-06 09:39:05.000,Post Execute phase is beginning.
2014-02-06 09:39:05.000,Insert ODS - Src - State wrote 0 rows.
2014-02-06 09:39:05.000,Cleanup phase is beginning.
2014-02-06 09:39:05.000,End of package execution.
2014-02-06 09:39:05.000,
2014-02-06 09:39:05.000,"<?xml version=""1.0""?><DTS:ProcessMemoryUsage xmlns:DTS=""www.microsoft.com/SqlServer/Dts""><DTS:PeakWorkingSetSize>66457600</DTS:PeakWorkingSetSize></DTS:ProcessMemoryUsage>"
2014-02-06 09:39:05.000,End of package execution.
Look at the times. It's going up in exact 5sec increments! The package takes about 3 sec to run when the master package is executed via dtexecUI outside SSDT. The log looks "normal" in those instances as well; the end times are more randomised
This is with SP1 CU8 installed.
Jakub @ Adelaide, Australia

Similar Messages

  • Execute package task to call external package

    Hi all,
    I'm developing a SSIS project, I need an "execute package task" to execute a packahe stored in the SSISDB catalog of a SQL Server 2014 Instance.
    If I choose "External Reference" I can choose from "File System" and "SQL Server" but if I point to the last one I can't see package stored in the SSISDB, it seems to me that the new SSISDB catalog is not supported in Visual
    Stusio (I'm using VS 2013).
    Hey Microsoft developers, when you are at the coffe break, please talk each other!!! 

    Hi Andrea,
    As per my understanding, we can store packages to SSISDB database from SQL Server 2012. But we can also store package to MSDB database from SQL Server 2012. This is not a deprecated feature.
    But the Execute Package task can run a child package from the project with Project Reference, can also run packages stored in the SQL Server MSDB database and packages stored in the file system with External Reference. There is no such feature that
    support run the package from SSISDB database in the Execute Package task. For more details, please see:
    https://technet.microsoft.com/en-us/library/ms137609(v=sql.110).aspx
    Besides, if you want to also select a package from SSISDB database, I recommend you that submit the requirement at
    https://connect.microsoft.com/SQLServer/.  If the requirement mentioned by customers for many times, the product team may consider to add this feature in the next SQL Server version. Your feedback
    is valuable for us to improve our products and increase the level of service provided.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Failed to load Execute Package Task error.

    Hello Experts,
    Excuse my newbie question.
    Execute Package Task: Error: The task has failed to load. The contact information for the task is "Microsoft Corporation; Microsoft SQL Server; Microsoft corporation; All Rights Reserved http://www.microsoft.com/sql/support/default.asp:1"
    We have a windows service running on our app server that deploys a package that executes other packages using execute package task. This package and rest of the packages are in the same folder at the same level under integration services
    catalog. The package fails to execute with the above said error.
    Execute package task is configured to run in 'in-process' mode. Changing delay validation to true/false in child packages didn't help.
    ironically if the package is deployed using a console application from the app server the package gets executed just fine. I checked permissions, everything seem fine.
    Is there anyway to get details of this error to understand the problem better? I am looking at reports > standard reports > all executions.
    Thanks.

    As it turns out, the 63bit ExecPackageTask.dll was not found under "C:\Program Files\Microsoft SQL Server\110\DTS\Binn". Installing "SQL Server Data Tools" installed only 32 bit version of the component under "C:\Program Files (x86)\Microsoft SQL
    Server\110\DTS\Binn".
    Is there a way to get 64bit version of this component without installing "Integration Services" (to avoid possible license issues)?
    Hi Jayakarthiks,
    To make use of the 64-bit ExecPackageTask.dll, you also need the 64-bit SSIS runtime and 64-bit DTExec utility tool. To obtain the 64-bit SSIS runtime and tools, you have to install the SQL Server Integration Services service from your SQL Server 2012 install
    media.
    Reference:
    http://msdn.microsoft.com/en-us/library/ms143731.aspx
    Regards,
    Mike Yin
    If you have any feedback on our support, please click here
    Mike Yin
    TechNet Community Support

  • Execute Package Task Not Working After New Installation VS2012

    After installing VS2012 Update 4, the execute package task runs the child package out of order.
    The same version of VS is being used as the old computer (same everything, SQL 2012, OS...).  The solution works correctly on machines where VS2012 was installed prior to July 2014.  I've downloaded the BIDS, SSDT.
    Anyone else have the same problem?  I have yet to find anywhere on the internet where someone has reported the same issue.  It has happened on multiple computers since August 2014.

    How to reproduce?
    It is not very likely, but did you try searching MS Connect for this issue?
    Arthur
    MyBlog
    Twitter

  • Error: The Task has failed to load on Execute Package Task using TFS Build and MSBuild.exe on dedicated Build Machine.

    I'm trying to automate my building and deployment of SSIS packages using TFS build definitionss in conjunction with MSBuild.exe on a separate build machine.
    I have successfully built and deployed my projects but receive the following error:
    Error: Failed to load task "etl_UWP_Stage_tstg_7_Large",type "".  The contact information for this task is "Microsoft Corporation: Microsoft SQL Server .... "
    This error happens at runtime as well as if I perform a "Validate..." on the SSIS Catalog on the server.
    This only occurs when trying to execute a package that contains an Execute Package Task.  Any other packages that are doing the actual work are working just fine.
    In addition, when I do the MSBuild from my developer machine with the same deployment process everything works fine, even the Execute Package Task.
    I can even open up the project on the build machine in VS2012, rebuild all, then deploy and everything works fine.
    Some things I've considered/verified:
    1) Visual Studio 2012 and Visual Studio Data Tools 2012 are both on the build machine and my local machine.
    2) The .NET version being used by VS2012 is the same on the build machine and my local machine.
    3) I find it suspicious that the "Type" referenced in the error above is blank...but not sure where to go with that.
    4) Obvious difference is that during the TFS Build and Deploy on the build machine, everything is running under the TFS Service account, rather than my account...once again...not sure where to go with this, but it is a significant difference.
    Has anyone had similar issues or have any thoughts on things to try?
    Thanks,
    Jason

    Not sure why I couldn't find this answer in my earlier searches...
    But this thread solved it for me:
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/3aa1e324-da32-4ea0-865c-c69f610ef262/failed-to-load-execute-package-task-error?forum=sqlintegrationservices
    Essentially I had to install SSIS on the build machine so that the 64-bit version for
    ExecPackageTask.dll, was available.  I did this from the original media CD selection Integration Services as a feature.
    Hope this helps someone in the future.

  • The process cannot access the file because it is being used by another process with Execute package task

    Hi,
    I've a master package that calls other packages with an Execute Package Task. Sometimes we have an error: "The process cannot access the file because it is being used by another process" and sometimes not. It seems random.
    We are working on a Terminal Server and the SQL Server database engine and the files are placed on another server. It seems that the errors doesn't occu when we run the packages on the server with a job. We can't log onto the windows server on this machine..
    Hennie

    I've seen this myself. On most occasions an immediate rerun would fix the issue. As stated this happens only when we try to run this from BIDs. From SQL agent job it always runs fine. 
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Execute Package Task

    1. I have a CSV file which contains (  PackageLocationWithName, source info (table/file/accesdb) & other control info) 
    2. I load the CSV file into the SQL Server Table ConfigurationTab
    3. In the SSIS package I assign the values returned from select * from ConfigurationTab to
    a system object : SubjectStagelist ( system.object) Result Set :  ResultName: 0 ..1..2.. , Variable : user:: SubjectStagelist
    4. In the ForEachLoopContainer I use : collection : For Each ADO Enumerator
                                                              ADO Object Source Variable : user:: SubjectStagelist
                                                              Enumeration Mode : Rows in the first table
    5. Created  variables equals to number of column in the table(csv file):  
    6. Assigned these variables as index 0,1,2... in Variable Mappings in the ForEachLoopContainer ( Assuming that the values from the SubjectStagelist will be assigned to the new variables )
    7. Under the ForEachLoopContainer I have the execute process task. Here I want to n number of child packages to perform the data load from source to target ( n = no of rec in csv file which is fixed)
    In the execute package task I use the location as File, point to one of the files as connection and under expressions  - I have connection as @user::PackagePath (this variable will hold the complete package path with name e.g c\desktop\....\pkgName.dtsx).
    When I try so save the master package in the execute package task, I use the location as file which
    automatically changed to SQLSERVER from file and also throws the error by saying
    The connection "c\desktop\....\pkgName.dtsx"
    is not found. This error is thrown by Connections collection when the specific connection element is not found.
    "@user::PackagePath" on property "Connection" cannot be written to the property. The expression was evaluated, but cannot be set on the property.  
    But it keeps on giving me this  error ? Am I following the right steps ? 
    Please advice....
    Neil

    Hi Visakh,
    Now I am getting the below error.  In SQL table I have the data type as datetime for
    Date_Endtime(in Pkg it is datetime) variable and Integer for
    Int_Batch_Nbr(in pkg it is Int32) variable
    How can I typecast this if required, as I am capturing entire record into Object variable for further processing
    Error: 0xC001C012 at ForEach Loop: ForEach Variable Mapping number 10 to variable "User::Date_Endtime" cannot be applied.
    Error: 0xC001F009 at MSTR_Pkg: The type of the value being assigned to variable "User::Int_Batch_Nbr" 
    differs from the current variable type. Variables may not change type during execution. Variable types are strict, except for variables of type Object.
    Neil
    I think the issue is you've variables jumbled up in the mapping within for each loop. You need to make sure the index mapped to Date_EndTime corresponds to resultset column with datetime values and that to Int_Batch_Nbr corresponds to integer values. As
    of now its not correct which is why it complains about data type mismatch.
    Also didnt understand why you unmarked original answer as it was the solution to your actually posted problem. Now what you're asking is an unrelated issue. Anyone looking at this would have benefited with previous answer if their original issue was with
    execute package task property expression.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Execute package task is not able to pickup the package to execute it from a Foreach Loop Container

    Hi,
    I want to execute 5 packages using EXECUTE PACKAGE Task in FOR EACH LOOP Container. I followed the instructions as per the below link
    http://microsoft-ssis.blogspot.in/2013/01/master-child-packages-part-1-file-based.html
    When I execute the main package it is going to folder where i have that 5 packages and picking up the first one then it is not completing the package keep on executing the same.
    What i need to check here, kindly help me here.
    Regards,
    Sridhar V
    Sridhar

    Add a dummy execute sql task as first step inside loop and add a breakpoint to that for OnPreExecute event and check value of variable during each iteration to see if it gets value of each child package witin folder correctly.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Dynamic Execute package Task

    Hi ETL,
    I have Master package consisting of 10 child packages, i'm calling child packages using execute package task
    Now the issue is, child package location should be configurable .
    Plz help me

    Thats easy.
    Add the connection string property of package connection manager used within Execute Package Task as a configuration within package and pass it through XML file or SQL server table value
    http://biatlink.wordpress.com/2013/11/22/ssis-dynamic-location-for-execute-package-task/
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • HELP: SSIS 2012 - Execute SQL Task with resultset to populate user variables produces DBNull error

    I am experiencing an unexplainable behavior with the Execute SQL Task control in SSIS 2012. The following is a description of how to simulate
    the issue I will attempt to describe:
    1. Create a package and add two variables User::varTest1 and User::varTest2 both of String type with default value of "Select GetDate()"
    for each, have shown this to not matter as the same behavior occurs when using a fixed string value or empty string value.
    2. Add a new Execute SQL Task to the control flow.
    3. Configure an OLE DB connection.
    4. Set SQLSourceType = "Direct Input"
    5. Set the ResultSet property of the task to "Single Row"
    6. In the ResultSet tab add two results as follows: 
    Result Name: returnvalue1, Variable Name: User::varTest1
    Result Name: returnvalue2, Variable Name: User::varTest2
    7. Set an expression for the SqlStatementSource property with a string value of "Select 'Test' returnvalue1, 'Testing' returnvalue2'"
    The idea is that the source would be dynamically set in order to run a t-sql statement which would have dynamic values for database name
    or object that would be created at runtime and then executed to set the user variable values from its resultset. Instead what occurs is that a DBNull error occurs.
    I am not sure if anyone else has experienced this behavior performing similiar actions with the Execute SQL Task or not. Any help would be
    appreciated. The exact message is as follows:
    [Execute SQL Task] Error: An error occurred while assigning a value to variable "varRestoreScript": "The type of the value
    (DBNull) being assigned to variable "User::varRestoreScript" differs from the current variable type (String). Variables may not change type during execution. Variable types are strict, except for variables of type Object.
    User::varRestoreScript is the first return value. And even with the a dummy select the same result occurs. I have narrowed the issue down
    to the T-SQL Statement structure itself. 
    The following works just fine within the execute sql task control as long as no resultset is configured for return:
    "Declare @dynamicSQL nvarchar(max)
    Select @dynamicSQL = name 
    From sys.databases 
    Where name = 'master'
    Select atest_var1=@dynamicSQL, atest_var2='static'"
    I have tried various iterations of the script above with no success. However, if I use the following derivative of it the task completes
    successfully and variables are set as expected.  This will not work for my scenario however as I need to dynamically build a string spanning multiple resultsets to build one of the variable values.
    "Select atest_var1=name, atest_var2='static'
    From sys.databases
    Where name = 'master'
    I have a sample package which can reproduce this issue using the above code.  You can get to that through the post on www.sqlservercentral.com/Forums/Topic1582670-364-1.aspx
    Scott

    Arthur,  the query when executed doesn't return a null value for the @dynamicSQL variable.  It returns "master" as a string value.  Even the following fails which implements that suggestion:
    Declare @dynamicSQL as nvarchar(max)
    Select @dynamicSQL = name
    From master.sys.databases 
    Where name = 'master' -- (or any other DB name)
    I believe I have found the cause of the issue.  It is datatype and size related.  The above script will properly set the variables in the resultset as long as you don't use nvarchar(max) or varchar(max).  This makes the maximum data size for
    a String variable 4000 characters whether unicode or non-unicode typed.

  • Time Capsule external drive very slow for backup

    I am experiencing very, very slow backup times using an external drive attached to my Time Capsule. When I recently added a 3 TB Rocstor external drive to my mid-2011 iMac using FireWire, I included it in the drives to be backed up. The internal 2 TB of storage in my Time Capsule was no longer enough, so I plugged a 4 TB WD My Book Studio drive into the Time Capsule's USB 2.0 port and specified this WD drive as the only backup storage. The first backup took almost 48 hours. Now the Time Machine keeps trying to back up about 400 GB - 500 GB (though there haven't been any changes) and is transferring about 400 KB/sec over Ethernet from my iMac to the Time Capsule's external drive. What am I doing wrong? (I recently upgraded the iMac to OS X 10.10.2 Yosemite.)

    Frankly, I have no way to gauge your network setup, so I can't really tell you what you should expect. I can tell you from my own experiences using a local network based on a 3rd Gen. Gigabit Ethernet AEBS that using Ethernet the average throughput (not interface) was in the neighborhood of 16-20 MBs/sec. The average throughput for a USB 2.0 drive connected directly was about the same.
    This works out to an average throughput of around 60 GBs/hour. Depending upon file sizes and number of files being transferred a USB 2.0 drive might reach 100 GBs/hour, but that is a rare outcome. The USB 2.0 drive interface speed is about the same as FireWire 400 - 400 Mb/s or roughly 50 MBs/sec. This is what I know from measurement. Do not expect Ethernet to be any faster for throughput.

  • Windows Server 2012 Storage Spaces Simple RAID 0 VERY SLOW reads, but fast writes with LSI 9207-8e SAS JBOD HBA Controller

    Has anyone else seen Windows Server 2012 Storage Spaces with a Simple RAID 0 (also happens with Mirrored RAID 1 and Parity RAID 5) virtual disk exhibiting extremely slow read speed of 5Mb/sec, yet write performance is normal at 650Mb/sec in RAID 0?
    Windows Server 2012 Standard
    Intel i7 CPU and Motherboard
    LSI 9207-8e 6Gb SAS JBOD Controller with latest firmware/BIOS and Windows driver.
    (4) Hitachi 4TB 6Gb SATA Enterprise Hard Disk Drives HUS724040ALE640
    (4) Hitachi 4TB 6Gb SATA Desktop Hard Disk Drives HDS724040ALE640
    Hitachi drives are directly connected to LSI 9207-8e using a 2-meter SAS SFF-8088 to eSATA cable to six-inch eSATA/SATA adapter.
    The Enterprise drives are on LSI's compatibility list.  The Desktop drives are not, but regardless, both drive models are affected by the problem.
    Interestingly, this entire configuration but with two SIIG eSATA 2-Port adapters instead of the LSI 9207-8e, works perfectly with both reads and writes at 670Mb/sec.
    I thought SAS was going to be a sure bet for expanding beyond the capacity of port limited eSATA adapters, but after a week of frustration and spending over $5,000.00 on drives, controllers and cabling, it's time to ask for help!
    Any similar experiences or solutions?

    Has anyone else seen Windows Server 2012 Storage Spaces with a Simple RAID 0 (also happens with Mirrored RAID 1 and Parity RAID 5) virtual disk exhibiting extremely slow read speed of 5Mb/sec, yet write performance is normal at 650Mb/sec in RAID 0?
    Windows Server 2012 Standard
    Intel i7 CPU and Motherboard
    LSI 9207-8e 6Gb SAS JBOD Controller with latest firmware/BIOS and Windows driver.
    (4) Hitachi 4TB 6Gb SATA Enterprise Hard Disk Drives HUS724040ALE640
    (4) Hitachi 4TB 6Gb SATA Desktop Hard Disk Drives HDS724040ALE640
    Hitachi drives are directly connected to LSI 9207-8e using a 2-meter SAS SFF-8088 to eSATA cable to six-inch eSATA/SATA adapter.
    The Enterprise drives are on LSI's compatibility list.  The Desktop drives are not, but regardless, both drive models are affected by the problem.
    Interestingly, this entire configuration but with two SIIG eSATA 2-Port adapters instead of the LSI 9207-8e, works perfectly with both reads and writes at 670Mb/sec.
    I thought SAS was going to be a sure bet for expanding beyond the capacity of port limited eSATA adapters, but after a week of frustration and spending over $5,000.00 on drives, controllers and cabling, it's time to ask for help!
    Any similar experiences or solutions?
    1) Yes, being slow either on reads or on writes is a quite common situation for storage spaces. See references (with some of the solutions I hope):
    http://social.technet.microsoft.com/Forums/en-US/winserverfiles/thread/a58f8fce-de45-4032-a3ef-f825ee39b96e/
    http://blogs.technet.com/b/askpfeplat/archive/2012/10/10/windows-server-2012-storage-spaces-is-it-for-you-could-be.aspx
    http://social.technet.microsoft.com/Forums/en-US/winserver8gen/thread/64aff15f-2e34-40c6-a873-2e0da5a355d2/
    and this one is my favorite putting a lot of light on the issue:
    http://helgeklein.com/blog/2012/03/windows-8-storage-spaces-bugs-and-design-flaws/
    2) Issues with SATA-to-SAS hardware is also very common. See:
    http://social.technet.microsoft.com/Forums/en-US/winserverClustering/thread/5d4f68b7-5fc4-4a3c-8232-a2a68bf3e6d2
    StarWind iSCSI SAN & NAS

  • Can SSIS 2012 execute packages built on 2008 R2 BI for 2008 R2 SQL Server?

    We have a query regarding SSIS instalattion on application server. As SQL Server 2008 R2 is getting out of life, DBAs are suggesting to install SSIS 2012. Now our concern is
    that our SQL Server DB Server is on 2008 R2, and developers are building packages using SSIS 2008. 
    Considering we deploy file packages, and backward compatibility, theoretically it shall work. There are 3 things here -
    1. {2008 R2 SSIS Package} ------ Deployed On -------> 2. {Server with 2012 SSIS installation} ------- Connects to ------->  3. {SQL Server 2008 R2 DB Instance}
    Any recommendation on this. Would it work? Anything which could cause issue?
    As we have to pay for license, do not want be in a situation where we pay for SSIS 2012 installation on server (#2 above) and then realize that there is some issue
    Regards, Amit Gupta

    Hi Amit,
    Just as Visakh and hsbal said, there will be no issues as SSIS 2012 will work with standard previous versions of SQL Server packages. But some custom SSIS Tasks can always throw error after upgrading packages from SQL Server 2008 R2 to SQL Server 2012.
    About this, please remember to do the following things:
    Change your custom tasks and components to target the .NET 4.0 Framework in each of the project properties, in the Application Tab, under Target framework.
    Update the assembly references in each of your projects to point to the appropriate SqlServer dlls.
    Build and copy your project dlls to the appropriate directory, depending on what kind of components you are updating, i.e. Program Files (x86)\Microsoft SQL Server\110\DTS\PiplelineComponents
    Register your components in the GAC.
    Close Visual Studio and then reopen; open your SSIS packages and refresh the SSIS Toolbox, by right clicking in the tool box and selected Refresh Toolbox.
    Besides, if some of your script task throw errors to you, we can open the script task in VSTA, insert Breakpoint to debug the scripts, then find the root cause. For more details, please see:
    Debug a Script by Setting Breakpoints in a Script Task and Script Component
    For more detailed information regarding to the license issue, please call 1-800-426-9400, Monday through Friday, 6:00 A.M. to 6:00 P.M. (Pacific Time) to speak directly to a Microsoft licensing specialist. For international customers, please use the Guide
    to Worldwide Microsoft Licensing Sites to find contact information in your locations.
    You can also visit the following site for more information and support on licensing issues:
    http://www.microsoft.com/licensing/mla/default.aspx
    Thanks,
    Katherine Xiong
    If you have any feedback on our support, please click
    here.
    Katherine Xiong
    TechNet Community Support

  • Configuration Manager 2012 SP1 CU4 Console - Content Status very slow

    I have a SCCM 2012 SP1 CU4 environment running.
    When using the console and viewing the Monitoring / Distribution Status / Content Status it took about 10-15 minutes to get the information.
    How can this be speed up a bit ? By adding additional indexes in SQL ?
    Any tips, hints are very welcome.

    Hi,
    Have you seen this previuous thread on the topic?
    https://social.technet.microsoft.com/Forums/en-US/7935143e-21a8-4770-83bf-7e1938d9defd/configuration-manager-2012-content-status-slow?forum=configmanagergeneral
    I would run the script from Ola that reports back on the indexe defragementeation and what I normally see as well is that more than one TempDB file increases performance greatly. It is most likely covered in the session mentioned above by Tim.
    Regards,
    Jörgen
    -- My System Center blog ccmexec.com -- Twitter
    @ccmexec

  • Microsoft SQL Server 2012 Management Studio Express silent installation very slow

    Iam doing silent installations often on several Systems.
    I install the Microsoft SQL Server 2012 Management Studio Express with this command:
    "SQLManagementStudio_x64_DEU.exe /ACTION=INSTALL /QUIET /IAcceptSQLServerLicenseTerms="True" /FEATURES=CONN,BC,SSMS /indicateprogress"
    The installation will take almost an hour to finish. 
    I never get error messages or so. The installation afterwards is fine.
    Is there a way to speed up the installation process?
    Thanks!

    Hi Dennis,
    When installing SQL Server Management Studio, it is normal for us taking an hour to finish the installation.
     Personally, if you want to speed up the installation process, I recommend improve your hardware device on your computer, such as CPU, RAM, hard disk and so on.
    Or you can submit an submit the requirement at https://connect.microsoft.com/SQLServer/. 
    If the requirement mentioned by customers for many times, the product team may consider to add or improve this requirement in the future SQL Server version. Thanks for your understanding.
    Thanks,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

Maybe you are looking for

  • LIST_TOO_MANY_LPROS

    Hi .          I am getting this error. pls send me solution Runtime Errors         LIST_TOO_MANY_LPROS Date and Time          22.07.2008 17:43:27 Short text No further list processing possible. What happened? You requested too many consecutive nested

  • Updating large pdf file on multi devices but do I have to download the entire file every time?

    So I have a pdf file that is 500mb (already on acrobat.com cloud) and I want to read this across tablet, desktop, and phone. I will be updating this file bit by bit on every one of those devices. Does this mean that every time I want to see the updat

  • B85-G43 overclocking i5-4570K ?

    What is the situation with overclocking (by adjusting multiplier) "K"-line CPUs on B85-G43 ? Is it possible ? Does it depend on BIOS version or CPU's production date ? I want to buy B85-G43 with i3-4130 but change it in the future to some i5-4...K Th

  • Flicking colors !!

    Hello, I have a concern that takes my mind for several days I made a 3D animation in 3DSMax, I recorded in sequence TGA. So imported into After Effects for post-production, and there, horror, watching the footage, I see that the colors "flash" in pla

  • How good is the built in backup?

    We are going to upgrade or server to Windows Server 2012 Essentials. Today we are using Symantec Backup Exec. Will the built in backup in WS be enough? We will be using MS SQL Server and it's crucial that that data will be backed up properly.