Execute package log status customization

Hi gurus!
I would like to customize the result of a package execution in BPC.
From the result status (set it to error based on my own parameters) to  append lines from external BADIS result or delete lines from standart log.
¿Is it posible? ¿Any idea on how to achieve this?
Thanks a lot in advance.
it0

First of all BAdI call has a return table parameter ET_MESSAGE. If you put your messages in that table they will be appended to the log.
If you want process whatever you have in the log try accessing Class cl_ujk_logger. All important Methods there are static, so you will be able to access the log and make allowed operations on it, like GET_LOG, CLEAR, etc.
Hope this helps,
Gersh

Similar Messages

  • Package Log - Error checking work status

    Hi, Experts,
    We are using SAP BPC 10 NW.
    When we run a package manually it is ok. But when we include it at package link (with the same prompt) and run it, we received error message (Error checking work status). Look at the log below:
    ++++++++++++++++++++++++++++++++++++++++++++++
    /CPMB/MODIFY finished in 0 seconds
    /CPMB/ALLOCATION_LOGIC finished in 449 seconds
    /CPMB/CLEAR finished in 0 seconds
    [Selection]
    SELECTION = /CEREAL/EXPENSE/PRIVATEPUBLICATIONS/BPCADMINBIP/TempFiles/FROM.TMP@@@SAVE@@@@@@EXPAND@@@|DIMENSION:CATEGORY|PROJETADO,REVISAO|DIMENSION:ENTITY|CC_11
    [Message]
    RUN_LOGIC:
    Error checking work status
    +++++++++++++++++++++++++++++++++++++++++++++++
    There is no work status activity at REVISAO and PROJETADO category members at this model.
    Would anybody help us to understand what is happening?
    Best Regards,
    Ana Teresa

    Hi Ana
    I found a related note.
    1945016 - "Error checking work status" when executing package
    Symptom
    There is "Error checking work status" in the status log after executing clear package:/CPMB/CLEAR.
    Environment
    SAP BusinessObjects Planning and Consolidation, version 10 for SAP NetWeaver
    Reproducing the Issue
    Run clear package:/CPMB/CLEAR
    Set check work status when clearing as YES. ("CHECKLCK"=YES)
    Check detail status log
    CauseWork status is not set or not found for the "to be cleared" data.
    Resolution
    Try to change the work status of the data in EPM and check if system pops up with "work status not found".
    If yes, check in work status table (you can get the table name by running program UJ0_GET_GEN_TABNAME) if there is default work status assigned.
    If yes, check UJW_STATCODE table in SE16, incase the line which contain field ORDERN=0 is missing, run program: UJW_UP_INSERT_0_STATE to insert this line.
    See Also
    1862768 - EPMWorkStatus function displays wrong status
    Hope this helps.
    Regards
    Juan Pablo

  • The process cannot access the file because it is being used by another process with Execute package task

    Hi,
    I've a master package that calls other packages with an Execute Package Task. Sometimes we have an error: "The process cannot access the file because it is being used by another process" and sometimes not. It seems random.
    We are working on a Terminal Server and the SQL Server database engine and the files are placed on another server. It seems that the errors doesn't occu when we run the packages on the server with a job. We can't log onto the windows server on this machine..
    Hennie

    I've seen this myself. On most occasions an immediate rerun would fix the issue. As stated this happens only when we try to run this from BIDs. From SQL agent job it always runs fine. 
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • 2012: Execute Package Task (external reference) Very Slow in SSDT

    Hi All,
    I have 4 packages which I've deployed using the package deployment model to the MSDB package store.
    Configuration setup is System Env Var pointing at SSIS Configurations table that populates the connection string of my SSISConfig data source. Source/Dest configurations are in the SSIS Configuration table.
    I have a  'master' package that executes the 4 packages in parallel using the Execute Package control flow task (external reference, SSIS Package Store)
    When I execute this master package in SSDT it takes 1minute 56 seconds to finish. All the execute package tasks seem to hang until the last few sec then finish one after the other. It's the same if execute one of the Execute Package Task items individually.
    It just hangs there for nearly 2 min then rushes through its work (i've verified by looking at the destination tables; nothing happens until the end).
    But the packages executed directly in ssdt take 2-3 sec to execute.
    I've deployed the master package to the server and executed it via SSMS. It takes <3 sec.
    I've RDPed into the ssis box and it's the same. Executing the packages individually or the master package in dtexecui takes less than 3 sec.
    Something's going wrong inside SSDT (some kind of timeout? but there are no error msgs or warning) that's making it take nearly 2 minutes.
    Have others experienced this? Have I misconfigured something? The env var has been working on my local machine since last week and i've restarted many times since then, so I'm positive SSDT is using the 'current' version of the env var.
    Jakub @ Adelaide, Australia

    Very strange. The sysssislogs are the same for any method of execution, but there's a 5 sec delay for each step when executing via SSDT
    endtime,message
    2014-02-06 09:37:15.000,Execute phase is beginning.
    2014-02-06 09:37:20.000,PrimeOutput will be called on a component. : 104 : Merge Join : Merge Join
    2014-02-06 09:37:20.000,PrimeOutput will be called on a component. : 213 : Unchanged : Unchanged
    2014-02-06 09:37:25.000,A component has returned from its PrimeOutput call. : 213 : Unchanged : Unchanged
    2014-02-06 09:37:25.000,A component has returned from its PrimeOutput call. : 104 : Merge Join : Merge Join
    2014-02-06 09:37:30.000,PrimeOutput will be called on a component. : 143 : ODS - Src - State : ODS - Src - State
    2014-02-06 09:37:30.000,PrimeOutput will be called on a component. : 17 : Src - dbo - State : Src - dbo - State
    2014-02-06 09:37:35.000,A component has returned from its PrimeOutput call. : 143 : ODS - Src - State : ODS - Src - State
    2014-02-06 09:37:35.000,A component has returned from its PrimeOutput call. : 17 : Src - dbo - State : Src - dbo - State
    2014-02-06 09:37:35.000,Rows were provided to a data flow component as input. : : 264 : OLE DB Source Output : 182 : Source Checksum : 185 : ChecksumInput : 7 : Source Checksum : Paths[Src - dbo - State.OLE DB Source Output] : Source Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:35.000,Rows were provided to a data flow component as input. : : 263 : OLE DB Source Output : 53 : Destination Checksum : 56 : ChecksumInput : 7 : Destination Checksum : Paths[ODS - Src - State.OLE DB Source Output] : Destination Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:40.000,A component will be given the end of rowset signal. : 53 : Destination Checksum : 56 : ChecksumInput : Destination Checksum : Destination Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:40.000,Rows were provided to a data flow component as input. : : 258 : ChecksumOutput : 104 : Merge Join : 109 : Merge Join Left Input : 7 : Merge Join : Paths[Source Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Left Input]
    2014-02-06 09:37:40.000,A component will be given the end of rowset signal. : 182 : Source Checksum : 185 : ChecksumInput : Source Checksum : Source Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:45.000,Rows were provided to a data flow component as input. : : 259 : ChecksumOutput : 104 : Merge Join : 118 : Merge Join Right Input : 7 : Merge Join : Paths[Destination Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Right Input]
    2014-02-06 09:37:45.000,Rows were provided to a data flow component as input. : : 264 : OLE DB Source Output : 182 : Source Checksum : 185 : ChecksumInput : 0 : Source Checksum : Paths[Src - dbo - State.OLE DB Source Output] : Source Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:45.000,Rows were provided to a data flow component as input. : : 263 : OLE DB Source Output : 53 : Destination Checksum : 56 : ChecksumInput : 0 : Destination Checksum : Paths[ODS - Src - State.OLE DB Source Output] : Destination Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:50.000,A component has finished processing all of its rows. : 53 : Destination Checksum : 56 : ChecksumInput : Destination Checksum : Destination Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:50.000,A component has finished processing all of its rows. : 182 : Source Checksum : 185 : ChecksumInput : Source Checksum : Source Checksum.Inputs[ChecksumInput]
    2014-02-06 09:37:55.000,A component will be given the end of rowset signal. : 104 : Merge Join : 118 : Merge Join Right Input : Merge Join : Merge Join.Inputs[Merge Join Right Input]
    2014-02-06 09:38:00.000,Rows were provided to a data flow component as input. : : 259 : ChecksumOutput : 104 : Merge Join : 118 : Merge Join Right Input : 0 : Merge Join : Paths[Destination Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Right Input]
    2014-02-06 09:38:05.000,A component has finished processing all of its rows. : 104 : Merge Join : 118 : Merge Join Right Input : Merge Join : Merge Join.Inputs[Merge Join Right Input]
    2014-02-06 09:38:10.000,A component will be given the end of rowset signal. : 104 : Merge Join : 109 : Merge Join Left Input : Merge Join : Merge Join.Inputs[Merge Join Left Input]
    2014-02-06 09:38:15.000,Rows were provided to a data flow component as input. : : 258 : ChecksumOutput : 104 : Merge Join : 109 : Merge Join Left Input : 0 : Merge Join : Paths[Source Checksum.ChecksumOutput] : Merge Join.Inputs[Merge Join Left Input]
    2014-02-06 09:38:20.000,A component has finished processing all of its rows. : 104 : Merge Join : 109 : Merge Join Left Input : Merge Join : Merge Join.Inputs[Merge Join Left Input]
    2014-02-06 09:38:20.000,Rows were provided to a data flow component as input. : : 261 : Merge Join Output : 2 : Add audit columns : 3 : Derived Column Input : 7 : Add audit columns : Paths[Merge Join.Merge Join Output] : Add audit columns.Inputs[Derived Column Input]
    2014-02-06 09:38:25.000,A component will be given the end of rowset signal. : 2 : Add audit columns : 3 : Derived Column Input : Add audit columns : Add audit columns.Inputs[Derived Column Input]
    2014-02-06 09:38:25.000,Rows were provided to a data flow component as input. : : 260 : Derived Column Output : 195 : Type 1 SCD : 196 : Conditional Split Input : 7 : Type 1 SCD : Paths[Add audit columns.Derived Column Output] : Type 1 SCD.Inputs[Conditional Split Input]
    2014-02-06 09:38:30.000,Rows were provided to a data flow component as input. : : 265 : Unchanged : 213 : Unchanged : 214 : Union All Input 1 : 7 : Unchanged : Paths[Type 1 SCD.Unchanged] : Unchanged.Inputs[Union All Input 1]
    2014-02-06 09:38:30.000,Rows were provided to a data flow component as input. : : 261 : Merge Join Output : 2 : Add audit columns : 3 : Derived Column Input : 0 : Add audit columns : Paths[Merge Join.Merge Join Output] : Add audit columns.Inputs[Derived Column Input]
    2014-02-06 09:38:35.000,A component has finished processing all of its rows. : 2 : Add audit columns : 3 : Derived Column Input : Add audit columns : Add audit columns.Inputs[Derived Column Input]
    2014-02-06 09:38:40.000,A component will be given the end of rowset signal. : 195 : Type 1 SCD : 196 : Conditional Split Input : Type 1 SCD : Type 1 SCD.Inputs[Conditional Split Input]
    2014-02-06 09:38:45.000,Rows were provided to a data flow component as input. : : 260 : Derived Column Output : 195 : Type 1 SCD : 196 : Conditional Split Input : 0 : Type 1 SCD : Paths[Add audit columns.Derived Column Output] : Type 1 SCD.Inputs[Conditional Split Input]
    2014-02-06 09:38:50.000,A component has finished processing all of its rows. : 195 : Type 1 SCD : 196 : Conditional Split Input : Type 1 SCD : Type 1 SCD.Inputs[Conditional Split Input]
    2014-02-06 09:38:55.000,A component will be given the end of rowset signal. : 66 : Insert ODS - Src - State : 79 : OLE DB Destination Input : Insert ODS - Src - State : Insert ODS - Src - State.Inputs[OLE DB Destination Input]
    2014-02-06 09:38:55.000,A component will be given the end of rowset signal. : 220 : Update ODS - Src - State : 225 : OLE DB Command Input : Update ODS - Src - State : Update ODS - Src - State.Inputs[OLE DB Command Input]
    2014-02-06 09:38:55.000,A component will be given the end of rowset signal. : 213 : Unchanged : 214 : Union All Input 1 : Unchanged : Unchanged.Inputs[Union All Input 1]
    2014-02-06 09:39:00.000,Rows were provided to a data flow component as input. : : 262 : New : 66 : Insert ODS - Src - State : 79 : OLE DB Destination Input : 0 : Insert ODS - Src - State : Paths[Type 1 SCD.New] : Insert ODS - Src - State.Inputs[OLE DB Destination Input]
    2014-02-06 09:39:00.000,Rows were provided to a data flow component as input. : : 257 : Changed : 220 : Update ODS - Src - State : 225 : OLE DB Command Input : 0 : Update ODS - Src - State : Paths[Type 1 SCD.Changed] : Update ODS - Src - State.Inputs[OLE DB Command Input]
    2014-02-06 09:39:00.000,"The final commit for the data insertion in ""Insert ODS - Src - State"" has started."
    2014-02-06 09:39:00.000,Rows were provided to a data flow component as input. : : 265 : Unchanged : 213 : Unchanged : 214 : Union All Input 1 : 0 : Unchanged : Paths[Type 1 SCD.Unchanged] : Unchanged.Inputs[Union All Input 1]
    2014-02-06 09:39:00.000,"The final commit for the data insertion in ""Insert ODS - Src - State"" has ended."
    2014-02-06 09:39:05.000,A component has finished processing all of its rows. : 213 : Unchanged : 214 : Union All Input 1 : Unchanged : Unchanged.Inputs[Union All Input 1]
    2014-02-06 09:39:05.000,A component has finished processing all of its rows. : 220 : Update ODS - Src - State : 225 : OLE DB Command Input : Update ODS - Src - State : Update ODS - Src - State.Inputs[OLE DB Command Input]
    2014-02-06 09:39:05.000,A component has finished processing all of its rows. : 66 : Insert ODS - Src - State : 79 : OLE DB Destination Input : Insert ODS - Src - State : Insert ODS - Src - State.Inputs[OLE DB Destination Input]
    2014-02-06 09:39:05.000,Post Execute phase is beginning.
    2014-02-06 09:39:05.000,Insert ODS - Src - State wrote 0 rows.
    2014-02-06 09:39:05.000,Cleanup phase is beginning.
    2014-02-06 09:39:05.000,End of package execution.
    2014-02-06 09:39:05.000,
    2014-02-06 09:39:05.000,"<?xml version=""1.0""?><DTS:ProcessMemoryUsage xmlns:DTS=""www.microsoft.com/SqlServer/Dts""><DTS:PeakWorkingSetSize>66457600</DTS:PeakWorkingSetSize></DTS:ProcessMemoryUsage>"
    2014-02-06 09:39:05.000,End of package execution.
    Look at the times. It's going up in exact 5sec increments! The package takes about 3 sec to run when the master package is executed via dtexecUI outside SSDT. The log looks "normal" in those instances as well; the end times are more randomised
    This is with SP1 CU8 installed.
    Jakub @ Adelaide, Australia

  • How to write custom messages to Data Manager package log from within SSIS?

    Just wonder in BPC 7.0/7.5 MS how do you write custom messages to Data Manager package log from within SSIS?  I mean I want to log output of specific custom tasks (non-BPC) in SSIS control flow to the resultant BPC package log created when running SSIS package in Data Manager.  Can this be done in a Script Task or some other way?  Not much documentation out there on this.
    Thanks,
    Brian

    Hi Brian
    In order to achieve what you want, it can get tricky in your package, ultimately It would depend on the task, for example if you had an execute SQL task, you could use a RAISEERROR command in your SQL statement and BPC would return the error message that you specified.
    If you had other tasks, then it would be a bit more tricky, you would have to have custom messages based on event handlers.
    Please see below links for examples :
    [Custom messages for logging|http://msdn.microsoft.com/en-us/library/ms345174.aspx]
    [Custom Logging Using Event Handlers |http://consultingblogs.emc.com/jamiethomson/archive/2005/06/11/SSIS_3A00_-Custom-Logging-Using-Event-Handlers.aspx]
    Hope this helps
    Kind Regards
    Daniel

  • Data manager package log does not show Global variables!!

    Hello Experts,
    We are using BPC 10 sp14 Microsoft version with SQL Server 2008 R2. We are seeing an issue where datamanager package log does not show the Global variables defined in package script whereas in BPC 7.5 Global variable were visible in package log.
    Please let us know is this behavior changed in BPC 10?
    Below is the package script having Global variables and Package log not showing any of them.
    Thanks & Regards,
    Rohit
    Package Script:
    Package Log:

    Hi Ergin,
    As far as I remember it's by design...
    Vadim

  • Error: The Task has failed to load on Execute Package Task using TFS Build and MSBuild.exe on dedicated Build Machine.

    I'm trying to automate my building and deployment of SSIS packages using TFS build definitionss in conjunction with MSBuild.exe on a separate build machine.
    I have successfully built and deployed my projects but receive the following error:
    Error: Failed to load task "etl_UWP_Stage_tstg_7_Large",type "".  The contact information for this task is "Microsoft Corporation: Microsoft SQL Server .... "
    This error happens at runtime as well as if I perform a "Validate..." on the SSIS Catalog on the server.
    This only occurs when trying to execute a package that contains an Execute Package Task.  Any other packages that are doing the actual work are working just fine.
    In addition, when I do the MSBuild from my developer machine with the same deployment process everything works fine, even the Execute Package Task.
    I can even open up the project on the build machine in VS2012, rebuild all, then deploy and everything works fine.
    Some things I've considered/verified:
    1) Visual Studio 2012 and Visual Studio Data Tools 2012 are both on the build machine and my local machine.
    2) The .NET version being used by VS2012 is the same on the build machine and my local machine.
    3) I find it suspicious that the "Type" referenced in the error above is blank...but not sure where to go with that.
    4) Obvious difference is that during the TFS Build and Deploy on the build machine, everything is running under the TFS Service account, rather than my account...once again...not sure where to go with this, but it is a significant difference.
    Has anyone had similar issues or have any thoughts on things to try?
    Thanks,
    Jason

    Not sure why I couldn't find this answer in my earlier searches...
    But this thread solved it for me:
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/3aa1e324-da32-4ea0-865c-c69f610ef262/failed-to-load-execute-package-task-error?forum=sqlintegrationservices
    Essentially I had to install SSIS on the build machine so that the 64-bit version for
    ExecPackageTask.dll, was available.  I did this from the original media CD selection Integration Services as a feature.
    Hope this helps someone in the future.

  • Failed to load Execute Package Task error.

    Hello Experts,
    Excuse my newbie question.
    Execute Package Task: Error: The task has failed to load. The contact information for the task is "Microsoft Corporation; Microsoft SQL Server; Microsoft corporation; All Rights Reserved http://www.microsoft.com/sql/support/default.asp:1"
    We have a windows service running on our app server that deploys a package that executes other packages using execute package task. This package and rest of the packages are in the same folder at the same level under integration services
    catalog. The package fails to execute with the above said error.
    Execute package task is configured to run in 'in-process' mode. Changing delay validation to true/false in child packages didn't help.
    ironically if the package is deployed using a console application from the app server the package gets executed just fine. I checked permissions, everything seem fine.
    Is there anyway to get details of this error to understand the problem better? I am looking at reports > standard reports > all executions.
    Thanks.

    As it turns out, the 63bit ExecPackageTask.dll was not found under "C:\Program Files\Microsoft SQL Server\110\DTS\Binn". Installing "SQL Server Data Tools" installed only 32 bit version of the component under "C:\Program Files (x86)\Microsoft SQL
    Server\110\DTS\Binn".
    Is there a way to get 64bit version of this component without installing "Integration Services" (to avoid possible license issues)?
    Hi Jayakarthiks,
    To make use of the 64-bit ExecPackageTask.dll, you also need the 64-bit SSIS runtime and 64-bit DTExec utility tool. To obtain the 64-bit SSIS runtime and tools, you have to install the SQL Server Integration Services service from your SQL Server 2012 install
    media.
    Reference:
    http://msdn.microsoft.com/en-us/library/ms143731.aspx
    Regards,
    Mike Yin
    If you have any feedback on our support, please click here
    Mike Yin
    TechNet Community Support

  • Execute Package Task Not Working After New Installation VS2012

    After installing VS2012 Update 4, the execute package task runs the child package out of order.
    The same version of VS is being used as the old computer (same everything, SQL 2012, OS...).  The solution works correctly on machines where VS2012 was installed prior to July 2014.  I've downloaded the BIDS, SSDT.
    Anyone else have the same problem?  I have yet to find anywhere on the internet where someone has reported the same issue.  It has happened on multiple computers since August 2014.

    How to reproduce?
    It is not very likely, but did you try searching MS Connect for this issue?
    Arthur
    MyBlog
    Twitter

  • Error invoking Package Schedule Status in BPC 5.1

    Hi BPC Gurus,
    We are on BPC 5.1 system and have been live since Apr 2008. We schedule all our Data Manager jobs (Import packages and other packages) every month using the Package Schedule status option under eData.
    We were performing a clean-up of the older job entries on this screen by choosing the "Delete Job" option in Package Schedule Status screen.
    A few of them got successfully deleted but then we got the below error.
    "(RunQueryToServer)(GetScheduleList) An error occurred while getting a schedule list. The system cannot find the file specified."
    We are now getting this error even if we try to bring up the Package Schedule Status screen.
    Could you pls. advise if anybody has seen this issue and what was the resolution.
    We are seeing this issue in our Production environment so your help would be greatly appreciated.
    Thanks,
    Abhay Shanbhag

    We had the same problem with the installation I'm working on right now.
    In our case, I went into the SQL Server database for my application set and looked at the entries in tblSchedule. There were entries in there that were very strange in the "CreateTime" field. Dates like "2008-43-30 <time>".
    We cleared out the records in tblSchedule and this solved the problem.
    Hope this helps,
    Jason

  • Execute Package Task

    1. I have a CSV file which contains (  PackageLocationWithName, source info (table/file/accesdb) & other control info) 
    2. I load the CSV file into the SQL Server Table ConfigurationTab
    3. In the SSIS package I assign the values returned from select * from ConfigurationTab to
    a system object : SubjectStagelist ( system.object) Result Set :  ResultName: 0 ..1..2.. , Variable : user:: SubjectStagelist
    4. In the ForEachLoopContainer I use : collection : For Each ADO Enumerator
                                                              ADO Object Source Variable : user:: SubjectStagelist
                                                              Enumeration Mode : Rows in the first table
    5. Created  variables equals to number of column in the table(csv file):  
    6. Assigned these variables as index 0,1,2... in Variable Mappings in the ForEachLoopContainer ( Assuming that the values from the SubjectStagelist will be assigned to the new variables )
    7. Under the ForEachLoopContainer I have the execute process task. Here I want to n number of child packages to perform the data load from source to target ( n = no of rec in csv file which is fixed)
    In the execute package task I use the location as File, point to one of the files as connection and under expressions  - I have connection as @user::PackagePath (this variable will hold the complete package path with name e.g c\desktop\....\pkgName.dtsx).
    When I try so save the master package in the execute package task, I use the location as file which
    automatically changed to SQLSERVER from file and also throws the error by saying
    The connection "c\desktop\....\pkgName.dtsx"
    is not found. This error is thrown by Connections collection when the specific connection element is not found.
    "@user::PackagePath" on property "Connection" cannot be written to the property. The expression was evaluated, but cannot be set on the property.  
    But it keeps on giving me this  error ? Am I following the right steps ? 
    Please advice....
    Neil

    Hi Visakh,
    Now I am getting the below error.  In SQL table I have the data type as datetime for
    Date_Endtime(in Pkg it is datetime) variable and Integer for
    Int_Batch_Nbr(in pkg it is Int32) variable
    How can I typecast this if required, as I am capturing entire record into Object variable for further processing
    Error: 0xC001C012 at ForEach Loop: ForEach Variable Mapping number 10 to variable "User::Date_Endtime" cannot be applied.
    Error: 0xC001F009 at MSTR_Pkg: The type of the value being assigned to variable "User::Int_Batch_Nbr" 
    differs from the current variable type. Variables may not change type during execution. Variable types are strict, except for variables of type Object.
    Neil
    I think the issue is you've variables jumbled up in the mapping within for each loop. You need to make sure the index mapped to Date_EndTime corresponds to resultset column with datetime values and that to Int_Batch_Nbr corresponds to integer values. As
    of now its not correct which is why it complains about data type mismatch.
    Also didnt understand why you unmarked original answer as it was the solution to your actually posted problem. Now what you're asking is an unrelated issue. Anyone looking at this would have benefited with previous answer if their original issue was with
    execute package task property expression.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • When/how do the scheduled packages switch status from 'scheduled' to 'ready' in the package Scheduler Status?

    Hi,
    I've scheduled FX Restatement packages to run each day in two different versions (6 times a day starting at 0615) for a year or so. However when I check out the 'View Package Status' the jobs I set up in the first version doesn't run when they are supposed to. I checked out the 'Package Scheduler Status' and I found that the jobs that DIDN'T run was stuck in a status called 'scheduled' while the jobs running had status 'ready'.
    So the question is how can I make sure all the jobs are running? Do I have to or can I do something about this myself? If I try to set up a package in two different versions, can the fact that I set them both to run at 0615 generate a conflict resulting in just one of them running?
    Hope someone has answers to these questions, would be so great to get this out of the way. Thanks!
    K

    Hi Kristoffer,
    remember to click on enable schedule when you schedule a package.
    To know if a scheduled package is ok you can see first if present in the tblschedule of your appset and also if present in the windows task schedulerif with status "ready".
    After you just have to check the view status if you found the package execution completed you're sure it's running. " version at the same time can cause issues also with different packages better to schedule with a sufficient delay to be sure that the first has ended, in your case 1 package is ok.
    I would delete the 2 packages from bpc, checking also the windows scheduler to be sure that no packages are running and rebuild one from scratch setting repeat every 4 hours and checking "enable schedule".
    Regards
         Roberto

  • Execute package task is not able to pickup the package to execute it from a Foreach Loop Container

    Hi,
    I want to execute 5 packages using EXECUTE PACKAGE Task in FOR EACH LOOP Container. I followed the instructions as per the below link
    http://microsoft-ssis.blogspot.in/2013/01/master-child-packages-part-1-file-based.html
    When I execute the main package it is going to folder where i have that 5 packages and picking up the first one then it is not completing the package keep on executing the same.
    What i need to check here, kindly help me here.
    Regards,
    Sridhar V
    Sridhar

    Add a dummy execute sql task as first step inside loop and add a breakpoint to that for OnPreExecute event and check value of variable during each iteration to see if it gets value of each child package witin folder correctly.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Email BPC data manager Package log to a team for BPC10 NW

    Hi iI found documentation on how to send Data manager package by email. Contained 2 transports but the documentation is for 7.0 NW do you know if they have a documentation for BPC 10 NW ?

    Hi Frederic,
    Please check the below documnet.
    How to Email BPC Data Manager- Formula Log or Package Log
    Thanks,
    Dinesh.v

  • Standby redo log status

    Can anyone explain what the status of clearing means when viewing standby redo log status?
    SQL> select status from v$log;
    STATUS
    CLEARING
    CLEARING
    CURRENT
    This is a bit different to what we'd see on an open database.

    Hello;
    Log is being re-created as an empty log.
    In a Standby database the regular redo are not used ( when not in PRIMARY MODE) only the standby redo logs are used. Sometimes as buffers for the redo coming from the primary.
    You should be able to change the status by doing a ALTER SYSTEM SWITCH LOGFILE on your primary.
    This note may help:
    Online Redo Logs on Physical Standby [ID 740675.1]
    Also
    Take a look at v$standby_log too.
    http://docs.oracle.com/cd/B28359_01/server.111/b28320/dynviews_3068.htm
    Best Regards
    mseberg

Maybe you are looking for

  • Acrobat PDF Maker freezes at 90% done when converting from Word

    ...and here's the weird thing. When I cancel, the PDF appears on my desktop, named "Microsoft Word - (filename).doc.pdf'. It was working fine up until a few hours ago, exporting the PDF in a few seconds, to the same file path, and with the same name,

  • Java and  Internet Explorer ?

    Hello every one. Could someone tell me how to call Internet Explorer from a Java Application, passing a string containing the URL? thank a lot in advance.

  • Help on SEPA COR1 Solution

    Hi Experts, One of our teams is working on SEPA Direct Debit implementation for the Europe region sites and looking for help in designing COR1 solution. To give a context, as we know, for the German, Austrian and Spanish markets, the CORE scheme has

  • New meeting notifications

    I am trying to troubleshoot a problem with one of my users systems. when a new meeting request is sent, it will not show in the inbox or calendar until outlook is closed and reopened. the system has outlook 2003/sp2 and ocfo 10.1.2.

  • Can we change ATP confirm qty in sales order?

    Hi All, Is it possible to change confirm qty in sales order when run ATP check? This is a rounding issue. We have a material with sales unit box and base unit EA, 1BOX =  20 EA, but in storage we just have 8 EA. When we create a sals order, the qty w