Extended method in the task is not executing

Hello Friends,
I had extended the business object BUS2105 and added a new synchronous method which will return the PR items as objects in the form of a table. This method does not have an import parameter as I am using OBJECT-KEY-NUMBER and has a multiline export element of type BUS2009. Then I had Implemented, Released & Delegated my Z business object. All these activities were completed perfectly.
Then in my workflow I had used this method in an activity step.  I had set the binding from Workflow container to the Task container( in the workflow container I am getting the EVTOBJECT into a workflow container element of type BUS2105 by binding from Event container to Workflow container) and then from the Task container to the Method container also. There were no inconsistencies during binding.
During test run this works perfectly and returns the workflow with the PR items as objects into a multiline container element of type BUS2009. Then I tested using SWUE and triggered against the 'CREATED' event of BUS2105. There also this works perfectly as it should be.
My problem is, when I create a PR from ME51N, execution of my workflow stops by this activity step which is using my custom method. From SWEL I can see that workflow workitem has been created. In the workflow log, the created PR number is coming up to the WIOBJECT_ID of the task container. But the method will not trigger. If binding was the problem, then this workflow should not work from SWUE or during test run.
Can anyone help me to solve this issue? What else I should do to solve this?
Thanks & Regards,
Abijith

Hi,
"My problem is, when I create a PR from ME51N, execution of my workflow stops by this activity step which is using my custom method. From SWEL I can see that workflow workitem has been created. In the workflow log, the created PR number is coming up to the WIOBJECT_ID of the task container. But the method will not trigger. If binding was the problem, then this workflow should not work from SWUE or during test run."
1. I hope the task would be a background task and that is the reason even after inserting a break point the execution of the method has not interrupted. When ever you want to debug a background task first make it as a fore ground task and make agent assignment as general task and then assign the recipient as yourself, so that the workitem comes to your inbox then you try to execute the worktiem after inserting the break point in the method. This way you can debug the method from the workflow.
2. The other part would be as you said that the workflow hangs at this activity step , so make sure the PR is not locked, I mean sometimes it may be possible that the object might be locked and if any changes are done in the method. it fails.
3. The reason I feel it is working from SWUE is that you have inserted values which are already existing in the system and they might not have locked by the system.
so please check the above three points.
Regards
Pavan

Similar Messages

  • The task could not be successfully executed. Reason: 'java.lang.StackOverfl

    Hi,
    I have a presentation that contains a group of attributes. I created a screen flow that invokes this presentation.
    When I execute the project in a browser, and click on the "_Append a new row to the end of this group_", this error message appears:
    The task could not be successfully executed. Reason: 'java.lang.StackOverflowError'.
    I used the same structure of the group of attributes that is applied in the Tutorial Guide of BPM Studio.
    The difference is that I used the group in the second activity not in the global activity.
    I really appreciate your help.
    Regards,
    Eva
    Edited by: user9976328 on 08-abr-2009 10:29

    Dan
    I appreciate your answer, but i still have problems because when i re-installed BPM Studio 10.3.1.0.0, the installer does not let me change the JVM parameters.
    The installer by default sets the following parameters:
    Product Name:
        Oracle BPM Studio 10.3.1.0.0
    Install Folder:
        C:\OraBPMStudioHome
    Shortcut Folder:
        C:\Documents and Settings\Eva Escobar\Menú Inicio\Programas\Oracle\BPM\Studio
    Install Set
        Typical
    Java VM Installation Folder:
        C:\OraBPMStudioHome\eclipse\jre
    Disk Space Information (for Installation Target):
        Required: 945.089.189 bytes
        Available: 32.327.565.312 bytes
    Please your feedback.
    Regards,
    Eva
    Edited by: Eva E on 24/07/2009 12:25 PM

  • Update Row into Run Table Task is not executing in correct sequence in DAC

    Update Row into Run Table Task is not executing in correct sequence in DAC.
    The task phase for this task is "Post Lost" . The depth in the execution plan is 19 but this task is running some times in Depth 12, some times in 14 and some time in Depth 16. Would like to know is this sequence of execution is correct order or not? In the out of the Box this task is executed at the end of the entire load. No Errors were reported in DAC log.
    Please let me know if any documents that would highlight this issue
    rm

    Update into Run table is a task thats required to update a table called W_ETL_RUN_S. The whole intention of this table is to keep the poor mans run history on the warehouse itself. The actual run history is stored in the DAC runtime tables, however the DAC repository could be on some other database/schema other than warehouse. Its mostly a legacy table, thats being carried around. If one were to pay close attention to this task, it has phase dependencies defined that dictate when this task should run.
    Apologies in advance for a lengthy post.... But sure might help understanding how DAC behaves! And is going to be essential for you to find issues at hand.
    The dependency generation in DAC follows the following rules of thumb!
    - Considers the Source table target table definitions of the tasks. With this information the tasks that write to a table take precedence over the tasks that reads from a table.
    - Considers the phase information. With this information, it will be able to resolve some of the conflicts. Should multiple tasks write to the same table, the phase is used to appropriately stagger them.
    - Considers the truncate table option. Should there be multiple tasks that write to the same table with the same phase information, the task that truncates the table takes precedence.
    - When more than one task that needs to write to the table that have similar properties, DAC would stagger them. However if one feels that either they can all go in parallel, or a common truncate is desired prior to any of the tasks execution, one could use a task group.
    - Task group is also handy when you suspect the application logic dictates cyclical reads and writes. For example, Task 1 reads from A and writes to B. Task 2 reads from B and writes back to A. If these two tasks were to have different phases, DAC would be able to figure that out and order them accordingly. If not, for example those tasks need to be of the same phase for some reason, one could create a task group as well.
    Now that I described the behavior of how the dependency generation works, there may be some tasks that have no relevance to other tasks either as source tables or target tables. The update into run history is a classic example. The purpose of this task is to update the run information in the W_ETL_RUN_S with status 'Completed' with an end time stamp. Because this needs to run at the end, it has phase dependency defined on it. With this information DAC will be able to stagger the position of execution either before (Block) or after (Wait) all the tasks belonging to a particular phase is completed.
    Now a description about depth. While Depth gives an indication to the order of execution, its only an indication of how the tasks may be executed. Its a reflection of how the dependencies have been discovered. Let me explain with an example. The tasks that have no dependency will have a depth of 0. The tasks that depend on one or more or all of depth 0 get a depth of 1. The tasks that depend on one or more or all of depth 1 get a depth of 2. It also means implicitly a task of depth 2 will indirectly depend on a task of depth 0 through other tasks in depth 1. In essence the dependencies translate to an execution graph, and is different from the batch structures one usually thinks of when it comes to ETL execution.
    Because DAC does runtime optimization in the order in which tasks are executed, it may pick a task thats of order 1 over something else with an order of 0. The factors considered for picking the next best task to run depend on
    - The number of dependent tasks. For example, a task which has 10 dependents gets more priorty than the one whose dependents is 1.
    - If all else equal, it considers the number of source tables. For example a task having 10 source tables gets more priority than the one that has only two source tables.
    - If all else equal, it considers the average time taken by each of the tasks. The longer running ones will get more preference than the quick running ones
    - and many other factors!
    And of course the dependencies are honored through the execution. Unless all the predecessors of a task are in completed state a task does not get picked for execution.
    Another way to think of this depth concept : If one were to execute one task at a time, probably this is the order in which the tasks will be executed.
    The depth can change depending on the number of tasks identified for the execution plan.
    The immediate predecessors and successor can be a very valuable information to look at and should be used to validate the design. All predecessors and successors provide information to corroborate it even further. This can be accessed through clicking on any task and choosing the detail button. You will see all these information over there. As an alternate method, you could also use the 'All/immediate Predecessors' and 'All/immediate Successor' tabs that provide a flat view of the dependencies. Note that these tabs may have to retrieve a large amount of data, and hence will open in a query mode.
    SUMMARY: Irrespective of the depth, validate
    - if this task has 'Phase dependencies' that span all the ETL phases and has a 'Wait' option.
    - click on the particular task and verify if the task does not have any successors. And the predecessors include all the tasks from all the phases its supposed to wait for!
    Once you have inspected the above two you should be good to go, no matter what the depth says!
    Hope this helps!

  • I am part of a family sharing setup.  Can I have my own payment method so the organizer does not get charged for my purchases?

    I am part of a family sharing setup.  Can I have my own payment method so the organizer does not get charged for my purchases?  If so, how do I set this up?

    Yes.  Just redeem gift cards so that your own account always has a sufficient credit balance.  Here is how it works:
    Family purchases and payments
    After you set up your family, any time a family member initiates a new purchase it will be billed directly to your account unless that family member has gift or store credit. First, their store credit will be used to pay the partial or total bill. The remainder will bill to the family organizer's card.

  • The process could not execute 'sp_repldone/sp_replcounters' error for Log Reader Agent and SQL Server Assertion 17066 & 3624 errors in SQL Logs

    One of our SQL Server started creating SQLDUMP file and and on investigation I found the error longs are filled with Errors 3624 & 17066. There is transnational replication configured on one of the databases is the LogReader Agent is failing error "The
    process could not execute 'sp_repldone/sp_replcounters' on XXXXX". 
    Not sure if both these Assertion & Logreader Agent errors are related. Before I remove and put the replication, I wanted to check if anyone has experienced the same issues or aware of what the cause. 
    ***********Error messages from SQL Logs******
    **Dump thread - spid = 0, EC = 0x0000000111534460
    Message
    A system assertion check has failed. Check the SQL Server error log for details. Typically, an assertion failure is caused by a software bug or data corruption. To check for database corruption, consider running DBCC CHECKDB. If you agreed to send dumps to
    Microsoft during setup, a mini dump will be sent to Microsoft. An update might be available from Microsoft in the latest Service Pack or in a QFE from Technical Support.
    Error: 3624, Severity: 20, State: 1.
    SQL Server Assertion: File: <logscan.cpp>, line=2123 Failed Assertion = 'UtilDbccIsInsideDbcc () || (m_ProxyLogMgr->GetPru ()->GetStartupState () < RecoveryUnit::Recovered)'. This error may be timing-related. If the error persists after rerunning
    the statement, use DBCC CHECKDB to check the database for structural integrity, or restart the server to ensure in-memory data structures are not corrupted.
    Error: 17066, Severity: 16, State: 1.
    External dump process return code 0x20000001.
    External dump process returned no errors.
    Thank you in advance.

    You need to determine if this error is a transient one or a show stopper one.
    It sounds like your log reader agent has crashed and can't continue.
    If so your best bet is to call Microsoft CSS and open a support incident.
    It also sounds like DBCC CHECKDB was running while the log reader agent crashed.
    If you need to get up and running again run sp_replrestart, but then you might find that replicated commands are not picked up. You will need to run a validation to determine if you need to reinitialize the entire publication or a single article.
    I have run into errors like this, but they tend to be transient, ie the log reader agent crashes, and on restart it works fine.
    looking for a book on SQL Server 2008 Administration?
    http://www.amazon.com/Microsoft-Server-2008-Management-Administration/dp/067233044X looking for a book on SQL Server 2008 Full-Text Search?
    http://www.amazon.com/Pro-Full-Text-Search-Server-2008/dp/1430215941

  • HT1918 I have a new visa debit card, I have tryed several times to put in this new information into my IPad. The program keeps telling me. There's a billing problem with a previous purchase. Please update your payment method. The program will not let me u

    I have a new visa debit card, I have tryed several times to put in this new information into my IPad. The program keeps telling me.
    There's a billing problem with a previous purchase. Please update your payment method.
    The program will not let me update, what am I doing wrong. rose165

    I would love to have a phone number to call someone with apple on this issue.

  • I am trying to download the update IOS 5 version for mi ipad 2 and at the end when it finishes downloading it shows an error "connection has been interrupted and that the task could not be completed", my internet service is fine, what can it be?

    I am trying to download the update IOS 5 version for mi ipad 2 and at the end when it finishes downloading it shows an error "connection has been interrupted and that the task could not be completed", my internet service is fine, what can it be?

    The 2nd generation can not be upgraded beyond 4.2.1.

  • Error 1003 - The vi is not executable​. Simulated DAQ-cards.

    Hi all,
    I am updating an application written in LabVIEW 7.1 for a customer. The code is not made by my company and is pretty horrendous but it works. I can not develop the code on the machine that it is supposed to run on so I have to do it on my own computer which means that I have to simulate two DAQ-cards that the program needs. I have pretty much finished the program and I want to test if I can build an exe-file. This is where I run in to trouble an get the error message:
    Error 1003 occurred at \\Server1\Users\martinh\LabVIEW Data\app\internal.llb\_Main.vi
    Possible reason(s):
    LabVIEW:  The VI is not executable.
    I know that this has been discussed at length here and that one has to look out for global variables, dynamically loaded VI:s etc. My question is, does anyone think that the simulated DAQ-cards give me trouble? Do they use some VI:s placed somewhere odd which I need to include when I build the application?
    Sincerely
    Martin

    Hi Martin,
    I suggest you try what is suggested in this link, some of the information is already covered in the posts above but some might be new.
    http://digital.ni.com/public.nsf/allkb/705C2ECA081​F3C7986256C0F00559B02?OpenDocument
    If you are using the office toolkit, it might be an idea to masscompile the _office folder and you might also need to uncheck the "Disconnect type definitions and remove unused polymorphic VI instances".
    Good Luck
    Andreas E
    Applications Engineer
    National Instruments

  • The VI is not executable. The full development version...

    "The VI is not executable. The full development version of LabVIEW is required to fix the errors." That is the error I've been dealing with all day. Actually it is repeated 3 to 4 times in the error message, seen when first launching the EXE. Also of note is that the run-arrow is broken. This is in an EXE that was developed well before version 2013, but I recently upgraded it to 2013 to hopefully fix some random LVRT crashes. There is obviously no issue running the code on my development machine, only once it's deployed.
    The good news is I've been able to solve the problem (mostly). My EXE now runs without the errors. I found that the problem wasn't actually in the EXE.
    We have a separate project with no files, but in the build specification we have the LabVIEW 2013 Run-Time Engine, as well as NI-VISA for serial support. We then bundle the deployment in a self-extracting RAR and it does its business. In the past this has worked just fine, but with LV2013, it seems that there is a component missing from the deployment.
    I found that if I download the Run-Time Engine directly from NI.com and run it, the only things I need to install are the Run-Time Engine, and the DEPLOYABLE LICENSE. If I do this, my EXE runs great, but with the custom installer, it is not included for some reason!
    Putting aside the fact that I don't know why the Deployable License package is needed (we aren't connecting to remote panels), I can't seem to modify my custom installer to install the Deployable License. There are versions for LabVIEW 2009 and 2011 SP1 in the Additional Installers list, but I don't see 2013.
    Can anyone explain why this is??? For now I will just distribute the 257MB Run-Time Engine and the NI-VISA installers separately I guess. But I'd really like to get back to a single, simple package to install all the prerequisites we need. HELP!

    iannicholson wrote:
    Yes, but then every installer I build will contain the runtime engine. That gets very large very quick. So I have an installer for JUST the runtime engine and VISA runtime.
    Just remove them from the "Additional Installers" tab of the Installer dialog.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • When I touch my iPad screen to execute an app, the screen goes dark then the app icon reappears. The app does not execute.

    When I touch my iPad screen to execute an app, then the screen goes dark then the app icon reappears. The app does not execute. Any thoughts?

    Try a Reset...
    Press and hold the Sleep/Wake button and the Home button at the same time for at least ten seconds, until the Apple logo appears. Release the Buttons.
    Basic troubleshooting
    From Here
    http://www.apple.com/support/ipad/basics/

  • Error "The application could not execute the required function"

    Hi All,
    When I try to activate the DTP the following error occurs."the application could not execute the required function".
    I was able to activate the DTP for the same DataSource 2 days back but now I am unable to activate .
    Kindly suggest what could be the possible reason for the same.
    Regards,
    Santosh Mani

    If you have transformation, check it is active. Is the object 0plant is acitve or not?
    Ravi Thotahdri

  • Please help suggestions for solving built-only compiler error: 'The VI is not executable. The full development version of LabView is required tofix this error.'

    We have develoepd a software tool and build it on regular basis. It currently runs error free when compiled in the editor, but when we built it and run the executable stand alone we get the error.
    'The VI is not executable. The full development version of LabView is required tofix this error.' plus a broeken error.
    This menas an compiler error that is not present in the editor but in the stand alone version. We tried to identify errors as suggested in several posts in this forum, but so far unsuccesfull.
    As the editor and its compiler do not see the error  and are running fine and the stand alone version just syas 'find the error in the editor' in this case LabView is of no help.
    Can anyone suggest a sensible or 'good practice ' way of searching for the source of this error?
    Our project  comprises hundreds of Vis over several libraries.
    Thanks,
    Chris

    Thanks Craig for all your suggestions.
    We seem to have located the problem in a new vi just added to the package causing conflicts by using the same vi names as other vis already present in the package. Excluding this vi removed the error.
    It seems related to a conflict by having two vis with the same name, which was mentioned by LabView and interactively resolved when running the main vi from the editor. When successfully building the main vi the builder did not mention this conflict and reported a successful build, but when trying to run the executable it gave the cryptic error. The error caused us problems because there was no hint for the cause, just the suggestion to solve this in the editor, while at the same time in the editor the VI was running fine.
    We will post about this in detail after we have positively proven that this actually was the case.
    The .net version issue was already checked.
    Performance was the same on all machines we tested on including the dev machine.
    Debugging was tried nut did not help as the vi could not run (broken arrow). We assumed that debugging only helps in running faulty functioning vi''s. We did not check for broken arrows in sub-vis (after connecting), that could have helped, although our application has hundreds of our own vis.
    In relation to your remark: 
    'Are you using many classes? Have you verified that the proper access scopes are set for functions calling those vis?'
    Could you elaborate on setting access scopes. We were not aware of this option in LabView, although we realize this is a basic element of the underlying c code.
    Ragrds,
    Chris

  • LabVIEW: The VI is not executable. Most likely the VI is broken or one of its subVIs cannot be located. Open the VI in LabVIEW using File Open and verify that it is runnable.

    When trying to run a sequence on a deployment machine I get this error. When I switch the Adaptor back to the Development System the problem goes away however I do not have the luxury of keeping the full Labview development on the test machine. Is there a know fix for this problem and what is the cause of this error. Here is the software and versions I am using. It seems that the vi's that are broken are using Invoke and Property Nodes within them, I am not sure if this is just a coincidence. It appears it is more related to the invoke node than the property node. Any help is greatly appreciated. Why does it seem everything works great until you try and deploy it on a test machine. This is not the first time I have had these kinds of problems using Nation Instruments software.
    Labview 8.2.1
    TestStand 4.0
    Solved!
    Go to Solution.

    I have a similar problem.
    A bit of background - I had developed a test suite with a Labview front panel for manual operations, and a Teststand project which used some of the same lower level vi's, and specific test step vi's. It was developed on LV 2010 32-bit and Teststand 2010. All was fine.
    Recently i had to rebuild my development machine and installed only LV 64-bit 2011 and the latest (2010) Teststand from the 2012 Dev Suite.
    I knew that i wanted to run some new hardware drivers in 64-bit versions due to another problem, so i replaced the .dll in the very lowest level VI with the 64 bit version. And ran the manual interface. All ok.
    But when i try the Teststand execution, i get one of two errors:
    - If i set Teststand to use the LV runtime engine 11.0.1 (2011 SP1), the i get a "Unable to load VI xxxxx with the runtime engine version '11.0'" Despite the fact that each VI does run correctly in LV dev when loaded individually, and i've saved them as version 11.
    - If i set it to use the development system (which shows in the config menu 'active version:32-bit' even though 32-bit LV is not installed), then i get the "The VI is not executable. Most likely the VI is broken or one of its subVIs cannot be located." error, even though i when i open each VI in the LV 2011 it runs ok.
    Any ideas?
    How can i set TS dev system to use the 64-bit version? Or is there a separate 64-bit deployemrnt of TS?
    Labview 2010, TestStand 2010

  • Worflow : Method with in the task is not working.

    Hello,
    i am saving a parked document and change event is triggereing ( i saw this in SWEL).
    i have a created a single step task that contains the trigering event (Change for parked document)  and custom method which contains the code to update a z table.
    i see in swel event is fired and also the task is fired with that.
    but the table is not getting updated.
    i tested the object and debuggd it manually through swo2. the program is working fine when executed through swo2. but doesnot work when executed through an event via transaction.
    please let me know if there is any additional setting for this to work.
    Thanks.

    I went in to worflow log and found that,
    1. Status :  Ready.
    2. After pressing details button i get details for step table below.
    3. In this table there are two entries : a)Work item Created after event. b)Dailog work item created.
    can somebbody please explain me why the method is not executed (or is it executed).
    thanks

  • Error: The task could not be successfully executed.

    I am working on a project that is using ALBPM 6.0 Studio for local development. During the local (PC) process execution of Interactive items, ComponentExecutionException is encountered sporadically.
    The details of the error message given in the browser are below:
    fuego.xobject.util.InvalidContextKey: Invalid BPM Object context key.
    Reason: 'fuego.xobject.util.InvalidContextKey: Invalid BPM Object context key.'
    Caused by: Invalid BPM Object context key
    fuego.lang.ComponentExecutionException: This task could not be successfully executed.
    Reason: fuego.xobject.util.InvalidContextKey: Invalid BPM Object context key.
    at fuego.web.execution.InteractiveExecution.retry(Interactive.java:251)
    at fuego.web.papi.TaskExecutor.runInstanceTask(TaskExecutor.java:375)
    at fuego.web.papi.TaskExecutor.execute(TaskExecutor.java:98)
    This error occurs sporadically. Any input on how to resolve would be greatly appreciated.

    Check whether Requires HTTP Authentication is checked in the External Resources and enter the password and username for it if required..

Maybe you are looking for

  • How do I add a step to run at the beginning of an existing task sequence?

    I need a way to add a step to run at the beginning of an existing task sequence like you see up above. We need to do so to all task sequences that install software in our environment. I understand PowerShell very well but, I am not sure of how I woul

  • BAPI or FM for t-code IW34

    Hi all, i need create service order for existing service notifications (like transaction IW34).  I have already used bapi BAPI_ALM_ORDER_MAINTAIN to create service order, but i dont know link or connecting the generated order with the notification. P

  • Error when accessing CMP EJB methods

    I get a strange error, when I access a getter method in my remote interface to query the variables in my CMP EJB. I use Jdev 3.2 agains Oracle 8.1.7. I can see that the error must be originating from the persistance manager BC4J XML, but not sure how

  • Consumed  quantity shouldnt change in GRN for sub contracting

    Dear Gurus, While making GRN for subcontracting activity system allows users  to change the consumption quantity manually. . System has to capture whatever quantity is there in BOM. how can we do this, as per my knowledge if we make greyed out the co

  • Regarding company code in sales order

    Hi guys, From where we will be getting the <b>company code</b> into the sales order? And in sales order where can we find the company code for respective sales order? Regards Venkat