Parallel task for comments

Hi,
Is is possible to create a parallel task to let a few user enter comments while the workflow is still running and tasks being completed by other users ? (I mean the users who receive the comment task in their SAP inbox does not block the rest of the WF and wether they have completed their task or not, others can still perform their decision task)
Where would the comments be stored in that case ?
Thanks in advance for you help.
T

Hello,
You'd have to decide beforehand who can add the comments - they will be the ones receiving the workitem.
The commnts can be stored in the workflow container, or the users could add attachments. In both cases you could have the workflow do something with them at the end (eg send them somewhere).
You could have a 1-from-2 fork, one with the comments and one with the rest of the workflow. The comments branch could loop around, waiting each time for more comments.
regards
Rick Bakker
Hanabi Technology

Similar Messages

  • How to create parallel tasks using parallel for loops

    Hi,
    I am setting up a program that communicates with six logic controllers and has to read the system status every 100 ms. We are using OPC datasockets for this, and they appear a little slow. 
    I have created a uniform comm. method for all controllers, and now I find myself programming this method six times to communicate with each system. I am wondering if this could be done more elegant using the parallel for loop, in which case I would program an exchange once and then have six workers running simultaneously. Since a picture is more clear that a thousand words, what I am asking is this:
    Is it possible to replace something like
    by
    and have this for loop running these tasks in parallel (on different cores / in different threads)?
    I have configured the loop to create 8 instances at compile, so I would have 2 instances surplus available at runtime if I find I need an additional system.
    The benefits of the method show in the second picture to me are:
    * takes less space
    * modifications have to be made only once
    * less blocks, wires and stuff makes it more clear what's going on.
    * flexibility in the actual number of tasks running (8 instances available at runtime)
    * if more tasks are required, I need only to update the maximum number of instances and recompile, i.e. no cutting and pasting required. 
    Unfortunately, I don't have those system available yet, so there's no way to test this. Yet, I would like to know if the above works as I expect - unfortunately the labview help is not completely clear to me on this.
    Best regards,
    Frans 
    Solved!
    Go to Solution.

    Dear mfletcher,
    First of all: thanks for confirming that my intuition was right in this case.
    As for your question on the help: below is a copy/paste from the help on the 'configure parallelism dialog box' 
    Number of generated parallel loop instances—Determines the number of For Loop instances LabVIEW generates at compile time. The Number of generated parallel loop instances should equal the number of logical processors on which you expect the VI to execute. If you plan to distribute the VI to multiple computers, Number of generated parallel loop instances should equal the maximum number of logical processors you expect any of those computers to ever contain. Use the parallel instances terminal on the For Loop to specify how many of the generated instances to use at run time. If you wire a larger number to the parallel instances terminal than you specify in this dialog box, LabVIEW only executes as many loop instances as you specify here.The reason for me doubting if what I programmed would work the way I intended lies in the fact that the help only mentions processors here, which would be interpreted as actual cores. Thus on a dual core machine, the number should be 2.
    I think it would be helpful to mention something about threads here, because in some case one would like to have more parallel threads than there are cores in a system.
    In mu case I would like to create six threads, which on my dual core processor would be spread over only two cores. Then these six threads run in parallel.I know that in case of heavy math that would not help, but since I am doing communications, which have timeouts and such, and that probably runs smoother in six parallel tasks even though I only have two cores. 
    Hope this helps in improving the help of the for loop.
    Regards,
    Frans 

  • Parallel Task Error resolution for Marketing Campaign Load

    Hi,
    During a Marketing campaign load, we encountered the "Parallel Task Error".
    Description of error:
    Error/Warning : Error invoking service 'Mktg Data Load Service', method 'CampaignLoad' at step 'Load Segment Tree Cells/Segments'.(SBL-BPR-00162)*
    +*
    One or more of the parallel tasks may have encountered an error.  Any queued tasks that were scheduled to follow the failed task have been cancelled.(SBL-MKT-00453)*
    Further investigation revealed that the unique constraint on the S_CAMP_CON table was not being met. As per a suggestion by one of the Oracle Support Web's SR, they had asked to change the value of the field 'Token Number' (in the load file format) to a randomly generated number using the formula:
    CAST((RCOUNT(1) + 001) as varchar(40)) || '001')
    However, the issue arises when we assign more than one nodes of a segment tree to a campaign. Due to this, the token number would be repetitive between records of the two nodes.
    Hence, to overcome this error, instead of defaulting the value of Token Number to a system generated field, or using the above function, we made it default to a unique value (for example: account id), coming from a database column. This way we made sure that the token number is unique for all the records being loaded by the campaign.
    This resolved the error, however we would like to confirm if this is the right approach? Can someone let us know if there is any other solution to avoid the error?
    Thanks in advance

    Hi,
    During a Marketing campaign load, we encountered the "Parallel Task Error".
    Description of error:
    Error/Warning : Error invoking service 'Mktg Data Load Service', method 'CampaignLoad' at step 'Load Segment Tree Cells/Segments'.(SBL-BPR-00162)*
    +*
    One or more of the parallel tasks may have encountered an error.  Any queued tasks that were scheduled to follow the failed task have been cancelled.(SBL-MKT-00453)*
    Further investigation revealed that the unique constraint on the S_CAMP_CON table was not being met. As per a suggestion by one of the Oracle Support Web's SR, they had asked to change the value of the field 'Token Number' (in the load file format) to a randomly generated number using the formula:
    CAST((RCOUNT(1) + 001) as varchar(40)) || '001')
    However, the issue arises when we assign more than one nodes of a segment tree to a campaign. Due to this, the token number would be repetitive between records of the two nodes.
    Hence, to overcome this error, instead of defaulting the value of Token Number to a system generated field, or using the above function, we made it default to a unique value (for example: account id), coming from a database column. This way we made sure that the token number is unique for all the records being loaded by the campaign.
    This resolved the error, however we would like to confirm if this is the right approach? Can someone let us know if there is any other solution to avoid the error?
    Thanks in advance

  • Enable Reader Rights  for Commenting and Markup in Acrobat X Pro?

    I am sure there is a nice long thread somewhere on this but I cannot find the old "Enable for Commenting and Review" in Acrobat X Pro (Windows 7, 32 bit). I get basic comments (sticky and highlight) but i need to make a PDF where my Reader X users can make all the comments and markups. I also tried doing the Send for Review task and it still did not have all the tools in Reader X.

    File > Save As > Reader Extended PDF > Enable Commenting & Measuring

  • Create Multiple tasks for Single Item in List using state machine workflow in sharepoint

    Hi,
    I want to create multiple create tasks for Single Item in List based on Assigned to column using state machine Workflow through visual studio
    Here Assigned to column allows multiple users. so i have to create task for every user based on column .
    I'm trying for this but i didn't got any solution
    Please provide solution for this.

    Hi,
    According to your post, my understanding is that you wanted to allow multiple users to approve.
    There are some articles about creating parallel tasks in state machine workflow, you can have a look at them.
    http://www.codeproject.com/Articles/477849/Create-Parallel-Task-in-State-Machine-Workflow-in
    http://msdn.microsoft.com/en-us/library/office/hh128697(v=office.14).aspx
    http://social.technet.microsoft.com/Forums/office/en-US/b16ee858-4360-479a-a686-4ee35b7be9db/sharepoint-2010-workflow-creating-multiple-tasks?forum=sharepointdevelopmentprevious
    Thanks & Regards,
    Jason
    Jason Guo
    TechNet Community Support

  • Problem passing comments between human tasks - InputText Comment Invisible

    Hi,
    I want to pass the comments between three human tasks, i discover that it to works if i check reinitiate checkbox
    and link all human tasks with the antecedent.
    But, now i have another issue: When i create in the new button in the second human task, the inputText comment is invisible, see the image bellow:
    http://shareimage.org/images/ehmo4rrj80dbswt8fpms.png
    Why inputText comment is invisible ?
    Thanks.
    Victor

    Hi,
    I resolve partially this issue.
    Was a problem in my pageDef.xml. It works and now the inputText is visible.
    The only problem now is: when i make a CLAIM in the second human task, the filed comments is cleaned and the previous comments is losted.
    Why i have to do for comments don't disappear when i make a CLAIM ?
    Thanks
    Victor Jabur

  • Changing OSM task state/status from another parallel task flow

    Hi, guys!
    I have a problem with implementing parallel task flow. Let's consider such situation: there are two parallel task flows, the first flow contain manual task, upon completion which I need to change second flow's task state/status and vice-versa. I studied XMLAPI functions, wsapi operations, api from automation package but couldn't find appropriate aproach.
    Thanks!
    Edited by: serj129 on 15.08.2011 2:50

    Have you any suggestions?
    I tried to implement 2 automation task, which contain automation plugins. First task implemented xquery sender as completed event, the second task implemented xquery reciever, which changes status to success, but this approach isn't working.
    All properties of automation plugins configured properly, but I always receive next exceptions:
    ####<Aug 22, 2011 4:39:52 PM ALMT> <Warning> <EJB> <car07-eth0.telecom> <AdminServer> <ExecuteThread: '13' for queue: 'oms.automation'> <oms-automation> <> <> <1314009592567> <BEA-010065> <MessageDrivenBean threw an Exception in onMessage(). The exception was:
    java.lang.RuntimeException: No transaction associated with request.
    java.lang.RuntimeException: No transaction associated with request
         at oracle.communications.ordermanagement.cluster.message.ClusterMessageHandlerBean.onMessage(Unknown Source)
         at weblogic.ejb.container.internal.MDListener.execute(MDListener.java:466)
         at weblogic.ejb.container.internal.MDListener.transactionalOnMessage(MDListener.java:371)
         at weblogic.ejb.container.internal.MDListener.onMessage(MDListener.java:327)
         at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:4585)
         at weblogic.jms.client.JMSSession.execute(JMSSession.java:4271)
         at weblogic.jms.client.JMSSession.executeMessage(JMSSession.java:3747)
         at weblogic.jms.client.JMSSession.access$000(JMSSession.java:114)
         at weblogic.jms.client.JMSSession$UseForRunnable.run(JMSSession.java:5096)
         at weblogic.work.ExecuteRequestAdapter.execute(ExecuteRequestAdapter.java:21)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:145)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:117)
    >
    ####<Aug 22, 2011 4:39:52 PM ALMT> <Warning> <EJB> <car07-eth0.telecom> <AdminServer> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1314009592571> <BEA-010216> <The Message-Driven EJB: BiTTLautomation.BiTTLautomation.OSMManualListenerMDB is throwing exception when processing the messages. Delivery failed after 185 attempts. The EJB container will suspend the message delivery for 60 seconds before retry.>
    ####<Aug 22, 2011 4:40:09 PM ALMT> <Info> <oms> <car07-eth0.telecom> <AdminServer> <Timer-8> <admin> <> <> <1314009609038> <BEA-000000> <impl.OrchestrationCascadingLRUPolicy: Evicted order /order/474 from orchestration cache due to cache entry expiry>
    ####<Aug 22, 2011 4:40:24 PM ALMT> <Info> <oms> <car07-eth0.telecom> <AdminServer> <ExecuteThread: '13' for queue: 'oms.automation'> <oms-automation> <BEA1-08945C597092560E5D6A> <> <1314009624879> <BEA-000000> <plugin.AbstractAutomator: Creating automation plugin [class oracle.communications.ordermanagement.automation.plugin.XQuerySender_iults7_Impl] built using SDK version [7.0.2.437]>
    I suggest that it is not correct approach, because of automation context restored from DB is related to other task.
    Have any ideas of changing task status from OSM DB directly. This approach is temporary, for compatibility with performing manual tasks in 2 systems: OSM and other
    Thanks!
    Edited by: serj129 on 22.08.2011 3:59
    Edited by: serj129 on 22.08.2011 3:59

  • How to enforce order of execution of parallel tasks?

    How do I set the order of execution of parallel tasks in a container?
    The execution isn't truly random. It appears to be set to a random constant at package creation.
    Demo
    I set up a dummy package.
    Max concurrent executions = 1
    sequence container with four script tasks that do a thread.sleep
    The tasks always execute in the order 4, 2, 1, 3. Closing/opening the solution makes no difference to the order of parallel execution.
    Is there a way i can change the execution order? Do I need to change the guid/id of the object (tried this but it invalidates the xml) or is there an order somewhere else in the XML?
    I've found a hack/workaround but it's a bit ugly so I'm wondering if there's a way to do it without adding dummy tasks.
    Basically if i create a dummy start task and then drag off "on success" constraints to each of the four threads, the threads will execute in parallel in the order in which i created the constraints
    Why do i want to do this? I have a number of parallel tasks in a sequence container where three will take a long time and the rest won't take as long.
    I'll minimise the time required for the execution of container by using parallel tasks AND controlling the queue order (longest first). It won't matter whether I have 1 logical cpu or as many logical cpus as i have tasks.
    Thanks
    Jakub @ Adelaide, Australia

    I think the problem you are trying to say is 
    a) You have multiple tasks( i assume 10) now out of these 3 execute slowly.
    b) You done have any constraints for task execution, so what you want to achieve is that the 7 task should continue without waiting for the other 3 in your case the 3 are starting up earlier and other 7 are not able to start as a result.
    AS arthur pointed out these are 2 diff things constraints and parallel execution, and there would always be some task which would be started at some point of time.
    But if you know in advance some tasks are going to take time, i would suggest keep 2 diff sequence containers in 1 keep the 7 tasks which runs fast in another one keep the other 3 which tend to run slow
    Abhinav http://bishtabhinav.wordpress.com/

  • Parallels Server For Mac configuration (or any VM product), best practices

    What is the recommended practice for configuring Parallels Server For Mac on an XServe?   The xserve has 3 hard drive bays.   What do most people do to maximize uptime, failover, disaster recovery, etc.  We do have 3 different physical xserves with PSfM also.
    Is a software RAID recommended for the boot OS drive? for the drive storing the VMs if not the same drive?
    How are your hard drives configured?
    I'm concerned that mirroring two 1TB hard drives and using that for both the host OS and for the VMs will present problems:
    When creating new VMs, copying large amounts of data, etc, I'd guess that moving 10-100 GB of data is tasking for the software raid and will noticeably affect performance for all VMs.
    If one of the mirrored drives fails and the raid needs to be rebuilt, this probably would bring the entire system to a crawl, at best, and possibly be unusable while it works to repair the raid.  Thus it would have to be done during off peak hours.  This would prevent a complete outage though.
    are these correct assumptions?

    Hello there,
    There is no upgrade path from Server 2012 Essentials to R2.  R2 is not an upgrade, it is a new build.  Essentials does not upgrade easily, you will want to perform a migration of the server.  This will require additional hardware.  You
    can possibly backup the 2012 Essentials, build a temp server on workstation hardware just to do the migration from,  restore the backup to the new server and then perform the migration to R2. 

  • Create a Task, for use in "Deferred Task"

    I am in the processing of creating Deferred Tasks for my Update-User workflow.
    I know how to create these easily. My problem is : the actual "Task" itself, which will be called upon from the Workflow.
    I don't know the proper syntax for it.
    For instance, below is the code I am using for my "Send Email Task". This task is supposed to send an automatic email notification.
    *<?xml version='1.0' encoding='UTF-8'?>*
    *<!DOCTYPE TaskDefinition PUBLIC 'waveset.dtd' 'waveset.dtd'>*
    *<!-- MemberObjectGroups="#ID#Top" authType="UserAdminTask" createDate="Thu Jul 16 15:22:08 EEST 2009" extensionClass="WFProcess" name="Send Email Task" visibility="invisible"-->*
    *<TaskDefinition name='Send Email Task' creator='%STARTUP%Configurator' createDate='1247746928617' lastModifier='%STARTUP%Configurator' lastModDate='1247746928617' repoMod='1247746928170' primaryObjectClass='TaskDefinition' wstype='ProvisioningTask' taskType='Workflow' executor='com.waveset.workflow.WorkflowExecutor' syncControlAllowed='true' execMode='sync' execLimit='0' resultLimit='3600' resultOption='delete' visibility='invisible' progressInterval='0'>*
    *<Extension>*
    *<WFProcess maxSteps='0' audit='true'>*
    *<Variable name='user' input='true'>*
    *<Comments>&#xA; A user view. This is required.&#xA; </Comments>*
    *</Variable>*
    *<Variable name='backgroundProvisioning' input='true'>*
    *</Variable>*
    *<Variable name='provisioningRetryButton' value='true' input='true'>*
    *<Comments>&#xA; Set to "true" to enable the inclusion of a Retry link&#xA; in the task result if the provisioning fails.&#xA; </Comments>*
    *</Variable>*
    *<Variable name='approvals'>*
    *<Comments>&#xA; Returned from the Lighthouse Approval process. Contains the&#xA; approval structure. The approvals.approved variable set if&#xA; all approvals were successful. This is passed into the Notify&#xA; process later for post-provisioning notifications.&#xA; </Comments>*
    *</Variable>*
    *<Variable name='userCreated'>*
    *<Comments>&#xA; Set to "true" from the Provision workflow when the&#xA; Lighthouse account has been created.&#xA; </Comments>*
    *</Variable>*
    *<Variable name='sunset'>*
    *<Comments>&#xA; Object containing information for registering a sunset date/time.&#xA; </Comments>*
    *</Variable>*
    *<Variable name='error'>*
    *<Comments>Set in the event of unusual processing errors.</Comments>*
    *</Variable>*
    *<Variable name='options'>*
    *<Comments>Options to pass to the provisioning task regarding resource&#xA; provisioning.&#xA; </Comments>*
    *</Variable>*
    *<Transition to='end'>*
    *<Comments>&#xA; Terminate if we encounter unusual errors (not provisioning errors).&#xA; </Comments>*
    *<ref>error</ref>*
    *</Transition>*
    *<Activity id='0' name='start'>*
    *<Transition to='Notify'/>*
    *</Activity>*
    *<Activity id='1' name='Notify'>*
    *<Action id='0' name='Email' application='com.waveset.provision.WorkflowServices'>*
    *<Argument name='op' value='notify'/>*
    *<Argument name='template' value='Send Email Now'/>*
    *</Action>*
    *<Transition to='End'/>*
    *</Activity>*
    *<Activity id='2' name='end'>*
    *<WorkflowEditor x='428' y='275'/>*
    *</Activity>*
    *</WFProcess>*
    *</Extension>*
    *<MemberObjectGroups>*
    *<ObjectRef type='ObjectGroup' id='#ID#Top' name='Top'/>*
    *</MemberObjectGroups>*
    *</TaskDefinition>*
    I am not sure if my syntax is correct?
    Any tips?

    Thanks again, sbalu.
    I may have found the error (although, I don't understand what it means).
    Something about *"Missing View"*
    Catalog#format() Entry locale=null, key=SES_WORKITEM_MISSING_VIEW, defVal=SES_WORKITEM_MISSING_VIEW
    20091126 08:28:22.701 TaskThread(0x0052fecf) Catalog#format() Info parameters=null
    20091126 08:28:22.701 TaskThread(0x0052fecf) Catalog#format() Info tmp=null
    20091126 08:28:22.701 TaskThread(0x0052fecf) Catalog#format() Info pattern=Missing view id.
    20091126 08:28:22.701 TaskThread(0x0052fecf) Catalog#format() Info msg=Missing view id.
    20091126 08:28:22.701 TaskThread(0x0052fecf) Catalog#format() Exit returned= Missing view id.
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#addTaskResult() Entry name=Send Email Task, ti=TaskInstance:Send Email Task tdalton
    20091126 08:28:23.092 TaskThread(0x0052fecf) Catalog#format() Entry locale=null, key=SES_WORKITEM_MISSING_VIEW, defVal=SES_WORKITEM_MISSING_VIEW
    20091126 08:28:23.092 TaskThread(0x0052fecf) Catalog#format() Info parameters=null
    20091126 08:28:23.092 TaskThread(0x0052fecf) Catalog#format() Info tmp=null
    20091126 08:28:23.092 TaskThread(0x0052fecf) Catalog#format() Info pattern=Missing view id.
    20091126 08:28:23.092 TaskThread(0x0052fecf) Catalog#format() Info msg=Missing view id.
    20091126 08:28:23.092 TaskThread(0x0052fecf) Catalog#format() Exit returned= Missing view id.
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#addTaskResult() Exit void
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#launchTask() Data executeOnce= false
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#launchTask() Exit returned= false
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Exit returned= false
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#endObject() Entry no args
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#endObject() Exit void
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#processObject() Exit returned= false
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#processDeferredTasks() Data idOrName== #ID#6BF7-:B3702F65421:491FD113-:D0E4376B71977D1D
    20091126 08:28:23.092 TaskThread(0x0052fecf) DeferredScanner#processDeferredTasks() Data nameOrId== ukaila
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processDeferredTasks() Data object lastMod= 24
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processObject() Entry obj=User:ukaila
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processObject() Data tasks= [ { accountId=ukaila, task=Send Email Task, name=Send Email Task, date=11/25/2009 11:15:00 } ]
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processObject() Data task list object= { accountId=ukaila, task=Send Email Task, name=Send Email Task, date=11/25/2009 11:15:00 }
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Entry obj=User:ukaila, trig={ accountId=ukaila, task=Send Email Task, name=Send Email Task, date=11/25/2009 11:15:00 }
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Data start= 11/25/09 11:15 AM
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Data _now= 11/26/09 8:28 AM
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Data task definition= CSC Send Email Task
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#startObject() Entry obj=User:ukaila
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#startObject() Entry type=User, nameOrId=ukaila
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#startObject() Exit void
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#startObject() Exit void
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Data task template= null
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Data task definition= TaskDefinition:Send Email Task
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#processTrigger() Data task template= TaskTemplate:null
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#launchTask() Entry obj=User:ukaila, trig={ accountId=ukaila, task=Send Email Task, name=Send Email Task, date=11/25/2009 11:15:00 }, tt=TaskTemplate:null
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#launchTask() Data taskName= null
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#launchTask() Data generated taskName= Send Email Task ukaila
    20091126 08:28:23.123 TaskThread(0x0052fecf) TaskTemplate#setTaskName(String) Entry no args
    20091126 08:28:23.123 TaskThread(0x0052fecf) TaskTemplate#setTaskName(String) Exit void
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#launchTask() Data orgName= null
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#launchTask() Data owner= null
    20091126 08:28:23.123 TaskThread(0x0052fecf) DeferredScanner#launchTask() Data subjectLength= 29,650
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#create(Element) Entry element=[Map: null]
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Entry element=[Map: null]
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info elname=Map
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info classAtt=null
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info classAtt=null
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info regName=Map
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Exit returned= com.waveset.util.XmlObjectFactory$Registration@1edf84f
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#create(Element) Info handler=com.waveset.util.XmlHashMap@493dca
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#create(Element) Entry element=[Locale: null]
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Entry element=[Locale: null]
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info elname=Locale
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info classAtt=null
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info classAtt=null
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Info regName=Locale
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#getConstructor() Exit returned= com.waveset.util.XmlObjectFactory$Registration@77a748
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#create(Element) Info handler=com.waveset.util.XmlLocale@1876e5d
    20091126 08:28:23.123 TaskThread(0x0052fecf) XmlObjectFactory#create(Element) Exit returned= en_US

  • Dynamic Parallel Approval for HCM Process and Forms

    Hi everyone,
    I have a scenario where I need to use the "Dynamic Parallel Approval" (or to keep it simple, initially I tried using the "Parallel Approval" wizard)for a workflow used in the HCM Process and Forms.
    The standard task for approval in process and forms is TS17900101. I have mentioned a multiline container in the Miscellaneous tab of this task. However,I was unable to use this task in the wizard. There are no results attahced to this task unlike any other standard approval task (like TS30200147). I need to use the task TS17900101 in the workflow assigned to process and forms, but not sure how to handle this scenario (parallel approval).
    If this is not the right way of doing it, Is there any workaround for "Parallel Approval" in HCM Process and Forms.
    Could anybody throw some light around this area.
    Thanks for your help.
    - MM

    Thanks Anuj. But I believe, the container element that I add in the miscellaneous tab does not necessarily have to be used in the agent assignment. The multiline container is just to instantiate the workitem 'n' number of times. Correct me if I am wrong.
    My concern is that I was unable to use this approval task (TS17900101) in the workflow wizard for dynamic paralle/parallel approval.
    Arghadip - Thanks for your suggestion. I have seen some your nice contributions in the WF forum.
    I actually tried using the 'Blocks'. But this is what I ran into. When I send multiple approval requests (say 3), if one person has approved it and the second has rejected it,I need to take out the workitem from the third person's list (because it has been rejected by someone in the group). I am not sure if this is possible using Blocks. And in my case the third person is still having the workitem, but gets a dump/error when he tries to open it.
    Also, if any one has rejected the request, I do not have to wait for the rest to take any action on the workitem and proceed further. But I guess in 'Blocks' it will not let you go out unless every workitem has been processed.
    To summarize,here's what I need - I need to come out of the block for two conditions. One, if everyone has approved, comeout of the block with an apprval flag. Two, if anyone has rejected (even if some have not processed their workitem), delete the workitems from others inbox and come out of the block with a rejection flag.
    So, any kind of input or suggestions on how this could be handled would be highly appreciated.
    Thanks
    MM

  • DAC Lite maximum number of parallel tasks cannot be more than 5

    Hi,
    I am getting below error in my DAC server log.
    13 SEVERE Wed Feb 13 17:16:50 IST 2008 For DAC Lite the maximum number of parallel tasks cannot be more than 5
    14 SEVERE Wed Feb 13 17:16:50 IST 2008 For DAC Lite ETL can be run on only one Informatica Server.
    What does this mean? I have just done Oracle BI Apps 7.9.4 installation as per installation document on linux server. Do i need to do any configuration changes to rectify this? I don;t understand why this error is coming. I restarted my server also but still i am getting the same error.
    I have only one informatica server installed on DAC server.
    Thanks,
    Prasad

    Check the settings for your Informatica Server on the DAC Setup View.
    There is a Max Sessions variable. What is set to?
    To be honest I have never seen this "DAC Lite" entry. Something has not been installed or set up properly.
    Is this an installation on Windows or Linux?
    If Linux/UNIX I suspect a file was not change to Unix format when it was copied over from Windows and that may be having an impact on things here.

  • How do you programmatically "enable for commenting"?

    I have seen a million forum posts where people want to programmatically enable their pdfs for commenting, and the answer is: use LiveCycle.  We are considering buying LiveCycle for our server JUST for this purpose, since there doesn't seem to be any other way to do it.  Before we do, I want to know how it would work.  Since I don't have livecylce, I can't very well poke around and figure it out, and I don't see any free trial downloads.  Can anyone explain how you would go about this task?  On the other hand, isn't the pdf an open document standard?  Is there any way to modify the pdf "under the hood" to enable it for commenting, without using the Acrobat menu item?

    TO programmatically apply uage rights to a PDF document, see the topic named Applying Usage Rights to PDF Documents located at the following URL:
    http://livedocs.adobe.com/livecycle/8.2/programLC/programmer/help/000411.html
    This will explain what to do and will lead you to the following Java code example:
    * This Java Quick Start uses the following JAR files
    * 1. adobe-reader-extensions-client.jar
    * 2. adobe-livecycle-client.jar
    * 3. adobe-usermanager-client.jar
    * 4. adobe-utilities.jar
    * 5. jbossall-client.jar (use a different JAR file if LiveCycle ES is not deployed
    * on JBoss)
    * These JAR files are located in the following path:
    * <install directory>/Adobe/LiveCycle8/LiveCycle_ES_SDK/client-libs
    * For complete details about the location of these JAR files,
    * see "Including LiveCycle ES library files" in Programming
    * with LiveCycle ES
    import com.adobe.livecycle.readerextensions.client.*;
    import java.util.*;
    import java.io.File;
    import java.io.FileInputStream;
    import com.adobe.idp.Document;
    import com.adobe.idp.dsc.clientsdk.ServiceClientFactory;
    import com.adobe.idp.dsc.clientsdk.ServiceClientFactoryProperties;
    public class ApplyUsageRights{
         public static void main(String[] args) {
           try{
                //Set connection properties required to invoke LiveCycle ES                                        
                Properties connectionProps = new Properties();
                connectionProps.setProperty(ServiceClientFactoryProperties.DSC_DEFAULT_EJB_ENDPOINT, "jnp://localhost:1099");
                connectionProps.setProperty(ServiceClientFactoryProperties.DSC_TRANSPORT_PROTOCOL,ServiceClientFactoryProperties.DSC_EJB_PROTOCOL);         
                connectionProps.setProperty(ServiceClientFactoryProperties.DSC_SERVER_TYPE, "JBoss");
                connectionProps.setProperty(ServiceClientFactoryProperties.DSC_CREDENTIAL_USERNAME, "administrator");
                connectionProps.setProperty(ServiceClientFactoryProperties.DSC_CREDENTIAL_PASSWORD, "password");
                //Create a ServiceClientFactory object
                ServiceClientFactory myFactory = ServiceClientFactory.createInstance(connectionProps);
                //Create a ReaderExtensionsServiceClient object
                ReaderExtensionsServiceClient reClient = new ReaderExtensionsServiceClient(myFactory);
                //Retrieve the PDF document to which to apply usage rights
                FileInputStream fileInputStream = new FileInputStream("C:\\Adobe\\Loan.pdf");
                Document inputPDF = new Document(fileInputStream);
                //Create a UsageRight object and specify specific usage rights
                UsageRights useRight = new UsageRights();
                useRight.setEnabledDynamicFormFields(true);
                useRight.setEnabledComments(true);
                useRight.setEnabledFormFillIn(true);
                useRight.setEnabledDigitalSignatures(true);
                //Create a ReaderExtensionsOptions object
                ReaderExtensionsOptionSpec reOptions = new ReaderExtensionsOptionSpec();
                //Set the usage rights
                reOptions.setUsageRights(useRight);
                reOptions.setMessage("This is a Rights-Enabled PDF Document");
                //Apply usage rights to a PDF document
                Document rightsEnabledPDF = reClient.applyUsageRights(
                     inputPDF,
                     "RE2",
                   null,
                   reOptions);
                //Create a new PDF file that represents the rights-enabled PDF document
                File resultFile = new File("C:\\Adobe\\Critique5.pdf");
                rightsEnabledPDF.copyToFile(resultFile);
              }catch (Exception e) {
                    e.printStackTrace();

  • Emailing Wiki Update Notifications also for comments

    Hi all,
    as described in the archived article
    https://discussions.apple.com/message/11651524?messageID=11651524
    mail-notification is configured and working via RSS.
    BUT not in the case of new comments.
    How can we realize this? Waiting for LION will not be a real alternative.
    Thanks in advance for support and reply.
    Robert D.
    http://www.hd-edv.de

    UPDATE:
    we have also contacted the apple-supprt and they told us,that is designed it that way, that means you can not search for comments and also you can not collect them to send them via RSS.
    does anybody know how to parse the XML-plist-files with the comments and transform them into RSS-XML-format, that should be possible solution...maybe as scheduled task (every 5 min)....?
    thanks in advance...

  • Job fail with Timeout for parallel process (for SID Gener.): 006000

    Hello all,
    Im getting below error and not able to find any issue with Basis side. Please anyone help on this!
    Job started
    Data package has already been activated successfully (will be skipped)
    Process started
    Process started
    Process started
    Process started
    Process started
    Import from cluster of the data package to be activated () failed
    Process 000001 returned with errors
    Process 000002 returned with errors
    Process 000003 returned with errors
    Process 000004 returned with errors
    Background process BCTL_4XU7J1JPLOHYI3Y5RYKD420UL terminated due to missing confirmation
    Process 000006 returned with errors
    Data pkgs 000001; Added records 1-; Changed records 0; Deleted records 0
    Log for activation request ODSR_4XUG2LVXX3DH4L1WT3LUFN125 data package 000001...000001
    Errors occured when carrying out activation
    Analyze errors and activate again, if necessary
    Activation of M records from DataStore object CRACO20A terminated
    Activation is running: Data target CRACO20A, from 1,732,955 to 1,732,955
    Overlapping check with archived data areas for InfoProvider CRACO20A
    Data to be activated successfully checked against archiving objects
    Parallel processes (for Activation); 000005
    Timeout for parallel process (for Activation): 006000
    Package size (for Activation): 100000
    Task handling (for Activation): Backgr Process
    Server group (for Activation): No Server Group Configured
    Parallel processes (for SID Gener.); 000002
    Timeout for parallel process (for SID Gener.): 006000
    Package size (for SID Gener.): 100000
    Task handling (for SID Gener.): Backgr Process
    Server group (for SID Gener.): No Server Group Configured
    Activation started (process is running under user *****)
    Not all data fields were updated in mode "overwrite"
    Data package has already been activated successfully (will be skipped)
    Process started
    Process started
    Process started
    Process started
    Process started
    Import from cluster of the data package to be activated () failed
    Process 000001 returned with errors
    Process 000002 returned with errors
    Process 000003 returned with errors
    Process 000004 returned with errors
    Errors occured when carrying out activation
    Analyze errors and activate again, if necessary
    Activation of M records from DataStore object CRACO20A terminated
    Report RSODSACT1 ended with errors
    Job cancelled after system exception ERROR_MESSAGE

    Thanks for the link TSharma I will try that today.
    UPDATE:
    I ran a non-parallel Data Pump and just let it run overnight. This time it finished after 9 hours.  In this run I set the STATUS=300 parameter in the PARFILE which basically echos STATUS updates to standard out every 300 seconds (5 minutes).
    And as before after 2 hours it finished 99% of the export and just spit out WAITING status for the last 7 hours until it finished.  The remaining TABLES it exported (a few hundred) were all very small or ZERO rows.  There clearly is something going on that is not normal.  I've done this expdp before on clones of this database and it usually takes about 2-2.5 hours to finish.
    The database is about 415 Gigabytes in size.
    I will update what the TRACE finds and I'm also opening a case with MOS.

Maybe you are looking for