Callbacks for Process States

Hi all,
Are there any callbacks available for Process states(faulted,closed,completed) etc similar to callbacks for human workflow tasks in BPEL 1013? I need to handle the scenarios when a process is faulted or is aborted (from BPELConsole) and propagate this information to my application.
thanks,
Ravikiran

This is normal exception handling. You just use the Fault handling in the SOAP response.
Create a Fault message in your BPEL process and send this back to the client.

Similar Messages

  • Callback for a state change on digital Input 6025E

    Hello All,
    I need to have a callback function invoked when a state change occurs on serveral digital inputs (interlocks, kill button etc.).
    I am using a PCI-6025E and traditional NI-DAQ. For test purposes I have connected the output P3.0 to the input P4.0 using a SCB-68
    terminal block and initialize the ports in the following manner:
    void initfunc()
    long nPattern;
    DIG_Prt_Config(1, 3, 0, 1); /* port 3 output */
    DIG_Prt_Config(1, 4, 0, 0); /* port 4 input */
    DIG_Out_Prt(1, 3, 0xff); /* set all outputs high */
    DIG_In_Prt(1, 4, &nPattern); /* returns 0xff */
    Config_DAQ_Event_Message (1, 1, "DI4", 8, 0xffff, 0xffff, 0, 0, 0, 0, 0, (unsigned long)(&myCallback)); /* callback on state change */
    Now later in my program I perform the following (again as a test) when a button is pushed on a dialog box.
    CDigioDlg:nKillBtn()
    DIG_Out_Line(1, 3, 0, 0);
    DIG_In_Prt(1, 4, &nPattern); /* returns 0xfe */
    My callback funtion never gets called on the state change. Now looking at the documentation for the Config_DAQ_Event_Message it says that
    it notifies traditional NI-DAQ applications when the status of an asynchronous DAQ operation (initiated by a call to DAQ_Start, DIG_Block_Out, WFM_Group_Control, and so on) meets certain criteria you specify.
    Eventually this routines will need to monitor these input lines continuously within a DLL. I want to call an initialization function within
    the DLL to configure the board and callbacks then return. I want the callback functions within the DLL to handle these state changes.
    Any help would be greatly appreciated,
    John

    Hello John. Thank you for contacting National Instruments. It looks like you are using the callback incorrectly. I found an example that uses the callback in Visual C++ with Traditional DAQ. This isn't the same setup as you are using, but it should give you a good idea of how to setup the callback. Please let me know if you have any questions. Have a great day!
    Marni S.
    National Instruments
    Attachments:
    DAQEvent3VC6.zip ‏5 KB

  • Process in FIBF for account statement...

    In FIBF ,click on settings-process module of sap application..we could see many process .I need one process which could be used for account statement.That is customer or vendor account statement.
    Thanks and Regards

    perhaps try 2810 or 2820?
    A.

  • Oracle + Unix question , Error validation for SQL statements in co process

    Hi,
    If I use co process in a shell script , what is the best way to do the error validation of the execution of any sql statements . I was trying to change the following code to make use of co process concept. When I tried echo $? after the execution of a "wrong statements"
    it is still giving Zero output ( meaning success)
    <<<<<<<<<<<<<<<<<<<<<<<<<<<
    cat ${CFG_DIR}/srs_indx.lst | egrep -v '^#' | egrep -v '^[    ]*$' | while read x
    do
    echo " Processing values :: pre $pre tbl $tbl indx $indx tblspc $tblspc cols $cols param $param" >> ${LOGS_DIR}/srs_indx.log
    sqlplus ${BIZSTG} << EOT >> ${LOGS_DIR}/srs_indx.log
    set verify on timing on term on echo on feedback on serveroutput on
    WHENEVER SQLERROR CONTINUE
    drop index $indx;
    WHENEVER SQLERROR EXIT FAILURE
    alter session set query_rewrite_enabled = true;
    create $pre index $indx on $tbl ($cols)
    tablespace $tblspc
    $param;
    exit
    EOT
    RC=$?
    if ( test $RC -ne 0 )
    then
         ERR_MSG="ERROR in creating index $indx for table $tbl from srs_indx.ksh of $ENVIR : $APP by `whoami`@`hostname` on `date` "
         echo $ERR_MSG >> ${LOGS_DIR}/srs_indx.log
         process_warning ${LOGS_DIR}/srs_indx.log
         exitstat=1
    else
         echo "$indx created at `date`" >> ${LOGS_DIR}/srs_indx.log
    fi
    done
    >>>>>>>>>>>>>>>>>>>>
    Any help will be appreciated .
    with thanks and regards
    Ranjeesh K R

    Hi,
    Thanks for the response, I guess people misunderstood my question .
    My question was about "error handling in case of CO PROCESS". For those who don't know about this, in case CO PROCESS you just need to login once to Oracle. In my above code It is logging in & out for each entry in the list. CO process statements may look Similar to SQL statements in pro*C. In PRO*C we use EXEC, but here print -p statements..
    a sample is given for those who are new to co process.
    sqlplus -s /nolog |& # Open a pipe to SQL*Plus
    print -p -- 'connect user/password@instance'
    print -p -- 'set feed off pause off pages 0 head off veri off line 500'
    print -p -- 'set term off time off'
    print -p -- "set sqlprompt ''"
    print -p -- "select sysdate from dual;"
    read -p SYSDATE
    print -p -- "select user from dual;"
    read -p USER
    print -p -- "select global_name from global_name;"
    read -p GLOBAL_NAME
    print -p -- exit
    echo SYSDATE: $SYSDATE
    echo USER: $USER
    echo GLOBAL_NAME: $GLOBAL_NAME
    But I got stuck during error handling of these SQL statements, especially the $? / $* returns 0 even after an erroneus execution.
    Regards

  • BPEL process state "Off"

    Hi,
    I had created one BPEL process which reads from FTP Remote server and writes to local file directory.
    After reading from Remote FTP server I am deleting file from the server. In the remote server I had created on user to interact with FTP server with all permissions. When the BPEL process executed FTP Adapter is deleting the File successfully from remote location but BPEL state is changed to “Off”.
    When I checked the logs, I found there is problem with Delete permission. Actually as mentioned above file is deleted successfully in the remote location.
    OPMN logs:
    <File Adapter::Outbound> Processing file : /click/xelusparts/feeds/jabilparts/outbound/process/\po_28.txt
    <File Adapter::Outbound> Sending message to Adapter Framework for posting to BPEL engine: {
    file=/click/xelusparts/feeds/jabilparts/outbound/process/\po_28.txt
    <2007-06-11 19:00:42,677> <DEBUG> <default.collaxa.cube.engine.deployment> <CubeProcessHolder::loadAndBind> Check if process FTPRetrieveTOFileWriteProcess, revision 1.0 needs to be reloaded
    <2007-06-11 19:00:42,709> <DEBUG> <default.collaxa.cube.ws> <WSDLManager::getWSDL> registered wsdl at http://dt1521-localhost.com:80/orabpel/default/FTPRetrieveTOFileWriteProcess/1.0/_FTPRetrieveTOFileWriteProcess.wsdl
    <2007-06-11 19:00:42,709> <DEBUG> <default.collaxa.cube.ws> <WSDLManager::getWSDL> got wsdl at: http://dt1521-localhost.com:80/orabpel/default/FTPRetrieveTOFileWriteProcess/1.0/_FTPRetrieveTOFileWriteProcess.wsdl
    <2007-06-11 19:00:42,724> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Sucessfully sent message : /click/xelusparts/feeds/jabilparts/outbound/process/\po_28.txt to Adapter Framework.
    <2007-06-11 19:00:42,724> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Deleting file : /click/xelusparts/feeds/jabilparts/outbound/process/\po_28.txt after processing.
    <2007-06-11 19:00:42,724> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Host name is 'DT1521-localhost.com'.
    <2007-06-11 19:00:42,724> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Reading reply from DT1521-localhost.com
    <2007-06-11 19:00:42,818> <DEBUG> <default.collaxa.cube.ws> <WSInvocationManager::invoke> operation: Write, partnerLink: <partnerLink name="FileWriter" partnerLinkType="{http://xmlns.oracle.com/pcbpel/adapter/file/Write/}Write_plt">
    <myRole name="null">
    <ServiceName>null</ServiceName>
    <PortType>null</PortType>
    <Address>null</Address>
    </myRole>
    <partnerRole name="Write_role">
    <ServiceName>null</ServiceName>
    <PortType>{http://xmlns.oracle.com/pcbpel/adapter/file/Write/}Write_ptt</PortType>
    <Address>null</Address>
    </partnerRole>
    <conversationId>bpel://localhost/default/FTPRetrieveTOFileWriteProcess~1.0/10610-BpInv1-BpSeq1.3-3</conversationId>
    <properties>{}</properties>
    </partnerLink>
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <WSDLManager::getWSDL> registered wsdl at D:\OraHome_1\integration\orabpel\domains\default\tmp\.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar\Write.wsdl
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <WSDLManager::getWSDL> got wsdl at: D:\OraHome_1\integration\orabpel\domains\default\tmp\.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar\Write.wsdl
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <WSInvocationManager::invoke> def is file:/D:/OraHome_1/integration/orabpel/domains/default/tmp/.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar/Write.wsdl
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <WSIFInvocationHandler::invoke> opName=Write, parnterLink=<partnerLink name="FileWriter" partnerLinkType="{http://xmlns.oracle.com/pcbpel/adapter/file/Write/}Write_plt">
    <myRole name="null">
    <ServiceName>null</ServiceName>
    <PortType>null</PortType>
    <Address>null</Address>
    </myRole>
    <partnerRole name="Write_role">
    <ServiceName>{http://xmlns.oracle.com/pcbpel/adapter/file/Write/}Write</ServiceName>
    <PortType>{http://xmlns.oracle.com/pcbpel/adapter/file/Write/}Write_ptt</PortType>
    <Address>null</Address>
    </partnerRole>
    <conversationId>bpel://localhost/default/FTPRetrieveTOFileWriteProcess~1.0/10610-BpInv1-BpSeq1.3-3</conversationId>
    <properties>{}</properties>
    </partnerLink>
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <WSIFInvocationHandler::doShortCut> Parner Property optShortCut
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> ResourceAdapterMetaData Connection Created
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <AdapterFramework::Outbound> file:/D:/OraHome_1/integration/orabpel/domains/default/tmp/.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar/Write.wsdl [ Write_ptt::Write(opaque) ] - Looking up Resource Adapter JDNI location 'eis/FileAdapter'
    <2007-06-11 19:00:42,834> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Managed Connection Created
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Creating Connection
    <2007-06-11 19:00:42,834> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Connection Created
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Creating FileInteraction
    <2007-06-11 19:00:42,834> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> FileInteraction Created
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <AdapterFramework::Outbound> Instantiating outbound JCA interactionSpec oracle.tip.adapter.file.outbound.FileInteractionSpec
    <2007-06-11 19:00:42,834> <DEBUG> <default.collaxa.cube.ws> <AdapterFramework::Outbound> Populating outbound JCA interactionSpec oracle.tip.adapter.file.outbound.FileInteractionSpec with properties: {FileNamingConvention=spob230_%SEQ%.txt, OpaqueSchema=true, ElapsedTime=60, NumberMessages=1, PhysicalDirectory=c:\inbound_dir, FileSize=1024000}
    07/06/11 19:00:42 file:/D:/OraHome_1/integration/orabpel/domains/default/tmp/.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar/fileAdapterOutboundHeader.wsdl
    07/06/11 19:00:42 file:/D:/OraHome_1/integration/orabpel/domains/default/tmp/.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar/fileAdapterOutboundHeader.wsdl
    07/06/11 19:00:42 file:/D:/OraHome_1/integration/orabpel/domains/default/tmp/.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar/Write.wsdl
    <2007-06-11 19:00:42,849> <DEBUG> <default.collaxa.cube.ws> <AdapterFramework::Outbound> file:/D:/OraHome_1/integration/orabpel/domains/default/tmp/.bpel_FTPRetrieveTOFileWriteProcess_1.0.jar/Write.wsdl [ Write_ptt::Write(opaque) ] - Invoking JCA outbound Interaction
    <2007-06-11 19:00:42,849> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> File Adapter received an input Record
    <2007-06-11 19:00:42,849> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Waiting outside sync object.....
    <2007-06-11 19:00:42,849> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Batcher created using staging : true with number of messages : 1
    <2007-06-11 19:00:42,849> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Input Record: javax.xml.transform.dom.DOMSource@e5394c
    <2007-06-11 19:00:42,849> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Translating....
    <2007-06-11 19:00:42,849> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Outbound Translation done.
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Refreshed (staging/num Messages) file count : 1
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Host DT1521-localhost.com FTP command: USER orabpel
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Wrote to Control File
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Batching staging File: staging0
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Batcher sending to sender for output file writing
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Sending File to output Dir
    <2007-06-11 19:00:42,865> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> FileAgent.putFile() puts spob230_17.txt in c:\inbound_dir begins ...
    <2007-06-11 19:00:42,865> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Done Writing to output file spob230_17.txt in dir c:\inbound_dir !!
    <2007-06-11 19:00:42,881> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> End clearing Control File
    <2007-06-11 19:00:42,881> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Deleting staging File: staging0
    <2007-06-11 19:00:42,881> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> End clearing Control File
    <2007-06-11 19:00:42,881> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Interaction gets a ok response
    <2007-06-11 19:00:42,881> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Done with Input Record
    <2007-06-11 19:00:43,256> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> FTP Command: USER, reply:
    331 Password required for orabpel
    <2007-06-11 19:00:43,256> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Host DT1521-localhost.com FTP command: PASS
    <2007-06-11 19:00:43,506> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> FTP Command: PASS, reply:
    230 Login OK
    <2007-06-11 19:00:43,506> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Host DT1521-localhost.com FTP command: CWD /click/xelusparts/feeds/jabilparts/outbound/process/
    <2007-06-11 19:00:43,615> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> FTP Command: CWD, reply:
    250 Directory successfully changed
    <2007-06-11 19:00:43,615> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Host DT1521-localhost.com FTP command: DELE po_28.txt
    <2007-06-11 19:00:43,709> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> FTP Command: DELE, reply:
    200 Deleted file "po_28.txt".
    <2007-06-11 19:00:43,709> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> FTPClient.deleteFile(): 200 Deleted file "po_28.txt".
    <2007-06-11 19:00:43,709> <ERROR> <default.collaxa.cube.ws> <File Adapter::Outbound> FTPClient.deleteFile(): Unable to delete file po_28.txt', reply code = 200
    <2007-06-11 19:00:43,709> <DEBUG> <default.collaxa.cube.ws> <File Adapter::Outbound> Processer thread calling onFatalError with exception Error deleting file from FTP Server.
    Unable to delete file from server.
    Please ensure whether the remote file has delete permission.
    <2007-06-11 19:00:43,709> <FATAL> <default.collaxa.cube.activation> <AdapterFramework::Inbound> [Get_ptt::Get(opaque)] - Resource Adapter requested Process shutdown!
    <2007-06-11 19:00:43,709> <INFO> <default.collaxa.cube.activation> <AdapterFramework::Inbound> Adapter Framework instance: OraBPEL - performing endpointDeactivation for portType=Get_ptt, operation=Get
    <2007-06-11 19:00:43,709> <INFO> <default.collaxa.cube.ws> <File Adapter::Outbound> Endpoint De-activation called in adapter for endpoint : /click/xelusparts/feeds/jabilparts/outbound/process/
    <2007-06-11 19:00:43,725> <DEBUG> <default.collaxa.cube.engine.deployment> <CubeProcessHolder::loadAndBind> Check if process FTPRetrieveTOFileWriteProcess, revision 1.0 needs to be reloaded
    <2007-06-11 19:00:44,037> <INFO> <default.collaxa.cube.activation> <AdapterFramework::Inbound> JCAActivationAgent::onStateChanged State is changed for process 'bpel://localhost/default/FTPRetrieveTOFileWriteProcess~1.0/', state=OFF
    <2007-06-11 19:00:44,037> <DEBUG> <default.collaxa.cube.engine.deployment> <CubeProcessHolder::updateRevision> Updated state for process FTPRetrieveTOFileWriteProcess (revision 1.0) to 1
    Thanks,
    Vidya

    Hi SR,
    You can do this by having a BPEL process with a java embedding containing below code.
    * Input Variables
    String processName   = (String)getVariableData("processName");                                                                                 
    String revision      = (String)getVariableData("revision");             
    * Declare variables.
    IBPELProcessHandle procHandle = null;            
    Locator loc = null;            
    BPELProcessMetaData bpelProcessMetadata = null;             
    try{               
         * get Locator Instance
        loc = getLocator();                            
         * Lookup Process. Revision optional.
        if(revision == null || revision.trim().equals(""))            
          procHandle = loc.lookupProcess( processName );             
        else            
          procHandle = loc.lookupProcess(processName, revision);             
        if(procHandle == null)          
          throw new Exception("Unable to find process: " + processName);          
         * Get Metadata of the process.
        bpelProcessMetadata = procHandle.getMetaData();             
         * Set Lifecycle to Retired.
         * Use setState(IBPELProcessConstants.STATE_OFF) to change process state to off.
        bpelProcessMetadata.setLifecycle(IBPELProcessConstants.LIFECYCLE_RETIRED);    
         * Finally update the process with the modified metadata.
        procHandle.updateMetaData(bpelProcessMetadata);                   
    }catch (Exception e)                            
      addAuditTrailEntry(e);              
    }If you want you can schedule this process.
    Regards
    PavanKumar.M

  • Background job is taking lot of time for processing the job.

    One background job - which is processing Idocs is processing a job for more than 2000+ seconds.. and completed tho.
    But this is happening for the past few days.. is there any way to trouble shoot, or find from any logs of this completed job, to find out why it was taking lot of time for processing.
    Can you please tell me the steps of analyzing / trouble shooting why it was taking lot of time daily.
    Regards,
    Satish.

    Hi Satish,
    Run DB stat from db13 it will improve the performance.
    Check number of idocs. You can send part by part, instead of sending huge data.
    Check SM58 logs
    Suman

  • Design Time for Processes and Forms

    Hi experts,
    When trying to open  "Design Time for Processes and Forms" in order to create a process, I get a dump.
    Could you let me know what could be wrong ? config? etc..
    Hereunder is a sample of what I found in T.code ST22
    The following checkpoint group was used: "No checkpoint group specified"
    If in the ASSERT statement the addition FIELDS was used, you can find
    the content of the first 8 specified fields in the following overview:
    " (not used) "
    " (not used) "
    " (not used) "
    " (not used) "
    " (not used) "
    " (not used) "
    " (not used) "
    " (not used) "
        referenced node must exist
          READ TABLE object_hierarchy_nodes TRANSPORTING NO FIELDS
                     WITH KEY node_id = object_hierarchy_node_wa-ref_node_id
    >>>      ASSERT sy-subrc = 0.
        ENDIF.
      provider has to be specified (except in case of recursion)
        IF object_hierarchy_node_wa-ref_node_id IS INITIAL.
          ASSERT object_hierarchy_node_wa-provider_id IS NOT INITIAL.
        ENDIF.
    Thanks in advance,
    Louis

    note: Note 1075650
    Re: Assertion_Failed dump when accessing Design Time for Processes & Forms EP3
    regards,
    Prakesh.

  • Replacement code for Translate statement in latest version

    Replacement code for
    TABLES XXXX  USING <r> LIKE <r>
    TRANSLATE <r>-<fieldname> TO UPPER CASE.
    <r> is a structure

    Hi,
    TRANSLATE ... TO UPPER/LOWER CASE
    TRANSLATE ... USING
    The arguments of these instructions must be single fields of type C, N, D, T or STRING or purely character-type structures. There is a syntax or runtime error if arguments of a different type are passed. A subset of this function is provided with the addition IN BYTE MODE for processing byte strings – that is, operands of type X or XSTRING. A statement such as CONCATENATE a x b INTO c is thus no longer possible when a, b, and c are all character-type, but x is of type X.
    TRANSLATE ... CODEPAGE ...
    TRANSLATE ... NUMBER FORMAT ...
    The above statements are not allowed in Unicode programs
    <b>Reward points</b>
    Regards

  • PO Processing State

    Dear All,
    Can u tell me all the States of PO Processing as
    Purchasing document processing state    03
    here what it Mean by 03.
    Please help me
    Regards,
    Pardeep malik

    Dear,
    The system always creates a new version when the purchasing document has reached a final state.
    This means one of the following processing states:
    ●     Active
    If you do not use the document release functionality
    ●     Release completed
    If you use the document release functionality
    For a new version of a purchasing document to be created, all messages must have also been transmitted.
    No new versions are generated for purchase orders that are on hold.
    Not every change to an external purchasing document results in a new version. The following changes lead to a new version:
    ●     All print-relevant changes
    You define which changes are print-relevant in Customizing for Purchasing under Messages ® Fields Relevant to Printouts of Changes.
    ●     Changes in document assignment
    ●     Changes in address
    Address changes result in a new version only if the field ADRC-NAME1 has been defined as a field that is relevant to printouts of changes.
    ●     Changes to master conditions
    Changes in master conditions result in a new version only if the field EKPO-NETPR has been defined as a field that is relevant to printouts of changes.
    Version Data
    You will find the version data:
    ●     In the item details on the Versions tab page in Enjoy purchase requisitions (ME51N, ME52N, ME53N)
    ●     In the header data on the Versions tab page of Enjoy purchase orders (ME21N, ME22N, ME23N)
    ●     Under Header ® Versions in RFQs, conventional purchase orders (ME21, ME22, ME23), contracts, and scheduling agreements.
    There you can maintain version data (e.g. reason for change) and complete the version. In addition, the Value change field shows you whether and to what extent the value of a requisition or external purchasing document item has changed in comparison with the previous version.
    You can display the change documents for one or more versions by selecting the relevant versions and choosing  Display Changes.
    Please check accordingly.
    Regards,
    Syed Hussain.

  • Modification of std form F140_ACC_STAT_01 for account statement

    Hi
    I want to do modification of std form F140_ACC_STAT_01 for account statement .
    Now can any body tell me how to find or what is the Application type for transaction NACE Output type and processing routine for this std form and how to find out all these info .
    So in NACE>Application Type>Output Type>->processing routine>
    Pl help , answers will be awarded
    rgds
    mojib

    it was actually thrown in the wrong forum

  • Block Customer for Account Statement : Urgent pls

    Can anyone suggest if there is any option to block a customer from being sent Account Statement.
    We currently have a Z program, where we give a range of customers in the selection screen and it will print statements for all the customer.
    But if we dont want to send statement to a particular customer, how can we do that..
    Is there any similar option like Dunning block, to block a customer from being sent a statement.
    Please let me know

    In the customer master data, company code tab. In the Account statement field you can select the Indicator for periodic account statements 1 or 2. You can also add one for no statement. When you run F.27 it will process statements based on the Indicator for periodic account statements.
    pls assign points to say thanks.

  • Updates getting stuck in V2 processed state

    Hi,
    In our ERP production system EP0, update entries are getting stuck
    in 'V2 Processed' state.
    Details of the same show that the collective run state is in initial
    state for functional module 'MCEX_UPDATE_03'.
    Please find the attached screenshot of details.
    These updates are gettin stuck for specific transactions like MIGO.
    We have already tried running the report RSM13005 but it has not helped.
    Please look into it urgently as the system is a production system.
    Thanks in advance.
    Regards,
    Suresh
    Edited by: suresh k on Sep 26, 2008 2:38 PM

    Hi
    As you checked in transaction SM13 they are stucked in Initial state. There was various states associated with the update like INIT and ERR.
    What is INIT status:
    Update records which have not yet been processed have the status INIT or AUTO. INIT refers to a complete update record (header, function module call of update and data elements are present), which has not yet been processed however.
    What to do: SM13 -> Update records -> Update -> All records/Single
    you can set the update administration so that all records with the status 'INIT' or AUTO' (all update records which have not yet been updated) are processed (as far as possible).
    Warnings: please check with the user who posted the updates in the first place - make sure that they have not already reposted the update manually. If you are in doubt, please re-enter the data manually rather than using this update function.
    Hope this helps!!
    Kind Regards
    Esha
    Edited by: Esha Rajpal on Sep 30, 2008 5:06 AM

  • Technical content activation for BI STAT and Admin cockpit

    Hi friends,
    I am going to setup the IMG call for BI STAT and Admin cockpit. If i run the technical content activation will i need a precaution for other data loading. If i run them in background does it affects other data loads. did i need to take down time for the whole system?
    After the implemntation did i need to load data for all infoproviders in quality system.

    Hello Suresh,
         I am not clear on all your questions.
    In general, technical content process chains load data similiar to other BI process chains.
    if your system is productive for a longer period already ( > 6 months ), there may be a lot of
    BI statistics data generated and this will take a longer time to load the technical content cubes during the initial load.
    Otherwise, it should not affect system performance.
    It is better to schedule them daily to avoid huge data loads.
    But these are only guidelines. It depends on your need.
    You would need to do separately for quality and production system, if you want the BI stats in those systems. This is similiar to any other Data load and cube and content in BI.
    Thanks,
    Bilal

  • Is there any System setting for Update statement

    Is there any setting for update statement related to implicit commit because Update statement in my code behaves differently in different systems.
    My code is
    Data:  it_eaus like eaus occurs 0 with header line,
            wa_eaus type eaus.
    select * from eaus into table it_eaus where auszbeleg = '000000000001'  and storausz = 'X'.
    if sy-subrc = 0 .
    wa_eaus-aedat = ''.
    wa_eaus-aenam = ''.
    wa_eaus-storausz = ''.
    modify it_eaus from wa_eaus transporting aedat aenam storausz where auszbeleg = '000000000001'.
    update eaus from table it_eaus.
    endif.
    Data:  it_eausv like eausv occurs 0 with header line,
            wa_eausv type eausv.
    select * from eausv into table it_eausv where auszbeleg = '000000000001'  and storausz = 'X'.
    if sy-subrc = 0 .
    wa_eausv-aedat = ''.
    wa_eausv-aenam = ''.
    wa_eausv-storausz = ''.
    modify it_eausv from wa_eausv transporting aedat aenam storausz where auszbeleg = '000000000001'.
    update eausv from table it_eausv.
    endif.

    Hi
    It isn't an alternative way: u can only explicit the COMMIT just as I said before:
    Update DBTAB set COL.
    COMMIT WORK.
    Anyway just I said before the debugger can execute a COMMIT, from SAP help:
    The New Debugger can run in exclusive and non-exclusive mode. Exclusive mode means that the application that is being analyzed exclusively occupies a work process of the application server during debugging. In non-exclusive mode, the Debugger functions are limited. In non-exclusive mode, after each Debugger interaction, the system requests a roll-out in the application. Therefore, an implicit database commit must be executed. This has the following consequences:
    u25CF      Debugging is not possible between the statements SELECTand ENDSELECTbecause the database cursor needs to be closed when using an implicit database COMMIT statement.
    In this case, program execution is terminated.
    ·        Debugging is not possible for conversion or field exits.
    ·        Due to the implicit database commit, inconsistent datasets can occur in the database.
    For this reason, non-exclusive mode is not possible in productive systems.
    Max

  • Callback for specific step group

    Hi,
    I have a sequence, Main sequence. In this sequence, there is Setup, Main and Cleanup step group. Setup is to initialise my devices, Main is to perform the tests on my UUTs, Cleanup is to stop and clear task for my devices. I would like to have a TS sequence in such a way that it runs Setup and Cleanup once and continue to loop Main. With this setting, I would still want to have the UUT Information window popup.
    I have tried using SequenceLoadFile callback and put my vis for initialising my deivces (found in Setup) in this callback. It works when I first open my sequence file. However, I would like to know if I can do it in such a way that everytime I click Run (F5), it will run my initialise vis (found in Setup) once and keep looping Main with popup of UUT Information window for each UUT and run Cleanup with I click Stop on the UUT Information window.
    Could anyone give me some advise on how I can do so??
    Thank you & Regards
    Lee

    Process Setup Callback and Process Cleanup Callback.  They are model callbacks that you can add to your client sequence and put your initialize and cleanup code in.  You add them the same way that you added the Sequence File Load callback (which, by the way, is an engine callback). 
    The Purple sequences are engine callbacks.  The Green sequences are Model callbacks.
    It doesn't matter if you override engine callbacks like you are doing.  It just matters what kind of behavior you want.  In your case you need the two model callbacks I mentioned at the first.
    Hope that helps,
    jigg
    CTA, CLA
    teststandhelp.com
    ~Will work for kudos and/or BBQ~

Maybe you are looking for

  • Learn Photoshop CS5.1 or upgrade to CC?

    Hey Guys, I need some advice... I purchased Photoshop CS5.1 a few years ago and never had the resources to learn the program.  I found the videos on YouTube really did not help me for the most part.    I just recently purchased a Canon Rebel T5i and

  • How to set a bind variable into a vo query from the current vo ?

    I would like to exclude from my list of values defined in a view object the already referenced items in the current view object content. Let's say i have dept and emp with a many to many relationships, I would like that the already existing dept for

  • ITunes not updating sorting in iTunes Music folder

    I have enabled "Keep iTunes Music folder organized". iTunes will properly sort files based on the ID3 data when I first import them, but it does not update the filenames or folder names when I change song information from within iTunes. I don't know

  • HOW TO SEND THE REPORTS AND  THROUGH XI TO NON -SAP SYSTEM

    Hi Experts,              I want to know how to send the reports and  smartform through xi to another system. All the post blogs are   just explaning only about the FILE-TO-FILE, FILE-TO-IDOCS scenarios only.

  • Account Hierarchy error while creating Funds,

    Hi All, I am working on CRM 7.0. I have created Account Hierarchy tree, Hierarchy Node & Subnode and assigned Accounts to node and subnode using transaction BPH. Now while i select this hierarchy in front end to create Fund in TPM, I am getting error