Map data from class CL_PT_REQ_WF_ATTRIBS with workflow

Hi All,
I am using the standard class CL_PT_REQ_WF_ATTRIBS with DUMMY method for approval workitem which uses webdynpro ABAP. The webdynpro person is passing one variable RESULT after the approval, and RESULT with its value is shown in the log but there is no workflow container element named RESULT in the workflow. so, I created one with that name. But when I use this vaiable in the task after mapping, it is not showing any value. Am I missing something or doing anything wrong?

Hi
After creating the  container element in Approve Leave request Task did you try to refresh the buffer and Refresh Org assignement.
And also make sur ethat the webdynpro developer is writing  back the decision to Approve Leave request Task
and also make sure you do a proper bidning between Task container to workflow container. Check in the workflow log whteher in atleast the task container the value is populate. Also please let us know how the webdynpro developer is assigning the result value is it something like he  is using SAP_WAPI_WRITE_CONTAINER function module then make sure that the container element is CAse Sensitive. so ask him to define as you decalred in the task/workflow container.
Regards
Pavan
Edited by: Pavan Bhamidipati on May 17, 2011 2:18 PM
Edited by: Pavan Bhamidipati on May 17, 2011 2:18 PM
Edited by: Pavan Bhamidipati on May 17, 2011 2:19 PM

Similar Messages

  • Pass the data from a form in workflow to z table

    Hi all,
    How can I pass the data from a form in workflow to z table?

    Thanks for your answer, I tried the solution 2, I create "Submit" button, and ser the mapping scope to  be "All data rows", it only works when I select at least one row, otherwise the data would not be passed.
    Another question is I have serveral imported table parameter, for each table I have one "submit" event, I want these tables to be submitted at the same time, but if I click the submit button in one table toolbar, I can only submit the table data which has a submit button clicked, for other tables, the data is not passed, how can I achieve it?
    Thanks.

  • Removing complete Challan Mapping data from system

    Dear Consultants,
    My client has entered erroneous data for challan mapping. When I checked the data, I found that there is some difference in the Challan amount & available amount. It means there are some employees mapped for these challans. but the main page of CMAP shows employees mapped = 0. When I tried to delete the challan data, system says, there are some employees mapped to these challans.
    When on the last node of CMAP, I tried to select all employees on the right hand & click on the arrow from right to left, no amount of challans was changed.
    May I know how to delete the entire mapping data from the system?
    regards,
    Bindumadhav

    Dear MVR,
    The reply by Mr.Deepak is as follows. :
    Dear,
    Regarding your query, the issue reported by you basically happens due to inconsistent mapping or unmapping of the employees to the Challan.
    We would suggest you to kindly delete of the negative entry from the table 'T7INTE'. You also need to delete the entries from the table PCL4 as well for the corresponding challan for which the negative entry has been created. Performing the above suggested steps will remove the reported inconsistencies. After deleting the entries completely, kindly perform the Challan mapping for the employees again. Please make sure that you are not deleting the entry from the table for which Form24Q
    has already been generated. The deletion process can be explained as such:
    1. Execute the transaction 'se16'. Go to the tables mentioned above. You can execute the table 'T7INTE' with the TAN number. On execution, a negative entry will be displayed for a particular challan. Select that entry and delete of the entry from the table.
    2. Go to table PCL4. Execute the table with relid as 'I1' to 'I2' and the TAN number in the field 'SRTFskindly performing the mapping action again. Hopefully, following the above mentioned steps will resolve the reported inconsistency and you will be able to perform the mapping successfully.
    Best Regards,
    Deepak..
    I can see some entries. But the system does not allow me to delete the table entry from T7INTE & PCL4.
    Regards,
    Bindumadhav

  • Regarding how to pass the data from web dynpro to workflow

    hi gurus,
    how to pass the data from web dynpro to workflow.
    Regards
    vijay

    Check this [thread|SAP_WAPI_START_WORKFLOW;

  • How to pass the data from web dynpro to workflow.

    hi gurus,
    how to pass the data from web dynpro to workflow.
    Regards
    vijay

    Hi
    you can use function module
    data   ls_input_container  TYPE swr_cont.
    data   lt_input_container  TYPE TABLE OF  swr_cont.
    CALL FUNCTION 'SAP_WAPI_START_WORKFLOW'
        EXPORTING
          task            = ptask
        IMPORTING
          return_code     = lv_return_code
          new_status      = lv_new_status
        TABLES
          input_container = pinput_container
          message_lines   = lt_message_lines
          message_struct  = lt_message_struct.
    where you pass the data in imnternal table "pinput_container" as
      ls_input_container-element = 'KUNNR'.
      ls_input_container-value = ls_skna1-kunnr ."wd_this->lv_kunnr.
      APPEND ls_input_container TO lt_input_container .
      ls_input_container-element = 'CLUSER'.
      ls_input_container-value = lv_cluser.
      APPEND ls_input_container TO lt_input_container .
      ls_input_container-element = 'BUKRS'.
      ls_input_container-value = lv_bukrs. " youe value as per requirement.
      APPEND ls_input_container TO lt_input_container .
      ls_input_container-element = 'VKORG'.
      ls_input_container-value = ls_sknvv-vkorg. " youe value as per requirement
      APPEND ls_input_container TO lt_input_container .
      ls_input_container-element = 'VTWEG'.
      ls_input_container-value = ls_sknvv-vtweg. "youe value as per requirement.
      APPEND ls_input_container TO lt_input_container .
      ls_input_container-element = 'SPART'.
      ls_input_container-value = ls_sknvv-spart. "youe value as per requirement.
      APPEND ls_input_container TO lt_input_container .
    *Also Forgot to mention where ptask is your workflow ID *
    Regards,
    Arvind
    Edited by: Arvind Patel on May 14, 2010 7:38 AM

  • Reading Data from a Table With an Expert

    Hi all - I'm trying to write an Expert in OWB that would read data from a table and create a new table based on that information. I have the table creation part down, but how can I read data from a table with an Expert? From what I've read, Experts only deal with metadata, and there is no mechanism in Experts to actually read data in the tables. Does OraTcl work in Experts? Has anyone ever came across this problem, and if so, how did you solve it? Thanks in advance for your time! - Don

    Hi Don
    Can also use Java and JDBC from within Tcl, see the routines in this file below, take modify or whatever;
    http://blogs.oracle.com/warehousebuilder/ombora.tcl
    source <dir>\ombora.tcl
    set g_user scott
    set g_upwd zzzzzz
    set g_hostname localhost
    set g_port 1521
    set g_srvname ora111
    oraconnect $g_user $g_upwd $g_hostname:$g_port:$g_srvname
    oraselect "select * from emp" ""
    As well as dumping the results to stdout (you probably want to comment that part out and change any other part you don't like), g_res is a ResultSet object you can use to do whatever.
    Cheers
    David

  • Cartesian of data from two tables with no matching columns

    Hello,
    I was wondering – what’s the best way to create a Cartesian of data from two tables with no matching columns in such a way, so that there will be only a single SQL query generated?
    I am thinking about something like:
    for $COUNTRY in ns0: COUNTRY ()
    for $PROD in ns1:PROD()
    return <Results>
         <COUNTRY> {fn:data($COUNTRY/COUNTRY_NAME)} </COUNTRY>
         <PROD> {fn:data($PROD/PROD_NAME)} </PROD>
    </Results>
    And the expected result is combination of all COUNTRY_NAMEs with all PROD_NAMEs.
    What I’ve noticed when checking query plan is that DSP will execute two queries to have the results – one for COUNTRY_NAME and another one for PROD_NAME. Which in general results in not the best performance ;-)
    What I’ve noticed also is that when I add something like:
    where COUNTRY_NAME != PROD_NAME
    everything is ok and there is only one query created (it's red in the Query plan, but still it's ok from my pov). Still it looks to me more like a workaround, not a real best approach. I may be wrong though...
    So the question is – what’s the suggested approach for such queries?
    Thanks,
    Leszek
    Edited by xnts at 11/19/2007 10:54 AM

    Which in general results in not the best performanceI disagree. Only for two tables with very few rows, would a single sql statement give better performance.
    Suppose there are 10,000 rows in each table - the cross-product will result in 100 million rows. Sounds like a bad idea. For this reason, DSP will not push a cross-product to a database. It will get the rows from each table in separate sql statements (retrieving only 20,000 rows) and then produce the cross-product itself.
    If you want to execute sql with cross-products, you can create a sql-statement based dataservice. I recommend against doing so.

  • How to extract data from XML file with JavaScript

    HI All
    I am new to this group.
    Can anybody help me regarding XML.
    I want to know How to extract data from XML file with JavaScript.
    And also how to use API for XML
    regards
    Nagaraju

    This is a Java forum.
    JavaScript is something entirely different than Java, even though the names are similar.
    Try another website with forums about JavaScript.
    For example here: http://www.webdeveloper.com/forum/forumdisplay.php?s=&forumid=3

  • Select max date from a table with multiple records

    I need help writing an SQL to select max date from a table with multiple records.
    Here's the scenario. There are multiple SA_IDs repeated with various EFFDT (dates). I want to retrieve the most recent effective date so that the SA_ID is unique. Looks simple, but I can't figure this out. Please help.
    SA_ID CHAR_TYPE_CD EFFDT CHAR_VAL
    0000651005 BASE 15-AUG-07 YES
    0000651005 BASE 13-NOV-09 NO
    0010973671 BASE 20-MAR-08 YES
    0010973671 BASE 18-JUN-10 NO

    Hi,
    Welcome to the forum!
    Whenever you have a question, post a little sample data in a form that people can use to re-create the problem and test their ideas.
    For example:
    CREATE TABLE     table_x
    (     sa_id          NUMBER (10)
    ,     char_type     VARCHAR2 (10)
    ,     effdt          DATE
    ,     char_val     VARCHAR2 (10)
    INSERT INTO table_x (sa_id,  char_type, effdt,                          char_val)
         VALUES     (0000651005, 'BASE',    TO_DATE ('15-AUG-2007', 'DD-MON-YYYY'), 'YES');
    INSERT INTO table_x (sa_id,  char_type, effdt,                          char_val)
         VALUES     (0000651005, 'BASE',    TO_DATE ('13-NOV-2009', 'DD-MON-YYYY'), 'NO');
    INSERT INTO table_x (sa_id,  char_type, effdt,                          char_val)
         VALUES     (0010973671, 'BASE',    TO_DATE ('20-MAR-2008', 'DD-MON-YYYY'), 'YES');
    INSERT INTO table_x (sa_id,  char_type, effdt,                          char_val)
         VALUES     (0010973671, 'BASE',    TO_DATE ('18-JUN-2010', 'DD-MON-YYYY'), 'NO');
    COMMIT;Also, post the results that you want from that data. I'm not certain, but I think you want these results:
    `    SA_ID LAST_EFFD
        651005 13-NOV-09
      10973671 18-JUN-10That is, the latest effdt for each distinct sa_id.
    Here's how to get those results:
    SELECT    sa_id
    ,         MAX (effdt)    AS last_effdt
    FROM      table_x
    GROUP BY  sa_id
    ;

  • How to retrive data from selected checkboxes with fieldnames

    hi experts,
    how to retrive data from selected checkboxes with fieldnames into another alv grid report.(here the fieldnames selected from  table names is dynamically).
    thankx in advance
    rani.k.

    Hi,
    Use user_command in the alv grid and then
    do the follwoing code
    FORM user_command1 USING lv_ucomm LIKE sy-ucomm
                      rs_selfield TYPE slis_selfield.
    Declaration of local Variables
      DATA : lv_ref1 TYPE REF TO cl_gui_alv_grid.
      DATA lv_cnt TYPE i.                                    "+INS SUHESH 12.07.2008
    Check function code
      CASE lv_ucomm.
        WHEN 'ONLI'.
          CALL FUNCTION 'GET_GLOBALS_FROM_SLVC_FULLSCR'
            IMPORTING
              e_grid = lv_ref1.
          CALL METHOD lv_ref1->check_changed_data.
    now loop ur final internal table where check = 'X'.
    now pass data to other internal table..Now the internal table will be having values that the user selcetd on the screen.
    Hope this helps.
    Regards,
    Nagaraj

  • How does iCloud manage app data like whatsapp ? Does it overwrite the data everytime I backup to I cloud with the data on the phone or it merges the data from previous backups with the new data?

    How does iCloud manage app data( like whatsapp) ? Does it overwrite the data everytime I backup to I cloud with the data on the phone when i backup or it merges the data from previous backups with the new data?

    You need to subscribe to iTunes Match to store your music on iCloud:
    http://www.apple.com/itunes/itunes-match/
    To transfer iTunes Store purchases made on your phone to your computer, connect your phone then choose "Transfer purchases..." from the File menu in iTunes.
    Photostream transfers photos taken on your iOS devices to other devices and your computer. For troubleshooting Photostream see: http://support.apple.com/kb/TS3989

  • PASS DATA FROM ABAP PROGRAM TO WORKFLOW CONTAINER

    Hello Everybody,
    I am a workflow beginner and I want to pass a data from my ABAP code to a workflow container. This workflow container is used in the Send Mail step.
    The ABAP  Program is:
    REPORT  zdb_work_flow.
    tables : vbak.
    DATA : k TYPE sweinstcou-objkey,
                e TYPE swetypecou-event.
    DATA : st TYPE swcont,
                itab TYPE STANDARD TABLE OF swcont.
    PARAMETERS : vbeln TYPE vbak-vbeln.
    start-of-selection.
      st-element = 'VBELN'.   "This is my WORKFLOW CONTAINER
      st-tab_index = 1.
      st-elemlength = 10.
      st-type = 'C'.
      st-value = vbeln.
      append st to itab.
      k = vbeln.
      e = 'TRIGGER'               "This is an event in my BUS OBJ ZDB_WFLOW
      CALL FUNCTION 'SWE_EVENT_CREATE'
        EXPORTING
          objtype                               = 'ZDB_WFLOW'   "My BUS OBJ
          objkey                                 = k
          event                                  = e    "Event TRIGGER in ZDB_WFLOW
        tables
          event_container               = itab
        exceptions
          objtype_not_found             = 1
          OTHERS                        = 2
      IF sy-subrc <> 0.
        MESSAGE ID sy-msgid TYPE sy-msgty NUMBER sy-msgno
              WITH sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
      ENDIF.         .
      COMMIT WORK.
    Now I have created a workflow in SWDD as ZDB_WF_SW.
    In this WF I have created a container as VBELN.
    In the SEND MAIL I have given a header as "Worflow Triggered for &VBELN&"
    PROBLEM.
    The Workflow is getting triggered. I am also getting a mail "Workflow Triggered for" in my mailbox but the &VBELN& is missing from the header.
    I have set the IMPORT EXPORT flag of the WF CONTAINER "VBELN"
    Please Help.
    Edited by: Deepankar.B on Feb 22, 2011 4:08 PM
    Edited by: Deepankar.B on Feb 22, 2011 4:10 PM

    You wouldn't use a SMA as an assistance class - but refer to one within it.
    I would use Shared Memory Area classes - but I'd also build a SMA that could talk across multiple application servers.
    Thomas Jung proposed this in a forum reply some time ago.
    It shouldn't be too difficult to use RFC's to access the content of SMA's in multiple app servers - ensuring synchronisation across multi app servers. I've not yet had the need, so haven't attempted.
    The long polling in 7.02 apparently uses this approach, so in a 7.02 system there may even be standard components/ SMAs that you could use.
    I do not think global areas in FM's are available across sessions - I don't think you could use that as a data transport.

  • Passing data from user exit to workflow.

    HI,
    Please consider the below scenario.
    In the user exit 'MV50AFZ1' I have created an include in the form 'USEREXIT_SAVE_DOCUMENT_PREPARE.'.
    In this include based on some certain conditions a workflow is triggered.
    I am using the function module 'SWE_EVENT_CREATE' to trigger the workflow.
    Now the requirement is to pass some data from this user exit ('USEREXIT_SAVE_DOCUMENT_PREPARE') into the workflow.
    Could someone please explain me how to pass this data from the user exit into the workflow ?
    Thanks in advance.
    With regards
    Abhishek

    Hello everybody,
    Thanks a lot for your responses.
    I am using the fuction module 'SAP_WAPI_CREATE_EVENT' to trigger the workflow, I am also able to pass the delivery number through the paramater 'OBJECT_KEY' of the function module in to the workflow container.
    But I am unable to pass the line item number and the material number into the workflow.
    I am following the below process:
    I have defined a stucture ZST_DELIVERY containing the POSNR and MATNR in the data dictionnary.
    I have pouplated the internal table of this structure inside my user exit.
    Then I m passing the values of the above internal table into another internal table - 'Container''which is of structure type 'SWR_CONT'.
    And I am passing the internal table 'Container' into the paramater 'INPUT_CONTAINER' of the function module 'SAP_WAPI_CREATE_EVENT'.
    And in the workflow I have created a container 'ST_DELIVERY' which is of the structure ZST_DELIVERY and I have made it 'import', 'export.'
    And in my business object ZLIKP in the custom event CheckPrice I have defined a paramater 'ST_DELIVERY' of the structure  ZST_DELIVERY and I have made it multiline.
    And in I have binded my workflow container 'ST_DELIVERY' and the ZLIKP paramater with the same name.
    Please do let me know if the above steps that I have followed is correct.
    Please do let me know if any further information is required.
    Thanks in advance,
    ~Regards
    Abhishek

  • Approver & approval date from Requisition & PO Approval workflow tables

    Dear All,
    My req is to create a report that calculates the lead time taken by a Purchase Requisition to be converted into a PO and released. In the output they want the APPROVER NAME, DATE APPROVED against the REQUISITION NUMBER (EBAN-BANFN) and also they want APPROVER NAME, DATE APPROVED against the Purchase Order (so all in all 2 fields of approvers & 2 fields of approval dates). But the thing is they want these APPROVER NAME, DATE APPROVED from the Requisition Approval Workflow tables.
    Please update the table & fields that will have the Approver and date approved from workflow corresponding to the Purchase Requisition Number and Purchase Order. or What is the link between the "Approver, approval date & Purchase Requisition Number" and "Approver, approval date & Purchase Order".
    Thanks in advance. The solution will be rewarded the points.
    Warm Regards,
    Farooq.

    We have HR/Payroll and Purchasing working fine for 6 years with the following schedule settings. 0.1 seems too frequent:
    A. Schedule one concurrent request to process deferred activities:
    1. Item Type :
    2. Min Threshold :
    3. Max Threshold :
    4. Process Deferred : Yes
    5. Process Timeout : No
    6. Process Stuck : No
    (Once every five minutes)
    B. Schedule one concurrent request to process timeout activities:
    1. Item Type :
    2. Min Threshold :
    3. Max Threshold :
    4. Process Deferred : No
    5. Process Timeout : Yes
    6. Process Stuck : No
    (Once every six hours)
    C. Run for stuck activities periodically
    1. Item Type :
    2. Min Threshold :
    3. Max Threshold :
    4. Process Deferred : No
    5. Process Timeout : No
    6. Process Stuck : Yes
    (Once every day)
    D. Run forall three activities periodically
    1. Item Type :
    2. Min Threshold :
    3. Max Threshold :
    4. Process Deferred : Yes
    5. Process Timeout : Yes
    6. Process Stuck : Yes
    (Once every day)
    By the way, we also have Synchronize WF Local Tables run once every 24 hours
    Hope this helps.
    Edited by: DBA115102 on Feb 1, 2011 10:41 AM

  • Help with accessing same keychain data from two apps with bundled appid

    I have two ios apps that I'd like to share the same keychain data.  
    According to the apple docs, this should be possible as long as I've set up the appid with a wild card, which I did from the provisioning portal:
    appid:   bundleID.*
    Then, in the AIR for IOS settings of the flash ide, in the App ID field, I made the two apps be members fo the this appid:  bundleID.app1 and bundleID.app2
    But when I use the EncriptedLocalStore class to access the data from one of the apps that the other app stored, it isn't available.
    But each app, independently, is successfully writing to/reading data from the keychain.
    From what I've read, I'm doing everything right.  But clearly I'm not because it isn't working.

    hi dear
    the link which you have send is really too helpful
    but when i am using the way shown there and as  directed it shows an error that
                 'DataSource - Table not found'
    i think this is because the temprary table we r using is neither user defined nor SBO table.
    so plz tell me some way.

Maybe you are looking for