ABAP for Master Data lookup

I am trying ABAP for master data lookup in update rule.
Here is the scenario ---
There is a Master Data object MDABC  with attributes A1 , A2 . I need to Map IO1 to A1 and IO2 to A2.
What should be the Start Routine and Update Routine . Please help with a working code.
Thanks

>
sap_newbee wrote:
> Thanks Aashish ,
> Here is the code I am usind but Its not populating any result May be you could help me out in debugging
>
>
> Start Routine -
>
> DATA: BEGIN OF ITAB_MDABC OCCURS 0,
> MDABC LIKE /BIC/PMDABC-/BIC/MDABC,
>  A1 LIKE /BIC/PMDABC-/BIC/A1,
> A2 LIKE /BIC/PMDABC-/BIC/A2,
> END OF ITAB_NMDABC.
>
>  SELECT
> /BIC/MDABC
> /BIC/A1
> /BIC/A2
>  FROM /BIC/PMDABC INTO TABLE ITAB_MDABC
>  FOR ALL ENTRIES IN DATA_PACKAGE
> WHERE /BIC/MDABC = DATA_PACKAGE-/BIC/MDABC.
> ENDSELECT.
>
>
> In Update Routine for Infoobject IO1 The code Iam using is
>
>
> READ TABLE ITAB_MDABC   WITH KEY
>  MDABC   =  COMM_STRUCTURE-/BIC/MDABC
>  BINARY SEARCH.
> IF sy-subrc = 0.
>  RESULT = ITAB_MDABC-A1.
> ENDIF.
>   RETURNCODE = 0.
>
>   ABORT = 0.
>
> Please help.
> Thanks
Please use table in the select statement. Modifications in BOLD
Edited by: Ashish Gour on Oct 17, 2008 2:57 PM

Similar Messages

  • Master data Lookup missing for sold to party

    Master data Lookup missing for sold to party?
    The look up flow is from E_BOM to E_BOM1
    regards

    Hi,
    Check if SIDs are generated for the infoobject, if not try to do the change run attribute and then load the data to Cube.
    Hope this helps...
    Rgs,
    Ravikanth.

  • 0CALDAY for time dependent master data lookup unknown when migrated to 7.0

    I am in the process of migrating a number of InfoProviders from the 3.x Business Content to the new methodology for BI 2004s.
    When I try to create a transformation from the Update Rules for 0PA_C01, all of the rules that use a master data lookup into 0EMPLOYEE give me the error such as "Rule 41 (target field: 0PERSON group: Standard Group): Time char. 0CALDAY in rule unknown". 
    How do I fix the transformation rule that is generated from the update rule for these time-dependent master data attributes?

    Hi Mark,
    look at http://www.service.sap.com/. I guess you need to implement some corrections or newer supp-packages.
    kind regards
    Siggi
    PS: take a look: https://websmp104.sap-ag.de/~form/handler?_APP=01100107900000000342&_EVENT=REDIR&_NNUM=941525&_NLANG=EN
    Message was edited by:
            Siegfried Szameitat

  • Creation of OLAP Variable for Master data

    Dear BW Experts,
    How to Create OLAP Variable for Master data  in SAP BW? What is the use of OLAP Variable?

    The entry column Type (Variable) on the Select Data tabpage indicates a variable time frame in which the data should be requested. The variables are not replaced by concrete values until a request is made. There are eight options:
    Variable            Data request        Description
    0
                          Yesterday
                                            (12am – 12am)
    1
                         Last week
                                            (Monday-Sunday)
    2
                           Last month
                                            (1. (First day - last day of previous month)
    3
                           Last quarter
                                           (1. (First day - last day of previous quarter)
    4
                           Last year
                                          (01. (January 1 – Dec 31 of previous year)
    5
                      User time selection
                                               See below
    6
                       ABAP routines
                                               See below
    7
                         <b>OLAP variables</b>
                                                See below
    <b>Time selections with the help of the variables 0 to 4.</b>
    The variables 0-4 only refer to date fields of the data type DATS.
    <b>User time selection</b>
    Type 5 gives you a free selection of all fields.
    If you select this variable for a field, confirm the selection, and then choose Detail for Type, an additional dialog box appears. In this dialog box you can freely limit the values of the fields and determine the next period value and the number of periods until a repetition.
    Variables are placeholders for values and are not replaced by concrete values until a data request is made. You can also use a Variable for selections when requesting data for an InfoObject, if you choose type 7, confirm the selection, and then choose Detail for Type.

  • Poor MDX performance on F4 master data lookup

    Hi,
    <P>
    I've posted this to this forum as it didn't get much help in the BW 7.0 forum. I'm thinking it was too MDX oriented to get any help there. Hopefully someone has some ideas.
    <P>
    We have upgraded our BW system to 7.0 EHP1 SP6 from BW 3.5. There is substantial use of SAP BusinessObjects Enterprise XI 3.1 (BOXI) and also significant use of navigational attibutes. Everything works fine in 3.5 and we have worked through a number of performance problems in BW 7.0. We are using BOXI 3.1 SP1 but have tested with SP2 and it generates the same MDX. We do however have all the latest MDX related notes, including the composite note 1142664.
    <P>
    We have a number of "fat" queries that act as universes for BOXI and it is when BOXI sends a MDX statement that includes certain crossjoins with navigational attributes that things fall apart. This is an example of one that runs in about a minute in 3.5:
    <P>
    SELECT { [Measures]. [494GFZKQ2EHOMQEPILFPU9QMV], [Measures].[494GFZSELD3E5CY5OFI24BPCN], [Measures].[494GG07RNAAT6M1203MQOFMS7], [Measures]. [494GG0N4P7I87V3YBRRF8JK7R] } ON COLUMNS , NON EMPTY CROSSJOIN( CROSSJOIN( CROSSJOIN( CROSSJOIN( CROSSJOIN( [0MAT_SALES__ZPRODCAT]. [LEVEL01].MEMBERS, EXCEPT( { [0MAT_SALES__ZASS_GRP]. [LEVEL01].MEMBERS } , { { [0MAT_SALES__ZASS_GRP].[M5], [0MAT_SALES__ZASS_GRP].[M6] } } ) ), EXCEPT( { [0SALES_OFF]. [LEVEL01].MEMBERS } , { { [0SALES_OFF].[#] } } ) ), [0SALES_OFF__ZPLNTAREA].[LEVEL01].MEMBERS ), [0SALES_OFF__ZPLNTREGN]. [LEVEL01].MEMBERS ), [ZMFIFWEEK].[LEVEL01].MEMBERS ) DIMENSION PROPERTIES MEMBER_UNIQUE_NAME, MEMBER_NAME, MEMBER_CAPTION ON ROWS FROM [ZMSD01/ZMSD01_QBO_Q0010]
    <P>
    However in 7.0 there appear to be some master data lookups that are killing performance before we even get to the BW queries. Note that in RSRT terms this is prior to even getting the popup screen withe "display aggregate".
    <P>
    They were taking 700 seconds but now take about 150 seconds after an index was created on the ODS /BIC/AZOSDOR0300. From what I can see, the navigational attributes require BW to ask "what are the valid SIDs for SALES_OFF in this multiprovider". The odd thing is that BW 3.5 does no such query. It just hits the fact tables directly.
    <P>
    SELECT "SID" , "SALES_OFF" FROM ( SELECT "S0000"."SID","P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000"."SALES_OFF" = "S0000"."SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BI0/D0PCA_C021" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BIC/DZBSDBL018" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BIC/DZBSDOR028" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BIC/DZBSDOR038" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BIC/DZBSDOR058" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BIC/DZBSDOR081" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "D"."SID_0SALES_OFF" AS "SID" FROM "/BIC/DZBSDPAY016" "D" ) UNION SELECT "S0000"."SID" ,"P0000"."SALES_OFF" FROM "/BI0/PSALES_OFF" "P0000" JOIN "/BI0/SSALES_OFF" "S0000" ON "P0000" . "SALES_OFF" = "S0000" . "SALES_OFF" WHERE "P0000"."OBJVERS" = 'A' AND "P0000"."SALES_OFF" IN ( SELECT "O"."SALES_OFF" AS "KEY" FROM "/BIC/AZOSDOR0300" "O" ) ) ORDER BY "SALES_OFF" ASC
    <P>
    I had assumed this had something to do with BOXI - but I don't think this is a MDX specific problem, even though it's hard to test in RSRT as it's a query navigation. Also I assumed it might be something to do with the F4 master data lookup but that's not the case, because of course this "fat" query doesn't have a selection screen, just a small initial view and a large number of free characteristics. Still I set the characteristic settings not only to do lookups on the master data values and that made no difference. Nonetheless you can see in the MDXTEST trace that event 6001: F4: Read Data. Curiously this is an extra one that sits between event 40011: MDX Initialization and event 40010: MDX Execution.
    <P>
    I've tuned this query as much as I can from the Oracle perspective and checked the indexes and statistics. Also checked Oracle is perfectly tuned and parameterized as for 10.2.0.4 with the May 2010 patchset for AIX. But this query returns an estimated 56 million rows and runs an expensive UNION join on them - so no suprise that it's slow. As a point of interest changing it from UNION to UNION ALL cuts the time to 30 seconds. Don't think that helps me though, other than confirming that it is the sort which is expensive on 56m records.
    <P>
    Thinking that the UNORDER MDX statement might make a difference, I changed the MDX to the following but that didn't make any difference either.
    <P>
    SELECT { [Measures].[494GFZKQ2EHOMQEPILFPU9QMV], [Measures].[494GFZSELD3E5CY5OFI24BPCN], [Measures].[494GG07RNAAT6M1203MQOFMS7], [Measures].[494GG0N4P7I87V3YBRRF8JK7R] } ON COLUMNS ,
    NON EMPTY UNORDER( CROSSJOIN(
      UNORDER( CROSSJOIN(
        UNORDER( CROSSJOIN(
          UNORDER( CROSSJOIN(
            UNORDER( CROSSJOIN(
              [0MAT_SALES__ZPRODCAT].[LEVEL01].MEMBERS, EXCEPT(
                { [0MAT_SALES__ZASS_GRP].[LEVEL01].MEMBERS } , { { [0MAT_SALES__ZASS_GRP].[M5], [0MAT_SALES__ZASS_GRP].[M6] } }
            ) ), EXCEPT(
              { [0SALES_OFF].[LEVEL01].MEMBERS } , { { [0SALES_OFF].[#] } }
          ) ), [0SALES_OFF__ZPLNTAREA].[LEVEL01].MEMBERS
        ) ), [0SALES_OFF__ZPLNTREGN].[LEVEL01].MEMBERS
      ) ), [ZMFIFWEEK].[LEVEL01].MEMBERS
    DIMENSION PROPERTIES MEMBER_UNIQUE_NAME, MEMBER_NAME, MEMBER_CAPTION ON ROWS FROM [ZMSD01/ZMSD01_QBO_Q0010]
    <P>
    Does anyone know why BW 7.0 behaves differently in this respect and what I can do to resolve the problem? It is very difficult to make any changes to the universe or BEx query because there are thousands of Webi queries written over the top and the regression test would be very expensive.
    <P>
    Regards,
    <P>
    John

    Hi John,
    couple of comments:
    - first of all you posted it in the wrong forum. This belongs into the BW forum
    - MDX enhancements in regards to BusinessObjects are part of BW 7.01 SP05 - not just BW 7.0
    would suggest you post it in the BW forum
    Ingo

  • About delta loading for master data attribute

    Hi all,
    We have a master data attribute loading failed which is a delta loading. I have to fix this problem but I have two questions:
    1. how can I find the those delta requests because I need to delete all these failed requests first, am I right ? Master data is not like cube or ods that we can find the requests in Manage, for master data how can we find them ?
    2. Could you please let me know the detailed procedures to perform this delta loading again ?
    Thanks a lot

    Hi...
    1. how can I find the those delta requests because I need to delete all these failed requests first, am I right ? Master data is not like cube or ods that we can find the requests in Manage, for master data how can we find them ?
    Look.....for master data.....no need to delete request from the target..........just make the status red.......and repeat the load.....But problem is that master data sometimes does'nt support Repeat delta..........if u repeat......then load will again fail with Update mode R.........in that case u hav to do re-init.......
    1) delete the init flag.......(In the IP scheduler >> in the top Scheduler tab >> Initialization option for source system)
    2) Init with data transfer(if failed load picks some records)..........otherwise .....init without data transfer.....if the last delta failed picking 0 records.......
    3) then Delta.......
    2. Could you please let me know the detailed procedures to perform this delta loading again ?
    1) Make the QM status red.........to set back the init pointer.......
    2) Repeat the load.....
    After that.........if again load failed with Update mode R.....
    1) delete the init flag.......
    2) Init with data transfer(if failed load picks some records)..........otherwise init without data transfer.....
    3) then Delta.......
    Regards,
    Debjani.....

  • Unable to generate "Export data source" for Master data infoobject

    Hi Gurus,
    I'm in BW 3.5
    I need to generate "Export data source" for a master data info object.
    I followed below steps:
        I ticked the checkbox "Characterist. is export data source" in the master data Info object maintenance screen.
        In the Infosource area, for the desired Infoobject --> Right click --> Generate export datasource --> a dialog box displays saying "The generation of DataSource  was successful"
    Problem:
    Even after refreshing, etc.. I couldn't find the generated datasource 8**** for the above master data. I also tried to replicate the datasource from the myself source system, but couldn't find it even there. I don't know where the generated export data source is hiding ???
    Could anyone please help ????
    Thanks,
    Venkat

    Hi Venkat,
    I guess that after creating an export datasource for infoobject this DS should appear in the RSA1-Source systems - MYSELF - datasource overview (with name starting with 8). Since it's a datamart, it should appear also as system generated infosource. And in this case you would be able to assign in URs of data target the appropriate infosource (make sure that in Infoproviders tab of RSA1 you have inserted your infoobject as a data target).
    However, there are some bugs related to export data source. CHeck if you see you export data source in RSA6.
    If you don't see it - consider applying OSS Note #816892 - "30BSP26: Export DS for master data is not generated".
    Note #876845 - "30BSP29: InfoObject change: Export DataSource not adjusted" is also may be useful.
    -Vikram

  • Error while creating Data Source for master data attributes

    Hi BI Experts,
    Well its been some time for me that I have been part of Extraction in BI.I primarily handled reporting in my last assignments.
    I was trying extraction with flat files in SAP BI 7(new to sap bi 7 but very much familiar with BW3.5) but failed in the activity during master data attributes and text upload in infoobject (say IOSP_Mat).
    Here is the procedure that I did after creation of characteristic IOSP_Mat.I created a source system for flat file followed by data source for Master data attributes, i selected all the parameters correctly.i.e. csv file format, data seperator as   ,
    and other settings, now when i am trying to look at the proposed data in the next tab using Load example data.its not showing the desired result.The columns that I have maintained in Flat File is  MAT_NUMBER and MAT_NAME (with say 100 data in the file)
    same is the result when I am trying to load the text data too columns maintained are
    (LANGUAGE MAT_NUMBER Short Description)(same 100 data).
    now i used to rsa1old transaction  to upload the file using 3.5 version.i created info source for master data/text/hierarchies for  IOSP_Mat
    now when trying to upload it using info package for master and text data,I observe its(the data) not maintained in the characteristic IOSP_Mat.
    When I monitored ,I figured the data has not been even uploaded to the PSA level.
    Can you BI experts tell me the answer for this.
    Thanks,
    Srijith

    apologies to all of you for late response,
    was busy with some other activities.
    I don't remember the exact message,but I remember it was not loaded to even the PSA level.I will try it again and post the exact message.
    Thanks again for your quick response.
    Once again sorry to all of you for my late response
    Thanks,
    Sri

  • Error handling for master data with direct update

    Hi guys,
    For master data with flexible update, error handling can be defined in InfoPackege, and if the load is performed via PSA there are several options - clear so far. But what about direct update...
    But my specific question is: If an erroneous record (e.g invalid characters) occur in a master data load using direct update, this will set the request to red. But what does this mean in terms of what happens to the other records of the request (which are correct) are they written to the master data tables, so that they can be present once the masterdata is activated, or are nothing written to masterdata tables if a single record is erroneous???
    Many thanks,
    / Christian

    Hi Christian -
    Difference between flexible upload & Direct upload is that direct upload does not have Update Rules, direct upload will have PSA as usual & you can do testing in PSA.
    second part when you load master data - if error occurs all the records for that request no will be status error so activation will not have any impact on it i.e. no new records from failed load will be available.
    hope it helps
    regards
    Vikash

  • Why to use chnage  pointers for  master data idoc why not  the  change idoc

    Hi Gurus,
    I have  one doubt about  Idoc.
    When changes to master  data has to be sent Change pointers are configured and  used. (CDPOS & CDHDR). In case of  the  transaction data change change idoc  is  used as is the case with orders (we  use ORDCHG for  ORDERS message type to send  the  chnage details to a  order which was already sent to other system.
    Why we can't use  change  idoc or  message type whatever, instead of  Chnage pointers in case of  master data or  vice versa in case  of transaction data.
    Your valuable  input  will be rewarded with suitable  points!!
    -B S B.

    Hi,
    It was a good question ... SAP seems never designed the change pointer to handle transactional data IDocs. Only for master data IDocs distribution purposes.
    It may due to volume data ... the transactional data can change many times within a day compare with master data which hardly change or once a while.
    Regards,
    Ferry Lianto

  • Effect of 'No update' for master data texts

    Hello all,
    I am using flexible update - update rules - for master data text. I want to confirm effect of 'No update' in following scenario
    In the fisrt set of update rules -from first source system -I have mapped both short text and long text
    Key  |  short text| long text
    ......| overwrite| overwrite
    K1    | some SHTX| some LGTX
    In the second set of update rules, I have selected 'No update' for Long text
    Key  |  short text | long text
    ......|overwrite| NO UPDATE
    K1    | some SHTX |
    My loading sequence is first source system followed by second source system
    what should be my ultimate long text and why?
    Many thanks in advance!
    Regards
    Sanjyot

    Hi surekha,
    your long text will not get changed.
    Take the example you mentioned
    Key | short text.....| long text
    ........| overwrite......| overwrite
    K1 | some SHTX.. | some LGTX
    In the second set of update rules, I have selected 'No update' for Long text
    Key | short text... | long text
    ........|overwrite......| NO UPDATE
    K1 | some SHTX. |
    In the second set of update rules, you mentioned "No UPdate", So, when a record with the same key is found n Masterdata, it is not going to make any changes to the field Long text.
    Where as short text will get overwritten.
    Take some data for ex with the above mentioned type
    data from source 1 : 2000 Maggi Noodles
    data from source 2 : 2000 Feasters
    when the data from first source system arrives, data will be 2000 maggi noodles
    when the data from second source suystem arrives, as there exist this records with same key(2000), it checks for the other fiels
    short text will be over written with the value in second system (according to rules its in overwrite mode)
    Long text will not be updated as the rule says no update.
    hope this helps,
    Cheers,
    Srinath.
    Cheers,
    Srinath.

  • Why we are not using "Message Control" for Master data ?

    Hi friends,
    We are using Message control as output determination for Transactional data, <b>what is the main reason for not using the msg control for Master data.</b>Please explain with details...
    I am very thankful to them...
    Thanks and Regards.
    <i><b>Naveen Kumar. Chamala</b></i>

    Hi Naveen,
       Listing down those I know -
       1. Master data is expected to be changed/created less frequestly.
       2. Transactional data is expected to be chnaged/created more frequestly.
       3. Transactions to create/change/display master data do no have message control integration, since you are not supposed to print/fax/mail master data to your partners.
       4. Distribution of master data in generally supposed to happen within Enterprise boundry and typically you will have one central server maintaining master data which will be distributed to multiple downstream operation system (Some thing where MDM fits). So ideally when distributing master data using ALE you will be sending it to multiple systems.
       5. Distributing transactional data PO/SO etc is same as sending fax/print which fits with the requirement of message control.
    Hope this helps,
    Regards,
    Sanjeev

  • What type of questions can i expect for master data?

    hi,
    am preparing for the interview. pls... let me know those questions for master data.
    thanks to all

    HI arun,
    Pl. be clear on
    what is master data, purpose, customer master data fields and material master data imp fields.
    common distribution channels and divisions.
    Org. master data. ( interms of one to many , many to many relations of org elements ).
    condition master data
    How master data is uploaded and maintained.
    Hope it helps. Pl. reward if useful.
    Thanks
    Sadhu Kishore

  • Aggregates for master data & Line item Dimension

    Hi i have question
    1     Can we create Aggregate for Master Data?
    2     Can we create Aggregate with Navigational Attributes?
    3     Can we create Aggregate with Line Item Dimension?
    4     How many dimension we can add with how many characteristics in it if we create a new Aggregate

    1 Can we create Aggregate for Master Data? - <b>No</b>
    2 Can we create Aggregate with Navigational Attributes? - <b>No</b>
    3 Can we create Aggregate with Line Item Dimension? - <b>No</b>
    4 How many dimension we can add with how many characteristics in it if we create a new Aggregate - <b>You can create a maximum of 13 dimensions in the cube.You can add as many characteristics for an aggregate,adding many characteristics in the aggregate is as good as having the query fetch data from the Cube instead of the aggregates.</b>
    Aggregates are subsets infocubes derived from the main InfoCube to increase the query performance.You can create as many aggregates you want in an InfoCube.
    For each aggregate,internally dimensions are created based on the selected characteristics.
    Take a look at this link for detail information...
    http://help.sap.com/saphelp_nw04/helpdata/en/7d/eb683cc5e8ca68e10000000a114084/content.htm

  • Some segments are missing in the idocs for master data zdebmas

    hi guru's,
    can any one hlep me here we facing the probelumm
    some segments are missing in the idocs for master data zdebmas
    , there is some issue on the generation of the Site Master IDoc (Message type: ZDEBMAS, Basic type: DEBMAS06).
    This is using the SAP standard program (RBDMIDOC) which reads the Site master change pointers.
    There is  some segments below is missing in the IDoc:
    how to chcek this probelumm...

    hi
    i got the function module. it is  triggerig whne i do changepointer running.
    what ever changes i made only that segments are onlycomming in to the idoc. but remaing segments are not comming.
    my req is to show all segments  even if i do changes in one segmet fields  dont change theay have send to the interfece all athe segments.i ahve to do some enhancemetns for that
    can u plse help me the login  or any function module which will fill the alla the segmetns .

Maybe you are looking for