IDoc to file. Grouping and generation of files based on condition

Hi All,
I am working on a scenario, Idoc to File. I need to create number of files based on  occurrence of a field(vendor) in the idoc. Also the condition is to group by the field(vendor) and create as many files as the number of groups formed.
For e.g.: In an Idoc if there are ten items present and 3 vendors who are supplying it. <b>The vendor and items can repeat in any order within the same idoc. </b>
I need to group the items by the vendors and create a file separately for each vendor.  Like :   
Item         Vendor
Item1       V1
Item2         V1
Item3         V2
Item4       V2
Item5        V2
Item6         V2
Item1      V2
Item2         V2
Item3        V3
Item6        V3
What we have to achieve is, to group by the vendors irrespective of the order of the items or number of reoccurrence of the items.  As you see, the Item1 is supplied by both V1 and V2. I need to generate 3 files as follows:
V1.xml with Item1 and Item2 details.
V2.xml with Item3, Item4, Item5, Item6, Item1 Item2 details.
V3.xml with Item3 and Item6 details.
Can someone please help me with achieving this without using BPM?

Hi Shashi,
For this scenario you have to use multimaping without BPM using this blog.
/people/jin.shin/blog/2006/02/07/multi-mapping-without-bpm--yes-it146s-possible
And while doing your message mapping you have to use <b>removeContext</b> & <b>SplitByValue</b> (double clike the on splitbyvalue and select <b>value changed</b>) function. This function you have to use with the node in which your Vendor is coming. Example:
if required to sort the vendor, then first sort the vendor in the below mapping.
Vendor (NODE) -
> removeContext -
> SplitByValue (value Changed) -
> IDOC.
By doing this your vendor will be grouped and you will get new IDOC on change of  Vendor.
Regards,
Sarvesh

Similar Messages

  • IDOC TO FILE based on condition

    Hi,
        I am having IDOC to Ftp scenario, where i need to send the IDOC on condition.
    1. In IDOC there is a field Location, based on this field the target files has to be created.
    ex:-
    1. for Location India ---> file has to be created in INDIAN folder
    2. for Location USA ---> file has to be created in USA folder
    3. 2. for Location Uk ---> file has to be created in UK folder etc.......
    Like this there are nearly 20 folders where i need to create at the target side, based on the location field the fied has to be created at the respective folder automatically.
    When i trigger the IDOC with the Location INDIA, automatically the file has to get created in INDIAN FOLDER.
    Please help me...
    cheers
    nissi

    Hi Roy,
               Thanks for your code, its very useful and suits to my requirement. But i am having folders already in my target file system.
    Based upon the location ID . directly the the file should be placed in the respective folder.
    please find the below code and suggest me if any changes to be done.
    public String dynamicDirectoryName(String Location, Container container) throws StreamTransformationException
             String Directory;
             String  inputFileName;
            DynamicConfiguration conf = (DynamicConfiguration) container
                                            .getTransformationParameters().get(StreamTransformationConstants.DYNAMIC_CONFIGURATION);
            //DynamicConfigurationKey key1 =  DynamicConfigurationKey.create("http://sap.com/xi/XI/System/File","Directory");
            DynamicConfigurationKey key2 =  DynamicConfigurationKey.create("http://sap.com/xi/XI/System/File","FileName");
            //Directory = conf.put(key1,location);
            inputFileName = con.put(key2,location);
            //Directory = conf.get(key1);
            inputFileName = conf.get(key2);
            If(inputFileName.equals("INDIA")
                con.put(key2,"INDIA");
            else if(inputFileName.equals("AUSTRALIA"))
                con.put(key2,"AUSTRALIA");
            else
               return "";

  • How to create additional Line in file based on condition available as part of ZINVOIC02 Idoc segment

    Scenario Details:
    Receiving Zinvoic02 Idoc in PI. Idoc to file translation creates comma separated file with .csv extn
    The logic was kept in such a way that how many E1EDP01 (items) are available in IDoc that many no of records will be created in csv file.
    The file logic for some the fields is as below:
    No of records
    InvNumber
    InvDate
    CusNumber
    LineitemDesc
    Tax1Type
    Tax1%
    for 1st E1EDP01
    E1EDK01-BELNR
    E1EDK03-DATUM
    E1EDK01-PARTN
    Populate when E1EDP04/MSKWZ=O2 or O4 with E1EDP19/KTEXT
    Hardcode when E1EDP04/MSKWZ=O2 or O4
    Sum all E1EDP04 /MSATZ when E1EDP04/MSKWZ=O2 or O4
    for 2nd E1EDP01
    E1EDK01-BELNR
    E1EDK01-DATUM
    E1EDK01-PARTN
    same as above
    same as above
    same as above
    for 3rd E1EDP01
    E1EDK01-BELNR
    E1EDK03-DATUM
    E1EDK01-PARTN
    same as above
    same as above
    same as above
    Additional Line to be created when one or more of E1EDP01 is having E1EDP04/MSKWZ = O3
    same as above
    same as above
    same as above
    Hardcode "REIM for USE TAX"
    Hardcode ""
    Hardcode ""
    Now we have got addition requirement to add a new lineitem when tax code is equal to O3 for any of the E1EDP01.
    Is it possible to create additional lineitem based on condition. If yes, please share what should be the approach.
    How we can create the additional lineitem?
    Currently we are using E1EDP01 to do context handling.
    The target structure is :
    MT_FILE
         INVOICE     0..unbounded
              InvNumber     0..1
              InvDate          0..1
              CusNumber     0..1
              LineitemDesc     0..1
              Tax1Type          0..1
              Tax1%               0..1

    Hello,
    Please add one extra field in the data structure of the target mapping and let its occurrence be 0..unbounded under the root node 'MT_ADP_Invoice'.
    Apply the condition, if tax code MSKWZ (with its context changed to E1EDP01) equalsS to '03', then map it to the newly created target field whose occurrence is 0..unbounded.
    This will then create an additional field which is your requirement.
    The above is one way.
    But if you want to have the same target field name as ADP_File appended for tax field being '03'.
    Then in that case you can you two message mapping for one common operational mapping / interface.
    In the first message mapping you need to have one target data structure created with the source data structure remaining the same as the one shown by you in scrn shot. Now this target data structure will be similar to the source, except that you need to add one more field at the end(name different from other fields) in target (whose occurrence should be 0..unbounded), and it needs to be mapped to E1EDP01 provided the tax code field MSKWZ(its context changed to E1EDP01) equalsS to constant '03'.
    In the second message mapping you need to map the target structure of previous message mapping to the actual required structure. The newly added field should be mapped to ADP_File of your final target structure.
    This will then create the same structure as required.
    Note : Please change the occurence of ADP_File to 0..unbounded.
    Regards,
    Souvik

  • IDOC-File - records in Idoc to file based on some condition

    Hello experts,
    I have a idoc to file scenario. the incoming Idoc can have multiple records in it and i have to map these records to a csv file. Now the problem is not all records has to be mapped to the file. Based on the value of a perticular file (which is not root), i need to decide if the entire record has to be processed or not. Like the idoc structure is say:
    IDOC
       E1WPA01            0....9999
          E1WPA04         0...99
              KONDART     0..1
    Now for all valid E1WPA01 where the value of KONDART equals to some spacified value (known) , there has to be one record in the target csv file.
    How can this be done at the root level?
    One way of doing is we give empty values in the target file where the condition of KONDART is not fulfilled. Say if there are 10 records of E1WPA01 and only 4 satisfies the condition, we will ahve 10 records in the output file but only 4 records will have value and the rest 6 will be blank like (,,,,,,,). But I dont want this. I want only 4 records in the output file.
    I tried mapping like this:
    E1WPA01----
    >Advanced -
    > Root
    E1WPA04-KONDART---->UDF
    The problem i am facing here is if 4 records fulfill the condition, the first 4 are taken and the revelent 4.
    Please help.
    Regards,
    Yash

    Hi Chirag,
    I wrote the following code in UDF:
    for(i=0;i<a.length;i++)
    if(a{i}.equals("Specified Values"))
    result.addValue("a{i}");
    else
    result.addValue(ResuktList.SUPPRESS);
    And its working partially. I mean the queue of the UDF looks like
    1. AAAA                      SUPPRESS
    2. <Specified value>     <Specified value>     
    3. AAA                        SUPPRESS
    4. AAA                        SUPPRESS
    5. AAA                        SUPPRESS
    6. AAA                        SUPPRESS
    7. <Specified value>     <Specified value>
    8. AAA                        SUPPRESS
    9. AAA                        SUPPRESS
    and it creates 2 ROOT nodes. But the root nodes are created for line 2 and line 8 insteed of 7.
    What can be problem??
    Yash

  • Grouping and sorting the data based on the parameter

    Hi,
    I need to display the total of item Cost i.e sub total based on parameter in Excel format
    if value = 1 then (Subtotal by Manufacture) and sorting by Manufacture, Receipt Date, Organization, Item Code
    if value = 2 then (Subtotal by Organization ) and sorting by Organization, Item, Serial Number
    and at the end of report need to display full total
    is it possible in excel output using the below xml , please guide me if i need to change the way of getting XML ouput or can be acheived using this.
    Please find the xml , thanks in advance
    <?xml version="1.0"?>
    <!-- Generated by Oracle Reports version 6.0.8.27.0 -->
    <ASWRECDTREP>
    <LIST_G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>10</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PARTS</ORGANISATION_NAME>
    <SERIAL_NUMBER>90699100802262</SERIAL_NUMBER>
    <ITEM_CODE>OM906LA.007</ITEM_CODE>
    <DESCRIPTION>MERCEDES SERIES 900 OFF HWY ENGINE 205 KW 0852</DESCRIPTION>
    <MANUFACTURER>MER</MANUFACTURER>
    <APPLICATION>INDL</APPLICATION>
    <ITEM_COST>15009.09</ITEM_COST>
    <MATERIAL_COST>685.25</MATERIAL_COST>
    <PO_NUMBER>1001395</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>22-MAY-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>EUR</CURRENCY_CODE>
    <PO_UNIT_PRICE>7431.3</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>10</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PARTS</ORGANISATION_NAME>
    <SERIAL_NUMBER>90699100802285</SERIAL_NUMBER>
    <ITEM_CODE>OM906LA.007</ITEM_CODE>
    <DESCRIPTION>MERCEDES SERIES 900 OFF HWY ENGINE 205 KW 0852</DESCRIPTION>
    <MANUFACTURER>MER</MANUFACTURER>
    <APPLICATION>INDL</APPLICATION>
    <ITEM_COST>15009.09</ITEM_COST>
    <MATERIAL_COST>685.25</MATERIAL_COST>
    <PO_NUMBER>1001395</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>22-MAY-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>EUR</CURRENCY_CODE>
    <PO_UNIT_PRICE>7431.3</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>30</ORGANIZATION_CODE>
    <ORGANISATION_NAME>Melbourne</ORGANISATION_NAME>
    <SERIAL_NUMBER>6520107896</SERIAL_NUMBER>
    <ITEM_CODE>3500.010</ITEM_CODE>
    <DESCRIPTION>ALLISON 3000 SERIES ON HWY TRANSMISSION E018179</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>11126.11</ITEM_COST>
    <MATERIAL_COST>664.23</MATERIAL_COST>
    <PO_NUMBER>971515</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>14-APR-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>8063.3</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>30</ORGANIZATION_CODE>
    <ORGANISATION_NAME>Melbourne</ORGANISATION_NAME>
    <SERIAL_NUMBER>6510869062</SERIAL_NUMBER>
    <ITEM_CODE>3200.010</ITEM_CODE>
    <DESCRIPTION>ALLISON 3000 SERIES ON HWY TRANSMISSION E017944</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>11853.57</ITEM_COST>
    <MATERIAL_COST>707.66</MATERIAL_COST>
    <PO_NUMBER>982120</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>12-MAY-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>8251.6</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>40</ORGANIZATION_CODE>
    <ORGANISATION_NAME>Brisbane</ORGANISATION_NAME>
    <SERIAL_NUMBER>D6002 G6003</SERIAL_NUMBER>
    <ITEM_CODE>520 SERIES.002</ITEM_CODE>
    <DESCRIPTION>KONRAD 520 SERIES MARINE STERN DRIVE</DESCRIPTION>
    <MANUFACTURER>OTH</MANUFACTURER>
    <APPLICATION>MARI</APPLICATION>
    <ITEM_COST>9296.95</ITEM_COST>
    <MATERIAL_COST>482.91</MATERIAL_COST>
    <PO_NUMBER>1009062</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>24-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>6939.29</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>40</ORGANIZATION_CODE>
    <ORGANISATION_NAME>Brisbane</ORGANISATION_NAME>
    <SERIAL_NUMBER>ABC123</SERIAL_NUMBER>
    <ITEM_CODE>GM20541-KP1.001</ITEM_CODE>
    <DESCRIPTION>KOHLER GENSET</DESCRIPTION>
    <MANUFACTURER>KOH</MANUFACTURER>
    <APPLICATION>PGEN</APPLICATION>
    <ITEM_COST>756.83</ITEM_COST>
    <MATERIAL_COST>.78</MATERIAL_COST>
    <PO_NUMBER>1015156</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>14-OCT-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>AUD</CURRENCY_CODE>
    <PO_UNIT_PRICE>20</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>40</ORGANIZATION_CODE>
    <ORGANISATION_NAME>Brisbane</ORGANISATION_NAME>
    <SERIAL_NUMBER>ABC124</SERIAL_NUMBER>
    <ITEM_CODE>GM20541-KP1.001</ITEM_CODE>
    <DESCRIPTION>KOHLER GENSET</DESCRIPTION>
    <MANUFACTURER>KOH</MANUFACTURER>
    <APPLICATION>PGEN</APPLICATION>
    <ITEM_COST>756.83</ITEM_COST>
    <MATERIAL_COST>.78</MATERIAL_COST>
    <PO_NUMBER>1015156</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>14-OCT-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>AUD</CURRENCY_CODE>
    <PO_UNIT_PRICE>20</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>06R1017763</SERIAL_NUMBER>
    <ITEM_CODE>6062HK30.005</ITEM_CODE>
    <DESCRIPTION>DETROIT SERIES 60 MARINE ENGINE 615 KW 2949322</DESCRIPTION>
    <MANUFACTURER>MDD</MANUFACTURER>
    <APPLICATION>MARI</APPLICATION>
    <ITEM_COST>99125.34</ITEM_COST>
    <MATERIAL_COST>2028.31</MATERIAL_COST>
    <PO_NUMBER>883339</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>05-MAR-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>63768</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310935001</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1000221</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>01-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310935002</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1000221</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>01-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310935003</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1000221</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>01-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6510876301</SERIAL_NUMBER>
    <ITEM_CODE>T350R.004</ITEM_CODE>
    <DESCRIPTION>ALLISON TORQMATIC ON HWY TRANSMISSION</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>9465.69</ITEM_COST>
    <MATERIAL_COST>597.33</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>6936.3</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6510876302</SERIAL_NUMBER>
    <ITEM_CODE>T350R.004</ITEM_CODE>
    <DESCRIPTION>ALLISON TORQMATIC ON HWY TRANSMISSION</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>9465.69</ITEM_COST>
    <MATERIAL_COST>597.33</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>6936.3</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6510876303</SERIAL_NUMBER>
    <ITEM_CODE>T375R.002</ITEM_CODE>
    <DESCRIPTION>ALLISON TORQMATIC ON HWY TRANSMISSION</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>10224.19</ITEM_COST>
    <MATERIAL_COST>645.19</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>7433.3</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310940354</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310940355</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310940356</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>6310940357</SERIAL_NUMBER>
    <ITEM_CODE>2500.003</ITEM_CODE>
    <DESCRIPTION>ALLISON 2000 SERIES ON HWY TRANSMISSION E016189</DESCRIPTION>
    <MANUFACTURER>ATD</MANUFACTURER>
    <APPLICATION>ONHY</APPLICATION>
    <ITEM_COST>4855.33</ITEM_COST>
    <MATERIAL_COST>306.39</MATERIAL_COST>
    <PO_NUMBER>1005777</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>23-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>3574.9</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    <G_ASW_REC_DATE>
    <ORGANIZATION_CODE>70</ORGANIZATION_CODE>
    <ORGANISATION_NAME>NDC - PRODUCT</ORGANISATION_NAME>
    <SERIAL_NUMBER>5272003543</SERIAL_NUMBER>
    <ITEM_CODE>T1637K33.002</ITEM_CODE>
    <DESCRIPTION>DETROIT SERIES 4000 INDUSTRIAL ENGINE 1865 KW 2456291</DESCRIPTION>
    <MANUFACTURER>MDD</MANUFACTURER>
    <APPLICATION>MNIG</APPLICATION>
    <ITEM_COST>420083.16</ITEM_COST>
    <MATERIAL_COST>3785.85</MATERIAL_COST>
    <PO_NUMBER>921170</PO_NUMBER>
    <TRANSACTION_RECEIPT_DATE>30-JUN-09</TRANSACTION_RECEIPT_DATE>
    <VENDOR_LOT_NUM></VENDOR_LOT_NUM>
    <CURRENCY_CODE>USD</CURRENCY_CODE>
    <PO_UNIT_PRICE>301549</PO_UNIT_PRICE>
    </G_ASW_REC_DATE>
    </LIST_G_ASW_REC_DATE>
    </ASWRECDTREP>
    Best Regards,
    Mahi

    Hi Vetri,
    I tried to implement the solution you have given and i am getting error,
    I tried to see the output by loading the xml to the template given in BIP Blog in the following Link,
    http://blogs.oracle.com/xmlpublisher/2007/05/left_up_down_right_group.html
    I am getting the same error,Please help me how to overcome this.If possible send me the template that was working for you.
    The error log is as follows:
    ConfFile: C:\Program Files\Oracle\BI Publisher\BI Publisher Desktop\Template Builder for Word\config\xdoconfig.xml
    Font Dir: C:\Program Files\Oracle\BI Publisher\BI Publisher Desktop\Template Builder for Word\fonts
    Run XDO Start
    Template: C:\Users\MAHESH\Desktop\Grouping.rtf
    RTFProcessor setLocale: en-us
    FOProcessor setData: C:\Users\MAHESH\Desktop\EmployeeListing.xml
    FOProcessor setLocale: en-us
    java.lang.reflect.InvocationTargetException
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at oracle.apps.xdo.common.xml.XSLT10gR1.invokeProcessXSL(Unknown Source)
         at oracle.apps.xdo.common.xml.XSLT10gR1.transform(Unknown Source)
         at oracle.apps.xdo.common.xml.XSLT10gR1.transform(Unknown Source)
         at oracle.apps.xdo.common.xml.XSLTWrapper.transform(Unknown Source)
         at oracle.apps.xdo.template.fo.util.FOUtility.generateFO(Unknown Source)
         at oracle.apps.xdo.template.fo.util.FOUtility.generateFO(Unknown Source)
         at oracle.apps.xdo.template.FOProcessor.createFO(Unknown Source)
         at oracle.apps.xdo.template.FOProcessor.generate(Unknown Source)
         at RTF2PDF.runRTFto(RTF2PDF.java:629)
         at RTF2PDF.runXDO(RTF2PDF.java:439)
         at RTF2PDF.main(RTF2PDF.java:289)
    Caused by: java.lang.ArrayIndexOutOfBoundsException: -1
         at oracle.xdo.parser.v2.XSLTContext.peekExprValue4Grouping(XSLTContext.java:871)
         at oracle.xdo.parser.v2.XPathFunctionCall.evaluate(XPathFunctionCall.java:536)
         at oracle.xdo.parser.v2.XPathFunctionCall.evaluate(XPathFunctionCall.java:583)
         at oracle.xdo.parser.v2.XSLVariable.getValue(XSLVariable.java:205)
         at oracle.xdo.parser.v2.XSLVariable.processAction(XSLVariable.java:117)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLResultElement.processAction(XSLResultElement.java:180)
         at oracle.xdo.parser.v2.XSLNode.processChildren(XSLNode.java:417)
         at oracle.xdo.parser.v2.XSLTemplate.processAction(XSLTemplate.java:191)
         at oracle.xdo.parser.v2.XSLStylesheet.execute(XSLStylesheet.java:512)
         at oracle.xdo.parser.v2.XSLStylesheet.execute(XSLStylesheet.java:489)
         at oracle.xdo.parser.v2.XSLProcessor.processXSL(XSLProcessor.java:271)
         at oracle.xdo.parser.v2.XSLProcessor.processXSL(XSLProcessor.java:155)
         at oracle.xdo.parser.v2.XSLProcessor.processXSL(XSLProcessor.java:192)
         ... 15 more
    Best Regards,
    Mahi

  • Automatically merging PDF files based on conditions

    Hello
    I need to be able to setup a job with some basic parameters to automatically merge multiple PDF’s.
    For the following group of files:
    1_a.pdf
    1_b.pdf
    2_a.pdf
    2_b.pdf
    3_a.pdf
    3_b.pdf
    I need to be able to append the pages in the "_b" files to the start of the "_a" files using the number at the start as the match identifier.
    I think the safest method would be to output a new "_c" file and maintain the original files.
    I can see functionality to manually do this or merge a group of files into 1 but I need to automate this to deal with larger volumes.
    Do I need additional software?  Any suggestions on how to do this?
    Appreciate any help.
    Cheers

    You're looking for a PDF Parser or PDF miner tool (PDFminer) as a starting framework, and you'll almost certainly be writing custom code around that as parsing a text file that's effectively free-form and originating from multiple different sources almost always (always?) involves writing customized processing code and an on-going series of tweaks as the suppliers of the PDF change their ticket formats.  (Even apparently-simple details such as the time and date formats, for instance, can vary by geography and language and by supplier, and can derail common processing.)
    In some cases that I can envision, it'd be entirely possible that the data you're after is actually located in an embedded image and not in text that can be parsed.
    The best approach is to get folks to send you JSON or XML or some other format intended for interchange, and avoid the whole mess that is parsing or mining a printer-oriented format.
    The other obvious option is to use something like Amazon's Mechanical Turk or some other explicitly outsourced help.  Depending on how often the formats change and how many of these PDF files you're dealing with and how varied the formats are, sometimes throwing staff at the problem can be the most cost-effective approach.

  • Process files based on condition

    Hi,
    Iam looking for some possible suggestions in desiging an interface for the below requirement.
    Let's assume there are two source files,  'FileA' and 'FileB' in an FTP Server.
    Now, we would need to Process 'FileB' in PI (no IR objects are needed, we need to zip 'FileB' and post it to some other location), only if certain condition gets satisfied in 'FileA'. If not ,we need to send an email saying that the condition in 'FileA' didnot get satisfied on that day.
    Is it possible to design this interface with out ccBPM.
    Let me know your thoughts.
    Thanks
    Pradeep

    the best and the correct way is to use BPM
    If you dont use BPM,
    1. you can code a module for your sender File adapter, do the validation/check and then in the module put in a snippet of code that will access the other file, zip the file and transfer it to a different location
    2. use a sender java proxy and do the required logic there.
    Both the solutions are not clean. So i recommend BPM

  • Restrict records in CSV file based on condition

    Greetings. I searched the forum for my issue and didn't find an answer or the right search terms. I am running APEX 4.1.1
    I have a report on a page. I am showing all records for a given SQL statement in my table, but in the "Export to CSV" I need to limit the output based on one of the fields. So for example, if I had a report like:
    SELECT x, y, Z
    FROM Table1
    I would like the Export to CSV for that table to limit the output such as:
    SELECT x, y, Z
    FROM Table1
    WHERE Z = 1
    Is that possible?
    Thanks!
    John

    Hi John,
    I have tested the login now and the above works for Interactive report only and for Classic report you need to do something like this because the request is set to different value
    SELECT x, y, Z
    FROM Table1
    WHERE ((Z = 1 and nvl(:REQUEST,'X') like 'FLOW_EXCEL_OUTPUT%') OR nvl(:REQUEST,'X') not like 'FLOW_EXCEL_OUTPUT%')See this working example http://apex.oracle.com/pls/apex/f?p=46417:25
    When you export to csv you will only get dept 10 records

  • Spliting files based on condition using multi mapping with BPM

    Hi All,
    Can any one please let me know How to <b>Splite the outbound records based on condition using multi mapping with Integration Process in BPM</b>?
    Thanks
    Govindu.

    Hi All,
    Sorry for mistake this question for Exchange infrastructure guys.
    Thanks,
    Govindu

  • SQL query and Delete the data based on condition

    Hi All,
    I have one requirement like I need to fetch the data from the table, In that table one filed (XXXXX) domain type is STAMP.
    delete the data if the sy-datum ( length 8) is greater than field XXXXX  which has length 14.
    Please suggest me how to delete or fetch the data with the same length.
    I have used below code..but did not work..
    "Concatenate sy-datum '%'  into gv_var1.
    select *
       from DD03l
       into itab
    where XXXX like gv_var1."
    and tried this one as well
    1. DELETE itab WHERE XXXXX GT sy-datum.
    2. DELETE itab WHERE XXXXX GT gv_var1.

    Hi,
    When dates are in different formats you have to always convert them into one common format and then do the checks. Its better to convert the date in the variable into timestamp(using CONVERT DATE statement-look for abapdocu for its usage) and then compare it with the date on database.
    data: gv_var1  TYPE timestamp.
    convert date sy-datum into time stamp gv_var1.
    select *
    from DD03l
    into itab
    where XXXX like gv_var1.
    should work now.
    Hope this helps.
    Shwetha

  • How to disable and enable Check box based upon condition

    Hi,
    I wants to disable/hide the check box field before the required field is filled.   Kindly help me

    Hi Mohammed,
    Try to use customize InfoPath form, add rules. Go to list tab, click customize Form, then this opens the list form in InfoPath
    format.
    1.Select the check box field control, click add rules, select if is blank, show validation error action.
    2.Then go to conditions, click column name is blank, change it to the required column you want.
    3.Go to rule type, click validation, change it to formatting, select hide this control.
    4.Publish the InfoPath form.
    Best Regards.
    Kelly Chen
    TechNet Community Support

  • Sql grouping and summing impossible?

    I want to create an sql query to sum up some data but i'm starting to think it is impossible with sql alone. The data i have are of the following form :
    TRAN_DT     TRAN_RS     DEBT     CRED
    10-Jan     701     100     0
    20-Jan     701     150     0
    21-Jan     701     250     0
    22-Jan     705     0     500
    23-Jan     571     100     0
    24-Jan     571     50     0
    25-Jan     701     50     0
    26-Jan     701     20     0
    27-Jan     705     0     300The data are ordered by TRAN_DT and then by TRAN_RS. Tha grouping and summing of data based on tran_rs but only when it changes. So in the table above i do not want to see all 3 first recods but only one with value DEBT the sum of those 3 i.e. 100+150+250=500. So the above table after grouping would be like the one below:
    TRAN_DT     TRAN_RS     DEBT     CRED
    21-Jan     701     500     0
    22-Jan     705     0     500
    24-Jan     571     150     0
    26-Jan     701     70     0
    27-Jan     705     0     300The TRAN_DT is the last value of the summed records. I undestand that the tran_dt may not be selectable. What i have tried so far is the following query:
    select tran_dt,
             tran_rs,
             sum(debt)over(partition by tran_rs order by tran_dt rows unbounded preceding),
             sum(cred)over(partition by tran_rs order by tran_dt rows unbounded preceding) from that_tableIs this even possible with sql alone, any thoughts?
    The report i am trying to create in BI Publisher.Maybe it is possible to group the data in the template and ask my question there?

    915218 wrote:
    Is this even possible with sql alone, any thoughts?It sure is...
    WITH that_table as (select to_date('10/01/2012', 'dd/mm/yyyy') tran_dt, 701 tran_rs, 100 debt, 0 cred from dual union all
                         select to_date('20/01/2012', 'dd/mm/yyyy') tran_dt, 701 tran_rs, 150 debt, 0 cred from dual union all
                         select to_date('21/01/2012', 'dd/mm/yyyy') tran_dt, 701 tran_rs, 250 debt, 0 cred from dual union all
                         select to_date('22/01/2012', 'dd/mm/yyyy') tran_dt, 705 tran_rs, 0 debt, 500 cred from dual union all
                         select to_date('23/01/2012', 'dd/mm/yyyy') tran_dt, 571 tran_rs, 100 debt, 0 cred from dual union all
                         select to_date('24/01/2012', 'dd/mm/yyyy') tran_dt, 571 tran_rs, 50 debt, 0 cred from dual union all
                         select to_date('25/01/2012', 'dd/mm/yyyy') tran_dt, 701 tran_rs, 50 debt, 0 cred from dual union all
                         select to_date('26/01/2012', 'dd/mm/yyyy') tran_dt, 701 tran_rs, 20 debt, 0 cred from dual union all
                         select to_date('27/01/2012', 'dd/mm/yyyy') tran_dt, 705 tran_rs, 0 debt, 300 cred from dual)
    , brk AS (
    SELECT     tran_dt,
            tran_rs,
         debt,
         cred,
            CASE WHEN Nvl (Lag (tran_rs) OVER (ORDER BY tran_dt, tran_rs), 0) != tran_rs THEN tran_rs || tran_dt END brk_tran_rs
      FROM that_table
    ), grp AS (
    SELECT     tran_dt,
            tran_rs,
             debt,
             cred,
            Last_Value (brk_tran_rs IGNORE NULLS) OVER  (ORDER BY tran_dt, tran_rs) grp_tran_rs
      FROM brk
    SELECT     Max (tran_dt),
            Max (tran_rs),
             Sum (debt),
             Sum (cred)
      FROM grp
    GROUP BY grp_tran_rs
    ORDER BY 1, 2
    Boneist
    MAX(TRAN_    TRAN_RS       DEBT       CRED
    21-JAN-12        701        500          0
    22-JAN-12        705          0        500
    24-JAN-12        571        150          0
    26-JAN-12        701         70          0
    27-JAN-12        705          0        300
    Me
    MAX(TRAN_ MAX(TRAN_RS)  SUM(DEBT)  SUM(CRED)
    21-JAN-12          701        500          0
    22-JAN-12          705          0        500
    24-JAN-12          571        150          0
    26-JAN-12          701         70          0
    27-JAN-12          705          0        300Edited by: BrendanP on 17-Feb-2012 04:05
    Test data courtesy of Boneist, and fixed bug.
    Edited by: BrendanP on 17-Feb-2012 04:29

  • PI 7.1 scenario -- File to File , File to IDOC  and IDoc to File Scenario

    Hi All,
    *I have PI 7.0 . I have in that File to File , File to IDOC  and IDoc to File Scenarios*
    File to File -- This scenario is working fine.
    File to IDOC   --  Setting done in PI 7.0 like SM59,  IDXI, IDX2  and in R/3 like SALE (logical system)---  This scenario is working fine.
    IDoc to File  --  Seeting done in R/3 like SALE (Logical system, Distrubtion Model (Port, Partner Profile)) -- This scenario is working fine.
    Now I need to move to PI 7.1  -- for File to File , File to IDOC  and IDoc to File Scenarios.
    Please let me know do I require to make any changes in Design and Configuration and settings in PI 7.1 and R/3 for above scenarios or it will work fine as it is already working for PI 7.0
    If changes are required please let me know what changes to do for each scaenario and pls also send blogs for that.
    Regards

    >
    Rickkk wrote:
    > Hi,
    >
    > Idoc metadeta is updated through idx2?
    >
    > Regards
    If there is no metadata maintained for an idoc type, the metadata for that idoc type gets loaded when an idoc of that type is received by PI for the first time. If you are looking at loading metadata of all idoc types before the first run, its really not required. Just make sure that the port is configured properly. Although, there is no harm in preloading the metadata using IDX2.
    You need to use IDX2 mandatorily when the the metadata changes and the metadata in ECC and PI is different. In that case, you need to delete metadata for the changed IDoc type and then either you can update the metada through IDX2 or let the system take care of.

  • Why we dont need sender agreement and comm channel in IDOC to file scenario

    Hi,
    there are 2 queries:
    first query :
    Scenario is "Idoc to file transfer"
    why we dont need sender agreement and communication channel for IDOC to file ,while we do need
    both of these for "File to RFC "
    Answer : which i got is because IDOC sits at ABAP stack but same is valid for RFC also.
    Then why For "IDOC to file" we dont need that.
    Second query:
    Also when i did scenario for File1-RFC-File2 scenario using BPM i have created a interface
    between XI-RFC synchronous to send data which we got from file1 in XI to RFC importing parameter
    and getting data from RFC back to File2.
    Here in IDOC to file scenario why we dont need any interfface for IDOC-XI Asynchronous.
    Regards,
    Mayank
    Edited by: mayank verdia on Jan 13, 2009 11:49 AM
    Edited by: mayank verdia on Jan 13, 2009 11:50 AM

    Dear All,
    For file to idoc scenario is there any possibility to get line items details or xml details i.e the segments and its related field details using reference id, transaction id or interface name or message id in SAP PI 7.0. I know we need to click each and every message in sxi_monitor and look for details.
    For SAP(R3 System) I can create a report and set the job for specific time period so automatically it throws the details(like reference no, document date, invoice no from) in ftp path as .csv file. The same ftp path is maintained in program.
    I wanted to check FTP--->PI postings and I have set the job at r3 system it is working fine and Im monitoring it too.
    Now the end to end scenario is FTP--->PI--->ECC(R3 system). Please help.Many Thanks.

  • IDOC to File Scenario - Configuration steps need to be done in R3 and XI

    Hi Experts,
    I am working with IDoc to File scenario. Can someone please explain, what are the steps or configuration things need to be done in R3 and as well as in XI too.
    Thanks in advance.
    Regards,
    Sree

    Hi Sree,
    I thought i will just share a few basic thoughts that you need to take care while doing an IDOC to File Scenario. Just see if it helps you in anyways.
    While doing an IDOC -> XI -> File scenario, The Points to noted are:
    1. You don’t need a DT, MT or a message interface for the IDOC as it itself acts as the Message Interface.
    2. You import the IDOC and use the same in your mapping.
    3. In this configuration note that you don’t have a sender agreement as you don’t have a sender IDOC adapter.
    If it is that you wanted to send an IDOC from XI (File to IDOC) , then in this case:
    points 1 and 2 will remain, but the 3 will not. You will need a sender agreement for the file and a receiver agreement for the IDOC.
    IDOC -> File scenario ref:
    <a href="/people/prateek.shah/blog/2005/06/08/introduction-to-idoc-xi-file-scenario-and-complete-walk-through-for-starters scenario</a>
    File -> IDOC ref:
    <a href="/people/anish.abraham2/blog/2005/12/22/file-to-multiple-idocs-xslt-mapping to IDOC</a>
    To be able to trigger your IDOC from the SAP ISU system, you will have to set the partner profile in we20. Select your Business System (mostly under Logical system) and then create Outbound entries for whichever IDOC you want to trigger.
    You define your basic type also in your partner profile settings, Please go thru the following links to get a better idea about partner profile:
    http://help.sap.com/saphelp_nw04/helpdata/en/dc/6b833243d711d1893e0000e8323c4f/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/dc/6b7cd343d711d1893e0000e8323c4f/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/32/692037b1f10709e10000009b38f839/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/5e/b8f8bf356dc84096e4fedc2cd71426/frameset.htm
    Also, I would suggest that you go through this blog by michal if any issue arises,
    /people/michal.krawczyk2/blog/2005/03/29/xi-error--unable-to-convert-the-sender-service-to-an-ale-logical-system
    /people/michal.krawczyk2/blog/2005/06/28/xipi-faq-frequently-asked-questions
    Also have a look at these forum threads...
    IDOC TO FILE SCENARIO
    idoc to file scenario
    I hope this answers your query.
    Regards,
    Abhy Thomas

Maybe you are looking for

  • Run Time Error on Saving Sales Order

    Hello, We are facing problem when saving sales order. When order is created and save button is pressed, it saves sales order and gives Order  Number. But while saving the order during update it gives dump due to Data duplication in VBFA while inserti

  • Ignore Save as Draft just consider Add Sales Order

    Dear all, I got a problem that the following code: If BusinessObjectInfo.FormTypeEx = "139" And BusinessObjectInfo.EventType =                              SAPbouiCOM.BoEventTypes.et_FORM_DATA_ADD Then If BusinessObjectInfo.BeforeAction = False And B

  • Few questions about Photoshop CS4 Student Edition

    I'm thinking of getting a DSLR camera in the summer, so (since I'm a geek) I want to make sure I've got all the software I'll need to edit the photos once I've taken them. The catch is that I graduate from university this year, so if I'm going to buy

  • Install NW 2004s SP9 - extremely limited callable objects in GP

    I have installed the sneak preview for NW 2004s SP9 and find the selection of callable object types in the Guided Procedures design time to be extremely limited. The only callable object types are Service > Web Service, User Interface > Web Pages, Fo

  • How to group by year

    Hi, I have the following simple query: SELECT pubdate, title FROM table ORDER BY date DESC Which returns hundreds of records. Instead of outputing every single date within a year for pubdate, I would like to group the output based on the year. So the