SFTP Sender - content conversion to read Header & detail and control record

Hi Experts,
    I'm just beginner in SAP PI. Just would like to know possibility for below requirement for my further reseach
Soure file format fix length  with 3 different format(H,D,T)
As example
H001TestHeader1
D001001TestDetail11
D001002TestDetail21
T2
H002TestHeader2
D002001TestDetail21
T2
The requirement detail :
1. There are multiple set of H,D,T per files
2.  T line is the record count of total D line per each H set.  for example as green font we have 2 D line, So T show T2
3.  If number in T line is not match with actual D line, then only these set should be filter out and write to error file and send email alert. for example as red font we have 1 D line, but T show T2
4.  If number in T line is match with actual D line, then these set will be write to result file.
As I know that content  conversion may be support this requirement for separate records set for H/D/T but I don't have idea how to handle requirement no.3 and 4, Is it need ccBPM? Please help to provide some overview guideline for my further research. Thank you.

As I know that content  conversion may be support this requirement for separate records set for H/D/T
>>> For the  FCC, you might be using Message Transformation bean and unfortunately MTB doesn't provide the option to split the message i.e. " Records set per message".
If you go for two step approach i.e.  then you can use this option to split the records.
1) pass through interface -> SFTP -> NFS
2) NFS(Standard FCC) -> Your target system
I don't have idea how to handle requirement no.3 and 4, Is it need ccBPM? Please help to provide some overview guideline for my further research. Thank you.
>>> If you can split the file for each recordset i.e.  one set of H/D/T then you can handle the validation part in message mapping.
If T has valid number of records then execute the mapping set the filename  and directory to "result file" locations using dynamic configuration.
If T has invalid number of records then execute the mapping set the filename  and directory to "error file" location.
To trigger the email from the message mapping refer to the below blogs -
Mail without email adapter? Part - I - Process Integration - SCN Wiki
How to send Alerts using UDF or JAVA Mapping

Similar Messages

  • Multi-Header, Detail and Trailer Record

    Hi,
    I have to plants, For Each plant i need Header, no.of Details and one trailer Record i need.
    Like
    Plant 1000
    Header1
    Deatail Record1
    Deatail Record2
    Deatail Record3
    Trailer1
    for plant 2000
    Header2
    Deatail Record1
    Deatail Record2
    Deatail Record3
    Trailer2
    Thanks!
    Regards,
    SReddy

    Hi,
    If you could give us the source and target structure with occurrences, it could be easy for us to suggest.
    Best regards,
    raj.

  • How to Seggragate Header, Footer and Trailer Records in Informatica Cloud ?

    Hi All, This is my source file Structure which is Flat File. Source File : "RecordType","Creation Date","Interface Name""H","06-08-2015","SFC02""RecordType","Account Number","Payoff Amount","Good Through Date""D","123456787","2356.14","06-08-2015""D","12347","2356.14","06-08-2015""D","123487","235.14","06-08-2015""RecordType","Creation Date","TotalRecordCount""T","06-08-2015","5" The Source File has to be loaded into three targets for Header , Detail and Trailer records separately. Target Files: File 1 : Header.txt
    "RecordType","Creation Date","Interface Name""H","06-08-2015","SFC02" File 2 : Detail.txt "RecordType","Account Number","Payoff Amount","Good Through Date""D","123456787","2356.14","06-08-2015""D","12347","2356.14","06-08-2015""D","123487","235.14","06-08-2015" File 3 : Trailer.txt "RecordType","Creation Date","TotalRecordCount""T","06-08-2015","5"  I tired this solution below :  1.  Source ---> Expression ]-----filter 1---Detail.txt                                        -----filter 2---Trailer.txt  In source , I will read the records starting from 3rd row. This is because, if i read from first row, the detail part contains more fields when compared to Header part. Header Part contains only three fields. So it is taking only first three fields records in Detail Section as well.That's why I am skipping the first two records(Header fields and header record) .. refer the example.. In filter 1, condition is Record_Type = 'D'. In Filter 2 , condition is Record_Type = 'T'.So, the filter 1 will load to Detail.txt and Filter 2 Will load to Trailer.txt In task , pre session command, Calling the windows .bat script to fetch the first two lines and load into Header.txt  This solution is working fine..  My query is can we use two pipeline flow in a same mapping in Informatica cloud.?  Pipeline Flow 1  Source ---> Expression ]-----filter 1---Detail.txt                                    -----filter 2---Trailer.txtPipeline Flow 2  Source ---> Expression ]-----filter 3(Record_Type='H')---Header.txt Source file is same in two flows. In first flow, I will read from the third row, skipping the header section as I mentioned earlier.In second flow , I ll read the entire content and take only header record and load into target.  If I add the flow 2 to existing flow 1.  I am getting the below error..TE_7020 Internal error. The Source Qualifier [Header_Source] contains an unbound field [Creation_Date]. Contact Informatica Global Customer Support.  1. Do informatica Cloud supports, two parallel flow in a same mapping ?2. Any other best solution for my requirement?   Since I am new to Informatica Cloud, Can anyone suggest any other solution if you have?? It will be more helpful if you guys suggest a good solution ..  ThanksSindhu Ravindran

    We are using a Webservices Consumer Tranformation in our mapping to connect to RightFax Server using a WSDL url via Business Service in a mapping.Here in the mapping, where we are sending the input parameters through a flat file and then connecting to Rightfax Server via WS Consumer transformation and then fetching the data and writing to a Flat File.  07/28/2015 10:10:49 **** Importing Connection: Conn_000A7T0B0000000000EM ...07/28/2015 10:10:49 **** Importing Source Definition: SRC_RightFax_txt ...07/28/2015 10:10:49 **** Importing Target Definition: GetFax_txt ...07/28/2015 10:10:49 **** Importing Target Definition: FaultGroup_txt ...07/28/2015 10:10:49 **** Importing SessionConfig: default_session_config ...    <Warning> :  The Error Log DB Connection value should have Relational: as the prefix.    <Warning> :  Invalid value  for attribute Error Log DB Connection. Will use the default value     Validating Source Definition  SRC_RightFax_txt...    Validating Target Definition  FaultGroup_txt...    Validating Target Definition  GetFax_txt...07/28/2015 10:10:49 **** Importing Mapping: Mapping0 ...    <Warning> :  transformation: RIghtfax - Invalid value  for attribute Output is repeatable. Will use the default value Never    [transformation< RIghtfax > ] Validating transformations of mapping Mapping0...Validating mapping variable(s).07/28/2015 10:10:50 **** Importing Workflow: wf_mtt_000A7T0Z00000000001M ...    <Warning> :  Invalid value Optimize throughout for attribute Concurrent read partitioning. Will use the default value Optimize throughput   [Session< s_mtt_000A7T0Z00000000001M >  --> File Reader< File Reader > ]     <Warning> :  The value entered is not a valid integer.    <Warning> :  Invalid value NO for attribute Fail task after wait time. Will use the default value Successfully extracted session instance [s_mtt_000A7T0Z00000000001M].  Starting repository sequence id is [1048287470] Kindly provide us a solution. Attached are logs for reference.

  • Problem in message mapping due to file sender content conversion

    Hi all
    basically i m using file sender content conversion and couldn't make file in a proper nested structure where weight was the nested field of detail as there is a limitation of content conversion in file adapter:
    My file came out like this:
    <header>
    </header>
    <Detail>
    </Detail>
    <Weight>
    </Weight>
    <Weight>
    </Weight>
    <Detail>
    </Detail>
    <Weight>
    </Weight>
    <Weight>
    </Weight>
    <trailer>
    </trailer>
    and i wanted it like:
    <header>
    </header>
    <Detail>
    <Weight>
    </Weight>
    <Weight>
    </Weight>
    </Detail>
    <Detail>
    <Weight>
    </Weight>
    <Weight>
    </Weight>
    </Detail>
    <trailer>
    </trailer>
    so now the problem is i want to generate an idoc from it....but theres no way out to determine which weight is related to what detail except that the weight related to detail 1 comes after start and end of detail 1 tag and then weights of second detail comes after start and end of detail 2 tag in my source xml file.
    The structure of the target idoc is
    <header>
    </header>
    <Detail>
    <Weight>
    <flag>D1</flag>
    </Weight>
    <Weight>
    <flag>D1</flag>
    </Weight>
    </Detail>
    <Detail>
    <Weight>
    <flag>D2</flag>
    </Weight>
    <Weight>
    <flag>D2</flag>
    </Weight>
    </Detail>
    <trailer>
    </trailer>
    so i was wondering if i can make a user defined function in which i feed detail and weight and result a flag d1/ d2 inside the tag field. My logic is smth like
    while detail is 1 make all the tag weights coming after detail 1 inside <detail1> tag then resultlist.CC for context change  then while detail is 2 make a loop to fill the weight coming after detail 2 inside the <detail2>
    Is this possible by any means. I really need help
    Thnks
    Naina

    Hi
    My sender file content conversion have this setting:
    HEADER,1,DETAIL,*,Weights,*,TRAILER,*
    And plz give me more idea of the adapter module ....is there any blog available which is related to my problem.
    Naina

  • ...file Sender content conversion 'lastFieldsOptional'  error

    Hi All,
    I am working on a File Sender content conversion--> flat file to XI.
    it's fixed length field all the rows with same column.
    090627 344535AFDFG+GBP65433 ASDSSD GFD dFSSGFD 6757532
    090627 344535AFDFG-GBP65433 ASDSSD GFD dFSSGFD 6757532
    090627 344535AFDFG-GBP65433 ASDSSD GFD dFSSGFD 6757532
    090627 344535AFDFG-GBP65433 ASDSSD GFD dFSSGFD 6757532
    090628 344536AFDFG+GBP45434 ASDSSD GFD dFSSGFD 6757532
    090628 344536AFDFG-GBP45434 ASDSSD GFD dFSSGFD 6757532
    **EOF**
    I am sucessfully able to handle the file if i remove "**EOF*" from incoming file. But i get "lastfieldsOptional" error with "*EOF**"
    >
    Parameters-
         Recordset Structure - ROW,*
         Recordsets per message - *
        ROW.fieldFixedLengths - 6,8,20,10,30,1,3,1,11,1,11,6,10,10,20,10,10,2,6,6,2,6,6,3
        ROW.fieldNames - INV_DATE,INV_NO,PAYMENT_REF,CUST_NO,CUST_NAME...etc etc
       ignoreRecordsetName - true
    There are parameters available to ignore Last field(last column) of a Row/Rocerdset
    but how to ignore the last row of the file.
    Plz suggest some parameter to ignore the last row/field of the file.
    Regs,
    Ansh

    Ansh,
    .lastFieldsOptional  is obselete and you cannot use. Please see sap help:
    http://help.sap.com/saphelp_nw04/helpdata/en/2c/181077dd7d6b4ea6a8029b20bf7e55/content.htm
    The best thing comes to my mind is use:
    .keepIncompleteFields as YES
    Also give a try with:
    .missingLastfields as add
    So with this I think your last line in the file will read into XI. But the value of last line is **EOF**. So the length is 9. Since your the first two fixed lengths are 6 and 8. You will have these values as ***EOF and ***. So while mapping you can use doesnot satrt with * and map it, so that your last line will be ignored.
    Note: with this apporach you have to make sure your first two columns never start with *. If you have a doubt then you can use the condition doesnot equal to ***EOF for first field and doesnot equal to *** for the second field when you map. I hope it makes sense to you.
    Regards,
    ---Satish

  • File sender content conversion

    hello
    I have a txt file sender content conversion in SAP PI.
    I define in the content conversion 6 fields names(field1, field2, field3,field4,field5,field 6) but the fie that I am loading has only three field(field1,field2,field3). the file is still loading though there is a different in the structure between what I define in the PI to what is loading to the PI.
    is there a way to send an error with out loading the file?
    I read about the count function in the mapping. isnt there a build in function that suppose to load only the structure that is define?
    Thanks
    Kfir

    If you define 6 fields and only 3 are present in the test file then you wont get an error (AFAIK)
    However, if you define 3 fields and there are 6 present in the source file then you will get an error saying more number of parameters found.
    isnt there a build in function that suppose to load only the structure that is define?
    I dont think so....this may however be achieved using a custom adapter module.
    Update:
    Just remembered that SAP PI (7.1 and above) comes with inbuilt XML-validation function....check if that helps in your design.
    Regards,
    Abhishek.
    Edited by: abhishek salvi on Dec 15, 2010 1:12 PM

  • Sender content conversion parameters

    what wld be the sender content conversion parameters for the following structure?
    MessageType        1-1
        Records            1-1
           File                1-unbounded
    thanks

    Hi,
    If you have not come across:
    http://help.sap.com/saphelp_nw04/helpdata/en/e3/94007075cae04f930cc4c034e411e1/frameset.htm
    would be of help.
    Regards,
    Manjusha.

  • File Sender Content Conversion: Help needed

    Hello Experts,
    i need help with file sender content conversion:
    i have a file which looks like this:
    12329460  24.01.09/07:01  167     Y010122851  136086  43300007            E70115  1L2_96_1
    12329660  25.01.09/07:02  157     Y010122851  136086  43390007            E711J5  1L2_96_1
    as you can see 8 fields, separated with whitespaces
    and i want and xml file which looks like this:
    <DT_DATA_FILESENDER>
      <Recordset>
          <Data> 
            <field1>12329460</field1>
            <field2>24.01.09/07:01</field2>
            <field3>167</field3>    
            <field4>Y010122851</field4> 
            <field5>136086</field5> 
            <field6>43300007</field6>
            <field7>E70115</field7> 
            <field8>1L2_96_1</field8
         </Data>
          <Data> 
            <field1>12329660</field1>
            <field2>25.01.09/07:02</field2>
            <field3>157</field3>    
            <field4>Y010122851</field4> 
            <field5>136086</field5> 
            <field6>43390007</field6>
            <field7>E711J5</field7> 
            <field8>1L2_96_1</field8
         </Data>
      </Recordset>
    Would you please let me know how the datatype has to look like?
    And especially how the FCC has to be configured for this scenario?
    Thanks in advance,
    Chris

    > With fixed lengths i get it working,
    >
    > but with de fieldseparator 0X09 it does not work,
    > would you please tell me the complete FCC config not only the fieldSeparation line?
    Well Christian,
    There is one thing, eithere you can use fieldSeparator or fieldFixedLengths you can not use both together. So in your case do not use fieldSeparator.
    The complete FCC is almost same as given in the blog in my previous reply. See the final output in that blog and create your data type accordingly. e.g.
    Recordset
    ------Item 0...unbound
    --------Field1 0..1
    --------Field2 0..1
    --------Field3 0..1
    --------Field8 0..1
    Regards,
    Sarvesh

  • Challenge in File Sender Content Conversion

    Hi,
    I have a real challenge concerning File Sender content conversion in SAP XI.
    My flat file looks like this:
    ##H   300
    MAR   206
    KAS 1
    DAT 01.03.08
    ART 1.129
    KUN 118
    EAN 4.499
    REL 5.0j 16.05.06
    SER             1
    ##E   300
    ##H   301
    DAT 01.03.08
    ZEI 07:54
    KAS 1
    ##E   301
    Each row represents a data field and has two values: The first one defines the field name, the second represents the field value. E.g. 'DAT' stands for Date and has the value 01.03.08 in the example.
    The fields belonging together in one data set are enclosed by a start qualiefier (##H)and an end qualifier (##E).
    The value after these qualifiers (i.e. '300' and '301' in the example above) represent a certain record type, e.g. '300' represents Customer data and '301' represents Account Data.
    Is it possible with file content conversion to create the following XML structure:
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:MT_DATA xmlns:ns0="http://sap.com/xi/account">
         <300>
              <MAR>206</MAR>
              <KAS>1</KAS>
              <DAT>01.03.08</DAT>
              <ART>1.129</ART>
              <KUN>118</KUN>
              <EAN>4.499</EAN>
              <REL>5.0j 16.05.06</REL>
              <SER>1</SER>
         </300>
         <301>
              <DAT>01.03.08</DAT>
              <ZEI>07:54</ZEI>     
              <KAS>1</KAS>
         </301>
    </ns0:MT_DATA>
    If it is not possible with content conversion, what could be an alternative? Adapter Module? MultiMapping?
    I'd really appreciate your input as I am working on this since several days without a solution.
    Thanks in advance.
    Alex

    It is not possible thru adapter..
    There is one blog to convert such files to XML sing java mapping. Search for any flat file to Idoc or xml in the forum.
    VJ

  • Fixed width file outgoing with header,detail and trailer in one map

    we have a scenario where in we have to create an output file. the file is fixed width with three different kinds of records namely header, details and trailer. if the specification of the metadata of the file is specified how will we incorporate the creation of the file in one map since the map produces only one call to UTL_FILE and writes it.
    there is a work around in a crude way where in we can define the file as a CSV and pass everything in an appanded string in one column. but i wanted to know how we can get the three logical records in one map..
    thanks
    kamal

    The problem is not with generating the records.. i mean header , detail and trailer.
    it is how we put them into a fixed width sampled file. if we look at the code generated we will see only one call being made to generate the file.

  • How to define a schema for fixed length consisting of multiple records of different types having header detail and footer at positions 22 and 23 namely having 10 for header 20 for detail and 30 for footer

    how to define a schema for fixed length consisting of multiple records of different types having header detail and footer at positions 22 and 23 namely having 10 for header ,20 for detail and 30 for footer.
    here is the sample file
    DDWTYBILL13092502572 1000000112\\NHLSFS\ORACLE\ORACLE\BILLING\VERIFY\            2013/09/25 01:19:08 BILLEND     
    DDWTYBILL13092502572 2000000212PRIVATE             PRIVATE             STY5572142     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731543                                           Suspect TB ?CA, (L) EMPIEEM                                           Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000312PRIVATE             PRIVATE             STY5572149     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731548                                           Empyema BPF                                                           Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.9                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000412PRIVATE             PRIVATE             STY5572351     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731584                                           Suspect TB ?EMPIEEM, ?CA                                              Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000512PRIVATE             PRIVATE             STY5572352     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731585                                           Suspect TB ?EMPIEEM, ?CA                                              Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000612MAJOR               MAJOR               STY5577530     010700         INV                              12ZTYG000009                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNS    ZOZI,S,SIYAMBONGA                                                                                                                 27-MAR-2010                              TYG       D3        DR MAHOMED                                                                                          5602181             113799662      TB731663                                           Lymphadenitis                                                         Ward D3                                           TYGERBERG HOSPITAL                                                                                  I88.9                                                       Z03.9                                                       H1 07-AUG-201307-AUG-201324-SEP-2013MEND     58.52
    DDWTYBILL13092502572 2000000712MAJOR               MAJOR               STY5577537     010700         INV                              12ZTYG000009                  06-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNS    DLAMINI,N,NTLANTLA                                                                                                                09-OCT-1976                              TYG       F1        DR WZESE                                                                                            5602181             135524676      TB731666                                           Suspect TB                                                            Ward F1                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                       H1 07-AUG-201314-AUG-201324-SEP-2013MEND     58.52
    DDWTYBILL13092502572 30003034120000003032END

    You can create an nXSD using Native Format Builder and use lookAhead attribute to read values at specific position and use that as the choice. You can find an example here
    http://docs.oracle.com/cd/E11036_01/integrate.1013/b28994/nfb.htm#BGBBAJFD

  • Re: to create outbound file with multiple header ,detail and trailor in informatica

    Dear Mohan Prakash, Please mentation The Target Which Format you need. Please let me know. Thanks & RegardsKasireddy+966545281845

    hi I want to create a file with muiltple header detail and trailorSample of my file  1F99500094959                    5F99000000999911025F99000000999912025F99000000999913029F99500094959                    1G83341002729803                5G83000000999918020109G83341002729803                1G83910377940                    5G830000009999190201011050003522029G83910377940                    T20110720000000900000006750{  In this file 1 is header and 5 is detail and 9 is trailer and in last T segment is also trailerI have any idea to create seperate 3 files and in lastconcatenate all file with the help of unixbut in my case data is so large (in millions) this approach is not use full please suggest my any ideaHow to create VSAM file

  • File Sender, Content Conversion - how to define variable length last field?

    XI 3.0 SP17
    With a File Sender communication channel, that uses Content Conversion - how do I define a 'variable length' last field?
    The scenario - the input file has four fields, of which the first three are a known fixed length, and the last (fourth, trailing) field is variable in length.
    Using a Message Protocol of 'File Content Conversion', how do I define that last variable length field (field name 'WOData' below) in the Content Conversion Parameters section?
    My current parameters are:
    Recordset Structure  -  Row,*
    ignoreRecordsetName  -  true
    Row.fieldFixedLengths  -  1,12,5,99999
    Row.fieldNames  -  WOType,WONum,WOLine,WOData
    I've tried the following for 'Row.fieldFixedLengths' to no avail -
    '1,12,5,*'
    '1,12,5,0'
    '1,12,5,'
    '1,12,5'
    The last two were grasping at straws )
    The only thing I've got to work is specifying a 'large' value for the final field (99999 above).
    In addition, does anyone know if specifying a large value (e.g. 99999) for the final trailing field will give rise to performance issues when the file is being processed?
    In the help for "Converting File Content in a Sender Adapter", it states -
    <Begin Quote>
    NameA.fieldFixedLengths
    If you make a specification here, the system expects a character string that contains the lengths of the structure columns as arguments separated by commas.
    If you also specify a separator for the columns, you must not add its length to the length of the columns.
    This entry is mandatory if you have not made an entry for NameA.fieldSeparator.
    <End Quote>
    http://help.sap.com/saphelp_nw04/helpdata/en/2c/181077dd7d6b4ea6a8029b20bf7e55/content.htm

    << note that fieldFixedLengths will not take any wildcard entries like *. So in these case it is ideal to provide a maximum char length.  But note that while the file is being created that many spaces will be created in your file !!! >>
    Hi Shabarish,
    Yes, no wildcard is the conclusion I came to, hence my maximum )
    The message size did not increase by any 'blank padding'.  When I look in [Message Display Tool (Detail Display)] 'Audit Log for Message: X'  -
    2006-10-17 18:22:42 Success Channel X: Entire file content converted to XML format
    2006-10-17 18:22:42 Success Send binary file  "X" from FTP server "X", size 103290 bytes with QoS EO
    2006-10-17 18:22:42 Success Application attempting to send an XI message asynchronously using connection AFW.
    2006-10-17 18:22:42 Success Trying to put the message into the send queue.
    2006-10-17 18:22:42 Success Message successfully put into the queue.
    2006-10-17 18:22:42 Success The application sent the message asynchronously using connection AFW. Returning to application.
    The input flat file in non-XML format was 92,132 bytes and the message payload into XI was 103,290 bytes.
    My understanding is that trailing spaces are stripped from XML nodes.

  • Content Conversion issue for header record

    Hi,
    We have a very urgent question on an issue here with one of our XI objects. 
    This is an inbound interface from an external system into R/3 & BW.  The inbound file has a header record (with about 8 fields) and detail records (about 900 fields per detail record). Data going into R/3 & BW don't have header records and everything goes in as detail records. One field from the header of this source file should be passed to the target structure at the detail level. Also, we are NOT using BPM.
    Can someone help us how we could define the file content conversion parameters for File adapter.
    Thanks in advance ......
    Prashant

    I'm so sorry, I wasn't subscribed to this thread and I didn't realize there were responses.
    If you have a message type made up of a Header with 1 occurence and Detail with 1 to unbounded occurunces, you'd want to do the following in content conversion:
    Document Name - your message type
    Document Namespac - your message type namespace
    Recordset Structure - Header,1,Detail,*
    Recordset Structure - Ascending
    Then you'll need to set some of the parameters, depending on the layout of your incoming file. 
    As for the problem of having hundreds of fields, I'm less sure about that.
    Would it be possible to break your detail data type down into smaller data types.  Each with fewer fields.  You'd still have to maintain every field in content conversion, but at least they'd be in seperate parameters, instead of all 900 in one tiny box.
    Here's a very rough example of what I mean:
    If you have 900 fields, instead of making 1 data type of detail, you could make 9 data types, Detail1, Detail2, Detail3, Detail4, Detail5, Detail6, Detail7,Detail8, Detail 9, each with 100 fields in them (or more with even less fields).
    Setting things up the file content conversion would be more complex in this scenario, so it might be a toss up if it's worth it to break it up this way or not if it meant configuring quite a few more parameters.
    For example,
    You'd have to declare your recordset structure like Header,1,Detail1,,Detail2,,Detail3,* etc, and you'd have to make sure to set the .endSeparator to '0' for all of the first 8 details, so it would recognize that they were all on one line.
    I hope this helps a little bit.

  • Ignore record in JMS sender content conversion

    Hi,
    I am using JMS sender adapter with content conversion.
    In the input text file, I have to ignore the first record.
    for this, in Module tab I have given parameter like this:
    Plain2XML -> documentOffset -> 1
    so that it will not read the first record in the text file.
    But it is not working
    Please suggest the right way.
    Regards,
    Anil.

    It should be:
    xml.documentSkipFirstRows
    Regards
    Stefan

Maybe you are looking for

  • Zen Touch Sensitivity & Carrying Case Is

    I have had the pleasure of owning the Zen Touch for a little over two months and I am consistently impressed with the quality of playback. I do, however, have to agree with some reviews and posts that talk of the over sensitivity of the touch pad. I

  • Multiple values selection implementation in VC

    Hi I am working with VC. We are trying to implement an multiple input selection. Our first option was to use the Help value wizard but i saw in this forums that is just possible with BI systems (we are using BAPI thus R/3 system). Now, we are trying

  • How-to: Servlet -- Servlet comunication.

    The scheme is about some servlets comunicating with others that realize the data set of the structure. How is the best way to let the servlets comunicate between them? Java is offering different techniques for let comunicating like http requests, dat

  • U20: confusing keyboard layout within X-Server

    Solaris x86 on U20 preinstalled using type 6 usb country-kit first start after automatic sys-unconfig while selecting iso8859-1 german keyboard/keystrokes are fine in console-mode keyboard is bad when using x-server (every key is right shifted for 1

  • Notepad Opening By Default Instead of MS Word

    Dear Experts, When I try to enter text field in Ex: ME51N,ME21N notepad is opening by default instead of MS Word. When I double click on the text tab by default notepad is opening, I want MS word to open. Please let me know how to make the changes. T