Flat file conversion issue

Hi all,
I'm doing a IDOC to Flat file scenario, now after the mapping and FCC I have an output in this way:
SOLD_TO,SHIP_TO,BILL_TO,0000300623,0000300622,0000500569
But I want it in this way:
SOLD_TO,0000300623,SHIP_TO,0000300622,BILL_TO,0000500569
What should I do in FCC to get the desired output?
Thanks,
Srinivas

Suman,
MT
DT
ADDR_TYPE
EXT_ADDR_ID
These two fields I have duplicated twice each in the mapping:
MT
DT
ADDR_TYPE
SOLD_TO
ADDR_TYPE
SHIP_TO
ADDR_TYPE
BILL_TO
EXT_ADDR_ID
0000300623
EXT_ADDR_ID
0000300622
EXT_ADDR_ID
0000500569
Srini

Similar Messages

  • Receiver CC File - Complex Flat File Conversion

    Dear forum users,
    I want your opinion about the feasibility of the content conversion described below using Receiver CC.
    The structure of my file is the following
    - A File header segment (FH)
    - 1...N Business Object that is described by many segments
    - A file footer segment (FF)
    Exemple of business Object : Purchase Order :
    1 Header Segment (POH)
    N Item Segments (POI)
    Exemple of file :
    FH
    POH (PO 1)
    POI (PO 1)
    POI (PO 1)
    POH (PO 2)
    POI (PO 2)
    POI (PO 2)
    FF
    Is content conversion standard tool can deal with such a structure ?
    Note that in the file created, the order of each segment has a big importance :
    PO Items of PO 1 must be after PO Header of PO 1 but before PO Header of PO 2 !
    I thnik the structure requested by converter described in documentation do not fit with my case :
    <root>...
      <nameA>
            <value1>value</value1>
            <value2>value</value2>
            <value3>value</value3>
      </nameA>
      <nameB>
            <value4>value</value4>
      </nameB>
    </root>
    I should have :
    <root>
    <FH>
      <PO>
        <POH>...</POH>
        <POI>...</POI>
        <POI>...</POI>
      </PO>
      <PO>
        <POH>...</POH>
        <POI>...</POI>
        <POI>...</POI>
      </PO>
    </FH>
    </root>
    Thanks very much for your help and opinions
    Jean-Charles.
    Edited by: Jean-Charles CARRET on Jan 3, 2011 6:12 PM

    Thank you Stephan for your answer.
    Firstly, could you confirm whether the "row" elements under the root tag are on the same level than other elements (POH, POI, FF and FH) or not ?
    I just check the both following parts of the help documentation called :
    - [Inserting MessageTransformBean in the Module Processor|http://help.sap.com/saphelp_nwpi711/helpdata/en/57/0b2c4142aef623e10000000a155106/frameset.htm]
    - Using the dispatcher : Example
    For me, it is not clear.
    1. Do I need to create an external class (XSLT Library for example) to manipulate my XML data before passing it to the standard Flat File Conversion tool ? (Here the manipulation would be to remove the row XML elements)
    2. Where is located the famous "Data" directory in which I must put my xsl file ?
    Thanks for clarifications.
    Jean-Charles

  • XML to flat file conversion

    Hi,
    Iam working on Idoc to file scenatio.Using java mapping I have mapped the Idoc to an XML format. Now I have a problem in converting the XML format to flat file format.
    My XML is as follows:--
    <Record_Idoc>
      <Record>
        <Header>
            <Order_no>10000398</Order_no>
            <Order_Type>ZSTQ</Order_Type>
            <Material_Head_no>1800406</Material_Head_no>
            <Material_Head_Desc>MARC MOL MEZ R16 FY02.</Material_Head_Desc>
            <Production_version>0003</Production_version>
            <Test_Order/>
            <Quantity_planned>801.00</Quantity_planned>
            <unit>PCE</unit>
         </Header>
        <operations>
             <operation_number>0010</operation_number>
             <Work_center>P_SIG</Work_center>
             <setup_start_date>20020619</setup_start_date>
             <Setup_start_time>060000</Setup_start_time>
             <Processing_start_date>20020619</Processing_start_date>
             <Processing_Start_Time>060000</Processing_Start_Time>
             <End_date>20020619</End_date>
            <End_Time>061600</End_Time>
         </operations>
        <operations>
             <operation_number>0020</operation_number>
             <Work_center>P_BFB</Work_center>
             <setup_start_date>20020619</setup_start_date>
             <Setup_start_time>060000</Setup_start_time>
                <Processing_start_date>20020619</Processing_start_date>
             <Processing_Start_Time>060000</Processing_Start_Time>
             <End_date>20020619</End_date>
             <End_Time>074926</End_Time>
          </operations>
      </Record>
    </Record_Idoc>
    My output flat file should be as follows:--
    Each record contains the following lines
    Header.........
    Operations......
    Operations......
    I have used File content Conversion in message Protocol
    These are my setting:--
    Header.addHeaderLine  nl
    Header.fieldSeparator  ,
    Header.endSeparator   nl
    operations.addHeaderLine  nl
    operations.fieldSeparator  ,
    operations.endSeparator   nl
    But seems that these settings does not work.
    Could some one help in solving this issue.
    regards
    Praveen

    try,
    Recordset Structure - Header,operations
    Header.fieldSeparator ,
    Header.endSeparator nl
    Header.addHeaderLine X  *****
    operations.fieldSeparator ,
    operations.endSeparator nl
    and in case it doesnt work try removing the Record_Idoc node from the XML in mapping and then try the Cont. Conv.
    X can be
    0 – No header line
    1 – Header line with column names from the XML document
    2 – As for 1, followed by a blank line
    3 – Header line is stored as NameA.headerLine in the configuration and is applied
    4 – As for 3, followed by a blank line

  • Flat file load issue-Urgent

    Hello
    I'm trying to load a master data infoobject with a flat file and getting errors.
    The infoobject ZCOLLECT_DATE has two fields - ZQUARTER and ZCOLLECTDATE
    ZCOLLECT_DATE is of Data type NUMC, Length 5, conversion routine PERI5 and Output length 6
    I created a flat file in Excel the following way:
    Left the first row blank
    Second column - 1/2008 (for ZQUARTER)
    Thrid column - 01/11/2008 (for ZCOLLECTDATE)
    I saved it as a CSV (save as-> csv).
    In the infopackage, I've selected this file in the External data tab, File type CSV, Data separator ; , Escape sign "
    When I try to load the infoobject, I get the following error:
    "Error in conversion exit CONVERSION_EXIT_PERI5_INPUT"
    I thought this was because the Length in my infoobject was 5, whereas ZQUARTER is of length 6 in my flat file (eg-1/2008) and ZCOLLECTDATE is of format mm/dd/yyyy.
    Any thoughts on what the issue is?
    Thanks
    Alex

    Hi Hemant
    I changed it to 12008, 22008 etc.. but still get the error ""Error in conversion exit CONVERSION_EXIT_PERI5_INPUT"
    Also, when I save the xls file as CSV, I get two pop-ups saying some features are not compatible and only those features that are compatible will be saved. Are these normal warnings, or could they be causing the issue?
    Also, in the infoobject, since it only consists of dates, do I need to specify conversion routine PERI5 in Infoobject properties/maintenance?
    Thanks

  • RFBIBL00 Flat File Creation issue

    Hello All,
    I'm using "RFBIBL00" to post the FI documents, and the input to the FB01 transaction screen is obtained from a flat file from AL11 directory.
    The issue here is regarding the date format of the Document date and Posting date. If the date format is as YYYYMMDD in the flat file, and if I'm using the date parameter (in the default option of one's SAP own data) other than starting with YYYY, the document is not getting posted and it is throwing error saying that date format is not correct. If I have chosen the date format starting with YYYY, then the document is getting posted successfully.
    Any suggestions/ideas to make the date format as independent?
    Thanks,
    Bharath.

    Hi Bharath
    I think you'll have to use the appropriate conversion exit for the same. Choose from the below options which suit you.
    SAP Internal format is 'YYYYMMDD' Example: '19990123'
    1. CONVERT_DATE_TO_EXTERNAL
    2. CONVERT_DATE_TO_INTERNAL
    3. CONVERSION_EXIT_PDATE_INPUT Conversion Exit for Domain GBDAT: DD/MM/YYYY -> YYYYMMDD
    4. CONVERSION_EXIT_PDATE_OUTPUT Conversion Exit for Domain GBDAT: YYYYMMDD -> DD/MM/YYYY
    5. CONVERSION_EXIT_IDATE_INPUT External date INPUT conversion exit
    6. CONVERSION_EXIT_IDATE_OUTPUT External date OUTPUT conversion exit
    7. CONVERSION_EXIT_LDATE_OUTPUT Internal date OUTPUT conversion exit (e.g. YYYYMMDD)
    8. CONVERSION_EXIT_SDATE_INPUT External date (e.g. 01.JAN.1994) INPUT conversion exit
    9. CONVERSION_EXIT_SDATE_OUTPUT Internal date OUTPUT conversion exit (e.g. YYYYMMDD)
    The date coming in from the flat file will certainly tick to one particular format i guess.
    Hope this helps.
    Harsh

  • Flat file truncation issue

    I am attempting to perform a fairly standard operation, extract a table to a flat file.
    I have set all the schemas, models and interfaces, and the file is produced how I want it, apart from one thing. In the source, one field is 100 characters long, and in the output, it needs to be 25.
    I have set the destination model to have a column physical and logical length of 25.
    Looking at the documentation presented at http://docs.oracle.com/cd/E25054_01/integrate.1111/e12644/files.htm - this suggests that setting the file driver up to truncate fields should solve the issue.
    However, building a new file driver using the string 'jdbc:snps:dbfile?TRUNC_FIXED_STRINGS=TRUE&TRUNC_DEL_STRINGS=TRUE' does not appear to truncate the output.
    I noticed a discrepancy in the documentation - the page above notes 'Truncates strings to the field size for fixed files'. The help tooltip in ODI notes 'Truncates the strings from the fixed files to the field size'. Which might explain the observed lack of truncation.
    My question is - what is the way to enforce field sizes in a flat file output?
    I could truncate the fields separately in each of the mapping statements using substr, but that seems counter-intuitive, and losing the benefits of the tool.
    Using ODI Version:
    Standalone Edition Version 11.1.1 - Build ODI_11.1.1.5.0_GENERIC_110422.1001

    bump
    If this is an elementary issue, please let me know what I've missed in the manual.

  • Handling flat file conversion with spaces inbetween  at sender side

    Hi,
       I am facing some problem in configuring the sender JMS adapter file content conversion. Please find the structure of my file below
    010AG  07/17/2007 000130800 TOZ07/17/200710:48:46
    010AU  07/17/2007 006682800 TOZ07/17/200710:48:46
    010-Record key
    AG-Metal code
    07/17/2007 -price Date
    000130800 -pricevalue
    TOZ-Unitofmessure
    07/17/200710:48:46-Unitofmessure
    there are 2spaces inbeween 1and 2nd fields and one space beween 2nd and 3rd ,one space between 3rd and 4th fileds
    I declared my source data strucute like below
    <source_MT>
    <PriceData>
         <Metal code>
          <price Date>
          <pricevalue>
          <Unitofmessure>
           <shipDate>
    </priceData>
    </source_MT>
    I am using this PDF to configure my serder communication channel https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/50061bd9-e56e-2910-3495-c5faa652b7
    . but i got struck up in declaring these two fields as i need deal with spaces.
    xml.NameA.fieldNames
    xml.NameA.fieldFixedLengths
    It would be great if sombody tell me how ican decalre content conversion rules for the file
    Thanks
    sudheer

    Hi,
    Please check some links on FCC.
    Introduction to simple(File-XI-File)scenario and complete walk through for starters(Part1)
    Introduction to simple (File-XI-File)scenario and complete walk through for starters(Part2)
    File Receiver with Content Conversion
    Content Conversion (Pattern/Random content in input file)
    NAB the TAB (File Adapter)
    Introduction to simple(File-XI-File)scenario and complete walk through for starters(Part1)
    Introduction to simple (File-XI-File)scenario and complete walk through for starters(Part2)
    How to send a flat file with various field lengths and variable substructures to XI 3.0
    Content Conversion (Pattern/Random content in input file)
    NAB the TAB (File Adapter)
    File Content Conversion for Unequal Number of Columns
    Content Conversion ( The Key Field Problem )
    The specified item was not found.
    File Receiver with Content Conversion
    http://help.sap.com/saphelp_nw04/helpdata/en/d2/bab440c97f3716e10000000a155106/content.htm
    Regards,
    Phani

  • Urgent: Flat file load issue

    Hi Guru's,
    Doing loading in data target ODS via flat file, problem is that flat files shows date field values correct of 8 characters but when we do preview it shows 7 characters and loading is not going through.
    Anyone knows where the problem is why in the preview screen it shows 7 characters of date when flat file has 8 characters.
    Thanks
    MK

    Hi Bhanu,
    How do i check if conversion is specified or not and another thing is it is not just one field we have 6 date fields and all of them are showing 7 characters in PSA after loading where as flat file have 8 characters in all of the 6 date fields.
    In PSA I checked the error message it shows 2 error messages:
    First Error message:
    The value '1# ' from field /BIC/ZACTDAYS is not convertible into the    
    DDIC data type QUAN of the InfoObject in data record 7 . The field      
    content could not be transferred into the communication structure       
    format.                                                                               
    em Response                                                                               
    The data to be loaded has a data error or field /BIC/ZACTDAYS in the    
    transfer structure is not mapped to a suitable InfoObject.                                                                               
    The conversion of the transfer structure to the communication structure 
    was terminated. The processing of data records with errors was continued
    in accordance with the settings in error handling for the InfoPackage   
    (tab page: Update Parameters).                                                                               
    Check the data conformity of your data source in the source system.                                                                               
    On the 'Assignment IOBJ - Field' tab page in the transfer rules, check    
    the InfoObject-field assignment for transfer-structure field              
    /BIC/ZACTDAYS .                                                                               
    If the data was temporarily saved in the PSA, you can also check the      
    received data for consistency in the PSA.                                                                               
    If the data is available and is correct in the source system and in the   
    PSA, you can activate debugging in the transfer rules by using update     
    simulation on the Detail   tab page in the monitor. This enables you to   
    perform an error analysis for the transfer rules. You need experience of  
    ABAP to do this.                                                                               
    2nd Error Message:
    Diagnosis                                                                               
    The transfer program attempted to execute an invalid arithmetic          
        operation.                                                                               
    System Response                                                                               
    Processing was terminated, and indicated as incorrect. You can carry out 
        the update again any time you choose.                                                                               
    Procedure                                                                               
    1.  Check that the data is accurate.                                                                               
    -   In particular, check that the data types, and lengths of the     
                transferred data, agree with the definitions of the InfoSource,  
                and that fixed values and routines defined in the transfer rules                                                                               
    -   The error may have arisen because you didn't specify the      
             existing headers.                                                                               
    2.  If necessary, correct the data error and start processing again.  
    Thanks
    MK

  • Regd. Error in Adapter module for IDOCxml to flat file conversion

    Hi Champs!
    I am using PI 7.1 and i am trying out the blog for direct conversion of IDOC xml to flat file and vice versa.
    /people/william.li/blog/2009/04/01/how-to-use-user-module-for-conversion-of-idoc-messages-between-flat-and-xml-formats
    has anyone tried this in Pi7.1
    i am getting the following error:
    Message processing failed.
    Cause: com.sap.engine.services.jndi.persistent.exceptions.NameNotFoundException:
    Path to object does not exist at SAP_XI_IDOC, the whole lookup
    name is localejbs/SAP_XI_IDOC/IDOCFlatToXmlConvertor.
    I have done all the setting required .. as per the blog.Is it srictly for PI7.11?
    Regards,
    Abhi

    Hi Abhi,
    This feature is a part of PI 7.11  new features , as it is mentioned in the begining of the Blog itself.
    So i feel , yes its strictly PI7.11 feature which may not be available in PI7.1 version which you are using currently.
    There may be workarounds...lets c  fellow SDNer's replys:)...
    Regards,
    Srinivas

  • XML to Flat File design issue

    Hi,
    A newbie to SSIS but was able to create an SSIS package which extracts XML data from one of the SQL server columns using "Execute SQL Task" and passes that to a for each loop container which contains an XML task for apply transform to each input
    xml and append it to a flat file(additonally using a script task within for each container).
    All good so far but now I want to apply conditional splitting to "Execute SQL Task" in the above and additionally pipe it to the exact similar additional process(For each container doing xml task as described above) for a different kind of
    flat file.
    If I alter the design to use the data flow approach and apply OOTB conditional split, I run into not knowing how to connect and execute more than one  foreach container and embedded XML and script task (dataflow to control flow connection)
    It is easy to put everything in a sequence container and repeat the Execute SQL Task . But to me that is very inefficient.
    Any creative ideas or pointers to some internet content which tells me how can this be done most efficiently.
    Hope my question makes sense and let me know if you need more clarification.
    Thanks in advance.
    SM

    As I understand what you're asking for is a way to create conditional branches to do different typeof processing for each subset of data. For this you can do like below 
    1. Add set of tasks to additional processing for each type of flat file and link them to Execute sql task.
    2. Make the precedence constraint option as Expression And Constraint for each of them. Make constraint as OnSuccess and expression based on your condition. You may need to create SSIS variables to capture value of fields to be used in expression manipulation.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • JCA flat file read issue

    Hi,
    We needs to read a flat file and transform it to destination xml format. Then send it to destination file location.
    Steps we have done:
    1. Create JCA adapter and configure the flat file schema
    2. Create proxy based on the jca
    3. create transformation for the source to target schema (this has no namespace)
    4. Create BS for sending the output
    Everything workins as expected when testing from OSB test console. But then the file is placed in the source folder, the output xml has the namespace xmlns:soap-env="http://schemas.xmlsoap.org/soap/envelope/" to the root node.
    e.g,
    <Root xmlns:soap-env="http://schemas.xmlsoap.org/soap/envelope/" >
    <Child>
    <Child1/>
    <Child2/>
    <Child3/>
    </Child>
    </Root>
    But expected output is
    <Root>
    <Child>
    <Child1/>
    <Child2/>
    <Child3/>
    </Child>
    </Root>
    We tried converting the xml to string then repalcing the xmlns:soap-env="http://schemas.xmlsoap.org/soap/envelope/" value with balnk.
    Also we tried with hadcorded xml using assign action instead of transformation. Even the harded xml also having the namespace to the root node.
    But the end system is failing due to this namespace value.
    Help me to resolve the issue.
    Thanks,
    Vinoth
    Edited by: Vinoth on Jun 8, 2011 10:12 PM

    Ideally your endsystem should not fail if you specify any number of namespace identifiers in the XML unless you are using them in elements or attributes within the XML. They should try to resolve this on their side.
    But to see whats going on in OSB, can you please paste log the $body variable in the before the publish action and paste the content here. Or send the sbconfig of the Proxy and business to me on my email mentioned in the profile if possible.

  • Flat File Conversion Error: Unable to initialize communication channel

    Dear experts.
    I have a file to flat file scenario. I just wanted to test this first because it's the first time. I followed this blog: /people/arpit.seth/blog/2005/06/02/file-receiver-with-content-conversion
    I get this message: Channel has not been correctly initialized and cannot process messages. I looked around on the forum and in other blogs, and it seems something is wrong with my Content Conversion Parameters. Could you please advise?
    The message is a small customer test message:
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:MT_TEST xmlns:ns0="urn://xi/csr/customer">
      <Customer_File>
        <Customer>
          <Name>xxx</Name>
          <Last>xxxxxxxx</Last>
        </Customer>
      </Customer_File>
    </ns0:MT_TEST>
    The parameters I am using are:
    Recordset structure: Customer_File,Customer
    Customer.addHeaderLine: 0
    Customer.fieldSeparator: #
    Customer_File.fieldSeparator: 'nl'
    Customer.endSeparator: 'nl'
    The mapping I am using is one-to-one.
    This is not working for me...
    Furthermore, I would like to set the Name and Last tag to set field lengths. Is this possible for these tags??
    Thanks in advance.
    *Will reward points for helpful answers

    hi
    PLz try with the following steps:
    chk the occurence of customer.
    if the customer occurence is (0-------unbounded) then
    ur recordset structure is  Customerfile,Customer*
    may helful.
    THanks
    MANas

  • Flat File Load Issue - Cannot convert character sets for one or more charac

    We have recently upgraded our production BW system to SPS 17 (BW SP19)
    and we have issues loading flat files (contain Chinese, Japanese and
    Korean characters) which worked fine before the upgrade. The Asian
    languages appear as invalid characters (garbled) as we can see in PSA. The Character Set
    Settings was previously set as Default Setting and it worked fine until
    the upgrade abd we referred to note 1130965 and with the code page 1100
    the load went through (without the Cannot convert character sets for one or more characters error) however the Asian language characters will appear
    as invalid characters. We tried all the code pages suggested in the note e.g.
    4102, 4103 etc on the info packages but it did not work. Note that the
    flat files are encoded in UTF-8.

    I checked lower case option for all IO
    When i checked the PSA failed log no of records processed is "0 of 0" my question is with out processing single record system is througing this error message.
    When i use same file loading from local workstation no error message
    I am thinking when  I FTP the file from AS/400 to BI Application Server(Linex) some invalid characters are adding but how can we track that invalid char?
    Gurus please share your thoughts on this I will assign full points.
    Thanks,

  • Flat File Load- issue

    Hi All,
    I am facing a problem while doing a deta load through Flat File; it's an .csv file.
    Problem arises whne I have some records like this £1,910.00 ; it's creating a separate column.
    I am having lot's of such amount field.
    is there any breakthrough for this issue ?? Your ans would be highly appriciated.
    Regards,
    Kironmoy Banerjee

    Hi,
    As satish has mentioned it is better to maintain two seperate columns as below
    1.0currency     
        GBP     
    2.0amount
        1910
    (GBP )Indicates British Pound.
    When ur maintaining amount as 1,910.00 it will go for next column because in info package  Extraction tabe Data Separator will be , (Comma).
    Regards
    Surendra
    Edited by: vaade surendra on May 7, 2010 11:37 AM
    Edited by: vaade surendra on May 7, 2010 11:37 AM
    Edited by: vaade surendra on May 7, 2010 11:38 AM
    Edited by: vaade surendra on May 7, 2010 11:41 AM

  • Oracle API for Extended Analytics Flat File extract issue

    I have modified the Extended Analytics SDK application as such:
    HFMCONSTANTSLib.EA_EXTRACT_TYPE_FLAGS.EA_EXTRACT_TYPE_FLATFILE
    instead of
    HFMCONSTANTSLib.EA_EXTRACT_TYPE_FLAGS.EA_EXTRACT_TYPE_STANDARD
    I am trying to figure out where the Flat file extract is placed once the Extract is complete. I have verified that I am connecting to HFM and the log file indicates all the steps in the application have completed successfully.
    Where does the FLATFILE get saved/output when using the API?
    Ultimate goal is to create a flat file through an automated process which can be picked up by a third party application.
    thanks for your help,
    Chris

    Never mind. I found the location on the server.

Maybe you are looking for

  • ODI File Datastore

    Hi All, I have a confusion regarding physical and logical length in ODI file datastore. I have a fixed width file where a col c2 had datatype as string(30). I defined that column in datastore as string>Physical length 30 >Logical length 30 My interfa

  • MS. Exchange Server 2007 Integration

    I want to integrate MS. Exchange Server 2007 with my WebCenter portal application. I read that I can do this with WebCenter Interaction. I have following questions:- 1. Is the license of WebCenter Interaction is convered under the license of WebCente

  • Ad blocking in Gmail using Pith Helmet

    How do i configure Pith Helmet to block the right hand side ads in Gmail.I've tried everything i can think of but it just doesn't work.

  • Cant add anything to desk top since I changed to Firefox!!

    When I clicked on the blue screen before I would get a big list of things to do, including add site to main menu or main screen. (I Forget the exact wording, )now cant do that anymore -all I can do is bookmark and that makes me go thru a whole long l

  • Viewing photos on Picasa Web after uploading-some are blank.Why?

    Viewing photos on Picasa Web after uploading-some are blank.Why Having trouble viewing photos in Picasa Web albums. When I upload them from my Picasa program to the web, some do not get loaded. Lots of blanks between photos. Help from Picasa told me