Flat file missing in datasource

Hello iam loading the flatfile record into the infocube in this process while iam creating datasource after naming the data source i need to select flat file option for source file field  but flat file option is missing.how to solve this problem.i think basis guys need to configure this .what do u experts suggest

Create a new source system for flat file and then try if you don't have one. Yes - Basis guy basically do it or if you have permission, you also can create it.
To create source system:
RSA1>Source Systems>right click and click Create
Choose radiobutton Symble with PC, File System(Manual Metadata,....)
Click Transfer and give the logical system name and source system name
click continue and activate it.
Now you can use this source system to load flat file.
Regards,
Ashok

Similar Messages

  • Error while loading CSV Flat file

    Hi,
    I am getting the following errors while loading a flat file into the DataSource:
    1. Error 'Unit CS is not created in language EN...' at conversion exit CONVERSION_EXIT_CUNIT_INPUT (field UNIT record 1, value CS)
    2. Error 'The argument '9,018.00' cannot be interpreted as anumber' on assignment field /BIC/ZTVSA record 1 value 9,018.00
    What can I do to correct this?
    Thanks

    I am taking the following (Headers) fields in this order in a CSV file :
    0LOGSYS     ,0MATERIAL,0UNIT,0BASE_UOM,0COMP_CODE,0PLANT,0FISCPER,0FISCVARNT,
    0CURRENCY,0PRICE_TYPE,ZTVSAKF,ZVOTVSKF,0PRICE_CTRL,ZPRICE_STKF,0PRICE_BASE,
    0VAL_CLASS,0CURTYPE,0VALUATION
    I have added all these fields in the DataSource as well. T-Code CUNIT has CS as a Unit. 
    I deleted and recreated the DataSource and am getting this error now:
    Error 'Unit CS is not created in language EN...' at conversion exit CONVERSION_EXIT_CUNIT_INPUT (field UNIT record 1, value CS)
    Error 'The argument '1,398.00' cannot be interpreted asa number' on assignment field /BIC/ZTVSAKF record 1 value 1,398.00
    Error 'The argument '8,10.00' cannot be interpreted as anumber' on assignment field /BIC/ZTVSKF record 1 value 8,10.00
    Error 'Unit 0 is not created in language EN...' at conversion exit CONVERSION_EXIT_CUNIT_INPUT (field BASE_UOM record 1, value )
    Error 'Unit CS is not created in language EN...' at conversion exit CONVERSION_EXIT_CUNIT_INPUT (field UNIT record 1, value CS)
    Error 'The argument '5,50.00' cannot be interpreted as anumber' on assignment field /BIC/ZTVSAKF record 1 value 5,50.00
    Does it matter what order the columns are in?

  • Flat file data upload error

    Hello,
    I am trying to load a flat file into a datasource. When I execute the infopackage, i am getting a lot of errors for diff fields in the datasource like this:
    Error 'An exception with the type CX_SY_CONVERSION_NO_NUM' at conversion exit RSDS_CONVERT_NUMBER (field DATE_OF_BIRTH record 1, value 12/)
    What does this mean and what do I have to do to load the data properly into the PSA?
    Thanks.

    Hi,
      This may be due to following,
    1. I guess you are using some Date data in your file. First check whether in your Datasource u have defined that field with correct data type( either Calday, calmonth...).
    2. If you have declared correctly, check whther you are giving the data for Date in your file in correct format: Ex: 012007, 022008....
    3. Check the sequence of the fields in your file and datasource, whether its correct or not.
    rgrds,
    v.sen

  • Loading data from flat file to ODS

    Hi Experts,
    I have a developed a report, in which Flat file is the datasource and data target as DSO. 
    Here I have one requriment to Increatment the counter value in TVARVC table after every load compleate.
    Can this be done thru Process chain?  I belive this can be done thru process chain by creating variable.
    If this can be done thru Process chains can  any experts send me sample examples code.
    Thanks In Advance.
    Sreedhar.M

    Hi sridhar,
         We have a similar scenario in which we will load data from flat file to datasource the file name (x_20081010.dat) of such format  being fetched from the low value of the TVARVC table ... once successfully executed we had program to increment the value of low which increments by one day adn the next day the file with that will be placed in the application server...
        U can create a abap program varaint and attribute for that and u can achieve that.... this is sample code which searches for the string and after some coding i.e once ur file has been loaded successful code u can write similar code to increment or counter values which u are asking for...
    search s_tvarvc-low for '20'.
            if sy-subrc ne 0.
              sy-msgty = 'E'.
              sy-msgid = 'ZWM1'.
              sy-msgno = '000'.
              sy-msgv1 = 'Unable to find DATE to increment'.
              raise unable_increment.
            else.
              w_date = s_tvarvc-low+sy-fdpos(08).
              w_nextdate = w_date + 1.
              replace w_date with w_nextdate into s_tvarvc-low.
              modify tvarvc from s_tvarvc.
            endif.
    Regards
    vamsi

  • Field in Flat File DataSource

    I created a Flat File DataSource.
    In the DataSource, I enetered InfoObject 0OUTAGTARGT in the Fields tab ->InfoObject Template column. I want this InfoObject to be mapped
    into field KTWRT (Target Value) in R/3 table EKKO.
    However, when I enter the the InfoObject 0OUTAGTARGT and hit Enter key, system prompts me to copy the properties. I choose Copy.
    And then it says:
    Reference field 0DER_CURR was added automatically.
    Now I have ORDER_CURR entered in the Field column, with InfoObject Template 0ORDER_CURR.
    In the Field column, I replaced the field ORDER_CURR with WAERS (Currency field from R/3).
    When I save it, I hit the following:
    Missing currency field / unit field ORDER_CURR
    System expect the field ORDER_CURR to be entered in the Field column. But in my source file, there is no field call
    ORDER_CURR. It is actually WAERS.
    How can I overcome this?

    Maintain 0ORDER_CURR as it is needed by 0OUTAGTARGT.
    Next, simply map WAERS to the ORDER_CURR field.
    But it doesnt allow me to change. In DataSource ->Fields tab, I try to replaced the field ORDER_CURR with WAERS. When I click on Activate, I
    hit the following:
    Missing currency field / unit field ORDER_CURR
    Please help.

  • Flat File DataSource doesn't load properly.

    Hi all,
    we have a data file, which is tab delimited, saved as CSV and text versions. i have created a flat file datasource, mapped all fields.
    I can load CSV version fine, but when use it on text version, it doesn't like it and can only return column one with the rest of the column blanked in the preview tab.
    Setting:
    In the "General" tab, everything is unchecked.
    In Extraction tab, Adapter - "Load Text-Type File from Local Workstation"
                             Character settings - "Default setting"
                             Data format - "Separated with Separator"
                             Data separator - tab (00BB Hex)
                             Escape Sign - CR (00B6 Hex)
                             Number format - Direct entry
                                                     Thousand Separator - ","
                                                     Decimal point - "."
    i have tried different setting, change number format to "User Master Records".
    With the text version, the result in Preview can only display first column and Proposal preview can only shows one column. There is no error returned from the text version.
    The CSV version can display all data and they are aligned as well.
    what do i miss? is the text file not saved correctly? i can see all tabs and CR in the text version. it seems to be the tab is not detected properly or something.
    Regards

    Hi,
    Have a look at the below link:
    Loading Text Files(.TXT) into BW with HEX option

  • Extracting data directly from a datasource to a flat file

    My requirements are to extract data from a SAP standard profit center accounting data extractor 0EC_PCA_1 directly to a flat file. Can this be done in a standard way or is custom ABAP development required ?
    I have tried the RSA3 based extractor, but as the result is displayed in packets, I would have to export the content of each extracted packet to separate flat file which is not very feasible.
    Br,
    HR

    Hi,
    1st option: create a dummy ods as well as a dummy infosource and dummy update rules. Assign the datasource to the dummy infosource. In the start routine of the transfer rules take each datapackage and download it to a file. Delete the datapackage. So you need almost no coding and you can also use a possible delta mechanism.
    2nd option: create your own abap and call the extraction module. Take care about the way it works like init mode, building packages... Possibly you need to find a way to remodel the delta capability.
    kind regards
    Siggi

  • Error while loading Flat file to the table (ORA-00936: missing expression)

    lat file Hi Gurus
    Receiving the following error while trying to load of flat file to the database :
    ODI-1228: Task test_file_load (Integration) fails on the target ORACLE connection DEMO_STAGE.
    Caused By: java.sql.SQLSyntaxErrorException: ORA-00936: missing expression
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:457)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:405)
         at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:889)
         at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:476)
         at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:204)
         at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:540)
         at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:217)
         at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1079)
         at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1466)
         at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3752)
         at oracle.jdbc.driver.OraclePreparedStatement.execute(OraclePreparedStatement.java:3937)
         at oracle.jdbc.driver.OraclePreparedStatementWrapper.execute(OraclePreparedStatementWrapper.java:1535)
         at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:163)
         at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)
         at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)
         at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:537)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
         at java.lang.Thread.run(Thread.java:662)
    The file which I have tried to load is : SRC_SALES_PERSON and teh table structure is
    CREATE table "TRG_SALES_PERSON"(
    "SALES_PERSON_ID" NUMBER(8,0) NOT NULL,
    "FIRST_NAME" VARCHAR2(80),
    "LAST_NAME" VARCHAR2(80),
    "DATE_HIRED" VARCHAR2(80),
    "DATE_UPDATED" DATE NOT NULL)
    Knowledge module used are
    LKM File to SQL
    IKM SQL Incremental Update
    We rae using ODI 11g R2 ...
    Thanks and reallty appreciate any help in thsi regard.

    HI there,
    I am facing the same issue while loading data from SRC_SALES_PERSON(flat file) to TRG_CUSTOMER.
    I dont see any errors in the steps however the data is not laoded finally. Here are the sql commands
    **On source**
    select     ID     C11_ID,
         LASTNAME     C9_LASTNAME
    from      TABLE
    /*$$SNPS_START_KEYSNP$CRDWG_TABLESNP$CRTABLE_NAME=SRC_SALES_PERSONSNP$CRLOAD_FILE=D:\Pratima\Softwares\ODI\ofm_odi_companion_generic_11.1.1.5.1_disk1_1of1[1]\demo\oracledi-demo\oracledi\demo\file/SRC_SALES_PERSON.txtSNP$CRFILE_FORMAT=FSNP$CRFILE_SEP_FIELD=0x0009SNP$CRFILE_SEP_LINE=0x000D0x000ASNP$CRFILE_FIRST_ROW=0SNP$CRFILE_ENC_FIELD=SNP$CRFILE_DEC_SEP=SNP$CRSNP$CRDWG_COLSNP$CRCOL_NAME=IDSNP$CRTYPE_NAME=STRINGSNP$CRORDER=1SNP$CRLINE_OFFSET=1SNP$CRLENGTH=11SNP$CRPRECISION=11SNP$CRACTION_ON_ERROR=0SNP$CRSNP$CRDWG_COLSNP$CRCOL_NAME=FIRSTNAMESNP$CRTYPE_NAME=STRINGSNP$CRLINE_OFFSET=12SNP$CRLENGTH=50SNP$CRPRECISION=50SNP$CRSNP$CRDWG_COLSNP$CRCOL_NAME=LASTNAMESNP$CRTYPE_NAME=STRINGSNP$CRLINE_OFFSET=62SNP$CRLENGTH=50SNP$CRPRECISION=50SNP$CRSNP$CRDWG_COLSNP$CRCOL_NAME=DATE1SNP$CRTYPE_NAME=STRINGSNP$CRLINE_OFFSET=112SNP$CRLENGTH=20SNP$CRPRECISION=20SNP$CR$$SNPS_END_KEY*/
    On Target
    insert into STAGING.C$_0TRG_CUSTOMER
         C11_ID,
         C9_LASTNAME
    values
         :C11_ID,
         :C9_LASTNAME
    The actual code at source fails however the step is in green.
    Thanks in Advance,
    Pratima

  • Flat File Hierarchy Datasources

    We are creating several flat file hierarchy datasources. The datasources will be maintained in an external database.
    I am familiar with the creation of flat file datasources for master and transactional data, but don't have good documentation for how the external flat file hierarchy is to be formatted.
    We have reviewed the 2004 Netweaver Document "How to Download Hierarchies in BW" and have used the program referenced there, Z_SAP_HIERARCHY_DOWNLOAD, to download our hierarchies to then test the upload.
    However, this has not worked.
    Does anyone have the record layout that the flat file must be in to upload a simple 2-level hierarchy (for example groupings of 0vendor)?
    Thanks, Doug

    I think the URL you mention is http://help.sap.com/saphelp_nw04/helpdata/en/fa/e92637c2cbf357e10000009b38f936/content.htm.
    This contains the record layout instructions as follows.
      5.      Maintaining the hierarchy:
    Choose Hierarchy Maintenance, and specify a technical name and a description of the hierarchy.
    PSA Transfer Method: You have the option here to set the Remove Leaf Value and Node InfoObjects indicator. As a result, characteristic values are not transferred into the hierarchy fields NODENAME, LEAFFROM and LEAFTO as is normally the case, but in their own transfer structure fields.  This option allows you to load characteristic values having a length greater than 32 characters.
    Characteristic values with a length > 32 can be loaded into the PSA, but they cannot be updated in characteristics that have a length >32.
    The node names for pure text nodes remain restricted to 32 characters in the hierarchy (0HIER_NODE characteristic).
    The system automatically generates a table with the following hierarchy format (for sorted hierarchies without removed leaf values and node InfoObjects):
    Description
    Field Name
    Length
    Type
    Node ID
    NODEID
    8
    NUMC
    InfoObject name
    INFOOBJECT
    30
    CHAR
    Node name
    NODENAME
    32
    CHAR
    Catalog ID
    LINK
    1
    CHAR
    Parent node
    PARENTID
    8
    NUMC
    First subnode
    CHILDID
    8
    NUMC
    Next adjacent node
    NEXTID
    8
    NUMC
    Language key
    LANGU
    1
    CHAR
    Description - short
    TXTSH
    20
    CHAR
    Description - medium
    TXTMD
    40
    CHAR
    Description- long
    TXTLG
    60
    CHAR

  • What are the settings for datasource and infopackage for flat file loading

    hI
    Im trying to load the data from flat file to DSO . can anyone tel me what are the settings for datasource and infopackage for flat file loading .
    pls let me know
    regards
    kumar

    Loading of transaction data in BI 7.0:step by step guide on how to load data from a flatfile into the BI 7 system
    Uploading of Transaction data
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( Transaction data )
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to create ODS( Data store object ) or Cube.
    • Specify name fro the ODS or cube and click create
    • From the template window select the required characteristics and key figures and drag and drop it into the DATA FIELD and KEY FIELDS
    • Click Activate.
    • Right click on ODS or Cube and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.
    4. Monitor
    Right Click data targets and select manage and in contents tab select contents to view the loaded data. There are two tables in ODS new table and active table to load data from new table to active table you have to activate after selecting the loaded data . Alternatively monitor icon can be used.
    Loading of master data in BI 7.0:
    For Uploading of master data in BI 7.0
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( master data attributes, text, hierarchies)
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to select Insert Characteristics as info provider
    • Select required info object ( Ex : Employee ID)
    • Under that info object select attributes
    • Right click on attributes and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.

  • FLAT FILE Datasource SERIOUS Problem

    Hi
    I want to read a header information in csv file to fill the FLAT FILE DATASOURCE structure.
    I cannot read it by default as I have to IGNORE THE HEADER ROW otherwise it wont come up correctly along with otherdetail  data that is below header row during extraction of CSV file.
    How can i do that or How can I CALL A Function module or any other to fetch that header row ?
    PLEASE HELP URGENTLY.
    Thanks

    you cannot read the header row only. But you can do a start routine within the transfer rules. There you can do whatever you like.
    Perhaps you might have some troubles getting your data that far. Your header row must comply to the rules defined in your datasource like size or data type. So you might need to change some definitions of your datasource.

  • Flat File Import, Ignore Missing Columns?

    The text files I'm importing always contain a fixed set of columns for example total number of full set of columns is 60, or a subset of those columns (some csv contain 40 columns, some contain 30 or 20 or any other number.) .  I would like to import
    these csv based on the column header inside the each csv, if it is a subset of full column set then the missing columns can be ignored with null value.
    At the moment in SQL 2012, if I import a subset of columns in the csv file, the data doesn't import...I assume because the actual file doesn't include every column defined in the flat file source object? 
    Is it possible to accomplish this without dynamically selecting the columns,  or using script component?
    Thanks for help.
    Sea Cloud

    If the columns coming are dynamic, then you might have to first get it into staging table with a single column reading entire row contents onto that single column and parse out the column information using string parsing logic as below
    http://visakhm.blogspot.in/2010/02/parsing-delimited-string.html
    This will help you to understand what columns are present based on which you can do the insertion to your table.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Infopackage Routine in flat file datasource

    Hi,
    There is a requirements to load the data from multiple flat files in folder into BI which source data is coming from other system. Each flat file has it own naming convention.
    There are around more the 200 flat file (CSV.file) in folder,data source is flat file datasource.
    Could any one help me in above issue.
    Regards
    RAM

    I assume they'll be using FTP to send the files... You have the layout defined and naming convention as well...
    One option would be to create a user and folder in the BW server for the FTP to use it and put all files in the same place... From here you can define different Infopackages to read those files at specific times during the day or whatever schedule you guys agree...
    A thing with the naming convention... Try to avoid using timestamps on the names and keep the names to be always the same... That way you just overwrite the existing files, the space usage won't grow much and you can make the loads automatic with the same file names... 
    Just some ideas...

  • Error in DTP creation for a flat file datasource

    Hi SDN friends,
    Please if somebody can give me a help with this, it will be very appreciate.
    I have a DTP to load attributes master data into an infoobject. The master data comes from a flat file (data source flat file) .
    The extraction mode is full, when I try to modify any selection in the "Update" tab, a message "Enter valid value" shows, but it doesn't give me any hint about which is the wrong value that must be checked.
    The message ID is 00 002 and the and the text is the following
    Enter a valid value
    Message no. 00002
    Procedure
    Display the allowed values with F4 and correct your entry.
    I think it may be something related to the flatfile datasource. Are there any special considerations to create a DTP to upload master data from a flat file?
    Thanks so much for your attention and best regards,
    Karim

    Hi Karim,
    I suspect that you might be entering values that are not in correct format as defined in BI DTP. Eg: date it could be DD.MM.YYYY in BI you might be giving MM.DD.YYYY similary you may be entering the wrong format for master data you are giving it in selections.
    Hope it helps...
    Best Regards,
    Madhu

  • Mapping Quantity keyfigure in BI 7- Flat file datasource

    Hi Guys,
    I have a problem in creating transformations for Quantity key figure for flat file datasource. In my csv file quantity is just  a number. i created a datasource and Infopackage and loaded data into PSA without any problem.
    In cube i have quantity key figure of QUAN type with 0UNIT unit of measure. So when I try to map Quantity in datasource with quantity InfoObject, 0UNIT is empty.  So I am unable to activate transformations.
    Can somebody suggest me how to proceed with htis problem? Should I define Quantity key figure as just number?
    Thank you in advance

    Hi Sneralla,
      you can take one  more field on you flatfile.. as units(ex: KG  or CM.. like that) beside of the  quanity filed..  and now load it into PSA and  now create transformations...  and now map this  units filed to  newly added field...
       you can try this option.. becasue.. when you are loading  Quantity   field data  from flat file means we need to mention the units also..
    regards
    @jay

Maybe you are looking for

  • Problem with Graphics and JPanel...

    I would like to draw a rectangle on a JPanel. Here is my code: JPanel panel = new JPanel(); Graphics g = panel.getGraphics(); g.drawRect(20,20,20,20); When I run my program, I have a NullPointerException... Do you any suggestions to fix the problem??

  • Print tile problem

    I have a one page pdf document I want to print enlarged on several pages. I'm running Acrobat Pro on a mac OS 10.6.7 with an epson workForce 840. I follow instructions (see below) and selected tiling options but the print out is reduced rather than e

  • How do I get the iMac to display thumbnails and not an icon of jpeg photos?

    Hi folks, I just purchased my first Mac, an Imac. I was browsing through a folder, using Finder, containing photos files from my latest trip. Under Windows, I used to tell Windows Explorer to show thumbnails of the photos. How do I get the iMac and F

  • [IDOC] Creating own IDOC

    Hey, playing with IDOCs, I would like to create an own one with customized structure. Found this thread: Creating own IDOC with segments In we31 I have to create a segment. - Do you uses segment names longer than 7 characters? I would not   communica

  • Why are lyrics not viewable in itunes in the same way as on the ipod?

    Why are lyrics not viewable in itunes in the same way as on the ipod? In itunes all you get is "Get Info" then "Lyrics" and that screen then locks itunes until you close it. Inflexible, annoying and surely unnecessary?