Flat file  load - Cube

Hi Experts,
I have loaded data to cube through Flat file in BI7. The request is in green at manage option and reportable but the transferred and added records are 0. Please help me out.
Regards,
Ravi

Did you run delta dtp from PSA to Cube ? which was already loaded I guess check in cube
If the request from PSA already loaded to cube and next time when you run delta dtp again it will fetch 0/0 records...
Are you sure you have loaded PSA with new request which was not yet loaded to cube?
Edited by: Srinivas on Aug 6, 2010 5:07 PM

Similar Messages

  • Data loading from flat file to cube using bw3.5

    Hi Experts,
                       Kindly give  me the detailed steps with screens  about Data loading from flat file to cube using bw3.5
           ...............Please

    Hi ,
    Procedure
    You are in the Data Warehousing Workbench in the DataSource tree.
           1.      Select the application components in which you want to create the DataSource and choose Create DataSource.
           2.      On the next screen, enter a technical name for the DataSource, select the type of DataSource and choose Copy.
    The DataSource maintenance screen appears.
           3.      Go to the General tab page.
                                a.      Enter descriptions for the DataSource (short, medium, long).
                                b.      As required, specify whether the DataSource builds an initial non-cumulative and can return duplicate data records within a request.
                                c.      Specify whether you want to generate the PSA for the DataSource in the character format. If the PSA is not typed it is not generated in a typed structure but is generated with character-like fields of type CHAR only.
    Use this option if conversion during loading causes problems, for example, because there is no appropriate conversion routine, or if the source cannot guarantee that data is loaded with the correct data type.
    In this case, after you have activated the DataSource you can load data into the PSA and correct it there.
           4.      Go to the Extraction tab page.
                                a.      Define the delta process for the DataSource.
                                b.      Specify whether you want the DataSource to support direct access to data.
                                c.      Real-time data acquisition is not supported for data transfer from files.
                                d.      Select the adapter for the data transfer. You can load text files or binary files from your local work station or from the application server.
    Text-type files only contain characters that can be displayed and read as text. CSV and ASCII files are examples of text files. For CSV files you have to specify a character that separates the individual field values. In BI, you have to specify this separator character and an escape character which specifies this character as a component of the value if required. After specifying these characters, you have to use them in the file. ASCII files contain data in a specified length. The defined field length in the file must be the same as the assigned field in BI.
    Binary files contain data in the form of Bytes. A file of this type can contain any type of Byte value, including Bytes that cannot be displayed or read as text. In this case, the field values in the file have to be the same as the internal format of the assigned field in BI.
    Choose Properties if you want to display the general adapter properties.
                                e.      Select the path to the file that you want to load or enter the name of the file directly, for example C:/Daten/US/Kosten97.csv.
    You can also create a routine that determines the name of your file. If you do not create a routine to determine the name of the file, the system reads the file name directly from the File Name field.
                                  f.      Depending on the adapter and the file to be loaded, make further settings.
    ■       For binary files:
    Specify the character record settings for the data that you want to transfer.
    ■       Text-type files:
    Specify how many rows in your file are header rows and can therefore be ignored when the data is transferred.
    Specify the character record settings for the data that you want to transfer.
    For ASCII files:
    If you are loading data from an ASCII file, the data is requested with a fixed data record length.
    For CSV files:
    If you are loading data from an Excel CSV file, specify the data separator and the escape character.
    Specify the separator that your file uses to divide the fields in the Data Separator field.
    If the data separator character is a part of the value, the file indicates this by enclosing the value in particular start and end characters. Enter these start and end characters in the Escape Charactersfield.
    You chose the; character as the data separator. However, your file contains the value 12;45 for a field. If you set u201C as the escape character, the value in the file must be u201C12;45u201D so that 12;45 is loaded into BI. The complete value that you want to transfer has to be enclosed by the escape characters.
    If the escape characters do not enclose the value but are used within the value, the system interprets the escape characters as a normal part of the value. If you have specified u201C as the escape character, the value 12u201D45 is transferred as 12u201D45 and 12u201D45u201D is transferred as 12u201D45u201D.
    In a text editor (for example, Notepad) check the data separator and the escape character currently being used in the file. These depend on the country version of the file you used.
    Note that if you do not specify an escape character, the space character is interpreted as the escape character. We recommend that you use a different character as the escape character.
    If you select the Hex indicator, you can specify the data separator and the escape character in hexadecimal format. When you enter a character for the data separator and the escape character, these are displayed as hexadecimal code after the entries have been checked. A two character entry for a data separator or an escape sign is always interpreted as a hexadecimal entry.
                                g.      Make the settings for the number format (thousand separator and character used to represent a decimal point), as required.
                                h.      Make the settings for currency conversion, as required.
                                  i.      Make any further settings that are dependent on your selection, as required.
           5.      Go to the Proposal tab page.
    This tab page is only relevant for CSV files. For files in different formats, define the field list on the Fields tab page.
    Here you create a proposal for the field list of the DataSource based on the sample data from your CSV file.
                                a.      Specify the number of data records that you want to load and choose Upload Sample Data.
    The data is displayed in the upper area of the tab page in the format of your file.
    The system displays the proposal for the field list in the lower area of the tab page.
                                b.      In the table of proposed fields, use Copy to Field List to select the fields you want to copy to the field list of the DataSource. All fields are selected by default.
           6.      Go to the Fields tab page.
    Here you edit the fields that you transferred to the field list of the DataSource from the Proposal tab page. If you did not transfer the field list from a proposal, you can define the fields of the DataSource here.
                                a.      To define a field, choose Insert Row and specify a field name.
                                b.      Under Transfer, specify the decision-relevant DataSource fields that you want to be available for extraction and transferred to BI.
                                c.      Instead of generating a proposal for the field list, you can enter InfoObjects to define the fields of the DataSource. Under Template InfoObject, specify InfoObjects for the fields in BI. This allows you to transfer the technical properties of the InfoObjects into the DataSource field.
    Entering InfoObjects here does not equate to assigning them to DataSource fields. Assignments are made in the transformation. When you define the transformation, the system proposes the InfoObjects you entered here as InfoObjects that you might want to assign to a field.
                                d.      Change the data type of the field if required.
                                e.      Specify the key fields of the DataSource.
    These fields are generated as a secondary index in the PSA. This is important in ensuring good performance for data transfer process selections, in particular with semantic grouping.
                                  f.      Specify whether lowercase is supported.
                                g.      Specify whether the source provides the data in the internal or external format.
                                h.      If you choose the external format, ensure that the output length of the field (external length) is correct. Change the entries, as required.
                                  i.      If required, specify a conversion routine that converts data from an external format into an internal format.
                                  j.      Select the fields that you want to be able to set selection criteria for when scheduling a data request using an InfoPackage. Data for this type of field is transferred in accordance with the selection criteria specified in the InfoPackage.
                                k.      Choose the selection options (such as EQ, BT) that you want to be available for selection in the InfoPackage.
                                  l.      Under Field Type, specify whether the data to be selected is language-dependent or time-dependent, as required.
           7.      Check, save and activate the DataSource.
           8.      Go to the Preview tab page.
    If you select Read Preview Data, the number of data records you specified in your field selection is displayed in a preview.
    This function allows you to check whether the data formats and data are correct.
    For More Info:  http://help.sap.com/saphelp_nw70/helpdata/EN/43/01ed2fe3811a77e10000000a422035/content.htm

  • What are the settings for datasource and infopackage for flat file loading

    hI
    Im trying to load the data from flat file to DSO . can anyone tel me what are the settings for datasource and infopackage for flat file loading .
    pls let me know
    regards
    kumar

    Loading of transaction data in BI 7.0:step by step guide on how to load data from a flatfile into the BI 7 system
    Uploading of Transaction data
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( Transaction data )
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to create ODS( Data store object ) or Cube.
    • Specify name fro the ODS or cube and click create
    • From the template window select the required characteristics and key figures and drag and drop it into the DATA FIELD and KEY FIELDS
    • Click Activate.
    • Right click on ODS or Cube and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.
    4. Monitor
    Right Click data targets and select manage and in contents tab select contents to view the loaded data. There are two tables in ODS new table and active table to load data from new table to active table you have to activate after selecting the loaded data . Alternatively monitor icon can be used.
    Loading of master data in BI 7.0:
    For Uploading of master data in BI 7.0
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( master data attributes, text, hierarchies)
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to select Insert Characteristics as info provider
    • Select required info object ( Ex : Employee ID)
    • Under that info object select attributes
    • Right click on attributes and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.

  • Flat File loading Initialize with out Data transfer is disabled in BI 7.0

    Hi experts,
              When loading through flat file in BI 7.0 for Info Package Level Initialization Delta Process with data Transfer is coming by default,but when i want to select Initialization Delta Process without Data transfer is disabled. (in the creation of Data Source (flat file) in the Extraction Tab Delta Process is changed to FIL1 Delta Data (Delta Images).
    please provide me Solution.
    regards
    Subba reddy.

    Hi Shubha,
    For flat file load please go throught he following link:
    http://help.sap.com/saphelp_nw70/helpdata/EN/43/03450525ee517be10000000a1553f6/frameset.htm
    This will help.
    Regards,
    Mahesh

  • Flat-File Loading problem

    Hi Friends,
    I am struggling with flat-file loading problem. I am trying to load a .csv file into data target. I took all pre-cautions while loading data. I look into preview and simulate the data. Everything is ok, but when i schedule the data, i found 0 records in the monitor. The following is the STATUS message of the above problem:
       No data available
    Diagnosis
    The data request was a full update.
    In this case, the corresponding table in the source system does not
    contain any data.
    System response
    Info IDoc received with status 8.
    Procedure
    Check the data basis in the source system.
             Can anybody help me what is the problem and procdure to resolve it?
    Regards,
    Mahesh

    Hi Eugene,
    Thanks for the quick reply. The following screen-shot tells you the messages of detail tab;
    OVER ALL STATUS MISSING WITH MESSAGES OR WARNINGS
    REQUEST: MISSING MESSAGES
    EXTRACTION
    EVERYTHING IS OK
    DATA REQUEST RECEIVED
    NO DATA AVAILABLE DATA ELECTION ENDED.
    PROCESSING
    NO DATA
               The above message was shown in details tab. Pls guide me to locate the problem.
    Regards,
    Mahesh

  • Transformation not generating-Flat file loading

    Hello guys, I hope you can help me with this little confusion I have on BI7 Flat file loading.
    I got a File (CSV) on my workstation. I am trying to load Master Data. Here is the example of my file and issues:
    Lets say, I have CSV file named "CarModel.CSV" on my PC.
    This excel file has 10 Records and No atributes for this Field.
    So the records should show like this and it is showing correctly in PSA level, DS level.
    A
    B
    C
    D
    E
    F
    My goal is to load Flat file data to InfoObject (inserted in Infoprovider)
    I created Source system, DS, all thats stuffs.
    I am now on Display DS screen under Proposal Tab. I am putting 10 Records to show and hitting Load example data....it works fine by showing all 10 records. However in the bottom part of this screen, what should show as a Field ?In my case , First it was showing the First Record ("A")...i didnt think it was correct but i prceeded anyways. Transformation could not be generated.
    I tried by deleting "1" from the field "No. of header rows to be ignored" and its the same result ..No transformations.. I mean i know its very simple to load this data but i am not sure what i am doing wrong..My question is:
    1) What should show under Field/propasal tab in my case?am i supposed to create this Field?
    2) Should the system to propse the field which is from my flat file header ?in my case i dont have any header..should i have include header in my csv file like "car model"? i am confused please give me some info thanks
    dk

    Hi,
    In filed tab, u have to enter ur infoobject names in an order...and press enter it'll automatically give the description and its other factors...
    i guess u shuld have some header in the sense eg: customer,cust ID like this..this only u have to enter as fields...in proposal tab...try that
    rgds,

  • Best practise around handling time dependency for flat file loads

    Hi folks,
    This is a fairly common situation - handling time dependency for flat file loads. Please can anyone share their experience around handling this. One common approach is to handle the time validity changes within the flat file where it is easily changeable by the user but then again is prone to input errors by the user. Another would be to handle this via a DSO. Possibly, also have this data entered directly in BI using IP planning layouts. There is a IP planning function that allows for loading flat file data but then again, it only works without the time dependency factor.
    It would be great to hear thoughts or if anyone can point to a best practise document for such a scenario.
    Thanks.

    Bump!

  • Unit Code, Commercial Code - Flat File loading vs BPS Loading

    Hi SDN Community,
    Recently, we have moved our flat file loading to be performed by BPS interfaces.
    During flat file loads, the Commercial code of all units of measures, are verified.
    eg.DAY
    But when loaded by BPS, the UNIT code is verified.
    eg. TAG.
    In the display of the BPS upload, it displays DAY which is the commercial code.
    The only thing, is the customer is forced to use TAG in the BPS upload files.
    They wish to create another record in the transaction
    CUNI to be
    Unit code = DAY
    Commercial code = DAY.
    However, i have found that we cannot allocate the same commercial code to another unit code.
    Is this a design constraint, or a process error that i am doing.
    Thank you.
    Simon

    04.01.2010 - 02:22:53 CET - Reply by SAP     
    Dear customer,
    As i do not fully understand how this works, but the base table T006A
    has 2 entries, 1 for english, 1 for German, should it not be that the
    English EN entry field be working rather that the German DE, hence DAY
    should be used? Can you please confirm my understanding on this?
    >>> If you check my attachment "SE16.xls", you can see that it's for
    language >>English<<, and the internal format is TAG while the external
    format is DAY.
    - Are there any plans to modify the BPS functionality in newer SAP BW
    versions to allow the customer to indicate the same UOM as per flat
    file laods? Or is an upgraded Support Pack that allows this?
    - If not, would it be possible to make any customer enhancements to
    allow this to take place depending on customer requirements.
    >>> There is no plan at this moment to change this BPS functionality
    in newer SAP BW versions or support packages.
    We really recommend you to use internal format TAG in the upload file
    in BPS which I think should be acceptable and feasible to you. All
    other ways trying to use external format is risky and we cannot assure
    you that it will work well as it's not SAP standard function. (I think
    it's not worth the risk as the standard function which requires
    internal format should not be too unacceptable)
    Thanks for your understanding.
    I also don't think my BC-SRV-ASF-UOM colleague would be able to help
    you a lot regarding this.
    Best Regards,
    Patricia Yang
    Support Consultant - Netweaver BW
    Global Support Center China
    SAP Active Global Support

  • One Time Flat File load in a Cube that extracts from ECC

    Hi,
    I have a cube that is already extracting data from ECC on a daily basis. I am asked to load some historical data into the cube from their legacy system. My question is, how do i prepare that file from the legacy system to match the fields in the Cube? What are the technicalities that i need to look at?
    For Example:
    0MATERIAL in the Cube:  Material Number in the Flat File
    ZAMOUNT in the Cube: Cost of Product in the Flat File
    How do I arrange to match these?
    Thanks

    Hi
    Create DS and enter the required objects
    Preview the Data
    Create Transformations (Map DS fields to Targets Objects - direct assignemnt)
    Create Infopack and extract data upto PSA
    Create DTP and run - it will extract data from PSA to Target
    No of Heade rows to be ignored : 1
    File type : CSV
    Data seperator : ;
    Escape Sign: "
    Thanks,

  • FLAT FILE LOADING INTO TRANSACTION CUBE

    Hi Guru's,
                 I'm working on bps to load flat file into trasaction cube.
    when i'm actiavting init Functin module, i'm getting the following error.
    'IS NO COMPONENT EXIST WITH THE NAME PARNM'.
    THE ERROR IS COMING  WHEN THE FM IS READING FILE NAME.
    tHANKS IN ADVANCE.

    Hi Vijay,
    Just check Function Module Parameters:
    IMPORT: I_AREA        TYPE     UPC_Y_AREA
                  I_PLEVEL        TYPE     UPC_Y_PLEVEL
                  I_PACKAGE   TYPE     UPC_Y_PACKAGE
                  I_METHOD        TYPE     UPC_Y_METHOD
                  I_PARAM        TYPE     UPC_Y_PARAM
                  IT_EXITP        TYPE     UPF_YT_EXITP
                  ITO_CHASEL TYPE     UPC_YTO_CHASEL
                  ITO_CHA        TYPE     UPC_YTO_CHA
                  ITO_KYF        TYPE     UPC_YTO_KYF
    EXPORT: ETO_CHAS  TYPE     ANY TABLE
                   ET_MESG       TYPE     UPC_YT_MESG

  • Error while creating process chain for flat file loading

    Hi All,
    I had Created a process Chain to load Transactiion data(Full load) form flat file which is in my computer.
    start>Load>DTP>DeleteIndex>DTP loading CUBE--> Create Index
    but the system is throwing an error as "An upload from the client workstation in the background is not possible."
    I dont know why this error is coming?
    Can some one help me
    Regards
    Mamta

    Hi Mamta,
    Basically if you want to load the DS through FF using process chain, the FF has to be placed in Application server. We cant load the FF when it is located in the client local workstation(FF On your PC).
    So better you remove the Infopackage step from the PC. Load the IP manually. Once it is completed you can start the process chain with the following steps:
    start>DTP>DeleteIndex>DTP loading CUBE> Create Index
    Hope it is clear & helpful!
    Regards,
    Pavan

  • Regarding flat file load to DSO...

    Hi Experts,
    I am trying to load transaction data from flat file to DSO in BI7.0 that comprises two characteristics and one keyfigure. The characteristics type is char and keyfigure's type is int.  Till PSA, data is fine.  In transformation, I have just done one-to-one mapping with source field and target field, but while executing DTP, the error popped up as follows,
    Data package 1:Errors during processing
    Extraction completed, processing termination
    Please help. I tried various combinations, but not able to schedule.  Valuable answers would be awarded remarkably.
    Thanks,

    Hi there...
    Here are some ideas to Ur problem
    Assuming that u defined info objects correctly, length and types... and u mapped transformations correctly
    And ur data in flat file is absolutely correct...
    Of course very common but suggestible is plz check the data lengths in flat file  and defined length of the info object.. If yes
    Come to Data source and check the settings for the flat file ,separater and escape sign and rows to ignored..and check the data preview...
    Plz compare the data in PSA and Flat file... plz do this very carefully... If yes...
    Go to data source and at the FIELDS tab and check whether u assigned info objects correctly...if yes...
    Check the conversion routines in the same screen of the tab FIELDS.. if for key figure u see PER16.. Remove it and try to load to PSA and do load the cube..
    Hope this can give u an idea...
    Regrds...KP

  • Issue with flat file loading timing out

    Hello
    I have a scenario, where I am loading a flat file with 65k records into a cube. The problem is, in the loading process, I have to look up the 0Material table which has a million records.
    I do have an internal table in the program, where I select a subset of the Material table ie around 20k to 30k records. But my extraction process takes more than 1 1/2 hrs and is failing (timed out).
    How can i address this issue? I tried building indexes on the Material table and its not helping.
    Thanks,
    Srinivas.

    Unfortunately, this is BW 3.5, so there is no END routine option here. And I tried both .csv and notepad file methods, and both are creating problems for us.
    This is the total code, do you guyz see any potential issues:
    Start Routine (main code)
    refresh i_oldmats.
    refresh ZI_MATL.
    data: wa_datapak type transfer_structure.
    loop at datapak into wa_datapak. (** I collect all the old material numbers from my flat file into an internal table i_oldmats**)
       i_oldmats-/BIC/ZZOLDMATL = wa_datapak-/BIC/ZZOLDMATL.
       collect i_oldmats.
    endloop.
    sort i_oldmats.
      SELECT /BIC/ZZOLDMATL MATERIAL (** ZI_MATL only has recs. where old materials exist, this gets about 300k records out of 1M**)
             FROM /BI0/PMATERIAL
             INTO ZI_MATL FOR ALL
             ENTRIES in i_oldmats WHERE
              /BIC/ZZOLDMATL = i_oldmats-/BIC/ZZOLDMATL .
                collect    ZI_MATL.
      Endselect.
      Sort ZI_MATL.
    Transfer rule routine (main code)
    IF TRAN_STRUCTURE-MATERIAL = 'NA'.
        READ TABLE ZI_MATL INTO ZW_MATL
        WITH KEY /BIC/ZZOLDMATL = TRAN_STRUCTURE-/BIC/ZZOLDMATL
        BINARY SEARCH.
        IF SY-SUBRC = 0.
          RESULT = ZW_MATL-MATERIAL.
        ENDIF.
      ELSE.
        RESULT = TRAN_STRUCTURE-MATERIAL.
      ENDIF.
    Regards,
    Srinivas.

  • BI7 Flat file loading

    Hello
    I am trying to load flat file from my workstation to BI. I have a transaction data, in which 4 chars IO filed and 3 KFs field. I am little confusion about loading in BI7 since Infosource , TR, UR not required.
    I am creating flat file source system and DS as well. I have info obj catalog for both Chars and kfs. do i have to load master data first ?before loading transactional? i have data in excel format all master data in different file , and also Transaction data (include MD and kfs) in one file. can i just create DS for transaction data and then load or i have to create DS for each Chars IO to load master data first?
    when i tried to create DS fro trans data , for some reason under proposal tab it only populates one field, FIELD1, i assume i ll have to go to Field tabe and manuall select all fields by adding new rows..
    Could someone give me some idea on how to do the loading to infocube? i also tried to create Transformation , it gave me an error..
    when we create DS for transactional data using Flat file , wouldnt system supposed to map all the DS fields to the cube using Transformation, but its not happening to mine, please help...
    thanks
    DK

    Hi........
    Before loading transaction data you have to load Master data...
    1)  For each master data....ie IO.....you have to create one datasource.........
    You have to create datasource using tcode : RSO2.......
    The General tab is same for all datasource........
    In the Extraction tab...
    Real Time>> RDA is not supported......
    Adapter >> Load text type file from work station
    Name of the file >> select your file
    Header rows to be ignored >> 1
    Character set settings >> Default
    Data format >> Seperated with seperator...
    Data seperator>>;
    Escape sign >> #
    Number format >>Direct entry
    Thousands seperator>>.
    Decimal point seperator>>,
    SAP suggests the use of header row ,as the system can use this header information to help define fields on the file.
    The Proposal tab reads the header row and proposes field names and type..........here you have to choose load example data and take a look at the proposed list.Make sure that all fields are flagged Copy to Fieldlist......This proposed field is derived from the header line and contents of your flatfile....
    These proposed fiels and types can be changed in Fields tab........At this point you need to check if the field types and lenth are correct......In some cases proposal will not be accurate,and you will need to adjust the field types and length on the fields tab.....
    You file should be saved as .csv format.........
    2) Then you have to create an infopackage and load data till PSA......
    3) Then create transformation between the datasource(PSA) and the target(IO).........
    4)Then create DTP and load data..........
    After this in the same way you have to create datasource for transaction data...........the infopackage..........transformation ........DTP........process is same...........
    Check this link :
    http://help.sap.com/saphelp_nw70/helpdata/EN/43/03450525ee517be10000000a1553f6/frameset.htm
    Hope this helps you.......
    Regards,
    Debjani.........
    Edited by: Debjani  Mukherjee on Sep 21, 2008 9:44 AM

  • No Data Available (Flat File Loading Problem))

    Hi Friends,
    I loaded the data from FlatFile(Application Server) I am able to see in the Preview
    but the data is not laoding into the Cube. Request status is green but with zero records.
    Following is the message
    No data available
    Diagnosis
    The data request was a full update.
    In this case, the corresponding table in the source system does not
    contain any data.
    System response
    Info IDoc received with status 8.
    Procedure
    Check the data basis in the source system
    Can anybody help
    Regards,
    CV

    Hi chakri,
    first u have cto check  ur flat file in the Presentation server OR Application server
    2. Ur flat file fields sequence same In the Trasfer structure .
    U should maintain the same sequence in Transferstructure  what u have maintained in the flat file.
    3. when ur loading date from flatfile source System u have to use FULLUPDATE  mode .
    check the above .
    Thanks,
    kiran

Maybe you are looking for