Data load from Legacy system to BW Server through BAPI

Requirements: We have different kind of legacy systems and SAP BW server. We want to load all legacy system data into SAP BW server using BAPI. Before loading we have to validate all data. If there are bad data, data missing we have to let the legacy system user/ operator knows to fix the data into their system with detail explanation. When it is fixed, we have to load the data again.
Load Scenario:  We have two options to load data from legacy systems to BW Server.
1.     We need to load data directly from legacy system to BW Server using BAPI program.
2.     Legacy Systems data would be in workstations or flash drive as .txt (one line separated by comma) or .csv file. Need to load from .txt /.csv file to BW Server using BAPI program.
What we want in the BAPI program code?
It will Read / Retrieve data from text / csv file and will put into the Internal table. Internal table structure would be based on BAPI InfoObject structure.
Call BAPI InfoObject function module ‘BAPI_IOBJ_CREATE’ to create InfoObject, include all necessary / default components, do the error check, load the data and return the status.
Could some one help me with the sample code please? I am new in ABAP / BAPI coding.
Is there any other better idea to load data from legacy system to BW server? BTW we are using BW 3.10. Is there any better option with BI 7.0 to resolve the issue? I appreciate your help.

my answers:
1. this is a scendario for a data push into SAP BW. You can only use SOAP-Based Transfer of Data.
http://help.sap.com/saphelp_nw04/helpdata/en/fd/8012403dbedd5fe10000000a155106/frameset.htm
(here for BW 3.5, but you'll find similar for 7.0)
In this scenario you'll have an RFC dinamically created for every Infosource you need to transfer data.
2. You can make a chain for each data load, so you can call the RFC "RSPC_API_CHAIN_START" to start the chain from external.
the second solution is more simply and available on every release.
Regards,
Sergio

Similar Messages

  • Dataload from legacy system to BW Server

    Requirements: We have different kind of legacy systems and SAP BW server. We want to load all legacy system data into SAP BW server using BAPI. Before loading we have to validate all data. If there are bad data, data missing we have to let the legacy system user/ operator knows to fix the data into their system with detail explanation. When it is fixed, we have to load the data again.
    Load Scenario: We have two options to load data from legacy systems to BW Server.
    1. We need to load data directly from legacy system to BW Server using BAPI program.
    2. Legacy Systems data would be in workstations or flash drive as .txt (one line separated by comma) or .csv file. Need to load from .txt /.csv file to BW Server using BAPI program.
    What we want in the BAPI program code?
    It will Read / Retrieve data from text / csv file and will put into the Internal table. Internal table structure would be based on BAPI InfoObject structure.
    Call BAPI InfoObject function module ‘BAPI_IOBJ_CREATE’ to create InfoObject, include all necessary / default components, do the error check, load the data and return the status.
    Could some one help me with the sample code please? I am new in ABAP / BAPI coding.
    Is there any other better idea to load data from legacy system to BW server? BTW we are using BW 3.10. Is there any better option with BI 7.0 to resolve the issue? I appreciate your help.
    Regards,
    Mau

    my answers:
    1. this is a scendario for a data push into SAP BW. You can only use SOAP-Based Transfer of Data.
    http://help.sap.com/saphelp_nw04/helpdata/en/fd/8012403dbedd5fe10000000a155106/frameset.htm
    (here for BW 3.5, but you'll find similar for 7.0)
    In this scenario you'll have an RFC dinamically created for every Infosource you need to transfer data.
    2. You can make a chain for each data load, so you can call the RFC "RSPC_API_CHAIN_START" to start the chain from external.
    the second solution is more simply and available on every release.
    Regards,
    Sergio

  • Data Migration from Legacy system to ECC systems via ETL through SAP PI

    Hi All,
    I wanted to know if we can migrate the data from Legacy systems to ECC systems via PI.
    What I understand is there is ETL tool is used to extract the data from legacy system, and what client is looking to load that data via PI?
    Can we do that ? I am concerned because this will involve mass data?
    If I have to use PI , I see option of  PROXY / IDOC /FILE as receiver in ECC system (take the data from ETL)?
    Can somebody has done this earlier , please share the approach.
    Thanks,
    Pushkar Patel

    Hi ,
    I require few details from you.
    1. What ETL tool you are using, If Informatica, it already have PowerConnect to connect to SAP. So you can create source and Target Structure and also you can use RFC's to send data to R/3. Else, for other ETL tools, can you prepare RFC's or any other way to send data to R/3. let me know the tool.
    2. Does R/3 contains the master data tables? If yes, then try to use LSMW for Mass upload of data to tables.
    If your client don't want to use either of these options please elaborate, what is the case.
    Regards
    Aashish Sinha

  • Data migration from legacy system to SAP

    hi all,
      could you please post some docs on migration from legacy system to SAP system., also the problems faced while migration and best practices in data migration
    regards
    sridharan

    Hi ,
    I require few details from you.
    1. What ETL tool you are using, If Informatica, it already have PowerConnect to connect to SAP. So you can create source and Target Structure and also you can use RFC's to send data to R/3. Else, for other ETL tools, can you prepare RFC's or any other way to send data to R/3. let me know the tool.
    2. Does R/3 contains the master data tables? If yes, then try to use LSMW for Mass upload of data to tables.
    If your client don't want to use either of these options please elaborate, what is the case.
    Regards
    Aashish Sinha

  • Data Migration from Legacy System in IS-U

    Hi All,
    We are going to Implement a new IS-U project,  I have a problem with LSMW(Legacy System Migration Workbench ),  I have some Conversion Objects of IS-U and I need to know whether Data Migration is possible or not, please tell me how to find these Objects and how to know Data Migration is possible or not.
    Objects are like.,
    1. Accounts
    2. Actuate Reports
    3. Business Partner Relationships
    4. Active Campaigns/Campaign Content/Dispositions
    5. Connection Object
    6. Contacts
    7. Contracts
    8. Opportunities
    9. Payment Arrangement History
    10. Payments
    11. Premises
    12. Rate Changes
    13. Security Deposits
    these are few and there are some more..,
    Thanks in Advance,
    Sai.

    Hi Ram,
    Use Transaction Code EMIGALL. It will ask for company code. By default Company Code is SAP. If you entered with the SAP you will get all the objects.Then goto menu IS-U Migration-->User Handbook. It will give you details Idea.
    Also Check the following Procedure
    You can find detailed documentation of EMIGALL in SAP itself. Use Transaction EQ81 to display it. It provides all the concepts and procedures to work with EMIGALL.
    Here are some points about EMIGALL :
    1. It Migrates data Business Object wise
    2. It uses Direct Input Technique
    3. It has more than 100 objects of IS-U
    and the steps for implementation goes like this:
    1)You have to create a user specially for migration which will have all the authorizations related to migration workbench, BASIS and IS-U
    2)You have to create your own company in EMIGALL. There is a default company called SAP.
    3)Company SAP contains all the Business Objects
    4)You have to figure out what business objects u need and then u have to copy those business objects to ur company from Standard Company SAP
    5)Each objects contains more than one structure and each structure can contain more than one fields. The relation goes like this
    Object ---> Structure ---> Field
    6)You have to define field rules for each required field of the object. You have to mark "Not required" for fields u don't need
    7)After field rules for a given object is set u have to generate load report i.e. actual Direct Input Program which will migrate data. This program is generated on basis of field rules set by u.
    8)After the load report is generated u have to prepare an input file (import File) for migration. The import file should be according to structure provided by SAP and must be in binary format. SAP Provides the structure of file according to your configurations. You have to write ur own Data conversion program(in any language) for this task.
    9)You take import file as input and migrate the data using generated load program
    10)Finally u can check the Migration Statistics and Error Log
    Thanks and Regards,
    Jyotishankar Dutta
    Message was edited by:
            Jyotishankar Dutta
    Message was edited by:
            Jyotishankar Dutta

  • Data loading from source system takes long time.

    Hi,
         I am loading data from R/3 to BW. I am getting following message in the monitor.
    Request still running
    Diagnosis
    No errors could be found. The current process has probably not finished yet.
    System response
    The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
    and/or
    the maximum wait time for this request has not yet run out
    and/or
    the batch job in the source system has not yet ended.
    Current status
    in the source system
    Is there anything wrong with partner profile maintanance in the source system.
    Cheers
    Senthil

    Hi,
    I will suggest you to check a few places where you can see the status
    1) SM37 job log (In source system if load is from R/3 or in BW if its a datamart load) (give request name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
    Also see if there is any 'sysfail' for any datapacket in SM37.
    2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. (In source system if load is from R/3 or in BW if its a datamart load). See if its accessing/updating some tables or is not doing anything at all.
    3) RSMO see what is available in details tab. It may be in update rules.
    4) ST22 check if any short dump has occured.(In source system if load is from R/3 or in BW if its a datamart load)
    5) SM58 and BD87 for pending tRFCs and IDOCS.
    Once you identify you can rectify the error.
    If all the records are in PSA you can pull it from the PSA to target. Else you may have to pull it again from source infoprovider.
    If its running and if you are able to see it active in SM66 you can wait for some time to let it finish. You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
    If you feel its active and running you can verify by checking if the number of records has increased in the data tables.
    SM21 - System log can also be helpful.
    Thanks,
    JituK

  • Data Migration from Legacy System in CRM

    Hi All,
    We are going to Implement a new CRM project, I have a problem with LSMW(Legacy System Migration Workbench ), I have some Conversion Objects of CRM and I need to know whether Data Migration is possible or not, please tell me how to find these Objects and how to know Data Migration is possible or not.
    Objects are like.,
    1. Accounts
    2. Actuate Reports
    3. Active Campaigns/Campaign Content/Dispositions
    4. Contacts
    5. Contracts
    6. Opportunities
    7. Payment Arrangement History
    8. Payments
    9. Premises
    10. Rate Changes
    11. Security Deposits
    12. Business Partner Relationships
    13. Web self-service information
    14. Usage that has been used for quotes
    15. Tax history information, including AXCIS
    17. Service Requests
    18. Service Order History
    19. Security Deposits
    20. Reference Values
    21. Rate Changes
    Can anybody please tell any transaction code in CRM where can I find the Data migration for the above objects is possible or not.
    Thanks in Advance,
    Sai.

    Hello,
    for migration into CRM I would suggest the 'XIF-Interfaces'; these are interfaces provided in CRM for connection to external systems. I'm not sure if this will cover all objects mentioned above, but at least some of them should exist.
    Some information about existing interfaces can be found in the integration repository under http://ifr.sap.com/index.html; from the start screen first select 'Enter the repository' and then have a look under 'generic components' and 'SAP CRM'.
    Regards, Katja Ohliger

  • Date migration from Legacy system (excel file) to CRM 7.0

    Hi,
    Can any one guide me on the below:
    I am provided with excel files having prospect data with 63 fields and most of them are not the part of standard fields available.
    The requirement is to create the same data in CRM. I am unable to use ELM as most of the fields are not available. I suppose i need to get the fields created by technical team. However need some knowledge sharing on this. I think even BAPI approach would need fields to be created first.
    Should i revert back to the client asking for detailed information on the tabs or sections under which the missing fileds have to be created?
    Also guide me with alternate solution to map the data from legacy to CRM
    Please mark your solution on this
    thanks

    Agarwal,
    In your situation you need to have the additional attributes created in CRM, before you can do an upload.  Now if the new attributes are marketing attributes for the business partner, you can use ELM.  Otherwise if you still want to do ELM then you need to use the BADI.
    Your other option is use XIF adapter with LSMW to perform the upload, but the custom fields must appear in the bdoc structure on the CRM for them to appear.  Normally this won't be a problem if the CRM system has been extended via EEWB or AET.
    Take care,
    Stephen

  • Upload master data tables from legacy systems with XI on BW

    Hi developers,
    Can you help me  to reconstruct the steps necessary for upload master data, deriving by legacy tables, on BW system using XI?
    Thanks in advance

    Hi,
    see the following link https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/52398597-0601-0010-868a-fd95f8061d85
    Rgs
    Antonino

  • AR & AP Data migration from legacy

    Hi,
    1.What are the precaution to be taken for Data migration from legacy system?
    2.We intend to use a conversion ac ( Dr conv ac cr Vendor ) Is it Correct way of doing it ?
    3.How do i ensure that all my invoices uploaded correctlt ?
    4.What is the best Standard program to get it doen ?.
    5.Is LSMW is best to do it or shall I have to go for custom development?
    Any piece of yr advice is appreciated.
    Thanks, Chitra

    Hi Chitra,
    Following link are helpful in understanding LSMW:
    Note 105244 - LSMW 1.0: known problems, tips and tricks
    Re: LSMW and error handling
    /community [original link is broken]
    For Data Migration:
    Note 400257 - Data Migration and Upload Report
    Note 739860 - Data migration from a legacy system to SAP
    Hope this helps.
    Please assign points as way to say thanks

  • Load the data from Legacy system into CRM

    I'm new to CRM,I would like to know  What is the procedure used to Load the data (ex: Business partners data ,Organization data,Sales data, etc.) from Legacy system into CRM system?
    Can anyone  explain me and send  any documents,it  would be very helpful.
    I highly appreciate your help.
    Thanks,
    [email protected]

    Hi Rani,
    You can also use the ALE IDOC method to load data from legacy to CRM system if you have some middlware tool(E.g. XI) in between. We have implemented this in our project where we were getting flat files from legacy, XI was converting the same into IDOC and the IDOC was sent to CRM where we read the IDOC data, populate the standard BAPI's or function modules to create the data in CRM.
    You need to define an IDOC TYPE similar to the data structure on the legacy system. Maintain the ALE settings. Write the function module to validate and create data in the CRM system.
    E.G. To create business partner SAP has provided BAPI_BUPA_CREATE_FROM_DATA bapi.
    <b>Reward points if it helps.</b>

  • Uploading data from Legacy system

    Hi
    i am shorthly going to work on data migration project. I have never worked on a data migration project and want to know the answers for the following questions.
    1] which technique i must first consider in data upload ( both master data and transaction data )
    2] What steps are involved?
    3] I know we can use LSMW, BAPI's , BDC , Direct Input , IDOC etc but want to know which one choose and when?
    4] What care must be taken in each data migration
    5] is there is std list of programs or a good document on when to choose what?
    ~SR

    Hi,
    Data migration can be done by the follwoing methods:
    LSMW : for master data migartion or small data migartion
    BAPI : its just like a FM , thru which u can create master aswell as Transactional data.
    u have to import some paremetre as well as tables if required, and Bapi will RETURN the message saying so and so material/delievry had cretaed.
    BDC : in BDC we got 2 methods.
    CAll transaction and session.
    CAll transacton : using this u can uplaod transcational data as well as master data.
    the updation can be synchronus or asynchronus updates.
    Session: using this u can uplaod transcational data as well as master data.
    The updation in session is synchronus, so its time taking process.
    Most of the time we go for Call transcation In real situations.
    Dircet input method: its out dated.no longer it is being used.
    in all the above cases, data is alraedy existing in flat file and u r going to migrate the data in one shot but IDOC is used where the dynamical data migration is necessary.
    suppose say ones the material is craeted in legacy sytem , that particulat matreial shud be migarted to R/3.this can be handled thru IDOC.
    Idoc: master data as well as Trnascational data is migarted thru IDOC( matmas, orders).
    Idoc uses ALE technology in migartion of idoc.
    Revert back if any issues.
    Check these links.
    SAP Data Migration with LSMW
    http://www.sap-img.com/sap-data-migration.htm
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    LSMW
    No ABAP effort are required for the SAP data migration. However, effort are required to map the data into the structure according to the pre-determined format as specified by the pre-written ABAP upload program of the LSMW.
    The Legacy System Migration Workbench (LSMW) is a tool recommended by SAP that you can use to transfer data once only or periodically from legacy systems into an R/3 System.
    More and more medium-sized firms are implementing SAP solutions, and many of them have their legacy data in desktop programs. In this case, the data is exported in a format that can be read by PC spreadsheet systems. As a result, the data transfer is mere child's play: Simply enter the field names in the first line of the table, and the LSM Workbench's import routine automatically generates the input file for your conversion program.
    The LSM Workbench lets you check the data for migration against the current settings of your customizing. The check is performed after the data migration, but before the update in your database.
    So although it was designed for uploading of legacy data it is not restricted to this use.
    We use it for mass changes, i.e. uploading new/replacement data and it is great, but there are limits on its functionality, depending on the complexity of the transaction you are trying to replicate.
    The SAP transaction code is 'LSMW' for SAP version 4.6x.
    For those with the older SAP version (4.7 and below), the data migration programs might not have been pre-loaded.
    You can download the LSMW at no cost from SAPNet under Services, SAP Methodology and Tools, category Tools.
    If you are an existing SAP customer with an OSS ID, you can access the SAP Service Marketplace to download the LSWM for your Basis teams to install to your SAP system: http://service.sap.com/lsmw
    The LSM Workbench carries out the following tasks:
    Reads the transfer data from one or more files (for example, spreadsheets, sequential files etc.)
    Converts the data from the source format into the target format
    Note that with background processing, the input file must not be located in the presentation server. Access to presentation server files is only possible when you are working onlineUpload Condition Pricing
    LSMW STEPS
    Data Cleansing
    Data cleansing allows you to compare, include and merge redundant business partner master records (potential duplicates) in data cleansing cases. Following the data cleansing process you can remove data records from the system using archiving.
    Integration
    Before you can carry out data cleansing, you must determine the redundant data in your system and include it in data cleansing cases.
    You have the following options for duplicate recognition and creation of data cleansing cases.
    The Business Address Services (BAS) provide interfaces for integrating the relevant external software (search machines).
    User-defined programs
    Services of data providers, who check their data for possible duplicates.
    In the options described above, the data cleansing cases are created via the system and put into the data cleansing worklist for further processing.
    In individual cases you can find potential business partner duplicates in the hitlist of the business partner search and create a separate data cleansing case.
    After connecting the non-SAP software (search engines), the system starts duplicate recognition (Delta Scan) with every change to or new entry of a business partner for this specific individual record. You should do a full scan through the complete dataset at the beginning of the consolidation process using the non-SAP software, this registers all potential duplicates in data cleansing cases and makes them available or further processing.
    Within the application basis, the referencing objects, which are taken into account by the data cleansing, are limited to the sub-objects of the business partner. You have the possibility to make modification-free enhancements to the referencing objects.
    Prerequisites
    You must have determined the redundant data records in your system.
    To use the interface of the Business Address Services, make the following Customizing settings: SAP Implementation Guide -> Basis -> Basis Services -> Address Management -> Activate Duplicate Check Index Pool
    Make the following system settings in the IMG of the SAP Business Partner: Basic Settings -> Data Cleansing -> Maintain Number Range / Define Priorities / Activate Data Cleansing.
    Activities
    Register data cleansing cases via BAPIs or create them in the dialog for business partner maintenance from the worklist or the business partner search.
    Process the created data cleansing cases in a separate process step afterwards.
    BATCH DATA COMMUNICATION
    About Data Transfer In R/3 System
    When a company decides to implement the SAP R/3 to manage business-critical data, it usually does not start from a no-data situation. Normally, a SAP R/3 project comes into replace or complement existing application.
    In the process of replacing current applications and transferring application data, two situations might occur:
    • The first is when application data to be replaced is transferred at once, and only once.
    • The second situation is to transfer data periodically from external systems to SAP and vice versa.
    • There is a period of time when information has to be transferred from existing application, to SAP R/3, and often this process will be repetitive.
    The SAP system offers two primary methods for transferring data into SAP systems. From non-SAP systems or legacy system. These two methods are collectively called “batch input” or “batch data communication”.
    1. SESSION METHOD
    2. CALL TRANSACTION
    3. DIRECT INPUT
    Advantages offered by BATCH INPUT method:
    1. Can process large data volumes in batch.
    2. Can be planned and submitted in the background.
    3. No manual interaction is required when data is transferred.
    4. Data integrity is maintained as whatever data is transferred to the table is through transaction. Hence batch input data is submitted to all the checks and validations.
    To implement one of the supported data transfers, you must often write the program that exports the data from your non-SAP system. This program, known as a “data transfer” program must map the data from the external system into the data structure required by the SAP batch input program.
    The batch input program must build all of the input to execute the SAP transaction.
    Two main steps are required:
    • To build an internal table containing every screen and every field to be filled in during the execution of an SAP transaction.
    • To pass the table to SAP for processing.
    Prerequisite for Data Transfer Program
    Writing a Data Transfer Program involves following prerequisites:
    Analyzing data from local file
    Analyzing transaction
    Analyzing transaction involves following steps:
    • The transaction code, if you do not already know it.
    • Which fields require input i.e., mandatory.
    • Which fields can you allow to default to standard values.
    • The names, types, and lengths of the fields that are used by a transaction.
    • Screen number and Name of module pool program behind a particular transaction.
    To analyze a transaction::
    • Start the transaction by menu or by entering the transaction code in the command box.
    (You can determine the transaction name by choosing System – Status.)
    • Step through the transaction, entering the data will be required for processing your batch input data.
    • On each screen, note the program name and screen (dynpro) number.
    (dynpro = dyn + pro. Dyn = screen, pro = number)
    • Display these by choosing System – Status. The relevant fields are Program (dynpro) and Dynpro number. If pop-up windows occur during execution, you can get the program name and screen number by pressing F1 on any field or button on the screen.
    The technical info pop-up shows not only the field information but also the program and screen.
    • For each field, check box, and radio button on each screen, press F1 (help) and then choose Technical Info.
    Note the following information:
    - The field name for batch input, which you’ll find in its own box.
    The length and data type of the field. You can display this information by double clicking on the Data Element field.
    • Find out the identification code for each function (button or menu) that you must execute to process the batch-input data (or to go to new screen).
    Place the cursor on the button or menu entry while holding down the left mouse button. Then press F1.
    In the pop-up window that follows, choose Technical info and note the code that is shown in the Function field.
    You can also run any function that is assigned to a function key by way of the function key number. To display the list of available function keys, click on the right mouse button. Note the key number that is assigned to the functions you want to run.
    Once you have program name, screen number, field name (screen field name), you can start writing.
    DATA TRANSFER program.
    Declaring internal table
    First Integral Table similar to structure like local file.
    Declaring internal table like BDCDATA
    The data from internal table is not transferred directly to database table, it has to go through transaction. You need to pass data to particular screen and to particular screen-field. Data is passed to transaction in particular format, hence there is a need for batch input structure.
    The batch input structure stores the data that is to be entered into SAP system and the actions that are necessary to process the data. The batch input structure is used by all of the batch input methods. You can use the same structure for all types of batch input, regardless of whether you are creating a session in the batch input queue or using CALL TRANSACTION.
    This structure is BDCDATA, which can contain the batch input data for only a single run of a transaction. The typical processing loop in a program is as follows:
    • Create a BDCDATA structure
    • Write the structure out to a session or process it with CALL TRANSACTION USING; and then
    • Create a BDCDATA structure for the next transaction that is to be processed.
    Within a BDCDATA structure, organize the data of screens in a transaction. Each screen that is processed in the course of a transaction must be identified with a BDCDATA record. This record uses the Program, Dynpro, and Dynbegin fields of the structure.
    The screen identifier record is followed by a separate BDCDATA record for each value, to be entered into a field. These records use the FNAM and FVAL fields of the BDCDATA structure. Values to be entered in a field can be any of the following:
    • Data that is entered into screen fields.
    • Function codes that are entered into the command field. Such function codes execute functions in a transaction, such as Save or Enter.
    The BDCDATA structure contains the following fields:
    • PROGRAM: Name of module pool program associated with the screen. Set this field only for the first record for the screen.
    • DYNPRO: Screen Number. Set this field only in the first record for the screen.
    • DYNBEGIN: Indicates the first record for the screen. Set this field to X, only for the first record for the screen. (Reset to ‘ ‘ (blank) for all other records.)
    • FNAM: Field Name. The FNAM field is not case-sensitive.
    • FVAL: Value for the field named in FNAM. The FVAL field is case-sensitive. Values assigned to this field are always padded on the right, if they are less than 132 characters. Values must be in character format.
    Transferring data from local file to internal table
    Data is uploaded to internal table by UPLOAD of WS_UPLOAD function.
    Population of BDCDATA
    For each record of internal table, you need to populate Internal table, which is similar to BDCDATA structure.
    All these five initial steps are necessary for any type of BDC interface.
    DATA TRANSFER program can call SESSION METHOD or CALL TRANSACTION. The initial steps for both the methods are same.
    First step for both the methods is to upload the data to internal table. From Internal Table, the data is transferred to database table by two ways i.e., Session method and Call transaction.
    SESSION METHOD
    About Session method
    In this method you transfer data from internal table to database table through sessions.
    In this method, an ABAP/4 program reads the external data that is to be entered in the SAP System and stores the data in session. A session stores the actions that are required to enter your data using normal SAP transaction i.e., Data is transferred to session which in turn transfers data to database table.
    Session is intermediate step between internal table and database table. Data along with its action is stored in session i.e., data for screen fields, to which screen it is passed, the program name behind it, and how the next screen is processed.
    When the program has finished generating the session, you can run the session to execute the SAP transactions in it. You can either explicitly start and monitor a session or have the session run in the background processing system.
    Unless session is processed, the data is not transferred to database table.
    BDC_OPEN_GROUP
    You create the session through program by BDC_OPEN_GROUP function.
    Parameters to this function are:
    • User Name: User name
    • Group: Name of the session
    • Lock Date: The date on which you want to process the session.
    • Keep: This parameter is passed as ‘X’ when you want to retain session after
    processing it or ‘ ‘ to delete it after processing.
    BDC_INSERT
    This function creates the session & data is transferred to Session.
    Parameters to this function are:
    • Tcode: Transaction Name
    • Dynprotab: BDC Data
    BDC_CLOSE_GROUP
    This function closes the BDC Group. No Parameters.
    Some additional information for session processing
    When the session is generated using the KEEP option within the BDC_OPEN_GROUP, the system always keeps the sessions in the queue, whether it has been processed successfully or not.
    However, if the session is processed, you have to delete it manually. When session processing is completed successfully while KEEP option was not set, it will be removed automatically from the session queue. Log is not removed for that session.
    If the batch-input session is terminated with errors, then it appears in the list of INCORRECT session and it can be processed again. To correct incorrect session, you can analyze the session. The Analysis function allows to determine which screen and value has produced the error. If you find small errors in data, you can correct them interactively, otherwise you need to modify batch input program, which has generated the session or many times even the data file.
    CALL TRANSACTION
    About CALL TRANSACTION
    A technique similar to SESSION method, while batch input is a two-step procedure, Call Transaction does both steps online, one after the other. In this method, you call a transaction from your program by
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>
    Messages into <MSGTAB>.
    Parameter – 1 is transaction code.
    Parameter – 2 is name of BDCTAB table.
    Parameter – 3 here you are specifying mode in which you execute transaction
    A is all screen mode. All the screen of transaction are displayed.
    N is no screen mode. No screen is displayed when you execute the transaction.
    E is error screen. Only those screens are displayed wherein you have error record.
    Parameter – 4 here you are specifying update type by which database table is updated.
    S is for Synchronous update in which if you change data of one table then all the related Tables gets updated. And sy-subrc is returned i.e., sy-subrc is returned for once and all.
    A is for Asynchronous update. When you change data of one table, the sy-subrc is returned. And then updating of other affected tables takes place. So if system fails to update other tables, still sy-subrc returned is 0 (i.e., when first table gets updated).
    Parameter – 5 when you update database table, operation is either successful or unsuccessful or operation is successful with some warning. These messages are stored in internal table, which you specify along with MESSAGE statement. This internal table should be declared like BDCMSGCOLL, a structure available in ABAP/4. It contains the following fields:
    1. Tcode: Transaction code
    2. Dyname: Batch point module name
    3. Dynumb: Batch input Dyn number
    4. Msgtyp: Batch input message type (A/E/W/I/S)
    5. Msgspra: Batch input Lang, id of message
    6. Msgid: Message id
    7. MsgvN: Message variables (N = 1 - 4)
    For each entry, which is updated in database, table message is available in BDCMSGCOLL. As BDCMSGCOLL is structure, you need to declare a internal table which can contain multiple records (unlike structure).
    Steps for CALL TRANSACTION method
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. UPLOAD or WS_UPLOAD function to upload the data from local file to itab. (Considering file is local file)
    4. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>.
    Refresh BDCTAB.
    Endloop.
    (To populate BDCTAB, You need to transfer each and every field)
    The major differences between Session method and Call transaction are as follows:
    SESSION METHOD CALL TRANSACTION
    1. Data is not updated in database table unless Session is processed. Immediate updation in database table.
    2. No sy-subrc is returned. Sy-subrc is returned.
    3. Error log is created for error records. Errors need to be handled explicitly
    4. Updation in database table is always synchronous Updation in database table can be synchronous Or Asynchronous.
    Error Handling in CALL TRANSACTION
    When Session Method updates the records in database table, error records are stored in the log file. In Call transaction there is no such log file available and error record is lost unless handled. Usually you need to give report of all the error records i.e., records which are not inserted or updated in the database table. This can be done by the following method:
    Steps for the error handling in CALL TRANSACTION
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. Internal table BDCMSG like BDCMSGCOLL
    4. Internal table similar to Ist internal table
    (Third and fourth steps are for error handling)
    5. UPLOAD or WS_UPLOAD function to upload the data from the local file to itab. (Considering file is local file)
    6. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tr.code> using <Bdctab>
    Mode <A/N/E>
    Update <S/A>
    Messages <BDCMSG>.
    Perform check.
    Refresh BDCTAB.
    Endloop.
    7 Form check.
    IF sy-subrc 0. (Call transaction returns the sy-subrc if updating is not successful).
    Call function Format_message.
    (This function is called to store the message given by system and to display it along with record)
    Append itab2.
    Display the record and message.
    DIRECT INPUT
    About Direct Input
    In contrast to batch input, this technique does not create sessions, but stores the data directly. It does not simulate the online transaction. To enter the data into the corresponding database tables directly, the system calls a number of function modules that execute any necessary checks. In case of errors, the direct input technique provides a restart mechanism. However, to be able to activate the restart mechanism, direct input programs must be executed in the background only. Direct input checks the data thoroughly and then updates the database directly.
    You can start a Direct Input program in two ways;
    Start the program directly
    This is the quickest way to see if the program works with your flat file. This option is possible with all direct input programs. If the program ends abnormally, you will not have any logs telling you what has or has not been posted. To minimize the chance of this happening, always use the check file option for the first run with your flat file. This allows you to detect format errors before transfer.
    Starting the program via the DI administration transaction
    This transaction restarts the processing, if the data transfer program aborts. Since DI document are immediately posted into the SAP D/B, the restart option prevents the duplicate document posting that occurs during a program restart (i.e., without adjusting your flat file).
    Direct input is usually done for standard data like material master, FI accounting document, SD sales order and Classification for which SAP has provided standard programs.
    First time you work with the Direct Input administration program, you will need to do some preparation before you can transfer data:
    - Create variant
    Define job
    Start job
    Restart job
    Common batch input errors
    - The batch input BDCDATA structure tries to assign values to fields which do not exist in the current transaction screen.
    The screen in the BDCDATA structure does not match the right sequence, or an intermediate screen is missing.
    On exceptional occasions, the logic flow of batch input session does not exactly match that of manual online processing. Testing the sessions online can discover by this.
    The BDCDATA structure contains fields, which are longer than the actual definition.
    Authorization problems.
    RECORDING A BATCH INPUT
    A B recording allows you to record a R/3 transaction and generate a program that contains all screens and field information in the required BDC-DATA format.
    You can either use SHDB transaction for recording or
    EDIT&#61614; BATCH INPUT &#61614; SERVICES &#61614;SYSTEM
    And from here click recording.
    Enter name for the recording.
    (Dates are optional)
    Click recording.
    Enter transaction code.
    Enter.
    Click Save button.
    You finally come to a screen where, you have all the information for each screen including BDC_OKCODE.
    • Click Get Transaction.
    • Return to BI.
    • Click overview.
    • Position the cursor on the just recorded entry and click generate program.
    • Enter program name.
    • Click enter
    The program is generated for the particular transaction.
    BACKGROUND PROCESSING
    Need for Background processing
    When a large volume of data is involved, usually all batch inputs are done in background.
    The R/3 system includes functions that allow users to work non-interactively or offline. The background processing systems handle these functions.
    Non-interactively means that instead of executing the ABAP/4 programs and waiting for an answer, user can submit those programs for execution at a more convenient planned time.
    There are several reasons to submit programs for background execution.
    • The maximum time allowed for online execution should not exceed 300 seconds. User gets TIMEOUT error and an aborted transaction, if time for execution exceeds 300 seconds. To avoid these types of error, you can submit jobs for background processing.
    • You can use the system while your program is executing.
    This does not mean that interactive or online work is not useful. Both type of processing have their own purposes. Online work is the most common one entering business data, displaying information, printing small reports, managing the system and so on. Background jobs are mainly used for the following tasks; to process large amount of data, to execute periodic jobs without human intervention, to run program at a more convenient, planned time other than during normal working hours i.e., Nights or weekends.
    The transaction for background processing is SM36.
    Or
    Define jobs&#61614; Jobs &#61614; Administration &#61614;Tools
    Or
    &#61614;System Jobs&#61614;services
    Components of the background jobs
    A job in Background processing is a series of steps that can be scheduled and step is a program for background processing.
    • Job name. Define the name of assigned to the job. It identifies the job. You can specify up to 32 characters for the name.
    • Job class. Indicates the type of background processing priority assigned to the job.
    The job class determines the priority of a job. The background system admits three types of job classes: A B & C, which correspond to job priority.
    • Job steps. Parameters to be passed for this screen are as follows:
    Program name.
    Variant if it is report program
    Start criteria for the job: Option available for this are as follows:
    Immediate - allows you to start a job immediately.
    Date/Time - allows you to start a job at a specific name.
    After job - you can start a job after a particular job.
    After event - allows you to start a job after a particular event.
    At operation mode - allows you to start a job when the system switches to a particular operation mode.
    Defining Background jobs
    It is two step process: Firstly, you define the job and then release it.
    When users define a job and save it, they are actually scheduling the report i.e., specifying the job components, the steps, the start time.
    When users schedule program for background processing, they are instructing the system to execute an ABAP/4 report or an external program in the background. Scheduled jobs are not executed until they are released. When jobs are released, they are sent for execution to the background processing system at the specified start time. Both scheduling and releasing of jobs require authorizations.
    HANDLING OF POP UP SCREEN IN BDC
    Many times in transaction pop up screen appears and for this screen you don’t pass any record but some indication to system telling it to proceed further. For example: The following screen
    To handle such screen, system has provided a variable called BDC_CURSOR. You pass this variable to BDCDATA and process the screen.
    Usually such screen appears in many transactions, in this case you are just passing information, that YES you want to save the information, that means YES should be clicked. So you are transferring this information to BDCDATA i.e., field name of YES which is usually SPOT_OPTION. Instead of BDC_OKCODE, you are passing BDC_CURSOR.
    BDC_CURSOR is also used to place cursor on particular field.
    AN EXAMPLE WITH SESSION METHOD
    Following program demonstrates how data is passed from flat file to SAP transaction and further to database table by using SESSION method.
    The transaction is TFBA (to change customer).
    A simple transaction where you are entering customer number on first screen and on next screen data is displayed for the particular customer number. Field, which we are changing here, are name and city. When you click on save, the changed record gets saved.
    Prerequisite to write this BDC interface as indicated earlier is:
    1. To find screen number
    2. To find screen field names, type of the field and length of the field.
    3. To find BDC_OKCODE for each screen
    4. Create flat file.
    Flat file can be created in your hard disk as follows:
    1 Vinod Krishna Hyderabad
    2 Kavitha Secunderabad
    3 Kishore Hyderabad
    (Where 1st character field is Customer number, 2nd field is Customer name and 3rd field is City.)
    To transfer this data to database table SCUSTOM following interface can be used.
    REPORT DEMO1.
    Following internal table is to upload flat file.
    DATA: BEGIN OF ITAB OCCURS 0,
    ID(10),
    NAME(25),
    CITY(25),
    END OF ITAB.
    *Following internal table BDCDATA is to pass date from internal table to session.
    DATA: BDCTAB LIKE BDCDATA OCCURS 0 WITH HEADER LINE.
    Variables
    DATA: DATE1 LIKE SY-DATUM. DATE1 = SY-DATUM - 1. “ This is for Hold Date
    To upload flat file to internal table.
    CALL FUNCTION UPLOAD
    EXPORTING
    FILE NAME = ‘C:FF.TXT’
    FILE TYPE = ‘ASC”
    TABLES
    DATA_TAB = ITAB
    EXCEPTIONS
    CONVERSION_ERROR = 1
    INVALID_TABLE_WIDTH = 2
    INVALID_TYPE = 3
    NO_BATCH = 4
    UNKNOWN_ERROR = 5
    OTHERS = 6.
    If sy-subrc = 0.
    Calling Function to Create a Session
    CALL FUNCTION ‘BDC_OPEN_GROUP’
    EXPORTING
    CLIENT = SY-MANDT
    GROUP = ‘POTHURI’
    HOLDDATE = DATE1
    KEEP = ‘X’
    USER = SY-UNAME
    EXCEPTIONS
    CLIENT_INVALID = 1
    DESTINATION_INVALID = 2
    GROUP_INVALID = 3
    GROUP_IS_LOCKED = 4
    HOLDDATE_INVALID = 5
    INTERNAL_ERROR = 6
    QUEUE_ERROR = 7
    RUNNING = 8
    SYSTEM_LOCK_ERROR = 9
    USER_INVALID = 10
    OTHERS = 11.
    If sy-subrc = 0.
    *-- MAIN Logic--
    LOOP AT ITAB
    PERFORM GENERATE_DATA. “ Populating BDCDATA Table
    CALL FUNCTION ‘BDC_INSERT’
    EXPORTING
    TCODE = ‘TFBA’
    TABLES
    DYNPROTAB = BDCTAB
    EXCEPTIONS
    INTERNAL_ERROR = 1
    NOT_OPEN = 2
    QUEUE_ERROR = 3
    TCODE_INVALID = 4
    PRINTING_INVALID = 5
    POSTING_INVALID = 6
    OTHERS = 7.
    REFRESH BDCTAB
    ENDLOOP.
    Calling function to close the session
    CALL FUNCTION ‘BDC_CLOSE_GROUP’
    EXCEPTIONS
    NOT_OPEN = 1
    QUEUE_ERROR = 2
    OTHERS = 3.
    Endif.
    Endif.
    *& Form GENERATE_DATA
    Create BDC Data
    FORM GENERATE_DATA
    Passing information for 1st screen on BDCDATA
    BDCTAB-PROGRAM = ‘SAPMTFBA’.
    BDCTAX-DYNPRO = 100.
    BDCTAP-DYNBEGIN = ‘X’.
    APPEND BCDTAB.CLEAR BDCTAB.
    Passing field information to BDCDATA
    BDCTAB-FNAM = ‘SCUSTOM-ID’
    BDCTAB-FVAL = ITAB-ID.
    APPEND BDCTAB.CLEAR BDCTAB.
    Passing BDC_OKCODE to BDCDATA
    BDCTAB-FNAM = ‘BDC_OKCODE’.
    BDCTAB-FVAL = ‘/5’.
    APPEND BDCTAB.CLEAR BDCTAB.
    Passing screen information for next screen to BDCDATA
    BDCTAB-PROGRAM = ‘SAPMTFBA’.
    BDCTAB-DYNPRO = 200.
    BDCTAB-DYNBEGIN = ‘X’.
    APPEND BDCTAB.CLEAR BDCTAB.
    Passing screen information to BDCDATA
    BDCTAB-FNAM = ‘SCUSTOM-NAME’.
    BDCTAB-FVAL = ITAB-NAME.
    APPEND BDCTAB.CLEAR BDCTAB.
    Passing screen information to BDCDATA
    BDCTAB-FNAM = ‘SCUSTOM-CITY’.
    BDCTAB-FVAL = ITAB-CITY.
    APPEND BDCTAB.CLEAR BDCTAB.
    Passing BDC_OKCODE to BDCDATA
    BDCTAB-FNAM = ‘BDC_OKCODE’.
    BDCTAB-FVAL = ‘SAVE’.
    APPEND BDCTAB.CLEAR BDCTAB.
    ENDFORM. “GENERATE_DATA
    AN EXAMPLE WITH CALL TRANSACTION
    Same steps to be repeated for CALL TRANSACTION
    The only difference between the two types of interface is in Session method, you create session and store information about screen and data into session. When session is processed the data is transferred to database. While in CALL TRANSACTION, data is transferred directly to database table.
    REPORT DEMO1.
    Follow above Code till MAIN Logic. Even the Subroutine should be copied
    LOOP AT ITAB
    PERFORM GENERATE_DATA, “Populating BDCDATA Table
    Call transaction ‘TFBA’ using BCDDATA Mode ‘A’ Update ‘S’.
    REFRESH BDCTAB
    ENDLOOP.
    check this link:
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sap-img.com/bdc.htm
    www.sappoint.com/abap/bdcconcept.pdf
    www.sap-img.com/abap/learning-bdc-programming.htm
    www.sap-img.com/abap/question-about-bdc-program.htm
    www.sapdevelopment.co.uk/bdc/bdchome.htm
    www.planetsap.com/bdc_main_page.htm
    Re: bdc mm01
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    http://www.sap-img.com/abap/bdc-example-using-table-control-in-bdc.htm
    http://help.sap.com/saphelp_erp2005/helpdata/en/fa/097119543b11d1898e0000e8322d00/frameset.htm
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    IDOC
    What is IDOC ?
    IDoc
    Standard SAP format for electronic data interchange between systems (Intermediate Document). Different message types (such as delivery confirmations or purchase orders) normally represent different specific formats, the IDoc types. However, multiple message types with related content can be assigned to one IDoc type: For example, the IDoc type ORDERS01 transfers the “logical” message types ORDERS (purchase order) and ORDRSP (order confirmation).
    Also check these link it will help you.
    idoc information
    http://help.sap.com/saphelp_nw2004s/helpdata/en/78/21785851ce11d189570000e829fbbd/frameset.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/xi/sapR3%28Idocs%29ToXI--Steps+Summarized&
    /people/prateek.shah/blog/2005/06/08/introduction-to-idoc-xi-file-scenario-and-complete-walk-through-for-starters
    ALE/ IDOC/ XML
    Troubleshooting of ALE Process - /people/raja.thangamani/blog/2007/07/19/troubleshooting-of-ale-process
    http://www.sapgenie.com/sapgenie/docs/ale_scenario_development_procedure.doc
    http://www.thespot4sap.com/Articles/SAP_XML_Business_Integration.asp
    http://help.sap.com/saphelp_srm30/helpdata/en/72/0fe1385bed2815e10000000a114084/content.htm
    IDOC Convertion
    /people/kevin.wilson2/blog/2005/12/07/changing-fields-in-an-idoc-segment
    Please check this online document for ALE and IDoc.
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCMIDALEIO/BCMIDALEIO.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCMIDALEPRO/BCMIDALEPRO.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/CABFAALEQS/CABFAALEQS.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCSRVEDISC/CAEDISCAP_STC.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCSRVEDI/CAEDI.pdf
    Also check this links for additional information.
    http://help.sap.com/saphelp_erp2004/helpdata/en/dc/6b835943d711d1893e0000e8323c4f/content.htm
    http://www.sapgenie.com/sapgenie/docs/ale_scenario_development_procedure.doc
    http://edocs.bea.com/elink/adapter/r3/userhtm/ale.htm#1008419
    http://www.netweaverguru.com/EDI/HTML/IDocBook.htm
    http://www.sapgenie.com/sapedi/index.htm
    serialization /people/alessandro.guarneri/blog/2006/11/26/content-based-serialization-dynamic-queue-name-in-xi
    /people/prateek.shah/blog/2005/06/08/introduction-to-idoc-xi-file-scenario-and-complete-walk-through-for-starters - IDoc to File
    IDOc testing - /people/suraj.sr/blog/2005/12/29/generate-test-case-for-an-idoc-scenario
    /people/ravikumar.allampallam/blog/2005/06/24/convert-any-flat-file-to-any-idoc-java-mapping - Any flat file to any Idoc
    /people/pooja.pandey/blog/2005/07/27/idocs-multiple-types-collection-in-bpm - Collection of IDoc to Single File
    /people/stefan.grube/blog/2006/09/18/collecting-idocs-without-using-bpm - collecting IDocs without BPM
    /people/prateek.shah/blog/2005/06/08/introduction-to-idoc-xi-file-scenario-and-complete-walk-through-for-starters - IDoc to File
    Hope it will be useful to u....plz don't forget to reward points...!!!!
    Regards
    Vasu

  • Data to be fetched from legacy system to SAP and doc needs to bee posted

    Hi Experts,
    I have a requirement where my client is using some application in Legacy system(Fuel Software) which receives fuel(Petrol, Deasel, CNG) in system and issue the same to company owned vehicles and wants that sap should be integrated and document also needs to be posted in sap for all the receivings and isssuences against their vehicles.
    Process Designed is:
    1. Required Data file will be picked from legacy system with some date stamp and will be placed in sap directory.
    2. There will be batch job which will check for new file based on date stamp.
    3. Will pick record n will pass to Goods_mvt_create bapi and then will post the document.
    4. We have created Vehicles as Internal orders in sap and whenever any issue will be there internal order no will be captured.
    5. By this way we are receiving and issuing in sap.
    Now my problem is i am stuck in designing step 1 and rest 2,3,4,5 steps have already been designed and working fine. Need your help in step 1. Else you may suggest some better way-out but again want to mention rest steps are already mapped
    Thanks in advance..

    Hi Ashutosh,
    You can use functions:
    FTP_CLIENT_TO_R3
    FTP_COMMAND
    FTP_COMMAND_LIST
    FTP_CONNECT
    FTP_COPY
    FTP_DISCONNECT
    FTP_R3_TO_CLIENT
    FTP_R3_TO_SERVER
    FTP_SERVER_TO_R3
    FTP_START_REG_SERVER
    FTP_START_SAPFTP
    FTP_STOP_REG_SERVER
    FTP_VERSION
    which allow you to connect from SAP ERP to any FTP server (if the legacy has it). You can validate the sample program RSFTP007.
    Regards,
    Juan

  • Upload data from legacy system to SAP through ALE IDOC

    Hello All,
    I have a requirement where i need to upload the data from legacy system to SAP. So i am using ALE IDOC.
    In my requirement i need to extend the Standard IDOC. I have extended the IDOC and even found the exit for the updation of the extended fields to SAP. My data would be be placed in the application server.
    Can anybody tell me how to retrieve the data and update in the tables for the extended fields.
    Could you please provide a sample program for retrieving data for IDOC.
    Thanks

    Hi,
    You can find the sample code in the following link.
    Re: calling idoc_input_creditor
    Regards
    Sajid

  • Extracting data from Legacy system - How do we get the data into a normaliz

    Hi All,
    Requirement that I have now is to get the data from 6 different tables ORACLE database to BI using DB Connect, where in each of those 6 tables has the transactional data as well - If it is only 2 tables, would have been creating a view and get the data. Each of the 6 tables are linked to one other by the Primary Key and Foreign key relation ship.
    If it is from ERP system we would have created a Functional Module.
    Please advice me on how to get the data to BI end from those tables in Legacy system.
    Thanks in advance.

    Anil:
    I recommend following Ralph Kimball's EDW data modeling principles.
    Have a staging layer of DSO's with data dump from source system. So have 6 different DSO's extracting data from 6 tables from source system independently.
    Then have an integeration layer of DSO's depending on requirement. You could combine data from 2 or more DSO's here based on requirement and integrate your data here. Use primary key foreign key relationships here to do look ups or direct loads to do your business logic.
    Then have an analytical layer of InfoCubes, MultiProviders, InfoSets as required.
    The 3 layered architecture is a standard best practice that i have seen followed in many organizations. You could search for this type of architecture online under EDW (Enterprise Data Warehouse). There should be plenty of information available.
    Good Luck.
    MP.

Maybe you are looking for