Implementing generic data migrator

Hi, I am planning to implement a data migrator in Java which migrate data from one DB to another. It’s not going to dump tables as it is. Source will have some different structure and destination will have different structure. I want to fix destination tables structure and expose some configuration by which I can map source table’s columns.
Please give me an idea or references to impalement this.
Edited by: cahvarahul on Mar 25, 2010 11:43 PM

cahvarahul wrote:
Hi, I am planning to implement a data migrator in Java which migrate data from one DB to another. It’s not going to dump tables as it is. Source will have some different structure and destination will have different structure. I want to fix destination tables structure and expose some configuration by which I can map source table’s columns.
Please give me an idea or references to impalement this.
As noted implicitly a lot of work. Presumably this is a one shot.
There are commecial products that I believe claim they help with this.
What do you mean "fix destination tables"? Best I can guess that is a different task.
What do you mean by "expose some configuration by which I can map source table’s columns"? As noted that isn't going to be a good idea to allow a novice user to use a tool like that and pointless for someone with enough experience to do it correctly.

Similar Messages

  • Problem with native SQL cursor in generic data source

    Hi, All!
    I am implementing generic data source based on FM.
    Because of complicated SQL I canu2019t use Open SQL and RSAX_BIW_GET_DATA_SIMPLE-example u201Cas isu201D.
    So, I have to use Native SQL. But Iu2019ve got a problem with a cursor. When I test my data source in RSA3, everything is Ok. But, if I start appropriate info-package, I get error u201CABAP/4 processor: DBIF_DSQL2_INVALID_CURSORu201D. It happens after selecting of 1st data package in line u201CFETCH NEXT S1 INTOu2026u201D. It seems to me that when system performs the second call of my FM the opened cursor has already been disappeared.
    Did anyone do things like this and what is incorrect?
    Is it real to make generic data source based on FM with using Native SQL open, fetch, closeu2026

    Hi Jason,
    I don't think this SQL is very valuable It is just an aggregation with some custom rules. This aggregation is performing on info-provider which consists of two info-cubes. Here we have about 2 billion records in info-provider and about 30 million records in custom db-table Z_TMP (certainly, it has indexes). I have to do this operation on 21 info-providers like this and I have to do this 20 times for each info-provider (with different values of host-variable p_GROUP)
    SELECT T.T1, SUM( T.T2 ), SUM( T.T3 ), SUM( T.T4 )
            FROM (
                    SELECT F."KEY_EVENT06088" AS T1,
                            F."/BIC/EV_COST" + F."/BIC/EV_A_COST" AS T2,
                            DECODE( D.SID_EVENTTYPE, 23147, 0,
                                                          23148, 0,
                                                          23151, 0,
                                                          23153, 0,
                                                          23157, 0,
                                                          23159, 0,
                                                          24896734, 0,
                                                          695032768, 0,
                                                          695029006, 0,
                                                          695029007, 0,
                                                          695036746, 0, F."/BIC/EV_COST") +
                              DECODE( D.SID_EVENTTYPE, 23147, 0,
                                                          23148, 0,
                                                          23151, 0,
                                                          23153, 0,
                                                          23157, 0,
                                                          23159, 0,
                                                          24896734, 0,
                                                          695032768, 0,
                                                          695029006, 0,
                                                          695029007, 0,
                                                          695036746, 0, F."/BIC/EV_A_COST") AS T3,
                            DECODE( D.SID_EVENTTYPE, 23147, F."/BIC/EV_DURAT",
                                                          23148, F."/BIC/EV_DURAT",
                                                          23151, F."/BIC/EV_DURAT",
                                                          23153, F."/BIC/EV_DURAT",
                                                          23157, F."/BIC/EV_DURAT",
                                                          23159, F."/BIC/EV_DURAT",
                                                          24896734, F."/BIC/EV_DURAT",
                                                          695032768, F."/BIC/EV_DURAT",
                                                          695029006, F."/BIC/EV_DURAT",
                                                          695029007, F."/BIC/EV_DURAT",
                                                          695036746, F."/BIC/EV_DURAT", 0) AS T4
                      FROM "/BIC/VEVENT0608F" F,
                           Z_TMP G,
                           "/BIC/DEVENT06085" D
                      WHERE F."KEY_EVENT06088" = G.ID
                            AND F."KEY_EVENT06085" = D.DIMID
                            AND G.GROUP_NO = :p_GROUP
                            AND ( F."/BIC/EV_COST" < 0 OR F."/BIC/EV_A_COST" < 0 )
                            AND D.SID_EVENTTYPE <> 695030676 AND D.SID_EVENTTYPE <> 695030678
                    UNION
                    SELECT F."KEY_EVNA06088" AS T1,
                            F."/BIC/EV_COST" + F."/BIC/EV_A_COST" AS T2,
                            DECODE( D.SID_EVENTTYPE, 23147, 0,
                                                          23148, 0,
                                                          23151, 0,
                                                          23153, 0,
                                                          23157, 0,
                                                          23159, 0,
                                                          24896734, 0,
                                                          695032768, 0,
                                                          695029006, 0,
                                                          695029007, 0,
                                                          695036746, 0, F."/BIC/EV_COST") +
                              DECODE( D.SID_EVENTTYPE, 23147, 0,
                                                          23148, 0,
                                                          23151, 0,
                                                          23153, 0,
                                                          23157, 0,
                                                          23159, 0,
                                                          24896734, 0,
                                                          695032768, 0,
                                                          695029006, 0,
                                                          695029007, 0,
                                                          695036746, 0, F."/BIC/EV_A_COST") AS T3,
                            DECODE( D.SID_EVENTTYPE, 23147, F."/BIC/EV_DURAT",
                                                          23148, F."/BIC/EV_DURAT",
                                                          23151, F."/BIC/EV_DURAT",
                                                          23153, F."/BIC/EV_DURAT",
                                                          23157, F."/BIC/EV_DURAT",
                                                          23159, F."/BIC/EV_DURAT",
                                                          24896734, F."/BIC/EV_DURAT",
                                                          695032768, F."/BIC/EV_DURAT",
                                                          695029006, F."/BIC/EV_DURAT",
                                                          695029007, F."/BIC/EV_DURAT",
                                                          695036746, F."/BIC/EV_DURAT", 0) AS T4
                    FROM "/BIC/VEVNA0608F" F,
                         Z_TMP G,
                         "/BIC/DEVNA06085" D
                    WHERE F."KEY_EVNA06088" = G.ID
                          AND F."KEY_EVNA06085" = D.DIMID
                          AND G.GROUP_NO = :p_GROUP
                          AND ( F."/BIC/EV_COST" < 0 OR F."/BIC/EV_A_COST" < 0 )
                          AND D.SID_EVENTTYPE <> 695030676 AND D.SID_EVENTTYPE <> 695030678
                 ) T
            GROUP BY T.T1

  • HELP ME ON DATA MIGRATION PLEASE

    Hi
    what is SAP data migration what kind of tools and techniques include in that and also please what are the roles and responsibilities include in that when we are working on project.(RELATED TO ABAP)

    Data Migration – Current Situation
    Data migration or data conversion is a persistent issue facing most organizations who have implemented SAP. Whether it is during the final phases of a new SAP implementation, during SAP upgrades and updates, during corporate restructurings, or during mergers and acquisitions, data migration continues to remain a challenging problem for IT.
    Most data migration projects are faced with a time and budget crunch. But, these projects usually require extensive support from programmers and other technical experts and therefore data migration projects tend to get expensive and time-consuming.
    The tools for data transfer and data migration provided by SAP are very technical and have a steep learning curve associated with them and they require technical experts to build and implement the data migration scripts.
    Data Migration – Ideal Scenario
    An ideal data migration scenario would let the data migration project be implemented by the end user departments themselves. A few super-users within these end-user departments that supply the data should have the ability to transfer and migrate data themselves without relying on technical experts. Such a scenario would significantly cut the time and effort required in a data migration project.
    Reaching such an ideal scenario would require data migration tools that are easy to learn and require no programming. Furthermore, these tools should work across all the different SAP modules and the SAP products, including the different versions of SAP.
    Data Migration
    Data Migration is the process of moving required (and most often very large) volumes of data from our clients’ existing systems to new systems. Existing systems can be anything from custom-built IT infrastructures to spreadsheets and standalone databases.
    Data Migration encompasses all the necessary steps to cleanse, correct and move the data into the new system *. Technological changes, change in providers, software updates or data warehousing / data mining projects make such delicate and critical operations necessary.
    <b>Data Migration can be done using BDC, LSMW, BAPI, Idoc's.</b>
    What is the difference between batch input and call transaction in BDC?
    Session method.
    1) synchronous processing.
    2) can tranfer large amount of data.
    3) processing is slower.
    4) error log is created
    5) data is not updated until session is processed.
    Call transaction.
    1) asynchronous processing
    2) can transfer small amount of data
    3) processing is faster.
    4) errors need to be handled explicitly
    5) data is updated automatically
    Differences between bdc session method and call transaction method.
    The most important aspects of the batch session interface are: - Asynchronous processing - Transfers data for multiple transactions - Synchronous database update During processing, no transaction is started until the previous transaction has been written to the database. - A batch input processing log is generated for each session - Sessions cannot be generated in parallel
    The most important aspects of the CALL TRANSACTION USING interface are: - Synchronous processing - Transfers data for a single transaction - Synchronous and asynchronous database updating both possible The program specifies which kind of updating is desired. - Separate LUW for the transaction The system performs a database commit immediately before and after the CALL TRANSACTION USING statement. - No batch input processing log is generated.
    Refer this
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    Check these link:
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://www.sap-img.com/abap/question-about-bdc-program.htm
    http://www.itcserver.com/blog/2006/06/30/batch-input-vs-call-transaction/
    http://www.planetsap.com/bdc_main_page.htm
    call Transaction or session method ?
    http://www.sapbrain.com/FAQs/TECHNICAL/SAP_ABAP_DATADICTIONARY_FAQ.html
    http://www.****************/InterviewQ/interviewQ.htm
    http://help.sap.com/saphelp_46c/helpdata/en/35/2cd77bd7705394e10000009b387c12/frameset.htm
    ALE/ IDOC
    http://help.sap.com/saphelp_erp2004/helpdata/en/dc/6b835943d711d1893e0000e8323c4f/content.htm
    http://www.sapgenie.com/sapgenie/docs/ale_scenario_development_procedure.doc
    http://edocs.bea.com/elink/adapter/r3/userhtm/ale.htm#1008419
    http://www.netweaverguru.com/EDI/HTML/IDocBook.htm
    http://www.sapgenie.com/sapedi/index.htm
    http://www.sappoint.com/abap/ale.pdf
    http://www.sappoint.com/abap/ale2.pdf
    http://www.sapgenie.com/sapedi/idoc_abap.htm
    http://help.sap.com/saphelp_erp2005/helpdata/en/0b/2a60bb507d11d18ee90000e8366fc2/frameset.htm
    http://help.sap.com/saphelp_erp2005/helpdata/en/78/217da751ce11d189570000e829fbbd/frameset.htm
    http://www.allsaplinks.com/idoc_sample.html
    http://www.sappoint.com/abap.html
    http://help.sap.com/saphelp_erp2004/helpdata/en/dc/6b835943d711d1893e0000e8323c4f/content.htm
    http://www.sapgenie.com/sapgenie/docs/ale_scenario_development_procedure.doc
    http://edocs.bea.com/elink/adapter/r3/userhtm/ale.htm#1008419
    http://www.netweaverguru.com/EDI/HTML/IDocBook.htm
    http://www.sapgenie.com/sapedi/index.htm
    http://www.allsaplinks.com/idoc_sample.html
    Check these step-by-step links
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/ccab6730-0501-0010-ee84-de050a6cc287
    https://sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/8fd773b3-0301-0010-eabe-82149bcc292e
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/3c5d9ae3-0501-0010-0090-bdfb2d458985
    regards,
    srinivas
    <b>*reward for useful answers*</b>

  • Data Migration from Legacy System in CRM

    Hi All,
    We are going to Implement a new CRM project, I have a problem with LSMW(Legacy System Migration Workbench ), I have some Conversion Objects of CRM and I need to know whether Data Migration is possible or not, please tell me how to find these Objects and how to know Data Migration is possible or not.
    Objects are like.,
    1. Accounts
    2. Actuate Reports
    3. Active Campaigns/Campaign Content/Dispositions
    4. Contacts
    5. Contracts
    6. Opportunities
    7. Payment Arrangement History
    8. Payments
    9. Premises
    10. Rate Changes
    11. Security Deposits
    12. Business Partner Relationships
    13. Web self-service information
    14. Usage that has been used for quotes
    15. Tax history information, including AXCIS
    17. Service Requests
    18. Service Order History
    19. Security Deposits
    20. Reference Values
    21. Rate Changes
    Can anybody please tell any transaction code in CRM where can I find the Data migration for the above objects is possible or not.
    Thanks in Advance,
    Sai.

    Hello,
    for migration into CRM I would suggest the 'XIF-Interfaces'; these are interfaces provided in CRM for connection to external systems. I'm not sure if this will cover all objects mentioned above, but at least some of them should exist.
    Some information about existing interfaces can be found in the integration repository under http://ifr.sap.com/index.html; from the start screen first select 'Enter the repository' and then have a look under 'generic components' and 'SAP CRM'.
    Regards, Katja Ohliger

  • Validation rules applied to data migration templates at import

    Hi everyone!
    First post here for me, so please bear with me if I missed something.
    My company has just started the initial implementation of ByDesign. We come from a set of disparate and partially home-grown systems that we outgrew a few years ago.
    As part of this initial phase, we are basically re-creating the data on customers, suppliers, etc. since none of our existing systems makes a good source, unfortunately. We will be using the XML templates provided by ByDesign itself to import the relevant data.
    It has become clear that ByDesign applies validation rules on fields like postal codes (zip codes), states (for some countries), and other fields.
    It would be really helpful if we could get access to the rules that are applied at import time, so that we can format the data correctly in advance, rather than having to play "trial and error" at import time. For example, if you import address data, the first time it finds a postal code in the Netherlands which is formatted as "1234AB", it will tell you that "there needs to a space in the 5th position, because it expects the format to be "1234 AB". At that point, you stop the import, go back to the template to fix all the Dutch postal codes, and try the import again, only to run into the next validation issue.
    We work with a couple of very experienced German consultants to help us implement ByDesign, and I have put this question to them, but they are unaware of a documented set of validation rules for ByDesign. Which is why I ask the question here.
    So just to be very celar on what we are looking for: the data validation/formatting rules that ByDesign enforces at the time the XML data migration templates are imported.
    Any help would be appreciated!
    Best regards,
    Eelco

    Hello Eelco,
    welcome to the SAP ByDesign Community Network!
    The checks performed on postal codes are country specific, and represent pretty much the information that you would find in places like e.g. the "Postal Codes" page in Wikipedia.
    I recommend to start with small files of 50-100 records that are assembled of a representative set of different records, in order to collect the validation rules that need reactions based on your data in an efficient way. Only once you have caught these generic data issues, I would proceed to larger files.
    Personnaly I prefer to capture such generic work items on my list, then fix the small sample file immediately by editing, and do an immediate resimulation of the entire file, so that I can drill deeper and collect more generic issues of my data sample. Only after a while when I have harvested all learnings that were in my sample file, I would then apply the collected learnings to my actual data and create a new file - still not too large, in order to use my time efficiently.
    Best regards
    Michael  

  • Data Migration for Open Purchase Order

    Hi, All,
    Is there anyone know how to Count the volume for Open Purchase Order. What's the normal strategy for the Data Migration and Cut-over stage?
    My client want to know how many Open Purchase Order in the legacy system and then determine manual or automatic data migration. If manual, how to do? If automatic, how to do? Because all materials and vendors, plants are different number. How to track? How to find out to match between new and old?
    Thank you very much

    JC,
    Sounds a bit early to be making decisions about the realization phase.  It doesn't sound like you have finished the Blueprinting phase yet, much less the testing phase.
    Anyhow, in my experience I typically use LSMW (Legacy system migration workbench) to load MM master data (material masters), Inventory (WIP, RM, FG, etc) Purchasing Master data (Vendors, Purchase Info Records, Source Lists, Quota Arrangements), and Purchasing transactional documents (POs, PurReqs, Scheduling Agreements, etc).  Depending on the complexity and volume of data, it  may be necessary to write custom programs to load the data.  You will find this out during your requirements gathering.
    It is uncommon but possible to load all of these data manually.  I have never run across a client that wants to pay a consultant's hourly rate to sit at a terminal to peck away loading master data, so if the client intends to have his own users enter the data manually, the project manager should make provision that there will be qualified TRAINED client employees available for this data entry.  I did help with a portion of a conversion once manually; of Sales Credits, but there were only about 30 SD docs to load.   I did this the evening before go-live day, while I was waiting for some of my LSMW projects to complete in the background.
    A good opportunity to 'practice' your data loads is right after you have completed your development and customization, and you have gotten the approval from the client to proceed from the pilot build to the full test environment.  Once you have moved your workbench and customization into the client's test environment, but before integration testing, you can mass load all, or a substantial portion of your conversion data into the qual system.  You can treat it like a dry run for go-live, and fine tune your processes, as well as your LSMW projects.
    Yes, it is good practice to generate comparisons between legacy and SAP even if the client doesn't ask for it. For Purchase orders on the SAP side, you could use any of the standard SAP Purchasing reports, such as ME2W, ME2M, ME2C, ME2L, ME2N.  If these reports do not meet the requirements of the client, you could write a query to display the loaded data, or have an ABAPer write a custom report.
    You didn't ask, but you should also do comparisons of ALL loaded data - including master data.
    It sounds like you are implying that the client wants YOU to extract the legacy data.  For an SAP consultant, this is not very realistic (unless the legacy system is another SAP system).  Most of us do not understand the workings of the myriad legacy systems.  The client is usually expected to produce one or more legacy system technical experts for you to liase with.  You normally negotiate with the technical expert about every facet of of the data migration.  In addition, you will liase with business users, who will help you and the implementation team to logically validate that the final solution (turnkey SAP production system, fully loaded with data) will meet the client's business needs.
    Finally, you mentioned how do you track the mapping of master data between legacy and SAP.  There are many ways to do this.  I normally try to get the legacy person do the conversion on his end, eg, when he gives you the load file, you would like to have already translated the master data and inserted the SAP relevant values into the file.  If this is not possible, I usually use MS Access databases to maintain a master map, and I perform the mapping on a PC.  If your data package is small, you can probably get by using MS Excel or similar.
    Good Luck,
    DB49

  • Data migration

    Dear all my friends,
    please help to share me about data migration before go-live FI system.
    i plan to migrate GL acct balance, AP/AR open item and last 3 months historical data.
    please help to show me : how should i do ?
    thanks !

    1. Normal GL Accounts: In this case line items are not required and totals are to be updated in SAP. This can be done using GL fast entry if the number is not too high.
    2. Open Item Managed GL Accounts: All the open line items are required to be uploaded in SAP from Legacy in order to clear them. Hence, a LSMW / BDC program has to be developed to upload the line items.
    BAPI_ACC_DOCUMENT_POST - for GL/AP/AR
    BAPI_ACC_INVOICE_RECEIPT_POST - for AP
    BAPI_ACC_GL_POSTING_POST - for GL
    GL Balance
    F-02 or RFBIBL00 or BAPI
    Dr GL a/c Credit Data migration Account or Clearing Account
    3. Customer & Vendor Open Items: All open line items in Customer and Vendor Accounts have to be brought in into SAP. For this also a LSMW / BDC program needs to be developed and upload the line items.
    Vendor Balances
    F-02 or RFBIBL00 or BAPI
    Credit Vendor a/c (individually) and Dr Data migration Account or Clearing Account
    Customer balances
    F-02 or RFBIBL00 or BAPI
    Dr Customers a/c (individually) and Dr Data migration Account or Clearing Account
    4. Asset Accounting: In case asset accounting is implemented the individual balances of each asset (Gross Value and Accumulated Depreciation) have to be uploaded using transaction AS91. For this also you need to develop a LSMW / BDC program. Further, you also need to update the GL Accounts related to Asset Accounts with totals for each asset class T.Code OASV.
    For uploading the Assets
    AS91 -> for uploading assets in AA
    OSAV --> for uploading assets in GL entry Dr Asset (individually) Cr Data migration Account or Clearing Account
    You can use LSMW for uploading all these things
    After uploading all, your data migration account will become zero

  • Data Migration from Legacy System in IS-U

    Hi All,
    We are going to Implement a new IS-U project,  I have a problem with LSMW(Legacy System Migration Workbench ),  I have some Conversion Objects of IS-U and I need to know whether Data Migration is possible or not, please tell me how to find these Objects and how to know Data Migration is possible or not.
    Objects are like.,
    1. Accounts
    2. Actuate Reports
    3. Business Partner Relationships
    4. Active Campaigns/Campaign Content/Dispositions
    5. Connection Object
    6. Contacts
    7. Contracts
    8. Opportunities
    9. Payment Arrangement History
    10. Payments
    11. Premises
    12. Rate Changes
    13. Security Deposits
    these are few and there are some more..,
    Thanks in Advance,
    Sai.

    Hi Ram,
    Use Transaction Code EMIGALL. It will ask for company code. By default Company Code is SAP. If you entered with the SAP you will get all the objects.Then goto menu IS-U Migration-->User Handbook. It will give you details Idea.
    Also Check the following Procedure
    You can find detailed documentation of EMIGALL in SAP itself. Use Transaction EQ81 to display it. It provides all the concepts and procedures to work with EMIGALL.
    Here are some points about EMIGALL :
    1. It Migrates data Business Object wise
    2. It uses Direct Input Technique
    3. It has more than 100 objects of IS-U
    and the steps for implementation goes like this:
    1)You have to create a user specially for migration which will have all the authorizations related to migration workbench, BASIS and IS-U
    2)You have to create your own company in EMIGALL. There is a default company called SAP.
    3)Company SAP contains all the Business Objects
    4)You have to figure out what business objects u need and then u have to copy those business objects to ur company from Standard Company SAP
    5)Each objects contains more than one structure and each structure can contain more than one fields. The relation goes like this
    Object ---> Structure ---> Field
    6)You have to define field rules for each required field of the object. You have to mark "Not required" for fields u don't need
    7)After field rules for a given object is set u have to generate load report i.e. actual Direct Input Program which will migrate data. This program is generated on basis of field rules set by u.
    8)After the load report is generated u have to prepare an input file (import File) for migration. The import file should be according to structure provided by SAP and must be in binary format. SAP Provides the structure of file according to your configurations. You have to write ur own Data conversion program(in any language) for this task.
    9)You take import file as input and migrate the data using generated load program
    10)Finally u can check the Migration Statistics and Error Log
    Thanks and Regards,
    Jyotishankar Dutta
    Message was edited by:
            Jyotishankar Dutta
    Message was edited by:
            Jyotishankar Dutta

  • Data Migration From Peoplesoft , JDEdwards To SAP.

    Hi,
    This is kiran here we are doing data Migration work from Peoplesoft And JDEdwards to SAP.in SAP side it involves Master data tables Related to Customer, Vendor, Material. and Meta data tables related to SD, MM, FI. We as SAP Consultant identified Fields from above tables and marked them as Required, Not required, And Mandatory. The Peoplesoft and JDEdwards flocks come up with the same from their side. Then we want map the Fields. as I am new to data Migration any body suggest me what are the steps involves in data Migration How to do Data Mapping in Migration Thanks in advance.
    Thanks
    Kiran.B

    Hi Kiran,
    Good... Check out the following documentation and links
    Migrating from one ERP solution to another is a very complex undertaking. I don't think I would start with comparing data structures. It would be better to understand the business flows you have currently with any unique customizations and determine how these could be implemented in your target ERP. Once this is in place, you can determine the necessary data unload/reload to seed your target system.
    A real configuration of an ERP system will only happen when there is real data in the system. The mapping of legacy system data to a new ERP is a long difficult process, and choices must be made as to what data gets moved and what gets left behind. The only way to verify what you need to actually run in the new ERP environment is to migrate the data over to the ERP development and test environments and test it. The only way to get a smooth transition to a new ERP is to develop processes as automatic as possible to migrate the data from the old system to the new.
    Data loading is not a project that can be done after everything else is ready. Just defining the data in the legacy system is a huge horrible task. Actually mapping it to one of the ERP system schemas is a lesson in pain that must be experienced to be believed.
    The scope of a data migration project is usually a fairly large development process with a lot of proprietary code written to extract legacy data, transform and load the data into the ERP system. This process is usually called ETL (extract, transform, load.)
    How is data put into the ERP?
    There is usually a painfully slow data import facility with most ERP systems. Mashing data into the usually undocumented table schema is also an option, but must be carefully researched. Getting the data out of the legacy systems is usually left to the company buying the ERP. These export - import processes can be complex and slow, sometimes specialized ETL tools can help, sometimes it is easier to use what ever your programmers are familiar with, tools such as C, shell or perl.
    An interesting thing to note is that many bugs and quirks of the old systems will be found when the data is mapped and examined. I am always amazed at what data I find in a legacy system, usually the data has no relational integrity , note that it does not have much more integrity once it is placed in an ERP system so accurate and clean data going in helps to create a system that can work.
    The Business Analysts (BAs) that are good understand the importance of data migration and have an organized plan to migrate the data, allocate resources, give detailed data maps to the migrators (or help create the maps) and give space estimates to the DBAs. Bad BAs can totally fubar the ERP implementation. If the BAs and management cannot fathom that old data must be mapped to the new system, RUN AWAY. The project will fail.
    Check these links
    http://pdf.me.uk/informatica/AAHN/INFDI11.pdf
    http://researchcenter.line56.com/search/keyword/line56/Edwards%20Sap%20Migration%20Solutions/Edwards%20Sap%20Migration%20Solutions
    http://resources.crmbuyer.com/search/keyword/crmbuyer/Scm%20Data%20Migration%20On%20Peoplesoft%20Peoplesoft%20Data%20Migration/Scm%20Data%20Migration%20On%20Peoplesoft%20Peoplesoft%20Data%20Migration
    Good Luck and Thanks
    AK

  • DMS Data Migration-Help required

    Hi,
    I don't have enough knowledge on DMS Data migration.
    In our recent Implementation, we are implementing SAP DMS.
    Our client has more legacy data which is in their server. So we have to upload all the data's into SAP DMS.
    I gone through few SCN threads and understand that LSMW & BDC will be used for Data migration.
    Which is the best method for Data migration?
    We are having 50+ document types and it has different processes (status network), Classification & few Object links.
    Please advice me and provide me some examples to upload legacy data in DMS.
    Thanks in Advance,
    Ravi

    Hi Ravi,
    There are few approach discussed in the thread https://scn.sap.com/thread/2119796
    You may check feasibility on these approaches for your scenario.
    We are having 50+ document types and it has different processes (status network), Classification & few Object links.
    Creating these configurations into SAP is one time and should not be a problem.
    I gone through few SCN threads and understand that LSMW & BDC will be used for Data migration.
    Which is the best method for Data migration?
    How much volume of data do you have in legacy system ?
    Have a thought on whether do you really need to migrate all these data into SAP DMS   ???
    OR
    You can configure the DMS system and start fresh leaving the document in legacy system as it is.
    Hope this helps.
    Regards,
    Deepak Kori

  • Data Migration techniques

    Hi Experts,
    I want to know about data migration techniques and how we can best use MDM while migrating old version of R/3 to new version of R/3....
    I have implemented SAP MDM in cases where we have number of SAP R/3 instances across different region and we were taking data from different R/3 one by one and doing data stan'zation,consolidation, Harmonization and all... I am not talking about all this...
    There was a good explanation from <b>Markus Ganser</b> about <b>Duplicate data and identical data</b>... I know all this..but my question is when I have <b>only one SAP R/3</b> and I still want to implement MDM solution while migrating my old R/3 instance to new one, How can I proceed in this scenario? What is the data migration technique...
    I know the common answer will be to use MDM as a middle ware..take Master data from old instance and after consolidation, send it back to new instance and at the same time send tran'data directly to new version... But is this worth doing this? Is there any other approach?
    If there is any document on this or any one have idea about data migration techniques while implementing MDM solution than send me documents on [email protected].......
    In short, I am looking for below 3 points while doing migration along with SAP MDM
    <b><b>Data Migration techniques</b>
    <b>Prerequisites</b>
    <b>Methodology in this kind of scenario</b>
    Step by step procedure</b>
    cheers,
    R.n

    hi..
    here i am sending u the link for complete Data Migration Life Cycle
    <a href="http://www.redwoodsystems.co.uk/dataMWhitePaper.html#links">http://www.redwoodsystems.co.uk/dataMWhitePaper.html#links</a>
    hope it might be of any use to u
    thank you & reward points if useful
    Message was edited by:
            Dasari Narendra

  • Data Migration from R3 to new ECC System

    Hello,
    We are doing a fresh implementation for SAP MDM. Currently we have 3 SAP R/3 systems and 1 non-R3 system (BPCS system). One of the project goals is to do data migration from all these 4 systems into a new SAP ECC system and to MDM for Vendor Master, Customer Master and Material Master. My manager says that we need Business Objects tool to Migrate these data. But, my question is that is it not possible to achive this through XI/PI itself ? Can anyone let me know the entire process in XI/PI as to how do I receive the Idocs from R/3, convert into XML and then send it to the new ECC system in Idoc format ?
    Thks & Rgds,
    Hema

    Hi Gandhi,
    You have SAP MDM, BPC and R/3 systems in your landscape, when you are migrating data to new ECC system, what is the role of mdm, did you moving the same data to MDM to avoid duplicate/redundant data??is there any data migration from R/3 to MDM(any integration required)or MDM to R/3,
    You have to develop different interfaces to transfer data from MDM to ECC, BPC to ECC.definelty you required one middleware, the option definitely SAP PI. it offers very easy of integration between MDM and R/3 system, you can use MDM-PI adapter to connect to directly MDM syndicator file port from there you can pull the XML file and convert it in to IDoc format use Receiver IDoc adapter in PI, the same way we can integrate BPC system with R/3.
    Search in sdn, there are some scenarios integration of R/3 and MDM using PI, refer those links, if still you have any doubts let me know,
    Regards,
    Raj

  • SAP Accelerated Data Migration (ADM) Tool

    Hi-
    Has anyone used ADM for SAP HR/HCM data migration?
    If yes, can you please share your experience?
    <REMOVED BY MODERATOR>
    Thanks,
    Edited by: Hunk2007 on Feb 19, 2008 6:11 PM
    Edited by: Alvaro Tejada Galindo on Feb 19, 2008 12:14 PM

    Hey
    Did anyone use ADM for SAP HCM Implementation ??? Are you planning to use this tool or use the standard ways to load data ?
    Please let me know
    Deepthi

  • Data Migration - SAP EC - Environmental Compliance

    Hi,
    I’m in a Project here in Brazil and they will start to use the portal SAP EC and I have to do the data migration.
    I read in some manuals about the “Export Excel Template”, but I don’t know if this is the correct way to do the data migration for this portal. If I use this way, is this possible I change the fields and create a new template?
    Thanks a lot!
    Regards,

    Hi Victor,
    EC30_XML_TransferTool.ZIP  
    PermitDataTemplateSample.zip
    This Zip File containing the  EC standard documentation of the XML Transfer Template and an example definition file.
    This is correct Template for export and import for mass data .
    for User Manual Click here
    EC_FP3_overview_part1-generic.pdf open this Pdf and see the page number 16. ( EC provides a separate help document for the new Excel tool.)
    All the Best
    Edward

  • RDS for SAP data migration through IDOC's

    Dear Experts
      Just now we have installed BODS 4.1 and RDS also installed now. I am in the process of exploring the RDS. I found all the migration objects is available in the form of IDOC's. Now my questions are
    How to work on custom IDOC's. Do i need to extensively code on R/3 side to upload the custom fields. Like user exit or BADI for custom fileds
    If yes then i would say implementation time through RDS will remain more or less same time.
    Through BODS or by RDS migration objects is there any way without coding on ABAP R/3 side, Can we achieve migrating custom fields
      Please let me how to achieve above requirements by RDS
    Thanks
    Vijay Mukunthan

    Hi Vijay,
    You are right, the standard content of the Rapid Data Migration RDS packages brings the content with the standard IDoc interface for SAP standard only. So custom objects are not pre-built with the package. However, we incorporated the foundation to include custom coding as easy as possible:
    1) IDoc
    We are using the standard interface technology IDoc. IDocs do have an enhancement concept which is described in the online help. There are three main cases and all of them are supported, however two need additional coding:
    Z-fields: Additional fields in existing tables lead to IDoc segment extensions
    With the IDoc segment versioning concept this is easy to realize without any ABAP coding
    Z-tables: Additional IDoc segments will be necessary and lead to a new IDoc version
    With the IDoc versioning concept you can easily add Z-segments with leveraging provided user exits
    Custom business objects:
    You can even create a complete own IDoc message type in Z-namespace for own stuff, leveraging the given ALE layer of the IDoc middleware
    2) Enhancement guide
    With the RDS package we offer an enhancement guide which helps you to modify the jobs and data flows in the SAP Data Services content according to the changes in the target IDoc structure to reflect additional fields and tables. We built it as a step-by-step guide following for a sample IDoc type but will work similarly for any IDoc.
    Best regards,
    Frank Densborn.

Maybe you are looking for

  • What video capture setting to use???

    I'm using FCP 4.1 for some non-broadcast, DVD applications. I'm looking for BetaSP quality output but have no idea what setting will give me good image quality without slowing my Mac. Since I use a Kona LS card, there seem to be almost an endless amo

  • Setting up the notes programmatically

    Hi Guys,  We are developing a SharePoint 2013 app for a Club. We have two groups (Coaches, Players) . As a Coach, I want to setup Notes about a particular player in my team, These notes would be Comments or notes or simple text about the behavior of

  • IP10 - Behaviour of "Start in cycle"

    Hi experts! We have a Maintenance plan where we have Scheduled  Calls every month. We have made several changes in routing and instead of modify the routing of every order generated in order to update the orders we have set status TECO (Technically c

  • Workflow Processes in OWB10g

    Looking forward to the forthcoming release of OWB10g, but I am surprised to see that the 10g database is missing workflow. Does this mean that OWB10g will require a 9i OEM home with 9i (2.6.2 standalone) workflow installed in order to execute process

  • Unable to Use Adjust Tool or Effects Tool

    I'm using iPhoto 7.1.5. On my MacBook Pro, when I go into edit mode and click on either the Adjust tool or the Effects tool, nothing happens. No windows open to allow me to edit with these tools. The other tools (red eye, enhance, retouch, crop) work