Complete HR data migration

Dear experts,
we have a hr server for total enterprise structure. But as per business requirement we have a new server for 1 of our business. In this server we want to transfer entire data for corresponding employees only.
We have tried to transfer data from existing server to the new server.
We used ALE and a PFAL transaction to transfer the data.
We have succesfully transfered employee's master data like standard infotypes and Org details.
But the problem is we are not able to transfer Custom Infotypes and any Clusters like Payroll cluster, Time clusters, Claim Clusters,etc.
Can anyone please help me transferring this data.
Thanks & Regards,
Deepak.

Hey,
Thanks a lot for the information. I am trying to explore the transaction PU12 as well as few more like PU12_EXPORT.
If possible can you please explain me in short, how to use this transaction please, it will be very appreciable.
Thanks & Regards,
Deepak.

Similar Messages

  • Data Migration for Open Purchase Order

    Hi, All,
    Is there anyone know how to Count the volume for Open Purchase Order. What's the normal strategy for the Data Migration and Cut-over stage?
    My client want to know how many Open Purchase Order in the legacy system and then determine manual or automatic data migration. If manual, how to do? If automatic, how to do? Because all materials and vendors, plants are different number. How to track? How to find out to match between new and old?
    Thank you very much

    JC,
    Sounds a bit early to be making decisions about the realization phase.  It doesn't sound like you have finished the Blueprinting phase yet, much less the testing phase.
    Anyhow, in my experience I typically use LSMW (Legacy system migration workbench) to load MM master data (material masters), Inventory (WIP, RM, FG, etc) Purchasing Master data (Vendors, Purchase Info Records, Source Lists, Quota Arrangements), and Purchasing transactional documents (POs, PurReqs, Scheduling Agreements, etc).  Depending on the complexity and volume of data, it  may be necessary to write custom programs to load the data.  You will find this out during your requirements gathering.
    It is uncommon but possible to load all of these data manually.  I have never run across a client that wants to pay a consultant's hourly rate to sit at a terminal to peck away loading master data, so if the client intends to have his own users enter the data manually, the project manager should make provision that there will be qualified TRAINED client employees available for this data entry.  I did help with a portion of a conversion once manually; of Sales Credits, but there were only about 30 SD docs to load.   I did this the evening before go-live day, while I was waiting for some of my LSMW projects to complete in the background.
    A good opportunity to 'practice' your data loads is right after you have completed your development and customization, and you have gotten the approval from the client to proceed from the pilot build to the full test environment.  Once you have moved your workbench and customization into the client's test environment, but before integration testing, you can mass load all, or a substantial portion of your conversion data into the qual system.  You can treat it like a dry run for go-live, and fine tune your processes, as well as your LSMW projects.
    Yes, it is good practice to generate comparisons between legacy and SAP even if the client doesn't ask for it. For Purchase orders on the SAP side, you could use any of the standard SAP Purchasing reports, such as ME2W, ME2M, ME2C, ME2L, ME2N.  If these reports do not meet the requirements of the client, you could write a query to display the loaded data, or have an ABAPer write a custom report.
    You didn't ask, but you should also do comparisons of ALL loaded data - including master data.
    It sounds like you are implying that the client wants YOU to extract the legacy data.  For an SAP consultant, this is not very realistic (unless the legacy system is another SAP system).  Most of us do not understand the workings of the myriad legacy systems.  The client is usually expected to produce one or more legacy system technical experts for you to liase with.  You normally negotiate with the technical expert about every facet of of the data migration.  In addition, you will liase with business users, who will help you and the implementation team to logically validate that the final solution (turnkey SAP production system, fully loaded with data) will meet the client's business needs.
    Finally, you mentioned how do you track the mapping of master data between legacy and SAP.  There are many ways to do this.  I normally try to get the legacy person do the conversion on his end, eg, when he gives you the load file, you would like to have already translated the master data and inserted the SAP relevant values into the file.  If this is not possible, I usually use MS Access databases to maintain a master map, and I perform the mapping on a PC.  If your data package is small, you can probably get by using MS Excel or similar.
    Good Luck,
    DB49

  • Data Migration from 11i to R12 Global - Open POs,lines, receipts & on hand upload, Is it possible to do the onhand qty upload with over riding of all receipts which uploaded against Open PO lines?

    Hi Friends,
    We are in a phase of data migration from 11i to R12 
    I was discussed with client & they wants extraction of all open POs which was generated after 01 Jan 2014 to till date in 11i.
    Condition for open POs is PO qty-received qty=>0
    critical Example for open PO is :PO no: 10 has 4 lines, 3lines full qty has been received & for 1 line partial qty(say 50 out of 100) received.
    in this case he wants in R12 uploading as PO no:10 should entered as open PO with all 4lines & 3 lines complete receipt should be done, for 4th line partial qty i.e 50 should be received.
    the question is if we upload on hand qty first, then open POs & receipts, it will increase the onhand qty in new system(mismatch of on hand qty's 11i to R12) 
    Is it possible to do the onhand qty upload with over riding of all receipts which uploaded against Open PO lines.
    Or Please advice best solution.
    Thanks & Regards
    Giri

    adetoye50 wrote:
    Dear Contacts Journal Support Team,
    FYI, this is a user to user support forum.  You are NOT addressing Apple here.
    Honestly, I doubt anyone is really going to take the time to read the novel you have written.

  • Data Migration from CSV file to Database Table.

    Hi,
    I have checked few answered threads on Data Migration Issues but didn't find the solution yet.
    I am reading data from an CSV file to internal table but the complete row is coming under one field of the internal table.
    How i can get values in different fields of the internal table from the CSV file ?
    Thanks & Regards.
    Raman Khurana.

    Hi,
    If you are using GUI_UPLOAD, you might have missed to make has_field_separator  as 'X'.
      EXPORTING
    CALL FUNCTION 'GUI_UPLOAD'
      EXPORTING
       filename                      = 'C:\File.csv'
       filetype                      = 'ASC'
       has_field_separator           = ' X'  "<= Set this as 'X'.
      TABLES
        data_tab                      = itab.
    Regards,
    Manoj Kumar P

  • Data Migration from MOSS 2007 to SharePoint 2010

    Hi,
    We have done a migration project for one of our client.
    This migration was done from MOSS 2007 to SharePoint Enterprise Server 2010.
    After the migration was completed we developed lot of new webparts, lists ,application pages in this newly migrated farm in SharePoint Enterprise Server 2010.
    However the entire project took around 1 year due to some unavoidable delays. During this one year our client was using MOSS 2007 sites for there work. hence the data in MOSS 2007 farm/sites was continuously updated.
    Our client needs to go live in next month and also needs the sites/lists/libraries which are migrated in new SP 2010 farm to be updated as per the MOSS 2007 farm.
    Kindly suggest best possible way to achieve this data migration since last 1 year from MOSS 2007 to newly migrated SharePoint Enterprise Server 2010 farm.
    Kindly note we have done lot of customization and new developments in newly created SP 2010 farm.
    Thanks and regards,
    Amien

    Hi there,
    If you have already taken care of customization on SP2010 server. Then use any of the migration tools available in market to migrate data to target SP server.
    Content Matrix gives free 25 GB to migrate your SharePoint data (https://www.metalogix.com/Resources/Promotions/Content-Matrix-Migration-Express-Download.aspx). 
    If you don't know what is changed in last 1 year, then best thing to do is add new content database and migrate your data into it.
    Hope that helps.
    Regards,
    Tapan

  • Data Migration techniques

    Hi Experts,
    I want to know about data migration techniques and how we can best use MDM while migrating old version of R/3 to new version of R/3....
    I have implemented SAP MDM in cases where we have number of SAP R/3 instances across different region and we were taking data from different R/3 one by one and doing data stan'zation,consolidation, Harmonization and all... I am not talking about all this...
    There was a good explanation from <b>Markus Ganser</b> about <b>Duplicate data and identical data</b>... I know all this..but my question is when I have <b>only one SAP R/3</b> and I still want to implement MDM solution while migrating my old R/3 instance to new one, How can I proceed in this scenario? What is the data migration technique...
    I know the common answer will be to use MDM as a middle ware..take Master data from old instance and after consolidation, send it back to new instance and at the same time send tran'data directly to new version... But is this worth doing this? Is there any other approach?
    If there is any document on this or any one have idea about data migration techniques while implementing MDM solution than send me documents on [email protected].......
    In short, I am looking for below 3 points while doing migration along with SAP MDM
    <b><b>Data Migration techniques</b>
    <b>Prerequisites</b>
    <b>Methodology in this kind of scenario</b>
    Step by step procedure</b>
    cheers,
    R.n

    hi..
    here i am sending u the link for complete Data Migration Life Cycle
    <a href="http://www.redwoodsystems.co.uk/dataMWhitePaper.html#links">http://www.redwoodsystems.co.uk/dataMWhitePaper.html#links</a>
    hope it might be of any use to u
    thank you & reward points if useful
    Message was edited by:
            Dasari Narendra

  • What kind of data migrate to CCM 5.0??

    Hi,
    I´ve read many documents about Data Migration Assistant (DMA) and i have some doubts about it. In our company each telephone user handles speed dials, services associated to telephone and own book telephone. i`m completly sure that speed dials are migrated with DMA, but services and books telephone too???
    I don´t know if when in cisco documentation says Call Manager data it´s refering to all kind of datas.
    thanks to all.

    Anyone???

  • SAP Accelarated Data migration

    Hi All,
             Could someone kindly provide more info about SAP ADM. I am uanble to get any info about the same. I would like to understand as to how exactly the tool works. I have the 4 page PDF that is posted on the site but its not clear on the actual tool. Can someone kindly provide me with some document or screen shots or any more info about it. My mail id is [email protected] Could someone kindly reply at the earliest.
    Thanks
    Prahlad

    Hi Prahlad,
    Go through this hope u can understand.
    With SAP Accelerated Data Migration, you can reduce migration
    costs by as much as 50% and avoid interruption of business
    processes. Moreover, shutting down the source system after
    migration reduces system administration costs and total cost
    of operations. In short, you realize the following benefits:
    • Significantly reduced cost and time to complete migration
    projects
    • Accurate, cost-effective data transfer applicable for any kind
    of source system
    • Better data quality because of preconfigured business objects
    that ensure data consistency
    • Improved end-user productivity and acceptance thanks to
    migration of historical data
    • Effective migration that avoids interruption of business
    processes
    • Full support services to avoid risks and ensure the optimum
    performance of your new business applications
    • Faster return on investment
    In short, a smoother, more cost-effective migration to a new
    technology solution ultimately positions your organization
    to lower your total cost of ownership, maintain competitive
    advantage, and pursue new business opportunities.
    Expertise in Action
    SAP Accelerated Data Migration applies a business object–oriented,
    two-step approach that uses a neutral interface as a staging area
    and predefined migration content for the conversion and upload
    of data. The neutral interface enables the SAP tool to generate
    predefined migration content and prevents all potential legal
    issues regarding the intellectual property of any source-system
    vendor. The whole data migration process from the source to
    the target system consists of just two steps, as follows:
    1. Data is extracted from the source system into the standard
    interface as XML files.
    2. Data migrates from the interface into the mySAP Business
    Suite database. The migration is based on a new “migration
    workbench” engine developed by SAP based on the SAP
    NetWeaver® platform. All requirements for mapping structures
    and fields and developing complex conversion rules are solved
    within this engine (see Figure 1).
    Once the migration is complete, business-unit end users have
    access to all the legacy data in the new applications as if it had
    originated there. They can continue to work on the existing
    business process items in the new applications and benefit from
    improved functionality.
    Lifting the Limitations
    Much of the cost and effort involved in classical data migrations
    are generated by migration content development, as follows:
    • Identifying business objects for migration to properly support
    the business
    • Defining the structure and field mapping for the relevant
    business objects
    • Developing conversion rules for all necessary value mapping
    Readily available migration content can simplify this effort.
    SAP Accelerated Data Migration provides preconfigured business
    content, helping you migrate it to your new system more efficiently
    and rapidly. The tool allows the migration of all types of
    data, independent of its current state within a business process.
    This includes master and dynamic data, as well as partially
    processed and historical data, to minimize data loss. Business
    processes are uninterrupted and normal operation procedures
    can be retained.
    By providing a standard, neutral interface and reading data as an
    XML file, SAP Accelerated Data Migration is applicable for any
    kind of source system. Preconfigured data migration objects built
    specifically for SAP applications significantly simplify the conversion
    from non-SAP software data into SAP software data objects,
    yielding far-reaching benefits. Besides reducing related IT costs,
    you can be certain of consistency across business-object boundaries.
    Through a direct insert into the database, you avoid the
    performance limitations of classical data migration.
    Reward points if helpful.
    Thanks

  • RDS for SAP data migration through IDOC's

    Dear Experts
      Just now we have installed BODS 4.1 and RDS also installed now. I am in the process of exploring the RDS. I found all the migration objects is available in the form of IDOC's. Now my questions are
    How to work on custom IDOC's. Do i need to extensively code on R/3 side to upload the custom fields. Like user exit or BADI for custom fileds
    If yes then i would say implementation time through RDS will remain more or less same time.
    Through BODS or by RDS migration objects is there any way without coding on ABAP R/3 side, Can we achieve migrating custom fields
      Please let me how to achieve above requirements by RDS
    Thanks
    Vijay Mukunthan

    Hi Vijay,
    You are right, the standard content of the Rapid Data Migration RDS packages brings the content with the standard IDoc interface for SAP standard only. So custom objects are not pre-built with the package. However, we incorporated the foundation to include custom coding as easy as possible:
    1) IDoc
    We are using the standard interface technology IDoc. IDocs do have an enhancement concept which is described in the online help. There are three main cases and all of them are supported, however two need additional coding:
    Z-fields: Additional fields in existing tables lead to IDoc segment extensions
    With the IDoc segment versioning concept this is easy to realize without any ABAP coding
    Z-tables: Additional IDoc segments will be necessary and lead to a new IDoc version
    With the IDoc versioning concept you can easily add Z-segments with leveraging provided user exits
    Custom business objects:
    You can even create a complete own IDoc message type in Z-namespace for own stuff, leveraging the given ALE layer of the IDoc middleware
    2) Enhancement guide
    With the RDS package we offer an enhancement guide which helps you to modify the jobs and data flows in the SAP Data Services content according to the changes in the target IDoc structure to reflect additional fields and tables. We built it as a step-by-step guide following for a sample IDoc type but will work similarly for any IDoc.
    Best regards,
    Frank Densborn.

  • Data Migration from CRM 5.0 to CRM 7.0

    Hi Friends,
    We are into a re-implementation on CRM 7.0. The old system was on CRM 5.0.
    Since this is not an upgrade, the data will not copied over from the old system into the new system.
    So, now I would like to know how to migrate data (master/transaction) from CRM 5.0 into 7.0.
    I have read that we can make use of BAPI/LSMW/IDOC/XIF adapters to push info into CRM.
    But, the customer wants all the historical data to be in the new system.
    I think master data can be handled....  but how to migrate the transaction data?.... i mean the number ranges/document flow.. and other relationships.. how can we maintain the same links into the new system?
    If it was a migration from legacy system into the SAP CRM, we could have gone ahead with new number ranges etc... but this migration is from SAP to SAP.
    Also, we will have ECC6.0 connected to CRM 7.0. ANy additional things to be taken care of when the connection is live in DEV/QA/PROd instances?
    Any pointers to this ?

    Hi Gary,
    As per my understanding, your data migration involves quite complex scenarios covering the CRM and ERP box. As the customer needs the old data to be present in the new implementation, the old data needs to be migrated from the CRM 5.0 to CRM 7.0. For the migration to happen, you need to setup the customization (like the number ranges,Condition Data Tables, etc ). Then only you can proceed with the data migration. Also, you need a proper backup plan as this involves proper backup of the data.
    Once the setup is complete, you need to configure the ECC6.0 with the CRM7.0 system. Here also, you need to do a lot of settings and performing the Initial loading of the Adapter Objects. But before you proceed with the entire operation, the entire system landscape must be properly defined with appropriate settings.
    My recomendation is to take help of SAP Consultants and data migration experts as they are the best persons to guide you on this. Also it would be great if you can go through the SAP Best Practices documents which are best repository documents to guide you on this.
    Have a look on this link for an overview of the CRM Data Migration
    http://help.sap.com/saphelp_crm60/helpdata/en/1a/023d63b8387c4a8dfea6592f3a23a7/frameset.htm
    Hope this helps.
    Thanks,
    Samantak.

  • Data migration ALE or idocs and bapis

    hi ,
    thank you guys for the support.
    i am into a data migration project.
    i need the initial setup for the idocs and bapis.
    i dont know anything about these so can u send me the process for the setup and the data migration process too with idocs and bapis.
    i didnt understand when they asked me if we can data migrate with a ALE, is that the same as using idocs and bapis or is it a different approach.
    u can send me at any inofrmation connected to my eamil addr.
    i would really appreciate that.
    thank you.

    Hi,
    Outbound:
    Step 1. Application document is created when transaction is saved.
    2. Message control is invoked.
    3. Messages are processed by system.
    4. Messages are Edited (if desired).
    5. Output (ALE / EDI) is checked
    6. Validate against Message control record from Partner Profile
    7. Application Document is saved.
    8. Entry NAST table is created for every selected output program
    along with Medium & Timing.
    9. Check for Process Immediately .
    If (yes)
    Determine Processing Program from TNAPR Table.
    ELSE
    Execute RSNASTED Program.
    10. Read Partner Profile to determine Process Code.
    11. Process Code points to the Function Module & Invoked.
    12. IDoc is generated.
    13. Check for ALE Request.
    if (Yes)
    Perform Filters, Conversions, Version Changes etc.
    Else.
    IDoc is stored in DATABASE.
    INBOUND:
    Step 1. EDI Subsystem creates an IDoc file from EDI Messages
    2. Subsystem calls Functional Module EDI_DATA_INCOMING from startRFC program.
    3. Data in Control Record is validate against the Partner Profile.
    4. IDoc is generated in Database and syntax check is carried out.
    5. IDoc file is deleted once file read.
    6. Event PROCESSSTATE REACHED is triggered in Idoc Object Workflow.
    7. Check for Process Immediately.
    If NO
    Execute RBDAPP01 Program
    Else
    Read Process Code from Partner Profile
    Process Code Points to Function Module
    Application Document Posted.
    further help:
    check url
    http://www.sappoint.com/abap/ale.pdf
    http://www.sappoint.com/abap/ale2.pdf
    http://www.sapgenie.com/ale/configuration.htm
    http://www.sappoint.com/abap/ale.pdf
    http://www.sappoint.com/abap/ale2.pdf
    http://www.sapdevelopment.co.uk/training
    And also u can get lots of inof from the below link.
    http://www.sapgenie.com/ale/why_ale.htm
    Just follow the procedure
    Sending System(Outbound ALE Process)
    Tcode SALE ? for
    a) Define Logical System
    b) Assign Client to Logical System
    Tcode SM59-RFC Destination
    Tcode BD64 ? Create Model View
    Tcode BD82 ? Generate partner Profiles & Create Ports
    Tcode BD64 ? Distribute the Model view
    Message Type MATMAS
    Tcode BD10 ? Send Material Data
    Tcode WE05 ? Idoc List for watching any Errors
    Receiving System(Inbound ALE )
    Tcode SALE ? for
    a) Define Logical System
    b) Assign Client to Logical System
    Tcode SM59-RFC Destination
    Tcode BD64 ? Check for Model view whether it has distributed or not
    Tcode BD82 -- Generate partner Profiles & Create Ports
    Tcode BD11 Getting Material Data
    Tcode WE05 ? Idoc List for inbound status codes
    ALE IDOC Steps
    Sending System(Outbound ALE Process)
    Tcode SALE ?3 for
    a) Define Logical System
    b) Assign Client to Logical System
    Tcode SM59-RFC Destination
    Tcode BD64 !V Create Model View
    Tcode BD82 !V Generate partner Profiles & Create Ports
    Tcode BD64 !V Distribute the Model view
    This is Receiving system Settings
    Receiving System(Inbound ALE )
    Tcode SALE ?3 for
    a) Define Logical System
    b) Assign Client to Logical System
    Tcode SM59-RFC Destination
    Tcode BD64 !V Check for Model view whether it has distributed or not
    Tcode BD82 -- Generate partner Profiles & Create Ports
    Tcode BD11 Getting Material Data
    Tcode WE05 !V Idoc List for inbound status codes
    Message Type MATMAS
    Tcode BD10 !V Send Material Data
    Tcode WE05 ( )
    The BAPIs Create() and CreateFromData() create an instance of an SAP business object type, for example, a purchase order. These BAPIs are class methods.
    Change( )
    The BAPI Change() changes an existing instance of an SAP business object type, for example, a purchase order. The BAPI Change () is an instance method.
    Delete( ) and Undelete( ) The BAPI Delete() deletes an instance of an SAP business object type from the database or sets a deletion flag.
    The BAPI Undelete() removes a deletion flag. These BAPIs are instance methods.
    Cancel ( ) Unlike the BAPI Delete(), the BAPI Cancel() cancels an instance of a business object type. The instance to be cancelled remains in the database and an additional instance is created and this is the one that is actually canceled. The Cancel() BAPI is an instance method.
    Add<subobject> ( ) and Remove<subobject> ( ) The BAPI Add<subobject> adds a subobject to an existing object inst! ance and the BAPI and Remove<subobject> removes a subobject from an object instance. These BAPIs are instance methods.
    BAPI-step by step
    http://www.sapgenie.com/abap/bapi/example.htm
    just refer to the link below
    http://www.sapmaterial.com/?gclid=CN322K28t4sCFQ-WbgodSGbK2g
    list of all bapis
    http://www.planetsap.com/LIST_ALL_BAPIs.htm
    for BAPI's
    http://www.sappoint.com/abap/bapiintro.pdf
    http://www.sappoint.com/abap/bapiprg.pdf
    http://www.sappoint.com/abap/bapiactx.pdf
    http://www.sappoint.com/abap/bapilst.pdf
    http://www.sappoint.com/abap/bapiexer.pdf
    http://service.sap.com/ale
    http://service.sap.com/bapi
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCMIDAPII/CABFAAPIINTRO.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/CABFABAPIREF/CABFABAPIPG.pdf
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/BCFESDE8/BCFESDE8.pdf
    http://www.planetsap.com/Bapi_main_page.htm
    http://www.topxml.com/sap/sap_idoc_xml.asp
    http://www.sapdevelopment.co.uk/
    http://www.sapdevelopment.co.uk/java/jco/bapi_jco.pdf
    Also refer to the following links..
    www.sap-img.com/bapi.htm
    www.sap-img.com/abap/bapi-conventions.htm
    www.planetsap.com/Bapi_main_page.htm
    www.sapgenie.com/abap/bapi/index.htm
    Checkout !!
    http://searchsap.techtarget.com/originalContent/0,289142,sid21_gci948835,00.html
    http://techrepublic.com.com/5100-6329-1051160.html#
    Example Code
    U need to give the step_nr, item_nr, cond_count and cond_type so the correct conditon will be updated. If no condition exists for the given parameters, a new condition will be created.
    U can find these parameters for a particular condition type in table KONV.
    *& Form saveTransactionJOCR
    text
    --> p1 text
    <-- p2 text
    FORM saveTransactionJOCR .
    data: salesdocument like BAPIVBELN-VBELN,
    order_header_inx like bapisdh1x,
    order_header_in like bapisdh1,
    return type standard table of bapiret2 with header line,
    conditions_in type standard table of bapicond with header line,
    conditions_inx type standard table of bapicondx with header line,
    logic_switch like BAPISDLS,
    step_nr like conditions_in-cond_st_no,
    item_nr like conditions_in-itm_number,
    cond_count like conditions_in-cond_count,
    cond_type like conditions_in-cond_type.
    salesdocument = wa_order_information-VBELN.
    LOGIC_SWITCH-COND_HANDL = 'X'.
    order_header_inx-updateflag = 'U'.
    conditions
    clear conditions_in[].
    clear conditions_inx[].
    clear: step_nr,
    item_nr,
    cond_count,
    cond_type.
    step_nr = '710'.
    item_nr = '000000'.
    cond_count = '01'.
    cond_type = 'ZCP2'.
    CONDITIONS_IN-ITM_NUMBER = item_nr.
    conditions_in-cond_st_no = step_nr.
    CONDITIONS_IN-COND_COUNT = cond_count.
    CONDITIONS_IN-COND_TYPE = cond_type.
    CONDITIONS_IN-COND_VALUE = 666.
    CONDITIONS_IN-CURRENCY = 'EUR'.
    append conditions_in.
    CONDITIONS_INX-ITM_NUMBER = item_nr.
    conditions_inx-cond_st_no = step_nr.
    CONDITIONS_INX-COND_COUNT = cond_count.
    CONDITIONS_INX-COND_TYPE = cond_type.
    CONDITIONS_INX-UPDATEFLAG = 'U'.
    CONDITIONS_INX-COND_VALUE = 'X'.
    CONDITIONS_INX-CURRENCY = 'X'.
    append conditions_inx.
    CALL FUNCTION 'BAPI_SALESORDER_CHANGE'
    EXPORTING
    SALESDOCUMENT = salesdocument
    ORDER_HEADER_IN = order_header_in
    ORDER_HEADER_INX = order_header_inx
    LOGIC_SWITCH = logic_switch
    TABLES
    RETURN = return
    CONDITIONS_IN = conditions_in
    CONDITIONS_INX = conditions_inx
    if return-type ne 'E'.
    commit work and wait.
    endif.
    ENDFORM. " saveTransactionJOCR
    Bdc to Bapi
    The steps to be followed are :
    1. Find out the relevant BAPI (BAPI_SALESORDER_CHANGE for VA02).
    for VA01 use BAPI_SALESORDER_CREATEFROMDAT2
    2. Create a Z program and call the BAPi (same as a Funtion module call).
    2. Now, if you see this BAPi, it has
    -> Importing structures.
    eg: SALESDOCUMENT: this will take the Sales order header data as input.
    -> Tables parameters:
    eg: ORDER_ITEM_IN: this will take the line item data as input.
    Note :
    Only specify fields that should be changed
    Select these fields by entering an X in the checkboxes
    Enter a U in the UPDATEFLAG field
    Always specify key fields when changing the data, including in the checkboxes
    The configuration is an exception here. If this needs to be changed, you need to complete it again fully.
    Maintain quantities and dates in the schedule line data
    Possible UPDATEFLAGS:
    U = change
    D = delete
    I = add
    Example
    1. Delete the whole order
    2. Delete order items
    3. Change the order
    4. Change the configuration
    Notes
    1. Minimum entry:
    You must enter the order number in the SALESDOCUMENT structure.
    You must always enter key fields for changes.
    You must always specify the update indicator in the ORDER_HEADER_INX.
    2. Commit control:
    The BAPI does not run a database Commit, which means that the application must trigger the Commit so that the changes are read to the database. To do this, use the BAPI_TRANSACTION_COMMIT BAPI.
    For further details... refer to the Function Module documentation for the BAPi.
    Bapi to VB(Visual Basic)
    Long back I had used the following flow structure to acheive the same.
    Report -> SM59 RFC destination -> COM4ABAP -> VB.exe
    my report uses the rfc destination to create a COM session with com4abap. com4abap calls the vb.exe and manages the flow of data between sap and vb exe.
    You need to have com4abap.exe
    If com4abap is installed you will find it in sapgui installatin directory , C:\Program Files\SAPpc\sapgui\RFCSDK\com4abap.
    else refer OSS note 419822 for installation of com4abap
    after making the settings in com4abap to point to the vb program and setting up rfc destination in sm59 to point to com4abap session , you can use the following function modules to call the vb code.
    for setting up com4abap and rfc destination please refer to the documentation for com4abap.
    Invoke NEW DCOM session
    call function 'BEGIN_COM_SESSION'
    exporting
    service_dest = service_dest "(this will be a RFC destination created in SM59)
    importing
    worker_dest = worker_dest
    exceptions
    connect_to_dcom_service_failed = 1
    connect_to_dcom_worker_failed = 2
    others = 3.
    call function 'create_com_instance' destination worker_dest
    exporting
    clsid = g_c_clsid
    typelib = g_c_typelib
    importing
    instid = g_f_oid
    exceptions
    communication_failure = 1 message g_f_msg
    system_failure = 2 message g_f_msg
    invalid_instance_id = 3
    others = 4.
    call function 'com_invoke' destination worker_dest
    exporting
    %instid = g_f_oid
    %method = 'UpdatePDF'
    sntemp = g_v_const_filent
    snsysid = sy-sysid
    snflag = 'N'
    tables
    rssaptable = g_t_pdfdetail1
    %return = g_t_pdfdetail1 "t_test
    exceptions
    communication_failure = 1 message g_f_msg
    system_failure = 2 message g_f_msg
    invalid_instance_id = 3
    others = 4.
    then close the com session , using
    FM delete_com_instance
    FM END_COM_SESSION
    Thanks and regards,
    Sarada

  • Delta data migration on SAP upgread from 4.6C to ECC 6.0

    Hi,
    We are planning for SAP upgrade in our Production environment and we are facing the below listed constraints and need your advice on the options to peform ECC Upgrade.
    1) Present HW resource is already facing bottleneck and so no additional load can be added.
    2) We have a completely new HW environment due to which we are unable to perform the test upgrade.
    Also, is there anyway to perform the Prepare Activities on new hardware (copy of Production system) first and then do the delta data migration for update of transactional data from present production system to new system (Prepare Activities done on this system) without affecting the Prepare Activities data.
    Thanks & Regards,
    Pratap

    Hi Pratap,
    You can first make a copy of you production system on new hardware by homogeneous system copy. Then do the test upgrade if everything will be fine. Then as suggested earlier migrate your production to new hardware and then do the upgrade.
    Thanks
    Sunny

  • SAP PS Data Migration Project

    Hi,
    This is my first data migration project. We have to migrate SAP PS transactional data from SAP 4.7 to SAP 6.0. My task is to migrate WBS element data to the new system. To do that I need to do field and Value mapping also. Please suggest how do we go about completing the task at hand and also suggest what else do we migrate in SAP PS.

    Hi,
    If it is a new system, you can use BDC/LSMW/BAPI for uploading the data to PS.
    Also topic have been already discussed in the forum. You can refer below threads for more info:
    http://scn.sap.com/thread/912834
    http://scn.sap.com/thread/3236416
    Regards
    Shishir

  • Treasury Data Migration

    Friends,
    One of our client has decided to use SAP Treasury and Risk Management. I have been asked to lead data migration within this area. For the same, I would like to know the generally used and needed master and transaction data for SAP Treasury Management. Also, let me know the various ways of migrating data into SAP TRM eg. BAPI's, Transaction Manager, LSMW etc.  
    Regards,
    Amey

    Hello Amey,
    Please find list of steps which might be helpful for data migration.
    Legacy Data Transfer
    1. Legacy data
    transfer always happens on a key date.
    2. The procedure is
    different for securities and OTC because they deal differently with positions
    on a key date.
    3. For securities, we
    simply need to copy the positions with all values as of the key date.
    4. For OTC, the
    entire transaction is copied as there is no explicit position management in the
    transaction manager.
    Customizing the Legacy Data Transfer:
    You primarily need
    customizing for the update types that the flows of the legacy data transfer
    business transactions are to carry. We can find the settings under Treasury and
    Risk Management à Transaction Management à General Settings à Tools à Legacy
    Data Transfer à Flow Data à Update Types.
    The update types are
    divided into those for valuation area independent & dependent. The
    independent one only relates to securities. à While the legacy data transfer is
    usually not intended to update the general ledger, nevertheless you must mark
    most of the update types as posting relevant and maintain an account
    determination in customizing. This is necessary to ensure that the data in the position
    management is complete and in particular the account assignment reference
    transfers can be performed correctly.
    Legacy Data Transfer for OTC:
    Step
    1: Existing OTC transactions enter the same way as new transactions.
    Enter either manually (or) use existing BAPIs to import them. BAPIs list à Use
    BAPI_FTR_FD_CREATE & BAPI_FTR_FXT_CREATE
    for creating FD & FOREX transactions respectively. Similarly there are many
    BAPIs for each product and we can get the list of BAPIs in BAPI transaction code
    under the node Financials à Financial Supply Chain management. Use FTR_BAPI
    Transaction to test run the Treasury BAPIs.We may define separate transaction
    type for legacy transactions so that settlement (or) posting release is not
    necessary.
    Step 2: Now we must change status of all the historical flows
    created in Step 1 to POSTED without actually generating accounting document. To do this we will use transaction TBB1_LC.
    This is similar to transaction TBB1. Unlike TBB1, TBB1_LC affects all the valuation areas without giving an option to affect only operational area. We can see the posted flows by TBB1_LC in posting journal (TPM20) without any reference to accounting documents.
    Step 3: Till now the historical flows do not yet explain the book value of OTC transactions. Hence we must first enter the historical valuation results. This is done using TPM63C which you can find under path Treasury and Risk Management à Transaction Management à General.Settings à Tools à Legacy
    Data Transfer à Flow Data à Enter Valuation Area Dependant Data for MM.After entering the
    valuation results use TPM63 to transfer values to Internal Position Management.
    Legacy Data Transfer for Futures & Listed Options:
    Step 1: Similar to OTC
    Step 2: Similar to OTC
    Step 3: Use transaction TPM63D (Valuation area dependant data for futures). We must establish the relationship between lot & the transaction on which it is based via the transaction number.
    Legacy Data Transfer for Securities:
    For securities position, it does not matter whether is arose in the past through a single purchase (or) a series of business transactions. The legacy data transfer thus seeks to build up the positions through One Business Transaction on the transfer key date.
    Step 1: Before we enter business transactions, we must first create the master data of security position i.e. their Position Indicator. I) Use TPM61A (or) menu path Treasury and Risk Management Transaction
    Management à General Settings  Tools à Legacy Data Transfer à Position Data Enter Position Information for Securities
    II) Now create the position indicator using TPM61.
    Step 2: Now we will enter the business transaction.
      I) Valuation area independent data for securities à For External Position
    II) Valuation area dependant data for securities à For Internal Position Nevertheless, we need to use transaction set the effects of update type on position to change the positions. If you want to enter LOT Positions with the transactions of legacy data transfer, use the Item Number. Ensure that business transaction generating the position and the one for valuation results carry the same item number.
    If you have built up
    lots via normal procedures, you must specify the corresponding transaction number
    when entering the valuation area dependant business transaction in the
    transaction column.
    Step 3: Now generate the business transactions using TPM63.We can initially build only valuation area independent data (external position) by checking “only valuation area independent data” button in TPM63. The external position can be checked in securities account cash flow i.e. external securities position in TPM40.
    Also we created LSMW for each product.
    Regards,
    Jain

  • LSMW data migration porcess from the beginning

    Hi gurus,
    I do understand that there are 14 steps to be completed during migration, but can anyone give me a basic idea as what needs to be done prior to the above steps when starting a data migration project, as i have been only involved in the LSMW(14 steps) process before.
    Thanks in advance, points will be awarded for all answers.
    Ganga

    Hello
    As an intro:
    Classical data migrations are typically limited to master data and
    tend to cause multiple negative impacts on workflow and oper-
    ating business units.
    To minimize costs, risks, and effort during a data migration and ensure a smooth changeover for operational business processs items, and better reporting
    capabilities SAP recently come with a Data Migration Solution called SAP Accelerated Data Migration tool.
    This advanced tool, coupled with support from the SAP Services organization and SAP partners, lets you easily migrate and integrate application data from any legacy system into SAP applications – and reduce migration costs significantly.
    Data migration is done in 4th phase i.e. the final preparation stage of ASAP methodology.
    As far as LSMW is concerned, BDC and LSMW both are use for loading data into SAP system. The difference is ,LSMW is loading tool by SAP where you can use various methods to upload, massage data or cleansed your data while loading into SAP from legacy system. The method can be BDC, BAPI, Idocs or SAP ABAP codes. Usually in BDC the user has to write an ABAP program or use other utility to load legacy data.
    LSMW has been developed mainly for the Functional people who has no knowledge of the technical aspects of report or BDC in ABAP because its very user friendly and handy to use. So, simply they can upload the master records using LSMW.
    Reg
    assign points if useful

Maybe you are looking for

  • Excel_reference

    report ZUPLOADEXCELSHEETS. INCLUDE ole2incl . tables : vbap, vbak. data: begin of itab occurs 0,       posnr like vbap-posnr,       vbeln like vbak-vbeln,       end of itab. select vbeln        posnr        from vbap        up to 20 rows        into

  • Starting over with G4

    Trying to figure out how to make my "cluttered" up G4 "new" again. We just got a new iMac and want to use the G4 for the kids. I've (somehow!) set up a network with an Airport extreme and express so it's connected to the internet. Now I want to wipe

  • Routing -operation

    Dear  pp gurus, In routing  , I have defined the operations . Although i know that in operation the set up , machine time, labour time is define, which helps in costing. but wat exactly is  set up, machine  and labour time? If suppose  10 people work

  • How do I find out how many people have seen my video?

    Wondering if it's possible to see some stats on my video? Views, uniques, average watch time, etc?

  • System Extension Issue

    I keep getting this message on my IMac 7.1 it says:  The System extension "/system/Library/Extensions/AppleUSBEthernetHost.kext" was installed improperly and cannot be used. Please try reinstalling it or contact the product's vendor for an update. I