Data Migration - Data Models Dependency

Hi,
In major ERP data migration and transformation project from legacy to SAP or from SAP to SAP, one should finalize the sequence in which data should be migrated. e.g. Vendor data cannot be migrated before customer data, or tax data should be uploaded first then customer data .. etc. Is there available any best practice or reference for such dependencies ? Primary models and secondary models which depend on primary models ?
Please share such content if you have, I'm sure it will be helpful to many.

The best way is to use iTunes. Back up your old iPad. Unplug it, plug your new one in and go through the set up process. When you get the question 'set up as new or restore from backup' choose to restore from the backup you just made. Your app data and information will be put onto your new iPad.
You may need to redownload apps, but since the app data is already there, you should not lose progress on games or the such.
http://support.apple.com/en-us/HT2109
has directions. You can also use iCloud backup, but I personally prefer iTunes since it takes the internet out of the equation and things are done locally (beyond redownloading apps if needed)

Similar Messages

  • ASAP Data Migration / Data Conversion / Data Transfer Templates

    Hi,
    I'm looking for the ASAP Data Migration / Data Conversion / Data Transfer Templates that used to be available in ASAP.
    Some time I ago, and I believe it was on this site and can no longer find it, I found a zip file which contained the above .xls template to manage all the data conversions on a project. Also included was the data mapping template .xls
    Many thanks.

    Hi,
    I'm looking for the ASAP Data Migration / Data Conversion / Data Transfer Templates that used to be available in ASAP.
    Some time I ago, and I believe it was on this site and can no longer find it, I found a zip file which contained the above .xls template to manage all the data conversions on a project. Also included was the data mapping template .xls
    Many thanks.

  • HELP ME ON DATA MIGRATION PLEASE

    Hi
    what is SAP data migration what kind of tools and techniques include in that and also please what are the roles and responsibilities include in that when we are working on project.(RELATED TO ABAP)

    Data Migration – Current Situation
    Data migration or data conversion is a persistent issue facing most organizations who have implemented SAP. Whether it is during the final phases of a new SAP implementation, during SAP upgrades and updates, during corporate restructurings, or during mergers and acquisitions, data migration continues to remain a challenging problem for IT.
    Most data migration projects are faced with a time and budget crunch. But, these projects usually require extensive support from programmers and other technical experts and therefore data migration projects tend to get expensive and time-consuming.
    The tools for data transfer and data migration provided by SAP are very technical and have a steep learning curve associated with them and they require technical experts to build and implement the data migration scripts.
    Data Migration – Ideal Scenario
    An ideal data migration scenario would let the data migration project be implemented by the end user departments themselves. A few super-users within these end-user departments that supply the data should have the ability to transfer and migrate data themselves without relying on technical experts. Such a scenario would significantly cut the time and effort required in a data migration project.
    Reaching such an ideal scenario would require data migration tools that are easy to learn and require no programming. Furthermore, these tools should work across all the different SAP modules and the SAP products, including the different versions of SAP.
    Data Migration
    Data Migration is the process of moving required (and most often very large) volumes of data from our clients’ existing systems to new systems. Existing systems can be anything from custom-built IT infrastructures to spreadsheets and standalone databases.
    Data Migration encompasses all the necessary steps to cleanse, correct and move the data into the new system *. Technological changes, change in providers, software updates or data warehousing / data mining projects make such delicate and critical operations necessary.
    <b>Data Migration can be done using BDC, LSMW, BAPI, Idoc's.</b>
    What is the difference between batch input and call transaction in BDC?
    Session method.
    1) synchronous processing.
    2) can tranfer large amount of data.
    3) processing is slower.
    4) error log is created
    5) data is not updated until session is processed.
    Call transaction.
    1) asynchronous processing
    2) can transfer small amount of data
    3) processing is faster.
    4) errors need to be handled explicitly
    5) data is updated automatically
    Differences between bdc session method and call transaction method.
    The most important aspects of the batch session interface are: - Asynchronous processing - Transfers data for multiple transactions - Synchronous database update During processing, no transaction is started until the previous transaction has been written to the database. - A batch input processing log is generated for each session - Sessions cannot be generated in parallel
    The most important aspects of the CALL TRANSACTION USING interface are: - Synchronous processing - Transfers data for a single transaction - Synchronous and asynchronous database updating both possible The program specifies which kind of updating is desired. - Separate LUW for the transaction The system performs a database commit immediately before and after the CALL TRANSACTION USING statement. - No batch input processing log is generated.
    Refer this
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    Check these link:
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://www.sap-img.com/abap/question-about-bdc-program.htm
    http://www.itcserver.com/blog/2006/06/30/batch-input-vs-call-transaction/
    http://www.planetsap.com/bdc_main_page.htm
    call Transaction or session method ?
    http://www.sapbrain.com/FAQs/TECHNICAL/SAP_ABAP_DATADICTIONARY_FAQ.html
    http://www.****************/InterviewQ/interviewQ.htm
    http://help.sap.com/saphelp_46c/helpdata/en/35/2cd77bd7705394e10000009b387c12/frameset.htm
    ALE/ IDOC
    http://help.sap.com/saphelp_erp2004/helpdata/en/dc/6b835943d711d1893e0000e8323c4f/content.htm
    http://www.sapgenie.com/sapgenie/docs/ale_scenario_development_procedure.doc
    http://edocs.bea.com/elink/adapter/r3/userhtm/ale.htm#1008419
    http://www.netweaverguru.com/EDI/HTML/IDocBook.htm
    http://www.sapgenie.com/sapedi/index.htm
    http://www.sappoint.com/abap/ale.pdf
    http://www.sappoint.com/abap/ale2.pdf
    http://www.sapgenie.com/sapedi/idoc_abap.htm
    http://help.sap.com/saphelp_erp2005/helpdata/en/0b/2a60bb507d11d18ee90000e8366fc2/frameset.htm
    http://help.sap.com/saphelp_erp2005/helpdata/en/78/217da751ce11d189570000e829fbbd/frameset.htm
    http://www.allsaplinks.com/idoc_sample.html
    http://www.sappoint.com/abap.html
    http://help.sap.com/saphelp_erp2004/helpdata/en/dc/6b835943d711d1893e0000e8323c4f/content.htm
    http://www.sapgenie.com/sapgenie/docs/ale_scenario_development_procedure.doc
    http://edocs.bea.com/elink/adapter/r3/userhtm/ale.htm#1008419
    http://www.netweaverguru.com/EDI/HTML/IDocBook.htm
    http://www.sapgenie.com/sapedi/index.htm
    http://www.allsaplinks.com/idoc_sample.html
    Check these step-by-step links
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/ccab6730-0501-0010-ee84-de050a6cc287
    https://sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/8fd773b3-0301-0010-eabe-82149bcc292e
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/3c5d9ae3-0501-0010-0090-bdfb2d458985
    regards,
    srinivas
    <b>*reward for useful answers*</b>

  • Alternative tools for Data Migration

    Hi,
    I am looking for any latest tool developed by SAP for Data migration activities other than LSMW.
    The requirements in my project include:
    Data Migration / Data Conversion covering :
                   Data cleansing / data quality check
                   Extraction, Transformation, Load
                   Reconciliation
    Information Lifecycle Management covering :
                   Data prevention
                   Data aggregation
                   Data deletion
                   Data archiving
    BAU data maintenance covering :
                   Master data maintenance
                   Test data maintenance
                   Data quality
                   Interface data journey
    Can some one let me know if there is any SAP product catering  to the above
    Thanks

    Did you find anything on this?

  • Data Migration for Open Purchase Order

    Hi, All,
    Is there anyone know how to Count the volume for Open Purchase Order. What's the normal strategy for the Data Migration and Cut-over stage?
    My client want to know how many Open Purchase Order in the legacy system and then determine manual or automatic data migration. If manual, how to do? If automatic, how to do? Because all materials and vendors, plants are different number. How to track? How to find out to match between new and old?
    Thank you very much

    JC,
    Sounds a bit early to be making decisions about the realization phase.  It doesn't sound like you have finished the Blueprinting phase yet, much less the testing phase.
    Anyhow, in my experience I typically use LSMW (Legacy system migration workbench) to load MM master data (material masters), Inventory (WIP, RM, FG, etc) Purchasing Master data (Vendors, Purchase Info Records, Source Lists, Quota Arrangements), and Purchasing transactional documents (POs, PurReqs, Scheduling Agreements, etc).  Depending on the complexity and volume of data, it  may be necessary to write custom programs to load the data.  You will find this out during your requirements gathering.
    It is uncommon but possible to load all of these data manually.  I have never run across a client that wants to pay a consultant's hourly rate to sit at a terminal to peck away loading master data, so if the client intends to have his own users enter the data manually, the project manager should make provision that there will be qualified TRAINED client employees available for this data entry.  I did help with a portion of a conversion once manually; of Sales Credits, but there were only about 30 SD docs to load.   I did this the evening before go-live day, while I was waiting for some of my LSMW projects to complete in the background.
    A good opportunity to 'practice' your data loads is right after you have completed your development and customization, and you have gotten the approval from the client to proceed from the pilot build to the full test environment.  Once you have moved your workbench and customization into the client's test environment, but before integration testing, you can mass load all, or a substantial portion of your conversion data into the qual system.  You can treat it like a dry run for go-live, and fine tune your processes, as well as your LSMW projects.
    Yes, it is good practice to generate comparisons between legacy and SAP even if the client doesn't ask for it. For Purchase orders on the SAP side, you could use any of the standard SAP Purchasing reports, such as ME2W, ME2M, ME2C, ME2L, ME2N.  If these reports do not meet the requirements of the client, you could write a query to display the loaded data, or have an ABAPer write a custom report.
    You didn't ask, but you should also do comparisons of ALL loaded data - including master data.
    It sounds like you are implying that the client wants YOU to extract the legacy data.  For an SAP consultant, this is not very realistic (unless the legacy system is another SAP system).  Most of us do not understand the workings of the myriad legacy systems.  The client is usually expected to produce one or more legacy system technical experts for you to liase with.  You normally negotiate with the technical expert about every facet of of the data migration.  In addition, you will liase with business users, who will help you and the implementation team to logically validate that the final solution (turnkey SAP production system, fully loaded with data) will meet the client's business needs.
    Finally, you mentioned how do you track the mapping of master data between legacy and SAP.  There are many ways to do this.  I normally try to get the legacy person do the conversion on his end, eg, when he gives you the load file, you would like to have already translated the master data and inserted the SAP relevant values into the file.  If this is not possible, I usually use MS Access databases to maintain a master map, and I perform the mapping on a PC.  If your data package is small, you can probably get by using MS Excel or similar.
    Good Luck,
    DB49

  • Data Migration from Microsoft Access to Oracle 9i

    Dear All
    This is my first request in your forum,My problem is that I have installed SQL developer but unable to migrate my MDB (Microsoft access file) into Oracle 9i,I have done the following activities in order to migrate data.
    open sql developer
    under migration,click quick migration
    quick migration wizard diaglog box appear,click new,
    in the oracle tab, input the following information
    input connection name 'zuhair',give user name and password (sys/s),check save password option,give role "sysdba" ,give connection type "basic",didnt check any any check box like os authorization and others,give host name 'zkhan-dfml',port '1521',sid 'dba' and then push the test button and then status show "success"
    in the access tab, input the following information
    input connection name ,input mdb file name by using browse button,after this step click test button then the following message display 'test fail odbc microsoft access driver not a valid password
    I have tried to create odbc link but i think i have done some thing wrong,kindly help
    Regards
    Zuhair

    Hi Zuhair,
    If you give details of the problems you are still having then we should be able to get SQL*Developer to work.
    However, if you don't want to use SQL*Developer then the following options are taken from note 393760.1 available in Metalink if you have access to that -
    1. Use a gateway that supports the non-Oracle database or generic connectivity if a third party ODBC driver is available to do a -
    'create Oracle_table as select from non_oracle_table@database_link'
    for each table. You would then have to manually create all the other objects (views, indexes, stored procedurs etc) and permissions etc that existed in the non-Oracle database.
    2. Use the SQL*Plus COPY command to copy data from the non-Oracle database to the
    Oracle database using a gateway or generic connectivity. The syntax is as follows:
    COPY FROM username/password@oracle_sid -
    INSERT destination_oracle_table -
    USING query;
    The following example selects all rows from the EMP table in the non-Oracle database and
    inserts them into the local Oracle EMP table:
    COPY FROM SCOTT/TIGER@ora10 -
    INSERT EMP -
    USING SELECT * FROM SCOTT.EMP@gateway;
    where "gateway" is the database link created for the gateway.
    3. Load the non-Oracle data into flat files and use SQL*Loader to put the data into Oracle tables. Again manual
    work is required for the other objects.
    4. Contact the Oracle Product Technology Services (PTS) group for assistance. See the webpage -
    http://www.oracle.com/technology/tech/migration//workbench/index_sqldev_omwb.html
    5. Use a third party migration tool. See -
    http://www.oracle.com/technology/tech/migration/isv/third_party.html
    6. Use JDBC to connect to the non-Oracle database and Java in the Oracle database. This method will work but it is not suppported by the JDBC team.
    =========
    The amount of work involved will depend on the number and size of Access tables that you want to migrate.
    Regards,
    Mike

  • Firewire to migrate data from MacBook Pro to imac?

    Would someone guide me as to what kind of firewire I need to buy to migrate data from my MacBook Pro (which is two years old, OS X 10.6.8) to an iMac (which is brand new, but also two years old, it's the older model with a 20" screen).  I've never even heard of firewire until I turned on the iMac for the first time, so would be very appreciative of guidance.  Thanks!

    Note if any of the data came from a PowerPC Mac originally, read this tip first:
    https://discussions.apple.com/docs/DOC-2295
    As that older data if it hasn't been through the Migration Assistant can give the Assistant a big headache.

  • Data migration for Securities

    Hello
    I have been going the suggestions on the  thread below,
    Data Migration Strategy for Securities
    I would like to clarify at which price do i  book securities on cut off date, here is an example
    my cut off date is 31/12/2009 and i have the following deals
    1 .   Purchase of R207 bond (04I) nominal 10million on 06/06/2009 price 98%
      2.      Sale of R207 bond   5million on 11/11/2009 price 97%
    My balance nominal as at cut off on R207 bond is 5Million so on the 31/12/2009 i will book on TS01 5 MILLION on what price, currently on QA  we booked all deals at original prices including all sales. Note that we do MTM on 04I and amortisation on  04X bonds.
    Thank you in advance
    Regards
    Victor
    Data Migration Strategy for Securities

    Hi Victor,
    looking into your example, it seems you have outstanding balance of 5 mio of certain bond and result (P&L) of difference between purchase (10 mio) and sales (5mio) should be booked in Y2009 and is already included in financial figures in Legacy system. If you say that you use plain Mart-to-Market Valuation for this particular type bond, so the Balance value is dependant on Market price on last date of the Y2009.
    Rgds,
    Renatas

  • Migration Date when migrating from classic to new g/l

    Hello gurus,
    I am planning a migration from classic to new g/l, but this fiscal year it will be not possible to finish phase 0 (from migration model plan). Is it possible to do it in the middle of the new fiscal year, what are the withdrawls or issues? what do you recommend me? What strategy can I use?
    Thanks,
    Silvia Guillen

    Hi Silvia,
    There is differnce in migration date & activation date. Migration date will always be the firts day of new fiscal year.
    But as on migration date you have to bring your config changes to PRD.
    For details please check the following link -
    https://websmp130.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=812919
    https://websmp230.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=1070629
    Regards,

  • Please help  ! Can I cut-off New G/L migration date during fiscal year ?

    Dear Experts
    I know that the principles of migration date of New G/L migration should be on first day of fiscal year and Year-end closing of the old fiscal year must be complete before New General Ledger Accounting is activated.
    However, if I need, can I cut-off new G/L migration date during fiscal year before year-end closing ? And Which points should be concern.
    Best regards,
    Suphawat

    Hi,
    First you should have installed the cockpit., you should have completed the testing and migrated the transports before the end of the Fiscal year which is 31st March in your case.  Depending on the scenario, this will take 20 - 40 days.
    You can begin the migration once your year end activity is performed.
    If you have not installed the cockpit and completed the configuration and tested it.. forget it for this year, do it next year.
    Hope this is clear now.
    Edited by: Jayaraman N on Apr 7, 2011 10:41 AM

  • Can I cut-off New G/L migration date during fiscal year ?

    Dear Experts
        I know that the principles of migration date of New G/L migration should be on first day of fiscal year and Year-end closing of the old fiscal year must be complete before New General Ledger Accounting is activated.
        However, if I need, can I cut-off new G/L migration date during fiscal year before year-end closing ? And Which points should be concern.  
    Best regards,
    Suphawat
    Moderator: Please, search SDN

    Hi,
    First you should have installed the cockpit., you should have completed the testing and migrated the transports before the end of the Fiscal year which is 31st March in your case.  Depending on the scenario, this will take 20 - 40 days.
    You can begin the migration once your year end activity is performed.
    If you have not installed the cockpit and completed the configuration and tested it.. forget it for this year, do it next year.
    Hope this is clear now.
    Edited by: Jayaraman N on Apr 7, 2011 10:41 AM

  • Clean install OSX+S 10.9 then migrate data from 10.6 Server?

    Howdy All,
    I've had OSX Server from the first release but stopped at Snow Leopard Server waiting for the "new server" to mature a bit.  I think it is time now to move to OSX Server 10.9 but also for a clean install to clear out the crud that has collected over all these years. 
    I have two main partitions: one for the OS with the Mail store and Wikis and another for user home directories and other files.  Again this is going from Snow Leopard Server 10.6.last to 10.9 Mavericks and then Mavericks Server
    My question is:  If I clean install Maverick and then Mavericks server on another partition will I be able to migrate data and services over manually? 
    Of course, I will need to setup users and directory services again from scratch (but that is part of the plan to clean out the unused stuff that has collected over the years).  I don't mind setting up fileshares, Web sites, DNS, and similar again.
    However, when this is done will I be able to copy the mail store (i.e. all users email) over and will it be detected, updated if needed, and work?  Similarly, when I copy over the wikis, will they be detected, updated if needed, and work?
    Further, when I setup the symbolic links for home directories again (onto the other partition), will the home directories be updated (if needed) and work?  I'm not sure what could have changed with regards to home directories but you never know.
    In summary, I guess my question is about what migration only happens when you update as opposed to what happens when an older version of a resource or service is detected after an update.  Hopefully there is more of the latter than the former. 
    Thanks in advance for any thoughts on this.
    Cheers,
    Ashley.

    Out of your services (AFP/SMB, Mail, OD, SWU, and web), the one with the most potential for disaster and headache is clearly mail.  If you are using the same host device, cutting services over in pieces will not be possible.  However, here are some suggestions and potential points of concern.
    AFP/SMB file services are cake.  The only thing you need to consider is the potential time to copy the data if you are moving it to new disks.  The other issue will be user's GUID values and the associated ACLs.  Let's take the following scenario based on what you've detailed.
    • You have data on /HDD/Shares/ and you are planning on moving them to the SSD.  Is the SSD drive large enough to accept this data? 
    • If you had a share /HDD/Shares/Data and this contained an ACL allowing the design group to have access, the design group from the 10.6.8 OD may have a different GUID than the one you create on the 10.9.1 system.  If this is the case, you can purge all ACLs with a sudo chmod -R -N /path/to/data.  (Server.app should remove and then add but older versions resulted in merged messes so I go nuclear on the old settings)  Then you can apply your new ACLs and allow the permissions to propagate.
    • If you are leaving the data were it is, you will simply need to reset permissions.  However, note that if you are exporting and then importing users (via an OD backup or via standard record format) then you are maintaining GUID and should not need to touch any permissions.
    Regarding SWU, I would suggest looking into Caching server.  If you are moving the entire environment to 10.9 and iOS 7, SWU is no longer needed.  Caching server is easy as pie, requires no client configuration, and is more economical on your internet connection and server storage requirements.
    Web is pretty easy also.  But, this is dependant on what you are doing with web.  If html/php/perl then you pretty much just move your site folders and you are up and running.  If you were using MySQL, note that Apple replaced it with Postgres.  You can either perform a conversion from MySQL to Postgres or you can just install MySQL again manually.  The choice is yours.  If you are not doing database backed sites, the migration should be cake.
    OD is one of those technologies that I always prefer to start clean.  In really large environments, this can be very tough due to passwords.  You can export an OD backup from 10.6 and attempt a restore in 10.9.1.  If you have a lot of MCX in 10.6.8, you may run into some trouble as Apple has deprecated MCX in 10.8 and above.  However, this ensures that you have everything, from password to GUID.  Test, test, and test some more if you go this route.  An alternate option, especially if you are embracing the move away from MCX and to Profiles, is to do a user and group export for 10.6's Workgroup Manager.  This will not provide passwords but it will provide editable text files of your account data.  You can strip out the MXC and other legacy values and then use the resulting file to import users into a clean 10.9.1 OD master.  Once again, you will not get passwords unless you add them to the import file.  You need to figure out how many accounts and how sensitive users are to password resets.
    The final piece is mail.  This is the one area I have very little experience.  I've been burned by Apple's mails solutions from way back in the AppleShare IP days and now make it policy to use anything else but Apple's mail solution.  In a perfect world, moving the mail data store to the new OS and triggering Server.app should be enough.  But Apple + mail never seem to enter the realm of a perfect world.
    And finally, make sure DNS is correct before you do anything.  Since you are dealing with mail, you should also shut firewall port forwards to prevent new mail from coming into the server while you work on the migration.  Nothing worse than stitching mail together after a blown migration attempt.
    R-
    Apple Consultants Network
    Apple Professional Services
    Author "Mavericks Server – Foundation Services" :: Exclusively available in the Apple iBooks Store

  • Using export/import to migrate data from 8i to 9i

    We are trying to migrate all data from 8i database to 9i database. We plan to migrate the data using export/import utility so that we can have the current 8i database intact. And also the 8i and 9i database will reside on the same machine. Our 8i database size is around 300GB.
    We plan to follow below steps :
    Export data from 8i
    Install 9i
    Create tablespaces
    Create schema and tables
    create user (user used for exporting data)
    Import data in 9i
    Please let me know if below par file is correct for the export :
    BUFFER=560000
    COMPRESS=y
    CONSISTENT=y
    CONSTRAINTS=y
    DIRECT=y
    FEEDBACK=1000
    FILE=dat1.dmp, dat2.dmp, dat3.dmp (more filenames here)
    FILESIZE=2048GB
    FULL=y
    GRANTS=y
    INDEXES=y
    LOG=export.log
    OBJECT_CONSISTENT=y
    PARFILE=exp.par
    ROWS=y
    STATISTICS=ESTIMATE
    TRIGGERS=y
    TTS_FULL_CHECK=TRUE
    Thanks,
    Vinod Bhansali

    I recommend you to change some parameters and remove
    others:
    BUFFER=560000
    COMPRESS=y -- This will increase better storage
    structure ( It is good )
    CONSISTENT=y
    CONSTRAINTS=y
    DIRECT=n -- if you set that parameter in yes you
    can have problems with some objects
    FEEDBACK=1000
    FILE=dat1.dmp, dat2.dmp, dat3.dmp (more filenames here)
    FILESIZE=2048GB
    FULL=y
    GRANTS=y -- this value is the default ( It is
    not necesary )
    INDEXES=y
    LOG=export.log
    OBJECT_CONSISTENT=y -- ( start the database in restrict
    mode and do not set this param )
    PARFILE=exp.par
    ROWS=y
    STATISTICS=ESTIMATE -- this value is the default ( It is
    not necesary )
    TRIGGERS=y -- this value is the default ( It is
    not necesary )
    TTS_FULL_CHECK=TRUE
    you can see what parameters are not needed if you apply
    this command:
    [oracle@ozawa oracle]$ exp help=y
    Export: Release 9.2.0.1.0 - Production on Sun Dec 28 16:37:37 2003
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    You can let Export prompt you for parameters by entering the EXP
    command followed by your username/password:
    Example: EXP SCOTT/TIGER
    Or, you can control how Export runs by entering the EXP command followed
    by various arguments. To specify parameters, you use keywords:
    Format: EXP KEYWORD=value or KEYWORD=(value1,value2,...,valueN)
    Example: EXP SCOTT/TIGER GRANTS=Y TABLES=(EMP,DEPT,MGR)
    or TABLES=(T1:P1,T1:P2), if T1 is partitioned table
    USERID must be the first parameter on the command line.
    Keyword Description (Default) Keyword Description (Default)
    USERID username/password FULL export entire file (N)
    BUFFER size of data buffer OWNER list of owner usernames
    FILE output files (EXPDAT.DMP) TABLES list of table names
    COMPRESS import into one extent (Y) RECORDLENGTH length of IO record
    GRANTS export grants (Y) INCTYPE incremental export type
    INDEXES export indexes (Y) RECORD track incr. export (Y)
    DIRECT direct path (N) TRIGGERS export triggers (Y)
    LOG log file of screen output STATISTICS analyze objects (ESTIMATE)
    ROWS export data rows (Y) PARFILE parameter filename
    CONSISTENT cross-table consistency(N) CONSTRAINTS export constraints (Y)
    OBJECT_CONSISTENT transaction set to read only during object export (N)
    FEEDBACK display progress every x rows (0)
    FILESIZE maximum size of each dump file
    FLASHBACK_SCN SCN used to set session snapshot back to
    FLASHBACK_TIME time used to get the SCN closest to the specified time
    QUERY select clause used to export a subset of a table
    RESUMABLE suspend when a space related error is encountered(N)
    RESUMABLE_NAME text string used to identify resumable statement
    RESUMABLE_TIMEOUT wait time for RESUMABLE
    TTS_FULL_CHECK perform full or partial dependency check for TTS
    VOLSIZE number of bytes to write to each tape volume
    TABLESPACES list of tablespaces to export
    TRANSPORT_TABLESPACE export transportable tablespace metadata (N)
    TEMPLATE template name which invokes iAS mode export
    Export terminated successfully without warnings.
    [oracle@ozawa oracle]$
    Joel P�rez

  • Data Migration of Excise Duty payables at initial upload

    Hi!
    Can any one tell me, how to upload initial Excise Duty payable in data migration procedure.
    for example if different excise duty payables like ED, CESS, SEC.CESS for normal goods and capital goods.
    thanks in advances.
    regs,
    ramesh b

    Hi Ramesh,
    From my exposure with indian company codes these are just another GL account to be migrated since there are
    no dependies you can do the mapping from legacy to the new sap system the GL account.
    Just my view
    regards
    pbb

  • Data migration methods

    Hi experts. I am working on a study for a data migration project where data from multiple legacy systems would be moved into SAP. Apart from LSMW and BDC, what are the other useful methods when the data migration involves huge volumes of upload data?

    Hi,
    Obviously the answer depends to a certain extent on the circumstances of your project (complexity of data, volume, etc) and on the object in question, but here are a few rules of thumb.
    If you are new to this I would stay away from IDOCs and BAPIs as the set up is often a little complicated and if there are errors the problem is often not too easy to identify. And IDOC processing is not always as quick as people assume. For example, the customer master IDOC is processed in the background by a call transaction of XD01.
    For you I would suggest definitely using LSMW, and where possible within LSMW use the standard uploads. There will be situations where the standard upload does not cover everything (again using customers as an example, it will not include the extended address data or the long texts) but I would still use the standard upload to load as much as you can, and then write a couple of simple update BDC recordings (XD02, XK02, etc) to add what's missing.
    Some examples of the standard upload programs are:
    Customers - RFBIDE00
    Vendors - RFBIKE00
    Materials - RMDATIND
    All finance transactions - RFBIBL00
    Fixed assets - RAALTD01
    The standard uploads are normally well documented; they often use BDC (not RMDATIND) so are quite transparent; and they are generally pretty flexible.
    There are a few standard uploads which are not so good, such as the sales order load, and to a lesser extent the purchase order load. In these cases you might need to look at using the BAPIs but get help from someone who has used them before.
    You might want to use recordings for very simple loads like bank master.
    If migrating SAP to SAP, where possible you would look to transfer master data via ALE using IDOCs. But again you need to know what you are doing and it sounds like you are moving from a non-SAP to a SAP system anyway.
    Good luck,
    Ben
    http://www.harlex-ltd.com

Maybe you are looking for