Export table data from production to quality system - Oracle 10.2

Hi,
I am trying to export a few table's data from our production system to the quality system.
Reason: To synchronize the table data between Production and Quality system
System info: SAP ECC 6.0 on Oracle 10.2 ; OS: Solaris
Description:
There are few standard SAP tables which hold the data of the various institutions of our client. Our requirement is to export the master data relevant for a particular institution from the production system to the quality system. We intend to use the Oracle Export and Import utility (Suggestions on the usage of which tool are welcome).
Constraints:
Lets say, there is a table NXXX in production system. This table contains the data for the institutions AA, BB, CC. I intend to export the data relevant to the institution AA from the table NXXX and import the data to the table NXXX in the Quality system.
Now I understand that the schema of these tables are different. In the production system, the schema would be SAP<Prod SID>.NXXX and in the quality system, the schema is SAP<Qas SID>.NXXX. I need your suggestions to overcome this constraint.
Regards,

Hi,
I do not think you can just export data in your scenario using the oracle imp/exp tool it's all or nothing...
as for the scema naming between QA/PROD you just need to use :
imp <dump file> fromuser=<PROD schema>  touser=<QA schema>  only thing you have to look out for is the mandt as it will not change.
I had something similar recently where I had to copy 14+ million rows from Prod to a Dev client without impacting on other clients data for this table in Dev.
So for data consistancy 
1) I restored a copy of prod on a new LPAR and exported the table.
2) Created a dummy prod client using client '000' as a base.
3) imported the table  using the syntax above enabling the change of schema
We have now a 14 + million rows in the dummy prod client on the new LPAR
Now on the Dev server
4) Created new client on dev as per point 2
5) SCC9 Remote client copy from the new LPAR using SAP_ALL
6) As I had 9+ million rows in the target client to get rid of I used delete_commit procedure as in metalink note 3777.1
    This enable  me to delete rows fast without caning the roll back segs.
7) using R3trans and a command file from the command line copied the single table into the target client.
I do not know if this is useful for you because your scenario is slighley different, but this is what I did for a whole table for a particular mandt.
Regards
Mark
Edited by: Mark Norman on Oct 22, 2009 11:16 AM
Edited by: Mark Norman on Oct 22, 2009 11:17 AM

Similar Messages

  • Export Master Data from Production to DEV

    Hi Gurus,
              I need to export All the Master Data and Configuration files from Production to Development System. If anyone can Advise me to know the step by step for exporting and importing. Our PRD server is on HP-UX/Oracle and ECC6.
    Eg Master Data like Customer Master, Vendor Master, Material Master...Etc.,
    Please Advice.

    Hi,
    I need to export All the Master Data and Configuration files from Production to Development System.
    I think you can have a option to do simple Remote Client copy with SAP_CUST or SAP_APPL client copy profile from your PRD system to DEV system.
    You can create a New Client in DEV system by performing Remote Client Copy as mentioned above from PRD system to ease the activity.
    You can get all the information in below mentioned Web Links.
    [Copying Clients Between Systems|http://help.sap.com/saphelp_nw04/helpdata/en/69/c24c684ba111d189750000e8322d00/content.htm]
    [SAP Client Copy Profiles|http://help.sap.com/saphelp_nw04/helpdata/en/99/0a2a887e2511d2a6250000e82deaaa/content.htm]
    [SAP Client Copy Data Types|http://help.sap.com/saphelp_nw04/helpdata/en/ff/9be1817d811c41985662da25061f98/content.htm]
    Regards,
    Bhavik G. Shroff

  • How to export the data from an excel file to Oracle database by using macro

    Hi,
    I want to export data from an excel file into Oracle database.There is one method that by creating the macros through excel we are able to do this.
    can anybody please send me the code to create macros?
    Thanks in advance.
    Regards,
    Surendra.V

    There are several resources on the web, but you could start analyzing if this code fits your needs --> VBA Macro for Excel 2003 Export of Text File with Comma and Quote Delimiters.
    ~ Madrid

  • Exporting table data from a JDBC rowset to an excel sheet

    Hi,
        I am developing an application in which I have to insert and display records on a view.
        I have used a webservice to display data.
        On the display view I need to provide a button which on clicking exports the table data to an excel sheet.
       I want to use the JDBC rowset to export data to an excel rather than the binary cache method.
       So could someone plz help me out with the code.

    Dear Abinas,
    Please read this tutorial.
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/edc2f3c2-0401-0010-8898-acd5b6a94353?QuickLink=index&…
    Thanks & Regards,
    Patralekha

  • Reg: Exporting the data from the Physical Standby database system.

    Hello all,
    We have an oracle standard edition one 11gR1 environment, in which i need to export the data from the physical standby system.
    So can anyone suggest me, How to do it safely (mount state).
    Regards,
    Konda.

    Hi,
    You can use "exp" utility to export the database from a read only database whereas you can't export the read only database using "expdp".
    Since you are using version 11g, it is not advisable to use exp. You can convert your standby to snapshot database, open read, write and export the database using expdp and then again convert it to standby.
    This is a very easy process which we follow in our company to do testing when required.
    Follow note 11g Using Snapshot Standby Database. (Doc ID 443720.1) for converting and usage.
    Thank you!!

  • Export table data only

    Hi Gurus,
    I am wondering if anyone can offer some advice in regards to export table data from a source and import it to another system.
    We have a development environment (say, Banner_dev) and a production one (say, Banner_Prod). We routinely perform database refresh from Banner_Prod to Banner_dev, using Oracle Rman cloning. However, the developers are not happy with this approach saying that other things (such as packages, functions, procedures) get overwritten.
    We (DBAs) have been asked to see if we can use export/import approach so that only data within a schema will be exported and imported, leaving other objects untouched in Banner_Dev when we refresh the database from Banner_Prod to Banner_Dev.
    Is this doable with Oracle Data Pump? If it is, how can we accomplish this?
    Thanks!

    You might consider a layered, scripted approach that takes into account the differences between the production and development environments. You might want to take into account things like differences between table DDL, data upgrading, coordinating code and so forth. The ideal is to be able to say "hey everybody, we're going to push this button tonight, be sure your work is saved into this system that will refresh everything."
    Of course, I'm used to doing this with exp/imp because of various app and dba requirements (like, I really don't want to use the same schema name between dev and prod) and the scripts have been refined over time, but there isn't any reason you couldn't have a general post-clone procedure. Describe it in terms that developers see as beneficial to them.

  • What are the major process to transfer the data from legacy to sap system.

    What are the major process to transfer the data from legacy to sap system using BDC at Real Time only?

    hi,
    BATCH DATA COMMUNICATION
    main methods are:
    1. SESSION METHOD
    2. CALL TRANSACTION
    3. DIRECT INPUT
    Advantages offered by BATCH INPUT method:
    1. Can process large data volumes in batch.
    2. Can be planned and submitted in the background.
    3. No manual interaction is required when data is transferred.
    4. Data integrity is maintained as whatever data is transferred to the table is through transaction. Hence batch input data is submitted to all the checks and validations.
    To implement one of the supported data transfers, you must often write the program that exports the data from your non-SAP system. This program, known as a “data transfer” program must map the data from the external system into the data structure required by the SAP batch input program.
    The batch input program must build all of the input to execute the SAP transaction.
    Two main steps are required:
    • To build an internal table containing every screen and every field to be filled in during the execution of an SAP transaction.
    • To pass the table to SAP for processing.
    Prerequisite for Data Transfer Program
    Writing a Data Transfer Program involves following prerequisites:
    Analyzing data from local file
    Analyzing transaction
    Analyzing transaction involves following steps:
    • The transaction code, if you do not already know it.
    • Which fields require input i.e., mandatory.
    • Which fields can you allow to default to standard values.
    • The names, types, and lengths of the fields that are used by a transaction.
    • Screen number and Name of module pool program behind a particular transaction.
    To analyze a transaction::
    • Start the transaction by menu or by entering the transaction code in the command box.
    (You can determine the transaction name by choosing System – Status.)
    • Step through the transaction, entering the data will be required for processing your batch input data.
    • On each screen, note the program name and screen (dynpro) number.
    (dynpro = dyn + pro. Dyn = screen, pro = number)
    • Display these by choosing System – Status. The relevant fields are Program (dynpro) and Dynpro number. If pop-up windows occur during execution, you can get the program name and screen number by pressing F1 on any field or button on the screen.
    The technical info pop-up shows not only the field information but also the program and screen.
    • For each field, check box, and radio button on each screen, press F1 (help) and then choose Technical Info.
    Note the following information:
    - The field name for batch input, which you’ll find in its own box.
    - The length and data type of the field. You can display this information by double clicking on the Data Element field.
    • Find out the identification code for each function (button or menu) that you must execute to process the batch-input data (or to go to new screen).
    Place the cursor on the button or menu entry while holding down the left mouse button. Then press F1.
    In the pop-up window that follows, choose Technical info and note the code that is shown in the Function field.
    You can also run any function that is assigned to a function key by way of the function key number. To display the list of available function keys, click on the right mouse button. Note the key number that is assigned to the functions you want to run.
    Once you have program name, screen number, field name (screen field name), you can start writing.
    DATA TRANSFER program.
    Declaring internal table
    First Integral Table similar to structure like local file.
    Declaring internal table like BDCDATA
    The data from internal table is not transferred directly to database table, it has to go through transaction. You need to pass data to particular screen and to particular screen-field. Data is passed to transaction in particular format, hence there is a need for batch input structure.
    The batch input structure stores the data that is to be entered into SAP system and the actions that are necessary to process the data. The batch input structure is used by all of the batch input methods. You can use the same structure for all types of batch input, regardless of whether you are creating a session in the batch input queue or using CALL TRANSACTION.
    This structure is BDCDATA, which can contain the batch input data for only a single run of a transaction. The typical processing loop in a program is as follows:
    • Create a BDCDATA structure
    • Write the structure out to a session or process it with CALL TRANSACTION USING; and then
    • Create a BDCDATA structure for the next transaction that is to be processed.
    Within a BDCDATA structure, organize the data of screens in a transaction. Each screen that is processed in the course of a transaction must be identified with a BDCDATA record. This record uses the Program, Dynpro, and Dynbegin fields of the structure.
    The screen identifier record is followed by a separate BDCDATA record for each value, to be entered into a field. These records use the FNAM and FVAL fields of the BDCDATA structure. Values to be entered in a field can be any of the following:
    • Data that is entered into screen fields.
    • Function codes that are entered into the command field. Such function codes execute functions in a transaction, such as Save or Enter.
    The BDCDATA structure contains the following fields:
    • PROGRAM: Name of module pool program associated with the screen. Set this field only for the first record for the screen.
    • DYNPRO: Screen Number. Set this field only in the first record for the screen.
    • DYNBEGIN: Indicates the first record for the screen. Set this field to X, only for the first record for the screen. (Reset to ‘ ‘ (blank) for all other records.)
    • FNAM: Field Name. The FNAM field is not case-sensitive.
    • FVAL: Value for the field named in FNAM. The FVAL field is case-sensitive. Values assigned to this field are always padded on the right, if they are less than 132 characters. Values must be in character format.
    Transferring data from local file to internal table
    Data is uploaded to internal table by UPLOAD of WS_UPLOAD function.
    Population of BDCDATA
    For each record of internal table, you need to populate Internal table, which is similar to BDCDATA structure.
    All these five initial steps are necessary for any type of BDC interface.
    DATA TRANSFER program can call SESSION METHOD or CALL TRANSACTION. The initial steps for both the methods are same.
    First step for both the methods is to upload the data to internal table. From Internal Table, the data is transferred to database table by two ways i.e., Session method and Call transaction.
    SESSION METHOD
    About Session method
    In this method you transfer data from internal table to database table through sessions.
    In this method, an ABAP/4 program reads the external data that is to be entered in the SAP System and stores the data in session. A session stores the actions that are required to enter your data using normal SAP transaction i.e., Data is transferred to session which in turn transfers data to database table.
    Session is intermediate step between internal table and database table. Data along with its action is stored in session i.e., data for screen fields, to which screen it is passed, the program name behind it, and how the next screen is processed.
    When the program has finished generating the session, you can run the session to execute the SAP transactions in it. You can either explicitly start and monitor a session or have the session run in the background processing system.
    Unless session is processed, the data is not transferred to database table.
    BDC_OPEN_GROUP
    You create the session through program by BDC_OPEN_GROUP function.
    Parameters to this function are:
    • User Name: User name
    • Group: Name of the session
    • Lock Date: The date on which you want to process the session.
    • Keep: This parameter is passed as ‘X’ when you want to retain session after
    processing it or ‘ ‘ to delete it after processing.
    BDC_INSERT
    This function creates the session & data is transferred to Session.
    Parameters to this function are:
    • Tcode: Transaction Name
    • Dynprotab: BDC Data
    BDC_CLOSE_GROUP
    This function closes the BDC Group. No Parameters.
    Some additional information for session processing
    When the session is generated using the KEEP option within the BDC_OPEN_GROUP, the system always keeps the sessions in the queue, whether it has been processed successfully or not.
    However, if the session is processed, you have to delete it manually. When session processing is completed successfully while KEEP option was not set, it will be removed automatically from the session queue. Log is not removed for that session.
    If the batch-input session is terminated with errors, then it appears in the list of INCORRECT session and it can be processed again. To correct incorrect session, you can analyze the session. The Analysis function allows to determine which screen and value has produced the error. If you find small errors in data, you can correct them interactively, otherwise you need to modify batch input program, which has generated the session or many times even the data file.
    CALL TRANSACTION
    About CALL TRANSACTION
    A technique similar to SESSION method, while batch input is a two-step procedure, Call Transaction does both steps online, one after the other. In this method, you call a transaction from your program by
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>
    Messages into <MSGTAB>.
    Parameter – 1 is transaction code.
    Parameter – 2 is name of BDCTAB table.
    Parameter – 3 here you are specifying mode in which you execute transaction
    A is all screen mode. All the screen of transaction are displayed.
    N is no screen mode. No screen is displayed when you execute the transaction.
    E is error screen. Only those screens are displayed wherein you have error record.
    Parameter – 4 here you are specifying update type by which database table is updated.
    S is for Synchronous update in which if you change data of one table then all the related Tables gets updated. And sy-subrc is returned i.e., sy-subrc is returned for once and all.
    A is for Asynchronous update. When you change data of one table, the sy-subrc is returned. And then updating of other affected tables takes place. So if system fails to update other tables, still sy-subrc returned is 0 (i.e., when first table gets updated).
    Parameter – 5 when you update database table, operation is either successful or unsuccessful or operation is successful with some warning. These messages are stored in internal table, which you specify along with MESSAGE statement. This internal table should be declared like BDCMSGCOLL, a structure available in ABAP/4. It contains the following fields:
    1. Tcode: Transaction code
    2. Dyname: Batch point module name
    3. Dynumb: Batch input Dyn number
    4. Msgtyp: Batch input message type (A/E/W/I/S)
    5. Msgspra: Batch input Lang, id of message
    6. Msgid: Message id
    7. MsgvN: Message variables (N = 1 - 4)
    For each entry, which is updated in database, table message is available in BDCMSGCOLL. As BDCMSGCOLL is structure, you need to declare a internal table which can contain multiple records (unlike structure).
    Steps for CALL TRANSACTION method
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. UPLOAD or WS_UPLOAD function to upload the data from local file to itab. (Considering file is local file)
    4. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>.
    Refresh BDCTAB.
    Endloop.
    (To populate BDCTAB, You need to transfer each and every field)
    The major differences between Session method and Call transaction are as follows:
    SESSION METHOD CALL TRANSACTION
    1. Data is not updated in database table unless Session is processed. Immediate updation in database table.
    2. No sy-subrc is returned. Sy-subrc is returned.
    3. Error log is created for error records. Errors need to be handled explicitly
    4. Updation in database table is always synchronous Updation in database table can be synchronous Or Asynchronous.
    Error Handling in CALL TRANSACTION
    When Session Method updates the records in database table, error records are stored in the log file. In Call transaction there is no such log file available and error record is lost unless handled. Usually you need to give report of all the error records i.e., records which are not inserted or updated in the database table. This can be done by the following method:
    Steps for the error handling in CALL TRANSACTION
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. Internal table BDCMSG like BDCMSGCOLL
    4. Internal table similar to Ist internal table
    (Third and fourth steps are for error handling)
    5. UPLOAD or WS_UPLOAD function to upload the data from the local file to itab. (Considering file is local file)
    6. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tr.code> using <Bdctab>
    Mode <A/N/E>
    Update <S/A>
    Messages <BDCMSG>.
    Perform check.
    Refresh BDCTAB.
    Endloop.
    7 Form check.
    IF sy-subrc <> 0. (Call transaction returns the sy-subrc if updating is not successful).
    Call function Format_message.
    (This function is called to store the message given by system and to display it along with record)
    Append itab2.
    Display the record and message.
    DIRECT INPUT
    About Direct Input
    In contrast to batch input, this technique does not create sessions, but stores the data directly. It does not simulate the online transaction. To enter the data into the corresponding database tables directly, the system calls a number of function modules that execute any necessary checks. In case of errors, the direct input technique provides a restart mechanism. However, to be able to activate the restart mechanism, direct input programs must be executed in the background only. Direct input checks the data thoroughly and then updates the database directly.
    You can start a Direct Input program in two ways;
    Start the program directly
    This is the quickest way to see if the program works with your flat file. This option is possible with all direct input programs. If the program ends abnormally, you will not have any logs telling you what has or has not been posted. To minimize the chance of this happening, always use the check file option for the first run with your flat file. This allows you to detect format errors before transfer.
    Starting the program via the DI administration transaction
    This transaction restarts the processing, if the data transfer program aborts. Since DI document are immediately posted into the SAP D/B, the restart option prevents the duplicate document posting that occurs during a program restart (i.e., without adjusting your flat file).
    Direct input is usually done for standard data like material master, FI accounting document, SD sales order and Classification for which SAP has provided standard programs.
    First time you work with the Direct Input administration program, you will need to do some preparation before you can transfer data:
    - Create variant
    - Define job
    - Start job
    - Restart job
    Common batch input errors
    - The batch input BDCDATA structure tries to assign values to fields which do not exist in the current transaction screen.
    - The screen in the BDCDATA structure does not match the right sequence, or an intermediate screen is missing.
    - On exceptional occasions, the logic flow of batch input session does not exactly match that of manual online processing. Testing the sessions online can discover by this.
    - The BDCDATA structure contains fields, which are longer than the actual definition.
    - Authorization problems.
    RECORDING A BATCH INPUT
    A B recording allows you to record a R/3 transaction and generate a program that contains all screens and field information in the required BDC-DATA format.
    You can either use SHDB transaction for recording or
    SYSTEM ? SERVICES ? BATCH INPUT ? EDIT
    And from here click recording.
    Enter name for the recording.
    (Dates are optional)
    Click recording.
    Enter transaction code.
    Enter.
    Click Save button.
    You finally come to a screen where, you have all the information for each screen including BDC_OKCODE.
    • Click Get Transaction.
    • Return to BI.
    • Click overview.
    • Position the cursor on the just recorded entry and click generate program.
    • Enter program name.
    • Click enter
    The program is generated for the particular transaction.
    BACKGROUND PROCESSING
    Need for Background processing
    When a large volume of data is involved, usually all batch inputs are done in background.
    The R/3 system includes functions that allow users to work non-interactively or offline. The background processing systems handle these functions.
    Non-interactively means that instead of executing the ABAP/4 programs and waiting for an answer, user can submit those programs for execution at a more convenient planned time.
    There are several reasons to submit programs for background execution.
    • The maximum time allowed for online execution should not exceed 300 seconds. User gets TIMEOUT error and an aborted transaction, if time for execution exceeds 300 seconds. To avoid these types of error, you can submit jobs for background processing.
    • You can use the system while your program is executing.
    This does not mean that interactive or online work is not useful. Both type of processing have their own purposes. Online work is the most common one entering business data, displaying information, printing small reports, managing the system and so on. Background jobs are mainly used for the following tasks; to process large amount of data, to execute periodic jobs without human intervention, to run program at a more convenient, planned time other than during normal working hours i.e., Nights or weekends.
    The transaction for background processing is SM36.
    Or
    Tools ? Administration ? Jobs ? Define jobs
    Or
    System ? services ? Jobs
    Components of the background jobs
    A job in Background processing is a series of steps that can be scheduled and step is a program for background processing.
    • Job name. Define the name of assigned to the job. It identifies the job. You can specify up to 32 characters for the name.
    • Job class. Indicates the type of background processing priority assigned to the job.
    The job class determines the priority of a job. The background system admits three types of job classes: A B & C, which correspond to job priority.
    • Job steps. Parameters to be passed for this screen are as follows:
    Program name.
    Variant if it is report program
    Start criteria for the job: Option available for this are as follows:
    Immediate - allows you to start a job immediately.
    Date/Time - allows you to start a job at a specific name.
    After job - you can start a job after a particular job.
    After event - allows you to start a job after a particular event.
    At operation mode - allows you to start a job when the system switches to a particular operation mode.
    Defining Background jobs
    It is two step process: Firstly, you define the job and then release it.
    When users define a job and save it, they are actually scheduling the report i.e., specifying the job components, the steps, the start time.
    When users schedule program for background processing, they are instructing the system to execute an ABAP/4 report or an external program in the background. Scheduled jobs are not executed until they are released. When jobs are released, they are sent for execution to the background processing system at the specified start time. Both scheduling and releasing of jobs require authorizations.
    HANDLING OF POP UP SCREEN IN BDC
    Many times in transaction pop up screen appears and for this screen you don’t pass any record but some indication to system telling it to proceed further. For example: The following screen
    To handle such screen, system has provided a variable called BDC_CURSOR. You pass this variable to BDCDATA and process the screen.
    Usually such screen appears in many transactions, in this case you are just passing information, that YES you want to save the information, that means YES should be clicked. So you are transferring this information to BDCDATA i.e., field name of YES which is usually SPOT_OPTION. Instead of BDC_OKCODE, you are passing BDC_CURSOR.
    BDC_CURSOR is also used to place cursor on particular field.
    A simple transaction where you are entering customer number on first screen and on next screen data is displayed for the particular customer number. Field, which we are changing here, are name and city. When you click on save, the changed record gets saved.
    Prerequisite to write this BDC interface as indicated earlier is:
    1. To find screen number
    2. To find screen field names, type of the field and length of the field.
    3. To find BDC_OKCODE for each screen
    4. Create flat file.
    generally  Batch Input usually are used to transfer large amount of data. For example you are implementing a new SAP project, and of course you will need some data transfer from legacy system to SAP system.
    CALL TRANSACTION is used especially for integration actions between two SAP systems or between different modules. Users sometimes wish to do something like that click a button or an item then SAP would inserts or changes data automatically. Here CALL TRANSACTION should be considered.
    2. Transfer data for multiple transactions usually the Batch Input method is used.
    check these sites for step by step process:
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    Check these link:
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://www.sap-img.com/abap/question-about-bdc-program.htm
    http://www.itcserver.com/blog/2006/06/30/batch-input-vs-call-transaction/
    http://www.planetsap.com/bdc_main_page.htm
    call Transaction or session method ?
    null

  • Exporting data from Production to Development in a R/3 System

    Good Morning,
    I have a landscape where i have a Productive and a Development system (SAP R/3 Enterprise Release 4.70 SR1 1.10 with MaxDB 7.5 running on a HP-UX 11.23) and i was asked to copy the data and customizing (not user master data) from Production to Development. Both systems have 7 clients (6 for each country and one for Human Resources) and i need to copy the productive data to the development data (the Productive data can overwrite the Development data).
    This is my first time doing this, and since i have a short deadline to do it, i need all the help i can get...
    I'm not sure if i should use SCC8 or SCC9 in order to create a transport request for each client copy.
    I've checked the OS space in sapdata and transport directories, i've also checked the database space and i need to change MAXDATAVOLUMES size and also create more volumes.
    Now that i've done that, i need to know what to do next... Does anyone knows where i can find a procedure or documentation on this? Some kind of step by step guide?
    About the Export Profiles, which default profile should i choose in order to copy
    All client specific data, customizing AND NOT user master data?
    Can i do any kind of test run before the export in order to understand the space i'm going to need in transport directory for example?
    Thanks in advance.
    Best Regards,
    Pedro Gaspar

    Hello Pedro,
    1. In SCC8 the SAP_APPL profile isn't available, so i guess i need to use SAP_ALL and change the user master data accordingly afterwards right?
    In that case, you should/can 'Export' the user master of the target system before the Client copy, so that you can re-import the same user master in the target system after sap_all copy from source.
    2. I also have several project teams developing in ABAP in my development system. If i export the productive client, will the ABAP code be overwritten (scripts, smartforms, etc...)? I ask this, because right now there are several programs with different versions in both systems (the landscape consists of only a DEV and PRD system NO quality). If the ABAP code in development will be overwritten, i need to tell the development teams to comment all the changes, then create transports to production to make the program versions the same, and then save the data an cofiles to import them again in development after the import. Will also RFC definitions change? Jobs will also be copied? What's the difference between this and a Homogeneous System Copy using backup/restore?
    This is the reason that normally the 'Production to Development' copy is not recommended. I would wait if our other techies here can help/advise us for this.
    A simple difference between a client copy and system copy is that the client copy is just 'client dependent' change and system copy is 'system wide' change - erasing everything on the target and just overwriting/duplicating the 'complete source' onto the target.
    Thanks

  • How to export some data from the tables of an owner with integrity?

    Hi to all,
    How to export some data from the tables of an owner with integrity?
    I want to bring some data from all tables in a single owner of the production database for development environment.
    My initial requirements are: seeking information on company code (emp), contract status (status) and / or effective date of contract settlement (dt_liq_efetiva) - a small amount of data to developers.
    These three fields are present in the main system table (the table of contracts). Then I thought about ...
    - create a temporary table from the query results table to contract;
    - and then use this temporary table as a reference to fetch the data in other tables of the owner while maintaining integrity. But how? I have not found the answer, because: what to do when not there is the possibility of a join between the contract and any other table?
    I am considering the possibility of consulting the names of tables, foreign keys and columns above, and create dynamic SQL. Conceptually, something like:
    select r.constraint_name "FK name",
    r.table_name "FK table",
    r.column_name "FK column",
    up.constraint_name "Referencing name",
    up.table_name "Referencing table",
    up.column_name "Referencing column"
    from all_cons_columns up
    join all_cons_columns r
    using (owner, position), (select r.owner,
    r.constraint_name fk,
    r.table_name table_fk,
    r.r_constraint_name r,
    up.table_name table_r
    from all_constraints up, all_constraints r
    where r.r_owner = up.owner
    and r.r_constraint_name = up.constraint_name
    and up.constraint_type in ('P', 'U')
    and r.constraint_type = 'R'
    and r.owner = 'OWNERNAME') aux
    where r.constraint_name = aux.fk
    and r.table_name = aux.table_fk
    and up.constraint_name = aux.r
    and up.table_name = aux.table_r;
    -- + Dynamic SQL
    If anyone has any suggestions and / or reuse code to me thank you very much!
    After resolving this standoff intend to mount the inserts in utl_file by a table and create another program to read and play in the development environment.
    Thinking...
    Let's Share!
    My thanks in advance,
    Philips

    Thanks, Peter.
    Well, I am working with release 9.2.0.8.0. But the planning is migrate to 10g this year. So my questions are:
    With Data Pump can export data just from tables owned for me (SCHEMAS = MYOWNER) parameterizing the volume of data (SAMPLE) and filters to table (QUERY), right? But parameterizing a contract table QUERY = "WHERE status NOT IN (2,6) ORDER BY contract ":
    1º- the Data Pump automatically searches for related data in other tables in the owner? ex. parcel table has X records related (fk) with Y contracts not in (2,6): X * SAMPLE records will be randomly exported?
    2º- for the tables without relation (fk) and which are within the owner (MYOWNER) the data is exported only based on the parameter SAMPLE?
    Once again, thank you,
    Philips
    Reading Oracle Docs...

  • System refresh from Production to Quality

    Hi,
    We are going for system refresh from Production to Quality. We are at SAP NetWeaver 2004s with 700 release and at 0021 level. Our data base system is DB6 with the release 09.07.0000
    I understand that there is a note 886102 available for the system copy. But I would like to know how that had been practically implemented from your ready documents like
    1. What would be the BI consultant role during the refresh (I mean where do we involve at). I have seen many links related to this but nothing answer my question, so please don't give me links available.
    2. How to identify tables that need to be copied and restored to retain the correct source systems for data/info sources.
    3. What should be the BI consultant task before refresh?
    4. What should be the BI consultant task post refresh?
    5. What are issues faced post refresh in quality system.
    I request, the consultant who had worked on these refresh can provide me correct solution.
    Thanks in advance.
    Regards.
    Raj

    Our prerefresh activities included
    Inform security team to do no user or authorisation changes for quality during the refresh.
    Set message in development to not release any transports anymore and set message in quality to inform users in quality not to manually import transports into quality and also not approve transports for production. This ensures no transports get moved to quality and production.
    Switch off cyclic import all job (like TMS_0000000038TMS_TP_IMPORT) and the cyclic  RSTMS_DIST_APPROVED_REQUESTS job 
    Prepare list of transports for re-import to quality after refresh and give this to BASIS.
    Post refresh activities included
    Tcodes SM37, SM35 and SP01. Check that BASIS had set all released jobs to status "Susp/Released"
    "All jobs are in 'Susp/Released' state. Set them all to 'Scheduled ' as follows:
    -  Run report BTCTRNS2 to change all to 'Released'.
    - Immediately use SM37 to change all to 'Scheduled' "
    IF ANY ARE NEEDED. Remember to change Exec Target in any job you need to release.
    "Schedule  RSTMS_DIST_APPROVED_REQUESTS to run at x:29 and x:59 - so every 30 minutes.
    Please schedule with DDIC as step user (and not your own user-id)."
    Check the STMS_QA and import queues to be sure that the tranports are correct- no extra ones during refresh???
    Once happy with the above request that Basis schedule the auto import to run every 30 minutes
    First ensure that BDLS has finished and system is ready for use.
    Post refresh issues faced in production
    Many reinit issues
    ACR issues.
    Master data issues.

  • Transport Table with Data from Development to Quality

    Hi
    Requirement:
    Transporting Table with Values from Development to Quality System.
    I have a Z* - table, appended one field at the end.
    I need to transport data along with the table structure from Development System to Quality System.
    In Development System, I have assigned the Transport request in Client 100(We can't add data to the table in this client).
    What needs to be done to migrate the new table structure with Values.
    Thanks in Advance
    Regards,
    Baburaj
    NOTE: the requirement is to transport table and table entries through a transport request.
    There should not be any manual update for the table in the target System (Quality System).

    Check the delivery class assigned to the table. Please see the documentation
    Delivery class
        The delivery class controls the transport of table data when installing
        or upgrading, in a client copy and when transporting between customer
        systems. The delivery class is also used in the extended table
        maintenance.
        There are the following delivery classes:
        o   A: Application table (master and transaction data).
        o   C: Customer table, data is maintained by the customer only.
        o   L: Table for storing temporary data.
        o   G: Customer table, SAP may insert new data records, but may not
            overwrite or delete existing data records. The customer namespace
            must be defined in table TRESC. (Use Report RDDKOR54 here).
        o   E: System table with its own namespaces for customer entries. The
    customer namespace must be defined in table TRESC. (Use Report
    RDDKOR54 here.)
    S: System table, data changes have the same status as program
    changes.
    W: System table (e.g. table of the development environment) whose
    data is transported with its own transport objects (e.g. R3TR PROG,
    R3TR TABL, etc.).
    or during client copy
    the data of client-specific tables is copied.
    Classes C, G, E, S: The data records of the table are copied to the
    target client.
    Classes W, L: The data records of the table are not copied to the
    target client.
    Class A: Data records are only copied to the target client if
    explicitly requested (parameter option). Normally it does not make
    sense to transport such data, but is supported to permit you to copy
    an entire client environment.
    or during installation, upgrade and language import
    behavior differs here for client-specific and cross-client tables.
    nt-specific tables
    Classes A and C: Data is only imported into client 000. Existing
    data records are overwritten.
    Classes E, S and W: Data is imported into all clients. Existing data
    records are overwritten.
    Class G: Existing data records are overwritten in client 000. In all
    other clients, new data records are inserted, but existing data
    records are not overwritten.
    Class L: No data is imported.
    Cross-client tables
    o   Classes A, L and C: No data is imported.
    o   Classes E, S, and W: Data is imported. Exisitng data records with
         the same key are overwritten.
    o   Classe G: Data records that do not exist are inserted, but existing
         data records are not overwritten.
    ehavior during transport between customer systems
    Data records of tables of delivery class L are not imported into the
    target system. Data records of tables of delivery classes A, C, E, G, S
    and W are imported into the target system (this is done for the target
    client specified in the transport for client-specific tables).
    se of the delivery class in the extended table maintenance
    The delivery class is also analyzed in the extended table maintenance
    (SM30). The maintenance interface generated for a table makes the
    following checks:
    o   You cannot transport the entered data with the transport link of the
        generated maintenance interface for tables of delivery classes W and
        L.
    o   When you enter data, there is a check if this data violates the
        namespace defined for the table in table TRESC. If the data violates
        the namespace, the input is rejected.
    Thanks,

  • Export data from the R/3 system from a ABAP report to BW system

    Hi Everyone,
    I have requirement to export data from the R/3 system from a ABAP report to BW system.
    Currently we are planning to create a Ztable to put the data into that, but I would like to know , is there a better way to do instead of going for a Ztable.
    Regards,
    Shobana.K

    Hi Shobha,
    In the second option as you mentioned sending the data to BI.
    1. Check when u click on second  option what is the target details that is showing?( Menas suppose if you send the data using Open Hub u will find the options like FF,Otherparty,etcc...).
    2. As you mentioned if you have option to upload the data to Table.Then you can load and u can create the Generic DS.
    Issues & Concern
                1. Need to think how the data will be poplated(Means how are going to maintain Delta info records table level before
                   populating to it)
                 2. whether Ztable should be refreshed before updating R/3 report data to Ztable.
    Based on the concerns verify your scenario.If  everything ok then u can proceed as per u r plan.
    Regards
    Ram.

  • What is the easiest way to export all tables data from Oracle to MS SQL Server?

    Hello MS,
    I would like to export all tables from Oracle 11.2 to MS SQL Server 2012 R1.
    Using the tool "Microsoft SQL Server Migration Assistant v6.0 for Oracle" did not work for me because there are too many warnings and errors regarding the schema creation (MS cannot know it because they are not the schema designer). My idea is
    to leave/skip the schema creation to the application designer/supplier and instead concentrate on the Oracle data export and MS SQL data import.
    What is the easiest way to export all tables data from Oracle to MS SQL Server quickly?
    Is it:
    - the „MS SQL Import and Export Data“ Tool
    - the “MS SQL Integration Services” Tool
    - not Oracle dump *.dmp format because it is a propritery binary format
    - flat file *.csv (delimited format)
    Thanks!

    Hi lingodingo,
    If you want to directly export all tables from Oracle database to SQL Server, I suggest you use SQL Server Import and Export Wizard. Because you just need to follow the wizard with GUI, this is the easiest way.
    If you want to make some modification for the tables‘ data before loading to SQL Server, I suggest you use SQL Server Integration Services package. For more details, please refer to the following similar thread:
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/38b2bed2-9d4e-48d4-a33f-1d9eed1c062d/flat-file-to-sql-server?forum=sqldatamining
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Table Technical settings for transfer data from developemnt to Quality

    Hi All,
    Please let me know which table techinical settings required for to transport table with data from development to quality?
    Thanks & Regards,
    padmaja.

    Hi,
    You do not require technical settings to transport data from dvlpmnt to quality.
    However, you can transport any tables entries with a simple trick:
    (1) Create a workbench request
    (2) Change the request (or its task) and add the following entry: R3TR TABU <name of table> =>
    (3) Enter the key values you want to transport, e.g.:
    - <client>* for all entries of the current client
    - add additional restrictions for the table entry (<client><bukrs>* for all entries of a given company code)
    Best regards,
    Prashant

  • R12 Interface table for loading data from 3rd part payroll system

    Hi All,
    Can anyone help me to have a lists and detailed technical information of available interface table on Oracle R12 for importing/loading data from third party payroll system. And what should be the best way of importing the data? It should be load first to AP then to GL or load it directly to GL?
    Any help much appreciated. Thanks.
    Cyrus

    Hi Cyrus,
    Can you please let us know your business requirements of this integration, i.e what business want to acheive out of this integration.
    It depends on what your business requirements are wether to send only accounting information from payroll system to your Oracle GL ( then you can integrate Payroll system to Oracle GL directly, by sending the accounting information from your payroll) or if your requirement is to create payroll related invoices in AP and then do payments in oracle AP and then pass accounting information to GL ( then integrate your payroll to AP)
    Regards,
    Madhav

Maybe you are looking for

  • AP that will work across subnets?

    Hello, We currently have a few 2106 wireless controllers and a dozen Aironet 1121 access points at our main location. We have radius authentication as well as certificates etc. This setup works great and has for years. We have a few satellite offices

  • Cursor Jumping around forms as type - now with 3.6.12 on windows 7 - makes impossible to use

    Have windows 7 installed with Firefox 3.6.12 - whenever bring up page with multiple field entries - as I type it jumps all around the page into other text areas or around inside the same text area and even seems to activate some navigation and submis

  • Pictures in Imovie

    I know this is a somewhat tired question; I'm trying to make slide shows using pictures from iphoto. I'm positive that the photos are a large size etc. There's a relatively large loss of quality once brought to imovie. I'm hesitant to use the newer i

  • CS6 - no Camera Raw plugin/filter

    Have upgraded from CS5 to CS6. Do not have Camera Raw plugin/filter. Have tried the 'updates' under help, no good. Do I have to purchase the plugin? If so, please direct me to the appropriate place on the Adobe site. thankyou in anticipation.

  • RDP has been disabled on ALL servers

    On Friday I came into work and all my servers had remote desktop disabled. This also included my terminal servers. There is no GPO for this. So group policy is out. I thought maybe a patch went out but could not locate any information on this. Has an