Tablespace export import via datapump

Friends ,
I want to export a particular tablespace using datapump "expdp" and also import it to a new tablespace of a new database . Using datapump , is it possible to do ?
Plz help .. ...

Maybe it's easier to use Transportable Tablespaces, see http://download.oracle.com/docs/cd/B19306_01/server.102/b14231/tspaces.htm#sthref1281
It's also possible to use datapump, but it takes longer, see http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_export.htm#sthref71
http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_import.htm#sthref254
HTH
Enrique

Similar Messages

  • Sequence behavior after importing via DataPump

    Hi Friends,
    I'm running Oracle DB 11.2.0.3 on Windows 2008 R2 SP1 Servers and I faced a strange sequences behavior after importing a schema via Data Pump.
    The export is done this way:
    EXPDP userid/password dumpfile= logfile= directory= remap_dumpfile=y (no news)
    The import is done this way
    IMPDP userid/password dumpfile= logfile= directory= remap_schema=(old_one:new_one) remap_tablespace=(old_ones:new_ones, so on...)
    The import works fine. There are no errors and the sequences are as well imported with no warnings.
    The strange behavior is that the sequences seems to "reset". When we call a sequence the NEXTVAL is just lower than the values already stored in the Database, and we get ORA-00001 a lot. The sequence should know that vale. I don't have this problem when using exp/imp, just via DataPump.
    So that when we create an order that should receive the value of 100, as an example, because we have 99 orders on the system, Oracle suggest a value lower than 99 or even the number one value (01).
    We then wrote a script to check the CURVAL of the sequences on the base schema to recreate the sequences using this initial value on the new imported schema.
    Does anyone faced this problem before?
    Any suggestions?
    Tks a lot

    Richard
    I've tried what you just said.
    Adding the parameter consistent=y makes Oracle to show a message like that at the beginning of the export
    "flashback_time=TO_TIMESTAMP('2013-09-03 12:18:12', 'YYYY-MM-DD HH24:MI:SS')"
    It warns me: Legacy Parameter CONSISTENT=TRUE, and replaces with flashback_time.
    Really, I did not know about this behavior with this "old" parameter. I'm very appreciated about your help.
    I was almost thinking it was a DataPump bug or something.
    Thanks a lot Richard. I'll now update my scripts and make lots of test.
    If you have more advices using this parameter please share us.
    Cheers

  • When I attempt to export/import via FTP for OOTP 11 I get a connectivity error message

    I participate in OOTP ver 11 fantasy baseball leagues which uses FTP to export and import league databases. For some reason, I have been unable to import/export using FTP. Each time I attempt to export/import I get a connectivity error message stating "am unable to connect". I have not had this problem with the previous versions of OOTP Fantasy Baseball. Plus, when I first upgraded to OOTP 11, I did not have this problem.
    == This happened ==
    Every time Firefox opened
    == I think it started after a firefox update ==
    == User Agent ==
    Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; SV1; .NET CLR 1.1.4322; .NET CLR 2.0.50727; .NET CLR 3.0.04506.648; MS-RTC LM 8; .NET CLR 3.0.4506.2152; .NET CLR 3.5.30729)

    If there are problems with updating or with the permissions then easiest is to download the full version and trash the currently installed version to do a clean install of the new version.
    Download a new copy of the Firefox program and save the disk image (dmg) file to the desktop
    *Firefox 6.0.x: http://www.mozilla.com/en-US/firefox/all.html
    *Firefox 3.6.x: http://www.mozilla.com/en-US/firefox/all-older.html
    * Trash the current Firefox application to do a clean (re-)install
    * Install the new version that you have downloaded
    Your profile data is stored elsewhere in the Firefox Profile Folder, so you won't lose your bookmarks and other personal data.
    * http://kb.mozillazine.org/Profile_folder_-_Firefox

  • SAP Export/Import via R3load

    Hello,
    SAP recommended a sorted export/import for oracle databases. A unsorted export/import ist very very quicker than a sorted export/import.
    very quicker means sorted = eg. 56 hours and unsorted eg. 12 hours.
    Is there any problems if i take the unsorted export?
    regards,
    stefan

    Hello Stefan,
    >> SAP recommended a sorted export/import for oracle databases
    That fact relies on database internal facts like clustering factor of indexes to the table. You will benefit of a sorted export in cases of primary indexes access and so on... but this topic is much bigger than your question.
    >> very quicker means sorted = eg. 56 hours and unsorted eg. 12 hours.
    Where did you get this values from? You can split big tables or packages ... or define parallel access paths and so on... and it depends on your database configuration (sort areas, PGA, etc..)
    I have done some migrations/exports with distmon... and optimized it manually ... so i got from round about 72 hours (normal export with some split packages) to only 12 hours. I have done only sorted exports.
    >> Is there any problems if i take the unsorted export?
    Some tables must be exported sorted (for example cluster tables) - have a look at sapnote #954268 ... but generally there are no "problems".
    If you have a BW system... some rules are changing...
    @ Markus:
    >> You mean the usage of "R3load -loadprocedure fast"? Or how do you do the "unsorted unload"?
    The option "loadprocedure fast" has nothing to do with the unload process. It speeds up the import (insert) by bypassing the buffer cache. Refer to sapnote #1045847 and the following link:
    http://download-east.oracle.com/docs/cd/B10501_01/server.920/a96524/c21dlins.htm
    Regards
    Stefan

  • Daily export backup (via datapump) of a 600GB production database

    Hi Guys,
    I have a 600GB DB.
    10G Database.
    Currently I have daily rman backup to tape.
    Based on your experience, should i bother to perform a daily export of production database of such size?
    Do u think it's useful?
    thanks

    Fran wrote:
    All depends, what do you want to do? Do you have enough space to save a backup and export daily?
    In my opinion with one backup is enough to save your database. I'm sorry, I can't agree with that. I've been in situations where the first, second and THIRD backup sources have been unavailable in the event of a Production restore. I had to go to Plan D and cross my fingers (and sweat a lot). That might be overkill, but you should never rely on just one backup method/source for Production data.
    Not only that, but, obviously, you can also use an export of the data and import into another database or a schema inside the database in case you need to address any logical corruption. We also use the exports to refresh test databases.
    We have many similar sizes of databases and we take exports weekly, instead of daily. An export will be an incomplete backup, obviously, though you can ensure consistency by giving it an SCN to use as a datapoint. You can also run Data Pump in parallel which does speed things up.
    If you use it as part of your backup strategy, I'd make sure I had regular backups of the parameter file and the controlfiles (both binary and trace). I'd also make sure I had regular txt files showing the necessary information for the tablespaces and the datafiles which you can use to recreate the files in case you needed to.
    Mark

  • Export / import tablespace with all objects (datas, users, roles)

    Hi, i have a problem or question to the topic export / import tablespace.
    On the one hand, i have a database 10g (A) and on the other hand, a database 11g (B).
    On A there is a tablespace called PRO.
    Furthermore 3 Users:
    PRO_Main - contains the datas - Tablespace PRO
    PRO_Users1 with a role PRO_UROLE - Tablespace PRO
    PRO_Users2 with a role PRO_UROLE - Tablespace PRO
    Now, i want to transfer the whole tablespace PRO (included users PRO_MAIN, PRO_USER1, PRO_User2 and the role PRO_UROLE) from A to B.
    On B, I 've created the user PRO_Main and the tablespace PRO.
    On A , i execute following statement:
    expdp PRO_Main/XXX TABLESPACES=PRO DIRECTORY=backup_datapump DUMPFILE=TSpro.dmp LOGFILE=TSpro.log
    On B:
    impdp PRO_Main/XXX TABLESPACES=PRO DIRECTORY=backup_datapump DUMPFILE=TSpro.dmp LOGFILE=TSpro.log
    Result:
    The User PRO_Main was imported with all the datas.
    But i 'm missing PRO_USER1, PRO_User2 and the role PRO_UROLE...
    I assume, i 've used wrong parameters in my expd and / or impdp.
    It would be nice, if anybody can give me a hint.
    Thanks in advance.
    Best Regards,
    Frank

    When you do a TABLESPACE mode export by specifying just the tablespaces, then all that gets exported are the tables and their dependent objects. The users, roles, and the tablespace definitions themselves don't get exported.
    When you do a SCHEMA mode export by specifying the schemas, you will get the schema definitions (if the schema running the export is privied) and all of the objects that the schema owns. The schema does not own roles or tablespace definitions.
    In your case, you want to move
    1. schemas - which you already created 1 on your target database
    2. roles
    3. everything in the tablespaces owned by multiple schemas.
    There is no 1 export/import command that will do this. This is how i would do this:
    1 - move the schema definitions
    a. you can either create these manually or
    b1. expdp schemas=<your list of schemas> include=user
    b2 impdp the results from b1.
    2. move the roles
    expdp full=y include=role ...
    remember, this will include all roles. If you want to limit what gets exported, then use:
    include=role:"in ('ROLE1', 'ROLE2', ETC.)
    impdo the roles just exported
    3. move the user information
    a. If you want to move all of the schema's objects like functions, packages, etc, then you need to use a schema mode
    export
    expdp user/password schemas=a,b,c ...
    b. If you want to move only the objects in those tablespaces, then use the tablespace export
    expdp user/password tablespaces=tbs1, tbs2, ...
    c. import the dumpfile generated in step 3
    impdp user/password ...
    Hope this helps.
    Dean

  • System copy via SAPINST (Export/Import)

    Hello!
    We are using SAP ERP EHP4 system on Windows with SAP MaxDB (Version: 7.7.07.09).
    We would like to copy this productive system to QAS system.
    Therefore we successfully run SAPINST (with running db and sap, because we cannot stop productive system) with option Export and have currently an valid export of productive system.
    After we read several sap notes (such as 1014782, 129352) we found that the system copy procedure Backup/Restore will be recomended.
    Questions
    1) Can we still copy productive system via export and import option with SAPINT?
    1a) Should/must the SAP Max DB instance be installed before runing SAPINT on the target server?
    1b)  Which SAP MaxDB version should be installed on the target server?
    Has some one experience with system copy of SAP systems with SPA MaxDB via Export/Import?
    It will be great to get some answers or recommendation.

    You're back with a new ID after using:
    Holger Thomasson
    Jgen Pfeiffer
    Thom Heinemann
    H. Lemke
    and now Tom Hermann.
    too bad it took me so long to find you!

  • (9I) EXPORT의 TABLESPACE OPTION, IMPORT의 STATISTICS OPTION

    제품 : ORACLE SERVER
    작성날짜 : 2004-08-13
    (9I) EXPORT의 TABLESPACE OPTION, IMPORT의 STATISTICS OPTION
    ===========================================================
    PURPOSE
    Oracle 9i에서의 새로운 IMPORT parameter 인 STATISTICS 와
    EXPORT 의 새로운 기능인 TABLESPACE parameter에 관한 설명이다.
    SCOPE
    Export Transportable Tablespace Feature는 8i~10g Standard Edition에서는
    지원하지 않는다.
    Explanation & Example
    1. Oracle 9i에서는 STATISTICS 라는 새로운 Import parameter 값을 제공한다.
    비교 ) Oracle 8i에서는 ANALYZE parameter와 RECALCULATE_STATISTICS
    parameter의 조합으로 사용한다.
    - ALWAYS : 미리 계산된 통계정보가 불안정하든 하지 않든 간에
    상관없이 통계정보를 항상 import. (Default)
    - NONE : 미리 계산된 통계정보를 import하지 않는다.
    - SAFE : 불안정하면 table을 import한후 통계정보를 재계산하고
    그렇지 않으면 미리 계산된 통계정보를 import.
    - RECALCULATE : 미리 계산된 통계정보를 import하는 대신에 import한 후
    table에 대한 통계정보를 재계산.
    여기서 잠깐 통계정보가 불안정하다는 것은 언제를 말하는지를 짚어본다.
    : Export time에 아래와 같은 경우에 미리 계산된 통계정보를 불안정하다
    고 표시한다.
    - Export하는 동안에 row error가 있을 때
    - Client character set 또는 NCHAR character set이 server의
    character set 또는 NCHAR character set과 다를 때
    - Query option을 사용할 때 ( 8i 이상에서 제공하는 새로운
    parameter)
    - 특정 partition 또는 subpartition만이 export될 때
    비교 ) Export 시에 STATISTICS parameter는 estimate, compute, none
    의 세가지 값을 가지며 estimate가 default이다.
    2. Oracle 9i 에서는 특정 tablespace에 있는 table을 export하기 위한
    TABLESPACE라는 새로운 parameter를 제공한다.
    - index가 어느 tablesapce에 속해 있는지의 여부에 상관없이 tablespace
    에 속한 table와 연관된 index는 모두 export된다.
    - TABLESPACE mode로 export을 받기 위해서는 EXP_FULL_DATABASE
    privilege를 가지고 있어야 한다.
    비교 ) 8i에서의 TABLESPACE option은 8i에서의 새로운 기능인
    TRANSPORTABLE TABLESPACE 기능을 사용하는 데 필요한 단지 meta
    data만을 export받는 것에 그친다. TABLESPACE 단위로 data을 받을
    수는 없다.
    9i에서도 믈론 TABLESPACE option은 TRANSPORTABLE TABLESPACE 기
    능을 위해서도 사용이 된다.
    RELATED DOCUMENTS
    <Note:159787.1>
    Oracle 9i New Features for Administrtors.

    Hi
    1)if I use the 3rd option it should gather stats for
    all the objects including indexes. However, when I
    used that option it didn't gather satas for Index. Do
    I need to use cascade=>TRUE option?
    Yes, default is FALSE, so need to force TRUE
    2) I need to backup my current stats and gather new
    stats for entire DB. If I export DATABASE stats will
    it export all stats including index, system?
    Yes
    3) If I collect stats using GATHER_DATABASE_STATS, Do
    I need to explicitly collect other stats for index,
    system etc?
    No

  • Upgrade from 8.1.6 to 9.2.0.7 via export/import

    Has anyone done this or know if it's possible to simply build a 9.2.0.7 instance, then use export/import to perform the upgrade from 8.1.6 to 9.2.0.7 since a direct manual upgrade from 8.1.6 is not supported??
    TIA

    Yes, building a database in a new release and then exporting the existing database using its version of the exp utility, performing a binary copy of the exp file to the new server, and then running the new version imp utility using the transferred dmp file was the standard method of moving databases to new platforms/versions for many years. The same process could always be used on the same server between equal or from a current to a new release.
    HTH -- Mark D Powell --

  • R/3 Export & Import during Unicod Conversion via SOCKET Method

    Hi
    We are in the process of upgrading our R/3 Enterprise to ECC 6.0. The size of the database is around 4 TeraBytes.
    Can somebody help us with necessary documentation/ othe rinputs to handle R/3 Export & Import using SOCKET method?  FYI we referred the SAP given doucments on the same. Any particular DOs & DO NOTs would be highly appreciated.
    Thank You
    Sai

    Hi Rahul
    regarding your 'Unicode doubt"' some ideas:
    1) The Upgrade Master Guide SAP ERP 6.0 and the Master Guide SAP ERP 6.0 include introductory information. Among other, these guides reference the SAP Service Marketplace-location http://service.sap.com/unicode@sap.
    2) In Unicode@SAP can you find several (content-mighty) FAQs
    Conclusion from the FAQ: First of all your strategy needs to follow your busienss model (which we can not see from here):
    Example. The "Upgrade to mySAP ERP 2005"-FAQ includes interesting remarks in section "DO CUSTOMERS NEED TO CONVERT TO A UNICODE-COMPLIANT ENVIRONMENT?"
    "...The Unicode conversion depends on the customer situation....
    ... - If your organization runs a single code page system prior to the upgrade to mySAP ERP 2005, then the use of Unicode is not mandatory. ..... However, using Unicode is recommended if the system is deployed globally to facilitate interfaces and connections.
    - If your organization uses Multiple Display Multiple Processing (MDMP) .... the use of Unicode is mandatory for the mySAP ERP 2005 upgrade....."
    In the Technical Unicode FAQ you read under "What are the advantages of Unicode ...", that "Proper usage of JAVA is only possible with Unicode systems (for example, ESS/MSS or interfaces to Enterprise Portal). ....
    => Depending on the fact if your systems support global processes, or depending on your use of Java Applications, your strategy might need to look different
    3) In particular in view of your 3rd option, I recommend you to take a look into these FAQs, if not already done.
    Remark: mySAP ERP 2005 is the former name of the application, which is named SAP ERP 6.0, now
    regards, and HTH, Andreas R

  • Migrate Database- Export/Import

    Hi,
    I need to migrate an Oracle database 9i from Sun Solaris to Linux. The final target database version would be 11g.
    Since this is a 9i database, I see that we have only option of export and database. We have around 15 schemas.
    I have some queries related to it.
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?
    Since there is a database version different I dont want to touch sys, system schema objects.
    Appreciate your thoughts.
    Regards
    Cherrish Vaidiyan

    Hi,
    Let me try to answer some of these questions you queried for:
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?Export won't export sys objects. For example, there are tables in sys, like obj$ that contain information for other metadata objects, like scott.emp, etc. This is not exported becuase when scott.emp is exported, the data from obj$ is essentially exported that way. When the dumpfile is imported and scott.emp is recreated, the data in sys.obj$ will be restored through the create table statement. As far as the SYSTEM schema is concerned, some objects are exported and some are not. There are tables in system that contain information about queues, jobs, etc. These will probably not make any sense on the target system so those types of tables are excluded from the export job. Other objects make sense to export/import so those are done. This is all figured out in the internals of export/import. Thre are other schemas that are not exproted. Some that I can think of are DMSYS, ORDSYS, etc. This would be for the same reason as SYS.
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?No, the dumpfiles are formatted differently. If you use exp, then you must use imp. If you use expdp, then you must use impdp. You can do exp on 9i and imp on 11g with the dumfile that was created on 9i.
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?This is case by case decision. It depends on what you want. If you want the complete database moved, then I would personally think that a full=y is what you would want to do. If you just did schema exports, then you would never export the tablespaces. This would mean that you would have to create the tablespaces on the source system before you ran imp. There are other objects that are not exported when a schema level export is performed that are exproted when a full is performed. This information can be seen in the utilities guide. Look to see what is exported when in user/schema mode vs full/database mode.
    Since there is a database version different I dont want to touch sys, system schema objects.This is all done for you with the internal workings of exp/imp.
    Dean
    Edited by: Dean Gagne on Jul 29, 2009 8:38 AM

  • Best choice for exporting / importing EUL

    Hi all
    I have been tasked with migrating an EUL from a R11 to a R12 environment. The Discoverer version on both environments is 10.1.2 and the OS is Solaris on oracle db's.
    I am unfortunately not experienced with Discoverer and there seems to be no one available to assist for various reasons. So I have been reading the manual and forum posts and viewing metalink articles.
    I tried exporting the entire eul via the wizard and then importing it to the new environment but i was not successfull and experienced the system hanging for many hours with a white screen and the log file just ended.
    I assumed this was a memory problem or slow network issues causing this delay. Someone suggested I export import the EUL in pieces and this seemed to be effective but I got missing item warnings when trying to open reports. This piece meal approach also worried me regarding consistency.
    So I decided to try do the full import on the server to try negate the first problem I experienced. Due to the clients security policies I am not able to open the source eul and send it to our dev. I was able to get it from their dev 11 system but I dismissed this as the dev reports were not working and the only reliable eul is the Prod one. I managed to get a prod eex file from a client resource but the upload to my server was extremely slow.
    I asked the dba to assit with the third option of exporting a db dump of the eul_us and importing this into my r12 dev environment. I managed this but had to export the db file using sys which alleviated a priviledge problem when logging in, I have reports that run and my user can see the reports but there are reports that were not shared to sysadmin in the source enviroment are now prefixed with the version 11 user_id in my desktop and the user cannot see her reports only the sysadmin ones.
    I refreshed the BA's using a shell script I made up which uses the java cmd with parameters.
    After some re reading I tried selecting all the options in the validate menu and refreshing in the discover admin tool.
    If I validate and refresh the BA using the admin tool I get the hanging screen and a lot of warnings that items are missing( so much for my java cmd refresh!) and now the report will not open and I see the substitute missing item dialogue boxes.
    My question to the forum is which would be the best approach to migrate the entire eul from a R11 instance to a R12 instance in these circumstances?
    Many thanks
    Regards
    Nick

    Hi Srini
    The os and db details are as follows:
    Source:
    eBus 11.5.2
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    SunOS 5.10 Generic_142900-11 sun4u sparc SUNW,Sun-Fire-V890
    Target:
    ebus 12.1.2
    Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production DEV12
    SunOS 5.10 Generic_142900-11 sun4u sparc SUNW,Sun-Fire-V890
    Yes the DBA initially did an exp for me using EUL_US as the owner but a strange thing happened with privileges and also, some of the imported tables appeared in the target environment under the apps schema(21 tables) even though the eul_us exp had been 48 tables.
    I also had a problem on the db with "eul_us has insufficient privileges on table space discoverer" type errors.
    I checked the eul_us db privileges and was unable to resolve this initial privilege error even though the privileges were granted to eul_us.
    The dba managed to exp as system and then import it with the full=y flag in the import command which seems to bring in the privileges.
    Then I ran the eul5_id.sql and then made up a list of the business areas and made a sh script to refresh the business areas as follows:
    java -jar eulbuilder.jar -connect sysadmin/oracle1@dev -apps_user -apps_responsibility "System Administrator" -refresh_business_area "ABM Activities" -log refresh.log
    This runs successfully and I can log in select business area and grant access to the users. The reports return data.
    Then one of the users said she can't see all her reports. I noticed some if I opened desktop that were sitting there prefixed with a hash and her version 11 user id.
    So back to the manuals and in the disco admin help the instructions are to first go to view > validate > select all options then go to the business area and click file refresh. This gives me a lot of warnings about items that are missing. I assume this is because the item identifiers brought across in the db dump are the version 11 ones and thus not found in the new system.
    Any suggestions?
    Many thanks
    Nick

  • Oracle 9 spatial index export/import

    Hi,
    when exporting/importing a user via exp/imp I encounter a problem with the numeric characters encoding during creation of a spatial index.
    Imp tool produces script like this:
    "BEGIN "
    "execute immediate 'INSERT INTO USER_SDO_GEOM_METADATA values (''POINTS'',''GEOMETRY'',mdsys.SDO_dim_array(MDSYS.SDO_DIM_ELEMENT(''X'',50000000,160000000,,005),MDSYS.SDO_DIM_ELEMENT(''Y'',450000000,600000000,,005)),,005) ' ; "
    "COMMIT; END;"
    Problem is with wrong representation of the numeric value '0.005' as ,005.
    Originally, the index was created via
    INSERT INTO USER_SDO_GEOM_METADATA
    VALUES (
    'points',
    'geometry',
    MDSYS.SDO_DIM_ARRAY(
    MDSYS.SDO_DIM_ELEMENT('X',-125000000,-115000000, '0.005'),
    MDSYS.SDO_DIM_ELEMENT('Y', 30000000, 42100000, '0.005')
    NULL
    Any hints how to reimport the index correctly?
    Thanks,
    Bogart

    You might need to set the NLS_LANG environment variable to get character set conversion done on import (I don't know this unfortunately - it has never been a problem for me).
    There is a section on character set conversions in the utilities manual.

  • Offline Instantiation of a Materialized View Site Using Export/Import

    Has anyone had any success performing offline instantiation of a materialized view site using export/import in Oracle9?

    This is what I wanted to ask in the forum. I want to use datapump for the initial instantiation because I believe that indexes (the actual indextables, not just the definition) are also replicated using datapump.
    So, is it possible to do the instantiation using datapump?

  • Export/Import and Client Copy

    Hi All,
    Could you please help me with what are the major differences between Export/Import...client copy...and a System Refresh.How they differ from each other.
    Regards
    Rajesh

    Hello Rajesh ,
    I have capture some data from SAP Help :
    Local Copy: Copying Clients Within a System:
    You can improve the performance of the client copy by, for example, by excluding tables or packages, with Edit -> Expert Settings.
    You can exclude tables from the client copy, for example if they are not relevant for the target client, in the Tables tab
    Copying Clients Between Systems (Remote Copy):
    The same product is installed, with the same release, in both systems
    The client copier can copy a client into another system. The systems can be on different platforms. You can change the client number.
    When you copy a client from one system to another, the data is transferred directly via the RFC interface - there is no intermediate storage on hard disk.
    Transporting Clients Between Systems ;Client export (SCC8):
    The client copier can copy a client into another system, which can be on a different platform. You can change the client number.
    You are no longer required to transport clients before you can copy them between systems. You can make a remote copy instead.
    Up to three transport requests are created, depending on the selected copy profile and the existing data.
    The transport request for texts is e.g. only created if the source client contains customer texts.
    <sid>KO<no>  cross-client data
    <sid>KT<no>  client-specific data
    <sid>KX<no>  texts and forms
    The data export is performed automatically asynchronously. The output of the export includes the names of the transport requests that are to be imported.
    Import transport requests into the target client (STMS)
    Choose one of the transport requests of the client transport in the Transport Management System (TMS). The other transport requests belonging to this client transport are then automatically added in the correct order.
    Import these transport requests into the target client.
    client import postprocessing (SCC7)
    You need to perform postprocessing activities to adapt the runtime environment to the current state of the data.
    Copy by Transport Request :
    This function transports customizing changes that have been recorded in a transport request between two clients in a system.
    You can choose whether you only copy the object list of the request or also the object lists of unreleased tasks in the request.
    Entries in the target client are overwritten or deleted according to the key entries in the transport request.
    Choose Administration -> System administration -> Administration -> Client admin. ->Special Functions -> Copy Transport Request.
    <a href="http://help.sap.com/saphelp_nw70/helpdata/en/69/c24c0f4ba111d189750000e8322d00/frameset.htm">For more info Click here</a>
    Regards ,
    Santosh Karadkar

Maybe you are looking for