Export-import if database is down

I had a query whether we can run the export and the import process if the database is down. Basically export process is the command, 'exp'. So, would this
command run even if the database is down. Even, import command is run with command, 'imp', so, would it be requiring database to be up.
I hope, my question is clear.
Please, help in solving the doubt.
regards

Hi,
Export is a Logical Backup in sense DB Objects, Think how come with out DB instance is Up and Running, does exp application can extract the things.
- Pavan Kumar N

Similar Messages

  • Export/Import of Database to increase the number of datafiles

    My BW system is close to 400 GB and it only has 3 datafiles.  Since we have 8 CPU cores, I need 8 datafiles.  What is the best way to export/import the database in order to achive 8 datafiles?

    With a BW system that size you can probably get away with it.  You most likely do at least a few full loads so all that data will be evenly distrubuted when you drop and reload.  If you can clean up some of your PSAs and log tables you can probably shrink the 3 files down a little anyway.  If you do little maintenance like that every few weeks, after locking auto growth, you can probably shave 2-3 GBs off each file each week.  Do that for a few months and your large files will love 20 GBs while the other ones start to grow.  Rebuilding indexes also helps with that.  You will be surprised how fast they will level out that way.
    With regard to performance you should be fine.  I certainly wouldnt do it at 10 am on Tuesday though :-).  You can probably get away with it over a weekend though.  It will take basically no time at all to create them and very little IO.  If you choose to clean out the 3 existing ones that will take some time.  I have found it takes about 1-3 hours to shrink a 150 GB datafile down to 100 GBs, that was with 8 CPUs, 30 GBs of RAM, and an SAN that I don't fully understand

  • Portal export/import  on database 8.1.7 on an other machine

    Hello,
    I'm running Oracle8i on AIX. I have problem when I try full export/import my database (version 8.1.7) on an other machine. I have a problem with import users PORTAL30 and PORTAL30_SSO and CREATE TABLE 'WWSEC_ENABLER_CONFIG_INFO$'
    The log when import full is:
    IMP-00017: following statement failed with ORACLE error 22973:
    "CREATE TABLE "WWSEC_ENABLER_CONFIG_INFO$" OF "SEC_ENABLER_CONFIG_TYPE" OID "
    "'BA6473B86EF248BEE0330A18780F48BE' ( PRIMARY KEY ("LS_LOGIN_URL")) OBJECT "
    "ID PRIMARY KEY PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 LOGGING STORAG"
    "E(INITIAL 131072 NEXT 131072 MINEXTENTS 1 MAXEXTENTS 4096 PCTINCREASE 0 FRE"
    "ELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "TBS_PORTAL""
    IMP-00003: ORACLE error 22973 encountered
    ORA-22973: size of object identifier exceeds maximum size allowed
    Thank you.
    Borjana Zekic

    When using export/import between different releases, you need to use the older version of export to create your dump file. In your case, you must use the 8.0.5 export utility to export from your 8.1.7 database. You need a tnsname in your 8.0.5 tnsnames.ora that points to your 8.1.7 database. Then you use SQL Net to connect and do the export. So, instead of doing exp system/manager you would do exp system/manager@tnsname. Just be sure to use the 8.0.5 export, and your import will work fine.

  • Issues with Export/Import using Database & Shared buffer

    Hi All,
    I have a method that calls a program via a job and I am having issues passing data between the two.
    Note that two different users process the method and program (via job) resp. This is how I am calling the second prog-
    SUBMIT ZPROGRAM
                 VIA JOB     l_jobname
                     NUMBER  l_jobcount
                     USER    i_user
                 AND RETURN.
    I need to pass data from method to  the second program and vice versa and then the method continues its processing with the data acquired from the second program.
    I have tried using Import/Export using database and also shared buffer. What I have found is that most of the times I am able to pass data from method to program. However the job takes a couple of min to execute and I think that the data is not making back to the method in time.
    I have looked at some useful forum links-
    Problem with export/import in back ground
    Re: EXPORT/IMPORT  to MEMORY
    but havent been able to find an answer yet. Any solution? Thanks in advance for your help!
    Liz

    Hi Suhas, Subhankar
    I have tested the scenario without the job previously itself and it works. Thats the reason, i am trying with the job now as my requirement is that I need to change the user while executing the second report.
    Here is an example of my import/export - I am passing the return value from the second report to the first.
    Code in second report-
    DATA: INDXKEY LIKE INDX-SRTFD VALUE 'RET1'.
    INDX-AEDAT = SY-DATUM.
    INDX-USERA = SY-UNAME.
    EXPORT RETURN1 TO SHARED BUFFER INDX(ST) ID INDXKEY.
    Code in first report -
    SUBMIT ZPROGRAM
                     VIA JOB     l_jobname
                         NUMBER  l_jobcount
                         USER    i_user
                     AND RETURN.
    Once Job close FM is executed successfully, I import the values as follows
    IMPORT RETURN1 TO RETURN1 FROM SHARED BUFFER INDX(ST) ID INDXKEY3.
    INDXKEY is having value RET1.
    However Return1 is not having any values in first report. It has some value in executed without the job
    Please note that I have tried Export/import with Database too and I am getting the same results.
    Thanks for your suggestions.
    Regards, Liz

  • Export/Import full database dump fails to recreate the workspaces

    Hi,
    We are studying the possibility of using Oracle Workspace to maintain multiple versions of our business data. I recently did some tests with the import/export of a dump that contains multiple workspaces. I'm not able to import the dump successfully because it doesn't recreate the workspaces on the target database instance, which is an empty Oracle database freshly created. Before to launch the import with the Oracle Data Pump utility, I have the "LIVE" workspace that exists in the WMSYS.WM$WORKSPACES_TABLE table. After the import is done, there is nothing in that table...
    The versions of the Oracle database and Oracle Workspace Manager are the same on source and target database:
    Database version (from the V$VERSION view):
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    PL/SQL Release 10.2.0.4.0 - Production
    CORE     10.2.0.4.0     Production
    TNS for 64-bit Windows: Version 10.2.0.4.0 - Production
    NLSRTL Version 10.2.0.4.0 - Production
    Workspace Manager version (from the WM_INSTALLATION view):
    OWM_VERSION: 10.2.0.4.3
    In order to recreate the tablespaces successfully during the full import of the dump, the directory structure for the tablespaces is the same on the target database's computer. I used the instructions given in this document, section "1.6 Import and Export Considerations" at page 1-19:
    http://www.wyswert.com/documentation/oracle/database/11.2/appdev.112/e11826.pdf
    Considering that the release of Oracle database used is version 10.2.0.4, I use the following command to import the dump since it doesn't contain the WMSYS schema:
    impdp system/<password>@<database> DIRECTORY=data_pump_dir DUMPFILE=expfull.dmp FULL=y TABLE_EXISTS_ACTION=append LOGFILE=impfull.log
    The only hints I have in the import log file are the following block of lines:
    1st one:
    ==============================
    Traitement du type d'objet DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    ORA-39083: Echec de la création du type d'objet PROCACT_SYSTEM avec erreur :
    ORA-01843: ce n'est pas un mois valide
    SQL en échec :
    BEGIN
    if (system.wm$_check_install) then
    return ;
    end if ;
    begin execute immediate 'insert into wmsys.wm$workspaces_table
    values(''LIVE'',
    ''194'',
    ''-1'',
    2nd one at the end of the import process:
    ==============================
    Traitement du type d'objet DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    ORA-39083: Echec de la création du type d'objet PROCACT_SCHEMA avec erreur :
    ORA-20000: Workspace Manager Not Properly Installed
    SQL en échec :
    BEGIN
    declare
    compile_exception EXCEPTION;
    PRAGMA EXCEPTION_INIT(compile_exception, -24344);
    begin
    if (system.wm$_check_install) then
    return ;
    end if ;
    execute immediate 'alter package wmsys.ltadm compile';
    execute immediate 'alter packag
    The target operating system is in french... here is a raw translation:
    1st one:
    ==============================
    Treatment of object type DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    ORA-39083: Object type creation failed PROCACT_SYSTEM with error :
    ORA-01843: invalid month
    failed SQL :
    BEGIN
    if (system.wm$_check_install) then
    return ;
    end if ;
    begin execute immediate 'insert into wmsys.wm$workspaces_table
    values(''LIVE'',
    ''194'',
    ''-1'',
    2nd one at the end of the import process:
    ==============================
    Treatment of object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    ORA-39083: Object type creation failed PROCACT_SCHEMA with error :
    ORA-20000: Workspace Manager Not Properly Installed
    failed SQL :
    BEGIN
    declare
    compile_exception EXCEPTION;
    PRAGMA EXCEPTION_INIT(compile_exception, -24344);
    begin
    if (system.wm$_check_install) then
    return ;
    end if ;
    execute immediate 'alter package wmsys.ltadm compile';
    execute immediate 'alter packag
    By the way, the computer of the source database is Vista 64 in english and the destination computer is Vista 64 in french, do you think this has anything to do with these error messages? The parameters of the NLS_SESSION_PARAMETERS view are the same in the source and target database...
    Thanks in advance for your help!
    Joel Autotte
    Lead Developer
    Edited by: 871007 on Jul 13, 2011 7:31 AM

    I tried to import the full database dump with the "imp" command and I had the same error message with more detail:
    . Import d'objets SYSTEM dans SYSTEM
    IMP-00017: Echec de l'instruction suivante avec erreur ORACLE 1843 :
    "BEGIN "
    "if (system.wm$_check_install) then"
    " return ;"
    " end if ;"
    " begin execute immediate 'insert into wmsys.wm$workspaces_ta"
    "ble"
    " values(''LIVE'',"
    " ''194'',"
    " ''-1'',"
    " ''SYS'',"
    " *to_date(''09-DEC-2009 00:00:00'', ''DD-MON-YYYY HH"*
    *"24:MI:SS''),"*
    " ''0'',"
    " ''UNLOCKED'',"
    " ''0'',"
    " ''0'',"
    " ''37'',"
    " ''CRS_ALLNONCR'',"
    " to_date(''06-APR-2011 14:24:57'', ''DD-MON-YYYY HH"
    "24:MI:SS''),"
    " ''0'',"
    " '''')'; end;"
    "COMMIT; END;"
    IMP-00003: Erreur ORACLE 1843 rencontrée
    ORA-01843: ce n'est pas un mois valide
    ORA-06512: à ligne 5
    The call to the TO_DATE function with the string of the date format is incompatible with the format and language set in the NLS parameters of my database. Since I know that the "imp" and "exp" commands work client side (unlike "impdp" and "expdp"), I did a full export of the source database from the computer on which I try to import the dump to have the same language exported in the dump and I was able to import the dump successfully.
    But, I would like to use the Data Pump tool instead of the old import and export tools. How would it be possible to set the NLS parameters that "impdp" must use before to launch the import of the dump?
    The NLS parameters are set to the "AMERICAN" settings in both source and destination database when I validate that with the "NLS_DATABASE_PARAMETERS" view. The dump I did with "expdp" on the source database exported the data with the "AMERICAN" settings as well...
    On the destination database, the computer language is in french and I guess it is considered by the Data Pump import tool since it outputs the log in the language of the computer and it tries to do a TO_DATE with the wrong date format because it works with the "CANADIAN FRENCH" settings...
    Any suggestions? I'll continue to read on my side and re-post if I find the solution.
    Thanks!
    Edited by: 871007 on Jul 19, 2011 8:42 AM
    Edited by: 871007 on Jul 19, 2011 8:43 AM
    Edited by: 871007 on Jul 19, 2011 8:43 AM

  • Exports / Imports on Databases with VPD

    I'm running Oracle 11.1.0.7, with VPD. I'm trying to use datapump to export and import the data from one machine to another.
    Not all the information in all the tables is making it across. Is there a special role or privilege that will allow SYSTEM (as opposed to SYS as SYSDBA) to grab all the data?

    I'm running Oracle 11.1.0.7, with VPD. I'm trying to use datapump to export and import the data from one machine to another.
    Not all the information in all the tables is making it across. Is there a special role or privilege that will allow SYSTEM (as opposed to SYS as SYSDBA) to grab all the data?

  • EXPORT/IMPORT DB PROVIDER NT-- SOLARIS IN PORTAL RELEASE 2

    Hi,
    I've been assigned to export/import a database provider with all the applications contained therein from NT platform to SUN SOLARIS platform. Can anyone please kindly tell how to do this or where to get the step-by-step guide for this task.
    Thanks.

    hi,
    i am using portal 9.0.4 in windows 2003 server. i am searching for the step-by-step import/export details. you told you have got it. can you share with me the same?
    my email id is [email protected] i am in-search of export/import commands. i am able to export through the transport set, but the import will show the importing details, but nothing is available. please help me.
    regards
    sudhir

  • Can data pump handle export/import of a sequence explicitly?

    Is it possible to just export/import a database sequence explicitly using data pump.
    I know we can export/import sequence by exporting/importing the user in which it is defined.
    However I am looking for a way to export/import individual sequnces without having to export/import the entire user.
    Thanks

    Hi, you can use the INCLUDE clause into expdp command, with this option you can specified the object for the export and import process. You can review the next documentation for more information about this topic.
    <br><br>
    INCLUDE
    <br><br>
    Regards.

  • Export/Import through R3SETUP.EXE

    Dear All,
    I am currently running on SAP 4.6C SR2 with 4.6D Kernel on SQL Database. The Dtabase size is 2 TB.
    I wish to Export/Import the Database through R3Setup.EXE.
    Can anybody help me with steps to do the same & other considerations to be kept in mind.
    I need to split Database fiels in 20 equal files.

    Hi,
    check following document  Chapter 8 R3load procedure on windows
    https://websmp206.sap-ag.de/~sapdownload/011000358700002949502001E/R3HOM.PDF
    regards,
    kaushal

  • Database refresh of a 10gb database through export-import

    hi ,
    I am using oracle 10g R2 . I need to refresh the development server from the production server .
    The size of the database is 10GB . Can I do it with export-import ?
    Thank you .

    besides these mentioned methods, i.e. 1) exp/imp; 2) data pump; 3) transportable tablespaces; 4) rman duplicate. you can also use manual duplication of the production database which is faster and easier I would say. Below is link, which can be used for detailed steps.
    http://www.dba-oracle.com/oracle_tips_db_copy.htm
    Please note that, if your source database can not be shut down, so you can copy your datafiles by putting tablespaces or database into hot backup mode.

  • System copy using SAPInst(Export Import database Independent prcoess failed

    Hello ,
    I am doing a System copy using SAPInst export/import process .
    Source system : SAP Netweaver'04( BW 3.5 , Kernel : UC 640 ,Patch level 196 )
    Export process fails at Phase 2 -  Database Export at R3load jobs running 1,waiting 0 . Below is the log details
    SAPSDIC.log
    (EXP) INFO:  entry for BAPICONTEN                        in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20000621155733
    (EXP) INFO:  entry for BAPICONTENT255                    in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20031127161249
    (EXP) INFO:  entry for BAPICONVRS                        in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20010131174038
    (EXP) INFO:  entry for BAPICREATORDATA                   in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20000621155733
    (EXP) INFO:  entry for BAPICRMDH1                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMDH2                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMEXP                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175513 > 20031211120627
    (EXP) INFO:  entry for BAPICRMEXT                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175514 > 20031211120627
    (EXP) INFO:  entry for BAPICRMKEY                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMKEY_T                      in DDNTT is newer than in DDNTT_CONV_UC: 20051229175835 > 20031211120803
    (EXP) INFO:  entry for BAPICRMMSG                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMMSG_T                      in DDNTT is newer than in DDNTT_CONV_UC: 20051229175835 > 20031211120803
    (EXP) INFO:  entry for BAPICRMOBJ                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175514 > 20031211120628
    (EXP) INFO:  entry for BAPICRMPAREX_T                    in DDNTT is newer than in DDNTT_CONV_UC: 20051229175452 > 20031211120305
    (EXP) INFO: limit reached, 5000 tables in DDNTT are newer than in DDNTT_CONV_UC
    (EXP) INFO: NameTab check finished. Result=2  #20100113131216
    (EXP) INFO: check for inactive NameTab entries: Ok.
    (RSCP) INFO: I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    (DB) INFO: disconnected from DB
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: job finished with 1 error(s)
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: END OF LOG: 20100113131216
    ***SAPCLUST.log ****
    (NT)  Warn:  EDIDOC: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "EDIDOC"
    (NT)  Warn:  PCDCLS: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "PCDCLS"
    (NT)  Warn:  SFHOA: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "SFHOA"
    (NT)  Warn:  SFHYT: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "SFHYT"
    (NT)  Warn:  UMG_TEST_C: normal NameTab from 20040211095029 younger than alternate NameTab from 20031113150115!
    (EXP) TABLE: "UMG_TEST_C"
    myCluster (55.22.Exp): 712: error when retrieving alternate nametab description for physical table UMG_TEST_F.
    myCluster (55.22.Exp): 713: return code received from nametab is 32
    myCluster (55.22.Exp): 299: error when retrieving physical nametab for table UMG_TEST_F.
    (CNV) ERROR: data conversion failed.  rc = 2
    (RSCP) INFO: I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    (DB) INFO: disconnected from DB
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: job finished with 1 error(s)
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: END OF LOG: 2010011312563
    Please suggest.
    Thanks & Regards
    Ganesh

    Is your DB unicode?  If so, did you select the unicode flag in sapinst?
    This [thread|System Copy Error while exporting ABAP; might offer some help.
    -Zach

  • Migrate Database- Export/Import

    Hi,
    I need to migrate an Oracle database 9i from Sun Solaris to Linux. The final target database version would be 11g.
    Since this is a 9i database, I see that we have only option of export and database. We have around 15 schemas.
    I have some queries related to it.
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?
    Since there is a database version different I dont want to touch sys, system schema objects.
    Appreciate your thoughts.
    Regards
    Cherrish Vaidiyan

    Hi,
    Let me try to answer some of these questions you queried for:
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?Export won't export sys objects. For example, there are tables in sys, like obj$ that contain information for other metadata objects, like scott.emp, etc. This is not exported becuase when scott.emp is exported, the data from obj$ is essentially exported that way. When the dumpfile is imported and scott.emp is recreated, the data in sys.obj$ will be restored through the create table statement. As far as the SYSTEM schema is concerned, some objects are exported and some are not. There are tables in system that contain information about queues, jobs, etc. These will probably not make any sense on the target system so those types of tables are excluded from the export job. Other objects make sense to export/import so those are done. This is all figured out in the internals of export/import. Thre are other schemas that are not exproted. Some that I can think of are DMSYS, ORDSYS, etc. This would be for the same reason as SYS.
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?No, the dumpfiles are formatted differently. If you use exp, then you must use imp. If you use expdp, then you must use impdp. You can do exp on 9i and imp on 11g with the dumfile that was created on 9i.
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?This is case by case decision. It depends on what you want. If you want the complete database moved, then I would personally think that a full=y is what you would want to do. If you just did schema exports, then you would never export the tablespaces. This would mean that you would have to create the tablespaces on the source system before you ran imp. There are other objects that are not exported when a schema level export is performed that are exproted when a full is performed. This information can be seen in the utilities guide. Look to see what is exported when in user/schema mode vs full/database mode.
    Since there is a database version different I dont want to touch sys, system schema objects.This is all done for you with the internal workings of exp/imp.
    Dean
    Edited by: Dean Gagne on Jul 29, 2009 8:38 AM

  • Export/import to clone database -- please advise

    Hi All,
    Need help especially from exports that are strong in export/import.
    can someone advise me how to use export/import to clone a database?
    I know how to clone database using cold backup and hot backup.
    But wish to know the full steps in using export/import to do cloning.
    I'm going to do a full export on my "TEST" database, drop the database, create a database and import using the full export.
    After doing a full database export, full=y. What are the steps to follow. Have a few doubts about it.
    1) export the database (full=y)
    2) drop the database
    3) re-create the database, will be using the same name "TEST".
    4) create the users (IDs created by DBA, eg: schemas), tablespace and datafiles. --> import doesn't do it right?
    5) do a full import (full=y). In this step, will i encounter any problems? As in understand that all the oracle-owned objects (example SYSTEM, SYS, AUX) are already in the newly created database, will import full=y cause any problems?
    6) do i have to re-create the roles, SYNONYMS and grant role,sys,obj privileges to IDs not created by oracle. (example schema owner)
    in short i would like to know what EXPORT will export and not export
    what IMPORT will import and not import.
    Please advise.

    Hi zetabouy,
    thanks for ur input, definately helpful for me.
    just to confirm. am i right to say that only a full database level (full=y) export/import will export/import out roles as well as public synonyms? because i have tried also exporting a user from the database, and imported that user into a different database, but public synonyms and roles were not imported.
    one more question:
    For example we are going to do port data from production to uat database using export and import for only the schema.
    After exporting from production with such syntax (owner=OWNER01),
    is is better to drop all objects beloging to the schema owner in UAT before importing it (touser=OWNER01, ignore=y) using the production dumpfile?
    I'm asking this question as I'm afraid that if we did not drop all the pl/sql objects in UAT database, the updated procedures/fuctions/triggers exported from the production database will not get imported to the UAT database as UAT database have the same procedures/fuctions/triggers names (but not updated code) .
    will profiles be exported and imported too duing full database level exp/imp?
    please kindly advise.
    thanks
    Message was edited by:
    chew

  • Export/import to/from database

    hi,
    i do not know what is this for. i read the help but still no idea.
    i know to use export/import to/from memory id and also set/get but not export/import to/from database.
    1) what help says it stores data cluster? what is data cluster
    2) can have example of export/import to/from database?
    3) what is the different for export/import to/from memory id and database?
    thanks

    Hi,
    1) A data cluster is a set of data objects grouped together for the purpose of storage in a storage medium(Like database, local memory or shared meomory etc), which can only be edited using ABAP statements like EXPORT, IMPORT and DELETE.
    2) Suppose you want to export the data in a internal table to database.
    Here TABLE_ID is the identifer of your data cluster. You are exporting your data to a SAP table INDX and giving it an ID (TABLE_ID in this case and this ID you are giving it an area by name XY where your data will be stored). EXPORT tab = itab
      TO DATABASE indx(XY)
      CLIENT '000'
      ID 'TABLE_ID'.   
    You can get this data as follows.
    IMPORT tab = itab
      FROM DATABASE indx(xy)
      CLIENT '000'
      ID 'TABLE_ID'.
    3) The difference is simple when you use MEMORY ID the data you export is stored in the application server MEMORY where as in the case of DATABASE it is stored in the database.
    Regards,
    Sesh

  • Export - Import in Oracle Application Database

    The export in Oracle Applications Database works fine.
    But the import always show some issue and sometimes never completes.
    Even if it completes it complete with error , how to ensure an error free export/import ?
    How to handle a table movement if the table consists of a LONG column ?

    The errors are cannot import this tables..
    Also mostly are of Long tables..Any IMP-XXXXX or ORA-XXXXX messages?
    Can you explain how long objects are exported and imported ?What type of objects?
    What is the database version?
    Are you doing a full database import?
    Also sometimes import hangs for 2/3 days?Does it hang when importing a specific object?
    Did you follow the steps outlined in the following notes when you did the Export/Import?
    Note: 362205.1 - 10g Release 2 Export/Import Process for Oracle Applications Release 11i
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=362205.1
    Note: 230627.1 - 9i Export/Import Process for Oracle Applications Release 11i
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=230627.1

Maybe you are looking for

  • How to assign a private IP address to a public IP address

    Hello. At the beginning sorry for my poor English. My company uses a Cisco 881 router and I have the following problem to resolve. I need to assign an local IP address from my private network to an public IP address (this is the public IP address of

  • GUI_DOWNLOAD fieldnames

    Hi, I have this code in my program: TYPES: BEGIN OF gtyp_cabeceras, column(50) TYPE c. TYPES: END OF gtyp_cabeceras. DATA: g_t_cabeceras TYPE STANDARD TABLE OF gtyp_cabeceras, g_w_cabeceras TYPE gtyp_cabeceras. g_w_cabeceras-column = 'Matr de coches'

  • Flash Player Security Settings per your D:/ drive...

    Hi, For anyone who read my post earlier in the day, I apologize for the redundancy; this is a new issue for me. I've been working on an e-learning product using Flash Builder; it was developed to be web-based, but the client has changed the requireme

  • Possible to use old company thinkpad?

    Hello, I'm new here to the forums. A company in the same building as I work was throwing away several old laptops, and a took 3 of them home.  I now have one T400, one T410 and one X201 lenovo thinkpad. I was owndering if it is possible for me to use

  • Web Service Error: java.sql.SQLException: Io exception: Broken pipe

    Hi, I am using JDeveloper PL/SQL web service generator. After some time we start receiving the "java.sql.SQLException: Io exception: Broken pipe" exception when we invoke the web service. I've noticed that the error starts occuring once all the conne