Export / Import generates IMP-00020

Hi all.
Today, I exported a schema using exp file=xxxxx.dmp direct=y statistics=none owner=XXXXXX. Then I recreated the database (for testing purposes).
When I tried to import the generated file, it gave me the IMP-00020 error, and after that several tables were not created. It is the same ORACLE_HOME, the same CHARACTERSET, the same server, the same release. Just the DB encarnation is different.
O.S.: Oracle EL 5.6
Oracle 11.2.0.2 64bit
Any ideas? Already tried increasing the buffer size (first to 1000000, then to 50000000), but the error persists.
Thanks in advance

CInglez wrote:
Hi all.
Today, I exported a schema using exp file=xxxxx.dmp direct=y statistics=none owner=XXXXXX. Then I recreated the database (for testing purposes).
When I tried to import the generated file, it gave me the IMP-00020 error, and after that several tables were not created. It is the same ORACLE_HOME, the same CHARACTERSET, the same server, the same release. Just the DB encarnation is different.
O.S.: Oracle EL 5.6
Oracle 11.2.0.2 64bit
Any ideas? Already tried increasing the buffer size (first to 1000000, then to 50000000), but the error persists.
Thanks in advanceWhy don't you use Data Pump?

Similar Messages

  • Export / import  exp / imp commands Oracle 10gXE on Ubuntu

    I have Oracle 10gXE installed on Linux 2.6.32-28-generic #55-Ubuntu, and I need soe help on how to export / import the base with exp / imp commands. The commands seems to be installed on */usr/lib/oracle/xe/app/oracle/product/10.2.0/server/bin* directory but I cannot execute them.
    The error message I got =
    No command 'exp' found, did you mean:
    Command 'xep' from package 'pvm-examples' (universe)
    Command 'ex' from package 'vim' (main)
    Command 'ex' from package 'nvi' (universe)
    Command 'ex' from package 'vim-nox' (universe)
    Command 'ex' from package 'vim-gnome' (main)
    Command 'ex' from package 'vim-tiny' (main)
    Command 'ex' from package 'vim-gtk' (universe)
    Command 'axp' from package 'axp' (universe)
    Command 'expr' from package 'coreutils' (main)
    Command 'expn' from package 'sendmail-base' (universe)
    Command 'epp' from package 'e16' (universe)
    exp: command not found
    Is there something I have to do ?

    Hi,
    You have not set environment variables correctly.
    http://download.oracle.com/docs/cd/B25329_01/doc/install.102/b25144/toc.htm#BABDGCHH
    And of course that sciprt have small hickup, so see
    http://ubuntuforums.org/showpost.php?p=7838671&postcount=4
    Regards,
    Jari

  • Schema Import failed :IMP-00008: unrecognized statement in the export file:

    Hi ,
    I was running a database schema import from a database DMP file.
    I encountered the following error in the log file that was getting generated :
    IMP-00008: unrecognized statement in the export file:
    All the tables got copied except for one. Now what I have thought of is to import that one table again after taking an export from the prod schema.
    But what about the functions,triggers,procedures ?? How do i bring all of them in the schema ?
    If reimporting the schema again the only option ?
    Thanks
    Kk

    Hello,
    In this order:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/exp_imp.htm#i1023988
    Table Objects: Order of Import
    Table objects are imported as they are read from the export file. The export file contains objects in the following order:
    1. Type definitions
    2. Table definitions
    3.Table data
    4.Table indexes
    5. Integrity constraints, views, procedures, and triggers
    6. Bitmap, function-based, and domain indexes
    The order of import is as follows: new tables are created, data is imported and indexes are built, triggers are imported, integrity constraints are enabled on the new tables, and any bitmap, function-based, and/or domain indexes are built. This sequence prevents data from being rejected due to the order in which tables are imported. This sequence also prevents redundant triggers from firing twice on the same data (once when it is originally inserted and again during the import).
    For example, if the emp table has a referential integrity constraint on the dept table and the emp table is imported first, all emp rows that reference departments that have not yet been imported into dept would be rejected if the constraints were enabled.
    When data is imported into existing tables, however, the order of import can still produce referential integrity failures. In the situation just given, if the emp table already existed and referential integrity constraints were in force, many rows could be rejected.
    A similar situation occurs when a referential integrity constraint on a table references itself. For example, if scott's manager in the emp table is drake, and drake's row has not yet been loaded, scott's row will fail, even though it would be valid at the end of the import.

  • Export/Import Process in the UI for Variations Content Translation is Generating CMP Files with No XML

    We have a SharePoint 2010 Publishing Website that uses variations to deliver contain to multiple languages. We are using a third-party translation company to translate publishing pages. The pages are
    exported using the  export/import using the UI process described here: "http://blogs.technet.com/b/stefan_gossner/archive/2011/12/02/sharepoint-variations-the-complete-guide-part-16-translation-support.aspx".
    Certain sub-sites are extremely content-intensive. They may contain many items in the Pages library as well as lists and other sub-sites. 
    For some sub-sites (not all), the exported CMP file contains no XML files. There should be a Manifest.XML, Requirements.XML, ExportSettings.XML, etc., but there are none. After renaming the CMP file
    to CAB and extracting it, the only files it contains are DAT files.
    The only difference I can see between the sub-sites that generate CMP files with no XML files is size. For example, there is one site that is 114 MB that produces a CMP file with no XML files. Small
    sites do not have this problem. If size is the problem, then I would think the process would generate an error instead of creating a single CMP file that contains only DAT files. However, I do not know exactly what the Export/Import Process in the UI is doing.
    This leads to two questions:
    1.
    Does anyone know why some CMP files, when renamed to *.CAB and extracted, would not contain the necessary XML files?
    2. Second, if exporting using the UI will not work, can I use PowerShell? I have tried the Export-SPWeb, but the Manifest.XML does not contain translatable
    content. I have not found any parameters that I can use with Export-SPWeb to cause the exported CMP to be in the same format as the one produced by the Export/Import process in the UI.
    As a next step, we could try developing custom code using the Publishing Service, but before doing this, I would like to understand why the Export/Import process in the UI generates a CMP that
    contains no XML files.
    If no one can answer this question, I would appreciate just some general help on understanding exactly what is happening with the Export/Import Process -- that is, the one that runs when you select
    the export or import option in the Site Manager drop down. Understanding what it is actually doing will help us troubleshoot why there are no XML files in certain export CMPs and assist with determining an alternate approach.
    Thanks in advance
    Kim Ryan, SharePoint Consultant kim.ryan@[no spam]pa-tech.com

    I wanted to bump this post to see about getting some more responses to your problem. I'm running into the same problem as well. We're running a SharePoint 2010 site and are looking at adding variations now. The two subsites with the most content take a
    while to generate the .cmp file (one to two minutes of the browser loading bar spinning waiting on the file). Both files are generated with a lot of .dat files but no .xml files. I was thinking like you that it must be a size issue. Not sure though. Did you
    ever happen to find a solution to this problem?

  • I have 100 groups in planning for those 100 groups i want to build roles like interactive,view user,planner etc.for those how to change in export -import folder .xml file  in that edit  how  to change user roles in that xml it will generate automatic id.h

    I have 100 groups in planning for those 100 groups i want to build roles like interactive,view user,planner etc.for those how to change in export -import folder .xml file  in that edit  how  to change user roles in that xml it will generate automatic id.how to do that in xml file ?

    Thanks john for you are reply.
    I had tried what you sad.I open shared service in that foundation project i had export shared service.after that in import-export file.In that role.csv,user.csv,group.csv.Like this file have.When i open user file added some users after i trying save in excel it shown messgse
    I click yes and save the .csv file and import from share servie. i got error like this
    am i doing right way john.or explain clearly

  • Parallel export/Import for exchange directory

    Hi,
    We have executed the homogeneous system copy(Win08/Ora 11) on ECC6 EHP4 and in order to speed up the process i have selected the parallel export/Import option on target system as export on source system is been executing by some other vendor.
    During Import abap phase it's displaying as" waiting -- mins for packahe data in exhange directory" as i have provided the path during input phase of "communication parameters for Import" in Exchange directory as <export directory>/Data.
    is this correct path and i have read some where that <TABART>.SQL  will be generated if on source system the same parallel expot/import option has enabled.And i couldn't find *.SQL files on export directory.
    Does it mean that guy who is executing on source system has not consider this option/not generated the SQl file with SMIGR_CREATE_DDL ?
    NB: i am executing import with SAPINST
    Regards,
    Dheeraj

    Dear Dheeraj,
    Kindly check in the export directory path, whether <package>.SGN files been generating or not .
    Also, we can check in exportmonitor.log.After export of individual pacakge, log says package transfered to network share directory.Even if it not merntions in log, parallel exp/imp not possible
    If it generates, parallel import will happen;If not, it will not.
    Its nothing to do with SQL files.
    Regards
    Ram

  • Imp-00020 long column too large for column buffer size (22)

    Hi friends,
    I have exported (through Conventional path) a complete schema from Oracle 7 (Sco unix patform).
    Then transferred the export file to a laptop(window platform) from unix server.
    And tried to import this file into Oracle10.2. on windows XP.
    (Database Configuration of Oracle 10g is
    User tablespace 2 GB
    Temp tablespace 30 Mb
    The rollback segment of 15 mb each
    undo tablespace of 200 MB
    SGA 160MB
    PAGA 16MB)
    All the tables imported success fully except 3 tables which are having AROUND 1 million rows each.
    The error message comes during import are as following for these 3 tables
    imp-00020 long column too large for column buffer size (22)
    imp-00020 long column too large for column buffer size(7)
    The main point here is in all the 3 tables there is no long column/timestamp column (only varchar/number columns are there).
    For solving the problem I tried following options
    1.Incresed the buffer size upto 20480000/30720000.
    2.Commit=Y Indexes=N (in this case does not import complete tables).
    3.first export table structures only and then Data.
    4.Created table manually and tried to import the tables.
    but all efforts got failed.
    still getting the same errors.
    Can some one help me on this issue ?
    I will be grateful to all of you.
    Regards,
    Harvinder Singh
    [email protected]
    Edited by: user462250 on Oct 14, 2009 1:57 AM

    Thanks, but this note is for older releases, 7.3 to 8.0...
    In my case both export and import were made on a 11.2 database.
    I didn't use datapump because we use the same processes for different releases of Oracle, some of them do not comtemplate datapump. By the way, shouldn't EXP / IMP work anyway?

  • BLOB importing with IMP

    Hello,
    I am trying to import a database dump to another location. This exercise gives an error when importing the table containing BLOB column.
    My original database was in tablespace 'R' and owned by user 'R'. Now I want to load it to tablespace 'M' owned by user 'M'. What I understand is that the loader (IMP) still tries to create table with a BLOB column in 'R' tablespace while I want it to be created in 'M'.
    What is the solution to this problem?
    Kindly reply asap.
    -- Hamid

    Well said Tom & Karen,
    However, I have extracted a piece of information from the metalink, it might be helpful to you.
    What are the errors you get and what type of export/import are you doing and have you tried SQL Loader which also supports loading of LOBs? The Utilities manual has information on how to accomplish export/import and load LOBs. If you are doing a direct path export then suggest running a conventional path export. Are you using any other features like Advanced Queuing and how big is the size of data, is it lesser than 4K? What kind of LOB specially you have and have you changed the database character set? If you are not using commit=y on import, suggest you try.
    130814.1 How to move LOB Data to Another Tablespace
    119897.1 How to Select a LOB Column over a Database Link
    66431.1 LOBS - Storage, Redo and Performance Issues
    If you are trying to set up a user/schema for one(on his own tablespace), into which he can import tables (have BLOB and CLOB columns) which are maintained in another schema (and tablespace) then its documented as an restriction described in the Oracle8i Utilities
    1. Import with INDEXFILE parameter to generate the script
    that can be used to modify the LOB's tablespace clause.
    $imp system/manager fromuser=scott tables=\(circul\)
    indexfile=create_lob_table
    2. Edit the generated script file create_lob_table.sql:
    REM CREATE TABLE "SCOTT"."CIRCUL" ("CIRCUL_ID" VARCHAR2(8) NOT NULL
    REM ENABLE, "FIC_CIRCUL" BLOB) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS
    REM 255 LOGGING STORAGE(INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAXEXTENTS
    REM 121 PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
    REM TABLESPACE "USERS" LOB ("FIC_CIRCUL") STORE AS (TABLESPACE "USERS"
    REM ENABLE STORAGE IN ROW CHUNK 2048 PCTVERSION 10 NOCACHE LOGGING
    REM STORAGE(INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAXEXTENTS 121
    REM PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)) ;
    REM ... 0 rows
    2.1 Remove the REM comments
    2.2 Change the owner SCOTT by the new owner U1
    2.2 Change the tablespace USERS by the TOUSER's tablespace NEW_TBS in the whole statement.
    3. Run the script create_lob_tabl
    4. Import the data only in the created table, ignoring the CREATE TABLE statement failure.
    $imp system/manager fromuser=scott tables=\(circul\) ignore=y

  • Oracle 9 spatial index export/import

    Hi,
    when exporting/importing a user via exp/imp I encounter a problem with the numeric characters encoding during creation of a spatial index.
    Imp tool produces script like this:
    "BEGIN "
    "execute immediate 'INSERT INTO USER_SDO_GEOM_METADATA values (''POINTS'',''GEOMETRY'',mdsys.SDO_dim_array(MDSYS.SDO_DIM_ELEMENT(''X'',50000000,160000000,,005),MDSYS.SDO_DIM_ELEMENT(''Y'',450000000,600000000,,005)),,005) ' ; "
    "COMMIT; END;"
    Problem is with wrong representation of the numeric value '0.005' as ,005.
    Originally, the index was created via
    INSERT INTO USER_SDO_GEOM_METADATA
    VALUES (
    'points',
    'geometry',
    MDSYS.SDO_DIM_ARRAY(
    MDSYS.SDO_DIM_ELEMENT('X',-125000000,-115000000, '0.005'),
    MDSYS.SDO_DIM_ELEMENT('Y', 30000000, 42100000, '0.005')
    NULL
    Any hints how to reimport the index correctly?
    Thanks,
    Bogart

    You might need to set the NLS_LANG environment variable to get character set conversion done on import (I don't know this unfortunately - it has never been a problem for me).
    There is a section on character set conversions in the utilities manual.

  • Upgrade from 8.1.6 to 9.2.0.7 via export/import

    Has anyone done this or know if it's possible to simply build a 9.2.0.7 instance, then use export/import to perform the upgrade from 8.1.6 to 9.2.0.7 since a direct manual upgrade from 8.1.6 is not supported??
    TIA

    Yes, building a database in a new release and then exporting the existing database using its version of the exp utility, performing a binary copy of the exp file to the new server, and then running the new version imp utility using the transferred dmp file was the standard method of moving databases to new platforms/versions for many years. The same process could always be used on the same server between equal or from a current to a new release.
    HTH -- Mark D Powell --

  • Error during export & import of rules in RAR 5.2

    Hi all,
    We followed the steps mentioned in exporting & importing the rules as per the Config guide. But we
    are receiving the below error during the import of rules. Can anybody please throw a limelight on why this error message is appearing?
    "Error in Table Data ==> VIRSA_CC_CRPROF
    SQL==>Insert into VIRSA_CC_CRPROF(VSYSKEY,PROFILE,RULESETID,RISKLEVEL,STATUS)Values(??????)
    Record==>D VIRSA_CC_CRPROF null"
    We also ensured that the downloaded file is not truncated and saved it in a separate folder. Does the file needs to saved in ANSI or Unicode text? We saved in ANSI
    Also the background ground job is not getting scheduled automatically during the import of ruels. Is that due to the above error?
    Thanks and Best Regards,
    Srihari.K

    Hello Sri,
    background job for generating the rules won`t be scheduled before you upload the file successfully.
    The most obvious reason for this error message is that you have line in your file - in table VIRSA_CC_CRPROF that is corrupted. Make sure all lines from table VIRSA_CC_CRPROF have all predefined fields - (VSYSKEY,PROFILE,RULESETID,RISKLEVEL,STATUS).
    If you keep hiting this problem just delete this table, after the upload you`ll add the critical profiles manually - I bet you don`t have many of them.
    Also from my experience I always choose to save the downloaded file in UNICODE UTF-8 format.
    Once the file is saved in other format, there`s no use of it, download it again and save directly in Unicode.
    Make sure you don`t have empty fields, even in the fields where you don`t have values, you must leave space, otherwise you`ll keep hitting the same issue.
    Regards,
    Iliya

  • Exported Motion Generator not rendering correctly in FCPX 10.0.1

    Hi
    I have a problem with exporting a generator from Motion 5 to FCPX 10.0.1.
    1) I created a sequence of a page of a book opening then closing. I created this in Motion 5, and exported it as a Generator to its own category and folder
    2) I have found that in this folder a movie of the sequence is created. This movie plays correctly. The movie is located at this path:
    HD/ Users/ Name/ Movies/ Motion Templates/ Generators/ Generator Title/ Individual Generator/ Gen Movie.mov  HD/ Users/ Name/ Movies/ Motion Templates/ Generators/ Generator Title/ Individual Generator/ Gen Movie.mov 
    NOTE: I also tried exporting the movie, from motion as a stand alone movie. This and the original motion file play the same and are correct
    3) Once in FCPX. I can either skim over the generator or place in the time line and let it render
    4) The results are the same, the front cover does not turn correctly. As it closes it gets half way, and then the page disappears. A few frames later it flies in from the left side of the screen
    5) This did not happen in the original version of FCPX (10.0), but even though I have an original copy on another machine, I cannot import this whole project, as the files are now too new.
    Can anyone help with sorting this out. I was going to upload a couple of small videos to show this, but the video upload button is greyed out.
    Thanks
    G

    Hi
    I have a problem with exporting a generator from Motion 5 to FCPX 10.0.1.
    1) I created a sequence of a page of a book opening then closing. I created this in Motion 5, and exported it as a Generator to its own category and folder
    2) I have found that in this folder a movie of the sequence is created. This movie plays correctly. The movie is located at this path:
    HD/ Users/ Name/ Movies/ Motion Templates/ Generators/ Generator Title/ Individual Generator/ Gen Movie.mov  HD/ Users/ Name/ Movies/ Motion Templates/ Generators/ Generator Title/ Individual Generator/ Gen Movie.mov 
    NOTE: I also tried exporting the movie, from motion as a stand alone movie. This and the original motion file play the same and are correct
    3) Once in FCPX. I can either skim over the generator or place in the time line and let it render
    4) The results are the same, the front cover does not turn correctly. As it closes it gets half way, and then the page disappears. A few frames later it flies in from the left side of the screen
    5) This did not happen in the original version of FCPX (10.0), but even though I have an original copy on another machine, I cannot import this whole project, as the files are now too new.
    Can anyone help with sorting this out. I was going to upload a couple of small videos to show this, but the video upload button is greyed out.
    Thanks
    G

  • How do you produce both Customs Export & Import documentation in an STO?

    Hello All -
    How do you produce both Customs Export & Import documentation in an STO scenario? For instance, if a company would like to create a STO in SAP to move materials from one plant to another and would like to produce both the Export & Import paperwork in GTS, what documents would be needed in R/3 to make this possible?
    Would the following be needed, or is there some other way?
    - STO
    - Outbound Delivery & F8 Invoice to generate the Customs Shipment in GTS for the Export documentation
    - Inbound Delivery to generate the Customs Declaration in GTS for the Import documentation

    Thanks Sameer -
    Yes, for the Import documentation, I think we will be declaring prior to Goods Receipt, so we will first use the Inbound delivery, and then the GR document later on once the materials have been received. (I believe you and I were talking about some of this in another thread actually.) Thank you for confirming this.
    As for the Note, I have also seen this - but thought that this was to more or less give visibility to the sending plant up-front at STO of what export licenses might be needed. (According to the note, currently they would only be able to check if licenses were needed at the outbound delivery, and not on the PO)
    Per the note: "Given a Stock Transport Purchase Order scenario: Plant A in Germany issues a STO PO against vendor plant US10 in the US. GTS looks at the PO in terms of importing from the US (departure country) coming into Plant A in Germany and performs compliance checks. Meanwhile, exporting plant US10 only has visibility at outbound delivery creation to see if am Export License is needed. Exporting plant US10 may want to know if they need a license to the country of destination, in this case Germany, at STO PO creation."
    However with the proper country-level export/import activation of the legal regulations, I believe that you can still have the STO check export-related licenses even though it is mapped to GTS as IMPORD. I have done some testing on this and it appears to work (you just need to ensure that you are activating the relevant countries for export/import properly in the legal regulation, and you can have both export & import documents check the same legal regulation and deliver the correct license determination). I was just wondering if you had any experience with this that was contrary to what I have found, or if there is something you think I am missing..
    Thanks

  • Migrate Database- Export/Import

    Hi,
    I need to migrate an Oracle database 9i from Sun Solaris to Linux. The final target database version would be 11g.
    Since this is a 9i database, I see that we have only option of export and database. We have around 15 schemas.
    I have some queries related to it.
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?
    Since there is a database version different I dont want to touch sys, system schema objects.
    Appreciate your thoughts.
    Regards
    Cherrish Vaidiyan

    Hi,
    Let me try to answer some of these questions you queried for:
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?Export won't export sys objects. For example, there are tables in sys, like obj$ that contain information for other metadata objects, like scott.emp, etc. This is not exported becuase when scott.emp is exported, the data from obj$ is essentially exported that way. When the dumpfile is imported and scott.emp is recreated, the data in sys.obj$ will be restored through the create table statement. As far as the SYSTEM schema is concerned, some objects are exported and some are not. There are tables in system that contain information about queues, jobs, etc. These will probably not make any sense on the target system so those types of tables are excluded from the export job. Other objects make sense to export/import so those are done. This is all figured out in the internals of export/import. Thre are other schemas that are not exproted. Some that I can think of are DMSYS, ORDSYS, etc. This would be for the same reason as SYS.
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?No, the dumpfiles are formatted differently. If you use exp, then you must use imp. If you use expdp, then you must use impdp. You can do exp on 9i and imp on 11g with the dumfile that was created on 9i.
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?This is case by case decision. It depends on what you want. If you want the complete database moved, then I would personally think that a full=y is what you would want to do. If you just did schema exports, then you would never export the tablespaces. This would mean that you would have to create the tablespaces on the source system before you ran imp. There are other objects that are not exported when a schema level export is performed that are exproted when a full is performed. This information can be seen in the utilities guide. Look to see what is exported when in user/schema mode vs full/database mode.
    Since there is a database version different I dont want to touch sys, system schema objects.This is all done for you with the internal workings of exp/imp.
    Dean
    Edited by: Dean Gagne on Jul 29, 2009 8:38 AM

  • Export/import to clone database -- please advise

    Hi All,
    Need help especially from exports that are strong in export/import.
    can someone advise me how to use export/import to clone a database?
    I know how to clone database using cold backup and hot backup.
    But wish to know the full steps in using export/import to do cloning.
    I'm going to do a full export on my "TEST" database, drop the database, create a database and import using the full export.
    After doing a full database export, full=y. What are the steps to follow. Have a few doubts about it.
    1) export the database (full=y)
    2) drop the database
    3) re-create the database, will be using the same name "TEST".
    4) create the users (IDs created by DBA, eg: schemas), tablespace and datafiles. --> import doesn't do it right?
    5) do a full import (full=y). In this step, will i encounter any problems? As in understand that all the oracle-owned objects (example SYSTEM, SYS, AUX) are already in the newly created database, will import full=y cause any problems?
    6) do i have to re-create the roles, SYNONYMS and grant role,sys,obj privileges to IDs not created by oracle. (example schema owner)
    in short i would like to know what EXPORT will export and not export
    what IMPORT will import and not import.
    Please advise.

    Hi zetabouy,
    thanks for ur input, definately helpful for me.
    just to confirm. am i right to say that only a full database level (full=y) export/import will export/import out roles as well as public synonyms? because i have tried also exporting a user from the database, and imported that user into a different database, but public synonyms and roles were not imported.
    one more question:
    For example we are going to do port data from production to uat database using export and import for only the schema.
    After exporting from production with such syntax (owner=OWNER01),
    is is better to drop all objects beloging to the schema owner in UAT before importing it (touser=OWNER01, ignore=y) using the production dumpfile?
    I'm asking this question as I'm afraid that if we did not drop all the pl/sql objects in UAT database, the updated procedures/fuctions/triggers exported from the production database will not get imported to the UAT database as UAT database have the same procedures/fuctions/triggers names (but not updated code) .
    will profiles be exported and imported too duing full database level exp/imp?
    please kindly advise.
    thanks
    Message was edited by:
    chew

Maybe you are looking for

  • All songs are gone, still taking up space

    Hi, just recently my iPod has been acting a bit weird and it has been having trouble connecting to my computer. I plugged it in last night and it failed to show up in iTunes or my Windows Explorer. I left it plugged in just in case it just had to tak

  • Burning coasters about 50% of the time recently

    Anyone else having this problem?  As far as I can tell it took place at the same time as the 2.6.19 upgrade and the switch to cdrkit.  My IDE dvd-rw drive is loaded as sr0/sg0 as I understand it should be after 2.6.19, and all of the required modules

  • Photoshop Script  Black and White Raster to Vector

    I do not have AI Adobe for I do not want to spend the money and have to learn an other complex Adobe application. I'm no artist either.  However at rimes I wanted  to have AI live trace feature to change black and white line drawing into vector graph

  • Cascading WRT54G to an existing network!!!

    Hi, The company bought a new WRT54G Wireless router and wants to cascade it to an existing network. Here's the setting: 1st Floor: ISP = Globe Broadband Modem/Router = Prolink Huricane 9300G IP=10.0.0.138 used as gateway Switch = Surecom IP=10.0.0.x

  • ON BUTTON.....

    Hi, my MacBook, it's "ON" button it isnot doing anything at all at all, can you please if you can help me with perhaps solutions, I thank you much Beijos, Isabella