Exporting index with partition export using expdp

Hi,
I am using EXPDP export In 11.1.0.7, is there a way i can export indexes along with the partition export of a table. With full table export, indexes are exported, but i don't see this with only a single partition export because i don't see indexes creation in following scenario.
1. Export a partition from the production table.
2. In development environment, drop all indexes on this table.
3. Drop same partition in development before importing it afresh from the export file created at first step.
4. Import the partition exported n first step (this does not automatically create indexes dropped in step 2.
5. Manually recrease indexes again.
Thanks

when you do a table mode export, indexes are included, unless you said exclude=index. Please list your expdp and impdp command and the log files to show that indexes are not included.
Thanks
Dean

Similar Messages

  • Error in export using expdp utility.

    Hi All,
    I am getting below error while doing expdp , version 11.2.0.2.0
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.GET_TABLE_DATA_OBJECTS []
    ORA-31642: the following SQL statement fails:
    BEGIN "SYS"."DBMS_CDC_EXPDP".SCHEMA_CALLOUT(:1,0,1,'11.02.00.02.00'); END;
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.DBMS_METADATA", line 1245
    ORA-01950: no privileges on tablespace 'XXXXX'
    Tried with ORA-1950 During Full or Schema Database Export [ID 454849.1] and DataPump Export Utility Fails With Errors ORA-39125 ORA-31642 PLS-201 on Package SYS.DBMS_CDC_EXPDP [ID 825680.1]
    But still the problem persists.
    Note: recently this db has migrated from 10g.
    Thanks in Advance.

    Hi;
    1. Please post your exact db version
    2. Please see:
    Export/Import DataPump: The Minimum Requirements to Use Export DataPump and Import DataPump (System Privileges) [ID 351598.1]
    ORA-1950 During Full or Schema Database Export [ID 454849.1]
    Regard
    Helios

  • ERROR when taking an export using EXPDP

    Hi There,
    I am receiving the following error when perform the scheduled nightly backup. The strange thing is the production instance has the same setup and it is not having this problem. The following parameters (memory_target, memory_max_target and streams_pool_size) are not being used in the production and non-production database base. They are all set to "0".
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    UDE-04031: operation generated ORACLE error 4031
    ORA-04031: unable to allocate 56 bytes of shared memory ("streams pool","unknown object","streams pool","fixed allocation callback")
    ORA-06512: at "SYS.KUPC$QUE_INT", line 431
    ORA-06512: at "SYS.KUPC$QUE_INT", line 1544
    ORA-06512: at "SYS.KUPC$QUE_INT", line 576
    ORA-25254: time-out in LISTEN while waiting for a message
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 3326
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 4551
    ORA-06512: at line 1
    UDE-04031: operation generated ORACLE error 4031
    ORA-04031: unable to allocate 56 bytes of shared memory ("streams pool","unknown object","streams pool","fixed allocation callback")
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 4373
    ORA-06512: at line 1
    UDE-31623: operation generated ORACLE error 31623
    ORA-31623: a job is not attached to this session via the specified handle
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 3326
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 4551
    ORA-06512: at line 1
    Thanks,
    J

    SQL> show parameter MEMORY_MAX_TARGET;
    NAME TYPE VALUE
    memory_max_target big integer 0
    SQL> SHOW PARAMETER SGA
    NAME TYPE VALUE
    lock_sga boolean FALSE
    pre_page_sga boolean FALSE
    sga_max_size big integer 768M
    sga_target big integer 768M
    SQL> SHOW PARAMETER PGA
    NAME TYPE VALUE
    pga_aggregate_target big integer 256M

  • Problems with Partition pruning using LIST option in 9i

    I am declaring the following partitions on a fact table and when
    I do an explain plan on the following SELECT statements it is
    doing a full table scan on the fact table. However if I use
    the "PARTITION" statement in the FROM clause it picks up the
    correct partition and runs great. I have used the analyze
    command and dbms_utility.analyze_schema command on all tables
    and indexes. I had simaliar problems when partitioning with the
    RANGE options too. I have looked through all of the INIT file
    parameters and don't see anything obvious. Is there something I
    am missing?
    Any help would be appreciated!
    Thanks,
    Bryan
    Facttable create statement....
    CREATE TABLE FactTable (
    ProductGID INTEGER NULL,
    CustomerGID INTEGER NULL,
    OrganizationGID INTEGER NULL,
    TimeGID INTEGER NULL,
    FactValue NUMBER NOT NULL,
    ModDate DATE NULL,
    CombinedGID INTEGER NOT NULL)
    PARTITION BY LIST (CombinedGID)
    (PARTITION sales1 VALUES(9999101),
    PARTITION sales2 VALUES(9999102),
    PARTITION sales3 VALUES(9999103),
    PARTITION model1 VALUES(9999204),
    PARTITION model2 VALUES(9999205),
    PARTITION model3 VALUES(9999206));
    Select statement....that is causing a full table scan.....the
    *tc tables are the equivelent to dimension tables in a star
    schema.
    SELECT tco.parentgid, tcc.parentgid, tcp.parentgid, sum
    (factvalue)
    FROM facttable f, custtc tcc, timetc tct, prodtc tcp, orgtc tco
    WHERE
    tco.childgid = f.organizationgid
    AND tco.parentgid = 18262
    AND tcc.childgid = f.customergid
    AND tcc.parentmembertypegid = 16
    AND tcp.childgid = f.productgid
    AND tcp.parentmembertypegid = 7
    AND tct.childgid = f.timegid
    AND tct.parentgid = 1009
    GROUP BY tco.parentgid, tcc.parentgid, tcp.parentgid;
    Select statement that works great....
    SELECT tco.parentgid, tcc.parentgid, tcp.parentgid, sum
    (factvalue)
    FROM facttable partition(sales1) f, custtc tcc, timetc tct,
    prodtc tcp, orgtc tco
    WHERE
    tco.childgid = f.organizationgid
    AND tco.parentgid = 18262
    AND tcc.childgid = f.customergid
    AND tcc.parentmembertypegid = 16
    AND tcp.childgid = f.productgid
    AND tcp.parentmembertypegid = 7
    AND tct.childgid = f.timegid
    AND tct.parentgid = 1009
    GROUP BY tco.parentgid, tcc.parentgid, tcp.parentgid;

    Hi Hoek,
    the DB version is 10.2 (italian version, then SET is correct).
    ...there's something strange: now I can INSERT rows but I can't update them!
    I'm using this command string:
    UPDATE TW_E_CUSTOMER_UNIFIED SET END_VALIDITY_DATE = TO_DATE('09-SET-09', 'DD-MON-RR') WHERE
    id_customer_unified = '123' and start_validity_date = TO_DATE('09-SET-09', 'DD-MON-RR');
    And this is the error:
    Error SQL: ORA-14402: updating partition key column would cause a partition change
    14402. 00000 - "updating partition key column would cause a partition change"
    *Cause:    An UPDATE statement attempted to change the value of a partition
    key column causing migration of the row to another partition
    *Action:   Do not attempt to update a partition key column or make sure that
    the new partition key is within the range containing the old
    partition key.
    I think that is impossible to use a PARTITION/SUBPARTITION like that: in fact the update of "END_VALIDITY_DATE" cause a partition change.
    Do u agree or it's possible an update on a field that implies a partition change?
    Regards Steve

  • Is it possible to export tables from diffrent schema using expdp?

    Hi,
    We can export tables from different schema using exp. Ex: exp user/pass file=sample.dmp log=sample.log tables=scott.dept,system.sales ...But
    Is it possible in expdp?
    Thanks in advance ..
    Thanks,

    Hi,
    you have to use "schemas=user1,user2 include=table:"in('table1,table2')" use parfileexpdp scott/tiger@db10g schemas=SCOTT include=TABLE:"IN ('EMP', 'DEPT')" directory=TEST_DIR dumpfile=SCOTT.dmp logfile=expdpSCOTT.log{quote}
    I am not able to perform it using parfile also.Using parfile it shows "UDE-00010: multiple job modes requested, schema and tables."
    When trying the below, i get error
    {code}
    bash-3.00$ expdp directory=EXP_DUMP dumpfile=test.dmp logfile=test.log SCHEMAS=(\'MM\',\'MMM\') include=TABLE:\"IN\(\'EA_EET_TMP\',\'WS_DT\'\)\"
    Export: Release 10.2.0.4.0 - 64bit Production on Friday, 15 October, 2010 18:34:32
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Username: / as sysdba
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYS"."SYS_EXPORT_SCHEMA_01": /******** AS SYSDBA directory=EXP_DUMP dumpfile=test.dmp logfile=test.log SCHEMAS=('MM','MMM') include=TABLE:"IN('EA_EET_TMP','WS_DT')"
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "MM"."EA_EET_TMP" 0 KB 0 rows
    ORA-39165: Schema MMM was not found.
    Master table "SYS"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    Dump file set for SYS.SYS_EXPORT_SCHEMA_01 is:
    /export/home/nucleus/dump/test.dmp
    Job "SYS"."SYS_EXPORT_SCHEMA_01" completed with 1 error(s) at 18:35:19
    {code}
    When checking expdp help=y shows :-
    {code}TABLES Identifies a list of tables to export - one schema only.{code}
    As per few testing,tables from different schemas are not possible to export using expdp in a single command.
    Anand

  • Can I export using a single dump directory to multiple locations in oracle

    I'm trying to do a full database export using expdp utility in oracle 10g. I have a single dump directory that is mapped to a particular file location say /export/dump. I don't want the entire dump file to be stored in the above path. Instead I want the dump to be distributed among multiple files. I know that this can be done using FILESIZE parameter which will distribute the contents to multiple files according to the size we have specified.
    My problems comes here, I want to export my data to multiple locations, the path is different than what I mentioned above, say /first/dump. Now my question is should I create multiple dump directories for each location before exporting or can I omit directory attribute in expdp and specify the complete path in the FILE parameter itself.

    No. EXPDP needs the server component LOGICAL DIRECTORY. If you don't specify the directory, it will go to the default EXPDP path which will be mostly /rdbms/log. It's defined by the parameters DATA_PUMP_DIR.
    You will have to specify directory attribute if you want to point your dumpfile to go to any specific location and you cannot give the directory path in the file name in EXPDP (unlike conventional exp)

  • Difference in file size between EM Export and Export using cmd.....

    Hi ,
    I use the following command in a command prompt to export a schema....
    C:\oracle\product\10.2.0\database10g\BIN\exp.exe system/*****@local FILE=C:\FILES.DMP ROWS=Y OWNER=USER_DB LOG=C:\TEST.LOG
    which produces a file around 900KB.
    When i exported on the same database , the same dataschema , both the data and the db objects ... using the EM the file produced was around 4000KB.....
    Why is this difference.....????
    I noticed in the log file of the exported file using EM some additional info... such as :
    'estimated USER_DB.TABLE_NAME 64KB'
    Can this info justify the difference in file size between the two methods..????
    Note: I use OraDb 10g v.2
    thanks a lot
    Simon

    The export done from EM is from Data Pump export.
    The manual export you did is original export.
    Try do a manual data pump export using expdp and compare the size.

  • Using expdp to export a mix of tables with data and tables without data

    Hi,
    I would like to create a .dmp file using expdp, exporting a set of tables with data and another set without data. Is there a way to do this in a single .dmp file? For example, I want all the tables in a schema with data, but for the fact tables in that schema, I only want the fact table objects, not the data. I thought it might be easier to create two separate .dmp files, one for each scenario, but would be nice to have one .dmp file that satisfies my requirement. Any help is appreciated.
    Thanks,
    -Rodolfo
    Edited by: user6902559 on May 11, 2010 12:05 PM

    You could do this with where clauses. Let's say you have 10 tables to export, 5 with data and 5 without data. I would do it like this
    tab1_w_data
    tab2_w_data
    tab3_w_data
    tab4_w_data
    tab5_w_data
    tab1_wo_data
    tab2_wo_data
    tab3_wo_data
    tab4_wo_data
    tab5_wo_data
    I would make one generic query
    query="where rownum = 0"
    and I would make 5 specific queries
    query=tab1_w_data:"where rownum > 0"
    query=tab2_w_data:"where rownum > 0"
    query=tab3_w_data:"where rownum > 0"
    query=tab4_w_data:"where rownum > 0"
    query=tab5_w_data:"where rownum > 0"
    The first query will be applied to all tables that don't have their own specific query and it will export no rows, the next 5 will apply to each of the corresponding table.
    Dean

  • Total number of records of partition tables exported using datapump

    Hi All,
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    OS: RHEL
    I exported a table with partitions using datapump and I would like to verify the total number of all the records exported of all these partition tables.The export has a query on it with WHERE ITEMDATE< TO_DATE (1-JAN-2010). I need it to compare with the exact amount of records in the actual table if it's the same.
    Below is the log file of the exported table. It does not show the total number of rows exported but only individually via partition.
    Starting "SYS"."SYS_EXPORT_TABLE_05": '/******** AS SYSDBA' dumpfile=data_pump_dir:GSDBA_APPROVED_TL.dmp nologfile=y tables=GSDBA.APPROVED_TL query=GSDBA.APPROVED_TL:"
    WHERE ITEMDATE< TO_DATE(\'1-JAN-2010\'\)"
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 517.6 MB
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q3" 35.02 MB 1361311 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q4" 33.23 MB 1292051 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q4" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q1" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q3" 30.53 MB 1186974 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q3" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q1" 30.44 MB 1183811 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q2" 30.29 MB 1177468 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2009_Q4" 30.09 MB 1170470 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q2" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q2" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2010_Q1" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q3" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2011_Q4" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q1" 5.875 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2006_Q3" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2006_Q4" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q1" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q2" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q3" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2007_Q4" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q1" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2008_Q2" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q2" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q3" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_2012_Q4" 0 KB 0 rows
    . . exported "GSDBA"."APPROVED_TL":"APPROVED_TL_MAXVALUE" 0 KB 0 rows
    Master table "SYS"."SYS_EXPORT_TABLE_05" successfully loaded/unloaded
    Dump file set for SYS.SYS_EXPORT_TABLE_05 is:
    /u01/export/GSDBA_APPROVED_TL.dmp
    Job "SYS"."SYS_EXPORT_TABLE_05" successfully completed at 12:00:36
    Edited by: 831134 on Jan 25, 2012 6:42 AM
    Edited by: 831134 on Jan 25, 2012 6:43 AM
    Edited by: 831134 on Jan 25, 2012 6:43 AM

    I assume you want this so you can run a script to check the count? If not and this is being done manually, then just add up the individual rows. I'm not very good at writing scripts, but I would think that someone here could come up with a script that would sum up the row count for the partitions of the tables from your log file. This is not something that Data Pump writes to the log file.
    Dean

  • How to export selected columns in a table using expdp of oracle10g

    Hi all..
    I have a table with 10 columns and i want to export only 4 columns(selected columns) data using expdp
    Pl. tell me if we can do this and if yes what is the syntax.
    Thanks..
    Sekhar

    That's not possible, you could use QUERY to specify where clause but not columns in select clause,
    for example,
    QUERY=employees:'"WHERE department_id > 10 AND salary > 10000"'
    you could use
    create table2export as select c1,c2,c3,c4 from tableof10columns;and export the temp table.

  • Using expdp to do a schema export is taking extremely long time

    I am using expdp in schema mode to export a 300 gigbyte database.
    The job status report 99% complete after about 2 hours .
    But now the job has been running 30 hours and is not finish.
    I can see that it is exporting the domain indexes and had been exporting
    the last index for the last 5 hours. Something is not working because I
    looked at the table the index is using and it has no data. So, why is it taking
    so long to export an index that has no data?
    Can someone tell if there is a way to bypass exporting indexes and a easy way
    to recreate the indexes if you do?
    I am using oracle 11g and expdp utility.

    I checked the log file and there are no errors in the file.
    There are no ORA- xxxx error messages.
    The last line in the log file is as follows:
    "Processing object type schema_export/table/index/domain_index/index "
    I just checked the export job run this morning and it is still on the same
    index object "A685_IX1" . This is a spatial index. It has been sitting at
    this same object according to the job status for at least 24 hours.

  • Exporting using Compressor failed, and other awekward issues with FCP

    Hi FCP gurus! I need your help!
    Everytime I try to export using compressor, my submission fails.
    I even try a sequence a exported as m2v on Monday, but today is not working.
    Compressor will some times say "Unable to view preview, FCP is busy". When I press submit, it goes to Batch Processor but it either will immediately show FAILED or it will just not do any processing. When I quit and restart, then the submission appears as failed.
    Also when exporting, FCP comes back with a pop message saying that 4 files went offline. When I try to reconnect them it says: "There are not clips selected that have files associated with them". So, why is FCP constantly not able to fine this files, if they are not been used?
    I need to export a file for my soundtrack guy, and he needs MOV, preferable DV-PAL or Motion-JPEG or similar frame-based codecs. What are my options?
    In any case, what the problem here is that I CANNOT export anything. What shall I do? I already restarted the computer twice, nothing.
    I realized that I have files using in this project, saved in two different external harddrives, do you think this has anything to do with the issue?
    I am using FCP 5.1.4
    Thanks so much in advance!

    First off, fwiw, I never "send to compressor."  I export a self-contained QT movie, not a conversion.  Then I run that through compressor with the appropriate settings.  I like the self contained file, it becomes a master from which I can make all sorts of variations.  The only difference, in theory between self contained and reference, is that if for any reason you lose the render files the reference movie will not work.
    Q1 No.
    Q2 see above
    Q3  There is no such thing as uncompressed.  It's just a different codec rather than a perfect copy of your timeline.
    Q4 lots of reasons for making a conversion, but none of them are valid for making DVD's
    Q5  Nothing wrong with letting DVDSP make the encoding-- just be sure to get into the preferences for DVDSP and set the destinations to a folder for the project, so you don't overwrite any other projects.  Depending on how long the project is, you can let it encode in the bg while you set up your project, build the buttons etc.  It's all the same encoding engine, whether is is done in compressor or dvdsp.   I find it much faster to use compressor then to take to dvdsp becaue you can use quickcluster.
    Now a couple of questions for you:  what sort of footage are you using and what are your fcp sequence settings?

  • I want to export with alpha channel using a codec that carries less data than gopro cineform

    Hi--I want to export some intermediate files with an alpha channel. i'm using gopro cineform, and these files end up being so large, meaning that i have to sit through hour long renders.  is there a a codec i can use that has an alpha chanel but creates a smaller file size?
    thanks

    Thanks for your help Patrick. I tell ya, I can't believe what's going on. As I said, this project until today has not caused me any issues what-so-ever. The other six sessions are totally fine. I also created a quick project with the same settings added text - then exported using the same settings I've used before, and it exports with the proper alpha channel.
    Yet, I rebooted the Mac, made sure to delete old versions in the trash can and everything. Still, no alpha.
    To keep moving forward, I literally had to use a color key in FCP to cut out the background color.
    Crazy!

  • I have Adobe Elements 9 and I used to be able to export HD 1080p videos with no issue. But recently, when I export using H.264 MPEG or MPEG 2, which I normally used, the high resolution videos come out very choppy.  How can i fix this?

    I have Adobe Elements 9 and I used to be able to export HD 1080p videos with no issue. But recently, when I export using H.264 MPEG or MPEG 2, which I normally used, the high resolution videos come out very choppy.  How can i fix this?

    mr
    Are you running the 9.0.1 Update of the program and on what computer operating system is it running?
    Is your video card/graphics card driver up to date according to the manufacturer of the card?
    Is your computer optimized, including maintenance of pile ups of preview files, conformed audio files, and conformed video files?
    Do you have the latest version of QuickTime installed on your computer with Premiere Elements 9.0/9.0.1?
    What are your project settings and what are the properties of your source media? When you say that "used to be able to export...." is that on the same computer, same brand/model/settings for the camera recording the video now in use, same Premiere Elements 9.0/9.0.1?
    Where player are you using for the playback of these exports?
    Please give us more details on the size of the projects (then and now) as well as details of the customized export settings in your Share/Computer/ route. Premiere Elements 9.0/9.0.1 is only a 32 bit application running in 32 bit or 64 bit system. Not sure if computer resources might be taxed more so now than before.
    More later.
    ATR

  • Export Database Link & Public Synonyms using Expdp

    Dear All,
    We are using expdp to do schema level exports:
    expdp dumpfile=<Directory>:expdp.dmp SCHEMAS=S1,S2,S3
    We also want to export the Public/Private Synonyms (created for the objects owned by above users), Public/Private Database Links as part of above command.
    How we can achieve this? We are using 11gR2.
    regards,
    Riaz

    HI,
    Public synonym can not be exported and never exported even at FULL Export also. . You can get them from "dba_synonyms". I suggest that you create so customer script .
    select      'create public synonym ' || table_name || ' for ' || table_owner || '.' || table_name || ';'
    from      dba_synonyms
    where
         owner='PUBLIC' and table_owner not in ('SYS', 'SYSTEM')
    order by
         table_owner;
    More: Information:
    https://forums.oracle.com/thread/855639?start=0&tstart=0
    Oracle Data Pump Schema Export and Public Synonyms
    Thank you

Maybe you are looking for

  • JDBC 2 on Oracle 8.0.5 on linux ?

    Hi there; I can't find oracle package for JDCB 2.0. In fact I need connection pooling and scrollable cursor features. Please save me ! null

  • Tomcat problem

    Hi, my problem is that after installing apache-tomcat-5.5.17.exe in xp, When i write http://localhost:8080/ in internet explorer after starting tomcat then it shows following error: error requested url could not retrieved While trying to retrieve the

  • List Output - how to position cursor back to Page 1, line 5, column 90.

    Hello, I have a report program that calculates and writes a Grand Total at the end of a multiple page report. The program can not calculate the Grand Total until the end of the program.  Pages are displayed as Subtotals are calculated. My user wants

  • Windows 8 Trackpad Driver

    I installed the release version of Windows 8 Pro, available through Microsoft DreamSpark Premium, on my late-2011 15" MacBook Pro through BootCamp. I had problems with the monitor and network communications, but after reinstalling the drivers from th

  • Recovering corrupted notes

    For some reason iCloud blew up my Notes (after blowing up my Numbers files earlier this week). So, after searching around these discussions, I found out that a Notes backup of the iCloud file exists in Containers or Mobile Documents, from which I sho