Importing partitioned table

Hi, everyone,
We have a partitioned table which has 18 partitiones. I have exported the partitioned table into one dump file. Now we want to import only 6 partitions of that partitioned table.
i want to clarify whether we need to precreate the table and import only those paritiones by specifiying like this:
imp xyz/nnn@orcl tables = (tablename:partitionname,tablename:partitionname,etc....) ignore = y
am i doing it correctly?
thanks in advance

Hi,
You are correct that you dont need to create the table prior to importing. Import command will create the table and required table and partitions for you. It will create all the partitions but will import the data of only partition which you are mentioning and you would then need to manually drop the partitions which are not needed.
I would suggest you to do the export of only partitions you want rather than exporting the full table and importing the only a few partitions.
Salman

Similar Messages

  • Importing partitioned table data into non-partitioned table

    Hi Friends,
    SOURCE SERVER
    OS:Linux
    Database Version:10.2.0.2.0
    i have exported one partition of my partitioned table like below..
    expdp system/manager DIRECTORY=DIR4 DUMPFILE=mapping.dmp LOGFILE=mapping_exp.log TABLES=MAPPING.MAPPING:DATASET_NAPTARGET SERVER
    OS:Linux
    Database Version:10.2.0.4.0
    Now when i am importing into another server i am getting below error
    Import: Release 10.2.0.4.0 - 64bit Production on Tuesday, 17 January, 2012 11:22:32
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "MAPPING"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
    Starting "MAPPING"."SYS_IMPORT_FULL_01":  MAPPING/******** DIRECTORY=DIR3 DUMPFILE=mapping.dmp LOGFILE=mapping_imp.log TABLE_EXISTS_ACTION=APPEND
    Processing object type TABLE_EXPORT/TABLE/TABLE
    ORA-39083: Object type TABLE failed to create with error:
    ORA-00959: tablespace 'MAPPING_ABC' does not exist
    Failing sql is:
    CREATE TABLE "MAPPING"."MAPPING" ("SAP_ID" NUMBER(38,0) NOT NULL ENABLE, "TG_ID" NUMBER(38,0) NOT NULL ENABLE, "TT_ID" NUMBER(38,0) NOT NULL ENABLE, "PARENT_CT_ID" NUMBER(38,0), "MAPPINGTIME" TIMESTAMP (6) WITH TIME ZONE NOT NULL ENABLE, "CLASS" NUMBER(38,0) NOT NULL ENABLE, "TYPE" NUMBER(38,0) NOT NULL ENABLE, "ID" NUMBER(38,0) NOT NULL ENABLE, "UREID"
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_TG_ID" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."PK_MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_UREID" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_V2" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."IDX_PARENT_CT" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    ORA-39112: Dependent object type CONSTRAINT:"MAPPING"."CKC_SMAPPING_MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"MAPPING"."PK_MAPPING_ITM" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_TG_ID" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."PK_MAPPING" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_UREID" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_V2" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_PARENT_CT" creation failed
    Processing object type TABLE_EXPORT/TABLE/COMMENT
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type COMMENT skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/REF_CONSTRAINT
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_MAPPING_MAPPING" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_MAPPING_CT" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_TG" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"MAPPING"."FK_TT" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_PART" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_TIME_T" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_DAY" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    ORA-39112: Dependent object type INDEX:"MAPPING"."X_BTMP" skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_TG_ID" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_V2_T" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."PK_MAPPING" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_PARENT_CT" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"MAPPING"."IDX_UREID" creation failed
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    ORA-39112: Dependent object type TABLE_STATISTICS skipped, base object type TABLE:"MAPPING"."MAPPING" creation failed
    Job "MAPPING"."SYS_IMPORT_FULL_01" completed with 52 error(s) at 11:22:39Please help..!!
    Regards
    Umesh Gupta

    yes, i have tried that option as well.
    but when i write one tablespace name in REMAP_TABLESPACE clause, it gives error for second one.. n if i include 1st and 2nd tablespace it will give error for 3rd one..
    one option, what i know write all tablespace name in REMAP_TABLESPACE, but that too lengthy process..is there any other way possible????
    Regards
    UmeshAFAIK the option you have is what i recommend you ... through it is lengthy :-(
    Wait for some EXPERT and GURU's review on this issue .........
    Good luck ....
    --neeraj                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • How to import partitioned tables in different tablespace

    Hi everyone,
    I try to import the partitioned tables in different tablespace.
    Consider the following situation:
    I have a dump file which is created by using "Export" utility. Some data are in partitioned tables, some of them are in non-partitioned tables. Also, all tables are located in "MYTBS" tablesapce. I try to import all data from this dump file to another database. I didn't get error messages when importing the data from non-partitioned tables. However, I got error message when importing the data from partitioned tables. The error message is: tablespace 'MYTBS' does not exist.
    I just want to how I can solve this problem other than create 'MYTBS' tablespace for my new database.
    Thanks in advance.
    Angel

    Hi,
    I got the following error message:
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE TABLE "FACILITYCONNECTION",....., "CONNECTIONTYPE" "
    "NUMBER(1, 0) NOT NULL ENABLE) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 25"
    "5 PARALLEL ( DEGREE DEFAULT INSTANCES 1) NOLOGGING STORAGE( PCTINCREASE 0) "
    "TABLESPACE "MYTBS" PARTITION BY RANGE ("CONNECTIONTYPE" ) (PARTITION "
    ""EXT" VALUES LESS THAN (1) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 S"
    "TORAGE(INITIAL 65536) TABLESPACE "MYTBS" NOLOGGING, PARTITION "FAC" VA"
    "LUES LESS THAN (2) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 STORAGE(I"
    "NITIAL 65536) TABLESPACE "MYTBS" NOLOGGING )"
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'MYTBS' does not exist
    Thanks.
    Angel

  • How to import partition table in 10 express edition database

    Hi all,
    I have 10g express edition installed on my laptop, I have a table with partition now I am unable to import that table in my oracle 10 express edition database.Anybody help me in this regard that how could I able to import this table in my database, even I dont want its partitions I want only data.
    Regards

    The phrase "awfully small table" was not intended as criticism. It merely notes that tables smaller than 2 GB would normally be too small to partition. Since you didn't specify the amount of data initially, it was reasonable to suspect that you might exceed the express edition data size limits.
    If you pre-create the table as a non-partitioned table (i.e. do a CREATE TABLE manually) before importing, I would expect that would allow the import to succeed (of course, you'd have to ignore the warning that the object already existed when it was time to import).
    Justin

  • Importing Partitioning tables, and Indexes

    Hello,
    during import of several partitioned tables, we had a strange problem. All non-PK indexes were imported into the partitioned tablespaces. PK indexes, did not seem to get created as local partitioned indexes. They have defaulted to a global PK in the a regular (read non-partitioned) tablespace.
    Is it normal ? Do we have to add any "special" commands or text to the import scripts for partitioned indexes ?
    Anyone, can you please post a reply to this, this has happened twice.
    Thanks, Sreedutt

    Note that this doesn't appear to be an ODBC-related question, but a question related to the Oracle import/export utility. Since there isn't an OTN forum that addresses this particular product, you're probably best served by contacting Oracle support.
    Justin Cave

  • Data Pump: Importing partitioned objects into target DB w/o partitioning

    Hello community,
    I have the following issue when importing partitioned tables from a source DB to a target DB without the partitioning option.
    In this scenario, the tables and indexes on the target are already present, solely without partitions. Release is 10gR2.
    <ul>
    <li> First try: impdp user/pw directory=dump_dir dumpfile=thedump tables=my_part_table TABLE_EXISTS_ACTION=truncate
    This stopped with "ORA-00439: feature not enabled: Partitioning", although the table already existed! Seems as if the partitioning is checked before the table_exists_action is executed.
    <li> Second try: impdp user/pw directory=dump_dir dumpfile=thedump tables=my_part_table TABLE_EXISTS_ACTION=append
    Table was truncated manually beforehand. Result: Same as above.
    <li> Third try: impdp user/pw directory=dump_dir dumpfile=thedump tables=my_part_table CONTENT=DATA_ONLY
    Table was truncated manually beforehand. Result: Success.
    BUT: This requires manual action on those objects and further means that you should manually de-/re-activate constraints and indexes on the table, at least when there's lots of data to load.
    </ul>
    So, option #3 is not that good of an option, as there's much more fiddling involved than it was earlier with "imp ... IGNORE=Y".
    Does anyone have a better idea how to handle this?
    I'd love to see a "use_partitions=N" option in impdp, but haven't found it so far. ;-)
    Regards,
    Uwe
    P.S.: Yes I know, the Index/Constraint part of it was more or less the same in the "imp age". Still, there's much more manual action needed than I am happy with.

    Hello community,
    I have the following issue when importing partitioned tables from a source DB to a target DB without the partitioning option.
    In this scenario, the tables and indexes on the target are already present, solely without partitions. Release is 10gR2.
    <ul>
    <li> First try: impdp user/pw directory=dump_dir dumpfile=thedump tables=my_part_table TABLE_EXISTS_ACTION=truncate
    This stopped with "ORA-00439: feature not enabled: Partitioning", although the table already existed! Seems as if the partitioning is checked before the table_exists_action is executed.
    <li> Second try: impdp user/pw directory=dump_dir dumpfile=thedump tables=my_part_table TABLE_EXISTS_ACTION=append
    Table was truncated manually beforehand. Result: Same as above.
    <li> Third try: impdp user/pw directory=dump_dir dumpfile=thedump tables=my_part_table CONTENT=DATA_ONLY
    Table was truncated manually beforehand. Result: Success.
    BUT: This requires manual action on those objects and further means that you should manually de-/re-activate constraints and indexes on the table, at least when there's lots of data to load.
    </ul>
    So, option #3 is not that good of an option, as there's much more fiddling involved than it was earlier with "imp ... IGNORE=Y".
    Does anyone have a better idea how to handle this?
    I'd love to see a "use_partitions=N" option in impdp, but haven't found it so far. ;-)
    Regards,
    Uwe
    P.S.: Yes I know, the Index/Constraint part of it was more or less the same in the "imp age". Still, there's much more manual action needed than I am happy with.

  • Getting error while importing a table partition

    Hi,
    I am trying to import a table partition from OEM and occurred with following error:
    Job IMPORT000042 has been reopened at Friday, 13 June, 2008 14:44
    Restarting "SYSMAN"."IMPORT000042":
    Processing object type TABLE_EXPORT/TABLE/TBL_TABLE_DATA/TABLE/TABLE_DATA
    ORA-31693: Table data object "SCOTT"."CONTAINER":"PARTITION_5" failed to load/unload and is being skipped due to error:
    ORA-06502: PL/SQL: numeric or value error
    LPX-00210: expected '<' instead of 'n'
    Job "SYSMAN"."IMPORT000042" completed with 1 error(s) at 14:44
    Job state: COMPLETED
    Thanks

    What's the source and target database Oracle version?
    What's the character set of both databases?

  • Error in import of partitioned table

    I have one partitioned table in one schema after taking export when I am trying
    to import in other database without having the same partition it is giving error
    ora-00959 tablespace "PART_FY06' does not exists.
    How to resolve it?

    Yes oracle is saying tablespace PART_FY06 does not exist. Create this tablespace. Export and Import are logical. For logical operation to get success, physical entities like tablespaces should exist.

  • Partition table export and import

    i am using Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - Production. And have a 133 partition in a single table. I want to take dump of each partition separately and restore it in a new schema. How to go about it.

    You can specify the partition you want to export by using the tables= parameter
    expdp user/password tables=your_table:your_partitoin ...
    If you eliminate your_partition, you will export the complete table. Do you want each partition to be a non-partitioned table? So if you have 133 partitions, do you want 133 tables? If so, just export the complete table and then on import use:
    impdp user/password partition_options=departitoin ...
    The result will be 1 table for each partition. The table will be called
    tablename_partitionname
    If it is too long, it gets truncated. The departition support was added in 11.1.0.6.
    Dean

  • Need help with importing data in partition table

    Hi,
    DB:8.1.7
    OS: win 2003 server
    I have a table which is 4.5GB in size. I created a new partition table,with local indexes. I exported the original table with all indexes,triggers and procedures.
    After creating the empty partitioned table,i imported the dump file. The data isn't loaded. My questions are:
    1) Will the indexes be overwritten?
    2) Will the data go automatically in allocated partitions?
    3) Do i need to export only table data ignoring indexes,triggers etc?
    Best Regards,

    ateeqrahman wrote:
    1) Will the indexes be overwritten?Not if they already exist with the same name, or with the same logical definition (ie. same columns are already indexed)
    2) Will the data go automatically in allocated partitions?Yes
    3) Do i need to export only table data ignoring indexes,triggers etc?Do you need the triggers? do you need indexes that you haven't created manually? What about grants?

  • Importing Some Partion of a Partitioned Table

    Hi Guru's,
    I import a partitioned table using original import option, during import i am successful to import 80% of data but after that import discontinued due to n/w error. I just want to know can i continue my import from where it is discontinued or i need to import the complete partitioned table again. The export dump file is complete partitioned table export not partition wise export.
    Any help will be greatly appreciated.
    Thank in advance.

    >
    I import a partitioned table using original import option, during import i am successful to import 80% of data but after that import discontinued due to n/w error. I just want to know can i continue my import from where it is discontinued or i need to import the complete partitioned table again. The export dump file is complete partitioned table export not partition wise export.
    >
    Use the TABLES clause in the import statement.
    See the Utilities doc
    http://docs.oracle.com/cd/B28359_01/server.111/b28319/dp_import.htm
    >
    TABLES
    Default: There is no default
    Purpose
    Specifies that you want to perform a table-mode import.
    Syntax and Description
    TABLES=[schema_name.]table_name[:partition_name]
    In a table-mode import, you can filter the data that is imported from the source by specifying a comma-delimited list of tables and partitions or subpartitions.
    If you do not supply a schema_name, it defaults to that of the current user. To specify a schema other than your own, you must either have the IMP_FULL_DATABASE role or remap the schema to the current user.
    The use of filtering can restrict what is imported using this import mode. See Filtering During Import Operations.
    If a partition_name is specified, it must be the name of a partition or subpartition in the associated table.
    >
    The doc has an example
    >
    The following example shows the use of the TABLES parameter to import partitions:
    impdp hr DIRECTORY=dpump_dir1 DUMPFILE=expdat.dmp TABLES=sh.sales:sales_Q1_2000,sh.sales:sales_Q2_2000

  • Importing a Partitioned Table with 10 Million Records.

    I've been trying to import from a dump file using:
    imp system/######@******** fromuser=fusr touser=tusr file=/f1/f2/expfl.dbf log=/o1/implg.log grants=N &
    import done in US7ASCII character set and UTF8 NCHAR character set
    import server uses UTF8 character set (possible charset conversion)
    This contains a Table 'Tab_Mil_Rec', with almost 10 millions of records and has 10 partitions.
    Done in 9i, on Solaris9.
    Problem is the process abruptly ends at 'Tab_Mil_Rec'. This table is created but nothing is imported. I checked the log file, but it has logged events before this table, but nothing (not evens errors or termination message) after that. No errors are thrown even at os level, I don't know exactly because this was done as backgruond job.
    Can anybody guess went wrong and whats the next step?

    Hi,
    Can you tried import partition by partition of this table ?
    Cheers

  • Importing into partitioned table from unpartitioned table

    I have taken export of a unpartitioned table using datapump.
    I want to import data using these dumpfiles into a partitioned table using impdp.
    Can you please provide me the basic script, I have tried running using the following script and I got the following error.
    noarg impdp $SCHEMA/$pw TABLES=\(ITPCS.ITPCS_BOM_GRT_SCAN\) EXCLUDE=STATISTICS parallel=4 dumpfile=dumpdir:exptable%U.dmp logfile=logdir:imptab2.log job_name=imptables_3 content=DATA_ONLY EXCLUDE=constraint,index,materialized_view
    Processing object type TABLE_EXPORT/TABLE/TBL_TABLE_DATA/TABLE/TABLE_DATA
    ORA-31693: Table data object "ITPCS"."ITPCS_BOM_GRT_SCAN" failed to load/unload and is being skipped due to error:
    ORA-00942: table or view does not exist
    ORA-06512: at "SYS.KUPD$DATA", line 1167
    ORA-14400: inserted partition key does not map to any partition
    Job "SYSTEM"."IMPTABLES_3" completed with 1 error(s) at 08:02

    pankaj2086 wrote:
    I have taken export of a unpartitioned table using datapump.
    I want to import data using these dumpfiles into a partitioned table using impdp.
    Can you please provide me the basic script, I have tried running using the following script and I got the following error.
    noarg impdp $SCHEMA/$pw TABLES=\(ITPCS.ITPCS_BOM_GRT_SCAN\) EXCLUDE=STATISTICS parallel=4 dumpfile=dumpdir:exptable%U.dmp logfile=logdir:imptab2.log job_name=imptables_3 content=DATA_ONLY EXCLUDE=constraint,index,materialized_view
    Processing object type TABLE_EXPORT/TABLE/TBL_TABLE_DATA/TABLE/TABLE_DATA
    ORA-31693: Table data object "ITPCS"."ITPCS_BOM_GRT_SCAN" failed to load/unload and is being skipped due to error:
    ORA-00942: table or view does not existWhat do you suppose the ora-00942 means? Looks pretty self explanatory to me. Have you looked it up? Have you verified that the target table exists in the target database?
    ORA-06512: at "SYS.KUPD$DATA", line 1167
    ORA-14400: inserted partition key does not map to any partitionAnd if the target table does exist, what do you suppose the ora-14400 means? Looks pretty self explanatory to me. Have you looked it up?

  • Import data from a partitioned table

    I am trying to import data from another server database. I need only data from two partitions from a partitioned table in that database. I have to use IMPORT wizard.
    The new server should also have the partitioned table with those two partitions data.

    I created the database on the destination server with a primary, log, FG1 and FG2 filegroups.
    Next, created the partition function and scheme
    create partition function pfOrders(int)
    as range right
    for values(34);
    create partition scheme psOrders
    as partition pfOrders
    to (FG1,FG2)
    go
    Next, created the table on primary with primary key.
    Next, used import wizard twice to import data related to partition_id 34 and 65.
    The problem is the data is on primary filegroup. FG1 and FG2 does not have any data.

  • Can I use gpt to recreate/unerase a partition table? (Rebuild the GPT/GUID partition table?)  I don't want to do FILE recovery.

    (Yes, I've  googled a bunch and read threads like this one already.)
    Can I use gpt or some other app to recreate/unerase a partition table?  That is, how can I rebuild a disk's GPT/GUID partition table?)  I don't want to do FILE recovery.
    What happened: Instead of erasing a single partition off a disk with many partitions, the entire partition table was erased (using Disk Utility, w/o deleting the underlying files).  Somehow the "Erasing a disk deletes all data on all its partitions." warning message was missed.
    I have a copy of the output of df, with the number of blocks in each partition, from just prior to the erasure, so I should be able to recreate the GPT/GUID partition table.  Editing the GPT with a hex editor is not feasible.  Simply recreating the partitions with Disk Utility will overwrite the key filesystem tables on each partition, and I don't want to do that, plus Disk Utility doesn't allow me to specify exact partition sizes anyway.
    Surely there's an app for rebuilding the partition table (other than emacs' hexl-mode!) for recreating/unerasing a partition table when the partition sizes and orders are known?  I've looked at the advertising for a bunch of recovery software and none of them clearly indicate that they will do what I want. 
    I guess I can try using gpt on a copy of the reformatted drive I've made with dd, and see what happens.  But perhaps someone knows of a tool that should do what I need, or knows if gpt is that tool or not.
    There are answers and tools that will do FILE recovery - search for files and recover the ones that aren't fragmented or deleted.  As far as I can find, they just look for files on the disk, and don't pay much, if any attention to the filesystem info or directory heirarchy, which in this case is valuable.  Of course I could send it in to DriveSavers, or the like.  But none of that seems necessary, and the scavenging file recovery apps won't do the job well,
    E.g. some are mentioned here:
    I don't want to do FILE recovery.
    Thanks for any help.
    The links in this post are to pages describing the underlined term, e.g. the man pages for df and gpt.
    dd output includes:
    Filesystem
    512-blocks 
    Used Available Capacity  Mounted on

    Aperture has the ability to work with files in their existing location. They are called "referenced masters." When you import images, you should select the "In their current location" in the "Store Files:" drop down box. Have a read of the documentation for full specifics. Unsure how you can resolve your duplication; might be some work but next time have a read of the manual first
    Information for versions is stored in the Aperture database (library file). The masters can be inside the library file itself, or they can be somewhere else.

Maybe you are looking for

  • How to set up and sync iTunes with different accounts??

    Is there any way to sync or setup multiple iTunes accounts that are already set up on the same computer so that we can all share apps and music?? I have been doing a lot of research and I can't find anything that works. Please help!

  • Dynamic endpoints

    all, i was required to manually edit a service's wsdl in order to declare faulting. by manually editing the wsdl it is no longer directly related to the service, requiring me to use the @jws:wsdl annotation. the problem that i'm having is that the en

  • Assets internally transfer Report

    Hi experts, My client is Transferring assets one storage location to other location in a one Company code. . Some time they transfer fully and some time partially. Now requirement is, client needs a Report regarding this assets transfer internally Is

  • Reduce Redo

    Hi Folks RDBMS: 11gR1 we have a finance controlling database in archivelog mode which performs huge calculations from time to time. the size of the db is 10gb, but when such a report run it generates ~100gb in a couple of hours. our goal is to reduce

  • Dynamic images in JSP

    Hi! I would like to generate dynamic images in a JSP which I would like do an "include" in other JSPs across the site. The reason for this is we show many emails/ phone numbers on various pages and with the prevalence of spiders/ robots, it would be