Import of Tables with CLOB problem.

Hi,
I am having problems with the import of tables with CLOB columns.
Tables with CLOB columns will force the use of the tablespace in which it (table) was created originally. It will not be created on the new schema's default tablespace. Is there a way to get around this problem?
More on this problem can be found here :
Re: Import error 1647, CLOB fields problem, need help!
I am using Oracle 9.2.0.6 standard edition, that is why the transportable tablespace option will not work.
Thanx
Rob

Sorry I have posted the wrong link, it should be:
Re: Import error 1647, CLOB fields problem, need help!
Regards
Rob

Similar Messages

  • Problem importing a table with blob's

    hi all, I'm facing the following situation.
    Source DB : 10.2.0.3 (client's DB)
    Destination DB (mine): 10.2.0.4
    I've a dump file (traditional) of a particular schema.
    I'm running import (imp) to import on my DB.
    It runs fine until it reaches one particular table. This table has 6 colums, 3 of them are BLOB.
    This table has 260000 rows (checked with export log).
    When import reaches row 152352 it stops loading data, but import is still running.
    what can I do to get more information from this situation in order to solve this problem?
    Any suggestion will be appreciated!
    Thanks in advance.

    Pl identify the source and target OS versions. Are there any useful messages in the alert.log ? How long did the export take ? Rule of thumb states import will take twice as long. Have you tried expdp/impdp instead ? Also see the following -
    How To Diagnose And Troubleshoot Import Or Datapump Import Hung Scenarios          (Doc ID 795034.1)
    How To Find The Cause of a Hanging Import Session          (Doc ID 184842.1)
    Import is Slow or Hangs          (Doc ID 1037231.6)
    Export and Import of Table with LOB Columns (like CLOB and BLOB) has Slow Performance          (Doc ID 281461.1)
    HTH
    Srini

  • Problem import excel table with photos

    When I import excel table with photos, the photos don't appear

    Welcome to Project Siena!
    If you're seeing an 'x' instead of your photo in the data source it's possible that Project Siena can't access the directory.  Try saving your images in another folder such as C:\Users\Public\Pictures.
    Others have also reported that if the Library isn't set up correctly in Windows 8 that they've had issues.
    Here are some posts to look at while we wait for additional information from you:
    Importing local images via an Excel file
    Siena Gallery unable to load image
    Images in Public Pictures directories showing
    up as X in Img from URL
    (fyi - this last post's screenshot will look different than yours as it was from Beta 1.)
    Thor

  • Understanding logminer results -- inserting row into table with CLOB field

    In using log miner I have noticed that inserts into rows that contain a CLOB (I assume this applies to other LOB type fields as well, have only tested with CLOB so far) field are actually recorded as two DML entries.
    --the first entry is the insert operation that inserts all values with an EMPTY_CLOB() for the CLOB field
    --the second entry is the update that sets the actual CLOB value (+this is true even if the value of the CLOB field is not being set explicitly+)
    This separation makes sense as there may be separate locations that the values are being stored etc.
    However, what I am tripping over is the fact the first entry, the Insert, has a RowId value of 'AAAAAAAAAAAAAAAAAA' which is invalid if I attempt to use it in a flashback query such as:
    SELECT * FROM PERSON AS OF SCN #####'  where RowId = 'AAAAAAAAAAAAAAAAAA'The second operation, the Update of the CLOB field, has the valid RowId.
    Now, again, this makes sense if the insert of the new row is not really considered "+done+" until the two steps are done. However, is there some way to group these operations together when analyzing the log contents to know that these two operations are a "+matched set+"?
    Not a total deal breaker, but would be nice to know what is happening under the hood here so I don't act on any false assumptions.
    Thanks for any input.
    To replicate:
    Create a table with CLOB field:
    CREATE TABLE DEVUSER.TESTTABLE
            ID NUMBER
           , FULLNAME VARCHAR2(50)
          , AGE NUMBER  
          , DESCRIPTION CLOB
           );Capture the before SCN:
    SELECT DBMS_FLASHBACK.GET_SYSTEM_CHANGE_NUMBER FROM DUAL;Insert a new row in the test table:
    INSERT INTO TESTTABLE(ID,FULLNAME,AGE) VALUES(1,'Robert BUILDER',35);
         COMMIT;Capture the after SCN:
    SELECT DBMS_FLASHBACK.GET_SYSTEM_CHANGE_NUMBER FROM DUAL;Start logminer session with the bracketing scn values and options etc:
    EXECUTE DBMS_LOGMNR.START_LOGMNR(STARTSCN=>2619174, ENDSCN=>2619191, -
               OPTIONS => DBMS_LOGMNR.DICT_FROM_ONLINE_CATALOG + DBMS_LOGMNR.CONTINUOUS_MINE + -
               DBMS_LOGMNR.COMMITTED_DATA_ONLY + DBMS_LOGMNR.NO_ROWID_IN_STMT + DBMS_LOGMNR.NO_SQL_DELIMITER)Query the logs for the changes in that range:
    SELECT
           commit_scn, xid,operation,table_name,row_id
           ,sql_redo,sql_undo, rs_id,ssn
           FROM V$LOGMNR_CONTENTS
        ORDER BY xid asc,sequence# ascResults:
    2619178     0C00070028000000     START                  AAAAAAAAAAAAAAAAAA     set transaction read write
    2619178     0C00070028000000     INSERT     TESTTABLE     AAAAAAAAAAAAAAAAAA     insert into "DEVUSER"."TESTTABLE" ...
    2619178     0C00070028000000     UPDATE     TESTTABLE     AAAFEXAABAAALEJAAB     update "DEVUSER"."TESTTABLE" set "DESCRIPTION" = NULL ...
    2619178     0C00070028000000     COMMIT                  AAAAAAAAAAAAAAAAAA     commitEdited by: 958701 on Sep 12, 2012 9:05 AM
    Edited by: 958701 on Sep 12, 2012 9:07 AM

    Scott,
    Thanks for the reply.
    I am inserting into the table over a database link.
    I am using the new version of HTML Db (2.0)
    HTML Db is connected to an Oracle 10 database I think, however the table I am trying to insert data into (via the database link) is in an Oracle 8 database - this is why we created a link to it as we couldn't have the HTML Db interacting with the Oracle 8 database directly due to compatibility problems (or so I've been told)
    Simon

  • Exporting Table with CLOB Columns

    Hello All,
    I am trying to export table with clob columns with no luck. It errors saying EXP-00011TABLE do not exist.
    I can query the table and the owner is same as what i am exporting it from.
    Please let me know.

    An 8.0.6 client definitely changes things. Other posters have already posted links to information on what versions of exp and imp can be used to move data between versions.
    I will just add that if you were using a client to do the export then if the client version is less than the target database version you can upgrade the client or better yet if possilbe use the target database export utility to perform the export.
    I will not criticize the existance of an 8.0.6 system as we had a parent company dump a brand new 8.0.3 application on us less than two years ago. We have since been allowed to update the database and pro*c modules to 9.2.0.6.
    If the target database is really 8.0.3 then I suggest you consider using dbms_metadata to generate the DDL, if needed, and SQLPlus to extact the data into delimited files that you can then reload via sqlldr. This would allow you to move the data with some potential adjustments for any 10g only features in the code.
    HTH -- Mark D Powell --

  • How to import a table with renamed name

    If I want to import a table from a DMP on UNIX towards my user where I already have a new version but I don't want to drop this new version 'cause I have only to move some records from one to the other, is there an import option for importing the table with another name? I mean I want to import on the same user, where there already exists the table MYTABLE, an old version of MYTABLE by impoting it with the new name MYTABLE_OLD, so afterwards I can update MYTABLE with some of old records of its old versione MYTABLE_OLD. Is it possible?

    You cannot directly do this.
    You can import the Table into a different schema and then rename the table.
    export and import back into the required schema.
    Import DataPump: How to Import Table Data into a Table that has Different Name ? [ID 342314.1]
    Regards
    Rajesh

  • Oracle 11g imp erroneously tries to recreate existing tables with CLOBs?

    I have a shell script for loading database dumps from both Datapump and the older exp/imp.
    Often when loading dumps, I need to rename the schema owner and tablespace names (which is handled by REMAP_SCHEMA and REMAP_TABLESPACE in Datapump).
    However I have a whole bunch of dumps created with exp at this point and not that many Datapump dumps yet. As such the old style dumps are handled by the shell script in this way:
    1) A first pass imp is run using INDEXFILE to generate a file with the SQL to create tables and indexes. Options also include FROMUSER and TOUSER.
    2) A series of sed command edit the SQL file to change the tablespace names (which are schema owner specific in our case).
    3) The editted SQL file is run with sqlplus to create the tables and indexes.
    4) A second pass imp is run to load the table rows as well as triggers, stored procedures, views, etc. Options include FROMUSER, TOUSER, COMMIT=Y, IGNORE=Y, BUFFER, STATISTICS=NONE, CONSTRAINTS=N
    This shell script has been working great for loading exp dump files into Oracle 9 and Oracle 10 databases, but now that I'm trying to load these dumps into Oracle 11, it fails.
    The problem is in step 4, the imp program is trying to create some of the tables that already were created with sqlplus in step 2. The problematic tables all seem to have CLOB columns in them. The table creation fails because it tries to use the tablespace names from the dump file, which do not exist in the destination database. And when the table creation fails, imp then decides not to load the rows for those table.
    This seems like a bug in the Oracle 11 imp program. I don't understand why it thinks it needs to recreate tables that already exist when those tables have CLOB columns. Is there something different about CLOB columns in Oracle 11 that I should know about that might be confusing imp into thinking that it needs to create tables when they already exist? Maybe I need to do something to those tables in SQL so that imp does not think it needs to recreate them?
    I know that the tables with the CLOBs were created correctly because I was trying to find some way to workaround this. For step 4, I tried using DATA_ONLY=Y, in which case imp does not try to create the tables and just loads the table rows. Of course using DATA_ONLY, I don't get a lot of other things like triggers, view and stored procedures. I started to try to get around that by doing 3 passes with imp, so that I could pick up the missing pieces by using an imp pass with ROWS=N, but strangely that has the same problem of trying to recreate the existing tables.

    The only solution I've found so far as a workaround is rather convoluted.
    1. I took an export using datapump's expdp of SCHEMA1 (in 10g it will skip the table with the xmltype).
    2. I imported the data to my empty schema (SCHEMA2) using impdp. To avoid the error that the type already exists with another OID, I used the TRANSFORM=oid:n parameter e.g.
    impdp user/pwd dumpfile=noxmltable.dmp logfile=importallbutxmltable.log remap_schema=SCHEMA1:SCHEMA2 TRANSFORM=oid:n directory=MYDUMPDIR
    3. I then manually created my xmltype table in the SCHEMA2 and did a select into to load it (make sure you have the select privileges to do so):
    INSERT INTO SCHEMA2.XMLTABLE2 SELECT * FROM SCHEMA1.XMLTABLE1;
    4. I am still taking an export with exp of the xmltable as well even though I'm not sure I can do anything with it.
    Thanks!
    Edited by: stacyz on Jul 28, 2009 9:49 AM

  • Import the table with 0 rows

    Hi
    I have a problem to import a dump that contains the tables with 0 rows.
    When i exported from ORACLE 11.2 64 bit on SERVER 2008 i noticed that log didn't confirm the tables with 0 rows.
    When i want to import to ORACLE 11.2 64 bit on other SERVER 2008 i have a lot of errors on this tables with 0 rows.
    In the log i get the same tables with 1 row at least, but no one with 0 rows.
    I open my dump in TEXTPAD and i see it contains "CREATE ....." these tables.
    I don't understand why it happens. I used FUll DUMP by SYS, it didn't help.
    This is not first time when i export and import dumps,no errors.
    I'm using command "EXP" and "IMP" and every time it's ok.(IF it's a releavent)
    Why it happens? any solutions for this issue?
    Thanks

    I've found (i guess so) solution to this issue
    here are two links to this new feature that is called deffered segment creation
    The reason for this behavior is 11.2 new feature ‘Deferred Segment Creation‘ – the creation of a table sent is deferred until the first row is inserted.
    As a result, empty tables are not listed in dba_segments and are not exported by exp utility
    http://www.nativeread.com/2010/04/09/11gr2-empty-tables-skipped-by-export-deferred-segment-creation/
    http://antognini.ch/2010/10/deferred-segment-creation-as-of-11-2-0-2/
    And this is i've found in official documentation from oracle
    Beginning in Oracle Database 11g Release 2, when creating a non-partitioned heap-organized table in a locally managed tablespace, table segment creation is deferred until the first row is inserted. In addition, creation of segments is deferred for any LOB columns of the table, any indexes created implicitly as part of table creation, and any indexes subsequently explicitly created on the table.The advantages of this space allocation method are the following:A significant amount of disk space can be saved for applications that create hundreds or thousands of tables upon installation, many of which might never be populated.Application installation time is reduced.There is a small performance penalty when the first row is inserted, because the new segment must be created at that time.
    To enable deferred segment creation, compatibility must be set to '11.2.0' or higher. You can disable deferred segment creation by setting the initialization parameter DEFERRED_SEGMENT_CREATION to FALSE. The new clauses SEGMENT CREATION DEFERRED and SEGMENT CREATION IMMEDIATE are available for the CREATE TABLE statement. These clauses override the setting of the DEFERRED_SEGMENT_CREATION initialization parameter.
    +Note that when you create a table with deferred segment creation (the default), the new table appears in the _TABLES views, but no entry for it appears in the SEGMENTS views until you insert the first row. There is a new SEGMENTCREATED column in _TABLES, _INDEXES, and _LOBS that can be used to verify deferred segment creation+
    Note:
    The original Export utility does not export any table that was created with deferred segment creation and has not had a segment created for it. The most common way for a segment to be created is to store a row into the table, though other operations such as ALTER TABLE ALLOCATE EXTENTS will also create a segment. If a segment does exist for the table and the table is exported, the SEGMENT CREATION DEFERRED clause will not be included in the CREATE TABLE statement that is executed by the original Import utility.

  • Error while importing a table with BLOB column

    Hi,
    I am having a table with BLOB column. When I export such a table it gets exported correctly, but when I import the same in different schema having different tablespace it throws error
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE TABLE "CMM_PARTY_DOC" ("PDOC_DOC_ID" VARCHAR2(10), "PDOC_PTY_ID" VAR"
    "CHAR2(10), "PDOC_DOCDTL_ID" VARCHAR2(10), "PDOC_DOC_DESC" VARCHAR2(100), "P"
    "DOC_DOC_DTL_DESC" VARCHAR2(100), "PDOC_RCVD_YN" VARCHAR2(1), "PDOC_UPLOAD_D"
    "ATA" BLOB, "PDOC_UPD_USER" VARCHAR2(10), "PDOC_UPD_DATE" DATE, "PDOC_CRE_US"
    "ER" VARCHAR2(10) NOT NULL ENABLE, "PDOC_CRE_DATE" DATE NOT NULL ENABLE) PC"
    "TFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 STORAGE(INITIAL 65536 FREELISTS"
    " 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "TS_AGIMSAPPOLOLIVE030"
    "4" LOGGING NOCOMPRESS LOB ("PDOC_UPLOAD_DATA") STORE AS (TABLESPACE "TS_AG"
    "IMSAPPOLOLIVE0304" ENABLE STORAGE IN ROW CHUNK 8192 PCTVERSION 10 NOCACHE L"
    "OGGING STORAGE(INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEF"
    "AULT))"
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'TS_AGIMSAPPOLOLIVE0304' does not exist
    I used the import command as follows :
    imp <user/pwd@conn> file=<dmpfile.dmp> fromuser=<fromuser> touser=<touser> log=<logfile.log>
    What can I do so that this table gets imported correctly?
    Also tell me "whether the BLOB is stored in different tablespace than the default tablespace of the user?"
    Thanks in advance.

    Hello,
    U can either
    1) create a tablespace with the same name in destination where you are trying to import.
    2) get the ddl of the table, modify the tablespace name to reflect the existing tablespace name in destination and run the ddl in the destination database, and run your import command with option ignore=y--> which will ignore all the create errors.
    Regards,
    Vinay

  • Fireworks Imports - Replacing Tables with Divs

    Hi folks,
    I love Fireworks to bits but it's intergration with DW is slowly driving me nuts.
    Take a look at the code for www.total-formula1.com
    I've imported a nav section, top bar and footer from Fireworks - which (IMHOP) look great until you look at the code.
    Then it becomes a mess of tables and spacer gif's.
    I've tried importing as CSS with images but it looses behaviours.
    Am I kidding myself. Do I have to split it all up and use DW to perform rollovers?
    Thoughts, help or a loaded shotgun welcome.

    Fireworks CSS and Images export does not support Javascript. Same with States, if I recall. CSS and Images will not export out anything but the current state. You would have to add that code (CSS or JS) in by using Dreamweaver or another editor.
    If properly set up, you may be able to export the entire page as CSS and images, then only have to add in the CSS to create the mouseover effect, after they've been exported as graphics.
    HTH
    Jim Babbage

  • Importing a table with a BLOB column is taking too long

    I am importing a user schema from 9i (9.2.0.6) database to 10g (10.2.1.0) database. One of the large tables (millions of records) with a BLOB column is taking too long to import (more that 24 hours). I have tried all the tricks I know to speed up the import. Here are some of the setting:
    1 - set buffer to 500 Mb
    2 - pre-created the table and turned off logging
    3 - set indexes=N
    4 - set constraints=N
    5 - I have 10 online redo logs with 200 MB each
    6 - Even turned off logging at the database level with disablelogging = true
    It is still taking too long loading the table with the BLOB column. The BLOB field contains PDF files.
    For your info:
    Computer: Sun v490 with 16 CPUs, solaris 10
    memory: 10 Gigabytes
    SGA: 4 Gigabytes

    Legatti,
    I have feedback=10000. However by monitoring the import, I know that its loading average of 130 records per minute. Which is very slow considering that the table contains close to two millions records.
    Thanks for your reply.

  • Dynamic region table with binding problem

    Hi. I'm using jdev 11.1.1.3 and ADF full stack.
    My app uses single page approach with ADF libraries and a dynamic region to show content. Main page includes a menu bar which is created on session startup and is used to load task flows from libraries.
    I have a class to provide some table functionality like CRUD, multi-row delete, etc and I need to reuse with every table included in libraries.
    To do this, in main web app I have a managed bean to hold table binding and to provide this common functions to all fragments if needed.
    I think it must be in request scope as fragment changes.
    Here's the problem: when I navigate to a fragment which contains a table with binding, first time goes well,
    but next time table, when fragment changes,new table displays without header or without data.
    If I go to another fragment without binding and return to a fragment with problem, it displays well. With fragments without table binding, it works OK.
    I've tried changing scope of this 'general' managed bean, but same thing. If I put a bean for each fragment, it works OK.
    So, how can I 'share' this managed bean through every table? Any suggestions? is it possible?
    Please, help me ....
    Thanks,
    DEMR.

    I am seeing a similar problem in my app. I have a main menu that drives a dynamic region. If the task flow that gets displayed in the dynamic region also contains a dynamic region, when the inner dynamic region is refreshed to point to a second task flow, the page does not render correctly. The errors we have seen include data in tables not being displayed in some cases, and PPR errors in other cases. In both cases, if the inner dynamic region is replaced (by a single bounded task flow with navigation logic in a template, for example) the errors go away. There appears to be issues involved with including a dynamic region within a dynamic region.
    Have you made any progress on solving your issue?

  • Export data from table with CLOB column

    DB: Oracle10g
    SQLDeveloper: 1.0
    OS: windows xp for sqldeveloper
    Linux for Oracle db
    we have a table with two CLOB columns. We want to export the data to either text or csv. But the CLOB columns don't show up in the columns that can be exported.
    How do I export CLOB data from SQL Developer?
    Regards,
    Lonneke

    I don't think this is a good protection: you can have these characters in VARCHAR anyway, and this way you "throw the baby out with the bath water". Not mentioning that right now also immune formats like HTML are also limited.

  • [CS4/JS] Import CALS-Tables with merged cells

    Hello scripters.
    I am actually trying to import CALS-tables (which is in my imported XML) into InDesign.
    I used the convertElementToTable for this.
    That worked, except for the merged cells (only cells which are merged to two or three columns).
    How could I script the automatic generation for merged cells?
    My xml-cals-table looks the following:
    <example>
        <table class="n" width="600" border="0">
            <colgroup>
                <col width="35%"/>
                <col width="30%"/>
                <col width="35%"/>
            </colgroup>
            <tr>
                <td colspan="3">
                    <ta>content content content content content content content content</ta>
                </td>
            </tr>
            <tr>
                <td align="left" valign="top">
                    <ta>content content content content </ta>
                </td>
                <td align="left" valign="top">
                    <ta>content content </ta>
                </td>
                <td align="right" valign="top">
                    <ta>content content </ta>
                </td>
            </tr>
            <tr>
                <td align="left" valign="top">
                    <ta>content content </ta>
                </td>
                <td align="left" valign="bottom">
                    <ta>content content </ta>
                </td>
                <td align="right" valign="bottom">
                    <ta>content content </ta>
                </td>
            </tr>
        </table>
    </example>
    Thanks in advance for any help!
    Franziska

    Hello scripters.
    I am actually trying to import CALS-tables (which is in my imported XML) into InDesign.
    I used the convertElementToTable for this.
    That worked, except for the merged cells (only cells which are merged to two or three columns).
    How could I script the automatic generation for merged cells?
    My xml-cals-table looks the following:
    <example>
        <table class="n" width="600" border="0">
            <colgroup>
                <col width="35%"/>
                <col width="30%"/>
                <col width="35%"/>
            </colgroup>
            <tr>
                <td colspan="3">
                    <ta>content content content content content content content content</ta>
                </td>
            </tr>
            <tr>
                <td align="left" valign="top">
                    <ta>content content content content </ta>
                </td>
                <td align="left" valign="top">
                    <ta>content content </ta>
                </td>
                <td align="right" valign="top">
                    <ta>content content </ta>
                </td>
            </tr>
            <tr>
                <td align="left" valign="top">
                    <ta>content content </ta>
                </td>
                <td align="left" valign="bottom">
                    <ta>content content </ta>
                </td>
                <td align="right" valign="bottom">
                    <ta>content content </ta>
                </td>
            </tr>
        </table>
    </example>
    Thanks in advance for any help!
    Franziska

  • Problem importing a table with data in an existing schema

    Hi,
    I am trying to import a schema and found an error while importing
    ORA-39083: Object type TABLE failed to create with error:
    ORA-01843: not a valid month
    Failing sql is:
    CREATE TABLE "AMBERSAP"."QUOTE" ("QUOTE_ID" VARCHAR2(15) NOT NULL ENABLE, "CONTRACT_ID" VARCHAR2(12), "QUOTE_OWNER" VARCHAR2(100) NOT NULL ENABLE, "TEMP_CONTRACT" VARCHAR2(1) DEFAULT 'N', "QUOTE_STATUS_CODE" VARCHAR2(1), "QUOTE_INITIATE_DATE" DATE NOT NULL ENABLE, "QUOTE_EXPIRY_DATE" DATE NOT NULL ENABLE, "EXTENDED_EXPIRY_DATE" DATE, "OFFER_DAYS" NUMBER(*,0), "OFFER_CLOSED_DATE" DATE, "BRAND_CODE" VARCHA
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Any feedback on this is highly apprciated.
    Thanks in advance .

    I've set the NLS_DATE_FORMAT='DD/MM/YYYY'
    by export NLS_DATE_FORMAT='DD/MM/YYYY'
    but still failing to create the quote table while i try to import the whole schema.
    here is the error message
    wlgora(oracle)[HDM_D]/db/HDM_D/dba/exp$: file=amberquota1.dmp logfile=impquote1.log TABLE_EXISTS_ACTION=APPEND <
    Import: Release 10.2.0.4.0 - 64bit Production on Tuesday, 11 October, 2011 14:52:58
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, Real Application Clusters, OLAP, Data Mining
    and Real Application Testing options
    Master table "AMBER"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
    Starting "AMBER"."SYS_IMPORT_FULL_01": amber/******** directory=EXPORT dumpfile=amberquota1.dmp logfile=impquote1.log TABLE_EXISTS_ACTION=APPEND
    Processing object type TABLE_EXPORT/TABLE/TABLE
    ORA-39083: Object type TABLE failed to create with error:
    ORA-01843: not a valid month
    Failing sql is:
    CREATE TABLE "AMBER"."QUOTE" ("QUOTE_ID" VARCHAR2(15) NOT NULL ENABLE, "CONTRACT_ID" VARCHAR2(12), "QUOTE_OWNER" VARCHAR2(100) NOT NULL ENABLE, "TEMP_CONTRACT" VARCHAR2(1) DEFAULT 'N', "QUOTE_STATUS_CODE" VARCHAR2(1), "QUOTE_INITIATE_DATE" DATE NOT NULL ENABLE, "QUOTE_EXPIRY_DATE" DATE NOT NULL ENABLE, "EXTENDED_EXPIRY_DATE" DATE, "OFFER_DAYS" NUMBER(*,0), "OFFER_CLOSED_DATE" DATE, "BRAND_CODE" VARCHAR2(
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type OBJECT_GRANT:"AMBER" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: Dependent object type INDEX:"AMBER"."QT_PK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QPPS_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QTC_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QT_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QL_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QP_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QF_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QTEXP_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QNC_ACTIVE_CHK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."QT_PK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type CONSTRAINT:"AMBER"."CON_AUTH_CON_CHECK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"AMBER"."QT_PK" creation failed
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/REF_CONSTRAINT
    ORA-39112: Dependent object type REF_CONSTRAINT:"AMBER"."QUOTE_QUOTESTATUS_FK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"AMBER"."QUOTE_BRAND_FK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    ORA-39112: Dependent object type REF_CONSTRAINT:"AMBER"."QCST_QUOTE_FK" skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    ORA-39112: Dependent object type TABLE_STATISTICS skipped, base object type TABLE:"AMBER"."QUOTE" creation failed
    Job "AMBER"."SYS_IMPORT_FULL_01" completed with 31 error(s) at 14:53:00
    wlgora(oracle)[HDM_D]/db/HDM_D/dba/exp$: sqlp
    SQL*Plus: Release 10.2.0.4.0 - Production on Tue Oct 11 14:53:09 2011
    Copyright (c) 1982, 2007, Oracle. All Rights Reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, Real Application Clusters, OLAP, Data Mining
    and Real Application Testing options
    Edited by: 790072 on 10/10/2011 18:54
    Edited by: 790072 on 10/10/2011 18:55

Maybe you are looking for

  • Access address book on iphone without syncing?

    The screen has stopped working on my girlfriends iphone and she needs to access her contacts before it's replaced tomorrow, is there a way of accessing the address book without having to sync it to my computer as her mac is at home? I dont want to ha

  • Transfer records to UNIX with different record length

    Hi all, I have the following problem. I have to transfer records to a UNIX file. The maximum length of a record is 2000. The records are TAB seperated. I also have records which have a length less then 2000. When i look at the file in Excel the last

  • How do I prevent the first line of text in a link from becoming indented?

    Hello.  In a div are two sentences.  They look fine until I change the second sentence into a link.  I tried adding a <span> tag, but doing so changed the class.  I tried placing the text of the second sentence into a separate div tag, but that didn'

  • Blue bounding box - Can't Remove

    www.cfhda.com/menu.html In Firefox and Navigfator, but "not Safari or Explore," there is a blue bounding box around a link (pricing) at the bottom. It shows up in the DW doc, but I can't get rid of it. The link is a png in a table cell but there are

  • Uploading two internal tables

    how to upload two internal tables at a time using standard class methods in to   internal tables. help required