ECC to HANA DB Migration, DMO error with invalid tables

Hello Migration Experts,
I am facing one issue during the uptime migration phase in DMO. Some of the Export and Import jobs are failing, and i see they are failing due to invalid tables. I have checked in HANA,those table are not imported in to HANA schema.
Should i be adding those tables as described in the BW first guidance document ?
Since the tables doesn't exist in HANA schema , what flag i should be using while adding the tables ?
Content from the first guidance doc
If you encounter critical or invalid tables you can do the following workaround:
Edit the file \\server\sapmnt\CIH\SUM\abap\bin\EUCLONEDEFS_ADD.LST (create if not available)
and add the affected tables here, depending the option you want to take, e.g.
/BIC/AZSPOB10300   noclone (Table seems not to exist)
TBTCS  igncount (Table count will be ignored)
TST01  igncount
DBABARL  nocontent (Table doesn’t exist on HANA)
REPOSRC  ignlargercount (Table might change during cloning)
I have also found OSS Notes 1747673 - R3load on SAP HANA: SQL error: SQL syntax error: near ";" which advises to repalce DDLHDB.TPL, but my kernel is higher than the note says, Can i still try ?
Please advise
Thanks in advance,
GR
Below are some of my error logs
Import Job One
R3load: START OF LOG: 20140906205313
R3load: sccsid @(#) $Id: //bas/741_REL/src/R3ld/R3load/R3ldmain.c#11 $ SAP
R3load: version R7.41/V1.9 [UNICODE]
Compiled Aug 16 2014 00:13:51
R3load -pipe -decluster -i /usr/sap/ECS/SUM/abap/migrate_ut/MIGRATE_UT_00032_IMP.CMD -datacodepage 4103 -dbcodepage 4103 -l /usr/sap/ECS/SUM/abap/migrate_ut/MIGRATE_UT_
00032_IMP.LOG -loadprocedure fast -table_suffix ~ -k 1gegdUM50D801eqteAAv1A94
-------------------- Start of patch information ------------------------
patchinfo (patches.h): (0.048) R3load stops in the dependency loop while creating tasks (note 2044380)
DBSL patchinfo (patches.h): (0.031) New DBCON syntax for HANA (note 1983389)
--------------------- End of patch information -------------------------
process id 17232
(DB) INFO: connected to DB
(DB) INFO: NewDB Kernel version 1.00.80.00.391861
                         SQLDBC 1.00.82.00.0394270
(GSI) INFO: dbname   = "DEV/00                                                                                                                          "
(GSI) INFO: vname    = "HDB                             "
(GSI) INFO: hostname = "saphana                                                         "
(GSI) INFO: sysname  = "Linux"
(GSI) INFO: nodename = "SAPECCBOX"
(GSI) INFO: release  = "3.0.13-0.27-default"
(GSI) INFO: version  = "#1 SMP Wed Feb 15 13:33:49 UTC 2012 (d73692b)"
(GSI) INFO: machine  = "x86_64"
(RTF) ########## WARNING ###########
        Without ORDER BY PRIMARY KEY the exported data may be unusable for some databases
(RDI) INFO: /usr/sap/ECS/SUM/abap/migrate_ut/MIGRATE_UT_00032_IMP.STR has format version 2
(RDI) INFO: /usr/sap/ECS/SUM/abap/migrate_ut/SAPDOKCLU.STR.logical has format version 2
(DCL) INFO: found logical cluster description for DOKCLU in /usr/sap/ECS/SUM/abap/migrate_ut/SAPDOKCLU.STR.logical
(DB) INFO: loading data in table "DOKCLU" with mass loader for LOBs #20140906205313
(DB) INFO: DOKCLU deleted/truncated
ERROR exec_ddl_stmt: (DB) ERROR: DDL statement failed
(DELETE FROM "DOKTL"WHERE   ('HY' < "ID" OR ("ID" = 'HY' AND 'SIMGISPAM_BELEGARTEN' < "OBJECT")  OR ("ID" = 'HY' AND "OBJECT" = 'SIMGISPAM_BELEGARTEN' AND 'D' <  "LANG
U") OR ("ID" = 'HY' AND "OBJECT" = 'SIMGISPAM_BELEGARTEN' AND  "LANGU" = 'D' AND 'E' < "TYP") OR ("ID" = 'HY' AND "OBJECT" =  'SIMGISPAM_BELEGARTEN' AND "LANGU" = 'D' A
ND "TYP" = 'E' AND '0013' <  "DOKVERSION")) AND ("ID" < 'TX' OR ("ID" = 'TX' AND "OBJECT" <  'SMP2ALR0000134') OR ("ID" = 'TX' AND "OBJECT" = 'SMP2ALR0000134' AND  "LAN
GU" < 'E') OR ("ID" = 'TX' AND "OBJECT" = 'SMP2ALR0000134' AND  "LANGU" = 'E' AND "TYP" < 'E') OR ("ID" = 'TX' AND "OBJECT" =  'SMP2ALR0000134' AND "LANGU" = 'E' AND "T
YP" = 'E' AND "DOKVERSION"  <= '0001')))
DbSlExecute: rc = 103
Another import Job
(RTF) ########## WARNING ###########
        Without ORDER BY PRIMARY KEY the exported data may be unusable for some databases
(RDI) INFO: /usr/sap/ECS/SUM/abap/migrate_ut/MIGRATE_UT_00063_IMP.STR has format version 2
(RDI) INFO: /usr/sap/ECS/SUM/abap/migrate_ut/SAPTERCL3.STR.logical has format version 2
(DCL) INFO: found logical cluster description for TERCL3 in /usr/sap/ECS/SUM/abap/migrate_ut/SAPTERCL3.STR.logical
(DDL) ERROR: no DDL for TERCL3
(build_ddl_stmt).
(IMP) INFO: a failed DROP attempt is not necessarily a problem
(DB) INFO: TERCL3 merged #20140906205315
(DDL) ERROR: no DDL for TERMC3
(build_ddl_stmt).
(IMP) INFO: a failed DROP attempt is not necessarily a problem
DbSl Trace:   prepare() of C_0002, rc=1, rcSQL=259
DbSl Trace: PREPARE C_0002 on connection 0, rc=259
ERROR HDBexistsRow: Prepare/Read failed (dbrc=103).
  (SQL error 259)
  error message returned by DbSl:
invalid table name:  Could not find table/view TERMC3 in schema SAPECS: line 1 col 58 (at pos 57)
(RDI) INFO: /usr/sap/ECS/SUM/abap/migrate_ut/SAPTERCL3.STR.logical has format version 2
(DCL) INFO: found logical cluster description for TERCL3 in /usr/sap/ECS/SUM/abap/migrate_ut/SAPTERCL3.STR.logical
(SQL) INFO: Searching for SQL file SQLFiles.LST
(SQL) INFO: found SQLFiles.LST
(SQL) INFO: Trying to open SQLFiles.LST
(SQL) INFO: SQLFiles.LST opened
(SQL) INFO: Searching for SQL file SSEXC.SQL
(SQL) INFO: SSEXC.SQL not found
(DB) INFO: TERCL3^0 dropped
(DB) INFO: TERCL3^0 created #20140906205315
(SQL) INFO: Searching for SQL file SDOCU.SQL
(SQL) INFO: SDOCU.SQL not found
ERROR exec_ddl_stmt: (DB) ERROR: DDL statement failed
(ALTER TABLE "TERMC3" DROP CONSTRAINT "TERMC3^0")
DbSlExecute: rc = 103
  (SQL error 259)
  error message returned by DbSl:
invalid table name: TERMC3: line 1 col 13 (at pos 12)
(IMP) INFO: a failed DROP attempt is not necessarily a problem
ERROR exec_ddl_stmt: (DB) ERROR: DDL statement failed
(ALTER TABLE "TERMC3" ADD CONSTRAINT "TERMC3^0" PRIMARY KEY ( "SPRAS", "TERM", "LINENUMBER" ) )
DbSlExecute: rc = 103
  (SQL error 259)
  error message returned by DbSl:
invalid table name: TERMC3: line 1 col 13 (at pos 12)
(RDI) INFO: /usr/sap/ECS/SUM/abap/migrate_ut/SAPTERCL3.STR.logical has format version 2
(DCL) INFO: found logical cluster description for TERCL3 in /usr/sap/ECS/SUM/abap/migrate_ut/SAPTERCL3.STR.logical
(DDL) ERROR: no DDL for TERCL3
(build_ddl_stmt).
(IMP) INFO: a failed DROP attempt is not necessarily a problem
(DB) INFO: TERCL3 unloaded #20140906205315
(DDL) ERROR: no DDL for TERMC3
(build_ddl_stmt).
(IMP) INFO: a failed DROP attempt is not necessarily a problem
ERROR exec_ddl_stmt: (DB) ERROR: DDL statement failed
(                                           ALTER TABLE "TERMC3" enable persistent merge )
DbSlExecute: rc = 103
  (SQL error 259)
  error message returned by DbSl:
invalid table name: TERMC3: line 1 col 56 (at pos 55)
(DB) INFO: disconnected from DB

Result of that table issue fix is that migration went ahead, finished both uptime and downtime run. Now i am trying to login to SAP and system is unable to log me in as DOKTL doesnt exist at all
But i see another table called DTELDOKTL in HANA, which same as DOKTL but no entries in it.
I have created a message to SAP , But what should i be doing now ?
Can i create this table in HANA ?
Can i do export/import of table ? Table structure should be there in system already for this option i guess.
Would i be able to replicate this table if i setup SLT ?
Please advise
C Mon Sep  8 11:02:30 2014
C  *** ERROR =>   prepare() of C_0341, rc=1, rcSQL=259
[dbsdbsql.cpp 1397]
C  {root-id=00505680721D1EE48DE09241702FF193}_{conn-id=00000000000000000000000000000000}_0
C  *** ERROR => PREPARE C_0341 on connection 0, rc=259
[dbslsdb.cpp  9138]
C  SQLCODE    : 259
C  SQLERRTEXT : invalid table name:  Could not find table/view DOKTL in schema SAPECS: line 1 col 97 (at pos 96)
C  sc_p=7f28f48cfac8,no=341,idc_p=7f28f3f87388,con=0,act=0,slen=254,smax=256,#vars=5,stmt=7a45980,table=DOKTL
C  stmtid = <0/SAPMSYST                                /50334994/20140907170329>
C  SELECT "ID" , "OBJECT" , "LANGU" , "TYP" , "DOKVERSION" , "LINE" , "DOKFORMAT" , "DOKTEXT" FROM "DOK\
C  TL" WHERE "LANGU" = ? AND "ID" = ? AND "OBJECT" = ? AND "TYP" = ? AND "DOKVERSION" = ? ORDER BY "ID"\
C   , "OBJECT" , "LANGU" , "TYP" , "DOKVERSION" , "LINE" ;
B  ***LOG BZA=> table DOKTL does not exist on database R/3 [dbdbslst     3580]
M  ThPerformDeleteThisSession: switch of statistics
D  GuiStatus clear generate inline ts >20140908030105,195<
M  ***LOG R47=> ThResFreeSession, delete () [thxxmode.c   1078]
M  *** WARNING => ThrtGuiDeleteScreen: rscpGetUserLoginLang failed (128), use system language [thrtGuiHandl 469]
Thanks,
GR

Similar Messages

  • Full Export/Import Errors with Queue tables and ApEx

    I'm trying to take a full export of an existing database in order to build an identical copy in another database instance, but the import is failing each time and causing problems with queue tables and Apex tables.
    I have used both the export utility and Data Pump (both with partial and full exports) and the same problems are occurring.
    After import, queue tables in my schema are unstable. They cannot be dropped using the queue admin packages as they throw ORA-24002: QUEUE_TABLE <table> does not exist exceptions. Trying to drop the tables causes the ORA-24005: must use DBMS_AQADM.DROP_QUEUE_TABLE to drop queue tables error
    As a result, the schema cannot be dropped at all unless manual data dictionary clean up steps (as per metalink) are done.
    The Apex import fails when creating foreign keys to WWV_FLOW_FILE_OBJECTS$PART. It creates the table ok, but for some reason the characters after the $ are missing so the referencing tables try to refer to WWV_FLOW_FILE_OBJECTS$ only.
    I am exporting from Enterprise Edition 10.2.0.1 and importing into Standard edition 10.2.0.1, but we are not using any of the features not available in standard, and I doubt this would cause the issues I'm getting.
    Can anyone offer any advice on how I can resolve these problems so a full import will work reliably?

    Thanks for the lead!
    After digging around MetaLink some more, it sounds like I'm running into Bug 5875568 (MetaLink Note:5875568.8) which is in fact related to the multibyte character set. The bug is fixed in the server patch set 10.2.0.4 or release 11.1.0.6.

  • ORA-06502 error with external table having long records

    I'm getting a strange error with the oci driver that I don't get with the thin driver.
    The basic situation is that we using external tables and both the oci and thin drivers have been working until we tested with a table that had longer records than the previous tables. The new table has 1800 byte records.
    The thin driver still works fine with the new table. However, the oci driver generates the following error with the new table:
    ORA-06502: PL/SQL: numeric or value error: character string buffer too small
    I suspect that with oci driver the oci DLLs are reading the external file and the DLLs can't handle the longer record length.
    Particulars
    - Oracle DB Server is 10.2.0.1 on SunOS 5.9
    - OCI Instant Client instantclient_10_2 on Windows XP
    - OCI client code from DB server installation running on DB server
    - Works with thin driver from Windows XP and on DB server machine for all record lengths
    - Works with both OCI drivers for records < 1800 bytes (don't know actual limit)
    - Fails with both OCI drivers for records = 1800 bytes.
    Does anyone out there have any thoughts.
    Thanks in advance.

    Your access parameters are in the wrong order. External tables are a bit fussy like that. Refer to the access_parameters section in the Utilities manual and follow the order there. From memory it will go something like this:
    RECORDS DELIMITED...
    LOG/BAD/DISCARDFILE...
    FIELDS TERMINATED...LDRTRIM
    MISSING FIELD VALUES...
       fields...
    )Regards...

  • Error with the table maintenance screens

    Hi Experts,
    I have created a maintenance view for two tables. The first table has only one field, which is the primary key. The second table is the text table for the first table, where the description of the primary key is maintained. Initially after creating the table maintenance for these two tables, all the fields were comming correctly. The first table showed only one field in the SM30 and the second table showed all fields in SM30(primary key, its description and the language key). The table maintenance was later deleted and regenerated, due to some changes in the table. After regenerating the TMG it is seen that the first table shows all the fields in SM30 same as the second table. I feel the function group program is not getting refreshed. Is there any method to regenerate the function group.
    Thanks in advance.
    Regards,
    Anu.

    Hi Dhina,
    I deleted the TMG for both the primary table and the text table, but when I check in SE16 I still find all the fields in the primary table just like the text table.
    My first table has
    MANDT
    EDUCATION_TYPE
    My second table has
    MANDT
    LANGU
    EDUCATION_TYPE
    DESCRIPTION
    I have create a maintanence view joining the above two tables(with the following fileds MANDT, EDUCATION_TYPE, DESCRIPTION), and deleted the table maintenance of the above tables. But still when I check the first table in SE16 I can see fields MANDT, EDUCATION_TYPE, DESCRIPTION.
    What is the problem with the table maintenance?
    Any idea?
    Thanks
    Anu.

  • Error with dynamic table

    Hi, I'm creating some Attributes on my context:
    IWDAttributeInfo a=wdContext.getNodeInfo().getChild("DataTable").addAttribute("XAxis","ddic:com.sap.dictionary.string");
    for(int i = 0; i < (wdContext.nodeEt_Coordinates().size()-i1); i+)
    {IWDAttributeInfo b=wdContext.getNodeInfo().getChild("DataTable").addAttribute("val"+Integer.toString(i),"java.lang.Boolean");}
    Next, I "filling" them with some values:     
    for(int u = 0; u < wdContext.nodeEt_Coordinates().size(); u++)
    {IWDNodeElement ne =  wdContext.getChildNode("DataTable", 0).createElement();
    ne.setAttributeValue("XAxis", wdContext.nodeEt_Coordinates().getEt_CoordinatesElementAt(u).getCoord_Short());
    for(int i = 0; i < wdContext.nodeEt_Coordinates().size(); i++)
    {ne.setAttributeValue("val"+Integer.toString(i), Boolean.FALSE);}     
    wdContext.getChildNode("DataTable", 0).addElement(ne);
    Out of this Node DataTable I´creating a table.
    I get an error message, when im checking two chekboxes from two different lines in the resulting table. The message is the following:
    " com.sap.tc.webdynpro.progmodel.context.ContextException: DataAttributeInfo(MainView.DataTable.val2): does not have a simple type "
    Can anyone help?
    Message was edited by: Alexander Reinke

    Hi,
       Replace this line
    {IWDAttributeInfo b=wdContext.getNodeInfo().getChild("DataTable").addAttribute("val"+Integer.toString(i),"java.lang.Boolean");}
    with
    {IWDAttributeInfo b=wdContext.getNodeInfo().getChild("DataTable").addAttribute("val"+Integer.toString(i),"ddic:com.sap.dictionary.boolean");}
    or
    {IWDAttributeInfo b=wdContext.getNodeInfo().getChild("DataTable").addAttribute("val"+Integer.toString(i),"java:java.lang.Boolean");}
    Regards,
    Satyajit.
    Message was edited by: Satyajit Chakraborty

  • Data pump import error with nested tables

    So the problem is somewhat long :)
    Actually the problems are two - why and how oracle are treating OO concept and why data pump doesn't work?
    So scenario for the 1st one:
    1) there is object type h1 and table of h1
    2) there is unrelated object type row_text and table of row_text
    3) there is object type h2 under h1 with attribute as table of row_text
    4) there is table tab1 with column b with data type as table of h1. Of course column b is stored as nested table.
    So how do you think - how many nested tables Oracle will create? The correct answer is 2. One explicitly defined and one hidden with system
    generated name for the type h2 which is under type h1. So the question is WHY? Why if I create an instance of supertype Oracle tries to adapt
    it for the subtype as well? Even more if I create another subtype h3 under h1 another hidden nested table appears.
    This was the first part.
    The second part is - if I do schema export and try to import it in another schema I got error saying that oracle failed to create storage table for
    nested table column b. So the second question is - if Oracle has created such a mess with hidden nested tables how to import/export to another
    schema?
    Ok and here is test case to demonstrate problems above:
    -- creating type h1 and table of it
    SQL> create or replace type h1 as object (a number)
      2  not final;
      3  /
    Type created.
    SQL> create or replace type tbl_h1 as table of h1;
      2  /
    Type created.
    -- creating type row_text and table of it
    SQL> create or replace type row_text as object (
      2    txt varchar2(100))
      3  not final;
      4  /
    Type created.
    SQL> create or replace type tbl_row_text as table of row_text;
      2  /
    Type created.
    -- creating type h2 as subtype of h1
    SQL> create or replace type h2 under h1 (some_texts tbl_row_text);
      2  /
    Type created.
    SQL> create table tab1 (a number, b tbl_h1)
      2  nested table b
      3  store as tab1_nested;
    Table created.
    -- so we have 2 nested tables now
    SQL> select table_name, parent_table_name, parent_table_column
      2  from user_nested_tables;
    TABLE_NAME                     PARENT_TABLE_NAME
    PARENT_TABLE_COLUMN
    SYSNTfsl/+pzu3+jgQAB/AQB27g==  TAB1_NESTED
    TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H2")."SOME_TEXTS"
    TAB1_NESTED                    TAB1
    B
    -- another subtype of t1
    SQL> create or replace type h3 under h1 (some_texts tbl_row_text);
      2  /
    Type created.
    -- plus another nested table
    SQL> select table_name, parent_table_name, parent_table_column
      2  from user_nested_tables;
    TABLE_NAME                     PARENT_TABLE_NAME
    PARENT_TABLE_COLUMN
    SYSNTfsl/+pzu3+jgQAB/AQB27g==  TAB1_NESTED
    TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H2")."SOME_TEXTS"
    SYSNTfsl/+pz03+jgQAB/AQB27g==  TAB1_NESTED
    TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H3")."SOME_TEXTS"
    TAB1_NESTED                    TAB1
    B
    SQL> desc "SYSNTfsl/+pzu3+jgQAB/AQB27g=="
    Name                                      Null?    Type
    TXT                                                VARCHAR2(100)OK let it be and now I'm trying to export and import in another schema:
    [oracle@xxx]$ expdp gints/xxx@xxx directory=xxx dumpfile=gints.dmp logfile=gints.log
    Export: Release 11.2.0.1.0 - Production on Thu Feb 4 22:32:48 2010
    <irrelevant rows skipped>
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "GINTS"."TAB1"                                  0 KB       0 rows
    . . exported "GINTS"."SYSNTfsl/+pz03+jgQAB/AQB27g=="         0 KB       0 rows
    . . exported "GINTS"."TAB1_NESTED"                           0 KB       0 rows
    . . exported "GINTS"."SYSNTfsl/+pzu3+jgQAB/AQB27g=="         0 KB       0 rows
    Master table "GINTS"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    ******************************************************************************And now import. In order to create types transformation of OIDs is applied and also remap_schema
    Although it fails to create the table.
    [oracle@xxx]$ impdp gints1/xxx@xxx directory=xxx dumpfile=gints.dmp logfile=gints_imp.log remap_schema=gints:gints1 transform=OID:n
    Import: Release 11.2.0.1.0 - Production on Thu Feb 4 22:41:48 2010
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Release 11.2.0.1.0 - Production
    Master table "GINTS1"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
    Starting "GINTS1"."SYS_IMPORT_FULL_01":  gints1/********@xxx directory=xxx dumpfile=gints.dmp logfile=gints_imp.log remap_schema=gints:gints1 transform=OID:n
    Processing object type SCHEMA_EXPORT/USER
    ORA-31684: Object type USER:"GINTS1" already exists
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    Processing object type SCHEMA_EXPORT/TYPE/TYPE_SPEC
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    ORA-39083: Object type TABLE:"GINTS1"."TAB1" failed to create with error:
    ORA-02320: failure in creating storage table for nested table column B
    ORA-00904: : invalid identifier
    Failing sql is:
    CREATE TABLE "GINTS1"."TAB1" ("A" NUMBER, "B" "GINTS1"."TBL_H1" ) SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39083: Object type INDEX_STATISTICS failed to create with error:
    ORA-01403: no data found
    ORA-01403: no data found
    Failing sql is:
    DECLARE I_N VARCHAR2(60);   I_O VARCHAR2(60);   c DBMS_METADATA.T_VAR_COLL;   df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS'; BEGIN  DELETE FROM "SYS"."IMPDP_STATS";   c(1) :=   DBMS_METADATA.GET_STAT_COLNAME('GINTS1','TAB1_NESTED',NULL,'TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H2")."SOME_TEXTS"',1);  DBMS_METADATA.GET_STAT_INDNAME('GINTS1','TAB1_NESTED',c,1,i_o,i_n);   INSERT INTO "
    ORA-39083: Object type INDEX_STATISTICS failed to create with error:
    ORA-01403: no data found
    ORA-01403: no data found
    Failing sql is:
    DECLARE I_N VARCHAR2(60);   I_O VARCHAR2(60);   c DBMS_METADATA.T_VAR_COLL;   df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS'; BEGIN  DELETE FROM "SYS"."IMPDP_STATS";   c(1) :=   DBMS_METADATA.GET_STAT_COLNAME('GINTS1','TAB1_NESTED',NULL,'TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H3")."SOME_TEXTS"',1);  DBMS_METADATA.GET_STAT_INDNAME('GINTS1','TAB1_NESTED',c,1,i_o,i_n);   INSERT INTO "
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "GINTS1"."SYS_IMPORT_FULL_01" completed with 4 error(s) at 22:41:52So any idea how to make export/import of such tables?
    TIA
    Gints

    Tom Kyte has said it repeatedly ... I will repeat it here for you.
    The fact that Oracle allows you to build object tables is not an indication that you should.
    Store your data relationally and build object_views on top of them.
    http://www.morganslibrary.org/reference/object_views.html
    If you model properly, and store properly, you don' have any issues.

  • Urgent Help Please, Error with Nested Table

    Hello All,
    I'm trying to create data mining models based on some medical data. I have combined two tables (Patients) and (tumors) into one table (cancer_patients). In the (cancer_patients) table there is one column (tumors) as nested table column. So I got the following error (ORA-22913) about the nested table. I need an advice about how to fix this error in ODM please
    This is the code of the tables
    CREATE TABLE "DSS2_MINING"."PATIENTS"
       (     "PATIENT_ID" VARCHAR2(10 BYTE) NOT NULL ENABLE,
         "REGISTRY_ID" NUMBER(10,0),
         "RACE" VARCHAR2(2 BYTE),
         "***" VARCHAR2(1 BYTE),
         "BIRTHDATE_YEAR" NUMBER(4,0),
         "NUMBER_OF_PRIMARIES" NUMBER(1,0),
         "VITAL_STATUS_RECORD" VARCHAR2(1 BYTE),
         "CAUSE_OF_DEATH" VARCHAR2(5 BYTE),
         "SURVIVAL_TIME" VARCHAR2(4 BYTE),
         "SURVIVAL_TIME_FINAL" NUMBER,
         "SURVIVAL_VARIABLE" VARCHAR2(1 BYTE),
          CONSTRAINT "PATIENTS_PK" PRIMARY KEY ("PATIENT_ID");
    CREATE TABLE "DSS2_MINING"."TUMORS"
       (     "TUMOR_ID" NUMBER NOT NULL ENABLE,
         "PATIENT_ID" VARCHAR2(10 BYTE),   -- FK
         "SEER_RECORD_NUMBER" NUMBER,       -- This column contain a sequance number of the records for each patients
         "MARITAL_STATUS" VARCHAR2(1 BYTE),
         "AGE" NUMBER,
         "DATE_OF_DIAGNOSIS" DATE,
         "HISTOLOGY_GROUP" VARCHAR2(2 BYTE),
         "BEHAVIOR" VARCHAR2(1 BYTE),
         "GRADE" VARCHAR2(1 BYTE),
         "DERIVED_AJCC_STAGE_GROUP" VARCHAR2(2 BYTE),
         "STAGE_OF_CANCER" VARCHAR2(2 BYTE),
         "RADIATION" VARCHAR2(1 BYTE),
         "CS_SCHEMA" VARCHAR2(2 BYTE),
         "FIRST_PRIMARY_IND" VARCHAR2(1 BYTE),
         "TUMOR_SIZE" NUMBER(4,1),
         "TUMOR_EXTENSION" VARCHAR2(2 BYTE),
         "LYMPH_NODES" VARCHAR2(1 BYTE),
         "NODES_POSITIVE" NUMBER,
         "ESTROGEN" VARCHAR2(3 BYTE),
         "PROGESTERONE" VARCHAR2(3 BYTE),
         "SURGERY" VARCHAR2(2 BYTE),
          CONSTRAINT "TUMORS_PK" PRIMARY KEY ("TUMOR_ID");
    create or replace type tumor_object AS
    object(
    tumor_id VARCHAR2(1),  
    marital_status VARCHAR2(1),  
    age NUMBER(3),  
    date_of_diagnosis DATE, 
    cs_schema VARCHAR2(2),  
    histology_group VARCHAR2(2),  
    behavior VARCHAR2(1),  
    grade VARCHAR2(1),  
    first_primary_ind VARCHAR2(1),  
    tumor_size NUMBER(4,   1),  
    tumor_extension VARCHAR2(2),  
    lymph_nodes VARCHAR2(1),  
    nodes_positive NUMBER(4),  
    surgery VARCHAR2(2),
    radiation VARCHAR2(1)
    create or replace type tumor_table as table of tumor_object;
      CREATE TABLE "DSS2_MINING"."CANCER_PATIENTS"
       (     "PATIENT_ID" VARCHAR2(10 BYTE) NOT NULL ENABLE,
         "RACE" VARCHAR2(2 BYTE),
         "***" VARCHAR2(1 BYTE),
         "NUMBER_OF_PRIMARIES" NUMBER(1,0),
         "TUMORS" "DSS2_MINING"."TUMOR_TABLE" ,
         "VITAL_STATUS_RECORD" VARCHAR2(1 BYTE),
         "CAUSE_OF_DEATH" VARCHAR2(5 BYTE),
         "SURVIVAL_TIME_FINAL" NUMBER,
         "SURVIVAL_VARIABLE" VARCHAR2(1 BYTE),
          CONSTRAINT "CANCER_PATIENTS_PK" PRIMARY KEY ("PATIENT_ID")
       NESTED TABLE "TUMORS" STORE AS "TUMORS_STOR_TABLE"Then, I have transferred the data using
    INSERT
      INTO CANCER_PATIENTS
      SELECT  PATIENT_ID,
              RACE,
              NUMBER_OF_PRIMARIES,
               SELECT  CAST(
                            COLLECT(
                                    TUMOR_OBJECT(
                                                 TUMOR_ID,
                                                 MARITAL_STATUS,
                                                 AGE,
                                                 DATE_OF_DIAGNOSIS,
                                                 CS_SCHEMA,
                                                 HISTOLOGY_GROUP,
                                                 BEHAVIOR,
                                                 GRADE,
                                                 FIRST_PRIMARY_IND,
                                                 TUMOR_SIZE,
                                                 TUMOR_EXTENSION,
                                                 LYMPH_NODES,
                                                 NODES_POSITIVE,
                                                 SURGERY ,
                                                 RADIATION
                            AS TUMOR_TABLE
                 FROM  "TUMORS" T
                 WHERE T.PATIENT_ID = P.PATIENT_ID
              VITAL_STATUS_RECORD,
              CAUSE_OF_DEATH,
              SURVIVAL_TIME_FINAL,
              SURVIVAL_VARIABLE
        FROM  PATIENTS PThanks
    A.L
    Edited by: user9003901 on Nov 30, 2010 12:23 AM

    Hi
    i was using AWM since 9i just because it creates OLAP metadata with a new API called CWM2, and Oracle Enterprise Manager was using a propietary API .
    With AWM 10.1.0.4 building your OLAP metadata is very simple, you can define your Dimensions, cubes, measures and even you can load your data from your source system.
    You can refer to the Oracle® OLAP Application Developer's Guide where you can find more information on this tool.
    And relative to your error i will recommend that you build on a dimension-by-dimension basis and try to validate your OLAP Metadata.
    Regards

  • AMS Setup Manager Diagnostics Test erroring with "invalid snapshot" error

    While running the Diagnostic Test for Setup Manager on the AMS, it's failing with errors in the below stages.
    * Testgroup: CheckEBSHome
    * Testgroup: CheckAgentStage
    * Testgroup: CheckAgentHome
    with below error
    Snapshot information is not populated
    Any idea?
    Regards
    Sree

    OID Setup Diagnostics Test failing due to this 11510 Home Page is redirecting to SSO page, if give login details, i am not able to login to 11510 instance.i created OID user same as 11510 user. i am getting internal server error in the browser.
    Thanks,
    Panneer.

  • GETTING AN ERROR WITH STANDARD TABLE

    I am working for the module CS (Customer Service)
    In <b>Development</b> for standard database table ILOA there is a field named TPLNR (Functional Location) , for this filed if i  provide a value <b>W*</b> it will provide the output of Functional Locations starting with letter <b>W</b>
    and found its working fine.
    But when it is transported to Quality we found an error for <b>W</b>  even if we provide a <b></b> it is showing an Error message of <b>Functional location * does not exist</b> but client need this requirement....also i have checked the <b>dataelement and domain</b> part every thing is same.....
    please let me know where it is going wrong

    Hi
    Have u compared with the data which is already inside the quality system whether it is having data in Quality or not???
    Regards
    Pavan

  • Error with invalid chars found in characteristics

    Hi,
       I am loading information for this field EKPO-TXZ01 (PO text) into ODS and then into infocube. I have no issue loading this field content into ODS. However from ODS to infocube, I encountered error that says invalid chars found. (This field was defined as chars 40 in the infoobject). At BW customizing, the chars like #, " [] are being maintained. Please advise how to overcome this.

    Hi,
    From my experience I can say that:
    1. Not always assigning permitted characters helps. Especially if chars that you want to pass to the ODS are special chars. And there is too little space in the string with allowable chars.
    2. Setting a characteristics as a lower case char helps for its master data. An ODS will not allow to accept data with lower case letters anyway.
    3. There are several topics related to your question in this forum with solutions, even with ABAP codes.
    Look, for example at this one:
    https://www.sdn.sap.com/sdn/collaboration.sdn?contenttype=url&content=https%3A//forums.sdn.sap.com/profile.jspa%3FuserID%3D136828
    Best regards,
    Eugene

  • Error with Pivot Table in 10.3.6 weblogic server

    Hi Everyone,
    I created an ADF application and in that i have a Pivot Table i developed using Jdeveloper 11.1.1.3 it was working fine when we deployed the application in Weblogic server 10.3.3.
    Now DBA's upgraded the Weblogic server to 10.3.6 so i downloaded Jdeveloper 11.1.1.6 and generated the EAR file and deployed in the new 10.3.6 server. The pivot table is not working i am getting the following error message
    Root cause of ServletException.
    java.lang.NoSuchMethodError: oracle.adfinternal.view.faces.bi.renderkit.imageView.ImageViewRendererUtils.isAttachmentMode()Z
    at oracle.adfinternal.view.faces.bi.renderkit.pivotTable.RichPivotTableRenderer.getRowBlockSize(RichPivotTableRenderer.java:1489)
    at oracle.adfinternal.view.faces.bi.renderkit.pivotTable.RichPivotTableRenderer.getPivotTableEncodeContext(RichPivotTableRenderer.java:1616)
    at oracle.adfinternal.view.faces.bi.renderkit.pivotTable.RichPivotTableRenderer.encodeAll(RichPivotTableRenderer.java:1789)
    Here is the code
    <dvt:pivotTable id="employeeDocuments_pt"
                            value="#{bindings.lum_EmployeeDocumentsReportView1.pivotTableModel}"
                            binding="#{backingBeanScope.backing_LUM_EMPLOYEE_DOCUMENTS_REPORT.employeeDocuments_pt}"
                            var="cellData" varStatus="cellStatus" sizing="auto" summary="Enrollments"
                            inlineStyle="width:1400px; height:900.0px;"
                            rowFetchSize="-1" columnFetchSize="75">
                <dvt:dataCell dataAttribute="DocumentFlag">
                <af:selectBooleanCheckbox id="idselectbooleancheckbox" label="SELECT" value="#{cellData.dataValue}"
                  readOnly="true"/>
                </dvt:dataCell>              
                </dvt:pivotTable>
    The code remains the same nothing changed from 10.3.3 to 10.3.6 and i even tried deploying the ear file generated from jdeveloper 11.1.1.3 even that did not work.
    Is this soemthing related to libraries issue?
    Can some one help me on this.
    Thanks,
    Harika.

    Timo,
    This the total code for Pivot table and to export that to Excel i think something in the below code is not supported in 11.1.1.6 and later versions. Can i use <dvt:exportPivotTableData> in commandButton ? because when i comment this code it is working fine.
    <af:commandButton text="EXPORT TO EXCEL" immediate="true"
                                binding="#{backingBeanScope.backing_LUM_EMPLOYEE_DOCUMENTS_REPORT.export_btn}"
                                id="export_btn" visible="false"
                                inlineStyle="font-family: Verdana, Arial, Helvetica, sans-serif; font-weight: bold; font-size: small;">
            <dvt:exportPivotTableData type="excelHTML" exportedId="employeeDocuments_pt"
               filename="Department_TrainingDocuments_Report.xls" title="PivotTable export"/>
                                </af:commandButton>
    <dvt:pivotTable id="employeeDocuments_pt"
                            value="#{bindings.lum_EmployeeDocumentsReportView1.pivotTableModel}"
                            binding="#{backingBeanScope.backing_LUM_EMPLOYEE_DOCUMENTS_REPORT.employeeDocuments_pt}"
                            var="cellData" varStatus="cellStatus" sizing="auto" summary="Enrollments"
                            inlineStyle="width:1400px; height:900.0px;"
                            rowFetchSize="-1" columnFetchSize="75">
                <dvt:dataCell dataAttribute="DocumentFlag">
                <af:selectBooleanCheckbox id="idselectbooleancheckbox" label="SELECT" value="#{cellData.dataValue}"
                  readOnly="true"/>
                </dvt:dataCell>              
                </dvt:pivotTable>
    Thanks,
    Harika.

  • Error With HTML Table

    Hi there -- I created an htm document in Dreamweaver 2004,
    and I now want to edit it in Fireworks. However, when I try to
    either open the file or reconstitute the HTML table, I get the
    following error message:
    Could not complete your request. Fireworks could not find one
    or more of the image files referenced by the HTML table. Please
    verify that the images are in the correct place and update the HTML
    file accordingly.
    I did a check of the images that are referenced by the HTML
    file, and they all did seem to appear in their correct locations.
    Is there a tool or specific procedure that I can use that can help
    me determine what file(s) in question are at fault? Thanks.

    Hi there -- I created an htm document in Dreamweaver 2004,
    and I now want to edit it in Fireworks. However, when I try to
    either open the file or reconstitute the HTML table, I get the
    following error message:
    Could not complete your request. Fireworks could not find one
    or more of the image files referenced by the HTML table. Please
    verify that the images are in the correct place and update the HTML
    file accordingly.
    I did a check of the images that are referenced by the HTML
    file, and they all did seem to appear in their correct locations.
    Is there a tool or specific procedure that I can use that can help
    me determine what file(s) in question are at fault? Thanks.

  • Email Migration Tool Error

    Hi,
    I get an error by using the Email Migration Tool Error with OCS Release 2 on a Linux RH AS 2.0 single box server:
    Exception in thread "main" java.lang.NoClassDefFoundError: oracle/mail/migration/migrate/Migrate
    Command:
    $ORACLE_HOME\bin\migrate.sh
    Please help.
    Bob de Leeuw

    I had an 9.2 database already installed on my laptop. I installed the migration tool in c:\oracle\ora92\oes\migration and set ORACLE_HOME=c:\oracle\ora92.

  • Error with forign key constraints

    hi
    I have a .sql file which contains about 180 tables scripts . TO disable the contraints i ran the below script first
    begin
    for v in (select constraint_name, table_name from user_constraints where constraint_type = 'R')
    loop
         execute immediate
         'alter table ' || v.table_name || ' disable constraint ' || v.constraint_name;
    end loop;
    end;
    and then the .sql file by giving @filename.sql .Some of the scripts are getting executed but the one's having forign key references it is showing containt error foreign key check. but when i ran the above script ie the one to disable the foreign key check again and then ran each of these table scripts which showed error before manually it gets executed .why is this so? can anyone help me in solving this problem
    Thanking you in advance
    Dinny

    I have a .sql file which contains table scripts and i some of these tables there is foreign key referencesing other tables so when i am ruuning the script it is showing such tables(ie the tables that i am refering to) is not existing for eg below is one of the script from the .sql file
    create table t_system_groups (
    group_id NUMBER(11) not null,
    group_name VARCHAR2(255) not null,
    group_desc VARCHAR2(255),
    parent_id NUMBER(11) default '0',
    full_path VARCHAR2(255) default NULL,
    CONSTRAINT pk_group_id_t_system_groups PRIMARY KEY (group_id),
    CONSTRAINT uk_group_name_t_system_groups UNIQUE (group_name),
    CONSTRAINT fk_t_system_groups FOREIGN KEY (parent_id) REFERENCES t_system_groups (group_id) ON DELETE CASCADE
    Since i have run the script for disabling the constraints before running the .sql file then why is it showing such as error. also showing errors with drop table having references
    for eg:
    SQL> begin
    2 for v in (select constraint_name, table_name from user_constraints where constraint_type = 'R'
    3 loop
    4 execute immediate
    5 'alter table ' || v.table_name || ' disable constraint ' || v.constraint_name;
    6 end loop;
    7
    8 end;
    9 /
    PL/SQL procedure successfully completed.
    SQL> drop table t_system_groups;
    drop table t_system_groups
    ERROR at line 1:
    ORA-02449: unique/primary keys in table referenced by foreign keys
    Regards
    dinny

  • OracleXMLQuery fails with invalid UTF-8 character

    All -
    While using OracleXMLQuery API to pull the data back as a DOM. The internal parser has a fatal error with invalid UTF-8 data stored in the XML data.
    Is it possible to use the setXSLT() method to transform the data before the DOM is attempted to be built. Or some other method within the OracleXMLQuery API to alter the invalid XML data before an attempt is made to parse it?
    Note: The database will have non-UTF-8 data; Presently beyond our control.

    Hi,
    Are the files from the Java-CD fully readable and fully copied, for instance the java-dumpfile ?
    Ensure JVM settings of note # 716604 area applied. Also, is connect possible to database by 'configtool' or does it throw an error ?
    Br,
    Javier

Maybe you are looking for