No data inserted in error table

hi,
I developed an ODI interface that inserts data from one table into other table , (both tables are in the same schema of oracle database.)
But on sending faulty data ,
I can’t find any records in error table (E$_ , SNP_CHK_TBL)
I tried these knowledge modules Combination
CKM SQL - IKM Oracle Incremental Update
CKM Oracle –IKM SQL Control Append
Data is inserted in Flow table … but PK Errors records are not inserted in error table.
query of PK Errors in Operator:
insert into ODI.E_EMP2
     ROW_ID,
     ERR_TYPE,
     ERR_MESS,
     ORIGIN,
     CHECK_DATE,
     CONS_NAME,
     CONS_TYPE,
     ENO,
     ENAME
select     distinct
     TRG.rowid,
     'F',
     'The primary key EMP2_PK is not unique.',
     '(7010)ERROR_CHK.EMP1_TO_EMP2',
     sysdate,
     'EMP2_PK',
     'PK',     
     TRG.ENO,
     TRG.ENAME
from     ODI.I_EMP2 TRG,
     ODI.I_EMP2 TRG2
where     TRG2.ENO = TRG.ENO and
     TRG2.rowid <> TRG.rowid
Do I need to modify CKM and IKM.... ?
Please help....

Hi, joping to help.
try using IKM SQL CONTROL APPEND.
on Control tab in the second tab you have to see the list of the check. Are they set to "yes"?
if not go to Model and add the costrants
Regards
DecaXD

Similar Messages

  • How is the data inserted into CST_INV_QTY_TEMP table?

    Hi All,
    How is the data inserted into CST_INV_QTY_TEMP table ?
    Thanks in advance,
    Mayur
    Edited by: 928178 on 17-Apr-2012 04:29

    How is the data inserted into CST_INV_QTY_TEMP table ?TABLE: BOM.CST_INV_QTY_TEMP
    http://etrm.oracle.com/pls/et1211d9/etrm_pnav.show_object?c_name=CST_INV_QTY_TEMP&c_owner=BOM&c_type=TABLE
    Thanks,
    Hussein

  • Need code for multi data insert..to table

    Hi All,
    I have input parameter data like ..
    [email protected],b.gamil.com,c.gamil.com ...
    and need to insert to table T_TABLE...LIKE
    NAME
    A
    B
    C
    Please let me know how to implement...

    anbarasan wrote:
    Hi All,
    I have input parameter data like ..
    [email protected],b.gamil.com,c.gamil.com ...
    and need to insert to table T_TABLE...LIKE
    NAME
    A
    B
    C
    Please let me know how to implement...Like this?
    SQL> ed
    Wrote file afiedt.buf
      1  with t as
      2   (select '[email protected],b.gamil.com,c.gamil.com ' txt from dual)
      3  select upper(regexp_replace(REGEXP_SUBSTR(txt, '[^,]+', 1, level), '[@.](.*)$')) replace_str
      4    from t
      5* connect by level <= length(regexp_replace(txt, '[^,]*')) + 1
    SQL> /
    REPLACE_STR
    A
    B
    CLogic for converting comma separated string to row borrowed from this post of Loknath Giri :
    {message:id=10028213}
    Edited by: Vivek L on 9 Mar, 2012 12:22 PM

  • Bulk text data insertion in Partitioned Table

    Hi All,
    I want to insert more than 200 crore records from 5 text data files in a single table say TAB1 (PARTITIONED table with RANGE type).
    Currently I am using EXTERNAL TABLE concept and providing 5 filenames (of total 45 crore records) at a time to external table definition access criteria LOCATION() with COMMA ",".
    After that using following sql statement:
    INSERT /*+ APPEND PARALLEL(6) */ INTO TAB1 SELECT * FROM TAB1_EXT;
    Presently The execution time statistics for 45 crore records is as follows:
    1)     With /*+ PARALLEL(6) */ hint, execution time : 81 Mins     
    2)     with /*+ APPEND PARALLEL */ hint, execution time : 73 Mins     
    3)     With cursor loop; bulk collect and FORALL with interval of 1000000 records, execution time : 98 Mins
    Earlier the execution time was too much but after changing and trying with variation in SQL logic, DB parameter, and H/w, now it is somewhat good.
    Now INDEXES are also removed from table. NOLOGGING is purposely not using as it will not be considerable at PROD environment.
    Also Database is not in ARCHIVE mode now.
    I also tried with change in DB parameters like SGA_MAX_SIZE which is now 16GB. I am thinking even to change DB_BLOCK_SIZE value from 8K to 32K.
    O/S: LINUX
    ORACLE DB: ORACLE 11g
    RAM: 256 GB
    48 CPU CORES, AMD Opteron (HP DL 585 Model)
    If you have any suggestion, please pass it to me.

    Instead of disturbing parallel coordinators you could just submit n jobs each doing a simple <b>insert into big_table select * from external_table_n</b> or something similar - all external_tables_i having the same structure just different location parameters.
    Regards
    Etbin
    Edited by: Etbin on 26.2.2011 12:25
    possibly grouping text data files according to partitions the data would end within

  • How to insert row in table control and save the data in tables

    Hi,
    I have one table control i am displaying data into table control ,
    my problem is : i want to display data into read mode in table control.
    but when i click on insert button on the same screen i want one blank line should inserted into table control , then i want to insert some data into table control on that row , when i click the save button . the new data inserted into the table control is insert that data into ztable ,
    please give me solution
    main problen is  how can know inserted line in table control and pass that data into ztable.

    Hi,
    Follow the below logic,
    PROCESS BEFORE OUTPUT.
      MODULE STATUS_0001.
      MODULE POPULATE_TABLE_CONTROL. --> Get the data from table store in 
                                                                          ITAB
      LOOP AT GT_CTRL_LP_D516 INTO GS_WA_CTRL_LP_D516
           WITH CONTROL CTRL_LP_D516
           CURSOR CTRL_LP_D516-CURRENT_LINE.
      The following module moves data to control
        MODULE MOVE_TO_CONTROL.--> Move data from ITAB to table control
      ENDLOOP.
    PROCESS AFTER INPUT.
      LOOP AT GT_CTRL_LP_D516.
      ENDLOOP.
      MODULE EXIT AT EXIT-COMMAND.
      MODULE USER_COMMAND_0001.  --> Here you have to take out the values from table control and update database table
    Reward points if helpful.
    Thanks and regards,
    Mallareddy Rayapureddy,
    Munich, Germany.

  • How do we use Data rules/error table for source validation?

    How do we use Data rules/error table for source validation?
    We are using OWB repository 10.2.0.3.0 and OWB client 10.2.0.3.33. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
    I reviewed the posting
    Re: Using Data Rules
    Thanks for this forum.
    I want to apply data rules to source table/view and rule violated rows should go to defined error table. Here is an example.
    Table ProjectA
    Pro_ID Number(10)
    Project_name Varchar(50)
    Pro_date Date
    As per above posting, I created the table in object editor, created the data rule
    NAME_NOT_NULL (ie project name not null). I specified the shadow table name as ProjectA_ERR
    In mapping editor, I have projectA as source. I did not find error table name and defined data rules in table properties. It is not showing up the ERR group in source table
    How do we bring the defined data rules and error table into mapping?
    Are there any additional steps/process?
    Any idea ?
    Thanks in advance.
    RI

    Hi,
    Thanks for your reply/pointer. I reviewed the blog. It is interesting.
    What is the version of OWB used in this blog?
    After defining data rule/shadow table, I deployed the table via CC. It created a error table and created the all the source coulmns in alphabatical order. If I have the primary key as 1st coulmn (which does not start with 'A') in my source, it will apprear middle of of columns in error table.
    How do we prevent/workaround this?
    If I have source(view) in sch A, how do we create Error table in Sch B for source(view)?
    Is it feasible?
    I brought the error table details in mapping. Configured the data rules/error tables.
    If I picked up 'MOVE TO ERROR' option, I am getting "VLD-2802 Missing delete matching criteria in table. the condition is needed because the operator contain at least one data rule with a MOVE TO ERROR action"
    On condition Loading - I have 'All constraints' for matching criteria.
    I changed to "no constraints' still I get the above error.
    If I change to 'REPORT' option instead of 'MOVE TO ERROR' option, error goes off.
    Any idea?
    Thanks in advance.
    RI

  • How to use error table in mapping level?

    Hi all
    Any one please tell me how to use error table in mapping level or how to handle the errors in mapping.
    I am creating one error table in oracle but i dont know how to use it in mapping.
    Thanks in advance.
    Kumar

    Hi Kumar,
    You need to use Error Table along with the Data Rule .
    Error tables are used with Logical error handlers (Data Rules). Detailed description on using data rules is explained in the following document
    http://www.oracle.com/technology/products/warehouse/pdf/OWB10gR2_ETLandBusinessRules.pdf
    In a nut shell usage of error tables is as follows.
    1. Create a data rule associate it with a table (Target table or table in which you load data).
    2. Use the table in a mapping as target table.
    3. Deploy the table and mapping.
    Now in the target schema one more table will be created with the name <target table name>_ERR. When a mapping is executed the data which violates the data rule gets into this error table.
    The following has to be done to have error table name of customers choice.
    1. Right click on the table and click configure.
    2. Give the table name in "shadow table name" section.
    3. Deploy the table.
    4. Use this table in a mapping. Now in the mapping on this table properties in "Error Table Name" section the name of the table specified in "shadow table name" will be populated.
    Thanks,
    Sutirtha

  • Constraint for insertion in a table if inserted rows should not more than 2

    Problem: A Teacher only can assigned for maximum 2 Subjects in the same class Standard.
    actually i want to create a table like following
    create table teaches
    teacherid int,
    standard varchar(10),
    subjectcode varchar(10),
    constraint cpk_teaches primary key(teacherid, standard, subjectcode)
    and i want that when any data inserted in the table,
    a constraint will work in the manner that there is
    NO POSSIBILITY TO INSERT the data for
    (teacherid, standard) columns THRICE (3rd time)
    actually i want that only maximum of 2 different data will insert in
    (subjectcode) for the same (teacherid, standard)

    Thanks for the suggestion,
    But, my original problem remains on the same stage.
    Actually this table has all the keys as foreign key of other tables.
    Why?, is not necessary this time.
    If you don't mind kindly give me the code of constraint, if any to solve my problem.
    Problem: I want to insert only maximum of two different Subjects for same Teacher and Standard and not more.
    In real life like following.
    Mr. TCHR-1 is teaching subject SUB-1 in standard STD-1.
    Mr. TCHR-1 is teaching subject SUB-2 in standard STD-1.
    But, can't Mr. TCHR-1 is teaching subject SUB-3 in standard STD-1.
    My table structure is not in the state of changing due to some reasons.
    It is, as previously described.
    create table teaches
    tchrid int,
    std varchar(10),
    subcode varchar(10),
    constraint cpk_teaches primary key(tchrid, subcode, std)
    )

  • Disabling data insertion in WFTASKHISTORY

    Hi
    In Orabpel.WFTaskPkg .. a procedure called 'updateTask' is defined;
    In the procedure following line is executed :
    IF p_IsVersionable = BOOLEAN_TRUE_STRING THEN
    createHistory(p_taskID);
    END IF;
    It inserts a record in the WFTASKHISTORY table if p_IsVersionable = true....
    Is it possible to set the value of 'p_IsVersionable' so that I can control the data insertion in WFTASKHISTORY table ? Any pointer will be helpful!!
    Thanks

    Hi Satish, this is an interesting question and as Michaela points out, Comments should be used to store non numeric data elements.
    However, your question makes me wonder what exectly do you want to enter characters for ? How would your use of characters for an amount be defined ? In case you would like to tell something about the amount (like the 100 for write-off is due to a customer not paying any bills), then you would use comments. I can imagine however that you want to use characters for sanother reason like adding new memebers to the database, but then you have to follow a different procedure.
    So please explain to what other use of characters you are thinking about.
    Best regards,
    Edwin van Geel

  • How is the data inserted into JTF_RS_SALESREPS?

    Hi,
    I wanted to know how is the data inserted into JTF_RS_SALESREPS.When we import the resources from CRM Foundation responsibilty the data will be inserted into jtf_rs_resource_extns.I think we need to run a concurrent program to transfer data from jtf_rs_resource_extns to JTF_RS_SALESREPS.
    Which Concurrent program should be run?
    Please help.
    Thanks.

    How is the data inserted into CST_INV_QTY_TEMP table ?TABLE: BOM.CST_INV_QTY_TEMP
    http://etrm.oracle.com/pls/et1211d9/etrm_pnav.show_object?c_name=CST_INV_QTY_TEMP&c_owner=BOM&c_type=TABLE
    Thanks,
    Hussein

  • How to read LONG RAW data from one  table and insert into another table

    Hello EVERYBODY
    I have a table called sound with the following attributes. in the music attribute i have stored some messages in the different language like hindi, english etc. i want to concatinate all hindi messages and store in the another table with only one attribute of type LONG RAW.and this attribute is attached with the sound item.
    when i click the play button of sound item the all the messages recorded in hindi will play one by one automatically. for that i'm doing the following.
    i have written the following when button pressed trigger which will concatinate all the messages of any selected language from the sound table, and store in another table called temp.
    and then sound will be played from the temp table.
    declare
         tmp sound.music%type;
         temp1 sound.music%type;
         item_id ITEM;
    cursor c1
    is select music
    from sound
    where lang=:LIST10;
    begin
         open c1;
         loop
              fetch c1 into tmp; //THIS LINE GENERATES THE ERROR
              temp1:=temp1||tmp;
              exit when c1%notfound;
         end loop;
    CLOSE C1;
    insert into temp values(temp1);
    item_id:=Find_Item('Music');
    go_item('music');
    play_sound(item_id);
    end;
    but when i'm clicking the button it generates the following error.
    WHEN-BUTTON-PRESSED TRIGGER RAISED UNHANDLED EXCEPTION ORA-06502.
    ORA-06502: PL/SQL: numeric or value error
    SQL> desc sound;
    Name Null? Type
    SL_NO NUMBER(2)
    MUSIC LONG RAW
    LANG CHAR(10)
    IF MY PROCESS TO SOLVE THE ABOVE PROBLEM IS OK THEN PLESE TELL ME THE SOLUTION FOR THE ERROR. OTHER WISE PLEASE SUGGEST ME,IF ANY OTHER WAY IS THERE TO SOLVE THE ABOVE PROBLEM.
    THANKS IN ADVANCE.
    D. Prasad

    You can achieve this in many different ways, one is
    1. Create another VO based on the EO which is based on the dest table.
    2. At save, copy the contents of the source VO into the dest VO (see copy routine in dev guide).
    3. commiting the transaction will push the data into the dest table on which the dest VO is based.
    I understand that if we attach VO object instance to region/page, we only can pull and put data in to only one table.
    if by table you mean a DB table, then no, you can have a VO based on multiple EOs which will do DMLs accordingly.Thanks
    Tapash

  • Fetch data from one table and insert into two tables in desired format

    I have similar to the following data in a table and it is not normalized. The groupID is being used to group two records of similar nature.
    DECLARE @OldDoc TABLE (oldDocID INT, groupID INT, deptID INT)
    INSERT INTO @OldDoc (oldDocID, groupID) VALUES (1, NULL, 111),(2,NULL,111),(3,1,111),(4,NULL,333),(5,1,222),(6,NULL,333),(7,2,222),(8,2,333),(9,NULL,111),(10,3,222),(11,NULL,333),(12,3,444)
    I need to process the data from the above table (@OldDoc) and write into two new tables (@NewDoc and @NewDocGroup) as follows.
    oldDocID should be stored as newDocID when inserting to @NewDoc table. Only records with groupID NULL and one record (first one) per group should be considered (For example, oldDocID 5 is not considered as 3 and 5 belong to the same groupID 1) for insertion. 
    DECLARE @NewDoc TABLE (newDocID INT)
    INSERT INTO @NewDoc (newDocID) VALUES (1),(2),(3),(4),(6),(7),(9),(10),(11)
    All records from @OldDoc should be considered for insertion into @NewDocGroup table. OldDocID is inserted as NewDocID and deptID is as-is. Instead of groupID, the ID of the first record in the 
    group should be considered as parentNewDocID (For example, 3 is considered as parentNewDocID for newDocID 5 as 3 and 5 belong to the same groupID in @OldDoc table) for the newDocID.
    DECLARE @NewDocGroup (newDocID INT, parentNewDocID INT, deptID INT)
    INSERT INTO @NewDocGroup (newDocID, parentNewDocID, deptID) VALUES (1,1,111),(2,2,111),(3,3,111),(4,4,333),(5,3,222),(6,6,333),(7,7,222),(8,7,333),(9,9,111),(10,10,222),(11,11,333),(12,10,444)
    How do I accomplish the above using SQL ? Thanks for the help.

    >> I have similar to the following data in a table and it is not normalized. The group_id is being used to group two records [sic] of similar nature. <<
    Rows are not records. Tables have to have a key by definition. You do not do math with identifiers, so they should not be numeric. Let's ignore that error for now. In short, you are posting garbage. If you had followed Forum Netiquette, would you have posted
    this? 
    CREATE TABLE Old_Documents
    (old_doc_id INTEGER NOT NULL PRIMARY KEY, 
     group_id INTEGER, 
     dept_nbr INTEGER NOT NULL
       REFERENCES Departments (dept_nbr));
    INSERT INTO Old_Documents(old_doc_id, group_id, dept_nbr) 
    VALUES  (1, NULL, 111), 
    (2, NULL, 111), 
    (3, 1, 111), 
    (4, NULL, 333), 
    (5, 1, 222), 
    (6, NULL, 333), 
    (7, 2, 222), 
    (8, 2, 333), 
    (9, NULL, 111), 
    (10, 3, 222), 
    (11, NULL, 333), 
    (12, 3, 444);
    >> I need to process the data from the above table (Old_Documents) and write into two new tables (New_Documents and New_Documents_Groups) as follows. <<
    Just like punch cards and mag tape data processing! Being old and being new are a status, not another kind of entity. But that is how mag tapes work. And you even use the verb "fetch" from tape files. This design flaw is called  attribute splitting.
    Do you have a Male_Personnel and Female_Personnel table? NO! It is just Personnel! 
    >> old_doc_id should be stored as new_doc_id when inserting to New_Documents table. Only records [sic] with group_id NULL and one record [sic] (first [sic; no ordering in a table] one) per group should be considered (For example, old_doc_id 5 is not considered
    as 3 and 5 belong to the same group_id =1) for insertion. <<
    Think about your punch card mindset. Why did you physically materialize that redundant New_Documents table? Let me answer that: this is how you work with punch cards! In SQL we use a VIEW:
    CREATE VIEW New_Documents (new_doc_id)
    AS 
    SELECT old_doc_id 
      FROM Old_Documents;
    >> All records [sic] from Old_Documents should be considered for insertion into New_Documents_Groups table. The old_doc_id is inserted as new_doc_id and dept_nbr is as-is. Instead of group_id, the ID [sic: which identifier??] of the first [sic: tables
    have no ordering like a deck of punch cards] record [sic] in the group should be considered as parent_new_doc_id (For example, 3 is considered as parent_new_doc_id for new_doc_id 5 as 3 and 5 belong to the same group_id in Old_Documents table) for the new_doc_id.
    <<
    Why not use 5 as the parent? My guess is that you are trying to form equivalence classes. See:
    https://www.simple-talk.com/content/print.aspx?article=2020
    --CELKO-- Books in Celko Series for Morgan-Kaufmann Publishing: Analytics and OLAP in SQL / Data and Databases: Concepts in Practice Data / Measurements and Standards in SQL SQL for Smarties / SQL Programming Style / SQL Puzzles and Answers / Thinking
    in Sets / Trees and Hierarchies in SQL

  • Inserting rows in table while logging errors...Urgent

    Hi all,
    i want to insert rows from temp table to main table.
    if temp table contains 100 rows.then all rows must be inserted
    in main table.if their is any exception while insertion,
    insertion must nt stop and erroring rows along with error no must be inserted in
    error log table.
    How to do this ?
    i used forall save exceptions to achieve this .
    but ws able to save only error no. but not errroing rows in error log table.
    is there any efficient way to achieve this?
    can forall be modified in way that it inserts erroring rows also?????

    Hi,
    1.) does dbms_errorlog works fine in this case?If all of your 807 tables having same columns then i think it will work (NOT TESTED).
    By the way are you also trying to insert rows for all tese 807 tables in a single PL/SQL block?
    Please post how you are trying to insert then we can think about it,
    Oracle ERROR_LOG table allows us to have any number of columns.
    From the Oracle Docs:
    The number of columns in this part of the error logging table can be zero, one, or more, up to the number of columns in the DML table. If a column exists in the error logging table that has the same name as a column in the DML table, the corresponding data from the offending row being inserted is written to this error logging table column. If a DML table column does not have a corresponding column in the error logging table, the column is not logged. If the error logging table contains a column with a name that does not match a DML table column, the column is ignored.
    Regards

  • How to insert data from file to table??

    I need to know that how can i insert data in multiple column through file. I can simply insert data in one column table but couldnt find out the way to put data in all column.
    My data store in a file
    ************************************************text.txt***************
    133, shailendra, nagina, 14/H, 45637, 9156729863
    **************************************************************my_data(table)**********
    trying to insert into below table...
    id, name, last_name, add, pin. mob
    Let me know if anything else needed..:)

    Hi Shailendra,
    Actually, in SQL Developer, you can open a connection to the target schema, right-click on the Tables node in the navigator tree view, select Import Data, then use the Data Import Wizard. It is extremely flexible. It looks like you have a comma separated variable file, so if you select Format: csv and Import Method: insert it will probably work just fine.
    To minimize the chance of errors during import, pick a preview limit value so the wizard can examine the size and data type of all columns in as many data rows as possible, then review the data type/size for each column in the next wizard page and override as necessary. For date columns it is also important to choose the appropriate format mask.
    Hope this helps,
    Gary
    SQL Developer Team

  • Read data from E$ table and insert into staging table

    Hi all,
    I am a new to ODI. I want your help in understanding how to read data from an "E$" table and insert into a staging table.
    Scenario:
    Two columns in a flat file, employee id and employee name, needs to be loaded into a datastore EMP+. A check constraint is added to allow data with employee names in upper case only to be loaded into the datastore. Check control is set to flow and static. Right click on the datastore, select control and then check. The rows that have violated the check constraint are kept in E$_EMP+ table.
    Problem:
    Issue is that I want to read the data from the E$_EMP+ table and transform the employee name into uppercase and move the corrected data from E$_EMP+ into EMP+. Please advise me on how to automatically handle "soft" exceptions in ODI.
    Thank you

    Hi Himanshu,
    I have tried the approach you suggested already. That works. However, the scenario I described was very basic. Based on the error logged into the E$_EMP table, I will have to design the interface to apply more business logic before the data is actually merged into the target table.
    Before the business logic is applied, I need to read each row from the E$EMP table first and I do not know how to read from E$EMP table.

Maybe you are looking for

  • Errors in Importing Web Services to Composite Application

    Hi Everyone, I have created a CAF and exposed its operations as Web Services using CE 7.1 EhP1 SP2 platform. Now I would like to import the web services into another CAF project as external service in the same CE AS. The following are the error messa

  • Non supported device issue

    Hi : I have a DELL Axim x30, that I have been using to read ebooks for years. I have Adobe reader for pocket pc installed on my device. I have been having problems with trying to load ePUB ebooks and PDF ebooks onto my device. I downloaded the latest

  • Express 2.0 Template Files

    (436.3 7-23-1997)) id 8525653D.005BAED9 ; Mon, 27 Oct 1997 11:41:25 -0500 X-Lotus-FromDomain: WORLDBANK Can anyone direct me on how to modify a Forte Express .tpl file? I found the corresponding .pex file but I need to know how to generate a new .tpl

  • Why can't i open file on ipad ?

    I have an old version of Numbers on my ipad and recently upgraded to the new version. Simultaneously, i have bought the new version of iworks on my mac. I have created a file on my mac and tried to open/download on my ipad, but it does not and te fol

  • HT1331 Why does this article still not indicate that Airport Disk Utility no longer exists?

    ADU stopped existing after 10.6. I see this article was updated in May; that means two revisions of OSX which don't contain this utility have been announced or released at that time yet it's still wrong. It's near the top of google rankings, making t