Join on flat files

Hi,
I would like to do join using flat files. Is it possible to do join in ODI?
Please let me if any possibilities.
Thanks in adv

As per Bhabani's response, yes - this is possible.
ODI will perform the following actions when you do this (for example, let's say there are 2 flat files - A & B):
- Creates two temporary flow tables in the database - one for each file
- Loads data from file A and B into their respective flow tables
- Optionally creates indexes on required join/filter fields on the flow tables
- Selects data from the two flow tables by applying any join and filter logic that you have defined, and finally,
- Loads the resulting data set into the target table
Hope this helps.

Similar Messages

  • What is the use of joiner that is only for combining tow flat file or other

    hi all
    joiner is use full for only combining tow or more flat files
    that purpose only or any other scenario is the for joiner transformation
    Thanks
    Sreedhar

    http://www.scribd.com/doc/24360471/Joiner-Transformation-Overview
    Please refer this.

  • How to join 5 different tables using SQL to make it to a flat file structur

    I am trying to load five differnt tables into one flat file structure table without cartesian product.
    I have five different tables Jobplan, Jobtask(JT), Joblabor(JL), Jobmaterial(JM) and Jpsequence(JS) and the target table as has all the five tables as one table.
    The data i have here is something like this.
    jobplan = 1record
    jobtask = 5 records
    joblabor = 2 records
    jobmaterial = 1 record
    jpsequence = 3 records
    The output has to be like this.
    JPNUM     DESCRIPTION     LOCATION     JT_JPNUM     JT_TASK     JL_JPNUM     JL_labor     JM_JPNUM     JM_MATERIAL     JS_JPNUM     JS_SEQUENCE
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     10     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     20     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     30     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     40     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     1001     50     NULL     NULL     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     1001     Sam     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     1001     Mike     NULL     NULL     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     1001     Hammer     NULL     NULL
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     1001     1
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     1001     2
    1001     Test Jobplan     USA     NULL     NULL     NULL     NULL     NULL     NULL     1001     3
    Please help me out with this issue.
    Thanks,
    Siva
    Edited by: 931144 on Apr 30, 2012 11:35 AM

    Hope below helps you
    CREATE TABLE JOBPLAN
    ( JPNUM NUMBER,
      DESCRIPTION VARCHAR2(100)
    INSERT INTO JOBPLAN VALUES(1001,'Test Jobplan');
    CREATE TABLE JOBTASK
    ( LOCATION VARCHAR2(10),
      JT_JPNUM NUMBER,
      JT_TASK  NUMBER
    INSERT INTO JOBTASK VALUES('USA',1001,10);
    INSERT INTO JOBTASK VALUES('USA',1001,20);
    INSERT INTO JOBTASK VALUES('USA',1001,30);
    INSERT INTO JOBTASK VALUES('USA',1001,40);
    INSERT INTO JOBTASK VALUES('USA',1001,50);
    CREATE TABLE JOBLABOR
    ( JL_JPNUM NUMBER,
      JL_LABOR VARCHAR2(10)
    INSERT INTO JOBLABOR VALUES(1001,'Sam');
    INSERT INTO JOBLABOR VALUES(1001,'Mike');
    CREATE TABLE JOBMATERIAL
    ( JM_JPNUM    NUMBER,
      JM_MATERIAL VARCHAR2(10)
    INSERT INTO JOBMATERIAL VALUES(1001,'Hammer');
    CREATE TABLE JOBSEQUENCE
    ( JS_JPNUM    NUMBER,
      JS_SEQUENCE NUMBER
    INSERT INTO JOBSEQUENCE VALUES(1001,1);
    INSERT INTO JOBSEQUENCE VALUES(1001,2);
    INSERT INTO JOBSEQUENCE VALUES(1001,3);
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             JT.LOCATION     AS LOCATION    ,
             JT.JT_JPNUM     AS JT_JPNUM    ,
             JT.JT_TASK      AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBTASK JT
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             JL.JL_JPNUM     AS JL_JPNUM    ,
             JL.JL_labor     AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBLABOR JL
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             JM.JM_JPNUM     AS JM_JPNUM    ,
             JM.JM_MATERIAL  AS JM_MATERIAL ,
             NULL            AS JS_JPNUM    ,
             NULL            AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBMATERIAL JM
    UNION ALL
    SELECT   JP.JPNUM        AS JPNUM       ,
             JP.DESCRIPTION  AS DESCRIPTION ,
             NULL            AS LOCATION    ,
             NULL            AS JT_JPNUM    ,
             NULL            AS JT_TASK     ,
             NULL            AS JL_JPNUM    ,
             NULL            AS JL_labor    ,
             NULL            AS JM_JPNUM    ,
             NULL            AS JM_MATERIAL ,
             JS.JS_JPNUM     AS JS_JPNUM    ,
             JS.JS_SEQUENCE  AS JS_SEQUENCE
    FROM JOBPLAN JP, JOBSEQUENCE JS;
         JPNUM DESCRIPTION     LOCATION      JT_JPNUM    JT_TASK   JL_JPNUM JL_LABOR     JM_JPNUM JM_MATERIA   JS_JPNUM JS_SEQUENCE
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        10         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        20         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        30         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        40         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    USA        1001        50         NULL       NULL       NULL       NULL    NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       1001       Sam        NULL       NULL    NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       1001       Mike       NULL       NULL    NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       1001       Hammer  NULL          NULL
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    1001          1
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    1001          2
          1001 Test Jobplan    NULL       NULL        NULL       NULL       NULL       NULL       NULL    1001          3
    {code}                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • How to Add Header and Footer in Flat file

    HI All,
    I am using ODI 11g.
    MY interface is populating data from joining to 2 tables through sql query and Loads in to Flat file.
    But i need to add Header and Footer in the Flat file.
    In the below 3 lines
    The 1st line is header ....2nd line is Data ....3rd line is Footer.
    Any ideas how to work on this.
    HSB9KT.KTPRTO.EBSDATA.D0000000 20120626020636
    CFFE00263 SOLGSR014991340000527894536058173880207SOLNE06242012072420120000009048382650000247MRS
    T00000000000000000000
    Thanks,
    Lony

    Hi Drieux,
    i am able to add Header and Footer in this.
    But i want the header like Its like HSB9KT.KTPRTO.EBSDATA.D0000000 || SYSDATE (YYYYMMDDHHMMSS)
    In my Out put i only see header as HSB9KT.KTPRTO.EBSDATA.D0000000 .
    I have added a Variable where i am using SELECT to_char(sysdate,'YYYYMMDDHHMMSS') FROM DUAL query.
    create header (HSB9KT.KTPRTO.EBSDATA.D0000000 #v_DATE)
    /*$$SNPS_START_KEYSNP$CRDWG_TABLESNP$CRTABLE_NAME=FFT_SOLSNP$CRLOAD_FILE=/home/tarak/FFT_SOL.datSNP$CRFILE_FORMAT=FSNP$CRFILE_SEP_FIELD=0x0009SNP$CRFILE_SEP_LINE=0x000ASNP$CRFILE_FIRST_ROW=0SNP$CRFILE_ENC_FIELD=SNP$CRFILE_DEC_SEP=SNP$CRSNP$CRDWG_COLSNP$CRCOL_NAME=descrSNP$CRTYPE_NAME=STRINGSNP$CRLINE_OFFSET=1SNP$CRLENGTH=994SNP$CRPRECISION=994SNP$CR$$SNPS_END_KEY*/
    Any idea why this sysdate part is not showing in Header.

  • How to delete extra tab space in the flat file??

    Hi all,
        I have a flat file on presentation server which has 20000 records. In this file there are 16 fields and distance between each field is different and what I want is each field should have one tab space between another.
    Thanks & Regards
    Jerry

    Hi,
    I dont want to join the fields but to seperate them with a tab space.
    Regards
    Jerry

  • Flat File as Target - MultiRecord Type

    How do I write from Oracle into a Flat File as Target where the data in Flat file is a multirecord type. For ex. Sales Order header and Sales Order Line info.
    I tried to do this by using a Joiner to the sales order header, line and items table and then mapping outgrp of joiner to the Flat file data with two record types in it. But when I generate and run this it could only create header records in out put file

    Dear 396217,
    As far as I am concerned you cannot really push out to a multi record structure without manual intervention.
    There are a few things you can do: on the target (not sure whether that was already available in 9.0.3... but it definitely is in 9.0.4) you can set the flat file as a target to generate field names as headers. If that is what you want... then great.
    The initial scenario: orders and order items in one model, is not supported. What you could do... is write a transformation to do this. Use calls to UTL_FILE to implement those and generate your own records. Primitive... but it works.
    Alternatively, create 2 files and have a separate external process merge these. Not ideal; I realize that.
    Mark.

  • Loading data file Flat file to Oracle DB

    I have 600/700 MB datafiles on AIX box which need to upload to Oracle using ODI.
    I am considering the below KMs ...which would be more appropriate in this case
    LKM File to Oracle
    LKM File to Oracle (EXTERNAL TABLE)
    Why should we use LKM File to Oracle (EXTERNAL TABLE) at all in any case, and do we need to create seperate table structure for that or ODI would take care internally?

    Hi,
    Loads data from a File to an Oracle staging area using the EXTERNAL TABLE SQL Command.Because this method uses the native EXTERNAL TABLE command, it is more efficient than the standard “LKM File to oracle” when dealing with large volumes of data. However,the loaded file must be accessible from the Oracle server machine.Note that the data of the source file is not duplicated in the Oracle staging area table. This table acts only as a"synonym" of the file. This may sometimes lead to Knowledge Modules if the same file is joined with other large tables in your Interface. For optimization purpose, you canenhance this LKM by adding an extra step that copies then file content to an actual physical table in the Oracle staging area Consider using this LKM if your source is a large flat file and your staging area is an Oracle database.
    Regards,
    surya

  • MAPPING FLAT FILE TO ORACLE ERROR!! Please help

    hello all
    I am using OWB on Oracle9i.
    Trying to map flat file (file.txt) to oracle table,
    Validate: okey, some warning
    Generate/Deploy: successful.
    But when I check to the database sqlplus there is no rows existed. I verify that I connect to the right database.
    My question is:
    When doing a map from flat file to oracle table. Do I have to have some thing in the middle such as (filter, join) or I just can do a straight mapping if I don't have any condition to filter out.
    Please response if you know the answer.
    Your answer greatly appreciated. Thank you for your help
    Regards,

    If you are using an OWB version which is 9.0.3.x or earlier, to load a flat file into the database, you should generate SQL loader files and run them either manually or by using a scheduler.
    The new version (9.0.4) supports external tables, so you don't have to use SQL loader. Please refer to the user manual for more details.
    Regards:
    Igor

  • How can I change from JDBC to flat files

    I have an application setup that runs on MySQL using JDBC. But, I'm having problem with the database(ConnectorJ to be exact). I just want check the functionality of this application. Can I use a flat file as opposed to the DB? How can I possibly do this?

    The text driver supports a very simple SQL. Inserts, deletes are allowed. Not sure about updates. Joins on tables are not.
    Of course it will still require a connection string.

  • Save flat file in bdc

    in bdc we have to save the flat file in desktop or c drive?plz tell
    Edited by: Alvaro Tejada Galindo on Feb 7, 2008 10:15 AM

    Hi,
    check the  code for better understanding.
    Regards
    Bala.M
    REPORT  zi_flatfile_zisobj.
    * Data Declaration
    * Internal Tables
    DATA : BEGIN OF itd_zisobj OCCURS 0 ,
    zinref        LIKE  zisobj-zinref,
    ziaccomtype   LIKE  zisobj-ziaccomtype,
    zietgnr       LIKE  zisobj-zietgnr,
    zioldnref     LIKE  zisobj-zioldnref,
    zicnttyp      LIKE  zisobjcnt-zicnttyp,
    END OF itd_zisobj.
    DATA : file_length TYPE i.
    START-OF-SELECTION.
    PERFORM data_selection_join.
    PERFORM gui_download_file.
    END-OF-SELECTION.
    *&      Form  data_selection_join
    *       text
    *  -->  p1        text
    *  <--  p2        text
    FORM data_selection_join .
    * Data selection and Inner join based on Contract Type Eq 'U'.
    SELECT   a~zinref
    a~ziaccomtype
    a~zietgnr
    a~zioldnref
    b~zicnttyp
    INTO  CORRESPONDING FIELDS OF TABLE itd_zisobj
    FROM ( zisobj AS a INNER JOIN zisobjcnt AS b ON a~zicountry  = b~zicountry
    AND a~ziplace    = b~ziplace
    AND a~zihouse    = b~zihouse
    AND a~ziaccom    = b~ziaccom
    AND b~zicnttyp   = 'U' ) .
    ENDFORM.                    " data_selection_join
    *&      Form  gui_download_file
    *       text
    *  -->  p1        text
    *  <--  p2        text
    FORM gui_download_file .
    * FM to download to the presentation server
    CALL FUNCTION 'GUI_DOWNLOAD'
    EXPORTING
    *   BIN_FILESIZE                    =
    filename                        = 'H:\TEST1.TXT'
    filetype                        = 'ASC'
    append                          = ' '
    write_field_separator           = ','
    header                          = '00'
    trunc_trailing_blanks           = ' '
    write_lf                        = 'X'
    col_select                      = ' '
    col_select_mask                 = ' '
    dat_mode                        = ' '
    confirm_overwrite               = ' '
    no_auth_check                   = ' '
    codepage                        = ' '
    *   IGNORE_CERR                     = ABAP_TRUE
    replacement                     = '#'
    write_bom                       = ' '
    trunc_trailing_blanks_eol       = 'X'
    wk1_n_format                    = ' '
    wk1_n_size                      = ' '
    wk1_t_format                    = ' '
    wk1_t_size                      = ' '
    *   WRITE_LF_AFTER_LAST_LINE        = ABAP_TRUE
    *   SHOW_TRANSFER_STATUS            = ABAP_TRUE
    IMPORTING
    filelength                      =  file_length
    TABLES
    data_tab                       = itd_zisobj
    *   FIELDNAMES                      =
    EXCEPTIONS
    file_write_error                = 1
    no_batch                        = 2
    gui_refuse_filetransfer         = 3
    invalid_type                    = 4
    no_authority                    = 5
    unknown_error                   = 6
    header_not_allowed              = 7
    separator_not_allowed           = 8
    filesize_not_allowed            = 9
    header_too_long                 = 10
    dp_error_create                 = 11
    dp_error_send                   = 12
    dp_error_write                  = 13
    unknown_dp_error                = 14
    access_denied                   = 15
    dp_out_of_memory                = 16
    disk_full                       = 17
    dp_timeout                      = 18
    file_not_found                  = 19
    dataprovider_exception          = 20
    control_flush_error             = 21
    OTHERS                          = 22
    IF sy-subrc NE 0.
    MESSAGE e398(00) WITH 'Error in Downloading File' 'TEST1.TXT' 'Retry'.
    ENDIF.
    ENDFORM.                    " gui_download_file

  • Query using flat file in the FROM Statement

    I have a flat file that contains about 10k rows that I need to join to a table with 2million plus rows. My problem is I cannot load the flat file into a Oracle table or create a new table to contain the 10k rows because I lack of permissions to create table on my user account. I have tried using MS-Access to query the two tables using ODBC, but the query times were horribly slow and would time out. I was curious if there is a way of calling the flat file as a table in my SQL statement from Oracle?
    I know the following doesn't work but I was wondering if there is something out there that would allow me to do something like the following?
    example:
    SELECT a.COL1, a.COL2, ff.COLa, FF.COLb
    FROM apple a, c:\temp\flat_file.csv ff
    WHERE a.COLa=ff.COLa
    db version - 10g
    OS - Windows XP Pro
    Thanks in Advance!

    user9339442 wrote:
    I have a flat file that contains about 10k rows that I need to join to a table with 2million plus rows. My problem is I cannot load the flat file into a Oracle table or create a new table to contain the 10k rows because I lack of permissions to create table on my user account. I have tried using MS-Access to query the two tables using ODBC, but the query times were horribly slow and would time out. I was curious if there is a way of calling the flat file as a table in my SQL statement from Oracle?
    I know the following doesn't work but I was wondering if there is something out there that would allow me to do something like the following?
    example:
    SELECT a.COL1, a.COL2, ff.COLa, FF.COLb
    FROM apple a, c:\temp\flat_file.csv ff
    WHERE a.COLa=ff.COLa
    db version - 10g
    OS - Windows XP Pro
    Thanks in Advance!Is there some reason you cannot get the data loaded into the local database?
    If you absolutely cannot (either through creating and loading objects, or having someone else do it for you) i would likely recommend installing your own database (Oracle XE is free and will easily handle your flat file)
    http://www.oracle.com/technetwork/database/express-edition/downloads/index.html
    I would install that, load your data into a schema you create, and then create a database link to the remote database and use that in your query.
    I would highly push you to work with the administrators of the database though and get them to set you up a schema where you can create objects (with access to the objects in other schemas that you require, if only read permitted).

  • How to import flat file as a source in DAC Client

    Dear All,
    I am doing some R&D,I have a problem that how can i import the flat file in DAC client which i am using as a source.
    Can any one please explain in detail step wise.
    Regards
    Tarang Jain

    Hi Tarang,
    How long will it take to run the sample execution plan? Since when i run 'CRM Automotive' EP, it runs for more than 3 hours and still it is running. My system config is 1GB RAM with windows os. I have also opened the DAC server and client in console mode and I m not able to see any errors. I have only got only the task 'Load row into run table ' completed successfully and next 10 are in Running stage from the past 2 hours.
    My DAC client shows continously the below shown message for 1 hr.
    'Sep 30, 2008 3:52:59 PM com.siebel.analytics.etl.client.data.model.ConnectionHan
    dler getConnection
    INFO: Getting connection for: Model structure for W_ETL_RUN_STEP (RUN_STEP) (cur
    rently 0)
    Sep 30, 2008 3:52:59 PM com.siebel.analytics.etl.client.data.model.ResultSetPars
    er logMinutesSeconds
    INFO: Query took 0 minutes, 0 seconds
    Sep 30, 2008 3:52:59 PM com.siebel.analytics.etl.client.data.model.ConnectionHan
    dler cleanUp
    INFO: Returning connection for: Model structure for W_ETL_RUN_STEP (RUN_STEP) (c
    urrently 1)
    Sep 30, 2008 3:52:59 PM com.siebel.analytics.etl.client.data.model.ResultSetPars
    er logMinutesSeconds
    INFO: Populating DACObjects took 0 minutes, 0 seconds
    Sep 30, 2008 3:52:59 PM com.siebel.analytics.etl.client.data.model.DACChildTable
    Model setParentID
    INFO: PARENT ROW ID IS: 95D45F88DBB268B3D19DB71B4E42D3FF
    Sep 30, 2008 3:52:59 PM com.siebel.analytics.etl.client.data.objectcomponents.Mo
    delStruct setQuery
    INFO: SELECT
    RUN_STEP.DEPTH
    ,RUN_STEP.STEP_NAME
    ,RUN_STEP.STATUS
    ,RUN_STEP.START_TS
    ,RUN_STEP.END_TS
    ,RUN_STEP.STATUS_DESC
    ,RUN_STEP.PHY_SRC_NAME
    ,RUN_STEP.PHY_TRGT_NAME
    ,RUN_STEP.PHY_FOLDER_NAME
    ,RUN_STEP.SERVER_NAME
    ,RUN_STEP.PHASE_NAME
    ,RUN_STEP.EXEC_TYPE_NAME
    ,RUN_STEP.SUCESS_ROWS
    ,RUN_STEP.FAILED_ROWS
    ,RUN_STEP.ERROR_CODE
    ,RUN_STEP.DEFN_STEP_WID
    ,RUN_STEP.ROW_WID
    ,RUN_STEP.LAST_UPD
    ,RUN_STEP.APP_WID
    ,RUN_STEP.RUN_WID
    ,RUN_STEP.STEP_WID
    ,RUN_STEP.PHY_SRC_WID
    ,RUN_STEP.PHY_TRGT_WID
    ,RUN_STEP.PHY_FOLDER_WID
    ,A_W_ETL_APP.ROW_WID
    ,A_W_ETL_APP.NAME
    FROM W_ETL_RUN_STEP RUN_STEP
    INNER JOIN W_ETL_APP A_W_ETL_APP ON RUN_STEP.APP_WID=A_W_ETL_APP.ROW_WID
    WHERE
    (RUN_STEP.RUN_WID='95D45F88DBB268B3D19DB71B4E42D3FF')
    ORDER BY
    RUN_STEP.START_TS ASC,
    RUN_STEP.DEPTH ASC'
    Any comments on this?
    I had also copied the CRM Analytics source file into the "D:\Informatica\PowerCenter8.1.1\server\infa_shared\SrcFiles"
    as metioned in Appendices ' Configuring Universal Adapter for CRM Analytics'.
    With Regards,
    Prads

  • Flat File to DB then back to Flat File

    I need to compare the contents of a flat file with two columns in a db table. Then do an insert into the db if there is a difference between the two and the do a select of the latest data on the db then write it to a file.
    I need help urgently, thats why i joined this forum today.
    Regards,
    Koupes

    571446 I am using a mind meld to extract the neccessary information (like the database you are using) directly from your brain. Please stand by while I write all the code for you.
    All hail Sun.

  • BDC (Flat File Data Validation) - Code

    I am trying to validate flat file data BEFORE performing BDC (Call Trans. or Session)..
    Pls help me out in below code for xk02..
    DATA: BEGIN OF itab occurs 0,  "ITAB having flat file data.
          lifnr(16) ,
          bukrs(4),
          ekorg(4),
          END OF itab.
    DATA: BEGIN OF int_final occurs 0,
          lifnr(16) ,
          bukrs(4),
          ekorg(4),
          status(6),
          message(6),
          END OF int_final.
    DATA: int_final TYPE TABLE OF int_final.
    DATA: wa_itab TYPE TABLE OF itab.
    DATA: validate_itab TYPE TABLE OF itab. "VALIDATE_ITAB having master data.
    DATA: wa_validate_itab TYPE TABLE OF itab.
    FORM data_validation .
    SELECT LFB1LIFNR LFB1BUKRS LFM1~EKORG INTO TABLE validate_itab
             FROM LFB1 INNER JOIN LFM1 ON LFB1LIFNR = LFM1LIFNR.
    IF sy-subrc = 0.
    SORT validate_itab BY lifnr bukrs ekorg.
    ENDIF.
    LOOP AT itab INTO wa_itab.
    READ TABLE validate_itab WITH KEY
    lifnr = itab-lifnr
    bukrs = itab-bukrs
    ekorg = itab-ekorg
    BINARY SEARCH.
    IF sy-subrc NE 0.
    PERFORM f_error_log USING text-005. "Invalid Value Set
    CONTINUE.
    ENDIF.
    ENDLOOP.
    ENDFORM.                    " data_validation
    *&      Form  f_error_log
    FORM f_error_log USING l_message TYPE string.
    CLEAR : fs_final.
    fs_final-lifnr = itab-lifnr.
    fs_final-bukrs = itab-bukrs.
    fs_final-ekorg = itab-ekorg.
    fs_final-status = text-014. "Error
    fs_final-message = l_message.
    APPEND fs_final TO int_final.
    ENDFORM.                    " f_error_log
    Thanks..

    Hi GAurav,
    I have a small question in th validation.
    In LFM1~LIFNR does not contian any value how u r comparing both and one more thing After getting the data using GUI_upload u will get the data into validate_tab.
    Loop at Vlidate_tab into wa_itab.
    SELECT LFB1LIFNR LFB1BUKRS LFM1~EKORG INTO TABLE validate_itab
    FROM LFB1 INNER JOIN LFM1 ON LFB1~LIFNR =  wa_itab-lifnr.
    endllop.
    Thanks,

  • Flat file (csv) to oracle database - get dictionary data from oracle table

    Hello,
    I need to develop following scenario. I have flat csv file with some data, for example:
    City;Address;Name
    In oracle schema, there is table where data from imported csv should be written. Also, in the same schema, there is table named CityDictrionary. After ODI loads each row from csv, it should translate text representation of City to numerric ID got from CityDictionary table. And this numeric value should be placed to destination table, instead of direct text value from csv.
    What is the simplest way to accomplish this task? Can you provide any tips?

    You can achieve this easily with a single ODI interface. The flat file and the CityDictionary tables are your sources. Identify the field on the CityDictionary table and the corresponding field on the flat file that will be used to join the data sets (e.g. the city name). Create a join between the sources using these fields. Because you're using a flat file as one of your sources, the join logic will have to be performed on either the staging or the target.
    The target is your destination table, and you should map the ID from the CityDictionary table to the appropriate field in the target, as well as any other required fields from the flat file.
    This type of interface is a fairly typical method of populating a normalized table.
    Alternatively, you can use a lookup - see the following blog for an example. In your case, the flat file will be your source and the CityDictionary table will be used for the lookup.
    http://www.odigurus.com/2012/02/lookup-transformation-using-odi.html
    Edited by: _Phil on Oct 1, 2012 11:52 PM
    Edited by: _Phil on Oct 1, 2012 11:57 PM                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

Maybe you are looking for