Error while reading data through External Table!!!

CREATE TABLE "COGNOS"."EXT_COGNOS_TBS9_TEST"
(     "ITEM_DESC" VARCHAR2(200 BYTE),
"EXT_CODE" VARCHAR2(20 BYTE),
"RC_DATE" DATE,
"RES_KD_AMNT" NUMBER(18,3),
"RES_FC_AMNT" NUMBER(18,3),
"NRES_KD_AMNT" NUMBER(18,3),
"NRES_FC_AMNT" NUMBER(18,3),
"TOTAL" NUMBER(18,3),
"OF_WHICH_OVR1" NUMBER(18,3)
ORGANIZATION EXTERNAL
( TYPE ORACLE_LOADER
DEFAULT DIRECTORY "EXTDATADIR"
ACCESS PARAMETERS
( RECORDS
DELIMITED BY NEWLINE LOAD WHEN *({color:#ff0000}EXT_CODE LIKE 'TBS9%'{color})* FIELDS TERMINATED BY ','
MISSING FIELD VALUES ARE NULL )
LOCATION
( 'TBS9_TEST.CSV'
External table creation went through successfully but am getting error while reading data. Am quite sure error is because of above line in red color. Could you please help me in transforming logic.
Thanks in Advance,
AP

Let's start with the basics...
1) You state that you are getting an error. What error do you get? Is this an Oracle error (i.e. ORA-xxxxx)? If so, please include the error number and the error message as well as the triggering statement. Or is the problem that rows are getting written to the reject file and errors are being written to the log file? If so, what record(s) are being rejected and what are the reasons given in the log file? Or perhaps the problem is something else?
2) You state that you are quite sure that the problem relates to the hilighted code. What makes you quite sure of this?
Justin

Similar Messages

  • Error while selecting date from external table

    Hello all,
    I am getting the follwing error while selecting data from external table. Any idea why?
    SQL> CREATE TABLE SE2_EXT (SE_REF_NO VARCHAR2(255),
      2        SE_CUST_ID NUMBER(38),
      3        SE_TRAN_AMT_LCY FLOAT(126),
      4        SE_REVERSAL_MARKER VARCHAR2(255))
      5  ORGANIZATION EXTERNAL (
      6    TYPE ORACLE_LOADER
      7    DEFAULT DIRECTORY ext_tables
      8    ACCESS PARAMETERS (
      9      RECORDS DELIMITED BY NEWLINE
    10      FIELDS TERMINATED BY ','
    11      MISSING FIELD VALUES ARE NULL
    12      (
    13        country_code      CHAR(5),
    14        country_name      CHAR(50),
    15        country_language  CHAR(50)
    16      )
    17    )
    18    LOCATION ('SE2.csv')
    19  )
    20  PARALLEL 5
    21  REJECT LIMIT UNLIMITED;
    Table created.
    SQL> select * from se2_ext;
    SQL> select count(*) from se2_ext;
    select count(*) from se2_ext
    ERROR at line 1:
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04043: table column not found in external source: SE_REF_NO
    ORA-06512: at "SYS.ORACLE_LOADER", line 19

    It would appear that you external table definition and the external data file data do not match up. Post a few input records so someone can duplicate the problem and determine the fix.
    HTH -- Mark D Powell --

  • Error in accessing data through external table

    Hi,
    I am getting the following error while trying to view data through external table,
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04062: no data source specified
    ORA-06512: at "SYS.ORACLE_LOADER", line 19
    according to this error "KUP-04062: no data source specified", location should be specified and i have specified the location as
    LOCATION (<diectory_name>:' '),
    then also it is giving the same error.
    Please Help!

    I am also got the same error when retrieving the data from external table.
    I resolve this problem, like this:
    1. I imported the flat file(.csv) into repository.
    2. Map that flat file fields to database table cols.
    3. Deployed the mapping using controlcenter, and run the mapping.
    4. if it fails there, i loaded the flatfile data through the command prompt using "sqlldr" statement and control file is our mapping control file.
    5 Then it loaded data into database table sucessfully.
    Thanks.

  • Error while loading  data into External table from the flat files

    HI ,
    We have a data load in our project which feeds the oracle external tables with the data from the Flat Files(.bcp files) in unix.
    While loading the data, we are encountering the following error.
    Error occured (Error Code : -29913 and Error Message : ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04063: un) while loading data into table_ext
    Please let us know what needs to be done in this case to solve this problem.
    Thanks,
    Kartheek

    Kartheek,
    I used Google (mine still works).... please check those links:
    http://oraclequirks.blogspot.com/2008/07/ora-29400-data-cartridge-error.html
    http://jonathanlewis.wordpress.com/2011/02/15/ora-29913/
    HTH,
    Thierry

  • Error while reading data for reporting

    Hi Gurus,
    I tried to get answer from the forums but not got exact answer to slove my issue..
    Q: I had generated 10 reports on Multiprovider(Virtual cube + std Cube) on Bex analyzer.
          But when i am Executing reports 3 reports are getting result where remaining 7 reports are throughing same error.
    Error: error reading the data of infoprovider
    error system error: no more storege space available for extending an internal table
    error massage from data source XXXXXX from Source system XXX
    error error occured during parallel processing
    error error while reading data, navigation is possible
    Where their are no parallel processing are going.
    the 3 reports are taking 15 min to generate the report..
    data is available in both sources with same selection options.
    Can any one please through some light, what exactly need to do to get all the reports and to reduce the response time.
    thanks in advance.
    Sandy

    Hi,
    The problem was solved.
    I had given restriction on 0COUNTRY with out calling Company code in the Column.
    Were 0COUNTRY is also an attribute of 0COMP_CODE.
    so i called both in the report and got the report with results...
    this is how i solved the issue.
    Thanks
    Sandy

  • Error while reading payroll area D0, table T549S

    Hi,
    would you know what this kind of error message could mean?
    "Error while reading payroll area D0, table T549S"
    This is an error while trying to extract payroll data to business warehouse. What should I look for to find the cause of this error?
    Thanks in advance,
    Mantas

    You have to contact the HR consultant in ur tem
    Next they might have not mapped the payroll which you are trying
    REWARD POINTS

  • Error while reading data for Virtual Infoprovider 0TCT_VC11 and 0TCT_VC12

    Hi ,
    The standard reports based on 0TCT_VC11 and 0TCT_VC12 are working fine in development system. When we moved the following objects to test system ,
    0TCT_VC11 - 0TCT_IS11 &  0TCT_DS11
    0TCT_VC12 - 0TCT_IS12 &  0TCT_DS12
    We are not able to view data either in the MP nor at the report level.We face the following issue
    There is still no data source assigned to VirtualProvider 0TCT_VC11
    Error reading the data of InfoProvider 0TCT_VC11
    Error while reading data; navigation is possible
    Can anybody help me on this.
    Regards,
    Lavanya.

    Hello Lavanya,
    Please activate the direct access for the virtual providers. Also note that it is only possible  to use the BI admin cockpit in a myself system.
    Please follow the below steps to activate the same:
    TX: RSA1 ->Infoproviders-> Business Information-> BI Statistics->select
    Virtual provider-> Go to context menu of the virtual cubes-> Activate
    Direct Access. ( Eg virtual cube name )
    Please also generate the reports by following the below steps:
    RSRT -> Query name -> Generate Report.
    Regards,
    Arvind

  • Error while Loading data through .csv file

    Hi,
    I am getting below date error when loading data through into Olap tables through .csv file.
    Data stored in .csv is 20071113121100.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TE_7007 Transformation Evaluation Error [<<Expression Error>> [TO_DATE]: invalid string for converting to Date
    ... t:TO_DATE(u:'2.00711E+13',u:'YYYYMMDDHH24MISS')]
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TT_11132 Transformation [Exp_FILE_CHNL_TYPE] had an error evaluating output column [CREATED_ON_DT_OUT]. Error message is [<<Expression Error>> [TO_DATE]: invalid string for converting to Date
    ... t:TO_DATE(u:'2.00711E+13',u:'YYYYMMDDHH24MISS')].
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TT_11019 There is an error in the port [CREATED_ON_DT_OUT]: The default value for the port is set to: ERROR(<<Expression Error>> [ERROR]: transformation error
    ... nl:ERROR(u:'transformation error')).
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> TT_11021 An error occurred moving data from the transformation Exp_FILE_CHNL_TYPE: to the transformation W_CHNL_TYPE_DS.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
    TRANSF_1_1_1> CMN_1086 Exp_FILE_CHNL_TYPE: Number of errors exceeded threshold [1].
    Any help is greatly appreciated.
    Thanks,
    Poojak

    1) Wrong format, you wont get much support loading OLAP cubes in here I think.
    2) Has your CSV file been anywhere near excel by chance the conversion of 20071113121100 to 2.00711E+13 looks very much like what I see when Excel has an invalid number mask / precision specified for a cell.
    *** We are using EBS. File was already generated by consultants through SQL query. Getting error while loading though Informatica. Target table is setup as date datatype and source is String(19).
    Expression in Informatica is setup as below.
    IIF(ISNULL(CREATED_ON_DT), NULL, TO_DATE(CREATED_ON_DT, 'YYYYMMDDHH24MISS'))
    Thanks,
    Poojak

  • Error while activating data from new table of DSO to active table

    HI,
    while activating data from new table of DSO to active table i am getting
    error message as "Error occurred while deciding partition number".
    Plz any idea hoe to resolve this one.
    thanks & regards
    KPS MOORTHY

    Hi
    You are trying to update/upload the Records which are already been there in the DSO Active Data Table which has the partition.
    Try to see the Record Nos already been activated and update/upload with selective, if possible.
    You can trace the changes at Change log table for the same.
    Hope it helps
    Edited by: Aduri on Jan 21, 2009 10:38 AM

  • Error while inserting data into a table.

    Hi All,
      I created a table.While inserting data into the table i am getting an error.Its telling "Create data Processing Function Module".Can any one help me regarding this?
    Thanx in advance
    anirudh

    Hi Anirudh,
      Seems there is already an entry in the Table with the same Primary Key.
    INSERT Statement will give short dump if you try to insert data with same key.
    Why dont you use MODIFY statement to achieve the same.
    Reward points if this Helps.
    Manish

  • Error 65575: overrun error while reading data from NI 9871

    Hello everyone,
    I am new here. I configured NI 9871 in RS-485 mode to acquire data from a sensor at the data rate of 8 samples per second. While reading data from sensor, sometimes I got the error 65575: "An overrun error has occurred in the data being read. Ensure that the module Flow Control is configured correctly." I didn't understand the error, how can I fix it, what dose the flow control means.
    I will be very thankful if anybody help me to fix this error.
    Thanks
    Wasif

    Hello Wasif,
    The error you are getting presents itself when the hardware buffer is full and new data is coming in. Data buffer is 64 bytes. One thing to try is to clear the RX FIFO after every read. Please refer to the following link that talks about the method.
    http://zone.ni.com/reference/en-XX/help/370984R-01/lvaddon11/crio-9871/
    Hope this help you solve the issue.
    Regards
    Luis S
    Application Engineer
    National Instruments

  • Error while loading data through s614

    Hi
    When I am trying to load data from structure s614..I am getting the Below error..
    I have gone through Transact.RFC..there i found that for ALEREMOTE..function module 'RSAR_TRFC_DATA_RECEIVED'..time limit exceeded..
    and also found that job was finished in the source system..
    Can any one give some suggestion for this...and also i have  a doubt..my client is having table s614 and structure 2lis_01_s614..do i need to run set up to extract data from this table..if it is so,how should i approach......
    Thanks
    Teja
    Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the source system.

    Hi,
    I have a similary problem.
    Have you help me??

  • Error while Inserting data into flow table

    Hi All,
    I am very new to ODI, I am facing lot of problem in my 1st interface. So I have many questions here, please forgive me if it has irritated to you.
    ========================
    I am developing a simple Project to load a data from an input source file (csv) file into a staging table.
    My plan is to achieve this in 3 interfaces:
    1. Interface-1 : Load the data from an input source (csv) file into a staging table (say Stg_1)
    2. Interface-2 : Read the data from the staging table (stg_1) apply the business rules to it and copy the processed records into another staging table (say stg_2)
    3. Interface-3 : Copy the data from staging table (stg_2) into the target table (say Target) in the target database.
    Question-1 : Is this approach correct?
    ========================
    I don't have any key columns in the staging table (stg_1). When I tried to execute the Flow Control of this I got an error:
    Flow Control not possible if no Key is declared in your Target Datastore
    With one of the response (the response was - "FLOW control requires a KEY in the target table") in this Forum I have introduced a column called "Record_ID" and made it a Primary Key column into my staging table (stg_1) and my problem has been resolved.
    Question-2 : Is a Key column compulsary in the target table? I am working in BO Data Integrator, there is no such compulsion ... I am little confused.
    ========================
    Next, I have defined one Project level sequence. I have mapped the newly introduced key column Record_Id (Primary Key) with the Project level sequence. Now I am got another error of "CKM not selected".
    For this, I have inserted "Insert Check (CKM)" knowledge module in my Project. With this the above problem of "CKM not selected" has been resolved.
    Question-3 : When is this CKM knowledge module required?
    ========================
    After this, the flow/interface is failing while loading data into the intermediar ODI created flow table (I$)
    1 - Loading - SS_0 - Drop work table
    2 - Loading - SS_0 - Create work table
    3 - Loading - SS_0 - Load data
    5 - Integration - FTE Actual data to Staging table - Drop flow table
    6 - Integration - FTE Actual data to Staging table - Create flow table I$
    7 - Integration - FTE Actual data to Staging table - Delete target table
    8 - Integration - FTE Actual data to Staging table - Insert flow into I$ table
    The Error is at Step-8 above. When opened the "Execution" tab for this step I found the message - "Missing parameter Project_1.FTE_Actual_Data_seq_NEXTVAL RECORD_ID".
    Question-4 : What/why is this error? Did I made any mistake while creating a sequence?

    Everyone is new and starts somewhere. And the community is there to help you.
    1.) What is the idea of moving data from stg_1 and then to stg_2 ? Do you really need it for any other purpose other than move data from SourceFile to Target DB.
    Otherwise, its simple to move data from SourceFile -> Target Table
    2.) Does your Target table have a Key ?
    3.) CKM (Check KM) is required when you want to do constraint validation (Checking) on your data. You can define constraints (business rules) on the target table and Flow Control will check the data that is flowing from Source File to Target table using the CKM. All the records that donot satisfy the constraint will be added to E$ (Error table) and will not be added to the Target table.
    4.) Try to avoid ODI sequences. They are slow and arent scalable. Try to use Database sequence wherever possible. And use the DB sequence is target mapping as
    <%=odiRef.getObjectName( "L" , "MY_DB_Sequence_Row" , "D" )%>.nextval
    where MY_DB_Sequence_Row is the oracle sequence in the target schema.
    HTH

  • Problem while selecting data from external tables

    Hi All,
    I am facing a problem with external tabels. I have created an external table
    create table ext_org_table
    ( id varchar2(10)
    , name varchar2(100)
    ,id_parent varchar2(10)
    ,name_parent varchar2(100))
    organization external
    ( type oracle_loader
    default directory MYDIR
    access parameters
    records delimited by newline
         nologfile
         nobadfile
    fields terminated by ','
    missing field values are null
    location ('Orgdata.csv')
    reject limit unlimited;
    The problem is that when i give select * from ext_org_table , i get following error
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04040: file Orgdata.csv in MYDIR not found
    ORA-06512: at "SYS.ORACLE_LOADER", line 19
    ORA-06512: at line 1
    Now , i have file Orgdata.csv in the specified directory.
    I have created directory in oracle using create directory command.
    I have granted read,write permission on the directory to user.
    I have given all read,write and execute permissions to folder where my files resides.My file resides in /appl/home/kk
    Now i don't understand why it is giving error file in Mydir not found.
    Kindly suggest.
    Regards
    Krish

    Hi,
    You are obviously doing someting wrong.
    May be the direcotry is not present or may be the file is not rpesent in the directory.
    I just tried ths
    Connected to Oracle Database 10g Enterprise Edition Release 10.2.0.1.0
    Connected as SYS
    create directory mydir as 'c:\csv';
    grant read,write on directory mydir to scott;From Scott
    create table ext_org_table
    ( id varchar2(10)
    , name varchar2(100)
    ,id_parent varchar2(10)
    ,name_parent varchar2(100))
    organization external
    ( type oracle_loader
    default directory MYDIR
    access parameters
    records delimited by newline
    nologfile
    nobadfile
    fields terminated by ','
    missing field values are null
    location ('Orgdata.csv')
    reject limit unlimited;
    select * from ext_org_table;Ofcourse i made up some dummy data to test and its all OK.
    again,
    You may want to check if
    1) If the directory exists (select * from all_directories where directory_name='MYDIR');
    2) If the user has the read/write permissions on that directory
    3) If the file orgdata exists in that directory.
    Regards,
    Bhushan

  • Error while downloading data from internal table into XML file

    hi all,
    i developed a program to download data from into internal table to xml file like this.
    tables: mara.
    parameters: p_matnr like mara-matnr.
    data: begin of itab_mara occurs 0,
                matnr like mara-matnr,
                ernam like mara-ernam,
                aenam like mara-aenam,
                vpsta like mara-vpsta,
          end of itab_mara.
    data: lv_field_seperator type c,     " value 'X',
          lv_xml_doc_name(30) type c,    " string value ‘my xml file’,
          lv_result type i.
          lv_field_seperator = 'x'.
          lv_xml_doc_name = 'my xml file'.
    types: begin of truxs_xml_line,
              data(256) type x,
          end of truxs_xml_line.
    types:truxs_xml_table type table of truxs_xml_line.
    data:lv_tab_converted_data type truxs_xml_line,
         lt_tab_converted_data type truxs_xml_table.
    data: lv_xml_file type rlgrap-filename value 'c:\simp.xml'.
    select matnr ernam aenam vpsta from mara into table itab_mara up to 5
           rows where matnr = p_matnr.
    CALL FUNCTION 'SAP_CONVERT_TO_XML_FORMAT'
    EXPORTING
       I_FIELD_SEPERATOR          = lv_field_seperator
      I_LINE_HEADER              =
      I_FILENAME                 =
      I_APPL_KEEP                = ' '
       I_XML_DOC_NAME             = lv_xml_doc_name
    IMPORTING
       PE_BIN_FILESIZE            = lv_result
      TABLES
        I_TAB_SAP_DATA             = itab_mara
    CHANGING
       I_TAB_CONVERTED_DATA       = lt_tab_converted_data
    EXCEPTIONS
      CONVERSION_FAILED          = 1
      OTHERS                     = 2
    IF SY-SUBRC <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    open dataset lv_xml_file for output in binary mode.
    loop at lt_tab_converted_data into lv_tab_converted_data.
    transfer lv_tab_converted_data to lv_xml_file.
    endloop.
    close dataset lv_xml_file.
    this program is syntactically correct and getting executed, but when i open the target xml file it is showing the following error.
    The XML page cannot be displayed
    Cannot view XML input using style sheet. Please correct the error and then click the Refresh button, or try again later.
    XML document must have a top level element. Error processing resource 'file:///C:/simp.xml'.
    will anyone show me the possible solution to rectify this error
    thanks and regards,
    anil.

    Hi,
    Here is a small sample program to convert data in an internal table into XML format and display it.
    DATA: itab  TYPE TABLE OF spfli,
          l_xml TYPE REF TO cl_xml_document.
    * Read data into a ITAB
    SELECT * FROM spfli INTO TABLE itab.
    * Create the XML Object
    CREATE OBJECT l_xml.
    * Convert data in ITAB to XML
    CALL METHOD l_xml->create_with_data( name = 'Test1'
                                         dataobject = t_goal[] ).
    * Display XML Document
    CALL METHOD l_xml->display.
    Here are some other sample SAP programs to handle XML in ABAP:
    BCCIIXMLT1, BCCIIXMLT2, and BCCIIXMLT3.
    Hope this helps,
    Sumant.

Maybe you are looking for

  • My Book appears greyed out in the Books section

    A purchased book, sometime back in September but I seem to have deleted the receipt e-mail, is now "greyed out" and seems to be a corrupted file. Can I recover somehow? I tried going into Purchased but it doesn't show there but I am sure I did purcha

  • Date chooser day and date not selectable

    Hello, Anyone knows how to make the day and date not selectable in the datefield or date chooser component??? It is possible to make any one day not available, but what i need is only a single day not selectable.. For example, i can make every tuesda

  • SRM 7 - Centralised Contract Management SAP XI inbound/outbound interfaces

    Hi there, We are busy implementing SRM 7 and I would like to get "centralised contract management" working between SRM 7 and the ECC backend/s (via SAP PI). Could you please let me know what the correct SAP PI CONTENT to load is? And what the correct

  • HT4972 I'm getting and error message while trying to update to ios6, what should I do?

    I'm getting and error message when I'm trying to update to ios6? What can I do?

  • Time setting loses time

    A long while ago I set the time on my iPod Classic 160GB. For reasons too complicated to explain, the iPod was not used for several months. It had a fully charged battery before it was left idle. Although we've had the iPod for a while, it has not be