Regarding Reading External Table into ADC

Hi ,
I was trying to read external table inmto ADC using Plans(tech Notes). I have completed the steps sucessfully.ADC has been generated without flaw. IF i am not wrong , one can use plans to update the ADC Dataobject from the database tables.
The problem i face is , if the database is updated , this change is not reflecting in the ADC.I followed the instruction given in Technotes,after plan creation, have changed plan settings,done request plan setting check, and restarted Plan moitoring service.I would really appreciate if anyone could let me know what should be done in order to get the updates in the ADC.
Expecting a reply soon.
Regards,
Lathika.

Hi Sanjay,
Many thanks for your reply.
Yes i have set the plan to run contuinuously .I have used BAM update Sink.YEs the source date is different.There is no error in the log....But still the change is not getting reflected in ADC object...
Is there any other solution to resolve this problem??
Regards,
Lathika

Similar Messages

  • Need to merge a csv file using external tables into a main table

    Hi,
    I have a csv file which contains the date(with time stamp), column1(number),column2(number), column3 (number). I am using external tables concept to load the data froom csv to this external table and then merging into the main table. Problem here is : the csv file is a system generated file and nothing can be edited under it. our aim is to automate this process of loading data from csv to the table. In this csv the date time stamp is not in the proper format.I mean the date is not visible and only minutes and seconds are visible.By changing the format in csv manually this can be overcome.but we donot need any manual intervention.
    how can i overcome this problem ?? please help mee...
    Excels data looks like:
    (PDH-TSV 4.0) (India Standard Time)(-330)     \\DISAPPSER01\Processor(_Total)\% Privileged Time     \\DISAPPSER01\Processor(_Total)\% Processor Time     \\DISAPPSER01\Web Service(_Total)\Current Connections
    56:59.0               47
    57:09.0     0.72379582     4.204561281     46
    57:19.0     0.916548537     4.006179927     44
    57:29.0     0.663034771     3.674662541     43
    57:39.0     0.750789844     4.093933999     42
    57:49.0     0.721538487     2.650858026     40
    57:59.0     0.594781604     3.333393703     40

    please format your sample data giving header to the column so that we can make sense out of the values, also since the minutes and seconds are only given, what is the date to be considered for records to be moved to the master table, sysdate or will the date be passed as a parameter?

  • How to register External table into Oracle APPS?

    hi
    can we register the External table in Oracle apps like SQL Loader?
    So if yes, please share the steps.
    Thx

    781261 wrote:
    okk..
    But i mean , we create Executable with Method type SQL Loader, to register in Oracle APPS.
    SO like what would be Executable Method for External Table?
    or can we call the External table in Procedure?
    ThanxI don't know oracle APPS (and this isn't the APPS forum), but an external table is just like another table. It's created once and then used by queries to fetch data. You would reference it in your applications just the same as you would any other table.

  • Regarding reading multiple records into a single internal table..

    hi experts,
    Need your help, i have a requirement like this.
    i will have a input file like this
    D 123   Suresh    12/01/2008
    E ven   sha         5432
    E ven   sha       5432
    D 153   Sachin    11/01/2008
    E ven   sha      5432
    Now all i need to consider from D to E  (i.e., before next D) as a single record and i need to preare a separate excel,
    So that E records can be max 9. So we cant expect this will be always 2 or 3 or 4.
    So how can i do this upload and process..
    give some idea.
    Thanks,
    Suresh

    Hi,
    Once you transfer the values from input file to internal table.
    loop through all records of internal table .
    check the first letter using offset
    e.g  if Itab-field+0(1)  = 'D'
           elseif  Itab-field+0(1)  = 'E'.
           endif.
         Populate the work area of internal table until you find the next 'D' once you find 'D' again ...append the work area to internal table......clear work area.
    Regards,
    Rahul

  • Reading internal table into File Adapter

    Hi,
    I would like to read an internal table from RFC function (sender), and write all the rows into the file by using File Adapter. Here is the scenerio;
    RFC Adapter -> XI -> File adapter
    Internal table content, below;
    FLD1  FLD2   FLD3
    1        A        B
    2        X        Y
    I expect the result, below;
    <?xml version="1.0" encoding="UTF8" ?>
    <rfc:Z_RFC xmlns:rfc="urn:sapcom:document:sap:rfc:functions">
    <ITAB_ZSUBS001>
      <item>
       <FLD1>1</FLD1>
       <FLD2>A</FLD2>
       <FLD3>B</FLD3>
      </item>
      <item>
       <FLD1>2</FLD1>
       <FLD2>X</FLD2>
       <FLD3>Y</FLD3>
      </item>
    </ITAB_ZSUBS001>
    </rfc:Z_RFC>
    I see all the records in payload, but cannot write into file. Anybody have any suggestion? Thank you.
    Regards,
    Orkun Gedik
    Message was edited by:
            Orkun GEDIK

    Hi,
    Thank you for the feedback. From now on I am getting the error, below;
    Error> occured [1] >Thu Apr 05 09:24:21,046<   
    RfcException:
        message: Commit fault: com.sap.aii.af.rfc.afcommunication.RfcChannelMismatchException: Wrong Sender Agreement:The Sender agreement does not have channel Rfc_Fiyat_Sender configured for the functionmodule Z_RFC
        Return code: RFC_CLOSED(6)
        error group: 108
        key: RFC_ERROR_INTERNAL
    Please see the last configuration, below;
    Message mappings:
    RFC Message:
    Messages 1..1
    Message1 1..1
    Z_RFC 1..1
    ITAB_ZSUBS001 1..1
    item 0..unbounded
    TXTKOD 0..1
    TXTAD 0..1
    TXTADDR 0..1
    Message Type: Malzeme
    Messages 1..1
    Message1 1..1
    Malzeme 0..unbounded
    TXTKOD 1..1
    TXTAD 1..1
    TXTADDR 1..1
    Interface mappings:
    Source Interface: Z_RFC Occurrence 1
    Target Interface: MI_MAlzeme_In_Asyn Occurrence 0..unbounded
    I bound the items below;
    item -> Malzeme
    TXTKOD -> TXTKOD
    TXTAD -> TXTAD
    TXTADDR -> TXTADDR
    Also,I refreshed the adapter cache and full cache by using SXI_CACHE and restart the RFC adapter thorugh Visual Admin in order to solve the problem, but it couldn't be a solution for the problem. Do you have any suggestion about it?
    Regards
    Message was edited by:
            Orkun GEDIK

  • Error while importing external table

    Hello,
    I am getting the following error message when trying to import an external table:
    SQL Exception
    SQL Exception
    Repository Error:SQL Exception..
    Class Name: CacheMediator.
    Method Name: getDDEntryFromDB.
    Repository Error Message: ORA-22992: cannot use LOB locators selected from remote tables
    I am using the "import..."- function of "external tables" in warehouse target module. The external table is working correctly when accessing it via sqlplus.
    We are using owb version 10.2.0.1.
    Any help would be appreciated.
    Best regards
    Alexander

    My friends, this a documented BUG into Metalink, but there is no solution.
    I used a simple workaround:
    1. I created the external table into the database;
    2. I created the external table into the OWB repository;
    3. I DIDN'T DEPLOY THE EXTERNAL TABLE TO THE DATABASE;
    4. I used the external table into my mappings with no problems;
    5. Finishing, I deployed my mappings to the database without any errors.
    Hope to help!,
    Gustavo.

  • How to ''give'' error for this case of an EXTERNAL TABLE?

    Our external table routine works fine:
    -- We have a csv file with 2 cols.
    -- When we create the table referring the csv it works fine.
    -- Even if the csv has more the 2 cols, the ET command only takes the 2 cols and it works fine.
    -- Now, users are saying that if the csv has more than 2 cols, the ET command should give an error
    I went through the command but cannot find any clause which will do this.
    Is there any other way or workaround?
    CREATE TABLE <table_name> (
    <column_definitions>)
    ORGANIZATION EXTERNAL
    (TYPE oracle_loader
    DEFAULT DIRECTORY <oracle_directory_object_name>
    ACCESS PARAMETERS (
    RECORDS DELIMITED BY newline
    BADFILE <file_name>
    DISCARDFILE <file_name>
    LOGFILE <file_name>
    [READSIZE <bytes>]
    [SKIP <number_of_rows>
    FIELDS TERMINATED BY '<terminator>'
    REJECT ROWS WITH ALL NULL FIELDS
    MISSING FIELD VALUES ARE NULL
    (<column_name_list>))\
    LOCATION ('<file_name>'))
    [PARALLEL]
    REJECT LIMIT <UNLIMITED | integer>;
    Is it possible to use the READSIZE?
    Edited by: Channa on Sep 23, 2010 2:28 AM

    -- Now, users are saying that if the csv has more than 2 cols, the ET command should give an error
    I went through the command but cannot find any clause which will do this.
    Is there any other way or workaround?I looked at Serverprocess' sql*loader script and did not see how that would answer your question - how to raise an error if the file has more than 2 columns. If I missed something can Serverprocess explain?
    I can't think of a direct way to do this with your external table either, but there may be indirect ways. Some brainstorming ideas of perhaps dubious usefulness follow.
    Placing a view over the external table can limit results to the first two columns but won't raise an error.
    A pipelined function can read the external table, check for data where there shouldn't be any, and raise an exception when you find data in columns where there should not be any.
    Similarly, you could ditch the external table and use utl_file to read the file, manually parsing and checking the data. LOTS more work but more control on your end. External tables are much easer to use :(
    Or, first load the external table into a work table before the "real" select. Check the work table for the offending data programatically and raise an error if data is where it should not be. You could keep the existing external table and not have to do a lot of recoding.
    Or, also load the data into an otherwise unneeded global temporary table first. Use a trigger on the load to look for the unwanted data and raise an error if offending data is there
    These ideas are boiling down to variations on validating the data before you use it.
    Good luck!

  • External Table Application

    Trying to install the External_Table_Simple_0.9 application but I am receiving errors.
    I installed the application but it could not see the flat file, so I de-installed and on re-install I got these errors.
    report error:
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04063: unable to open log file EXT_EMPLOYEES_EXTERNAL_5476_2344.log
    OS error The data is invalid.
    Has anybody experienced this.
    Gus

    Gus,
    Copy and paste will not do the work every time.
    I copied and pasted the package definition you posted and it won't compile. One
    of the declarations is not properly quoted:
    x :=
             q'! comment on table ext_employees is 'Data loaded: !'
          || TO_CHAR (SYSDATE, 'FMDay, Mon ddth, hh:mi am');
       x := x || q'! by !' || NVL (v ('APP_USER'), USER) || q'!' !';
       EXECUTE IMMEDIATE x;
    ...Other than that, you also had an inproper positioning specification here:
    FIRST_NAME position(8:22),
    last_name position(23:34),
    where you didn't consider whitspace between the two columns.
    I assume the file you use is of fixed column length. I used the information
    you pasted here and modified it accordingly to get a fixed lenght of the columns:
    EMP_ID FIRST_NAME LAST_NAME EMAIL    PHONE_NUMBER HIRE_DATE  JOB_ID  SALARY COMM_PCT MGR_ID DEPT_ID
    100    Steven     King      SKING    515.123.4567 06/17/1987 AD_PRES 24000           90
    101    Neena      Kochhar   NKOCHHAR 515.123.4568 09/21/1989 AD_VP   17000           100    90
    102    Lex        De Haan   LDEHAAN  515.123.4569 01/13/1993 AD_VP   17000           100    90
    103    Al         Hunold    AHUNOLD  590.423.4567 01/03/1990 IT_PROG 9000            102    60Saved it as employees.txt into the directory DATA_LOAD.
    Running the following script will create an external table:
    CREATE TABLE ext_employees_external
      employee_id     VARCHAR(20),
      first_name      VARCHAR(20),
      last_name       VARCHAR(20),
      email           VARCHAR(20),
      phone_number    VARCHAR(20),
      hire_date       VARCHAR(20),
      job_id          VARCHAR(20),
      salary          VARCHAR(20),
      commission_pct  VARCHAR(20),
      manager_id      VARCHAR(20),
      department_id   VARCHAR(20)
    ORGANIZATION EXTERNAL
      (  TYPE oracle_loader
         DEFAULT DIRECTORY data_load
         ACCESS PARAMETERS
           ( RECORDS DELIMITED BY NEWLINE
    SKIP 2
    FIELDS MISSING FIELD VALUES ARE NULL(
    employee_id POSITION(001:006),
    first_name POSITION(008:0017),
    last_name POSITION(019:027),
    email POSITION(029:036),
    phone_number POSITION(038:049),
    hire_date POSITION(051:060) ,
    job_id POSITION(062:068),
    salary POSITION(070:075),
    commission_pct POSITION(077:083),
    manager_id POSITION(086:092),
    department_id POSITION(093:100)
         LOCATION (data_load:'employees.txt')
    REJECT LIMIT 10
    NOPARALLEL
    NOMONITORING;This table is reading the data properly. You may now include this in your procedure.
    You certanly need to have a look at your .txt file and exactly determie the column
    widths. I was using varchar2 in the column definiton of the extenal table. Since you are
    loading the external table into your "real" table, you may want to perform data
    conversion at that point.
    The directory I specified in this script is DATA_LOAD. Your directory may be different.
    Denes Kubicek

  • SQL Loader versus External Table

    If anyone has worked on external tables please let me know your views.

    user637544 wrote:
    for sqlldr i follow the following approach
    1. truncate table
    2. call sqlldr and load records in staging table
    3. check for bad records
    how do i tuncate, call external table and check bad records for external table.As part of the SQL*Loader control file you can tell it to truncate the table as part of the process before it loads in the data.
    The key differences between SQL*Loader and External Tables are:
    1. SQL*Loader is an external utility run from the o/s command line, whereas External tables are usable from within SQL and PL/SQL.
    2. SQL*Loader can only load data up into the database, whereas External tables can allow you to read files and write files (10g onwards)
    3. SQL*Loader has limited control on skipping rows etc. whereas External tables can be queried using SQL and has all the flexibility that SQL can offer.
    4. SQL*Loader requires connection information to the database, whereas External tables are already part of the database
    5. SQL*Loader can only change the source filename through dynamic o/s level scripts, whereas External tables can have the filename changed using an ALTER TABLE ... LOCATION ... command.
    If you want to truncate your staging table before loading it with data from the external table, then you will just have to do that as a seperate step, because the external table is not associated with other tables. Simply truncate the other table and then query the data from the external table into the staging table.
    External tables offer a lot more flexibility than using SQL*Loader.

  • Sqlldr flat file to external table using shell scripts

    Hi,
    Has anyone done this before? Please give me a hand.
    Thanks!

    Thanks Justin.
    When do I need to create the external table EMP_STAGING ?
    These are my steps so far:
    - shell script to crate the flat file (but I need to change the table name to EMP_STAGING)
    - use a script to call sqlldr to load the flat file into the external table
    - then the script will call the MERGE sql script to merge the data from the external table into the database table
    Am I on the right track?
    In which stage should I create and drop the external table?
    Thanks!

  • Reading internal tables

    Hi all
    i am using internal table itab1,
    i want to add the similar contents of itab1 from itab2 into itab3
    how to do it?
    thanks
    raghvendra bhanap

    Hi,
    This is a documentation regarding reading internal tables.
    Hope this will be useful...
    When the size of the record or the number of records in the internal table is large, doing a linear search is time consuming. It is always a good practice to search a record by binary search (Always Sort the table before doing a binary search). The difference is felt especially in the production environment where the live data is usually huge. As of release 4.0 there are new types of internal tables like SORTED and HASHED which can be effectively used to further reduce the search time and processing time.
    e.g.     Do not use the following statement:-
    Select matnr from mara
    Into table i_mara
    Where matnr in s_matnr.
         Select matnr werks from marc
              Into table i_marc
              For all entries in i_mara
              Where matnr eq i_mara-matnr.
         Loop at I_mara.
              Read table i_marc with key matnr = I_mara-matnr.
         Endloop.
    Instead use the following statement:-
         Select matnr from mara
              Into table i_mara
              Where matnr in s_matnr.
                                       Select matnr werks from marc
                                            Into table i_marc
                                            For all entries in i_mara
                                           Where matnr eq i_mara-matnr.
                                       Sort I_marc by matnr.
                                       Loop at I_mara.
                                            Read table i_marc with key
    matnr = I_mara-matnr
    binary search.
                                       Endloop.
    It is a good practice to search records from internal tables using a binary search. But extreme caution needs to be applied as it may either increase the time or may cause run time termination if it is not sorted.
    Always the sort the internal table by the required keys before performing a binary search.                                                                               
    e.g.     Do not use the following statement:-
    Select matnr from mara
    Into table i_mara
    Where matnr in s_matnr.
         Select matnr werks from marc
              Into table i_marc
              For all entries in i_mara
              Where matnr eq i_mara-matnr.
         Loop at I_mara.
              Read table i_marc with key matnr = I_mara-matnr binary search.
         Endloop.
    Instead use the following statement:-
         Select matnr from mara
              Into table i_mara
              Where matnr in s_matnr.
                                       Select matnr werks from marc
                                            Into table i_marc
                                            For all entries in i_mara
                                           Where matnr eq i_mara-matnr.
                                       Sort I_marc by matnr.
                                       Loop at I_mara.
                                            Read table i_marc with key
    matnr = I_mara-matnr
    binary search.
                                       Endloop.
    It is a general practice to use           Read table <itab>…     This statement populates all the values of the structure in the workarea.
    The effect is many fold:-
    •     It increases the time to retrieve data from internal table
    •     There is large amount of unused data in work area
    •     It increases the processing time from work area later
    It is always a good practice to retrieve only the required fields. Always use the syntax      Read table <itab> transporting f1 f2  …  FN …          If just a check is being performed for existence of a record use      Read table <itab> transporting no fields …
    e.g.     Do not use the following statement:-
    data: i_vbak like vbak occurs 0 with header line.
    data: i_vbap like vbap occurs 0 with header line.
    Loop at i_vbak.
         read table i_vbap with key
    vbeln = i_vbak-vbeln binary search.
         If sy-subrc = 0 and i_vbap-posnr = ‘00010’.
         endif.
    Endloop.
    Instead use the following statement:-
    data: i_vbak like vbak occurs 0 with header line.
    data: i_vbap like vbap occurs 0 with header line.
    Loop at i_vbak.
                                  read table i_vbap transporting posnr with key
    vbeln = i_vbak-vbeln binary search.
                                  If sy-subrc = 0 and i_vbap-posnr = ‘00010’.
                                  endif.
    Endloop.
    There are many ways in which a select statement can be optimized. Effective use of primary and secondary indexes is one of them. Very little attention is paid especially to the secondary indexes. The following points should be noted before writing a select statement:-
    •     Always use the fields in the where clause in the same order as the keys in the database table
    •     Define the secondary indexes in the database for the fields which are most frequently used in the programs
    •     Always try to use the best possible secondary index in the where clause if it exists
    •     Do not have many secondary indexes defined for a table
    •     Use as many keys both primary and secondary as possible to optimize data retrieval
    As of release 4.5 it is now possible to define the secondary index in the where clause using %_HINT.
    e.g.     Do not use the following statement:-
         Assuming a secondary index is defined on the field vkorg in table vbak
         Select vbeln vkorg from vbak
              Into table i_vbak
              Where vbeln in s_vbeln.
         Loop at i_vbak.
              Case i_vbak-vkorg.
              When ‘IJI1’.
              When ‘IJI2’.
              Endcase.
         Endloop.
    Instead use the following statement:-
                                       Select vbeln vkorg from vbak
                                            Into table i_vbak
                                            Where vbeln in s_vbeln and
                                                 Vkorg in (‘IJI1’,’IJI2’).
                                       Loop at i_vbak.
                                            Case i_vbak-vkorg.
                                            When ‘IJI1’.
                                            When ‘IJI2’.
                                            Endcase.
                                       Endloop.

  • Join large external external Tables (best practice ?)

    Hi there,
    I have three external tables in a master-detail relation: table A (10000000 rows) is master of table B (20000000 rows), Table B is master of table C (100000000 rows). Can you tell my the best way:
    - directly join the external tables, or
    - copy the external tables into tables, create an index, and join them
    What is more efficient, and why ?
    Thanks for your help and ideas.
    Gerhard.

    In general, if the joins you are doing can benefit from indexes, you will want to copy the data to database tables. If the joins will end up doing full table scans anyway, it will matter far less.
    For data integrity, you will likely also want to be able to enforce foreign key constraints.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Reject the footer using the external table

    Hi,
    I have a flat file with fixed length which have a header and a footer.
    My file is something like this:
    HADF.TXT0309
    D12345ABCD
    D22345ABCD
    FOOTERHJ
    I want to create an external table based on that file, but
    I don't want to have the header and the footer in my table. To eliminate the header I used skip 1, but I don't know how to eliminate the footer.
    Any example and suggestions will be appreciated
    Thank you for your time and consideration
    Catalin

    Hi,
    This problem may be due to several reasons. I am aware of few reasons.
    1)Have you deployed the external table first?
    2) If you do not have your data base client and server running in the same machine, you should place the csv file in the database server machine's 'c:\CSV' folder in order to create the external table through database. Then do a select count(*) statement.
    3)Another reason may be as gerardnico said the file name you refered may be wrong.
    I don't know your requirement. If you could create the external table succefully and if you get value for the select count(*) from <external_table_name>, then try to import the external table into the Design Center and map it with the Table you need.
    If you are doing it purely with OWB then,
    Do you have the file Export_WithHeaders.csv in the Server machine's 'c:\CSV_FILE' folder?
    Because while importing the metadata of the CSV file the OWB will point to your local machine's 'c:\CSV_FILE' that is why your Validation and Deployment is success without errors.
    But while executing the map it will take the data from Server machine. It will search for the Location 'c:\CSV_FILE' in the server machine and will look for the file Export_WithHeaders.csv there. So create athe same folder setup which you have in your client machine and run this again.
    Try this if you do not get any better answers in this thread

  • Stuck on using an External Table in OWB 11.

    I am using Oracle 11, and OWB 11.
    I am trying to get some information from a CSV file, into an Oracle table. Here is what I have currently done:
    1. I have setup a location called CSV_FILE and this is pointing to my local drive in a folder called c:\CSV
    2. I have added the file to the Files tab, imported the metadata and the file is called Export_WithHeaders_CSV.csv
    3. I have setup a new external table called CSV and pointed it to the above File and set the default llocation to CSV_FILE
    4. I then right clicked on the external table and created the Data File pointing it to the CSV_FILE, and filling in the Data File Name with: Export_WithHeaders_CSV.csv
    I am trying to do 2 things, the first is to just view the data in the external table through Design Center, and the 2nd is too add it to an OWB map to import into an Oracle Database. When I try to view the data (Right click on CSV, and select Data) I get the following error:
    ORA-29913: error in exectuing
    ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04040: file
    Export_WithHeaders.csv in
    CSV_FILE not found
    I am able to generate and validate the external table without an erro, but I am not able to see any data, and I know the file contains close to 120 records?
    When I bring the external table into a simple OWB map, just mapping it directly to a table, the map validates fine, but when I run the map, the test data shows a row_count of 0? I have been beating my head against the wall for a good 2.5 days now, and any help would be greatly appreciated....

    Hi,
    This problem may be due to several reasons. I am aware of few reasons.
    1)Have you deployed the external table first?
    2) If you do not have your data base client and server running in the same machine, you should place the csv file in the database server machine's 'c:\CSV' folder in order to create the external table through database. Then do a select count(*) statement.
    3)Another reason may be as gerardnico said the file name you refered may be wrong.
    I don't know your requirement. If you could create the external table succefully and if you get value for the select count(*) from <external_table_name>, then try to import the external table into the Design Center and map it with the Table you need.
    If you are doing it purely with OWB then,
    Do you have the file Export_WithHeaders.csv in the Server machine's 'c:\CSV_FILE' folder?
    Because while importing the metadata of the CSV file the OWB will point to your local machine's 'c:\CSV_FILE' that is why your Validation and Deployment is success without errors.
    But while executing the map it will take the data from Server machine. It will search for the Location 'c:\CSV_FILE' in the server machine and will look for the file Export_WithHeaders.csv there. So create athe same folder setup which you have in your client machine and run this again.
    Try this if you do not get any better answers in this thread

  • Error loading local CSV file into external table

    Hello,
    I am trying to load a CSV file located on my C:\ drive on a WIndows XP system into an 'external table'. Everyting used to work correctly when using Oracle XE (iinstalled also locally on my WIndows system).
    However, once I am trynig to load the same file into a Oracle 11g R2 database on UNIX, I get the following errr:
    ORA-29913: Error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    error opening file ...\...\...nnnnn.log
    Please let me know if I can achieve the same functionality I had with Oracle XP.
    (Note: I cannot use SQL*Loader approach, I am invoking Oracle stored procedures from VB.Net that are attempting to load data into external tables).
    Regards,
    M.R.

    user7047382 wrote:
    Hello,
    I am trying to load a CSV file located on my C:\ drive on a WIndows XP system into an 'external table'. Everyting used to work correctly when using Oracle XE (iinstalled also locally on my WIndows system).
    However, once I am trynig to load the same file into a Oracle 11g R2 database on UNIX, I get the following errr:
    ORA-29913: Error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    error opening file ...\...\...nnnnn.log
    Please let me know if I can achieve the same functionality I had with Oracle XP.
    (Note: I cannot use SQL*Loader approach, I am invoking Oracle stored procedures from VB.Net that are attempting to load data into external tables).
    Regards,
    M.R.So your database is on a unix box, but the file is on your Windows desktop machine? So .... how is it you are making your file (on your desktop) visible to your database (on a unix box)???????

Maybe you are looking for

  • Scales definition in Scheduling agreement

    Hi, In a scheduling agreement, you can maintain condition with scales. When you do so, at goods receipt time, the system does not look at how many you just received to define the scale to use but look at the target quantity on your scheduling agreeme

  • How to send multiple attachemnt with a Email

    Hello experts, I have problem with sending multiple attachement with a Email . I have used program BCS_EXAMPLE_5. it is not for multiple attachment. Please help me. Thank you.

  • How do i turn off secure browsing

    how do i turn off secure browsing in firefox 7? some sites give me an error and wont show the full content because its on.

  • Seperate tax accounts by business area

    Hi All,    I had created input n outut taxes and assigned clearing accounts for them, how to link these to a particular business area 'X' so that i can get these balance sheet accounts selected when i use  business area 'X' in the report. Thanx, srav

  • SQL server 2005 and MDM components

    Question: 1) Could we make a repository with relevant tables in SQL server2005, and then import to MDM Console? 2) Could we see the tables of one repository in SQL server 2005.e.g. In MDM console, one repository created with the name Test001, Test 00