Loading 60 GB with sqlldr

Has anybody done something like this? I am wondering if sqlldr is a good way to import a 60GB csv file. Please comment about your experience.

My question here is, did you be able to store 60gb on csv file?
Regards
Vignesh Bala

Similar Messages

  • Error in Loading Data with SQLLDR in Oracle 10G

    Hello,
    Can any one suggest whats the problem in the below mentioned Control file used for loading data through SQL*LOADER
    LOAD DATA
    INFILE 'D:\test\temt.txt'
    BADFILE 'test.bad'
    DISCARDFILE 'test.dsc'
    INTO TABLE "TEST"
    INSERT
    (SRNO INTEGER(7),
    PROD_ID INTEGER(10),
    PROMO_ID INTEGER(10),
    CHANNEL_ID INTEGER(10),
    UNIT_COST INTEGER(10),
    UNIT_PRICE INTEGER(10)
    I am trying to load data in SCOTT schema as user scott.
    Why do i get such an error, please refer the attach Log file.
    SQL*Loader: Release 10.2.0.1.0 - Production on Fri Mar 20 14:43:35 2009
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Control File: D:\test\temt.ctl
    Data File: D:\test\temt.txt
    Bad File: test.bad
    Discard File: test.dsc
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array: 64 rows, maximum of 256000 bytes
    Continuation: none specified
    Path used: Conventional
    Table "TEST", loaded from every logical record.
    Insert option in effect for this table: INSERT
    Column Name Position Len Term Encl Datatype
    SRNO FIRST 7 INTEGER
    PROD_ID NEXT 10 INTEGER
    PROMO_ID NEXT 10 INTEGER
    CHANNEL_ID NEXT 10 INTEGER
    UNIT_COST NEXT 10 INTEGER
    UNIT_PRICE NEXT 10 INTEGER
    Record 1: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 2: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 3: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 4: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 5: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 6: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 7: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 8: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 9: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 10: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 11: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 12: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 13: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 14: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 15: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 16: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 17: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 18: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 19: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 20: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 21: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 22: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 23: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 24: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 25: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 26: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 27: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 28: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 29: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 30: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 31: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 32: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 33: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 34: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 35: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 36: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 37: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 38: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 39: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 40: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 41: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 42: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 43: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 44: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 45: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 46: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 47: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 48: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 49: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 50: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    Record 51: Rejected - Error on table "TEST".
    ORA-01460: unimplemented or unreasonable conversion requested
    MAXIMUM ERROR COUNT EXCEEDED - Above statistics reflect partial run.
    Table "TEST":
    0 Rows successfully loaded.
    51 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 3648 bytes(64 rows)
    Read buffer bytes: 1048576
    Total logical records skipped: 0
    Total logical records read: 64
    Total logical records rejected: 51
    Total logical records discarded: 0
    Run began on Fri Mar 20 14:43:35 2009
    Run ended on Fri Mar 20 14:43:43 2009
    Elapsed time was: 00:00:07.98
    CPU time was: 00:00:00.28
    Below is the method of using SQLLDR and table details,
    SQL> desc test
    Name Null? Type
    SRNO NUMBER(7)
    PROD_ID NUMBER(10)
    PROMO_ID NUMBER(10)
    CHANNEL_ID NUMBER(10)
    UNIT_COST NUMBER(10)
    UNIT_PRICE NUMBER(10)
    Method for using sqlldr is :
    cmd promt,
    d:\> sqlldr scott/tiger
    control = D:\test\temt.ctl
    SQL*Loader: Release 10.2.0.1.0 - Production on Fri Mar 20 15:55:50 2009
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Commit point reached - logical record count 64
    I even tried few examples,
    Which of the below mentioned Control file make sense,
    --1
    LOAD DATA
    INFILE 'D:\test\temt.txt'
    BADFILE 'test.bad'
    DISCARDFILE 'test.dsc'
    INTO TABLE "TEST"
    INSERT
    FIELD TERMINATED BY (,)
    (SRNO INTEGER(7),
    PROD_ID INTEGER(10),
    PROMO_ID INTEGER(10),
    CHANNEL_ID INTEGER(10),
    UNIT_COST INTEGER(10),
    UNIT_PRICE INTEGER(10)
    --2
    LOAD DATA
    INFILE 'D:\test\temt.txt'
    BADFILE 'test.bad'
    DISCARDFILE 'test.dsc'
    INTO TABLE "TEST"
    INSERT
    FIELD TERMINATED BY (,) optionally enclosed by '"'
    (SRNO INTEGER(7),
    PROD_ID INTEGER(10),
    PROMO_ID INTEGER(10),
    CHANNEL_ID INTEGER(10),
    UNIT_COST INTEGER(10),
    UNIT_PRICE INTEGER(10)
    *For code--1 i get below mentioned error..*
    D:\>sqlldr scott/tiger
    control = D:\test\temt.ctl
    SQL*Loader: Release 10.2.0.1.0 - Production on Fri Mar 20 16:36:00 2009
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    SQL*Loader-350: Syntax error at line 8.
    Expecting "(", found "FIELD".
    FIELD TERMINATED BY (,)
    ^
    *And for code--2 i get the below error,*
    D:\>sqlldr scott/tiger
    control = D:\test\temt.ctl
    SQL*Loader: Release 10.2.0.1.0 - Production on Fri Mar 20 16:39:22 2009
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    SQL*Loader-350: Syntax error at line 8.
    Expecting "(", found "FIELD".
    FIELD TERMINATED BY (,) optionally enclosed by '"'
    ^
    ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

    below is the data i am trying to load through sqlldr
    1,14,999,3,89098,111287
    2,14,999,2,88645,110899
    3,14,999,2,90418,117623
    4,14,999,3,89272,115999
    5,14,999,4,86364,117623
    6,15,999,3,87522,101399
    7,15,999,4,84671,99999
    8,15,999,4,84671,99999
    9,15,999,4,86364,101399
    10,15,999,4,88735,100399
    11,15,999,2,90418,101399
    12,15,999,3,89272,101399
    13,15,999,2,90418,101399
    14,17,999,2,9694,155099
    15,17,999,3,97618,155099
    16,17,999,3,97618,155099
    17,18,999,3,11333,169783
    18,18,999,2,11479,163279
    19,18,999,3,11333,163279
    20,18,999,3,115141,173504
    21,18,999,2,117086,165565
    22,18,999,2,116856,17532
    23,18,999,2,117086,169783
    24,19,999,4,489,6237
    25,19,999,3,4968,6302
    26,20,999,2,52876,60839
    27,20,999,3,52202,59999
    28,20,999,3,53246,59999
    29,20,999,3,54205,60209
    30,20,999,3,54205,60209
    31,21,999,4,76204,106773
    32,21,999,4,76204,106773
    33,21,999,3,7877,105299
    34,21,999,4,76204,106773
    35,21,999,4,77576,105668
    36,21,999,3,7877,105299
    37,21,999,4,76204,105299
    38,21,999,2,81377,107595
    39,21,999,2,81377,107595
    40,21,999,4,77728,107595
    41,22,999,3,2187,2656
    42,22,999,2,2216,2661
    43,22,999,3,2187,2656
    44,22,999,2,2251,2632
    45,22,999,3,2187,2656
    46,22,999,4,2154,2628
    47,22,999,3,2187,2656
    48,22,999,3,2231,2661
    49,22,999,3,2231,2624
    50,22,999,2,2296,2632
    51,22,999,3,2231,2661
    52,22,999,4,2158,2661
    53,23,999,3,1913,2408
    54,23,999,3,1951,2375
    55,23,999,3,1987,2383
    56,23,999,3,1951,2408
    57,24,999,4,3946,4943
    58,24,999,3,4073,4883
    59,24,999,2,4053,4934
    60,24,999,2,4053,4866
    61,24,999,4,3946,4943
    62,24,999,3,4001,4943
    63,24,999,3,4154,4892
    64,24,999,4,4025,4875
    65,24,999,4,4025,4875
    66,24,999,2,4134,4875
    67,24,999,3,4081,4943
    68,24,999,2,4134,4934
    69,24,999,4,4025,4943
    70,24,999,4,4025,4875
    71,24,999,3,4081,4943
    72,25,999,3,983,12655
    73,25,999,3,983,12655
    74,25,999,2,9958,12655
    75,25,999,3,983,12655
    76,25,999,2,9958,12832
    77,25,999,3,10027,12832
    78,25,999,2,10157,12774
    79,25,999,4,9888,12655
    80,25,999,2,10157,12832
    81,25,999,4,9888,12832
    82,25,999,4,9888,12832
    83,26,999,3,1305,17034
    84,26,999,3,1305,16799
    85,26,999,3,13551,16858
    86,27,999,3,3992,4927
    87,27,999,3,4064,4876
    88,29,999,3,43761,56175
    89,29,999,2,44942,55621
    90,29,999,4,42335,55399
    91,29,999,2,44322,55399
    92,29,999,2,45208,56175
    93,29,999,2,45208,56175
    94,29,999,4,43182,56175
    95,29,999,3,44636,56175
    96,29,999,4,43182,56175
    97,29,999,4,43182,56175
    98,30,999,3,869,1094
    99,30,999,4,857,1079
    100,30,999,2,898,1079
    ----------------------------------------------------------------------------------------------------------------------------------------------------------------------

  • Loading single digit months with SQLLDR

    How can I load dates with the month having a single digit using sqlldr??
    62305220507 in this example the date is 23-Jun0-05 22:05:07
    111804115452 in this example the date is 18-Nov-04 11:54:52
    I've tried the date mask mmddyyhh24miss but only the double digit months get loaded.
    Thanks,
    Paul

    How about something like this
    my_date CHAR "to_date(lpad(:my_date, 12, '0'), 'mmddyyhh24miss')"P.S. And maybe you should consider to use RR rather than YY for the two digit year?

  • Load Error with return code 7000 when call sqlldr in interface.

    Hi All,
    I have created a interface to load data from 'File' to 'Oracle' in ODI 11g.
    Below are the running order of this interface.
    1 - Loading - SrcSet0 - Drop work view
    2 - Loading - SrcSet0 - Drop Temp Table
    3 - Loading - SrcSet0 - Create Temp table
    4 - Loading - SrcSet0 - Create Work View
    5 - Loading - SrcSet0 - Generate CTL file
    6 - Loading - SrcSet0 - Call sqlldr
    We are getting the following error while running the step 6.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (most recent call last):
      File "<string>", line 22, in <module>
    Load Error: See C:\home\oracle\joc\odi/AIM.log for details
      at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
      at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:322)
      at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
    And it return code 7000 indicated in 'Execution Statistics'.
    If anyone has encountered a similar issue and able to solve it, please share your thoughts with me.
    Thanks in advance for your help.
    Dicksam

    This forum is dedicated to supporting the use of Maven with FMW and the new Oracle Maven Repository.  Given that ODI does not currently provide Maven support, you are unlikely to get a response in this forum.  You might want to try posting in a different forum dedicated to Data Integration, such as the Master Data Management forum.

  • Problem with sqlldr and commit

    Hi,
    i have a problem with sqlldr and commit.
    I have a simple table with one colum [ col_id number(6) not null ]. The column "col_id" is primary key in the table. I have one file with 100.000 records ( number from 0 to 99.999 ).
    I want load the file in the table with sqlldr ( sql*loader ) but i want commit only if all records are loaded. If one record is discarded i want discarded all record of file.
    The proble is that in coventional path the commit is on 64 row but if i want the same records of file isn't possible and in direct path sqlldr disable primary key :(
    There are a solutions?
    Thanks
    I'm for the bad English

    This is my table:
    DROP TABLE TEST_SQLLOADER;
    CREATE TABLE TEST_SQLLOADER
    (     COL_ID NUMBER NOT NULL,
         CONSTRAINT TEST_SQLLOADER_PK PRIMARY KEY (COL_ID)
    This is my ctlfile ( test_sql_loader.ctl )
    OPTIONS
    DIRECT=false
    ,DISCARDMAX=1
    ,ERRORS=0
    ,ROWS=100000
    load data
    infile './test_sql_loader.csv'
    append
    into table TEST_SQLLOADER
    fields terminated by "," optionally enclosed by '"'
    ( col_id )
    test_sql_loader.csv
    0
    1
    2
    3
    99999
    i run sqlloader
    sqlldr xxx/yyy@orcl control=test_sql_loader.ctl log=test_sql_loader.log
    output on the screen
    Commit point reached - logical record count 92256
    Commit point reached - logical record count 93248
    Commit point reached - logical record count 94240
    Commit point reached - logical record count 95232
    Commit point reached - logical record count 96224
    Commit point reached - logical record count 97216
    Commit point reached - logical record count 98208
    Commit point reached - logical record count 99200
    Commit point reached - logical record count 100000
    Logfile
    SQL*Loader: Release 11.2.0.1.0 - Production on Sat Oct 3 14:50:17 2009
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Control File: test_sql_loader.ctl
    Data File: ./test_sql_loader.csv
    Bad File: test_sql_loader.bad
    Discard File: none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 0
    Bind array: 100000 rows, maximum of 256000 bytes
    Continuation: none specified
    Path used: Conventional
    Table TEST_SQLLOADER, loaded from every logical record.
    Insert option in effect for this table: APPEND
    Column Name Position Len Term Encl Datatype
    COL_ID FIRST * , O(") CHARACTER
    value used for ROWS parameter changed from 100000 to 992
    Table TEST_SQLLOADER:
    100000 Rows successfully loaded.
    0 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 255936 bytes(992 rows)
    Read buffer bytes: 1048576
    Total logical records skipped: 0
    Total logical records read: 100000
    Total logical records rejected: 0
    Total logical records discarded: 0
    Run began on Sat Oct 03 14:50:17 2009
    Run ended on Sat Oct 03 14:50:18 2009
    Elapsed time was: 00:00:01.09
    CPU time was: 00:00:00.06
    The commit is on 992 row
    if i have error on 993 record i have commit on first 992 row :(
    Edited by: inter1908 on 3-ott-2009 15.00

  • Sql@loader-704  and ORA-12154: error messages when trying to load data with SQL Loader

    I have a data base with two tables that is used by Apex 4.2. One table has 800,000 records . The other has 7 million records
    The client recently upgraded from Apex 3.2 to Apex 4.2 . We exported/imported the data to the new location with no problems
    The source of the data is an old mainframe system; I needed to make changes to the source data and then load the tables.
    The first time I loaded the data i did it from a command line with SQL loader
    Now when I try to load the data I get this message:
    sql@loader-704 Internal error: ulconnect OCISERVERATTACH
    ORA-12154: tns:could not resolve the connect identifier specified
    I've searched for postings on these error message and they all seem to say that SQL Ldr can't find my TNSNAMES file.
    I am able to  connect and load data with SQL Developer; so SQL developer is able to find the TNSNAMES file
    However SQL Developer will not let me load a file this big
    I have also tried to load the file within Apex  (SQL Workshop/ Utilities) but again, the file is too big.
    So it seems like SQL Loader is the only option
    I did find one post online that said to set an environment variable with the path to the TNSNAMES file, but that didn't work..
    Not sure what else to try or where to look
    thanks

    Hi,
    You must have more than one tnsnames file or multiple installations of oracle. What i suggest you do (as I'm sure will be mentioned in ed's link that you were already pointed at) is the following (* i assume you are on windows?)
    open a command prompt
    set TNS_ADMIN=PATH_TO_DIRECTOT_THAT_CONTAINS_CORRECT_TNSNAMES_FILE (i.e. something like set TNS_ADMIN=c:\oracle\network\admin)
    This will tell oracle use the config files you find here and no others
    then try sqlldr user/pass@db (in the same dos window)
    see if that connects and let us know.
    Cheers,
    Harry
    http://dbaharrison.blogspot.com

  • Little problem with SQLLDR

    Hi,
    I've got great help here in the past so I thought this would be the right place to try and find the solution to my problem loading data using SQLLDR into an oracle database.
    Basically, upon loading I am using a to_date function. However, the parameters this to_date function must take depends on the data I'm trying to load (there are only two possible date formats I'm trying to load). Is there any way of using two to_date functions inside a conditional statement (if statement etc.).
    Thanks a lot in advance!!
    Dan

    Looking at what you posted, i think you may have some issues with the quoting.
    to_date(:DATE_FIELD,''DD-MM-YYYY'))"
    {code}
    You have a double quote before the DD there.
    Edited by: Tubby on Jun 14, 2010 10:10 AM
    I see someone else is quicker than me :)                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Again got stuck with sqlldr.

    Hi All,
    I am again got stuck with sqlldr. Please find below my ctl and log file.
    ----------- CTL FILE--------------
    load data
    discardfile 'c:\dis.log'
    infile 'c:\csv\home1.txt'
    into table home_expenses
    fields terminated by ","
    trailing nullcols
    (PURCHASE_DATE date "mm/dd/yy",PAYEE,RUPEE_SPENT,CATEGORY,SUB_CATEGORY,DESCRIPTION)
    ---------------- SAMPLE DATA-----------
    05/05/09,Gas,Withdrawal,620,Automobile,Gas,,2 cylinders,
    05/05/09,clothing,Withdrawal,900,Homeneeds,Clothing,,suits for usha,meenu,sheetu,mummy,mousi & me
    01/05/09,Vegitables,Withdrawal,250,Food,Vegitables,,Potota 5 kg tomoto 2kg& others ,
    05/05/09,Petorl,Withdrawal,500,Petrol,,,Petrol for CAR,
    05/08/09,Vegitables,Withdrawal,250,Food,Vegitables,,Vegitables for home.,
    01/09/09,Maintainance,Withdrawal,100,Maintainance,,,Curtain Maintainence,
    ------------------ PART OF LOG FILE--------------------
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array: 64 rows, maximum of 256000 bytes
    Continuation: none specified
    Path used: Conventional
    Table HOME_EXPENSES, loaded from every logical record.
    Insert option in effect for this table: INSERT
    TRAILING NULLCOLS option in effect
    Column Name Position Len Term Encl Datatype
    PURCHASE_DATE FIRST * , DATE mm/dd/yy
    PAYEE NEXT * , CHARACTER
    RUPEE_SPENT NEXT * , CHARACTER
    CATEGORY NEXT * , CHARACTER
    SUB_CATEGORY NEXT * , CHARACTER
    DESCRIPTION NEXT * , CHARACTER
    Data File home.ctl -
    Record 1: Discarded - all columns null.
    Record 3: Discarded - all columns null.
    Record 5: Discarded - all columns null.
    Record 7: Discarded - all columns null.
    Record 18: Discarded - all columns null.
    Record 25: Discarded - all columns null.
    Record 29: Discarded - all columns null.
    Record 38: Discarded - all columns null.
    Record 39: Discarded - all columns null.
    Record 41: Discarded - all columns null.
    Record 44: Discarded - all columns null.
    Record 47: Discarded - all columns null.
    Record 50: Discarded - all columns null.
    Record 53: Discarded - all columns null.
    Record 56: Discarded - all columns null.
    Record 59: Discarded - all columns null.
    Record 62: Discarded - all columns null.
    Record 65: Discarded - all columns null.
    Record 68: Discarded - all columns null.
    Record 71: Discarded - all columns null.
    Record 74: Discarded - all columns null.
    Record 77: Discarded - all columns null.
    Record 80: Discarded - all columns null.
    Record 83: Discarded - all columns null.
    Record 86: Discarded - all columns null.
    Record 89: Discarded - all columns null.
    Record 2: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 4: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 6: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 8: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 9: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 10: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 11: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    hare krishna

    Thanks Blu for your responce. I also tried to modify very first records, according to you, it should be loaded, because it does not contain many ",". It sill failed to laod.
    --------- sample data---------------------
    05/05/09,Gas,Withdrawal,620,Automobile,Gas
    ------------------part my log file---------------
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array: 64 rows, maximum of 256000 bytes
    Continuation: none specified
    Path used: Conventional
    Table HOME_EXPENSES, loaded from every logical record.
    Insert option in effect for this table: INSERT
    TRAILING NULLCOLS option in effect
    Column Name Position Len Term Encl Datatype
    PURCHASE_DATE FIRST * , DATE mm/dd/yy
    PAYEE NEXT * , CHARACTER
    RUPEE_SPENT NEXT * , CHARACTER
    CATEGORY NEXT * , CHARACTER
    SUB_CATEGORY NEXT * , CHARACTER
    DESCRIPTION NEXT * , CHARACTER
    Data File home.ctl -
    Record 1: Discarded - all columns null.
    Record 3: Discarded - all columns null.
    Record 5: Discarded - all columns null.
    Record 7: Discarded - all columns null.
    Record 18: Discarded - all columns null.
    Record 25: Discarded - all columns null.
    Record 29: Discarded - all columns null.
    Record 38: Discarded - all columns null.
    Record 39: Discarded - all columns null.
    Record 41: Discarded - all columns null.
    Record 44: Discarded - all columns null.
    Record 47: Discarded - all columns null.
    Record 50: Discarded - all columns null.
    Record 53: Discarded - all columns null.
    Record 56: Discarded - all columns null.
    Record 59: Discarded - all columns null.
    Record 62: Discarded - all columns null.
    Record 65: Discarded - all columns null.
    Record 68: Discarded - all columns null.
    Record 71: Discarded - all columns null.
    Record 74: Discarded - all columns null.
    Record 77: Discarded - all columns null.
    Record 80: Discarded - all columns null.
    Record 83: Discarded - all columns null.
    Record 86: Discarded - all columns null.
    Record 89: Discarded - all columns null.
    Record 2: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 4: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 6: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 8: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 9: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 10: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 11: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    Record 12: Rejected - Error on table HOME_EXPENSES, column PURCHASE_DATE.
    ORA-01858: a non-numeric character was found where a numeric was expected
    hare krishna

  • SQL*Loader problem with direct path load

    Hi all,
    Its on Oracle 9.2
    I have a sqlldr control file which has couple of columns like,
    my_column_1 ,
    my_column_2 "decode(:my_column_1,'ONE','AAA','TWO','BBB', :my_column_1)"
    The table I am loading to is in user X and I am running the load from
    user Y. Everything works fine with conventional path load (not direct
    path) as grants are made for the table to user Y.
    When I load the data with same control file with direct path, I get an
    error ,
    01031 - insufficient privileges
    Is this anything to do with the syntex I have used in the control file
    or its a privilege issue. If its a privilege issue which privilege is
    that ?
    I did following tests,
    1) Load is run with conventional path load, from user Y and the decode
    statement is in control file - Load works
    2) Load is run with direct path load, from user Y and decode statement
    is in control file - Load fails with above mentioned error
    3) Load is run with direct path load, from user Y and decode IS REMOVED
    from the control file - Load works (!!!)
    What can be the conclusion? Way out ?
    Thanks and Regards

    You need to grant
    grant lock any table to userY;
    For more information see MetaLink Note 1082550.6

  • File loaded successfully by SQLldr but external table failed

    Hi ,
    When i tried to create an external table to load data from file it failed with the below error
    "KUP-04018: partial record at end of file",
    When i query on the external table for rows with rownum<2500 i can see the records being properly loaded but when i tried to fetch all records it throws the above error.
    But the same file when i loaded with sqlldr it loaded all the records successfully.
    Can you please let me know the reason for this?
    Regards

    Can you post us your SQLLDR control file, your external table definition, a sample of your data (preferably including the last lines of data which you believe are erroring) and also let us know your database version.

  • UNICODE data files with SQLLDR

    how can i load UNICODE data files with SQLLDR.
    my Oracle instance is on UNIX with NLS_CHARACTERSET WE8ISO8859P1.
    I have .dat files extracted from SQL Server using bcp utility with -w option.
    When i use -c option i'm not getting the european characters correctly like the a and e with 2 dots on top....
    when i load UNICODE (-w) file with CHARACTERSET UTF8 in my control file, it doesnt go thru. Any solution for this ? Thanks !

    I just created a unicode textfile on windows with some westeuropean characters and imported it into we8iso8859p1 database on linux using controlfile parameter CHARACTERSET UTF16.
    They got all properly converted.
    As Justin mentioned, unicode on windows means generally UTF16 Little Endian.
    Best regards
    Maxim

  • Since updating to I.O.S7 on my ipad2 all i get when i turn it on is a loading screen with a picture of a charger going into an ITUNES sign? HELP nothing will work not interactive at all?

    all i get when i turn it on is a loading screen with a picture of a charger going into an ITUNES sign? HELP nothing will work not interactive at all.aince updating to I.O.S7.0.2???
    What should i do currently have it on charge for several hours not done a thing and have been locking and unlocking and resetting

    Please see: http://support.apple.com/kb/HT1808 for further info.  It sounds like you are already in Recovery mode so if the image in Step 4 is the one you see, skip to Step 5.
    Regards,
    Steve

  • Load testing with concurrent users on HTTP Web

    Hello,
    I am trying to do load testing my PHP Web Application.
    I record (with Firefox) and developed the testing script with OpenScript,
    but I found that I could only test with "Iteration" feature.
    Is there any way to do load testing with concurrent users ( multi users login at the same time ) which defined in databank?
    Thanks for help.

    Hi
    You need to load the script into OLT (Oracle Load Testing) you can download it from here you need the first link Oracle Application Testing Suite and run the OATSxxx.exe to install OLT.
    Regards
    Alex

  • Incremental load fails with the error LM_44127 Failed to prepare the task

    Guys,
    I have created a custom mapping and cretaed a execution plan for this mapping in the DAC. The full load completes successfully. But when ever the incremental lod is run , i am getting the below error and the task fails(The SDE load completes sucessfully , but the SIL load fails with the below error).
    LM_44127     Failed to prepare the task
    Please help!!!

    i googled it..
    http://datawarehouse.ittoolbox.com/groups/technical-functional/informatica-l/lm_44127-failed-to-prepare-task-when-running-workflow-in-informatica-86-on-aix-3199309
    you can try for better links now.. !!

  • Data load issue with export data source - BW 3.5

    Hi,
    We are facing issues in loading data with the help of export data source.
    We have created export data source of 0PCA_C01 cube. With the help of this export datasource,  we are loading data to other custom cube. Scenario is working fine in development server.
    But when we transported objects to quality server data is not getting loaded to custom target cube.
    It is extracting zero records.  All transports are ok and we have generated export datasource in quality before transports .Also regenerated export datasource after transport and activated infosource, update rule via RS* programs.  Every object is active but data is not getting extracted.
    RSA3 for 80PCA_C01 datasource isn't extracting any record in Quality. Records getting extracted in development.   We are in BW 3.5 with patch level 19.
    Please guide us to resolve the issue.
    Thanks,
    Aditya

    Hi
    Make sure that you have relevant Role & Authorization at Quality/PRS.
    You have to Transport the Source Cube first and then Create a Generate Export Data Source in QAS. Then, replicate data sources for BW QAS Soruce System. Make sure this replicated Data Source in QAS. Only then can transport new update rules for second cube.
    Hope it helps and clear

  • How can I convert my music, and CDs to MP3, for use in my vehicle's AM/FM/ Stereo? I have a 16 GB SDHC Card, that I would like to load up with some of my music.

    How can I convert my music, and CDs to MP3, for use in my vehicle's AM/FM/ Stereo? I have a 16 GB SDHC Card, that I would like to load up with some of my music.
    Thanks for any help you can furnish.
    Ed

    Well, this question really belongs in the iTunes forum, which is where you will get fast answers.
    Just go to iTunes Preferences.  Click on the General icon.  Set When you insert a CD to Ask to Import CD.  Click on Import Settings.  Set Import to MP3 decoder.  Make sure you're using Good Quality.
    This should deal with your CD's.
    With your current music, as long as it does not have DRM restrictions (older iTunes Music did), select the song, right click on the song (or group of songs), pull down to Create MP3 version. 

Maybe you are looking for