ORA-29275: partial multibyte character with 11.1.0.7

Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
SQL> create table test_char(c1 varchar2(100));
Table created.
SQL> insert into test_char values (chr(200));
1 row created.
SQL> select from test_char;*
ERROR:
**ORA-29275: partial multibyte character**
NLS_CHARACTERSET ==> AL32UTF8
But, with Oracle9i Enterprise Edition Release 9.2.0.7.0 - 64bit Production
Don't throw any error
SQL> select * from test_char;
C1
question is what is the reason behind throws this ORA-29275 and how to work around or suppress this error in 11.1.0.7(similar to 9.2)

And with CLOB it is ORA-600 :
Connected to:
Personal Oracle Database 11g Release 11.1.0.7.0 - Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
SQL> create table test_char(c1 clob);
Table created.
SQL> insert into test_char values (chr(200));
insert into test_char values (chr(200))
ERROR at line 1:
ORA-00600: internal error code, arguments: [kole_t2u], [34], [], [], [], [],
SQL>You should read carefully the Metalink note #788156.1 (AL32UTF8 / UTF8 (Unicode) Database Character Set Implications) part *B.4) The meaning of SP2-0784, ORA-29275 and ORA-600 [kole_t2u], [34] errors / losing characters when using convert.*+
And also
*ORA-600 [kole_t2u], [34] - description, bugs, and reasons*+
Doc ID: 734474.1+
Nicolas.

Similar Messages

  • Dynamic SQL run with error 'ORA-29275: partial multibyte character'

    Hi All,
    In daily work, we use daynamic SQL to load data from another database through DB Link. However, even if I run the following simple statement in our DB, it will raise the error.
    CREATE TABLE TEST_TMP
    AS
    SELECT TRANSACTION_REASON
    FROM [email protected] --it is using db link.
    WHERE LAST_UPDATE_DATE>=TO_DATE('18-12-2012 22:01:58', 'dd-mm-yyyy hh24:mi:ss');
    This statement had been workng correctly before OCT-2012 and there was no change with the source. Afterwards, there was just a change which upgraded the DB version from 10g to 11g on our side. I have browsed my technical articles, but I still cannot find the root cause. Is it due to the updgrading of the DB which practice stricter sanity check?
    I don't know.
    I'd appreciate of you can give me suggestion if you know such case.
    Thanks,
    David

    David Paul wrote:
    Hi,
    Can you give me a full explaination about the DL http://vibhorkumar.wordpress.com/2011/02/27/fix-of-ora-29275-partial-multibyte-character/ you mentioned?
    I don't understand the query in it well.
    Thanks,
    DavidDon't confuse yourself... It's just saying to convert data into your local DB type...
    Check these queries -
    1] /* check this in both your LOCAL and REMOTE databases */
    select dump('a',1010) from dual;
    I got --
    Typ=96 Len=1 CharacterSet=AL32UTF8: 97
    So, my charset encoding is AL32UTF8. Similarly, find yours for both Local & Remote DBs.
    2]
    /* Then do */
    CONVERT(transaction_reason,'<local_charset>','<remote_charset>')Edited by: ranit B on Dec 22, 2012 5:46 PM

  • A database error occured. The database error text is: ORA-29275: partial multibyte character . (WIS 10901)

    Hi,
    My Webi report is geeting failed with the error
    "A database error occured. The database error text is: ORA-29275: partial multibyte character . (WIS 10901)"
    may i know the root cause of the above error and how to resolve it. I am using BO 3.1.
    Its very important to provide the report. Please help urgently.
    Thanks in advance.
    Abid

    Hi Abid,
    Please see SAP Note 1556127.
    Symptom
    A database error occurs after refreshing a web intelligence report in java report panel or web intelligence in interactive mode
    The database error text is: ORA 29275 with partial multibyte character (WIS 10901)
    Environment
    windows 2003 Server
    Cause
    Environment variables are not set with value UTF-8:LC_ALL,LANG, and NLS_LANG
    Resolution
    Set following system environment variables: LC_ALL,LANG, and NLS_LANG with value UTF-8. For example, LC_ALL=EN_US.UTF-8

  • ORA-29275: partial multibyte character

    Hi,
    We have a database having three instances dev,stage and prod. We need to fetch data from other remote DB respective instances. for eg dev instance on our DB fetches data from dev instance of remote DB. But I am getting ORA-29275: partial multibyte character for one particular query in prod instance only.
    Please advice.
    Thanks,
    Sanchita

    Hi;
    What is DB version? What is OS? Is there any error message in alert log?
    Please see below notes:
    Unable to resync due to ORA-29275: partial multibyte character [ID 394166.1]
    Bug 6268409 - ORA-29275 error when querying the sql_redo/undo columns of V$LOGMNR_CONTENTS [ID 6268409.8]
    ORA-29275 When Querying V$LOGMNR_CONTENTS [ID 562991.1]
    Regard
    Helios

  • Partial Multibyte Character Error

    Hi All,
    I am getting following error while running a report in OBIEE 11 G.
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 17001] Oracle Error code: 29275, message: ORA-29275: partial multibyte character at OCI call OCIStmtFetch. [nQSError: 17012] Bulk fetch failed. (HY000)
    NLS_LANG Paramters is set at the Oracle database 11.
    [NLS_LANGUAGE].[NLS_TERRITORY][NLS_CHARACTERSET]----------->AMERICAN.AMERICA.AL32UTF8
    Any help in this.

    Hi,
    It's seems are you using char function in your report?
    If yes, Change the char size .
    ex: cast(column name as char).
    A. Cast(column name as char(300)).
    I am not sure this is what your looking so far.
    Award points it is useful.
    Thanks,
    Satya

  • BODI not able to read multibyte character data

    Hi,
    I have created a hashing function with MD5 algorithm in the oracle DB and it's returntype is varchar2. When I try to call this function in BODI, I get the error <OCIStmtFetch>: <ORA-29275: partial multibyte character. I pass a field as an input parameter to the function.
    My DB is oracle 11g, BODS version is 12.3.2. The datastores I use have been configured as utf-8 for both the source and the target.
    Please advise.
    Thanks

    Hi Saurabh
    How do you put the file on the application server?
    Might help to have a look at:
    [How to Work with Character Encodings in Process Integration (NW7.0)|https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/502991a2-45d9-2910-d99f-8aba5d79fb42]
    Best regards,
    Chris

  • ORA-29275

    Hi,
    I am creating a table like this:
    CREATE tableA
    As SELECT * FROM tableA@DBLINK;
    And I get this error:
    ERROR at line 2:
    ORA-29275: partial multibyte character
    Thanks.

    Pl post details of OS and database versions on the two databases, along with the characterset of the two databases.
    HTH
    Srini

  • Obiee server SQL throw exception :ORA-29275

    The obiee server throws oracle exception: ORA-29275 while execute in the sql in the bidw , the sql results have Chinese characters
    I have try to copy the sql to query from Jdeveloper adn SQL developer , the sql can run and no exception so does it the NSL_LANG issue in the obiee server side ?
    I also have try to set the NSL_LANG as AMERICAN_AMERICA.AL32UTF8 and AMERICAN_AMERICA.UTF8 , and they seems not worked!
    Odbc driver returned an error (SQLExecDirectW).
    Error Details
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 17001] Oracle Error code: 29275, message: ORA-29275: partial multibyte character at OCI call OCIStmtFetch. [nQSError: 17012] Bulk fetch failed. (HY000)

    Any one can help?

  • Multibyte character error in SqlLoader when utf8 file with chars like €Ää

    hello,
    posting from Germany, special charactes like german umlaute and euro sign in UTF8 Textfile, SqlLoader rejecting row with Multibyte character error
    Oracle Database 11g Release 11.2.0.2.0 - 64bit Production
    Database Characterset: WE8MSWIN1252
    OS: SLES 11 x86_64
    Testcase SqlDeveloper:
    CREATE TABLE utf8file_to_we8mswin1252 (
    ID NUMBER,
    text VARCHAR2(40 CHAR)
    can't enter euro symbol in this posting, end's in '€' (?)
    SELECT ascii(euro symbol) FROM dual;
    128
    SELECT chr(128) from dual;
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (1, '0987654321098765432109876543210987654321');
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (2, 'äüöäüöäüöäÄÖÜÄÖÜÄÖÜÄßßßßßßßßß߀€€€€€€€€€');
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (3, 'äüöäüöäüöäÄÖÜÄÖÜÄÖÜÄäüöäüöäüöäÄÖÜÄÖÜÄÖÜÄ');
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (4, 'ۧۧۧۧۧۧۧۧۧۧ1');
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (5, 'äüöäüöäüöäÄÖÜÄÖÜÄÖÜÄäüöäüöäüöä');
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (6, 'ßßßßßßßßß߀€€€€€€€€€1');
    INSERT INTO utf8file_to_we8mswin1252 (ID, text) VALUES (7, 'ßßßßßßßßß߀€€€€€€€€€äüöäüöäüöäÄÖÜÄÖÜÄÖÜÄ');
    commit;
    Select shows correct result, no character is wrong or missing!!!!
    put this in a UTF8 file without delimiter and enclosure like
    10987654321098765432109876543210987654321
    the SqlLoader controlfile:
    LOAD DATA characterset UTF8
    TRUNCATE                              
    INTO TABLE utf8file_to_we8mswin1252
    ID CHAR(1)     
    , TEXT CHAR(40)
    on a linux client machine, NOT the Oracle-Server
    export NLS_LANG=AMERICAN_AMERICA.WE8MSWIN1252
    sqlldr user/pwd@connectstring CONTROL=TEST01.ctl DATA=TEST01.dat LOG=TEST01.log
    Record 6: Rejected - Error on table UTF8FILE_TO_WE8MSWIN1252, column TEXT.
    Multibyte character error.
    Record 7: Rejected - Error on table UTF8FILE_TO_WE8MSWIN1252, column TEXT.
    Multibyte character error.
    Select shows missing characters in row 4 and 5, SqlLoader loads only the first 20 characters (maybe at random)
    and as shown above, row 6 and 7 never loaded
    Problem:
    can't load UTF8 Flatfiles with SqlLoader when german umlaute and special characters like euro symbol included.
    Any hint or help would be appreciated
    Regards
    Michael

    ## put this in a UTF8 file without delimiter and enclosure like
    The basic question is how you put the characters into the file. Most probably, you produced a WE8MSWIN1252 file and not an UTF8 file. To confirm, a look at the binary codes in the file would be necessary. Use a hex-mode-capable editor. If the file is WE8MSWIN1252, and not UTF8, then the SQL*Loader control file should be:
    LOAD DATA characterset WE8MSWIN1252
    TRUNCATE
    INTO TABLE utf8file_to_we8mswin1252
    ID CHAR(1)
    , TEXT CHAR(40)
    )-- Sergiusz

  • Creating a link to a Multibyte file with PDF 1.4

    I am trying to create a link to a file that contains multibyte characters (specifically Japanese) - unfortunately, I need to conform to the PDF 1.4 standard (I can make things work fine using PDF 1.7 and /UF).
    The 1.7 standard says that as of PDF 1.2, a file specification can contain multibyte character codes if you represent them in hexadecimal format. I cannot get this to work, and wonder if someone has any insight into why this might be.
    My multibyte file spec is UTF-8 encoded, and this is how I write it (in hex) to the document (without thr UTF-8 BOM). When I later click on that link, I get an error that the specified document does not exist - the part of the filespec that is English characters are displayed correctly, but the Japanese characters show up incorrectly.
    Does Acrobat expect these characters to be UTF-8 encoded, or otherwise? I've tried prepended the UTF-8 BOM, but that didn't make a difference. Do I need to UTF-16BE encode them (like with /UF)?
    Additional info: I've tried using Acrobat 5 to create a new PDF and inserted a link to a file whose name contains Japanese characters - when clicking on the link in this, I also get the 'file does not exist' error. Also, looking in the PDF, the filespec is NOT hex encoded, and the Japanese characters have been replaced by ???.

    Hmm... so, what does it mean when the 1.7 spec says 'In PDF 1.2 or higher, a file specification may contain multiple-byte character
    codes, represented in hexadecimal form between angle brackets (< and >).'? If Acrobat 5 supports PDF 1.4, I would have thought this would mean that Acrobat 5 would support multibyte character codes...
    I would guess that this puts the final nail in the coffin of creating links to filenames with non Roman characters from Acrobat 5.
    Just so I can report this upwards, I see that you are an Adobe employee - are you a support person?

  • Error while running package ORA-06553: PLS-553 character set name is not re

    Hi all.
    I have a problem with a package, when I run it returns me code error:
    ORA-06552: PL/SQL: Compilation unit analysis terminated
    ORA-06553: PLS-553: character set name is not recognized
    The full context of the problem is this:
    Previously I had a developing data base, then was migrated to a new server. After that I started to receive the error, so I began to check for the solution.
    My first move was compare the “old database” with the “new database”, so I check the nls parameters, and this was the result:
    select * from nls_database_parameters;
    Result from the old
    NLS_LANGUAGE     AMERICAN
    NLS_TERRITORY     AMERICA
    NLS_CURRENCY     $
    NLS_ISO_CURRENCY     AMERICA
    NLS_NUMERIC_CHARACTERS     .,
    NLS_CHARACTERSET     US7ASCII
    NLS_CALENDAR     GREGORIAN
    NLS_DATE_FORMAT     DD-MON-RR
    NLS_DATE_LANGUAGE     AMERICAN
    NLS_SORT     BINARY
    NLS_TIME_FORMAT     HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY     $
    NLS_COMP     BINARY
    NLS_LENGTH_SEMANTICS     BYTE
    NLS_NCHAR_CONV_EXCP     FALSE
    NLS_NCHAR_CHARACTERSET     AL16UTF16
    NLS_RDBMS_VERSION     10.2.0.1.0
    Result from the new
    NLS_LANGUAGE     AMERICAN
    NLS_TERRITORY     AMERICA
    NLS_CURRENCY     $
    NLS_ISO_CURRENCY     AMERICA
    NLS_NUMERIC_CHARACTERS     .,
    NLS_CHARACTERSET     US7ASCII
    NLS_CALENDAR     GREGORIAN
    NLS_DATE_FORMAT     DD-MON-RR
    NLS_DATE_LANGUAGE     AMERICAN
    NLS_SORT     BINARY
    NLS_TIME_FORMAT     HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY     $
    NLS_COMP     BINARY
    NLS_LENGTH_SEMANTICS     BYTE
    NLS_NCHAR_CONV_EXCP     FALSE
    NLS_NCHAR_CHARACTERSET     AL16UTF16
    NLS_RDBMS_VERSION     10.2.0.1.0
    As the result was identical, I decided to look for more info in the log file of the new database. What I find was this:
    Database Characterset is US7ASCII
    Threshold validation cannot be done before catproc is loaded.
    Threshold validation cannot be done before catproc is loaded.
    alter database character set INTERNAL_CONVERT WE8MSWIN1252
    Updating character set in controlfile to WE8MSWIN1252
    Synchronizing connection with database character set information
    Refreshing type attributes with new character set information
    Completed: alter database character set INTERNAL_CONVERT WE8MSWIN1252
    alter database character set US7ASCII
    ORA-12712 signalled during: alter database character set US7ASCII...
    alter database character set US7ASCII
    ORA-12712 signalled during: alter database character set US7ASCII...
    Errors in file e:\oracle\product\10.2.0\admin\orcl\udump\orcl_ora_3132.trc:
    Regards

    Ohselotl wrote:
    Hi all.
    I have a problem with a package, when I run it returns me code error:
    ORA-06552: PL/SQL: Compilation unit analysis terminated
    ORA-06553: PLS-553: character set name is not recognized
    The full context of the problem is this:
    Previously I had a developing data base, then was migrated to a new server. After that I started to receive the error, so I began to check for the solution.
    My first move was compare the “old database” with the “new database”, so I check the nls parameters, and this was the result:
    select * from nls_database_parameters;
    Result from the old
    NLS_LANGUAGE     AMERICAN
    NLS_TERRITORY     AMERICA
    NLS_CURRENCY     $
    NLS_ISO_CURRENCY     AMERICA
    NLS_NUMERIC_CHARACTERS     .,
    NLS_CHARACTERSET     US7ASCII
    NLS_CALENDAR     GREGORIAN
    NLS_DATE_FORMAT     DD-MON-RR
    NLS_DATE_LANGUAGE     AMERICAN
    NLS_SORT     BINARY
    NLS_TIME_FORMAT     HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY     $
    NLS_COMP     BINARY
    NLS_LENGTH_SEMANTICS     BYTE
    NLS_NCHAR_CONV_EXCP     FALSE
    NLS_NCHAR_CHARACTERSET     AL16UTF16
    NLS_RDBMS_VERSION     10.2.0.1.0
    Result from the new
    NLS_LANGUAGE     AMERICAN
    NLS_TERRITORY     AMERICA
    NLS_CURRENCY     $
    NLS_ISO_CURRENCY     AMERICA
    NLS_NUMERIC_CHARACTERS     .,
    NLS_CHARACTERSET     US7ASCII
    NLS_CALENDAR     GREGORIAN
    NLS_DATE_FORMAT     DD-MON-RR
    NLS_DATE_LANGUAGE     AMERICAN
    NLS_SORT     BINARY
    NLS_TIME_FORMAT     HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY     $
    NLS_COMP     BINARY
    NLS_LENGTH_SEMANTICS     BYTE
    NLS_NCHAR_CONV_EXCP     FALSE
    NLS_NCHAR_CHARACTERSET     AL16UTF16
    NLS_RDBMS_VERSION     10.2.0.1.0
    As the result was identical, I decided to look for more info in the log file of the new database. What I find was this:
    Database Characterset is US7ASCII
    Threshold validation cannot be done before catproc is loaded.
    Threshold validation cannot be done before catproc is loaded.
    alter database character set INTERNAL_CONVERT WE8MSWIN1252
    Updating character set in controlfile to WE8MSWIN1252
    Synchronizing connection with database character set information
    Refreshing type attributes with new character set information
    Completed: alter database character set INTERNAL_CONVERT WE8MSWIN1252
    *********************************************************************This is an unsupported method to change the characterset of a database - it has caused the corruption of your database beyond repair. Hopefully you have a backup you can recover from. Whoever did this did not know what they were doing.
    alter database character set US7ASCII
    ORA-12712 signalled during: alter database character set US7ASCII...
    alter database character set US7ASCII
    ORA-12712 signalled during: alter database character set US7ASCII...
    Errors in file e:\oracle\product\10.2.0\admin\orcl\udump\orcl_ora_3132.trc:
    RegardsThe correct way to change the characterset of a database is documented - http://docs.oracle.com/cd/B19306_01/server.102/b14225/ch11charsetmig.htm#sthref1476
    HTH
    Srini

  • Multibyte character (Chinese, Korean....) support problem

    I used OLE DB(C++) and JDBC(Java) in Oracle 8i server with no problem. I worked very well.
    but using ADO.NET(C#), there are the critical problem in selecting and inserting multibyte character (Chinese,Korean.. so on)
    Oracle 8i Server charset : US7ASCII
    Client NLS_LANG : AMERICAN_AMERICA.US7ASCII
    - SELECT multibyte column
    SELECT DUMP(multibyte) FROM table;
    DUMP(multibyte)
    Typ=1 Len=8: 199,209,177,219,195,226,183,194
    - value in OracleString
    OracleString.Value = ������� (broken character 0xfffd)
    I think Oracle Data Provider for .NET can handle only ASCII characters with setting charset US7ASCII.
    How I can get multibyte characters through byte array as binary data? (not String)

    Since your database character set is not UTF8 or Chinese character set, when inserting Chinese, if the character is not found/mapped in the database character set, a replacement character is used which will be the question marks you see.
    Ashish

  • Multibyte Character handling in Upload

    Hi Everyone,
    We had handled the multibyte character on uploading a CSV file converting BLOB to NCLOB instead of earlier BLOB to CLOB.  This worked fine as per our requirement but issue arised with the performance.  A file that was loading in 20 seconds, was loading in more than 7 minutes.
    Please let us know if there are any equivalent of NCLOB usage to handle multibyte character without compromising on the performance of the upload.
    Any suggestions appreciated
    DB Version used is 11g
    Regards,
    Sunil

    Hi Sunil,
    Even I am facing the same problem.
    Using BLOB to NCLOB is degrading the Performance.
    Can anyone help here.
    DB version is 11G.
    Regards,
    Animesh

  • SQL Loader Multibyte character error

    Hello,
    Oracle Database 11g Release 11.2.0.2.0 - 64bit Production
    Database Characterset: WE8MSWIN1252
    To load an utf8 File (UTF-8 Unicode text) i use option characterset UTF8 in controlfile.
    All went fine until in textfile the column is filled up with 40 chars and
    german umlaute included (ÜÖÄüöä...)
    Loader stops (errors=0) and gives:
    Record 146466: Rejected - Error on table 'TableName', column 'ColumnName'.
    Multibyte character error
    * use CHAR(40) and POSITION(start-end) for field description - no help
    * modify column, from VC2(40 char) to VC2(50 char) - no help
    * without characterset UTF8 option i got wrong characters for german umlaute - no help
    * manual insert the data from this row with NO problems !
    any hint or workaround?
    Regards
    Michael

    Hi Werner,
    on my linux desktop:
    $ file test.dat
    test.dat: UTF-8 Unicode text, with very long lines
    my colleague is working on a windows system.
    On both systems exact the same error from SQL Loader.
    Btw, try with different number of special characters (german umlaute and euro) and there is no chance to load without the error
    when to many (?) special characters or data is long as column length and special characters included.
    Regards
    Michael

  • SQL Loader Multibyte character error, LENGTH SEMANTICS CHARACTER

    Hi,
    startet SQL Loader Multibyte character error
    {thread:id=2340726}
    some mod locked the thread, why?
    the solution for others:
    add LENGTH SEMANTICS CHARACTER to the controlfile
    LOAD DATA characterset UTF8 LENGTH SEMANTICS CHARACTER
    TRUNCATE                              
    INTO TABLE utf8file_to_we8mswin1252
      ID    CHAR(1)     
    , TEXT  CHAR(40)
    )Regards
    Michael

    Hi Werner,
    on my linux desktop:
    $ file test.dat
    test.dat: UTF-8 Unicode text, with very long lines
    my colleague is working on a windows system.
    On both systems exact the same error from SQL Loader.
    Btw, try with different number of special characters (german umlaute and euro) and there is no chance to load without the error
    when to many (?) special characters or data is long as column length and special characters included.
    Regards
    Michael

Maybe you are looking for