Urgent :SQL Loader Arabic Character Set Issue

HI all,
I am loading arabic characters into my database using SQL Loader using a fixed length data file. I had set my characterset and NLS_LANG set to UTF8.When I try to load the chararacter 'B' in arabic data i.e. ' لا ' , it gets loaded as junk in the table. All other characters are loaded correctly. Please help me in this issue and its very urgent.
Thanks,
Karthik

Hi,
Thanks for the responses.
Even after setting the characterset to arabic and the problem continues to persist. This problem occurs only with the character "b".
Please find my sample control file,input file and nls_parameters below:
My control file
LOAD DATA
characterset UTF8
LENGTH SEMANTICS CHAR
BYTEORDER little endian
INFILE 'C:\sample tape files\ARAB.txt'
replace INTO TABLE user1
TRAILING NULLCOLS
name POSITION(1:2) CHAR(1),
id POSITION (3:3) CHAR(1) ,
salary POSITION (4:5) CHAR(2)
My Input file - Fixed Format
?a01
??b02
?c03
The ? indicates arabic characters.Arabic fonts must be installed to view them.
NLS_PARAMETERS
PARAMETER     VALUE
NLS_LANGUAGE     ARABIC
NLS_TERRITORY     UNITED ARAB EMIRATES
NLS_CURRENCY     ?.?.
NLS_ISO_CURRENCY     UNITED ARAB EMIRATES
NLS_NUMERIC_CHARACTERS     .,
NLS_CALENDAR     GREGORIAN
NLS_DATE_FORMAT     DD/MM/RR
NLS_DATE_LANGUAGE     ARABIC
NLS_SORT     ARABIC
NLS_TIME_FORMAT     HH12:MI:SSXFF PM
NLS_TIMESTAMP_FORMAT     DD/MM/RR HH12:MI:SSXFF PM
NLS_TIME_TZ_FORMAT     HH12:MI:SSXFF PM TZR
NLS_TIMESTAMP_TZ_FORMAT     DD/MM/RR HH12:MI:SSXFF PM TZR
NLS_DUAL_CURRENCY     ?.?.
NLS_COMP     BINARY
NLS_LENGTH_SEMANTICS     CHAR
NLS_NCHAR_CONV_EXCP     FALSE

Similar Messages

  • PL/SQL XML generation: Character set issues

    Hi,
    I am using the PL/SQL DOM (wrappers to the Java DOM), to generate XML from bits of database information. (On Oracle 8i).
    The output XML must be in UTF-8, and the database character set must be anything I want it to be. So I call
    setCharset(doc, 'UTF8')
    at the beginning, and I call
    writeToClob(doc, xmllob, 'UTF8')
    at the end, just to cover all eventualities.
    However, any character outside ASCII gets
    replaced with the character string "\xBF\xBF", which is rather tedious.
    If, instead, we go via
    writeToBuffer(doc, xmlbuf, 'UTF8')
    and then dump the buffer contents into a clob, the UTF8 encoding is preserved, and everything works.
    (This latter method is not good enough for my needs; I need more than 32K of data...)
    So I was wondering if any kind soul could tell me what I am doing wrong.
    Thanks,
    << Mike Alexander >>
    null

    I have the same problem. Is there any solution found?
    only xslprocessor.valueOf returns values of xml document not loosing special symbols.

  • SQL Loader Multibyte character error, LENGTH SEMANTICS CHARACTER

    Hi,
    startet SQL Loader Multibyte character error
    {thread:id=2340726}
    some mod locked the thread, why?
    the solution for others:
    add LENGTH SEMANTICS CHARACTER to the controlfile
    LOAD DATA characterset UTF8 LENGTH SEMANTICS CHARACTER
    TRUNCATE                              
    INTO TABLE utf8file_to_we8mswin1252
      ID    CHAR(1)     
    , TEXT  CHAR(40)
    )Regards
    Michael

    Hi Werner,
    on my linux desktop:
    $ file test.dat
    test.dat: UTF-8 Unicode text, with very long lines
    my colleague is working on a windows system.
    On both systems exact the same error from SQL Loader.
    Btw, try with different number of special characters (german umlaute and euro) and there is no chance to load without the error
    when to many (?) special characters or data is long as column length and special characters included.
    Regards
    Michael

  • Oracle 10g db character set issue

    I have a database 10g with database character set western
    european "WE8ISO8859P1" and we are receiving data from source
    database with database character set as "UTF8" during data load
    for one of the tables we receive the following error "ORA-29275:
    partial multibyte character" I understand this might be due to
    the fact western european character set is not a subset/superset
    of UTF8 .Am i right ? What would be the way around this ?

    It is certainly possible that the issue is that your database characterset is a subset of UTF8.
    How are you getting the data? Are we talking about a flat file? A query over a database link? Something else?
    Does the data you're getting contain characters that cannot be represented in the ISO-8859 1 character set? It is quite common to send UTF-8 encoded files even when the underlying data is representable in other 8-bit character sets (like ISO-8859 1).
    What are you trying to do with the data? Are you trying to load it into a CHAR/ VARCHAR2 column? A CLOB? A BLOB? An NCHAR/ NVARCHAR2? Something else?
    Justin

  • SQL Loader Multibyte character error

    Hello,
    Oracle Database 11g Release 11.2.0.2.0 - 64bit Production
    Database Characterset: WE8MSWIN1252
    To load an utf8 File (UTF-8 Unicode text) i use option characterset UTF8 in controlfile.
    All went fine until in textfile the column is filled up with 40 chars and
    german umlaute included (ÜÖÄüöä...)
    Loader stops (errors=0) and gives:
    Record 146466: Rejected - Error on table 'TableName', column 'ColumnName'.
    Multibyte character error
    * use CHAR(40) and POSITION(start-end) for field description - no help
    * modify column, from VC2(40 char) to VC2(50 char) - no help
    * without characterset UTF8 option i got wrong characters for german umlaute - no help
    * manual insert the data from this row with NO problems !
    any hint or workaround?
    Regards
    Michael

    Hi Werner,
    on my linux desktop:
    $ file test.dat
    test.dat: UTF-8 Unicode text, with very long lines
    my colleague is working on a windows system.
    On both systems exact the same error from SQL Loader.
    Btw, try with different number of special characters (german umlaute and euro) and there is no chance to load without the error
    when to many (?) special characters or data is long as column length and special characters included.
    Regards
    Michael

  • SQL LOADER and SHELL SCRIPT ISSUE

    Hello Guys,
    I know this not the right forum but i am not sure where i should post this.
    Pelase help
    I am running a shell script which is giving me error
    Username:SQL*Loader-128: unable to begin a session
    ORA-01017: invalid username/password; logon denied
    SQL*Loader: Release 10.2.0.4.0 - Production on Thu Nov 19 13:02:04 2009
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    SQL*Plus: Release 10.2.0.4.0 - Production on Thu Nov 19 13:02:06 2009
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Enter user-name:
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options
    SQL>
    0 rows updated.
    Commit complete.
    SQL> SQL> Disconnected from Oracle Database 10g Enterprise Edition
    SQL> SQL> Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options Thu Nov 19 13:02:06 EST 2009In shell script i have used the same username and passwd which i am using from command line
    the shell script is calling sql loader to load file
    and for that also the username and passwd is same.
    i am able to run sqlldr command from command line dont knw why here its giving error
    here is my shell script
    set -a
    . $HOME/.edw.env
    . $admlib/checklib.sh
    LOGDIR=$admsrc/sigma6/ppadala/copg
    LOGFILE=${LOGDIR}/log/test`date '+%m%d'`.xtr
    DB_USER=copg
    DB_PWD=copg
    set +a
    cd $LOGDIR
    if test ! -f $admsrc/sigma6/ppadala/copg/DM_Daily_EFolderCloseCancel_Report_11192009.txt
    then
       echo "Error: DM_Daily_EFolderCloseCancel_Report_11192009.txt does not exist and/or is not a regular file." >> ${LOGFILE}
       exit 1
    fi
    echo 'End of Checking for the existence of the file - Successful'>> ${LOGFILE}
    sqlldr control=$admsrc/sigma6/ppadala/copg/Close_Cancle.ctl log=$admsrc/sigma6/ppadala/copg/Close_cancle.log
    userid=${DB_USER}/${DB_PWD} silent=\(HEADER,FEEDBACK,DISCARDS\)>> ${LOGFILE} 2>&1
    case $? in 0) :;;1|3) echo "Error: SQL Loader" >> ${LOGFILE}
         exit 1;;
    esac
    sqlplus << EOD
    ${DB_USER}/${DB_PWD}
    @Close_Cancle.sql
    EOD
    if [ $? -ne 0 ]
    then
        echo "Error: SQL Plus for script Processing" >> ${LOGFILE}
        echo "Resi Unit Scheduling Report Refresh failed" >> ${LOGFILE}
    fi
    ) > ${LOGFILE} 2>&1
    echo `date` >> ${LOGFILE}
    if [ -f ${LOGFILE} ]
    then
    mail -s "Resi Unit Scheduling" "[email protected]" < ${LOGFILE}
    sleep 3
    `ck_error ${LOGFILE}`
    fiplease help guys
    thanks

    Thanks for the reply
    In Close_cancle.log also its the same msg which i posted.
    logon denied..............
    and this is the log file contents when i do set - X on
    + cd /u2144009/src/sigma6/ppadala/copg
    + test ! -f
    + /u2144009/src/sigma6/ppadala/copg/DM_Daily_EFolderCloseCancel_Report_1
    + 1192009.txt echo End of Checking for the existence of the file -
    + Successful
    + 1>> /u2144009/src/sigma6/ppadala/copg/log/test1119.xtr
    + sqlldr control=/u2144009/src/sigma6/ppadala/copg/Close_Cancle.ctl
    + log=/u2144009/src/sigma6/ppadala/copg/Close_cancle.log
    Username:SQL*Loader-128: unable to begin a session
    ORA-01017: invalid username/password; logon denied
    SQL*Loader: Release 10.2.0.4.0 - Production on Thu Nov 19 17:32:17 2009
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    + userid=copg/copg silent=(HEADER,FEEDBACK,DISCARDS)
    + 1>> /u2144009/src/sigma6/ppadala/copg/log/test1119.xtr 2>& 1
    + :
    + sqlplus
    + 0<<
    copg/copg
    @Close_Cancle.sql
    SQL*Plus: Release 10.2.0.4.0 - Production on Thu Nov 19 17:32:58 2009
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Enter user-name:
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options
    SQL>
    0 rows updated.
    Commit complete.
    SQL> SQL> Disconnected from Oracle Database 10g Enterprise Edition
    SQL> SQL> Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    + [ 0 -ne 0 ]
    Thu Nov 19 17:32:59 EST 2009Edited by: user10647455 on Nov 19, 2009 2:35 PM

  • Character Set issues.  Please advise

    I have a client who use a version 10gR2 DB that stores both English and French data. There are several times where they will send up .dmp file where we load it into ours.
    2 questions.
    What would be the best charater sets to use here in this setup?
    I am assuming we would use
    NLS_CHARACTERSET = WE8ISO9959P1
    NLS_NCHAR_CHARACTERSET = AL16UTF16
    Also if someone can confirm for me.
    NLS_CHARACTERSET = database character set ???
    NLS_NCHAR_CHARACTERSET = national character set???

    So is it better to say that I should use the AL32UTF8
    instead of AL16UTF16 ?It's not an instead of situation. AL32UTF8 is a valid setting for the database character set, which controls CHAR and VARCHAR2 columns. AL16UTF16 is a valid setting for the national character set which controls NCHAR and NVARCHAR2 columns.
    Could you tell me the difference?The difference between the two encodings comes down to how many bytes are required to store a particular code point (character). AL32UTF8 is a variable-length character set, so 1 character will require between 1 and 3 bytes of storage (4 for the supplemental characters but those are rather rare). AL16UTF16 is a fixed-width character set, so 1 character will require 2 bytes of storage (4 for the rare supplemental characters again).
    Also could you tell me the difference between
    WE8ISO8859P15 and WE8ISO8859P1 ? There's a Wikipedia article that discusses the differences and has links to the two different code tables.
    Werner's point is an excellent one as well. I was assuming that we were talking about how to set up both sides of this proposed system. If the source system already exists, there are additional considerations like ensuring that your target system supports a superset of the characters supported by the source system. Regardless, when doing imports & exports, as Werner points out, you need to ensure that NLS_LANG is set appropriately.
    Justin

  • Foreign character set issue

    Hi
    This might sound a bit silly, but stay with me.
    There's a database that decidedly supports UTF 8. I checked using this query
    select * from nls_database_parameters where parameter like '%CHARACTERSET'
    And got this result
    NLS_CHARACTERSET     UTF8
    NLS_NCHAR_CHARACTERSET     AL16UTF16
    It's Oracle 10g.
    In a particular table, some text is stored in multiple languages. There are seven languages (English, Mandarin, Japanese, German...). Every language has 2-3 rows to itself. There's a column where I have to get rid of the trailing few characters, the number of which depends on the content of the string.
    But I cannot see any of the Eastern languages in TOAD. The column is in VARCHAR2.
    My problem is twofold.
    1. What functions do I use to ensure that only the last few bytes are truncated, and there's no data loss (which many websites gravely warn of when dealing with foreign language data) ?
    2. How can I see this foreign language text in TOAD/SQLPlus?
    (Yes, I'm kind of new to the whole multiple-language-game. Please let me know if I've left out any important detail!)

    Do you have metalink access?
    If so, please see the notes below, there's a lot of good information in them:
    158577.1 - NLS_LANG explained
    260893.1 - Unicode Character Sets in the Database
    788156.1 - UTF8 implications
    With any character set situation there are at least two and a bit sides to the equation.
    First is whether you are storing the correct data.
    You're best using the DUMP function to inspect the stored data, e.g.
    SELECT DUMP(<column_name>) FROM <table_name> WHERE ....This function may help you with your truncation of the last few bytes - not sure why you need to do this?
    The "second and a bit" bit is having the correct client settings - NLS_LANG - and using a client which supports the characters required.
    SQL*Plus has it's limitations here. Toad I don't know well enough but it should support full UTF8 characters.
    SQL*Developer and iSQL*Plus both should support the full UTF8 - I tend to use the former, particularly for UTF8.

  • 9i Character Set Issues

    Hi All,
    I have a done a search through the forums on this topic, and revealed many answers to my question, however I need some more information to help me resolve an Issue.
    Here goes...
    Our company has a French/Canadian client that has installed 9i with the characterset of WE8DEC. The software that we wrote and provided them with does not support this characterset, and in turn they are seeing inverted question marks in place of some characters (Usually accented characters).
    Basically, I want to understand the options that our client has in terms of possible solutions.
    I understand that the characterset can only be altered in the database to a superset of the current characterset, so this affords us no option in this regard.
    My suggestions so far, is to create a new database and transfer the data accross, however will this actually cure the issue? Will the data transfer accross and convert into it's correct character? or will it just transfer as an inverted "?".
    Would an upgrade to 10g help them? Or would the issue remain?
    Is there anything that can be done on the O/S or Client level to rectify this?
    All help greatly appreciated!
    Thanks.

    user10498503 wrote:
    Hi All,
    I have a done a search through the forums on this topic, and revealed many answers to my question, however I need some more information to help me resolve an Issue.
    Here goes...
    Our company has a French/Canadian client that has installed 9i with the characterset of WE8DEC. The software that we wrote and provided them with does not support this characterset, and in turn they are seeing inverted question marks in place of some characters (Usually accented characters).
    Basically, I want to understand the options that our client has in terms of possible solutions.
    I understand that the characterset can only be altered in the database to a superset of the current characterset, so this affords us no option in this regard.
    My suggestions so far, is to create a new database and transfer the data accross, however will this actually cure the issue? Will the data transfer accross and convert into it's correct character? or will it just transfer as an inverted "?".Your issue has a client side component (the NLS_LANG setting already mentioned here) and a potential server side component (corrupted/invalid data in the database), depending on what happened so far.
    The client side issue could lead to corrupted data in the database, and there is no way of "correcting" this automatically, even if you move to a new database using a different character set on 9i, or upgrade to 10g.
    If a client that writes into the database uses a wrong NLS_LANG setting, you're potentially ending up with corrupted data in the database.
    If a client that reads data from the database uses a wrong NLS_LANG setting, you're potentially ending up with corrupted data on the front end application (the typical "?" question mark).
    All this has nothing to do yet if your database character set is actually capable of storing the characters you attempt to store from the client side.
    So I think you first need to determine if your database contains invalid data using the "CSSCAN" utility. For more information, check the manuals and MetaLink:
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96529/ch11.htm
    Depending on your findings you'll find the possible actions on MetaLink.
    Then you should determine if your client NLS_LANG is appropriate. There is a very good FAQ available here how to check the current setting and determine the correct setting:
    http://www.oracle.com/technology/tech/globalization/htdocs/nls_lang%20faq.htm
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • German character set issues on Solaris

    Hi,
    I am facing an issue with German character settings with my Java application on a solaris box.
    When I run my application on the box, and I pass an input file with German special characters they get converted as ?. However, other normal English characters are formed properly.
    When I run the same application on another Solaris box with a different JRE, the German characters are formed properly.
    I understand that there is a difference in the archiecture between the 2 boxes ie.e
    64 bit SPARC machine v/s 32 bit x86 machine
    the JRE
    1.4.2_03(64bit) v/s 1.4.1_01
    I am tryinbg to evaludate further differences between the 2 environments to pinpoint the issue, and get this resolved on the 1st box.
    Can anyone provide me any inputs?
    Lavin

    When you read the file, please point out what character set using. For example:
    FileInputStream fstream = new FileInputStream(url.getFile());
    DataInputStream in = new DataInputStream(fstream);
    BufferedReader br = new BufferedReader(new InputStreamReader(in, Charset.forName("ISO-8859-1")));
    br.readLine();
    This link possibly can help you.
    http://www.velocityreviews.com/forums/t126128-jdk-14-character-set-change.html

  • Arabic Character set conversion-help needed

    We have our main database running in 10g (Solaris o/s) & planning to move these to RAC 11g.
    One of our old oracle DB(8.0.5)/solaris, which is not used till recently need to upgrade to 10g Rel2.
    I know Supported direct upgrade 8.0.6/8.1.7/9i -> 10g
    Current DB: 8.0.5 (Character Set: AR8ISO8859P6)
    Target DB : 10g Rel 2 (Character Set: AR8MSWIN1256)
    I am thinking to go the following way by using exp/imp
    8.0.5(AR8ISO8859P6) -> 8.1.7(AR8ISO8859P6) -> 10G(AR8MSWIN1256)
    OR
    8.0.5(AR8ISO8859P6) -> 8.1.7(AR8MSWIN1256) -> 10G(AR8MSWIN1256)
    please advice
    thanks

    (1) At source db 8.0.5 (solaris)
    PARAMETER VALUE
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-YY
    NLS_DATE_LANGUAGE AMERICAN
    NLS_CHARACTERSET AR8ISO8859P6
    NLS_SORT BINARY
    NLS_NCHAR_CHARACTERSE T AR8ISO8859P6
    $set NLS_LANG=AMERICAN_AMERICA.AR8ISO8859P6
    $exp sys/dba file=full251109.dmp full=y
    (2):>> At target db 10g R2 (solaris)
    PARAMETER VALUE
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RRRR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_CHARACTERSET AR8ISO8859P6
    NLS_SORT BINARY
    NLS_NCHAR_CHARACTERSET UTF8
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    $export NLS_LANG=AMERICAN_AMERICA.AR8ISO8859P6
    $imp testdba/testdba file=full251105.dmp fromuser=PROFINAL touser=PROFINAL
    *$csscan testdba/testdba FULL=Y FROMCHAR=AR8ISO8859P6 TOCHAR=AR8ISO8859P6 LOG=P6check CAPTURE=Y ARRAY=100000 PROCESS=2*
    There is EXCEPTIONAL DATA in .err file+
    clients accessing 8.0.5 dataabase uses characterset AR8IS08859P6, which is SAME as 8.0.5 database
    -CSSCAN result->>[Database Scan Parameters]
    Parameter Value
    CSSCAN Version v2.1
    Instance Name MIG1
    Database Version 10.2.0.1.0
    Scan type Full database
    Scan CHAR data? YES
    Database character set AR8ISO8859P6
    FROMCHAR AR8ISO8859P6
    TOCHAR AR8ISO8859P6
    Scan NCHAR data? NO
    Array fetch buffer size 100000
    Number of processes 2
    Capture convertible data? YES
    [Scan Summary]
    Some character type data in the data dictionary are not convertible to the new
    haracter set Some character type application data are not convertible to the new characters
    [Data Dictionary Conversion Summary]
    Datatype Changeless Convertible Truncation Lossy
    VARCHAR2 2,235,403 0 0 *1,492*
    CHAR 1,097 0 0 0
    LONG 155,188 0 0 6
    CLOB 24,643 0 0 0
    VARRAY 21,352 0 0 0
    Total 2,437,683 0 0 1,498
    Total in percentage 99.939% 0.000% 0.000% 0.061%
    The data dictionary can not be safely migrated using the CSALTER script
    [Application Data Conversion Summary]
    Datatype Changeless Convertible Truncation Lossy
    VARCHAR2 16,986,594 0 0 *1,240,383*
    CHAR 164,114 0 0 0
    LONG 7 0 0 0
    CLOB 1 0 0 0
    VARRAY 1,436 0 0 0
    Total in percentage 93.256% 0.000% 0.000%
    6.744%
    [Distribution of Convertible, Truncated and Lossy Data by Table]
    USER.TABLE Convertible Truncation Lossy
    PROFINAL.BASE_MASTER_DATAS 0 0 *362,003*
    PROFINAL.CODE_ALLOW 0 0 *53*
    PROFINAL.CODE_ALLOWANCE_TYPES 0 0 *1*
    PROFINAL.CODE_BONUS_TYPES 0 0 *2*
    PROFINAL.CODE_BRANCHES 0 0 *2*
    PROFINAL.CODE_CERTIFICATES 0 0 *94*
    Kindly help,,,
    Edited by: userR12 on Nov 25, 2009 1:43 AM
    Edited by: userR12 on Nov 25, 2009 1:52 AM

  • (urgent) SQL*Loader Large file support in O734

    hi there,
    i have the following sqlloader error when trying to upload data file(s),
    each has size 10G - 20G to Oracle 734 DB on SunOS 5.6 .
    >>
    SQL*Loader-500: Unable to open file (..... /tstt.dat)
    SVR4 Error: 79: Value too large for defined data type
    <<
    i know there's bug fix for large file support in Oracle 8 -
    >>
    Oracle supports files over 2GB for the oracle executable.
    Contact Worldwide Support for information about fixes for bug 508304,
    which will add large file support for imp, exp, and sqlldr
    <<
    however, really want to know if any fix for Oracle 734 ?
    thx.

    Example
    Control file
    C:\DOCUME~1\MAMOHI~1>type dept.ctl
    load data
    infile dept.dat
    into table dept
    append
    fields terminated by ',' optionally enclosed by '"'
    trailing nullcols
    (deptno integer external,
    dname char,
    loc char)
    Data file
    C:\DOCUME~1\MAMOHI~1>type dept.dat
    50,IT,VIKARABAD
    60,INVENTORY,NIZAMABAD
    C:\DOCUME~1\MAMOHI~1>
    C:\DOCUME~1\MAMOHI~1>dir dept.*
    Volume in drive C has no label.
    Volume Serial Number is 9CCC-A1AF
    Directory of C:\DOCUME~1\MAMOHI~1
    09/21/2006  08:33 AM               177 dept.ctl
    04/05/2007  12:17 PM                41 dept.dat
                   2 File(s)          8,043 bytes
                   0 Dir(s)   1,165 bytes free
    Intelligent sqlldr command
    C:\DOCUME~1\MAMOHI~1>sqlldr userid=hary/hary control=dept.ctl
    SQL*Loader: Release 10.2.0.1.0 - Production on Thu Apr 5 12:18:26 2007
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Commit point reached - logical record count 2
    C:\DOCUME~1\MAMOHI~1>sqlplus hary/hary
    SQL*Plus: Release 10.2.0.1.0 - Production on Thu Apr 5 12:18:37 2007
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    As I am appending I got two extra rows. One department in your district and another in my district :)
    SQL> select * from dept;
        DEPTNO DNAME          LOC
            10 ACCOUNTING     NEW YORK
            20 RESEARCH       DALLAS
            30 SALES          CHICAGO
            40 OPERATIONS     BOSTON
            50 IT             VIKARABAD
            60 INVENTORY      NIZAMABAD
    6 rows selected.
    SQL>

  • Urgent: SQL*Loader-562: record too long

    with sqlldr73 , no problem
    but with sqlldr of 817 : i've got the problem !!??
    any help please ...
    Mourad from Paris

    Hi Sandeep,
    Oracle guru Dave Moore has many sample SQL*Loader control files published:
    http://www.google.com/search?&q=oracle+moore+sql%2aloader
    Here is a simple sample control file to load many tables:
    http://www.dba-oracle.com/t_sql_loader_multiple_tables_sqlldr.htm
    Hope this helps. . .
    Donald K. Burleson
    Oracle Press author

  • Sql Loader - Concurrent request id issue!

    Hello there:
    I want to load request id in my sql loader script please advice how can I do.
    Below is my scipt file.
    LOAD DATA
    INFILE '${CINT_TOP}/uploads/f_cost/inprocess.dat'
    BADFILE '/tmp/inprocess.CFL.bad'
    INTO TABLE cost_temp REPLACE
    FIELDS TERMINATED BY ',' OPTIONALLY ENCLOSED BY "'"
    (ITEM_COST DECIMAL EXTERNAL,
    CURRENCY CHAR,
    EFFECTIVITY_DATE DATE 'DD-MON-YYYY',
    ITEM_NUMBER CHAR,
    ORGANIZATION_ID CHAR,
    COST_TYPE_ID CHAR,
    SEQUENCE_ID CHAR
    Data file:
    51.24,USD,'30-JUL-2007',6346339,304,1001,fnd_global.conc_request_id
    100.5,USD,'30-JUL-2007',299150,304,1001,fnd_global.conc_request_id
    Thanks,
    Zeeshan

    I did get same error in my log file.
    LOAD DATA
    INFILE '${CINT_TOP}/uploads/f_cost/inprocess.dat'
    BADFILE '/tmp/inprocess.CFL.bad'
    INTO TABLE cost_temp REPLACE
    FIELDS TERMINATED BY ',' OPTIONALLY ENCLOSED BY "'"
    (ITEM_COST          DECIMAL EXTERNAL,
    CURRENCY          CHAR,
    EFFECTIVITY_DATE DATE 'DD-MON-YYYY',
    ITEM_NUMBER     CHAR,
    ORGANIZATION_ID CHAR,
    COST_TYPE_ID CHAR,
    SEQUENCE_ID     CHAR "fnd_global.conc_request_id"
    Column SEQUENCE_ID had SQL string
    "fnd_global.conc_request_id"
    applied to it.
    Record 1: Rejected - Error on table "CINT"."COST_TEMP", column SEQUENCE_ID.
    Column not found before end of logical record (use TRAILING NULLCOLS)
    Record 2: Rejected - Error on table "CINT"."COST_TEMP", column SEQUENCE_ID.
    Column not found before end of logical record (use TRAILING NULLCOLS)
    Table "CINT"."COST_TEMP":
    0 Rows successfully loaded.
    2 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    I need you advice please.
    Regards,
    Zeeshan

  • Character set issue after import?

    Hi,
    Source DB version:10.2.0.1
    OS:Red hat Linux
    Target DB version:10.2.0.1
    OS:Windows server
    source database character set:AL32UTF8
    Performed the export as below
    $export NLS_LANG=AMERICAN.AL32UTF8
    Performed the full database export and it finished successfully with out any warnings
    Export done in AL32UTF8 character set and AL16UTF16 NCHAR character set
    Now imported into the target database as below.
    target database character set:AL32UTF8
    c:\>set NLS_LANG=AMERICAN.AL32UTF8
    now run import command which imported successfully with out any warnings.
    However I’m having problems with Greek characters. Most of them are shown as ?, while some of them are converted to Latin chars
    For example:
    This was supposed to be Αγγελική ???e????
    And this Κουκουτσάκη ??????ts???
    While this one should be Δήμητρα ??µ?t?a
    From the import log file I can see that ‘import done in AL32UTF8 character set and AL16UTF16 NCHAR character set’ which I believe is correct.
    Can any one tell me how i can over come this problem of greek charecters.
    Thank you all.

    PARAMETER
    VALUE
    NLS_LANGUAGE
    AMERICAN
    NLS_TERRITORY
    AMERICA
    NLS_CURRENCY
    $
    PARAMETER
    VALUE
    NLS_ISO_CURRENCY
    AMERICA
    NLS_NUMERIC_CHARACTERS
    NLS_CHARACTERSET
    AL32UTF8
    PARAMETER
    VALUE
    NLS_CALENDAR
    GREGORIAN
    NLS_DATE_FORMAT
    DD-MON-RR
    NLS_DATE_LANGUAGE
    AMERICAN
    PARAMETER
    VALUE
    NLS_SORT
    BINARY
    NLS_TIME_FORMAT
    HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT
    DD-MON-RR HH.MI.SSXFF AM
    PARAMETER
    VALUE
    NLS_TIME_TZ_FORMAT
    HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT
    DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY
    $
    PARAMETER
    VALUE
    NLS_COMP
    BINARY
    NLS_LENGTH_SEMANTICS
    BYTE
    NLS_NCHAR_CONV_EXCP
    FALSE
    PARAMETER
    VALUE
    NLS_NCHAR_CHARACTERSET
    AL16UTF16
    NLS_RDBMS_VERSION
    10.2.0.1.0
    20 rows selected.

Maybe you are looking for

  • Quicktime is 'choppy' when playing MP4's

    I just recently purchased an HD camcorder and am having trouble playing an MP4 file in Quicktime. Windows Media player wont play the format at all and while I have no problem using Quicktime (I actually like the player better) the playback is 'choppy

  • How to share one itune library among 3 computer at home

    I have 2 desktop and one laptop in my home and would like to know how to setup so that all 3 computers uses the same itune library

  • How to use csv file to populate all the items in a form

    Our environment - Forms 6i/IAS 9i on Solaris/8.1.6 database on Solaris. In the web form I want the user to type in the name of an existing csv file(which contains 1 record). Then press a button and all the items on the form should be populated with t

  • Desktop.open() Problem

    I recently put together an application that relied heavily on the Desktop.open( ) method and several other methods from the Desktop class. However, all of a sudden the Desktop.open() does not work. I believe I read some where that this was due to the

  • Intrastat wrong file format

    Hi Gurus, I am customizing the INTRASTAT for Spain and everything works fine until the moment we create the file. This file should have different lines for every different information in the document, but what the system is doing is creating the file