Non-ascii character problem

hi
Scenario: Reading Flat-File and writing into Oralce Table.
In my database procedure, I have declared a column say X as VARCHAR2 to store string values. In the flat-file, some times the data for column X comes as Non-ASCII string values (e. SäHKOTOIM) and because of this the database procedure raises ORA-06502:PL/SQL:numeric or value error.
So, how I can identify that the flat file has non-ascii values so that I can reject that record and move with another record?
your suggestion will be greatly appreciated.
Regards
shakeel

Hi,
You set you nls_database_parameters which is compatiable with you input (non- ascii) characters.
Try to use the LCSSCAN in order to know the relevant character set with respective to input and set the nls_parameters
https://students.kiv.zcu.cz/doc/oracle/server.102/b14225/ch12scanner.htm#i1016925
- Pavan Kumar N

Similar Messages

  • Non-ASCII character in Email field

    Hi Guys,
    I am trying to enter non-english characters in Email field of user form, but OIM throws an error that "A non-Ascii character has been entered". I have also tried to turn off the AppFirewall Filter in xlConfig.xml file but no help. Is there any way thay I can enter non-Ascii characters in Email field?
    Regards,
    Rahul

    .oO(surfinIan)
    >I have a script that converts a ms word document to text
    then uploads that to a
    >blob field on a mysql db.
    > During the conversion some characters my not be
    recognised. When i then call
    >up the blob for display on the browser...those characters
    show up as unknown
    >characters with a ? or box. Is there a way to
    preg_replace those unknown
    >characters before displaying them.
    What about fixing the encoding problem instead? If chars get
    lost during
    such a transfer
    document->script->database->script->browser it's always
    an encoding problem somewhere down the road.
    The recommendation these days is to use UTF-8, which avoids
    most of
    these old problems. You just have to make sure that your
    documents are
    properly stored as UTF-8 in the database and delivered as
    such to the
    script and the browser, then you don't have to worry about
    special chars
    anymore.
    That's just the general idea. I can't be more specific, since
    I don't
    know your conversion script or the database structure.
    Micha

  • [Solved] no non-ASCII character input in rxvt-unicode

    Hello everyone,
    For some days now, I can't write any non-ASCII characters any more in rxvt-unicode and rxvt-unicode-patched. Unfortunately, downgrading the rxvt-unicode package doesn't seem to help. To have at least a temporary solution, I'd like to know at least which packages I could try to downgrade as well. Any ideas, anyone?
    greez,
    maxmin
    Last edited by Maximalminimalist (2011-03-12 13:12:26)

    When I try to type a non-ASCII-character I get nothing at all. This happens with my custom keyboard layout (modified programmer dvorak) and in some layouts I already tried (us: altgr-intl, ch, de and fr)
    When I paste a non-ASCII characters in rxvt-unicode I get
    maxmin ~ $ ?
    This happens only on my x86_64 desktop which is more up to date than my i686 laptop. (I'm afraid now to do any updates.)
    EDIT: I'm sorry, I don't know what you mean with locale settings. What do you mean with that?
    EDIT2: Maybe just typing locale in the terminal is what you mean:
    maxmin ~ $ locale
    locale: Cannot set LC_CTYPE to default locale: No such file or directory
    locale: Cannot set LC_MESSAGES to default locale: No such file or directory
    locale: Cannot set LC_ALL to default locale: No such file or directory
    LANG=en_US.utf8
    LC_CTYPE="en_US.utf8"
    LC_NUMERIC="en_US.utf8"
    LC_TIME="en_US.utf8"
    LC_COLLATE="en_US.utf8"
    LC_MONETARY="en_US.utf8"
    LC_MESSAGES="en_US.utf8"
    LC_PAPER="en_US.utf8"
    LC_NAME="en_US.utf8"
    LC_ADDRESS="en_US.utf8"
    LC_TELEPHONE="en_US.utf8"
    LC_MEASUREMENT="en_US.utf8"
    LC_IDENTIFICATION="en_US.utf8"
    LC_ALL=
    With other terminal emulators I get sometimes also nothing and sometimes right displayed but wrong interpreted character in vim. I didn't take notes while doing that but I'll try again if needed.
    Last edited by Maximalminimalist (2011-03-06 21:51:23)

  • ALV Grid bug when dealing with non-ASCII character

    Dear all,
    I have a requirement to display user's remarks on ALV.  The data element of the remarks column is TEXT200.  I know that each column in an ALV Grid can display at most 128 characters.  Since my SAP is an Unicode system, I expect that each column in my ALV Grid can display 128 Chinese characters, too.  However, the ALV Grid only display 42 Chinese characters at most.  Is this a bug in ALV Grid?  How can I fix it?
    I did a small experiment.  The results are listed below.  My version is Net Weaver 7.01.  The results show that the bug does not exist in ALV List.  However, my user prefers ALV Grid, which is more beautiful and elegant.
    Type of ALV
    Max number of
    ASCII character
    in an ALV column
    Max number of
    non-ASCII character
    in an ALV column
    REUSE_ALV_GRID_DISPLAY
    128
    42 Chinese characters
    CL_SALV_TABLE
    128
    42 Chinese characters
    CL_GUI_ALV_GRID
    128
    42 Chinese characters
    REUSE_ALV_LIST_DISPLAY
    132
    132 Chinese characters
    If you encounter the bug, please post your solution.  Thanks a lot. 

    It looks like limitation of ALV grid cell, which can contain up to 128 bytes in SAP gui.
    Your unicode characters are probably 3 bytes each.
    Check OSS Note 910300 for more detailed info.
    EDIT: Note 1401711 seems to be a correction for your issue It allows to use 128 characters (even if they take more than 128 bytes).

  • Is Linksys WRT54GH SSID can contains the non-ascii character?

    is Linksys WRT54GH SSID can contains the non-ascii character?
    we need to use it for our wireless testing, but i dont know if the SSID can contains non-ascii.
    anybody can help me? hurry, i will wait answer online.
    thanks in advance!
    Solved!
    Go to Solution.

    thank you  very much, Ricewind
    SSID cant contain non-ascii characters, it make me sad and disappointed
    why we can  set T-link router SSID with non-ascii characters?

  • Unicode value of a non-ASCII character

    Hi,
    Suppose, the unicode value of the character ् is '\u094d'.
    Is there any Java function which can get this unicode value of a non-ASCII character.
    Like:
    char c='्';
    String s=convertToUnicode(c);
    System.out.println("The unicode value is "+ s);
    Output:
    The unicode value is \u094d
    Thanks in advance

    Ranjan_Yengkhom wrote:
    I have tried with the parameter
    c:\ javac -encoding utf8 filename.java
    Still I am getting the same print i.e. \u3fIf it comes out as "\u3f" (instead of failing to compile or any other value), then your source code already contains the question mark. So you already saved it wrong and have to re-type it (at least the single character).
    >
    Then I studied one tutorial regarding this issue
    http://vietunicode.sourceforge.net/howto/java/encoding.html
    It says that we need to save the java file in UTF-8 format. I have explored most of the editors like netbean, eclipse, JCreator, etc... there is no option to save the java file in UTF-8 format.That's one way. But since that is so problematic (you'll have to remember/make sure to always save it that way and to compile it using the correct switch), the better solution by far is not to use any non-ASCII characters in your source code.
    I already told you two possible ways to achieve that: unicode escapes or externalized strings.
    Also read http://www.joelonsoftware.com/articles/Unicode.html (just because it's related, essential information and I just posted that link somewhere else).

  • Remove non ascii character

    i need a SQL or Procedure that will search non ascii character  in data and update the data by removing it
    Suppose there is table TABLE1 with Column NAME
    it contain number of row and few has non ascii character eg 'CharacterÄr'
    My sql or procedure should be able to search  'CharacterÄr' and update the row with 'Character'
    i.e. removing the non ascii character 'Ä' from the data

    Hi,
    Okay, in that case:
    SELECT str
    ,      REGEXP_REPLACE ( str
                          , '[^[:cntrl:] -~]'
                          )   AS new_str
    FROM    table_x
    or, to actually change the rows that contain the bad characters:
    UPDATE  table_x
    SET     str = REGEXP_REPLACE ( str
                                 , '[^[:cntrl:] -~]'
    WHERE   REGEXP_LIKE ( str
                        , '[^[:cntrl:] -~]'

  • Find the Special character and non Ascii character

    Hi:
    i have table ,this table column name contain some datas like
    sno name
    1 CORPORATIVO ISO, S.A. DE C.V.
    2 (주)엠투소프트
    3 TIMELESS
    4 南京南瑞集团公司
    5 PHOTURIS
    6 Ace Informática S/C ltda
    7 Computacenter AG & Co. oHG
    8 아이티앤씨
    9 MOCA
    10 anbarasan
    my requirement:
    1)i need to search the name column where contain the special character and non ascii character..if found any non ascii or spcial character ..need to say flag ''yes".if not found need to say "no"...kindly help on this issus...

    i need some example..i am not have any idea....
    i have table ,this table column name contain some datas like
    sno name
    1 CORPORATIVO ISO, S.A. DE C.V.
    2 (주)엠투소프트
    3 TIMELESS
    4 南京南瑞集团公司
    5 PHOTURIS
    6 Ace Informática S/C ltda
    7 Computacenter AG & Co. oHG
    8 아이티앤씨
    9 MOCA
    10 anbarasan
    my requirement:
    1)i need to search the name column where contain the special character and non ascii character..if found any non ascii or spcial character ..need to say flag ''yes".if not found need to say "no"...kindly help on this issus...

  • Keyword import fails on non-ascii character

      I recently tried to import a long set of keywords (about 4000 terms).  i set up the file in excel and then tried to import the records.  I kept getting this message:  
    only text files encoded with ascii or unicode UTF-8 are supported when importing keywords.
    I finally tracked down the problem when i converted the file to a MS word text file, broke it down into parts and eventually found the problem record.  for some reason, the apostrophe in the words "don't know" had been corrupted to a weird character.  after i corrected this, everything worked. 
    however, this took a long time.   It would have been helpful if  lightroom could have at least pinpointed the line where the import failed or offered to convert non-compliant charaters to some specific character or set of characters.

    Yeah, that didn't work so well since SuperDuper ran across repeated errors trying to do so; I suspect it's something to do with the drive. (SuperDuper complains about WD's MyBook, which is what the drive is.) Because SD stops the entire copy operation on single errors, it'd be a painstaking process.
    Besides that, I like doing fresh installs of all the bits.

  • Printer name with non-ASCII character

    Is it possible to name a printer using ISO-8859-1 (or WINLATIN1) characters?
    Our business system has logical printer names defined, sometimes using Norwegian (and Danish) letters (AE-lig, O-slash and A-ring), which we would like to map to the correct Windows UNC paths.
    Mapping to a UNC path containing one of these characters works fine. But if the printer name itself contains a non-ASCI ISO character it fails.
    Example using BÆR (B AE-lig R):
    !p BÆR * \\print\bærum * 50 *
    Don't know how the log will look after posting, but I'll try:
    jfserver.exe: [400]D:\Adobe\Central\Server\Data\Test.dat : ^job TEST -zBÆR -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos"
    jfserver.exe: [400] ^JOB formname='TEST' printername='B'
    jfserver.exe: [400]Scanning !f for candidates
    jfserver.exe: [400]Mapped printer name 'B' to id 'R' dev '*'
    jfserver.exe: [400]Other job tokens before substitution 'R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos"'
    jfserver.exe: [400]Other job tokens after substitution'R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos"'
    jfserver.exe: [400]Merge parms before substitution '\\printer\b'
    jfserver.exe: [400]Merge parms after substitution'\\printer\b'
    jfserver.exe: [400]Task parms before substitution '"test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'
    jfserver.exe: [400]Task parms after substitution '"test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'
    jfserver.exe: [400]"test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"
    jfserver.exe: [307]Launching task '"D:\Adobe\Central\Bin\jfmerge" "test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'.
    jfmerge: [125]* Processing data file: 'D:\Adobe\Central\Server\Data\Test.dat'.
    jfmerge: [289]MDF file `D:\Adobe\Central\Samples\Exprint\forms\test.mdf' opened.
    jfmerge: [400]The filename, directory name, or volume label syntax is incorrect.
    jfmerge: [2]Error opening output device/file '*'.
    jfmerge: [210]Nothing was printed.
    jfserver.exe: [314]Agent exit message: [2]Error opening output device/file '*'.
    jfserver.exe: [400]spawned '"D:\Adobe\Central\Bin\jfmerge" "test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'
    jfserver.exe: [400]--> returned [-2] [2]Error opening output device/file '*'..

    For others who might have the same problem. You can solve it by quoting the device name and/or printer name.
    !p """BÆR""" * """\\print\bærum""/q" * 50 *
    This, however, causes the printer name to be '"BÆR"' with the double quotes included in the name. Thus, the job card must contain the unfortunate tripple quotes as well: -z"""BÆR""".
    Vegard

  • Convert non-ASCII character to decimal value

    Hi all,
    I have the following problem:
    When reading a String containing the full spectrum of UTF-8 characters from a BufferedReader:
    How can I get the decimal value of the String's characters outside the ASCII range?
    When I try to do a MyBufferedReader.read(), I get the int value 65533 for these characters, instead of their real value in the UTF-8 set.
    Can anyone help?
    Thanks!
    Dre

    That's the character you get when you try to decode a character with no valid decoding. Of course there aren't any such characters for UTF-8. Therefore whatever you did, don't do that. More specifically: you've already got a String, don't do anything which involves converting it to bytes. That's completely unnecessary.
    Or perhaps the problem is the way you create the String in the first place. Who knows?

  • Extended ascii character problem

    Here is the message I am getting
    2004-10-12 10:04:44,207 ERROR [STDERR] org.postgresql.util.PSQLException: Invalid character data was found. This is most likely caused by stored data containing characters that are invalid for the character set the database was created in. The most common example of this is storing 8bit data in a SQL_ASCII database.
    this is the character that is causing the problem (ascii 0232 �)
    My question is how can I make java display this character. I am using Postgresql 7.4 as my database.

    I don't understand the connection between the database and "java display".
    And ASCII is a 7 bit character set, so the character you listed is not ASCII, rather it is a character in some other character set.

  • Extended ASCII character problems

    We are having problems with certain characters. If a character in the range ASCII 194-239 appears in a string it can cause the next character to be lost. Everything seems okay in SQL, but not in PL/SQL.
    A PL/SQL script to print all the ASCII characters outputs nothing for the above range.
    e.g.
    DECLARE
    v VARCHAR2(100):= 'Requiem por un PolicÃ|xxx|qqq';
    BEGIN
         dbms_output.put_line(instr(v,'|'));
         dbms_output.put_line(substr(v,instr(v,'|')));
    END;
    The expected results would be 22 and |xxx|qqq, but we actually get 25 and |qqq. The first '|' is being lost somehow.
    A test using a Java function retrieving the string from a table failed due to a conversion failure between UTF8 and UCS2.
    The NLS language is AMERICAN and the character set is UTF8.
    Thanks
    Steve

    I know this thread is old, but it took me forever to find this out...
    Use:
    SELECT CHR (<ASCII Value> USING NCHAR_CS) FROM DUAL;
    E.g.
    SELECT CHR (206 USING NCHAR_CS) FROM DUAL;
    CHR (206 USING NCHAR_CS)
    ÎEdited by: martinByrne on 04-Aug-2010 08:37

  • Non-English character problem in Oracle 10g Express Edition

    Hi There;
    I have a table. It's name is INSTITUTION. It has a NUMBER INS_ID and NVARCHAR2(50) INS_NAME . INS_NAME can contain Turkish characters, such as "ğ,ü,ş,ç,ö". According to business logic, there can not be a repetition on the INS_NAME.
    User will enter institution name from a textbox in ASP.NET , and I check this name in database from c sharp code, if there is no repetition, we will add this record.
    The problem is; when user enter a instition name that contains Turkish character, there is a duplication. If there is a instition name is *"su işleri"* , the both query; SELECT * FROM INSTITUTION WHERE INS_NAME = *'su işleri'*; and SELECT * FROM INSTITUTION WHERE INS_NAME = *'su isleri'*; returns no result, even though there it is.
    But if instition name is "oracle corporation" (there is no Turkish character) it query successfully. I have the same problem in Toad for Oracle 11.5.1.2. When I query database from toad SELECT * FROM INSTITUTION, the phrase *"su işleri"* has appeared. But when I query SELECT * FROM INSTITUTION WHERE INS_NAME = *'su işleri'*; , there is again no result.
    When I connect oracle database directly and perform the query SELECT * FROM INSTITUTION , the phrase *"su isleri"* (not *"su işleri"* ) has appeared.
    Here are the language settings of the database:
    National Language Support
    National Language Parameter Value
    NLS_CALENDAR______________GREGORIAN
    NLS_CHARACTERSET__________WE8MSWIN1252
    NLS_COMP__________________BINARY
    NLS_CURRENCY______________TL
    NLS_DATE_FORMAT__________DD/MM/RRRR
    NLS_DATE_LANGUAGE________TURKISH
    NLS_DUAL_CURRENCY_________YTL
    NLS_ISO_CURRENCY__________TURKEY
    NLS_LANGUAGE______________TURKISH
    NLS_LENGTH_SEMANTICS______BYTE
    NLS_NCHAR_CHARACTERSET___AL16UTF16
    NLS_NCHAR_CONV_EXCP______FALSE
    NLS_NUMERIC_CHARACTERS____ ,.
    NLS_SORT___________________TURKISH
    NLS_TERRITORY______________TURKEY
    NLS_TIME_FORMAT____________HH24:MI:SSXFF
    NLS_TIMESTAMP_FORMAT_______DD/MM/RRRR HH24:MI:SSXFF
    NLS_TIMESTAMP_TZ_FORMAT____DD/MM/RRRR HH24:MI:SSXFF TZR
    NLS_TIME_TZ_FORMAT__________HH24:MI:SSXFF TZR
    How can I resolve that problem? Thanks in advance.
    Edited by: 963344 on 05.Eki.2012 01:00
    Edited by: 963344 on 05.Eki.2012 01:01
    Edited by: 963344 on 05.Eki.2012 01:06

    This type of question/discussion belongs in {forum:id=50} forum.
    Very recently a thread there touched the topic of Turkish character support.
    Please read it: Western European Characterset to Turkish in sql
    >
    NLS_CHARACTERSET__________WE8MSWIN1252 Check the character set repertoire of win-1252 (look for the typical turkish language characters you've mentioned above).
    http://msdn.microsoft.com/en-us/goglobal/cc305145.aspx
    Look at character names, such as "... letter s with cedilla".

  • Php / mysql  replace non-ascii character in a string

    I have a script that converts a ms word document to text then
    uploads that to a blob field on a mysql db.
    During the conversion some characters my not be recognised.
    When i then call up the blob for display on the browser...those
    characters show up as unknown characters with a ? or box. Is there
    a way to preg_replace those unknown characters before displaying
    them.
    thanks
    ian

    .oO(surfinIan)
    >I have a script that converts a ms word document to text
    then uploads that to a
    >blob field on a mysql db.
    > During the conversion some characters my not be
    recognised. When i then call
    >up the blob for display on the browser...those characters
    show up as unknown
    >characters with a ? or box. Is there a way to
    preg_replace those unknown
    >characters before displaying them.
    What about fixing the encoding problem instead? If chars get
    lost during
    such a transfer
    document->script->database->script->browser it's always
    an encoding problem somewhere down the road.
    The recommendation these days is to use UTF-8, which avoids
    most of
    these old problems. You just have to make sure that your
    documents are
    properly stored as UTF-8 in the database and delivered as
    such to the
    script and the browser, then you don't have to worry about
    special chars
    anymore.
    That's just the general idea. I can't be more specific, since
    I don't
    know your conversion script or the database structure.
    Micha

Maybe you are looking for