Non-ascii character problem
hi
Scenario: Reading Flat-File and writing into Oralce Table.
In my database procedure, I have declared a column say X as VARCHAR2 to store string values. In the flat-file, some times the data for column X comes as Non-ASCII string values (e. SäHKOTOIM) and because of this the database procedure raises ORA-06502:PL/SQL:numeric or value error.
So, how I can identify that the flat file has non-ascii values so that I can reject that record and move with another record?
your suggestion will be greatly appreciated.
Regards
shakeel
Hi,
You set you nls_database_parameters which is compatiable with you input (non- ascii) characters.
Try to use the LCSSCAN in order to know the relevant character set with respective to input and set the nls_parameters
https://students.kiv.zcu.cz/doc/oracle/server.102/b14225/ch12scanner.htm#i1016925
- Pavan Kumar N
Similar Messages
-
Non-ASCII character in Email field
Hi Guys,
I am trying to enter non-english characters in Email field of user form, but OIM throws an error that "A non-Ascii character has been entered". I have also tried to turn off the AppFirewall Filter in xlConfig.xml file but no help. Is there any way thay I can enter non-Ascii characters in Email field?
Regards,
Rahul.oO(surfinIan)
>I have a script that converts a ms word document to text
then uploads that to a
>blob field on a mysql db.
> During the conversion some characters my not be
recognised. When i then call
>up the blob for display on the browser...those characters
show up as unknown
>characters with a ? or box. Is there a way to
preg_replace those unknown
>characters before displaying them.
What about fixing the encoding problem instead? If chars get
lost during
such a transfer
document->script->database->script->browser it's always
an encoding problem somewhere down the road.
The recommendation these days is to use UTF-8, which avoids
most of
these old problems. You just have to make sure that your
documents are
properly stored as UTF-8 in the database and delivered as
such to the
script and the browser, then you don't have to worry about
special chars
anymore.
That's just the general idea. I can't be more specific, since
I don't
know your conversion script or the database structure.
Micha -
[Solved] no non-ASCII character input in rxvt-unicode
Hello everyone,
For some days now, I can't write any non-ASCII characters any more in rxvt-unicode and rxvt-unicode-patched. Unfortunately, downgrading the rxvt-unicode package doesn't seem to help. To have at least a temporary solution, I'd like to know at least which packages I could try to downgrade as well. Any ideas, anyone?
greez,
maxmin
Last edited by Maximalminimalist (2011-03-12 13:12:26)When I try to type a non-ASCII-character I get nothing at all. This happens with my custom keyboard layout (modified programmer dvorak) and in some layouts I already tried (us: altgr-intl, ch, de and fr)
When I paste a non-ASCII characters in rxvt-unicode I get
maxmin ~ $ ?
This happens only on my x86_64 desktop which is more up to date than my i686 laptop. (I'm afraid now to do any updates.)
EDIT: I'm sorry, I don't know what you mean with locale settings. What do you mean with that?
EDIT2: Maybe just typing locale in the terminal is what you mean:
maxmin ~ $ locale
locale: Cannot set LC_CTYPE to default locale: No such file or directory
locale: Cannot set LC_MESSAGES to default locale: No such file or directory
locale: Cannot set LC_ALL to default locale: No such file or directory
LANG=en_US.utf8
LC_CTYPE="en_US.utf8"
LC_NUMERIC="en_US.utf8"
LC_TIME="en_US.utf8"
LC_COLLATE="en_US.utf8"
LC_MONETARY="en_US.utf8"
LC_MESSAGES="en_US.utf8"
LC_PAPER="en_US.utf8"
LC_NAME="en_US.utf8"
LC_ADDRESS="en_US.utf8"
LC_TELEPHONE="en_US.utf8"
LC_MEASUREMENT="en_US.utf8"
LC_IDENTIFICATION="en_US.utf8"
LC_ALL=
With other terminal emulators I get sometimes also nothing and sometimes right displayed but wrong interpreted character in vim. I didn't take notes while doing that but I'll try again if needed.
Last edited by Maximalminimalist (2011-03-06 21:51:23) -
ALV Grid bug when dealing with non-ASCII character
Dear all,
I have a requirement to display user's remarks on ALV. The data element of the remarks column is TEXT200. I know that each column in an ALV Grid can display at most 128 characters. Since my SAP is an Unicode system, I expect that each column in my ALV Grid can display 128 Chinese characters, too. However, the ALV Grid only display 42 Chinese characters at most. Is this a bug in ALV Grid? How can I fix it?
I did a small experiment. The results are listed below. My version is Net Weaver 7.01. The results show that the bug does not exist in ALV List. However, my user prefers ALV Grid, which is more beautiful and elegant.
Type of ALV
Max number of
ASCII character
in an ALV column
Max number of
non-ASCII character
in an ALV column
REUSE_ALV_GRID_DISPLAY
128
42 Chinese characters
CL_SALV_TABLE
128
42 Chinese characters
CL_GUI_ALV_GRID
128
42 Chinese characters
REUSE_ALV_LIST_DISPLAY
132
132 Chinese characters
If you encounter the bug, please post your solution. Thanks a lot.It looks like limitation of ALV grid cell, which can contain up to 128 bytes in SAP gui.
Your unicode characters are probably 3 bytes each.
Check OSS Note 910300 for more detailed info.
EDIT: Note 1401711 seems to be a correction for your issue It allows to use 128 characters (even if they take more than 128 bytes). -
Is Linksys WRT54GH SSID can contains the non-ascii character?
is Linksys WRT54GH SSID can contains the non-ascii character?
we need to use it for our wireless testing, but i dont know if the SSID can contains non-ascii.
anybody can help me? hurry, i will wait answer online.
thanks in advance!
Solved!
Go to Solution.thank you very much, Ricewind
SSID cant contain non-ascii characters, it make me sad and disappointed
why we can set T-link router SSID with non-ascii characters? -
Unicode value of a non-ASCII character
Hi,
Suppose, the unicode value of the character ् is '\u094d'.
Is there any Java function which can get this unicode value of a non-ASCII character.
Like:
char c='्';
String s=convertToUnicode(c);
System.out.println("The unicode value is "+ s);
Output:
The unicode value is \u094d
Thanks in advanceRanjan_Yengkhom wrote:
I have tried with the parameter
c:\ javac -encoding utf8 filename.java
Still I am getting the same print i.e. \u3fIf it comes out as "\u3f" (instead of failing to compile or any other value), then your source code already contains the question mark. So you already saved it wrong and have to re-type it (at least the single character).
>
Then I studied one tutorial regarding this issue
http://vietunicode.sourceforge.net/howto/java/encoding.html
It says that we need to save the java file in UTF-8 format. I have explored most of the editors like netbean, eclipse, JCreator, etc... there is no option to save the java file in UTF-8 format.That's one way. But since that is so problematic (you'll have to remember/make sure to always save it that way and to compile it using the correct switch), the better solution by far is not to use any non-ASCII characters in your source code.
I already told you two possible ways to achieve that: unicode escapes or externalized strings.
Also read http://www.joelonsoftware.com/articles/Unicode.html (just because it's related, essential information and I just posted that link somewhere else). -
i need a SQL or Procedure that will search non ascii character in data and update the data by removing it
Suppose there is table TABLE1 with Column NAME
it contain number of row and few has non ascii character eg 'CharacterÄr'
My sql or procedure should be able to search 'CharacterÄr' and update the row with 'Character'
i.e. removing the non ascii character 'Ä' from the dataHi,
Okay, in that case:
SELECT str
, REGEXP_REPLACE ( str
, '[^[:cntrl:] -~]'
) AS new_str
FROM table_x
or, to actually change the rows that contain the bad characters:
UPDATE table_x
SET str = REGEXP_REPLACE ( str
, '[^[:cntrl:] -~]'
WHERE REGEXP_LIKE ( str
, '[^[:cntrl:] -~]' -
Find the Special character and non Ascii character
Hi:
i have table ,this table column name contain some datas like
sno name
1 CORPORATIVO ISO, S.A. DE C.V.
2 (주)엠투소프트
3 TIMELESS
4 南京南瑞集团公司
5 PHOTURIS
6 Ace Informática S/C ltda
7 Computacenter AG & Co. oHG
8 아이티앤씨
9 MOCA
10 anbarasan
my requirement:
1)i need to search the name column where contain the special character and non ascii character..if found any non ascii or spcial character ..need to say flag ''yes".if not found need to say "no"...kindly help on this issus...i need some example..i am not have any idea....
i have table ,this table column name contain some datas like
sno name
1 CORPORATIVO ISO, S.A. DE C.V.
2 (주)엠투소프트
3 TIMELESS
4 南京南瑞集团公司
5 PHOTURIS
6 Ace Informática S/C ltda
7 Computacenter AG & Co. oHG
8 아이티앤씨
9 MOCA
10 anbarasan
my requirement:
1)i need to search the name column where contain the special character and non ascii character..if found any non ascii or spcial character ..need to say flag ''yes".if not found need to say "no"...kindly help on this issus... -
Keyword import fails on non-ascii character
I recently tried to import a long set of keywords (about 4000 terms). i set up the file in excel and then tried to import the records. I kept getting this message:
only text files encoded with ascii or unicode UTF-8 are supported when importing keywords.
I finally tracked down the problem when i converted the file to a MS word text file, broke it down into parts and eventually found the problem record. for some reason, the apostrophe in the words "don't know" had been corrupted to a weird character. after i corrected this, everything worked.
however, this took a long time. It would have been helpful if lightroom could have at least pinpointed the line where the import failed or offered to convert non-compliant charaters to some specific character or set of characters.Yeah, that didn't work so well since SuperDuper ran across repeated errors trying to do so; I suspect it's something to do with the drive. (SuperDuper complains about WD's MyBook, which is what the drive is.) Because SD stops the entire copy operation on single errors, it'd be a painstaking process.
Besides that, I like doing fresh installs of all the bits. -
Printer name with non-ASCII character
Is it possible to name a printer using ISO-8859-1 (or WINLATIN1) characters?
Our business system has logical printer names defined, sometimes using Norwegian (and Danish) letters (AE-lig, O-slash and A-ring), which we would like to map to the correct Windows UNC paths.
Mapping to a UNC path containing one of these characters works fine. But if the printer name itself contains a non-ASCI ISO character it fails.
Example using BÆR (B AE-lig R):
!p BÆR * \\print\bærum * 50 *
Don't know how the log will look after posting, but I'll try:
jfserver.exe: [400]D:\Adobe\Central\Server\Data\Test.dat : ^job TEST -zBÆR -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos"
jfserver.exe: [400] ^JOB formname='TEST' printername='B'
jfserver.exe: [400]Scanning !f for candidates
jfserver.exe: [400]Mapped printer name 'B' to id 'R' dev '*'
jfserver.exe: [400]Other job tokens before substitution 'R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos"'
jfserver.exe: [400]Other job tokens after substitution'R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos"'
jfserver.exe: [400]Merge parms before substitution '\\printer\b'
jfserver.exe: [400]Merge parms after substitution'\\printer\b'
jfserver.exe: [400]Task parms before substitution '"test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'
jfserver.exe: [400]Task parms after substitution '"test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'
jfserver.exe: [400]"test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"
jfserver.exe: [307]Launching task '"D:\Adobe\Central\Bin\jfmerge" "test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'.
jfmerge: [125]* Processing data file: 'D:\Adobe\Central\Server\Data\Test.dat'.
jfmerge: [289]MDF file `D:\Adobe\Central\Samples\Exprint\forms\test.mdf' opened.
jfmerge: [400]The filename, directory name, or volume label syntax is incorrect.
jfmerge: [2]Error opening output device/file '*'.
jfmerge: [210]Nothing was printed.
jfserver.exe: [314]Agent exit message: [2]Error opening output device/file '*'.
jfserver.exe: [400]spawned '"D:\Adobe\Central\Bin\jfmerge" "test.mdf" "D:\Adobe\Central\Server\Data\Test.dat" -l -apr"" -all"D:\Adobe\Central\Server\jfserver.log" -asl1 -amq0 -ams"D:\Adobe\Central\Server\Mst\R.mst" -m2T -z"*" \\printer\b R -afp"D:\Adobe\Central\Samples\Exprint\forms" -alp"D:\Adobe\Central\Samples\Exprint\logos" -aii"D:\Adobe\Central\Server\jfmerge.ini"'
jfserver.exe: [400]--> returned [-2] [2]Error opening output device/file '*'..For others who might have the same problem. You can solve it by quoting the device name and/or printer name.
!p """BÆR""" * """\\print\bærum""/q" * 50 *
This, however, causes the printer name to be '"BÆR"' with the double quotes included in the name. Thus, the job card must contain the unfortunate tripple quotes as well: -z"""BÆR""".
Vegard -
Convert non-ASCII character to decimal value
Hi all,
I have the following problem:
When reading a String containing the full spectrum of UTF-8 characters from a BufferedReader:
How can I get the decimal value of the String's characters outside the ASCII range?
When I try to do a MyBufferedReader.read(), I get the int value 65533 for these characters, instead of their real value in the UTF-8 set.
Can anyone help?
Thanks!
DreThat's the character you get when you try to decode a character with no valid decoding. Of course there aren't any such characters for UTF-8. Therefore whatever you did, don't do that. More specifically: you've already got a String, don't do anything which involves converting it to bytes. That's completely unnecessary.
Or perhaps the problem is the way you create the String in the first place. Who knows? -
Extended ascii character problem
Here is the message I am getting
2004-10-12 10:04:44,207 ERROR [STDERR] org.postgresql.util.PSQLException: Invalid character data was found. This is most likely caused by stored data containing characters that are invalid for the character set the database was created in. The most common example of this is storing 8bit data in a SQL_ASCII database.
this is the character that is causing the problem (ascii 0232 �)
My question is how can I make java display this character. I am using Postgresql 7.4 as my database.I don't understand the connection between the database and "java display".
And ASCII is a 7 bit character set, so the character you listed is not ASCII, rather it is a character in some other character set. -
Extended ASCII character problems
We are having problems with certain characters. If a character in the range ASCII 194-239 appears in a string it can cause the next character to be lost. Everything seems okay in SQL, but not in PL/SQL.
A PL/SQL script to print all the ASCII characters outputs nothing for the above range.
e.g.
DECLARE
v VARCHAR2(100):= 'Requiem por un PolicÃ|xxx|qqq';
BEGIN
dbms_output.put_line(instr(v,'|'));
dbms_output.put_line(substr(v,instr(v,'|')));
END;
The expected results would be 22 and |xxx|qqq, but we actually get 25 and |qqq. The first '|' is being lost somehow.
A test using a Java function retrieving the string from a table failed due to a conversion failure between UTF8 and UCS2.
The NLS language is AMERICAN and the character set is UTF8.
Thanks
SteveI know this thread is old, but it took me forever to find this out...
Use:
SELECT CHR (<ASCII Value> USING NCHAR_CS) FROM DUAL;
E.g.
SELECT CHR (206 USING NCHAR_CS) FROM DUAL;
CHR (206 USING NCHAR_CS)
ÎEdited by: martinByrne on 04-Aug-2010 08:37 -
Non-English character problem in Oracle 10g Express Edition
Hi There;
I have a table. It's name is INSTITUTION. It has a NUMBER INS_ID and NVARCHAR2(50) INS_NAME . INS_NAME can contain Turkish characters, such as "ğ,ü,ş,ç,ö". According to business logic, there can not be a repetition on the INS_NAME.
User will enter institution name from a textbox in ASP.NET , and I check this name in database from c sharp code, if there is no repetition, we will add this record.
The problem is; when user enter a instition name that contains Turkish character, there is a duplication. If there is a instition name is *"su işleri"* , the both query; SELECT * FROM INSTITUTION WHERE INS_NAME = *'su işleri'*; and SELECT * FROM INSTITUTION WHERE INS_NAME = *'su isleri'*; returns no result, even though there it is.
But if instition name is "oracle corporation" (there is no Turkish character) it query successfully. I have the same problem in Toad for Oracle 11.5.1.2. When I query database from toad SELECT * FROM INSTITUTION, the phrase *"su işleri"* has appeared. But when I query SELECT * FROM INSTITUTION WHERE INS_NAME = *'su işleri'*; , there is again no result.
When I connect oracle database directly and perform the query SELECT * FROM INSTITUTION , the phrase *"su isleri"* (not *"su işleri"* ) has appeared.
Here are the language settings of the database:
National Language Support
National Language Parameter Value
NLS_CALENDAR______________GREGORIAN
NLS_CHARACTERSET__________WE8MSWIN1252
NLS_COMP__________________BINARY
NLS_CURRENCY______________TL
NLS_DATE_FORMAT__________DD/MM/RRRR
NLS_DATE_LANGUAGE________TURKISH
NLS_DUAL_CURRENCY_________YTL
NLS_ISO_CURRENCY__________TURKEY
NLS_LANGUAGE______________TURKISH
NLS_LENGTH_SEMANTICS______BYTE
NLS_NCHAR_CHARACTERSET___AL16UTF16
NLS_NCHAR_CONV_EXCP______FALSE
NLS_NUMERIC_CHARACTERS____ ,.
NLS_SORT___________________TURKISH
NLS_TERRITORY______________TURKEY
NLS_TIME_FORMAT____________HH24:MI:SSXFF
NLS_TIMESTAMP_FORMAT_______DD/MM/RRRR HH24:MI:SSXFF
NLS_TIMESTAMP_TZ_FORMAT____DD/MM/RRRR HH24:MI:SSXFF TZR
NLS_TIME_TZ_FORMAT__________HH24:MI:SSXFF TZR
How can I resolve that problem? Thanks in advance.
Edited by: 963344 on 05.Eki.2012 01:00
Edited by: 963344 on 05.Eki.2012 01:01
Edited by: 963344 on 05.Eki.2012 01:06This type of question/discussion belongs in {forum:id=50} forum.
Very recently a thread there touched the topic of Turkish character support.
Please read it: Western European Characterset to Turkish in sql
>
NLS_CHARACTERSET__________WE8MSWIN1252 Check the character set repertoire of win-1252 (look for the typical turkish language characters you've mentioned above).
http://msdn.microsoft.com/en-us/goglobal/cc305145.aspx
Look at character names, such as "... letter s with cedilla". -
Php / mysql replace non-ascii character in a string
I have a script that converts a ms word document to text then
uploads that to a blob field on a mysql db.
During the conversion some characters my not be recognised.
When i then call up the blob for display on the browser...those
characters show up as unknown characters with a ? or box. Is there
a way to preg_replace those unknown characters before displaying
them.
thanks
ian.oO(surfinIan)
>I have a script that converts a ms word document to text
then uploads that to a
>blob field on a mysql db.
> During the conversion some characters my not be
recognised. When i then call
>up the blob for display on the browser...those characters
show up as unknown
>characters with a ? or box. Is there a way to
preg_replace those unknown
>characters before displaying them.
What about fixing the encoding problem instead? If chars get
lost during
such a transfer
document->script->database->script->browser it's always
an encoding problem somewhere down the road.
The recommendation these days is to use UTF-8, which avoids
most of
these old problems. You just have to make sure that your
documents are
properly stored as UTF-8 in the database and delivered as
such to the
script and the browser, then you don't have to worry about
special chars
anymore.
That's just the general idea. I can't be more specific, since
I don't
know your conversion script or the database structure.
Micha
Maybe you are looking for
-
I'm having trouble getting FCPX to share my entire video, it cuts off the opening titles and starts a few seconds in when I share to my MacBook. It's my first vid so I've done a few daft things like editing the audio after I put the visual together,
-
Saving a new .pdf doc that is (a)chapter(s) of a large doc
Is there a way to save only certain chapters from a .pdf doc? These documents have chapters. http://support.spss.com/ProductsExt/SPSS/Documentation/Manuals/16.0/SPSS%20Base%20User's%2 0Guide%2016.0.pdf There is an index on the left. It is possible to
-
My 5S appears to have zoomed in so much I can't see anything. Is there a way to reset?
please help, my 5S appears to be frozen
-
Error 32809 on some machines after Update December 2014
A user reported an error in a German forum that an Excel file throws the RTE 32809 on some machines. https://social.msdn.microsoft.com/Forums/de-DE/65025f6a-cf64-486a-b68d-9825e9d304d2/fehlermeldung-32809-seit-sicherheitsupdate-vom-09122014?forum=off
-
How to save a image from a document?
Hi there I was trying to save the image the from the "Welcome to Buzzword" document but when i right clicked on it allows me to print the document alone. Is it possible to download the picture to desktop like the way we do in a Adobe PDF file? Karthi