Encoding non english characters with utf 8 on jsp (Critical!!)

I am inserting hebrew characters from JSP into oracle db and everything is fine until this point. But when I try to retrieve the information from the database, the characters are not displayed properly (I get some garbage characters). I am sure that the data stored in the database is correct, but not sure why there is a problem in displaying the data in the JSP.
I came across a thread on TSS
http://www.theserverside.com/discussions/thread.tss?thread_id=28944
and followed the suggestions given there like having
<%@ page contentType="text/html; charset=UTF-8" pageEncoding="UTF-8" %>
<META http-equiv="Content-Type" content="text/html; charset=UTF-8">and also this
<%
//Some JDBC and sql statement query UTF-8 data and then ...
String str = rs.getString("utf8_data");
str = new String(str.getBytes("ISO-8859-1"),"UTF-8");
%>
<%= str %>Now, the data getting displayed is partly correct, I mean to say, some characters are still coming as squares.
Any ideas will be of great help.

even i doubt the database charset for this issue. But what I dont understand is how only certain hebrew characters are getting stored properly and why others are corrupted?
Also, can anyone let me know how i can view the Non-English characters present in the database directly, as TOAD is not able to display them

Similar Messages

  • SmartForms : Printing Non-English characters with SWIN output device?

    I have a Japanese Text (entered in Unicode at So10) and I want to print it from my SmartForms application using SWIN device type.
    I thought SWIN converted all the components into graphics before sending to printer. To my surprise, the Japanese text printed out as "########".
    I even tried duplicate the SWIN output device to another output device and changed the character set to 4103 UTF-16LE Unicode / ISO/IEC 10646 but then the print process hang.
    Anyone knows how to overcome the problem so that I can print Japanese with SWin output device?

    I have a Japanese Text (entered in Unicode at So10) and I want to print it from my SmartForms application using SWIN device type.
    I thought SWIN converted all the components into graphics before sending to printer. To my surprise, the Japanese text printed out as "########".
    I even tried duplicate the SWIN output device to another output device and changed the character set to 4103 UTF-16LE Unicode / ISO/IEC 10646 but then the print process hang.
    Anyone knows how to overcome the problem so that I can print Japanese with SWin output device?

  • Problem with the Non-English Characters

    Hello,
    I have been using Adobe Illustrator  but I have a huge problem with the non-english characters with Standart Fonts. With the Professional font's I have no problem with them. But when I'm using any standart font in font folio library I cannot type any "ğ-İ-ş". I can add those letters in fontlab with the glyphs (scedilla, idotaccent, gbreve). Most of the fonts have those letters already prepeared so I dont even have to redraw. But I can't add those glyph to every single font because I dont have that kind of time and patience. Is there any better solution for this? Or is there any font folio pack that all fonts are PRO.
    I'm looking forward for your answers
    Thanks.

    Joel wrote: I'm told that this is the exact difference between Adobe's Standard and Pro fonts — the Pro fonts have additional glyphs, including those necessary for extended Latin script.
    Exactly. The Pro fonts have at a minimum the Adobe Western 3 character set, which is essentially western European + Adobe CE.
    > Standard fonts just have the basic English character set, with maybe a bit of help for Spanish and French.
    A lot more than that!
    > You're doing Turkish, right? Adobe's coverage for Turkish in its fonts is not great - some of the Pro fonts have Turkish coverage, many do not.
    This is false. Every single Adobe Pro font supports Turkish.
    To be clear:
    All Adobe Standard fonts support the following languages: Afrikaans, Basque, Breton, Catalan, Danish, Dutch, English, Finnish, French, Gaelic, German, Icelandic, Indonesian, Irish, Italian, Norwegian, Portuguese, Sami, Spanish, Swahili and Swedish.
    Adobe Pro fonts support those languages, plus AT LEAST: Croatian, Czech, Estonian, Hungarian, Latvian, Lithuanian, Polish, Romanian, Serbian (Latin), Slovak, Slovenian and Turkish. Some Pro fonts have more language support than this, such as Greek and/or Cyrillic, and additional extended Latin.
    See: http://www.adobe.com/type/browser/info/charsets.html
    Cheers,
    T

  • URLEneQuery encoding is failing for some non english characters

    While creating a URLEneQuery we are getting error com.endeca.navigation.InternalException: No support for 8-bit urls.
    This error happens when the query string has some non english characters. (eg: Á).

    UrlENEQuery is designed around processing URL data, and URLs are not permitted non-ASCII characters in their production. To represent non-ASCII characters they must be %-encoded in URLs according to their byte(s) representation in a particular character-encoding, and you should prefer UTF-8 for URLs. So your LATIN CAPITAL LETTER A WITH ACUTE (U+00C1) should appear as %C3%81 in your URL, then UrlENEQuery should be able to process that character.

  • Non-English Characters (Encoding)

    Using the XMP Toolkit I'm having problems reading and writing non-English characters.  For example: keywords read which should be "casa campesina, cultivos agrí colas, zona cafetera, café, plátano" read as "casa campesina, cultivos agrÃcolas, zona cafetera, café, plátano".   I have the same problem with other languages such as Norwegian.
    Does anyone else have this type of problem?  Or perhaps a suggestion as to what I might be doing to cause such a problem?
    Best regards,
    Glenn Rogers
    Developer of DBGallery: Photo DATAbase System

    Hi Glen,
    if you write non ASCII characters using our toolkit, you have to make sure to encode your string in UTF8.
    If you see this while reading, the data in the file might not be valid UTF8. If it's local encoding (for example mac local encoding) our library will try to convert it based  on the OS you are running it on. So it you got mac local encoding in the EXIF of the file and you are using the toolkit on windows, this might cause the wrong characters you are seeing.
    In order to avoid this please always use UTF8 encoded strings.
    Regards,
    Samy
    XMP Team

  • My Firefox cannot display non-English characters, even though I have tried every language encoding I have!

    I am a big fan of Japanese songs and websites, so I was very disappointed when I saw that Firefox could not handle any non-English characters. I have tried every encoding I can, but none work and I just see boxes with numbers and letters inside. I have only just got this older laptop for my birthday - my old laptop which ran Windows Vista and had Firefox 4 had no trouble at all. Please help me!

    hello muoshui, please enter '''about:config''' into the firefox location bar (confirm the info message in case it shows up) & search for the preference named '''network.http.accept-encoding''' - right-click and reset that entry to the default value.
    if this does not resolve the issue already, please also go through the steps offered at [[Websites look wrong or appear differently than they should]].

  • [AS] Problem with non English characters in file path

    I wrote a script that exports a pdf file from ID, rasterizes it in PS, applies an action, saves it as another pdf file, and finally creates a Mail message, and attaches the file to it (the last part is written in AppleScript).
    The problem is that it doesn't work when the path to this file contains non English characters.
    This works:
    make new attachment with properties {file name:"/Volumes/Macintosh HD/BackUp Tetard/Test.pdf"}
    but this doesn't:
    make new attachment with properties {file name:"/Volumes/Macintosh HD/BackUp Têtard /Test.pdf"}
    I remember vaguely that I read somewhere that AppleScript can work with Unicode — in other words with such characters — starting from some version, don't remember which exactly, but it seems to me — Leopard.
    I am on Mac OS X 10.4.11 right now. Will updating solve this problem? Does anybody know any solution to this problem: a scripting addition, some hidden setting, etc.
    I made a little test: used a Russian character — ё and it works, but when I use — ê (Dutch) it doesn't. May it have something to do with the Region setting in International panel?
    Thanks in advance,
    Kasyan

    Kasyan, as of Leopard AppleScript treats all text as Unicode pre this you can specify 'as Unicode text'. Try a test with these.
    -- Leopard
    set x to POSIX path of (path to desktop)
    -- Pre Leopard
    set x to POSIX path of (path to desktop as Unicode text)
    -- Leopard
    set x to POSIX path of (choose file without invisibles)
    -- Pre Leopard
    set x to POSIX path of ((choose file without invisibles) as Unicode text)

  • Non-English characters in URL for rwservlet

    I'm having a problem when I try to use non-english characters in a URL request to generate a report.
    This works fine:
    http://...rwservlet?report=r1.jsp&m1=Fred
    But if I try Fréd (e with accent graph) the report does not return any data even though the SQL by itself would find data.
    I tried UTF-8 encoding
    http://...rwservlet?report=r1.jsp&m1=Fr%C3%A9d
    8859-1 encoding
    http://...rwservlet?report=r1.jsp&m1=Fr%E9d
    Or just spell it out (not sure what that gets encode as):
    http://...rwservlet?report=r1.jsp&m1=Fréd
    But noting works. Any ideas?
    Thanks, Andreas

    Suggestions
    1) Try with NLS_LANG as
    SWEDISH_SWEDEN.WE8DEC
    2) Make a paramform and enter via paramform (unencoded)
    (This is just for testing purpose)
    3) Change machine locale to swedish and try
    4) Which reports version is this ?
    Please see
    BUG 2713695 - NLS CHARACTERS FOR PARAMETERS CHANGE TO QUESTION MARKS WHEN PASSED ON URL BAR
    Get in touch with Support to see if this is the issue and if "yes" get a one-off patch.
    [    All Docs for all versions    ]
    http://otn.oracle.com/documentation/reports.html
    [     Publishing reports to web  - 10G  ]
    http://download.oracle.com/docs/html/B10314_01/toc.htm (html)
    http://download.oracle.com/docs/pdf/B10314_01.pdf (pdf)
    [   Building reports  - 10G ]
    http://download.oracle.com/docs/pdf/B10602_01.pdf (pdf)
    http://download.oracle.com/docs/html/B10602_01/toc.htm (html)
    [   Forms Reports Integration whitepaper  9i ]
    http://otn.oracle.com/products/forms/pdf/frm9isrw9i.pdf
    ---------------------------------------------------------------------------------

  • Wrong sorting of non-English characters (iTunes 11.2)

    Since version 11.2 of iTunes, there is a problem with sorting of names with non-English characters (such as ž, š, č, ř or other in Czech). It seems that iTunes completely omits these characters when sorting. As an example, I have Čajkovskij (which is the Czech spelling of Tchaikovsky) among artists in my library and it is currently sorted between Adele and Alan Parsons Project, i.e. it seems that the character Č has been completely omitted while sorting and thus the artist is sorted as Ajkovskij.
    All my ID3 tags are of version 2.3 and in Unicode, thus this should not be any encoding problem. I even tried to change the ID3 tags to version 2.4, i.e. UTF-8, and it did not help.
    I think that there must be many more people experiencing this problem. Not everyone has only English or ASCII names in the library and I am surprised that I could not find anything on this bug anywhere on the net yet.
    Thanks for any help on this.

    Looks like a bug in the latest build. See this recent thread.
    You could report the problem via iTunes Feedback or sign up for a free Apple Developer Connection account and make use of Apple Bug Reporter.
    tt2

  • Encrypting non-English characters

    Hi,
    I have this application which has to do the following
    Scenario (i)
    - Read ENCODED string SE from Network Source NS1 (Native,Non-JAVA)
    - Decode SE to SD using the same charset as NS1
    - Apply some transformation to SD to get SD2
    - Encode SD2 to get SE2 using the same charset as Network Source NS2
    (Native, Non-JAVA)
    - Send SE2 to NS2
    - NS2 gets what it expects without any problems :))
    Scenario (ii)
    - Read ENCODED string SE from Network Source NS1 (Native,Non-JAVA)
    - Decode SE to SD using the same charset as NS1
    - Apply some transformation to SD to get SD2
    - Get the bytes from SD2 as BSD2
    - Encode BSD2 to get BSE2 using the same charset as Network Source NS2
    (Native, Non-JAVA)
    - Encrypt BSE2 to BSE2_Enc
    - Send BSE2_Enc to NS2 (Native,Non-JAVA)
    - NS2 does not gets what it expects :((
    (It recieves English text OK but it gets ???? for non-English)
    The charset being used is windows-1256 (at NS1,NS2 and my application)
    Encryption is being done using BouncyCastle TwoFish w/ 256 bit keys
    Reading/Writing from/to network is being done over SocketChannel
    Get the bytes from SD2 as BSD2 => byte [] BSD2 = SD2.getBytes()It seems the non-Enlish characters are getting lost when I go SD2.getBytes()
    and they get encrypted as 'lost-non-English characters' ;)
    And when they get decrypted at NS2, they are displayed as 'lost-non-English
    characters' :)) i.e. ??????? .. so on
    Is there a way I can encrypt non-English plain text without losing information ?
    (without having to implement a TwoFish engine in my application itself)

    1) Bytes are not characters. Characters are UNICODE
    and have a byte representation defined by an encoding
    scheme. It is usually wrong to use the default
    encoding given by String.getBytes(). One should realy
    use String.getByte(encoding) eg
    "fred".getBytes("UTF-8");Awwllright ... got that :)) Thanks buddy
    2) Not having access to your code makes it difficult
    but make sure you are not converting encrypted bytes
    to a String using new String(encrypted bytes); No .. I am not doing that.
    3) Again, not having access to your code makes it
    difficult , but when you display your Strings make
    sure that you use a Font that has representations for
    all the UNICODE characters you wish to display. It is
    normal for any character that does not have a valid
    glyph in a gien font to display as a box.That infrastructure exists and is working fine ... as I mentioned
    this is working OK when plain text is being used.
    The problem was with using the getBytes() rather than getBytes("windows-1256")
    Its working now ... thanks alot .. again. I wonder how that never occurred to me.

  • Flex, xml, and non-English characters

    Hello! I have a Flex web app with AdvancedDataGrid. And I use httpService component to load some data to grid. The .xml file contains non-english characters in attributes (russian in my case) like this:
    <?xml version="1.0" encoding="utf-8" ?>
       <Autoparts>
        <autopart  DESCRIPTION="Барабан">
    </Autoparts>
    And when i run app, AdvancedDataGrid display it like "Ñ&#129;ПÐ". How can i fix it? I try to change encoding="utf-8" with some another charsets, bun unsuccesfully. Thank you.

    Try changing the xml structure by using CDATA instead of having the russian part as an attribute and see if that makes any difference.
    What I meant is use something like this:
    <?xml version="1.0" encoding="utf-8" ?>
       <Autoparts>
        <autopart>
           <description><![CDATA[Барабан]]></description>
      </autopart>
    </Autoparts>
    instead of the current xml.

  • Non-English characters

    Hello, I have read several times that since Java uses Unicode, it solves the problems of non-English characters automatically or something like that.
    But my app is not working as expected. Would someone help please?
    I have a client/server combo written in Java. The server can send messages in English or Japanese. The Japanese messages are hard-coded as String literals in the server source code. On the client side, they are displayed on a JEditorPane. But the Japanese characters are all garbled. The OS on the server side and client side are, of course, different.
    My supposition, which is obviously wrong as it is not working, is that since both ends of communication are Java app, I need not worry about any encoding conversions for String literals.
    Suggest me what is wrong here?

    How is the required encoding/decoding supposed to be done?
    When I didn't worry about non-English characters, I did the following, which WORKED.
    // SENDER side
    Socket socket ;
    PrintWriter     out = new PrintWriter(socket.getOutputStream(),true);
    String outMessage = "my message";
    out.println(outMessage);//RECEIVER
    Socket socket ;
    BufferedReader in = new BufferedReader(new InputStreamReader(socket.getInputStream()));
    String inMessage =  in.readLine();When non-English characters are involved, I did the following, which DID NOT WORK. Please someone correct me.
    // SENDER side
    Socket socket ;
    PrintWriter     out = new PrintWriter(socket.getOutputStream(),true);
    String outMessage = "my message";
    String utfString = new String(outMessage.getBytes(),"UTF-8");
    out.println(utfString);//RECEIVER
    Socket socket ;
    InputStreamReader ins = new InputStreamReader(clientSocket.getInputStream(),"UTF-8");
    BufferedReader in = new BufferedReader(ins);
    String inMessage =  in.readLine();The received message is still garbled.

  • Non-english characters in sqlplus

    Friends,
    9.0.1,10.2.0.1
    CAn the non-english characters be displayed in sqlplus prompt?
    I have to print the data with some hindi charcaters like name from the report generated via proc.
    What is teh workaround for this?
    Thanks

    Hi,
    I have RH Linux 7.3.
    Here is my situation.
    I have some tables that have some columns.In these columns i need to store data in hindi.
    The database character set(nls_characterset) is US7ASCII.
    The datatypes of the columns is varchar2 type.
    The tables are accessed through Pro*c and the proc code generates the report.
    The data in these reports needs to be printed out and the corresponding data has to be in hindi.
    So i have to do two things
    1) On query the tables, the user should see hindi data along with english data.
    2) The printed report should also contain hindi data.
    Questions:
    1) If i change the charcater set by the command:
    alter database character set AL32UTF8;
    and OS locale to UTF-8
    Would these chnages solve the purpose?
    2) Or do i need to recreate the database with AL32UTF8 and exp/imp all the data?
    3)Any other advice/option?
    Thanks

  • Non English characters conversion issue in LSMW BAPI Inbound IDOCs

    Hi Experts,
    We have some fields in customer master LSMW data load program which can
    contain non-English characters. We are facing issues in LSMW BAPI
    method with non-English characters Conversion. LMSW steps read and
    conversion are showing the non-English characters properly with out any
    issue. While creating inbound IDOCs most of the non-English characters
    replaced with '#' and its causing issues in creating customer master data in
    system. In our scenario customer data with non-English characters in
    the first name, last name and address details. Any specific setting
    needs to be done from our side? Please suggest me to resolve this issue.
    Thanks
    Rajesh Yadla

    If your language is a unicode tehn you need to change the options  like IN SAP you need to change it to unicode  in the initial screen Customize local layout(ALT F12) options 118  --> Encoding ....

  • Can't CF use non English characters in URLs ???? (critical for SEO)

    Hi all,
    I want to use non English characters (Greek characters) for folders of the URLs.
    eg   http://www.mysite.com/Φάκελος/index.cfm
    where "Φάκελος" is a non English word (Greek).
    When the called page is simple HTML  eg   www.mysite.com/Φάκελος/index.HTM
    it's displayed just fine.
    When the called page is CF page  eg   www.mysite.com/Φάκελος/index.CFM
    I get a "FILE NOT FOUND" error.
    In the page where the link exists everything is UTF-8.
    What's the problem ? Can't CF use non English characters in URLs ????
    It's critical for SEO issues.
    I use CF9.  Any ideas ???
    Thanks in advance.
    Anastassios

    I don't have this setting in the email application. But as I know, html with Exchange is working only with the 2007 version, my server is still 2003 so I think in my case it's plain text only.
    But I'm telling again: good old (and now starting to miss) E60 with MfE worked very well!

Maybe you are looking for

  • Dock not appearing & system freeze

    I'm runing OSX 10.6.8 and yesterday started having an issue where my dock pops up during start up and then goes away, comes back a few times, but ultimately will not come up at all. Then system slows down and then freezes with a spinning beach ball.

  • In MIGO wrong quantity entered

    Dear Users, In sap migo transaction user has entered more quantity than actual received physically and MIRO done for actual physical quantity. Now the issue is how to remove this excess quantity from the system so that PO quantity is balanced with ac

  • Where To Buy A Power Supply

    Hey just wonderin if any of you know somewhere to get a cheep Power Supply becouse im building a PowerMac G5 but i need a Power Supply for my june 2004 case. caleb

  • Word- pdf contains only images, no text

    Hi, i have created a Word 2010 Document with the typefaces Caslon and Kozuka. When creating the pdf with any setting (Standard, high quality, pdf/x) in Acrobat X (10.1.16) there is no text. Both "Acrobat-Create pdf-choose word file" and "Word-Acrobat

  • OPEN DATASET in ECC6.0

    Hi Guys, We are upgrading from 4.6c to ECC 6.0 and a lot of our programs are giving unicode compliance error on the OPEN DATASET statement. Even though we are moving to a unicode system, the systems we talk to are not unicode compliant yet so we dono