Special Character Encoding issue

Hi all
Am using OAS9i. i ve deployed a webservice. i submit a payload request data that has some unicode characters like "§". The data is base64binary encoded. The type of the element mentioned in the schema is base64binary. When i retrieve the payload in java implementation code the character is displayed as � in the console. Please advice how to fix this issue. I tried setting JVM option file.encoding=utf-8 it didnt work out.
Thanks
Shiny

When you use an UDF and you have programmed a Sax parser, then make sure, that the parser works with the correct encoding. So when the result of the webservice is  ISO-8859-1, then assign this to the parser.
In principle the encoding should be part of XML header. Make sure that the encoding of the response is the same as declared in XML header.

Similar Messages

  • Character encoding issue in sql server

    Hi Team,
    We have a table with more than 20 columns...In that we have some columns which will have data extracted from the Datawareshouse applicances and its one time load.
    The problem is we have two columns which will may have same set of values from some records and for different set of records
    the below values are the example for same set of records but the values are changed while importing into the sql server database..
    2pk Etiquetas Navide‰as 3000-HG                                
     2pk Etiquetas Navideñas 3000-H                           
    Is there anyway to change the first column values into the second column value ? 
    By looking at the data we can say its the code page issue..(Character encoding issue)..but how to convert it?
    Convertting(2pk Etiquetas Navide‰as 3000-HG)  
    to get   2pk Etiquetas Navideñas 3000-H   in the select query?

    Then it seems that you can do the obvious: replace it.
    DECLARE @Sample TABLE ( Payload NVARCHAR(255) );
    INSERT INTO @Sample
    VALUES ( N'2pk Etiquetas Navide‰as 3000-HG' );
    UPDATE @Sample
    SET Payload = REPLACE(Payload, N'‰', N'ñ');
    SELECT S.Payload
    FROM @Sample S;

  • Character encoding issue

    I'm using the below give code to send mail in Trukish language.
    MimeMessage msg = new MimeMessage(session);
    msg.setText(message, "utf-8", "html");
    msg.setFrom(new InternetAddress(from));
    Transport.send(msg);
    But my customer says that he gets sometime unreadable characters in mail. I'm not able to understand how to solve this character encoding issue.
    Should i ask him to change his mail client's character encoding settings?
    If yes which one he should set.

    Send the same characters using a different mailer (e.g., Thunderbird or Outlook).
    If they're received correctly, come the message from that mailer with the message
    from JavaMail. Most likely other mailers are using a Turkish-specific charset instead
    of UTF-8.

  • Special character display issue

    Hi,
    I've been struggling with this issue for some time now and
    have not yet found a solution. Basically I am trying to display the
    letter 'á' in bold within a dynamic text field, where the text
    is parsed from some internal script.
    There is nothing wrong with the script as it works for other
    special characters. I have embedded the correct glyphs, and even
    tried all glyphs.
    I have tried several methods of encoding the character in the
    scipt such as Unicode Escape, Numeric Entity and URL encoding; no
    luck with either. For other special characters I have simply
    embedded the glyph and typed the special character into my script
    using the charmap instructions and they have worked fine but for
    some reason when I want to make it bold it just wont display. The
    method I've used to embolden the text is show below, please note
    this has worked fine for other characters and the bold version of
    the font contains the glyph in the charmap.
    "<font face=\"AvantGarde LT
    Bold\">Lávese</font>"
    This results in: Lvese (in bold) skipping the á
    I'm starting to wonder if this is a problem with the software
    itself. If anyone has any suggestions, I'd love to give them a go.
    Thanks in advance for any assistance
    P.S. my e-mail is [email protected]

    Pl identify which exact version of 10g, along with the database characterset.
    Going by what you have stated, your database characterset is most likely WE8ISO8859P1, which does not recognize the single quote (or apostrophe) or the bullet point.
    You will need to use a characterset that supports these characters - Oracle recommends AL32UTF8
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14225/ch6unicode.htm
    If you can recreate this database, use AL32UTF8 then perform your conversion.
    HTH
    Srini

  • JSF myfaces character encoding issues

    The basic problem i have is that i cannot get the copyright symbol or the chevron symbols to display in my pages.
    I am using:
    myfaces 2.0.0
    facelets 1.1.14
    richfaces 3.3.3.final
    tomcat 6
    jdk1.6
    I have tried a ton of things to resolve this including:
    1.) creating a filter to set the character encoding to utf-8.
    2.) overridding the view handler to force calculateCharacterEncoding to always return utf-8
    3.) adding <meta http-equiv="content-type" content="text/html;charset=UTF-8" charset="UTF-8" /> to my page.
    4.) setting different combinations of 'URIEncoding="UTF-8"' and 'useBodyEncodingForURI="true"' in tomcat's server.xml
    5.) etc... like trying set encoding on an f:view, using f:verbatim, specifying escape attirbute on some output components.
    all with no success.
    There is a lot of great information on BalusC's site regarding this problem (http://balusc.blogspot.com/2009/05/unicode-how-to-get-characters-right.html) but I have not been able to resolve it yet.
    i have 2 test pages i am using.
    if i put these symbols in a jsp (which does NOT go through the faces servlet) it renders fine and the page info shows that it is in utf-8.
    <html>
    <head>
         <!-- <meta http-equiv="content-type" content="text/html;charset=UTF-8" /> -->
    </head>
    <body>     
              <br/>copy tag: &copy;
              <br/>js/jsp unicode: &#169;
              <br/>xml unicode: &#xA9;
              <br/>u2460: \u2460
              <br/>u0080: \u0080
              <br/>arrow: &#187;
              <p />
    </body>
    </html>if i put these symbols in an xhtml page (which does go through the faces servlet) i get the black diamond symbols with a ? even though the page info says that it is in utf-8.
    <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
    <html xmlns="http://www.w3.org/1999/xhtml"
         xmlns:ui="http://java.sun.com/jsf/facelets"
         xmlns:f="http://java.sun.com/jsf/core"
         xmlns:h="http://java.sun.com/jsf/html"
         xmlns:rich="http://richfaces.org/rich"
         xmlns:c="http://java.sun.com/jstl/core"
           xmlns:a4j="http://richfaces.org/a4j">
    <head>
         <meta http-equiv="content-type" content="text/html;charset=UTF-8" charset="UTF-8" />
         <meta http-equiv="X-UA-Compatible" content="IE=EmulateIE7" />
    </head>
    <body>     
         <f:view encoding="utf-8">
              <br/>amp/copy tag: &copy;
              <br/>copy tag: &copy;
              <br/>copy tag w/ pound: #&copy;
              <br/>houtupt: <h:outputText value="&copy;" escape="true"/>
              <br/>houtupt: <h:outputText value="&copy;" escape="false"/>
              <br/>js/jsp unicode: &#169;
              <br/>houtupt: <h:outputText value="&#169;" escape="true"/>
              <br/>houtupt: <h:outputText value="&#169;" escape="false"/>
              <br/>xml unicode: &#xA9;
              <br/>houtupt: <h:outputText value="&#xA9;" escape="true"/>
              <br/>houtupt: <h:outputText value="&#xA9;" escape="false"/>
              <br/>u2460: \u2460
              <br/>u0080: \u0080
              <br/>arrow: &#187;
              <br/>cdata: <![CDATA[©]]>
              <p />
         </f:view>               
    </body>
    </html>on a side note, i have another application that is using myfaces 1.1, facelets 1.1.11, and richfaces 3.1.6 and the unicode symbols work fine.
    i had another developer try to use my test xhtml page in his mojarra implementation and it works fine there using facelets 1.1.14 but NOT myfaces or richfaces.
    i am convinced that somewhere between the view handler and the faces servlet the encoding is being set or reset but i havent been able to resolve it.
    if anyone at all can point me in the right direction i would be eternally greatful.
    thanks in advance.

    UPDATE:
    I was unable to get the page itself to consume the various options for unicode characters like the copyright symbol.
    Ultimately the content I am trying to display is coming from a web service.
    I resolved this issue by calling the web service from my backing bean instead of using ui:include on the webservice call directly in the page.
    for example:
    public String getFooter() throws Exception
              HttpClient httpclient = new HttpClient();
              GetMethod get = new GetMethod(url);
              httpclient.executeMethod(get);
              String response = get.getResponseBodyAsString();
              return response;
         }I'd still love to have a solution to the page usage of the unicode characters, but for the time being this solves my problem.

  • FF character encoding issue in Mageia 2 ?

    Hi everyone,
    I'm running Mozilla Firefox 17.0.8 in a KDE distro of Linux called Mageia 2. I'm having problems in character encoding with certain web pages, meaning that certain icons like the ones next to menu entries (Login, Search box etc.) and in section headlines don't appear properly. Instead they appear either in some arabic character or as little grey boxes with numbers and letters written in it.
    I've tried experimenting with different encoding systems: Western (ISO 8859-1), (ISO 8859-15), (Windows 1252), Unicode (UTF-8), Central European (ISO 8859-2) but none of them does the job. Currently the char encoding is set to UTF-8. The same web page in Chrome (UTF-8) gives no such problem.
    Can you help me, please?

    Thank you!
    I solved my problem, however I find fonts are too small for certain web pages when compared to Chrome (see attached pictures of nytimes.com).
    Chrome's font size are set to "Medium".

  • Integration Gateway - JDBC - Character Encoding Issue

    Hello,
    I'm using SMP 3.0 SP06 and I'm getting data from MS SQL using JDBC interface and I can get all data successfully.
    The problem is:
    there is a column in database contain "Arabic" data "right-to-left" language.
    and when executing the OData service, for example, if the data in arabic is "هذه للتجربة" it is getting to me "هذه للتجربة"
    I think this is the same data but in a different encoding/decoding.
    Do you have any idea ?
    Thanks
    Hossam

    By the way, I have checked it again it is working fine when requesting data in XML format "default"
    The problem occurs only when requesting the service with format parameter "?$format=json"
    and it is even working fine when calling it from "Advanced REST client"
    so I think it is just a problem in the browser while displaying the data, specially chrome as it is working fine with IE, as chrome is displaying json files as plain text without any formatting or decoding, but IE is saving the file on PC and if I tried to open it by notepad++ I find data correctly decoded.
    It seems it is not an SMP nor Integration gateway issue, sorry for confusing

  • XML special character/encoding problem

    Hi
    I would like to store XML in a MSSQL database into a column with the datatype xml.
    It seems like the xml datatype in an xMII transaction allways is stored with encoding type UTF-8
    and the MSSQL xml datatype is UTF-16. This gives me some problem with special characters when inserting into the MSSQL database (in the example below is the MSSQL datatype xml):
    INSERT INTO
         VALUES
    The error returned is this:
    "com.microsoft.sqlserver.jdbc.SQLServerException: XML parsing: line 1, character 62, illegal xml character"
    If I replace the 'ä' with a normal 'a' the command executes ok.
    I am currently using a workaround that looks like this, when setting the parameter in my transaction:
    stringreplace(Local.test, " encoding=" & doublequote & "UTF-8" & doublequote, "")
    But I was hoping I could get rid of the stringreplace.
    Is there a solution / recommended way of doing this?
    Best Regards
    Simon Bruun
    Edited by: Simon Bruun on Mar 4, 2011 10:43 AM

    I solved this. I convert to Unicode UTF-8

  • Multi-byte character encoding issue in HTTP adapter

    Hi Guys,
    I am facing problem in the multi-byte character conversion.
    Problem:
    I am posting data from SAP CRM to third party system using XI as middle ware. I am using HTTP adapter to communicate XI to third party system.
    I have given XML code as UT-8 in the XI payload manipulation block.
    I am trying to post Chines characters from SAP CRM to third party system. junk characters are going to third party system. my assumption is it is double encoding.
    Can you please guide me how to proceed further.
    Please let me know if you need more info.
    Regards,
    Srini

    Srinivas,
    Can you go through the url:
    UTF-8 encoding problem in HTTP adapter
    ---Satish

  • Arabic character encoding issue

    Hi,
    Our JMS plugin receives xml text messages encoded in ISO-8859-1. Messages are originally in ISO-8859-6 and converted to ISO-8859-1 before put in the queue. I convert the message to ISO-8859-6 on receipt as below.
    For some unknown reason, some of the arabic characters show ? marks but some gets properly converted.
    ���� � ������ original
    ��?�� ������ after conversion
    Our plugin runs on Solaris 10. JDK used is 1.5. Solaris local set to en_US.ISO8859-15
    Code
    String in = ((TextMessage)message).getText ();
    String msgISO6 = new String (in.getBytes("ISO-8859-1"), "ISO-8859-6");
    Does anyone have any thoughts on the possible cause of the issue ?
    Thanks in advance
    Sohan

    String in = ((TextMessage)message).getText ();
    String msgISO6 = new String (in.getBytes("ISO-8859-1"), "ISO-8859-6"); That's wrong. Read the javadoc for the String class. in.getBytes gives a byte array with encoding ISO-8859-1, you are then trying to decode that array using "ISO-8859-6". The second argument in the string is not the encoding of the string that you are creating. A String is always encoded using Unicode/UTF-16.
    That means that you don't need to convert the encoding at all if you are gettings Strings, but you need to specify encoding when you are displaying them in e.g. a web page.
    Kaj

  • Special character deletion issue

    Hi all,
    I have a text field which contains special characters  which are not being entered by key board. I would like to repalce those chars. system is displaying these spl chars as '#'. I have used function module 'SCP_REPLACE_STRANGE_CHARS' , but it is replacing all the specail chars to dots(.), due to this if I have any valid chinese or japanese chars those are also getting replaced.
    pls any can help me out to resolve this issue.
    Thanks in advance
    reddi

    hi,
    if you want use this fm
    code as like below
    check this program
    RFFORI99
    or
    use like below
    form special_character  changing p_text_in p_text_out.
      call function 'SCP_REPLACE_STRANGE_CHARS'
        exporting
          intext                  = p_text_in
        INTEXT_LG               = 0
        INTER_CP                = '0000'
        INTER_BASE_CP           = '0000'
        IN_CP                   = '0000'
        REPLACEMENT             = 46
        importing
          outtext                 = p_text_out
        OUTUSED                 =
        OUTOVERFLOW             =
       exceptions
         invalid_codepage        = 1
         codepage_mismatch       = 2
         internal_error          = 3
         cannot_convert          = 4
         fields_not_type_c       = 5
         others                  = 6.
      if sy-subrc <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
      endif.
    endform.                                        " special_character
    Thanks and regards
    durga.K

  • Wrong special character encoding for window.open

    Hi all,
    I have a problem with special characters in urls.
    See the next code:
                      <h:outputLink onclick="window.open('#{webApplication.root}/dynamic/reports/user_efforts.jsf?i=1&person=#{item.person}','userefforts','width=600,height=300,resizable=yes,scrollbars=yes');return false" value="#" rendered="#{!empty item.projectNumber}">
                             <h:outputText value="#{item.person}"/>
                        </h:outputLink>In that, when item.person contains an �, then in the url it is converted to &eacute. The result is obviously that the person param is wrong for my popup,
    I suspect it has something to do with encoding, but both pages are UTF-8.
    Anyone can get me out of this mess? ;)
    Thanks

    Perhaps you should use "value" instead of "onclick".
    Think it makes no sense that an renderer should encode onclick - more an passthrough.
    In that case you also loose session id if cookie is disalbed.
    just try something like
    <h:outputLink value='yourURL' onclick='this.href;return false' ...

  • Special Character Encoding Preferences

    I develop web pages and html emails for multiple countries and prefer to use the decimal encoding rather than the alphanumeric encoding for special characters. see: http://www.freelancedesigners.com/tools/ascii_html_encoding.cfm
    Dreamweaver can either place the actual charcter (É) or the alphanumeric (&Egrave;), but not the decimal (&#200;)
    Is there a way to set the decimal as default? or convert an entire page without doing a find replace?
    thanks!

    Perhaps you should use "value" instead of "onclick".
    Think it makes no sense that an renderer should encode onclick - more an passthrough.
    In that case you also loose session id if cookie is disalbed.
    just try something like
    <h:outputLink value='yourURL' onclick='this.href;return false' ...

  • Shockwave to Javascript - character encoding issue !

    Hi !
    I have resigned from sending messages from JavaScript to
    Shockwave movie
    as I have found all existing methods unreliable (at worst
    scenario Flash
    blocker is installed and using localConnection trick with
    Flash gateway
    fails).
    But in the consequence, I have to send message (a search
    string) from
    Shockwave to JavaScript.
    That seems easy with the following Lingo:
    goToNetPage("javascript:void myJSfunction('" & aString
    But the problem is with encoding possible non-ASCII
    characters.
    I presume the browser page is using charset=UTF-8.
    Any idea how to properly encode 'aString' so it will preserve
    non-ASCII
    characters while being transfered to JavaScript?
    It is really urgent!
    Rgs,
    Ziggi

    > I have resigned from sending messages from JavaScript to
    Shockwave movie
    I replied a little late to your earlier thread, but take a
    look at
    <
    http://dasdeck.de/staff/valentin/lingo/dir_js/>

  • Special character loading issue with SQL Loader

    Hi,
    I am getting the special characters as part of my input file to SQL loader. The problem is that because the length of the input string (having special characters) is more than 1000 in input data, the converted bytes length is becoming more than 4000 and the defined substrb() function in control file is not working for it and the target 'ADDR_LINE' column in table 'TEST_TAB' is defined Varchar2(1024 char).
    Following is a sample ctl file and data i am using for it.
    LOAD DATA
    CHARACTERSET UTF8
    INFILE 'updated.txt'
    APPEND INTO TABLE TEST_TAB
    FIELDS TERMINATED BY "|~"
    TRAILING NULLCOLS
    INDX_WORD ,
    ADDR_LINE "SUBSTRB(:ADDR_LINE , 1, 1000)",
    CITY
    following is the actual data which i am receiving as part of input file to sql loader for the ADDR_LINE column:
    'RUA PEDROSO ALVARENGA, 1284 AƒAƒA‚AƒAƒA‚A‚AƒAƒAƒA‚A‚AƒA‚A‚AƒAƒAƒA‚AƒAƒA‚A‚A‚AƒAƒA‚A‚AƒA‚A‚AƒAƒAƒA‚AƒAƒA‚A‚AƒAƒAƒA‚A‚AƒA‚A‚A‚AƒAƒA‚AƒAƒA‚A‚A‚AƒAƒA‚A‚AƒA‚A‚A‚AƒAƒA‚AƒAƒA‚A‚AƒAƒAƒA‚A‚AƒA‚A‚AƒAƒAƒA‚AƒAƒA‚A‚A‚AƒAƒA‚A‚AƒA‚A‚A‚AƒAƒA‚AƒAƒA‚A‚AƒAƒAƒA‚A‚AƒA‚A‚A‚AƒAƒA‚AƒAƒA‚A‚A‚AƒAƒA‚A‚AƒA‚A‚A– 10 ANDAR'
    My database is having following settings.
    NLS_CALENDAR     GREGORIAN
    NLS_CHARACTERSET     AL32UTF8
    NLS_COMP     BINARY
    NLS_CURRENCY     $
    NLS_DATE_FORMAT     DD-MON-RR
    NLS_DATE_LANGUAGE     AMERICAN
    NLS_DUAL_CURRENCY     $
    NLS_ISO_CURRENCY     AMERICA
    NLS_LANGUAGE     AMERICAN
    NLS_LENGTH_SEMANTICS     BYTE
    NLS_NCHAR_CHARACTERSET     AL16UTF16
    NLS_NCHAR_CONV_EXCP     FALSE
    NLS_NUMERIC_CHARACTERS     .,
    NLS_SORT     BINARY
    NLS_TERRITORY     AMERICA
    NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_TIME_FORMAT     HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    Any help in this regard will be much appreciated.
    Thanks in advance.

    Is the data file created directly on the Unix server ? If not, how does it get to the Unix server ? And where does the file come from ? Is the UTF8 locale installed on the Unix server (check with the Unix sysadmin) ?
    HTH
    Srini

Maybe you are looking for

  • Audio playback stops short

    Has anyone had problems with audio playback in a previewed or published presentation where the playback stop short and advanced to the next slide?

  • Problem with Polycom PVX

    When using Polycom PVX windows based IP video conferencing software the remote site can see and here me but I am unable to see and hear them. The VC working correctly with an older Netgear wireless router and if I bypass TC completly and connect dire

  • How can I access the Setup Assistant on a new HD?

    How can I access the Setup Assistant on a new HD? We had a new HD installed but it didn't  show the Setup Assistant.

  • Broken LVM setup due to recent systemd upgrade

    So yeah, a couple of days ago I let the system upgrade and rebooted. My setup is basically two identical HDDs in a RAID 0 array + LVM. After the recent upgrade, boot hangs somewhere in the reading of the init ramdisk and tells me that it cannot find

  • Trying to burn sd disc get message HVDVD-TS founk in build Location

    what does this mean? I've used compressor to generate a mpeg file but cant burn