Problems with umlaut characters

Hi,
i´m on NW04 SP14 and ADS SP14.
I have implemented a interactive form which is bound to a context node of a webdynpro application. This context node contains some attributes of type string.
When I populate the attributes with some values containing umlaut characters the interactive form does not display them correctly (there is a space character instead of the umlaut character).
Is this a bug in the forms solution??
Thanks in advance.
Regards
Sebastian

Hi Markus,
´
the problem unfortunately still exists. Do you have any suggestions? Is it a bug or some misconfiguration in our environment?
Thanks for your support.
Regards
Sebastian

Similar Messages

  • Problems with special characters like "u00F6" "u00E4" "u00FC"

    Hi,
    we currently face some major problems with special characters like "ö" "ä" "ü". First I thought it is a SP02 specific problem but it still doesn't work with SP03.
    When I start an import the import manager shows correct converted value. "ö"s "ä"s "ü"s are displayed correctly. But after I've done the import I look into the Data manager and I see instead of these characters charcaters like "̈". (note: the character should show a little square box - but the forum cannot desplay it)
    How can I solve this problem?
    Regards
    Nico

    Hi Nico
      Just try it out the following steps, hope it solves ur issue.
    <b>To set one or more values back to their original values:</b>
    1. In the Import Manager Source Values grid, select the values that you want to set back to their original values.
    2. Right-click on one of the values and choose Set to Original Value from the context menu, or choose Values > Apply Value Conversion Filter > Set to Original Value from the main menu.
    3. MDME sets each value back to its original value, which appears in black in the Source Values grid.
    <b>NOTE : You can use the Set to Original Value filter only at the value-level and not at the field level.</b>
    thanks
    Alexander

  • Problems with special characters in InputStream or XPath

    Hello everyone,
    I am having problems with special characters such as ", ', - and so on, everything becomes ? (questionmark). (Not with special nordic characthers (å,æ,ø) though.)
    The encoding of the XML is UTF-8, and the server that holds the webservice is setting encoding to UTF-8 also, so no conflict there.
    What I have done is something like this:
    {code}
    String url = "http://www.formula1.com/rss/news/latest.rss"; // This is not the feed I use, it is a custom XML document
    InputSource is = new InputSource(url);
    DocumentBuilderFactory fct = DocumentBuilderFactory.newInstance();
    DocumentBuilder builder = fct.newDocumentBuilder();
    Document doc = builder.parse(is);
    Element rootElement = doc.getDocumentElement();
    XPathFactory factory = XPathFactory.newInstance();
    XPath xPath = factory.newXPath();
    out.print(xPath.evaluate("/rss/channel/title", rootElement)); // The xml is not a RSS feed, this is just an example.
    {code}
    Please let me know what I am doing wrong, or give me a hint of how to solve this in another way.
    Best regards,
    G.Hansen
    Edited by: G.Hansen on Mar 24, 2009 2:39 AM

    Hello, thanks for you reply. I had to rewrite your code a little, and I think that your code would not really compile, if you tried yourself :-)
    I ended up with something like this:
    our environment is running in 1.4 mode, so I could not use the for-each loop :-(
    public String printEntities(String s) {
         char[] sArray = s.toCharArray();
         StringBuffer sb = new StringBuffer();
           for (int i = 0; i < sArray.length; i++) {
             if (sArray[i] > 256)
                  sb.append("&#x" + Integer.toHexString(sArray) + ";");
         } else
              sb.append(sArray[i]);
         return sb.toString();

  • How to send Oracle rowid to servlet? | Problem with national characters.

    There is same possibility how to send rowid to servlet?
    I have now definition like this:
    <af:image source="/imageservlet?Par1=#{bindings.Col1.inputValue}"/>
    But If column contents national characters, servlet methods obtained changed these characters.
    My idea is to use not primary key for row, but use oracle rowid. It is simply possible?
    Use something like this:
    <af:image source="/imageservlet?Rowid=#{bindings.Rowid}"/
    Or Do you have ideas how to solve problem with national characters ?
    Thanks
    FiL

    Hi,
    Although your workaround works.
    I think this is a simple encoding problem.
    I simply need to make sure all parameters and pages are encoded with a char set which contains the national characters you mentioned.
    This is a bit dependent on the exact technology your using, but most can be done via the web.xml:
      <jsp-config>
          <jsp-property-group>
              <url-pattern>*.jsp</url-pattern>
              <page-encoding>UTF-8</page-encoding>
          </jsp-property-group>
      </jsp-config>     This forces all JSP pages to be encoded in UTF-8
    Adding the following parameter sometimes helps as well, although I think this one is a bit dated:
    You said your using a servlet so your servlet needs a similar block for its pattern
      <context-param>
        <param-name>PARAMETER_ENCODING</param-name>
        <param-value>UTF-8</param-value>
      </context-param>If you want to be 100% sure the encoding is set right make sure thepages contain:
    <%@ page contentType="text/html;charset=utf-8"%>Or depending on your view technology the syntax can be a bit different
    -Anton

  • Problem with national characters on windows client

    Hello there,
    I'am having problem with national characters on windows client.
    All national data stored in NVARCHAR2 colums, applications (.net) works fine,
    but in sqlplus:
    select city from test_table;
    - everything ok, sqlplus shows national characters
    select dump(N'<national symbols>') from dual
    - returns
    Typ=96 Len=12: 0,191,0,191,0,191,0,191,0,191,0,191
    select * from test_table where city = N'<national symbols> '
    - always returns nothing
    As i understand the problem in
    sql query text (and national literals) convertion
    to servers "WE8ISO8859P1" encoding, Is it possible
    to solve the issue?
    Thanks in advance
    PS.
    Console in right mode (chcp=1251)
    sqlplus shows russian messages well
    Server (oracle 9 on solaris):
    select * from nls_database_parameters
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_SAVED_NCHAR_CS WE8ISO8859P1
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET WE8ISO8859P1
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZH:TZM
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZH:TZM
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_RDBMS_VERSION 9.2.0.6.0
    Client (windows server 2003, oracle client 10):
    NLS_LANG = RUSSIAN_CIS.CL8MSWIN1251

    N'<national symbols>', being part of an SQL statement, will be converted to the database character set (WE8ISO8859P1) before being parsed. Only if the client and the database are both 10.2 or higher, the client can encode the literal appropriately so that it survives this conversion.
    In earlier versions, you can do the encoding yourself. Instead of the N'<national symbols>' literal use the UNISTR function: UNISTR('\xxxx\yyyy\zzzz'), where U+xxxx, U+yyyy, U+zzzz are Unicode code points of your national characters.
    -- Sergiusz

  • Loading a CSV file with Umlaut characters (àáä)

    Hai,
    We are uploading a CSV file though a Custom JSP page built based on Oracle JTF framework.
    The JSP page is loading the data into FND_LOBS table using JTF object, oracle.apps.jtf.amv.ServletUploader.
    The CSV file in the FND_LOBS table stored properly with the umlaut characters.
    Now the JSP page invokes a Java object to read and parse the data. We are selecting the data first into BLOB object and then using the Input Stream Reader to get the data.
    Here is the sample code:
    oraclepreparedstatement = (OraclePreparedStatement)oracleconnection.prepareStatement(" SELECT FILE_DATA FROM FND_LOBS WHERE FILE_ID = :1 ");
    oraclepreparedstatement.defineColumnType(1, 2004);
    oraclepreparedstatement.setLong(1, <file id>);
    oracleresultset = (OracleResultSet)oraclepreparedstatement.executeQuery();
    blob = (BLOB)oracleresultset.getObject(1);
    InputStreamReader inputstreamreader = new InputStreamReader(blob.getBinaryStream());
    lineReader = new LineNumberReader(inputstreamreader )
    lCSVLine = lineReader.readLine();
    I tried printing the character set used by the InputStreamReader and it returned as ASCII
    Now I tried setting the different character sets to read Umlaut characters(german chars) but nothing has worked.
    InputStreamReader inputstreamreader = new InputStreamReader(blob.getBinaryStream(),"UTF-8");
    Can someone please let me know where and how to set the Character Set to accept the Umlaut characters like àáä?
    Thanks,
    Anji

    Thank you for the quick response.
    Requirement:
    I need to retrive the BLOB object with umlaut characters from database, parse the data with comma delimeter into strings and store in database.
    I am viewing the umlaut data from the database table using TOAD utility tool.
    I tried the same code example provided above but it is not working as expected. The umlaut characters are translated to 'ýýý.
    CODE EXAMPLE:
    Input:
    test_umlaut (sno NUMBER, col1 VARCHAR2(100), col3 BLOB);
    insert into test_umlaut(sno,col3) values(200, utl_raw.cast_to_raw('äöüÄÖÜ' ))
    Note: Verified that the database is showing the umlaut characters on selecting the col3 and storing in a flat file
    --- code
    OraclePreparedStatement oraclepreparedstatement10 = null;
    OracleResultSet rs = null;
    oraclepreparedstatement10 = (OraclePreparedStatement)oracleconnection.prepareStatement(" SELECT col3 FROM test_umlaut WHERE sno = 200 ");
    oraclepreparedstatement10.defineColumnType(1, 2004);
    rs = (OracleResultSet)oraclepreparedstatement10.executeQuery();
    while(rs.next()) {
    BLOB b = (BLOB)rs.getObject(1);
    InputStream is = b.getBinaryStream();
    InputStreamReader r = new InputStreamReader(is,"utf8");
    BufferedReader br = new BufferedReader(r);
    String line;
    while( (line = br.readLine()) != null) {
         System.out.println(line);
         OraclePreparedStatement oraclepreparedstatement12 = null;
         OracleResultSet oracleresultset12 = null;
         oraclepreparedstatement12 = (OraclePreparedStatement)oracleconnection.prepareStatement(" INSERT INTO test_umlaut(sno,col1) VALUES (300,?) ");
         oraclepreparedstatement12.setString(1,line);
         oraclepreparedstatement12.executeUpdate();
    br.close();
    r.close();
    is.close();
    Output: Verified the output from the database table which is inserted in the loop above.
    select col1 from test_umlaut where sno=300
    ýýý

  • Application Hudba (Music) problem with Czech characters

    Hello,
    I have problem with Czech characters for Music files copied into application "Hudba" all Czech letters (š,č,ť...etc.) are translated into Japonies letters (which I could not read at all). Could someone help please.
    Thanks

     You might to consider contacting the app developers http://talk.sonymobile.com/t5/Music/Walkman-Beta-Test/m-p/953659#U953659

  • Problem with special characters (umlaut)

    Hi,
    I have a text box in my form.
    when I am entering the normal characters it working fine and the values are
    persisting quiet well in the database but when I am entering special characters (umlaut characters),
    it is not storing them properly in the database.
    I have checked in the logs at the application server side and the problem is while getting the values from
    the request. The JSP which is submitting the request I have explictly put the following line:
    request.setCharacterEncoding("iso-8859-1");
    But this also not seems to be working.
    Could you please let me know the resolution of this problem.
    Thanks in advance
    Amit

    Hi,
    Thanks for the reply.
    I have checked the database and there is no problem there.
    To check if I am getting the correct values from the JSP I have put the following code in my action class:
                                  if(columnName.equalsIgnoreCase("REMARKS"))
                                       log.debug("Character Encoding value of � is "+Character.getNumericValue('�'));
                                       log.debug("Character Encoding value of � is "+Character.getNumericValue('�'));
                                       log.debug("Character Encoding value of � is "+Character.getNumericValue('�'));
                                       log.debug("Character Encoding value of � is "+Character.getNumericValue('�'));
                                       log.debug("Character Encoding value of � is "+Character.getNumericValue('�'));
                                       log.debug("Character Encoding value of � is "+Character.getNumericValue('�'));
                                       String str = value.toString();
                                       char[] remarkArray = new char[str.length()];
                                       str.getChars(0, (str.length()-1), remarkArray, 0);
                                       for (int i = 0; i < remarkArray.length; i++)
                                            log.debug("Encoding value of "+remarkArray[i]+" is "+Character.getNumericValue(remarkArray));
    In the remarhs I am setting the value in the form as "������"
    and I am getting the following in the logs:
    Character Encoding value of &#9472; is -1
    Character Encoding value of � is -1
    Character Encoding value of &#9604; is -1
    Character Encoding value of � is -1
    Character Encoding value of � is -1
    Character Encoding value of � is -1
    Encoding value of &#9500; is -1
    Encoding value of &#9488; is -1
    Encoding value of &#9500; is -1
    Encoding value of &#9488; is -1
    Encoding value of &#9500; is -1
    Encoding value of &#9488; is -1
    Encoding value of &#9500; is -1
    Encoding value of &#9488; is -1
    Encoding value of &#9500; is -1
    Encoding value of � is -1
    Encoding value of &#9500; is -1
    Encoding value of is -1
    Kindly suggest what may be the issue??
    Amit

  • Problem with german characters (Umlaute) when working with JDev for Linux

    Hallo ...,
    i have a site in uix code under Linux. When i want to use ... confirm("...löschen...?") the JDeveloper can't handle the ö. When i want to use &ouml; i've got an error. It's only when i work under Linux. When i make the same under Windows i can use ä, ö, ü. What can i do?
    Any help is appreciated.
    Thank you

    Tom Gewecke wrote:
    w/ all the keyboards and language settings.
    What do you mean by that? The problem being talked about here is an inability to make French accented characters on a German keyboard. What is your problem exactly?
    I've posted here because I guess it may be the same bug.
    After upgrading I wasn't able to do special characters, including some accented vocals, like "ì", very common in italian language.
    Then I've added an english keyboard, removed the italian one and added it again, so the problem with missing accented vocals was solved.
    I still have the same problem with special consonants, like "z" and "c": if I hold my thumb on such letters, I have no options like, say, "ż" (such characters are very important for my field (linguistics) and if Apple won't solve this problem I'll be forced to look elsewhere).

  • Problems with Special Characters in DBs...

    We have just moved a set of tables from one db on one server to another db on another server. the only problem is that some special characters held within data on the tables aren't showing up properly on the new db. characters such as a circle with an 'R' in (as in registered trademark), or vowels with umlauts over them.
    THe db is used by a website and it is the development version of the website which uses the new db that is not displaying the characters properly. The characters are also displayed incorrectly within sql plus 3.3.
    If anyone could point out why the new db is not storing the characters correctly i would unmeasurably grateful.
    Thanking you in advance
    Ewan Gibb.

    Hello, thanks for you reply. I had to rewrite your code a little, and I think that your code would not really compile, if you tried yourself :-)
    I ended up with something like this:
    our environment is running in 1.4 mode, so I could not use the for-each loop :-(
    public String printEntities(String s) {
         char[] sArray = s.toCharArray();
         StringBuffer sb = new StringBuffer();
           for (int i = 0; i < sArray.length; i++) {
             if (sArray[i] > 256)
                  sb.append("&#x" + Integer.toHexString(sArray) + ";");
         } else
              sb.append(sArray[i]);
         return sb.toString();

  • Problem with some characters in complex objects

    Hi all,
    I've built a webservice which returns a complex object with several fields inside. All fields are public and accessable via getter and setter methods.
    The problem is, that some of these fields contains numbers or underscores in their names.
    For example:
    public int field_a;
    or
    public String house3of4;
    When I try to import these webservice as a model in a Web Dynpro project, it doesn't work until I remove these characters.
    Is this a known problem or is there any solution for it?
    Thanks
    Thomas

    NLS_LANG in registry is "ARABIC_UNITED ARAB EMIRATES.AR8MSWIN1256"
    I use oracle form 10g for developer
    oracle form 9i for database
    when I build a form in client side and make a text with farsi characters, when I run the form,all characters shows me correct in farsi except four characters(گ چ ژ پ)

  • Problem with international characters showing up as junk

    Hi All,
    Little question.
    I've made a xml data template which executes a query to fetch person names from the e-business suite tables.
    However there are international characters in the names which are showing up incorrectly. When executing the query in the database everything shows up correctly. But when the query is executed via XML publisher the produced XML contains junk characters.
    This is happening with for example o umlaut characters.
    The database characterset is: WE8ISO8859P1
    Version of XML publisher: 5.6.3
    Patrick

    This turned out to be an extra property which was set in the data template:
    property scalable_mode with value "on"
    This caused the special characters to be mangled.
    Patrick

  • Problems with Turkish characters not displaying correctly

    A user is having problems with certain Turkish characters in particular fonts not displaying correctly (as shown below).
    The two fonts (Gill Sans and Helvetica LT CondensedLight) are used for English language marketing material.  Ideally, for continuity, the same font would be preferred for our Turkish language marketing material rather than changing it to a similar font.
    Helvetica LT CondensedLight
    Turkish character “ İ ” does not display correctly in either Word or InDesign.
    Gill Sans
    Turkish characters “ ş ” and “ ğ ” display correctly in Word but not InDesign.
    My questions are:
    1. For Gill Sans, why are the characters not displaying in InDesign but are in Word?
    2. I've noticed here: http://www.webtype.com/font/gill-sans-family/#glyphs-tab that the required characters are available within the Gill Sans font.  Is this because the font is newer and includes the additional characters?  Would this work in InDesign?

    One of the big differences between InDesign and Word is that Word will happily substitute glyphs without telling you. So, unless your installs of Word and of InDesign are using different versions of Gill Sans (which is possible, BTW) then your s-with-cedilla and your g-with-macron are actually not in Gill Sans but in some other font. I'm guessing Arial.
    It's possible that you have more than one Gill Sans installed (e.g. one in your Document Fonts folder and another in your system fonts folder) that have different glyph complements. Gill Sans has been around a long time, and the version I have from the 90s has no Turkish support whatsoever.
    As far as the capital-I-with-dot, you can check your glyph coverage in a variety of ways. In both InDesign and Word, you can open up a window that shows you all of the glyphs in a font. In InDesign, find it in Type -> Glyphs. In Word, it's called "Insert Symbol" and you can find it in Word 2010 by going to Insert -> Symbol -> More Symbols. I am guessing that the answer to this question is simply that your cut of Helvetica LT has no cap-I-with-dot. You'd need to pick a font that actually has support for Turkish. On WIndows, I use a very full-featured freebie called BabelMap to check font coverage. They have a Web version here:
    http://www.babelstone.co.uk/Unicode/babelmap.html
    but I personally prefer the downloadable .exe file.

  • File XML Content Conversion: Problem with special characters

    Hello,
    in a file sender cc content conversion is used to transform a flat structure to XML. What we experiencecd is that the message mapping failed due to a character that was not allowed in XML:
    I was assuming that the file content conversion just creates XML messages with allowed characters. Is there any way to configure content conversion to remove control characters which are not allowed in XML? Unfortunately the sender system cannot be modified.
    Thank you.

    Hi Florian,
      Please use this UDF to remove special characters which prevent XML messages to form properly.
    public static String removeSpecialChar(String s)
              try
                   s=s.replaceAll("&","& amp ;");
                   s=s.replaceAll("<"  , "  & lt ;");
                   s=s.replaceAll(">", "& gt ;");
                   s=s.replaceAll("'", "& apos ;");
                   s=s.replaceAll("\"", "& quot ;");
              catch(Exception e)
                   e.printStackTrace();
              return s;
    Please remove spaces between characters within double quotes. I have added them because otherwise you can't see this code properly. Please check this below link , please replace the characters with proper values as the display is causing a problem here   
    http://support.microsoft.com/kb/316063
    regards
    Anupam
    Edited by: anupamsap on Jul 7, 2011 4:22 PM
    Edited by: anupamsap on Jul 7, 2011 4:23 PM

  • Problem with Polish Characters while downloading data using GUI_DOWLOAD

    Hi,
    I was trying to create a file on the presentation server using GUI_DOWLOAD.
    I am facing a problem with the data that contains polish characters.. like Ą , Ł..
    There is  a vendor numer in the system as FRĄCZEK, when i was tring to download, the vendor number is downloaded as FR¥CZEK, even i tried to download from SE16 also, still i am facing the same problem.
    I am facing the above problem with soem other polish characters..
    Is there any solution  to download the data properly.. ?
    Guys please help me.
    Thanks in advacnce.
    Regards,
    Sriram.
    Edited by: Srirama Murthy Maddirala on Oct 27, 2009 8:10 AM

    Hi Sriram,
    <li>Get the front-end code page using NLS_GET_FRONTEND_CP.
    <li>Pass the code page through GUI_DOWNLOAD.
    REPORT  ztest.
    TYPE-POOLS:abap.
    DATA:frontend_codepage TYPE  cpcodepage.
    DATA:codepage          TYPE  abap_encoding.
    "Get the Frontend Code page
    CALL FUNCTION 'NLS_GET_FRONTEND_CP'
      EXPORTING
        langu             = sy-langu
      IMPORTING
        frontend_codepage = frontend_codepage.
    "Pass the codepage through GUI_DOWNLOAD function module
    codepage = frontend_codepage.
    CALL FUNCTION 'GUI_DOWNLOAD'
      EXPORTING
        filename = 'C:\temp\data.xls'
        filetype = 'ASC'
        codepage = codepage
      TABLES
        data_tab = it_data.
    Thanks
    Venkat.O

Maybe you are looking for