No german umlauts in OHW 2.0.2 topic view

I used OHW 1.1.5 before. Now switching to OHW 2.0.2, in the table of contents the german umlauts are not displayed correctly. The rest of the installation is identical. Its the same OS, the same JDK, the same Tomcat servlet engine and the same client web browser.
This phenomenon is independent of using a navigator Alias or not. I can't find any option in the documentation, which allows to configure the TOC output. How can german umlauts displayed in the TOC with OHW 2.0.2?
Thanks in advance
Michael

Do they display correctly in the other navigators? (i.e. does the keywords show them correctly?). In addition, do they show up in your topic pages (the actual help content)?

Similar Messages

  • German Umlaute don't work after upgrading to Yosemite with Japanese Romaji Layout set on U.S.

    Hi everyone,
    previously i could just use [ALT] + [U] and then press [A] / [a] / [U] / [u] / [O] / [o] to generate the german umlaut version.
    After installing Yosemite this does not work anymore, which is kinda frustrating. Cause it forces me to add another keyboard layout and switch when i have to write a german umlaut.
    I was previously using 'Input Source' > 'Japanese' > Enabled Romaji (Romaji layout was set to U.S.) & Hiragana.
    For testing purpose i added the Standard U.S. Layout and writing the Umlaut with the keyboard shortcut still works there. So that's why i think this is a Romaji specific bug.
    Has anyone any idea what's wrong ? Did Apple change something ?
    Best Regards

    Before I stumbled upon this thread, there was no flag at all. I had the character viewer there because pre-Yosemite it was the only way to access emoji (and things like arrows and other special symbols) in many apps (and still is for my Adobe CS5 apps). Additionally, I would use the keyboard viewer from time to time as a way to learn which keys to use for other special symbols (bring up the keyboard viewer and press option or shift-option or other combinations and it would show you a preview of what each key would generate).
    While reading this thread, I added more input sources (US International and US Extended) and would see the flags. The flags replaced the previous character/keyboard viewer menu icon. I switched between each flag to test my issue under each, to no avail.  Then I tried adding the Spanish keyboard but still had the same issue.
    Now I've removed all the input sources except the primary US keyboard. There is no longer a flag in my menu, it's gone back to the character viewer icon.
    The issue is still there. When I type option-r, u, or g, in any app nothing happens at all.
    Unfortunately, neither auto-substitution nor press-and-hold work in Adobe CS5 apps, which is where I need to use these special characters the most.
    Thank you for reaching out

  • Numbers don't show german umlaut but quick view does

    I do have a problem opening CSV files in Numbers.
    I have a shared folder with VirtualBox. The virtual HD is formatted in FAT32. With Windows I created a CSV file which contains german umlauts like ÄÖÜ.
    When I open the CSV with the "quick view" (press space when file is selected in finder, the CSV is displayed as a nice Table and all Umlauts are OK.
    When I double click the CSV file to open it in Numbers, the Umlauts are not displayed. Opening in Text Edit is the same result, no Umlauts.
    ö = ^
    ß = fl
    ü = ¸
    ä = ‰
    I think this is an encoding problem, but why is "quick view" displaying it correct. I'm confused.
    Thanks for any help.

    Hello
    I cannot find any encoding settings for opening files in Numbers 09.
    A simple and obvious solution would be to save CSV file in UTF-8 in FAT32 volume in VirtualBox if possible.
    One other method is to copy the CSV file to HFS+ volume and set its extended attributes to specify its text encoding. This method is only changing the metadata of the file and so is less intrusive than to convert the data itself to different text encoding. When text encoding extended attribute is properly set, Numbers 09 (and TextEdit.app set to use "Automatic" encoding option in opening files) will open it in the specified text encoding.
    The following AppleScript script might help to set the extended attributes (OSX 10.5 or later only). Recipe is as follows.
    A1) Open /Applications/Utilities/AppleScript Editor.app, copy the code of set_latin1_xattr.applescript listed below and save it as application (bundle) in the folder where target CSV files reside.
    A2) Double click the saved applet and it will set the com.apple.TextEncoding extended attribute of the *.csv and *.txt files in the same folder where the applet resides to Latin 1 when they are recognised as 'ISO-8859 text' by file(1) command.
    -- set_latin1_xattr.applescript
    _main()
    on _main()
        set p2m to (path to me)'s POSIX path
        if p2m ends with "/" then set p2m to p2m's text 1 thru -2
        set sh to "
    # set current directory to parent directory of script (or die)
    cd \"${0%/*}\" || exit 1
    # add com.apple.TextEnncoding xattr for Latin-1
    # for *.csv and *.txt files, of which file info contains 'ISO-8859 text', in current directory
    for f in *.{csv,txt}
    do
        if [[ $(file \"$f\") =~ 'ISO-8859 text' ]]
        then
            xattr -w com.apple.TextEncoding \"WINDOWS-1252;$((0x0500))\" \"$f\"
        fi
    done
        do shell script "/bin/bash -c " & sh's quoted form & " " & p2m's quoted form
    end _main
    Another method is to convert the text encoding itself as you have already done using TextEdit.app.
    The following AppleScript might help to convert text encoding from latin-1 to utf-8 in bulk. Recipe is basically the same as the above.
    B1) Open /Applications/Utilities/AppleScript Editor.app, copy the code of convert_latin1_to_utf8.applescript listed below and save it as application (bundle) in the folder where target CSV files reside.
    B2) Double click the saved applet and it will convert the text encoding of the *.csv and *.txt files in the same folder where the applet resides to UTF-8 when they are recognised as 'ISO-8859 text' by file(1) command. Also it sets the extended attribute for text encoding accordingly.
    -- convert_latin1_to_utf8.applescript
    _main()
    on _main()
        set p2m to (path to me)'s POSIX path
        if p2m ends with "/" then set p2m to p2m's text 1 thru -2
        set sh to "
    # set current directory to parent directory of script (or die)
    cd \"${0%/*}\" || exit 1
    # make temporary directory
    temp=$(mktemp -d /tmp/\"${0##*/}\".XXXXXX) || exit 1
    # convert text encoding from ISO-8859-1 to UTF-8
    # for *.csv and *.txt files, of which file info contains 'ISO-8859 text', in current directory
    for f in *.{csv,txt}
    do
        if [[ $(file \"$f\") =~ 'ISO-8859 text' ]]
        then
            iconv -f ISO-8859-1 -t UTF-8 \"$f\" > \"$temp/$f\" \\
            && xattr -w com.apple.TextEncoding \"UTF-8;$((0x08000100))\" \"$temp/$f\" \\
            && mv -f \"$temp/$f\" \"$f\"
        fi
    done
    # clean up temporary directory
    rm -rf \"$temp\"
        do shell script "/bin/bash -c " & sh's quoted form & " " & p2m's quoted form
    end _main
    Note that the extended attribute is not supported in FAT32 and so the above methods only work in HFS+ formatted volume.
    Scripts are briefly tested with Numbers 2.0.5 under OSX 10.5.8 and 10.6.5. Please make sure you have backup CSV files before applying the above scripts.
    Good luck,
    H

  • Legend map request with german umlaut

    Hello,
    i have a problem sending a legend map request that contains a german umlaut.
    I'm trying to use the pl/sql function for sending and parsing a xml request from the mapviewer documentation (chapter 3.1.16, example 3-19).
    my request looks like this
    <map_request datasource="mvdemo" format="PNG_URL">
      <legend bgstyle="fill:#ffffff;stroke:#ffffff" profile="MEDIUM">
        <column>
          <entry style="C.YELLOW" text="Übernahme"/>    
        </column>
      </legend>
    </map_request>The resulting image does not contain the umlaut Ü but a little square instead.
    I tried replacing the Ü with Ü ;
    <map_request datasource="mvdemo" format="PNG_URL">
      <legend bgstyle="fill:#ffffff;stroke:#ffffff" profile="MEDIUM">
        <column>
          <entry style="C.YELLOW" text="Übernahme"/>    
        </column>
      </legend>
    </map_request>This works when i manually send the request on the mapviewer map request page, but it doesn't work when i use the pl/sql function.
    The documentation says that the request has to be url encoded, so i tried using utl_url.escape on the whole request and on only the umlaut. Both doesn't work.
    Thanks for help in advance,
    Dirk

    Hi Cleber,
    which MapViewer version are you using? This may be related with the PDF graphics library version. This library has been modified to better handle the legend texts. I tested your request and it appears to be OK with the current library. You can send me (by email) a picture of your legend, and I will confirm if this has been fixed.
    Joao

  • German umlaut in idoc to file scenario

    Hi,
    in our scenario we send MATMAS idocs to XI, map them and create a file using file adapter.
    Settings in file receiver communication channel: file type = text and  file encoding = UTF8. I also tried file encoding = ISO-8859-1 - both with the same result:
    German umlauts are not converted. E.g. the material short text "Hängematte" is shown as "Hängematte" in the created file. The receiving system errors out. Acceptable would be
    "H& #228;ngematte". (of course without the blank but I have to add it herer else the forum would replace this by "ä" )Any help is appreciated.
    Regards,
    Philipp

    Hi,
    Follow the steps:-
    1>In your adapter -> Set File type to TEXT -> use Encoding and provide ISO-8859-1 (http://en.wikipedia.org/wiki/ISO/IEC_8859-1)
    To know more about other encoding standards refer -
    http://en.wikipedia.org/wiki/Character_encoding
    2>file type==>binary
    Regards,
    AshwinM
    Reward If helpful

  • Adobe Dreamweaver FTP connection doesn't work with german "Umlaut" like "ö" in the severadress

    Hi
    Today I tried to establish a FTP connection to the serveradress "ftp.möbelverwandlung.com", but I get the error message that the program couldn't connect to the host. As Dreamweaver works perfectly with other serveradresses and I managed to establish a connection to "ftp.möbelverwandlung.com" with other programs, I think that the problem is the german "Umlaut" "ö". Did anyone recognize the same problem and does someone know a solution?
    Thank you very much for your help

    Here is the English translation of your previous post -
    Hello Maximilian,
    I got a even looked at my provider. Usually there must pay the original domain for the agreed period and agree a new. However, they show a accommodating for so short "duration" of a few hours like you. Try to contact your ISP phone support.
    MfG
    Hans-Günter
    Anyhow, I think you have covered the options!

  • IPhoto doesn`t work correctly with german "umlaut" (ä,ö,ü) in keywords

    iPhoto doesn`t work correctly with german "umlaut" (ä,ö,ü) in keywords. They get to ä̈,ö̈, ü̈ when I write the keyword a second time. Can anybody tell me why. Thank you. Heinz

    It apparently does not like non-standard English characters - use A-Z and 0-9 adn you should be OK
    report to Apple - iPhoto menu ==> provide iPhoto feedback
    LN

  • Encoding Problem: Losing German Umlaute from gathering data from Oracle 8i

    my problem does concerns the diplay of german Umlaute such as äöüß etc. The OS is NW65 out of the box with Apache 2.0.49 and tomcat 4.1.28, JVM 1.4.2_02 and JDBC-driver Oracle 8i 8.1.7 JDBC Driver.
    The Data containing Umlaute which are retreived from the Database does somehow lose it´s Umlaute. The Umlaut which are coded in the servlet directly in order to generate the regular HTML does display the Umlaute without any problem.
    The same servlet and request from a Unix Enviroment does work fine. We have checked all Codepage settings (Java, NetWare, Tomcat etc).

    Hi Sven and Ingmar,
    I will try to kill 2 birds with one stone here. First of all you should check the definition of your current database character set and make sure that it can support all the characters that you need to store in your database.
    Please check out
    http://www.microsoft.com/globaldev/reference/iso.asp
    WE8ISO8859P9 is for Turkish and WE8ISO8859P1 is for western European (without Euro support).
    Next you need to set your client NLS_LANG character set (eg. for your SQL*Plus seesion), so that Oracle can convert your client operating system characters correctly into your database character set. The NLS_LANG character set also determines how SQL*Plus interpret/display your characters upon retrieval from the db.
    In both of your cases , I believed that the client NLS_LANG setting was not defined , hence this was defaulting to AMERICAN_AMERICA.US7ASCII. This is telling SQL*PLUS that the your client operating system can handle ASCII data only , hence all the accented characters are converted to ASCII during insertion into the database. Likewise upon data retrieval.
    If you are running SQL*PLUS client on English Windows then you NLS_LANG character set should be WE8MSWIN1252 and for Turkish Windows it should set to TR8MSWIN1254 .
    If the client character set and the database character set are the same , Oracle does not perform any data conversion, that's why if you use a US7ASCII database and set you NLS_LANG character set to US7ASCII .then you can insert all the accented Latin characters , Japanese , Chinese etc. This configuration is not supported, and it can cause many issues.
    For more information on character set configuration and possible problems.
    Please check out the white paper Database Character set migration on http://otn.oracle.com/products/oracle8i/content.html#nls
    And the Globalization Support FAQ at:
    http://technet.oracle.com/products/oracle8i/
    Regards
    Nat
    null

  • German umlauts :Sender File Adapter

    Hi,
    I am trying to configure Sender File adapter,but it is not working properly for German umlauts.
    I have tried encoding UTF-8, UTF-16, ISO-8859-1,ISO-8859-2, ISO-8859-5 but none of them are working.
    Can anyone please help me out in this.
    Thanks a lot.
    Regards,
    Shweta

    Hi Stefan,
    I have already added this module exactly in the same way.
    But can you please confirm the sequence
    Module :
    localejbs/AF_Modules/PayloadSwapBean                     Local Enterprise Bean     swap
    localejbs/AF_Modules/MessageTransformBean     Local Enterprise Bean     Transform
    localejbs/AF_Modules/MessageTransformBean     Local Enterprise Bean     Transform
    Transform.ContentType                     text/plain;charset="UTF-8"
    Transform.ContentDisposition            attachment;filename="ABC.csv"
    Also,I noticed one thing
    The content is corrupted in SXI_MONITOR but it is correct in payload of File as well as Mail adapter in Message Monitor(RWB)
    Thanks.
    Regards,
    Shweta

  • German Umlauts OK in Test Environment, Question Marks (??) in production

    Hi Sun Forums,
    I have a simple Java application that uses JFrame for a window, a JTextArea for console output. While running my application in test mode (that is, run locally within Eclipse development environment) the software properly handles all German Umlauts in the JTextArea (also using Log4J to write the same output to file-- that too is OK). In fact, the application is flawless from this perspective.
    However, when I deploy the application to multiple environments, the Umlauts are displayed as ??. Deployment is destined for Mac OS X (10.4/10.5) and Windows-based computers. (XP, Vista) with a requirement of Java 1.5 at the minimum.
    On the test computer (Mac OS X 10.5), the test environment is OK, but running the application as a runnable jar, german umlauts become question marks ??. I use Jar Bundler on Mac to produce an application object, and Launch4J to build a Windows executables.
    I am setting the default encoding to UTF-8 at the start of my app. Other international characters treated OK after deployment (e, a with accents). It seems to be localized to german umlaut type characters where the app fails.
    I have encoded my source files as UTF-8 in Eclipse. I am having a hard time understanding what the root cause is. I suspect it is the default encoding on the computer the software is running on. If this is true, then how do I force the application to honor german umlauts?
    Thanks very much,
    Ryan Allaby
    RA-CC.COM
    J2EE/Java Developer
    Edited by: RyanAllaby on Jul 10, 2009 2:50 PM

    So you start with a string called "input"; where did that come from? As far as we know, it could already have been corrupted. ByteBuffer inputBuffer = ByteBuffer.wrap( input.getBytes() ); Here you convert the string to to a byte array using the default encoding. You say you've set the default to UTF-8, but how do you know it worked on the customer's machine? When we advise you not to rely on the default encoding, we don't mean you should override that system property, we mean you should always specify the encoding in your code. There's a getBytes() method that lets you do that.
    CharBuffer data = utf8charset.decode( inputBuffer ); Now you decode the byte[] that you think is UTF-8, as UTF-8. If getBytes() did in fact encode the string as UTF-8, this is a wash; you just wasted a lot of time and ended up with the exact same string you started with. On the other hand, if getBytes() used something other than UTF-8, you've just created a load of garbage. ByteBuffer outputBuffer = iso88591charset.encode( data );Next you create yet another byte array, this time using the ISO-8859-1 encoding. If the string was valid to begin with, and the previous steps didn't corrupt it, there could be characters in it that can't be encoded in ISO-8859-1. Those characters will be lost.
    byte[] outputData = outputBuffer.array();
    return new String( outputData ); Finally, you decode the byte[] once more, this time using the default encoding. As with getBytes(), there's a String constructor that lets you specify the encoding, but it doesn't really matter. For the previous steps to have worked, the default had to be UTF-8. That means you have a byte[] that's encoded as ISO-8859-1 and you're decoding it as UTF-8. What's wrong with this picture?
    This whole sequence makes no sense anyway; at best, it's a huge waste of clock cycles. It looks like you're trying to change the encoding of the string, which is impossible. No matter what platform it runs on, Java always uses the same encoding for strings. That encoding is UTF-16, but you don't really need to know that. You should only have to deal with character encodings when your app communicates with something outside itself, like a network or a file system.
    What's the real problem you're trying to solve?

  • Mail Adapter - PayloadSwapBean - MessageTransformBean - German umlauts

    Hi there,
    I'm receiving mails with an attachment (.csv / .txt) that I want to process to get IDocs. Everything works fine but the conversion of German umlauts. I tried to apply several charsets (i.e. iso-8859-1, iso-8859-2, utf-8) in the contentType parameter without success. The result in my payload after swapping and transforming is a message without umlauts. All these characters have been replaced by the same 'character' that looks like a quadrangle. Therefore even the earliest possible mapping comes too late to convert this character back into umlauts, because I don't know anymore the original ones.
    When I process the same attachment with a <u>file</u> <u>adapter</u> in the same manner (until getting an IDoc) there are no problems with umlauts, the payload looks fine!
    I even checked the note 881308 (although it's said to be for the mail receiver) but it's already in the system (XI 3.0, SP 14)
    Anyone an idea to solve my problem?
    Regards,
    Ralph

    Hi Ralph,
    now I got the solution:
    Allpy the MessageTransformBean twice.
    First you set the code page for the mail attachment, how it comes to the system.
    Then you do the conversion and set the code page how the target xml should be.
    Make two entries in Module configuration:
    localejbs/AF_Modules/MessageTransformBean - contenttype
    localejbs/AF_Modules/MessageTransformBean - tranform
    as paramters you set:
    contentType - Transform.ContentType - text/plain; charset=iso-8859-1
    transform     - Transform.ContentType - text/xml; charset=UTF-8
    transform     - Transform.Class           - com.sap.aii.messaging.adapter.Conversion
    and so on.
    The problem is that outlook does not provide the content type for the attachment, so the MailTransformBean assumes UTF-8, but the attachment has iso-8859-1, so you have to set this before the conversion.
    I have tested this with XI 3.0 SP17 with note 960501 included.
    Regards
    Stefan

  • Wrong email display name with german umlauts (MS Exchange 2003)

    We use 6 iPhones with Exchange 2003 and get wrong email display names with german umlauts (ä,ü,ö) - but the email-body is right.
    We get special characters instead of umlauts, so the display name split into pieces. Anwering is not posipble - we get a failure-message.
    We changed the standard-internet-mailformat on the exchange-server to unicode utf-8. First it works fine, after a few hours the names displays wrong again.
    So we use this hotfix:
    http://support.microsoft.com/?scid=kb%3Ben-us%3B916299&x=11&y=13
    Same result: First it works fine, after a few hours the names displays wrong again at the iPhone.
    Any ideas?

    Do you have commas in the display name? We used to have "Müller, Thomas" <[email protected]> and then got the split up and special characters you mention. Tests have shown that when leaving out the comma in the display name, e.g. "Thomas Müller" <[email protected]>, everything worked fine.
    Guess it's a question of whether a company wants to change its naming convention for a few iPhone users...
    HCD

  • Wrong coding of german umlaut in XML/XSLT output

    I'm using Oracle 8.1.5 with XSQL Servlet.
    The contents of my database contains german
    umlaut. Visiting the HTML output using
    Netscape/WinNT seems to be correct, but when
    I save the page the saved page contains a
    unknown (for me!) hex conversion of the
    umlaut and not '&auml;' etc. The stylesheet
    uses enconding ISO-8859-1.
    I think, the problem will be found in the
    stylesheet, but I don't know where.

    Hi,
    I test on my Lync client and there is no issue about it.
    Please update to the latest version for Lync Server 2013 (the latest version is 5.0.8308.815) and then test again.
    http://support2.microsoft.com/kb/2809243
    Best Regards,
    Eason Huang
    Eason Huang
    TechNet Community Support

  • Howto set up proper utf-8 locales and german umlaute?

    Hi there,
    have some issues here and i think it has something to do with my locales... I have them since one of the last updates i think...
    First problem:
    In Kopete, i can send the german "umlaute" (ä, ö, ü) to others and they are sent and displayed correctly at their side... But when they send me these characters, i get only a square and a deleted next-letter-in-the-word by them. This happens with every theme and font combination in kopete. Here is an example:
    Second problem:
    In Konsole, the "umlaute" are correctly displayed on my folders, but in the messages i get the chars are simply deleted and not visible. This happens in the VC and in Xorg too. Here is another example:
    (It should mean "Keine Handbuch Seite für pacman.conf")
    My config:
    I think i have configured my System properly, well at least i hope that Here are the relevant parts of my config files:
    /etc/rc.conf:
    LOCALE="de_DE.utf8"
    KEYMAP="de"
    CONSOLEFONT=
    CONSOLEMAP=
    /etc/profile:
    export LANG="de_DE.utf8"
    export LANGUAGE="de_DE.utf8"
    /etc/locale.gen:
    de_DE.UTF-8     UTF-8
    en_US.UTF-8     UTF-8
    And "locale -a" gives me:
    C
    POSIX
    de_DE.utf8
    en_US.utf8
    The fonts I use:
    KDE GUI: Bitstream Vera Sans (everywhere)
    Y-Terminal (Konsole) Font: Bitstream Vera Sans mono
    So, how can i configure a proper german utf-8? I have already searched the forums (both here and the DE one) and the wiki, but found no solution to this....
    THX
    Funkyou

    Thx for the suggestions, but none of them seems to solve it...
    baze:
    This line was already uncommented and i have generated my locales for several times now... I have updated my post with the uncommented lines in /etc/locale.gen, just to collect all information...
    Romashka:
    Ok, but what if the two variables contain the same content? I mean, when i define LANG="de_DE.utf8" in /etc/profile and it is then overwritten by /etc/profile.d/lang.sh with LANG="de_DE.utf8", then this is simply the same variable defined twice with the same content... I have tried to unset the one in /etc/profile, but it did not solve it...
    As for CONSOLEFONT, i had never one defined, can anyone suggest me a proper one? And the other question is: Its not working on the terminals, but it is also not working in Xorg, so is this really a CONSOLEFONT issue?
    I have tried another thing and switched my locales to de_DE@euro     ISO-8859-15, and with this setting the chars appear correctly... But i cannot be the only one where it does not work, i feel so excluded without UTF-8
    Have also tried another terminal in Xorg... In XTerm the chars are not deleted like in Kopete or on the VC, but i see an inverted question mark instead of them...

  • Characters With German Umlaut in webservice

    Hello,
    i have a requirement wherein i have to pass Characters With German Umlaut in my web service but when i see the trace they are getting converted into dots.
    does anyone have solution for it?
    Regards,
    Gunjan

    Hi,
    Ideally this should not be a problem. Please provide more information.
    Where you are checking trace? (PI?)
    Are you getting response or request as dots?
    Can you check encoding?
    Check this out, to me it seems problem related to encoding: http://www.sqldbu.com/eng/sections/tips/utf8.html
    The problem mainly arise because of mismatching of encoding: ISO-8859-1 and UTF-8.
    so double check encoding type provided in request and response.
    Regards,
    Gourav

Maybe you are looking for