Messages in Double Byte Language

I had created a message in english and then by logging into the required Double Byte Languages translated the text into Double Byte Language.The message is a warning message. Now the user wanted a pop-up to be dispalyed during the warning message, so in the options setting activated the "Dialog Box at Waring Message".The Activate Multibye Functions to support is also enabled. Now the issue is that when the message is displayed , the pop-up displays the correct text for all languages except the double byte languages, for double byte it displays "???????????".If I disable the "Activate Multibye Functions to support", garbage values are displayed for the double byte languages.Can someone tell me how do I solve this?

Hi!
This does not sound like a programming issue. First check / update the sapgui to the newest version, if this does not help ask OSS. I guess, in general your double byte characters are displayed correct, otherwise I would check language installation of SAP, maybe also font installation of the GUI computer.
Regards,
Christian

Similar Messages

  • Dose Indesign CS3's Data Merging List not surpport double byte language?

    Please help?
    Dose Indesign CS3's Data Merging List not surpport double byte language? But Indesign CS2 works well with Chinese Language!
    Thanks
    David

    I just did a test, in IDCS3 with a mixed English/Traditional Chinese/Simplified Chinese/Korean tab-delimited text file, and after some coaxing, it worked like a charm.
    1) I did it with a Unicode-encoded tab-delimited text file. When I went to "Select Data Source," I checked the "import options" box so that I could specify that the text file was Unicode. (What encoding is your data merge source file?)
    2) When I hit Preview, I realized that I'd set the generic form letter in a font that had no Korean support, so I had to select an appropriate font with Korean glyphs. (Are you using a font in ID that has Chinese support? I assume so.)
    So, yes, Data Merge works with Chinese characters. I didn't test it with csv, but I will if you're stuck using it. If you're starting with an Excel file, it's not hard to get it to spit out a tab-delimited Unicode text file.

  • UTF8 - Configuring two or more double-byte languages in the Same DB

    I would like to ask.
    a) Can the same DB configured to support more than one double-byte language, at the same time ? For example, I want the same set of tables to accept Simplified Chinese, Traditional Chinese, Korean and Japanese.
    b) If yes, then how are the code-bases of the WINNT would look like. I find that, changing the windows registry entry for various codebases, allows us to choose the latest entry only. Will the OS allow more than just one codebase to co-exisit. I dont know, if My question is correct or not. Pardon.
    Please help me in getting answers to these questions.
    Regards,
    Suganyaa

    Hi,
    Thanks for the reply. I just wanted to confirm the details. And, the answer to the second question...as far as I have gathered is ...
    " OS can support only one of the codebase. Not all at the same time. When I modify the codebase in the windows registry to support simplified chinese, i cannot use korean langpack. And, so on. But, I think this should not be a problem in developing the application. As, the multi-lingual users, will all have their OS codebases, in accordance to their language. The application , ( in java ) should be able to accept the double bytes and then store it in Oracle 8i. And, that is also not a problem.
    I am a new learner...sorry if I confused you.
    Regards,
    Suganyaa

  • Double byte language i.e Japanese or Chinese text in non Unicode System

    Hi,
    I have translated text into Chinese and Japanese in a Unicode system and want to move it into a non Unicode system. Would Chinese/Japanese characters display correctly in a non Unicode system when moved from Unicode system.  I am doing translation in ECC60 or SAP 4.7 Unicode system and moving to SAP 4.7 non Unicode system.
    Thanks
    Balakrishna

    Hi Balakrishna,
    in general the transport between Unicode and Non-Unicode systems is supported.
    However there are restrictions, which are outlined in SAP note 638357.
    In your case it is a prerequisite that the objects to be transported are language dependent (text lang. flag is set on the language key - see SAP note 480671) and the languages are properly setup in the target systems.
    For double byte data there is a specific issue when transferring data from Unicode to Non-Unicode:
    In a Non-Unicode system, one double-byte character needs two bytes, therefore e.g. in a 10 char field, 5 double byte chars are fitting. In a Unicode system, you can insert 10 double-byte chars in a 10 char field.  Hence there is a risk of truncating characters in case of Unicode --> Non-Unicode communication.
    Please also have a look at SAP notes 1322715 and 745030.
    Best regards,
    Nils

  • Double byte file name in KM NW04

    Prior to NetWeaver 04 using the HTML editor, we could name a file in a double byte language (Simplified/Traditional Chinese) by entering the English name in the "ID" field and then entering the double byte name in the "Name" field using "File/Details". While this was not elegant, it worked and would allow users to see the double byte name. Now in NetWeaver 04 we receive the error message "! name is not a canonical name" . Is there a way to name files in double byte using HTML or TEXT editors so our users can see the double byte name?

    Hi Boris,
    I'm working with Rick on the issue.
    I've tried the following (per OSS Note 507624), but we still get the "name is not canonical name" error.
    Added the following parameters via configtool to server instances:
    -Dhtmlb.useUTF8=X
    -Dfile.encoding=UTF8
    Added the following parameters via configtool to server instances:
    -Dhtmlb.useUTF8=X
    -Dfile.encoding=ISO-8859-1
    Set the following on <sid>adm account before starting portal (in combination with parameter settings above):
    setenv LC_ALL en_US.UTF-8
         locale
         LANG=en_US.UTF-8
         LC_CTYPE="en_US.UTF-8"
         LC_NUMERIC="en_US.UTF-8"
         LC_TIME="en_US.UTF-8"
         LC_COLLATE="en_US.UTF-8"
         LC_MONETARY="en_US.UTF-8"
         LC_MESSAGES="en_US.UTF-8"
         LC_ALL=en_US.UTF-8
    Unfortunately, these have not corrected the issue.
    Have I misinterpreted the recommendations in the Note or is there something else I needed to do?
    Thanks,
    Fred Bennett
    [email protected]

  • PDF acceleration F5 BigIP WA and double byte characters

    We have been trying to use the F5 appliance from BigIP to accelerate the delivery of PDF files from SharePoint over the WAN.  However, we encountered problems with the double-byte files many months ago and have been trying to resolve the problem with F5.  We have turned off PDF acceleration on the F5 because of the problems.  The problem occurs when PDF files have Kanji characters in the file name.  If the file names are English (single byte) the problem does not occur, even if the content of the PDF contains Kanji characters.
    After many months of working with F5, they are now saying that the problem is with the Adobe plug-in to Internet Explorer.  Specifically they say:
    The issue is a result of Adobe's (not F5's) handling of the linearization request of PDF’s with the Japanese character set over 300 KB when the Web Accelerator is enabled on the BigIP (F5) appliance.  We assume the issue exists for all double-byte languages, not only Japanese.  If a non-double byte character set is used, this works fine.  “Linearization” is a feature which allows the Adobe web plug-in to start displaying the PDF file while it is still being downloaded in the background.
    The F5 case number is available to anybody from Adobe if interested.
    The F5 product management  and the F5 Adobe relationship manager have been made aware of this and will bring this issue up to Adobe.  But this is as far as F5 is willing to pursue as a resolution.  F5 consider this an Adobe issue, not a F5 issue.
    Anybody know if this is truly a bug with the PDF browser plug-in?  Anybody else experienced this?

    Your searches should have also come up with the fact that CR XI R2 is not supported in .NET 2008. Only CR 2008 (12.x) and Crystal Reports Basic for Visual Studio 2008 (10.5) are supported in .NET 2008. I realize this is not good news given the release time line, but support or non support of cr xi r2 in .net 2008 is well documented - from [Supported Platforms|https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/7081b21c-911e-2b10-678e-fe062159b453
    ] to [KBases|http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/oss_notes_boj/sdn_oss_boj_dev/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/scn_bosap/notes.do], to [Wiki|https://wiki.sdn.sap.com/wiki/display/BOBJ/WhichCrystalReportsassemblyversionsaresupportedinwhichversionsofVisualStudio+.NET].
    Best I can suggest is to try SP6:
    https://smpdl.sap-ag.de/~sapidp/012002523100015859952009E/crxir2win_sp6.exe
    MSM:
    https://smpdl.sap-ag.de/~sapidp/012002523100000634042010E/crxir2sp6_net_mm.zip
    MSI:
    https://smpdl.sap-ag.de/~sapidp/012002523100000633302010E/crxir2sp6_net_si.zip
    Failing that, you will have to move to a supported environment...
    Ludek
    Follow us on Twitter http://twitter.com/SAPCRNetSup
    Edited by: Ludek Uher on Jul 20, 2010 7:54 AM

  • Double-byte parser

    Hey,
    I need to load in 400 character string of data into text using the FM SAVE_TEXT in chucks of 60 characters.  If the user logs in Japanese, they could be loading in a mix of single byte and double characters.  I need to be able to check if I break my string of data at the 60th character, will I be corrupting a double character. 
    Does anyone know how to perform this test in ABAP?
    Thanks for the feedback! Bill

    Hi Bill,
    If you are on version 6.2+ you could try using the NUMOFCHAR( ) command.
    I guess you could do something like:
    * First determine the number of actual
    * characters in 60th and 61st bytes.
    w_num_of_chars = NUMOFCHAR( w_text+60(2) ).
    * If there is only 1, then split on 59.
    IF w_num_of_chars = 1.
    * Otherwise, we have two characters so we can split at 60.
    ELSE.
    ENDIF.
    I must admit I'm a complete novice when it comes to double byte languages and their handling, so that may just be complete gobbledegook.
    Hope that helps.
    Cheers,
    Brad

  • What is double byte support for Materials?

    We recently upgraded to ECC5 Unicode form our earleir no unicode 46C.
    In this regard:
    How do we do a double byte language set up for Material master?
    What is meant by double byte support in the first place?
    Pl advise.
    - Ven

    If the system is Unicode no specific settings are necessary, you can enter the language code in the material text of any language and enter the text. If all people have the right fonts installed it will be displayed without problems.
    Markus

  • Double Byte Characters for Japanese Kanji Language

    Hello All,
    How can you identify the Double Byte characters and how we can check whether they are double byte characters  or not.
    Can you please tell me the double byte characters from Japanese Kanji language for the characters.
    This is very urgent.I will reward you with points.
    Thanks,
    Karan

    *& Report  YTEST_LOGIC
    REPORT  teched_unicode_solution_1               .
    *** Exercise 1: Distinction between byte and character length
    *** after Unicode enabling
    parameter: param type c.
    PERFORM test1 USING param.
    *  FORM test1
    FORM test1 USING text TYPE c.
      DATA: len1 TYPE i,
            len2 TYPE i,
            off TYPE i.
    DESCRIBE FIELD text LENGTH len1 IN BYTE MODE.
    DESCRIBE FIELD text LENGTH len2 IN CHARACTER MODE.
    WRITE:/ LEN1, LEN2.
    ENDFORM.                                                    "test1
    Use the above code to find which characters are double bytes and which are not...
    The double byte characters will have double the length when they are in the byte mode..
    reward points for helpful answers
    Edited by: Rahul Kavuri on Mar 26, 2008 6:35 PM

  • Text strings from VISA read don't match identical looking text constants - could it be double byte characters"

    Our RS232-enabled instrument sends ASCII strings to COM 1 and I read strings in. For example I get the string "TPM", or at least it looks like "TPM" if I display it. However, if I send that to the selector input of a Case structure, and create a case for "TPM", whether the two appear to match varies. Sometimes it matches, and measuring its length returns 3. Sometimes it measures 7 or 11 or 12 characters long, and it doesn't match. I can reproduce a match or a mismatch by my choice of the command that went to the instrument prior to the command that causes the TPM response, but have made no sense of this clue. I have run it through Trim Whitespace, with Both Ends (the default) explicitly selected. I have also turned the string into a byte array, autoindexed a For loop on that, and only passed the bytes if they don't equal 32, or if they don't equal 0, thinking spaces or nulls might be in there, but no better.
    The Trim Whitespace function's Help remarks that it does not remove "double byte characters". But I can't find anything else about "double byte characters". Could this be the problem? Are there functions that can tell whether there are "double byte characters", or convert into or out of them? By "double byte characters", do they just mean Unicode?
    Solved!
    Go to Solution.

    Cebailey,
    The double byte characters are generally used for characters specific to languages other than English.  If you display your message in  " '\' Codes Display"  in a string indicator do you see any other characters?   You could also use Hex Display to see count the number of bytes in the message.  You are probably getting messages with non-printable characters that might need to be trimmed before using your application.  If you want more information the '\' Codes Display, there's a detailed description found in the LabVIEW Help.  You can also find the same information on our website in the LabVIEW Help.  Backslash ('\') Codes Display
    Caleb WNational Instruments

  • Sapshcut and  double-byte characters trouble?

    Hi experts,
    I did try some commands to log-on to SAP system as the following examples and get some conclusions:
    (1) sapshcut.exe -sysname="今ウィ ちゃわ異"  -user="paragon1"  -pw="paragon1"  -client="800" -language=en  -maxgui "
    (2) sapshcut.exe -sysname="EC5"  -user="paragon1"  -pw="paragon1"  -client="800" -language=en  -maxgui "
    - For the (1) case, with any "double-byte characters" sysname (Janpanese,...), I can not log-on to SAP and get "Microsoft Visual C++ Runtime Library" error message.
    - For the (2) case, without "double-byte characters" sysname, I can log-on to SAP easily.
    Thus, I wan to know
    1) Does the sapshcut.exe support double-byte characters?
    2) Do we have a way to use sapshcut.exe with double-byte characters(Japanese,...)?
    Kindly Regards,

    The comments on the bytes/strings were helpful. Thanks.
    But I'm still confused as to what matching pattern could be used.
    For example a pattern like:
    [A-Za-z]
    I assume would not match any double byte characters.
    I also assume the following won't work either:
    [\\p{Alpah}]
    because it is posix - US-ASCII only.
    So how do you say "match the tag, then take any characters,
    double byte, ascii, whatever, then match the text tag - per the
    original example ?

  • Crystal XI R2 exporting issues with double-byte character sets

    NOTE: I have also posted this in the Business Objects General section with no resolution, so I figured I would try this forum as well.
    We are using Crystal Reports XI Release 2 (version 11.5.0.313).
    We have an application that can be run using multiple cultures/languages, chosen at login time. We have discovered an issue when exporting a Crystal report from our application while using a double-byte character set (Korean, Japanese).
    The original text when viewed through our application in the Crystal preview window looks correct:
    性能 著概要
    When exported to Microsoft Word, it also looks correct. However, when we export to PDF or even RPT, the characters are not being converted. The double-byte characters are rendered as boxes instead. It seems that the PDF and RPT exports are somehow not making use of the linked fonts Windows provides for double-byte character sets. This same behavior is exhibited when exporting a PDF from the Crystal report designer environment. We are using Tahoma, a TrueType font, in our report.
    I did discover some new behavior that may or may not have any bearing on this issue. When a text field containing double-byte characters is just sitting on the report in the report designer, the box characters are displayed where the Korean characters should be. However, when I double click on the text field to edit the text, the Korean characters suddenly appear, replacing the boxes. And when I exit edit mode of the text field, the boxes are back. And they remain this way when exported, whether from inside the design environment or outside it.
    Has anyone seen this behavior? Is SAP/Business Objects/Crystal aware of this? Is there a fix available? Any insights would be welcomed.
    Thanks,
    Jeff

    Hi Jef
    I searched on the forums and got the following information:
    1) If font linking is enabled on your device, you can examine the registry by enumerating the subkeys of the registry key at HKEY_LOCAL_MACHINEu2013\SOFTWARE\Microsoft\Windows NT\CurrentVersion\FontLink\SystemLink to determine the mappings of linked fonts to base fonts. You can add links by using Regedit to create additional subkeys. Once you have located the registry key that has just been mentioned, from the Edit menu, Highlight the font face name of the font you want to link to and then from the Edit menu, click Modify. On a new line in the dialog field "Value data" of the Edit Multi-String dialog box, enter "path and file to link to," "face name of the font to link".u201D
    2) "Fonts in general, especially TrueType and OpenType, are u201CUnicodeu201D.
    Since you are using a 'true type' font, it may be an Unicode type already.However,if Bud's suggestion works then nothing better than that.
    Also, could you please check the output from crystal designer with different version of pdf than the current one?
    Meanwhile, I will look out for any additional/suitable information on this issue.

  • JSF and Double Byte Character

    Hi,
    I wanted to know how to handle <h:outputText> with chinese character or double byte character.
    See sample code below :
    <%@ page language="java" contentType="text/html; charset=UTF-8"      pageEncoding="UTF-8"%>
    <meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
    <h:form styleClass="form" id="form1">
    <% request.setCharacterEncoding("UTF-8"); %>
    <h:inputText styleClass="inputText" id="text1"></h:inputText>
    <hx:commandExButton type="submit" value="Submit" styleClass="commandExButton" id="button1"      action="#{pc_SubmitTest.doButton1Action}"></hx:commandExButton>
    <h:outputText styleClass="outputText" id="text2"></h:outputText>
    </h:form>
    When you input with double byte character and submit,
    the output screen value did not render properly .
    I tried this similiar code at JSP, it work fine.
    Anybody know how to solve this problem ?
    Anything need to do at pagecode level ?
    Thank you.
    Reinardy

    Problem was due to the fact that I was trying to generate the excel file in char stream instead of byte stream

  • Right Truncation, varchar in SQL Server, double-byte in Oracle

    Hi.
    We have a DB Link using DG4MSQL from Oracle 11.1.0.7.0 to a SQL Server 2005 database. The Oracle database is set up to use UTF-8 so all character fields are double-byte on the oracle side.
    On the SQL Server table we have a column defined as varchar(32). When doing a select * from this table (from Oracle, over the DB link), everything works fine if that column contains values with a length of 16 characters or less. When the column contains 17 - 32 characters we get the following error:
    Error report:
    SQL Error: ORA-28500: connection from ORACLE to a non-Oracle system returned this message:
    [Oracle][ODBC SQL Server Driver]String data, right truncation {01004}
    ORA-02063: preceding 2 lines from SQLUAT
    28500. 00000 - "connection from ORACLE to a non-Oracle system returned this message:"
    *Cause:    The cause is explained in the forwarded message.
    *Action:   See the non-Oracle system's documentation of the forwarded
    message.
    It looks like Oracle (or the ODBC driver) is bringing the data back as double-byte even though it is declared as varchar. If we change the column on the SQL server side to nvarchar, everything works fine.
    Is this a bug? Or is there a setting that can be specified for the gateway so that it recognizes SQL Server varchar columns as single-byte, to prevent them from coming over as double-byte and messing up the ODBC driver? Or is there anything else we should do to prevent this problem?
    Thanks

    Hi,
    I don't think it's a problem with the hs or the dblink (but I may be wrong):
    UTF-8 is double-bytes as you said. when you declare varchar2(32) it actually means you can insert up to 32 bytes --> 16 characters.
    even if you do a simple insert to the table with more than 16 characters - it will fail.
    In order to overcome this you should declare the table with varchar2(32 char) which will tell Oracle to allocate enough bytes to hold 32 chars - not bytes. In UTF-8 this means 64 bytes (2*n). If you'd be using AL32UTF-8 this would mean 4*n.
    BTW- why are you using UTF-8 and not AL32UTF8?
    Hope this helps.

  • Asian double-byte characters

    Hello,
    When importing English data into the CRM field lengths are defined for example "Name" = 100 chartacters. When im importing some Asian languages these are defined as double byte. Can anyone confirm if this has any effects on the field lenghts i can import for fields or should i be able to expect to import the same number of charters as English?

    Jono, I would recommend trying it and see what happens. I have a document called "Steps to Import Asian Characters in a Data File". If you want it let me know your email address.

Maybe you are looking for