Japanese characters from args giving question marks on Japanese OS

Hi,
We are internationalising our product to japanese, and one of the features is to be able to open a file containing japanese characters from a double click on a japanese windows OS.
The double click is set up in the registry, and indeed it works properly for most characters. There are a few characters though (unicode character 20060 among them) that it doesn't work for, and what happens is that between the double clicking of the file, and the command line name of the file to open ("%1" in the registry / args[0] in java), the character in question is converted into the literal character for "?" (ascii 63), and java can't open the file.
Testing wordpad directly with this character is fine, the file opens. I've written a simple C++ app and a simple JAVA app which fork wordpad with the fileName param passed to it from the registry, and it didn't open the file passed because of that character.
So our java application, a simple java program and a simple C++ program can't resolve the fileName passed to it because of this character.
The thing is, wordpad is using the same regedit method to get its parameters as we do ("appName.exe" "%1" in shell/open/command) and it opens files containing this character without a problem.
Any ideas on what I'm missing?
Thanks very much
Jack

Hi,
We are internationalising our product to japanese,
and one of the features is to be able to open a file
containing japanese characters from a double click on
a japanese windows OS.
The double click is set up in the registry, and
indeed it works properly for most characters. There
are a few characters though (unicode character 20060
among them) that it doesn't work for, and what
happens is that between the double clicking of the
file, and the command line name of the file to open
("%1" in the registry / args[0] in java), the
character in question is converted into the literal
character for "?" (ascii 63), and java can't open the
file.
Testing wordpad directly with this character is fine,
the file opens. I've written a simple C++ app and a
simple JAVA app which fork wordpad with the fileName
param passed to it from the registry, and it didn't
open the file passed because of that character.
So our java application, a simple java program and a
simple C++ program can't resolve the fileName passed
to it because of this character.
The thing is, wordpad is using the same regedit
method to get its parameters as we do ("appName.exe"
"%1" in shell/open/command) and it opens files
containing this character without a problem.
Any ideas on what I'm missing?
Thanks very much
JackUnicode 20060 belongs to an extended part of JIS Kanji code set and there can be many
applications or systems that do not support those characters. Shift_JIS doesn't and
EUC-JP use 24 bit code for representig those chars which, unfortunately, aren't supported
by most exixting apps.

Similar Messages

  • Oracle 9i Support for multi language is not working.. Giving question mark

    HI,
    We have an application which uses oracle 9i as the database. Riight now we are supporting only english and there is a requirement to support multiple languages like korean, chineese and japaneese.
    But we are planning to migrate one part of the application to support multi languages. Means it may affect around 10 tables but with huge data. Totally we have around 100 tables.
    How to enable the database for supporting multiple langugages.?
    Is there any way to enable only the few tables supporting multiple languages. Because if we change the database level parameters for supporting languages, we may need to migrate entire tables. this will be a huge task.
    Even if want to set the parameters for supporting multiple languages.. how to set it. Is it possible set it in the existing database or do we need to re-create the table with these prameters.
    I have read in some documentation, that we can create table columns with nVarchar2 for supporting multi languages. I have created it. but if i copy some other language characters into those columns, it is giving question mark.
    Is it possible to do search using text in native langugage like chineese..
    Could somebody guide me on the above clarificationa and what would be the best approach..
    Thanks in advance
    Jino
    Regards,
    Jino George
    Ext: 6520

    You should not use any more Oracle 9.0.1 but at least Oracle 9.2.0.8 to get some extended support if you really cannot upgrade to 10g.
    I don't have any Oracle 9.x database available but I've run successfully following test with Oracle 10.2.0.1 in character mode under Linux:
    oracle@pbell:~$ export NLS_LANG=AMERICAN_AMERICA.AL32UTF8
    oracle@pbell:~$ sqlplus / @nls
    SQL*Plus: Release 10.2.0.1.0 - Production on Fri Aug 29 17:29:56 2008
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    SQL> drop table t;
    Table dropped.
    SQL> select * from v$nls_parameters where parameter like '%SET%';
    PARAMETER
    VALUE
    NLS_CHARACTERSET
    WE8ISO8859P1
    NLS_NCHAR_CHARACTERSET
    AL16UTF16
    SQL> create table t ( data nvarchar2(100));
    Table created.
    SQL> insert into t values(unistr('\76EE\7684\5730'));
    1 row created.
    SQL> select * from t;
    DATA
    目的地Try to make sure you have the right NLS_LANG setting on the client side (under Windows this is a registry setting).

  • Arabic characters are displaying as question marks in forms 10g

    We have migrated our application from forms 6i to forms 10g and now in forms 10g the arabic characters are displaying as question marks while it displays correctly in the old application using forms 6i. I have already set the character set to AR8MSWIN1256 in the registry, but it didn't help. Somebody please help.

    @ Sarah, Al-Salamu Alikum We Rahmatu Allah we Barakatu,
    Sarah Habibty, why new installation ? In order to select a new suitable character set !!!
    Then creating a new instance from the db is a better alternative since it saves time,effort and another back up of his current db is exist safely if needed for any purposes in the future.
    @Amer,honestly speaking...
    Modifing ur NLS_LANG to > AMERICAN_AMERICA.AR8MSWIN1256
    Works for me in both Arabic and English data in 2 applications.This works in my pc.But it didn't works at my boss pc this can happened don't have any reason for that.!!!!
    i spent lot's of time trying to search but what i had got is that solution i suggested by a friend of mine.
    Now please could you advise me, is it better to create a new instance of database as Amatu Allah has suggested or is it better to change the character set through sql as some others have suggested? Again i suggest to select the short cut way ; to reset the character set through sql after taking a back up from ur data that is currently exist.
    then retest again doing the select and test ur data input and retrieval.
    SQL> select * from v$nls_parameters
    2 where parameter in ('NLS_CHARACTERSET','NLS_LANGUAGE');watching the output if it works that's fine saving ur time & effort .
    if not working with the correct NLS_CHARACTERSET then use my previous solution.
    Hope this helps...
    Regards,
    Amatu Allah

  • Language:  Filename with characters for arabic turns question mark

    Language: Filename with characters for arabic turns question mark
    OS: Solaris 9
    Machine: Sun-Fire 25K
    There is an adobe distiller software that is configured and a java apps. There are postscript files that are being converted to .pdf format using the adobe distiller. Using the GUI (using the Exceed; for remote access), when they use GUI to convert the postscripts to pdf files, the long filenames have the corresponding characters for arabic reading purpose. This is OK.
    When we use the windows RUN to telnet the server and convert the postscripts to pdf, it gives a question marks characters in the filenames ( this; is a sample; filename; ?? ??? ??; right.pdf )
    We are not sure now if we have to add a package of arabic or a patch to resolve this problem.
    Message was edited by:
    yurioira32
    Message was edited by:
    yurioira32

    Solution found, I'll post the work around to those who might encounter the same problem.
    Somewhere in the layers of technology (webwork or weblogic I'd guess), the servlet response is encoded into UTF-8 regardless. The encoding in the database was ISO-8859-1. Sending ISO encoded bytes by UTF-8 caused the conflicting character codes (anything above 127) to show up as undefined.
    The fix is to decode the input byte array into ISO-8859 string, then encode that string into UTF-8, which can be send by Weblogic.
    isoConvert = new String(buf, "ISO-8859-1");
    out.write(isoConvert.getBytes("UTF-8"), 0, isoConvert.getBytes("UTF-8").length);

  • Japanese Characters are showing as Question Marks '?'

    Hi Experts,
    We are using Oracle Database with below nls_database_parameters:
    PARAMETER VALUE
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET WE8MSWIN1252
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_CSMIG_SCHEMA_VERSION 3
    NLS_RDBMS_VERSION 11.1.0.7.0
    When we are trying to view the Japanese characters (windows 7) in SQLdeveloper, toad or sqlPlus, we are getting data like '????'.
    Can anybody please explain us the setups required to view the Japanese characters from the local machine and database.
    Thanks in advance.

    user542601 wrote:
    [Note: If I insert the Japanese characters from Sql Developer or Toad, I am unable to see proper results.]For JDBC connections in Oracle SQL Developer, I believe a different parameter setting is required.
    Try running Sql Dveloper with jvm option: -Doracle.jdbc.convertNcharLiterals=true.
    I need to use this data in Oracle 6i Reports now.
    When I am creating reports using the table where I have Japanese characters stored in NVARCHAR2 column, the value is not displaying correctly in Report Regardless of Reports support for nchar columns, 6i is very very old and based on equally ancient database client libraries (8.0.x if memory serves me). Earliest version of Oracle database software that support the N literal replacement feature is 10.2. So, obviously not available for Reports 6i.
    I'm guessing only way to fully support Japanese language symbols is to move to a UTF8 database (if not migrating to a current version of Report Services).
    Please help to provide a workaround for this. Or do I need to post this question in any other forums?There is a Reports forum around here somewhere. Look in the dev tools section or maybe Middleware categories.
    Edit: here it is: {forum:id=84}
    Edited by: orafad on Feb 25, 2012 11:12 PM
    Edited by: orafad on Feb 25, 2012 11:16 PM

  • Some characters are replaced by question marks!

    All of a sudden my iMac (OS X 10.5.1 Leopard) is displaying question marks for some special characters.
    For example, on internet, look up a word "pediment" in Yahoo! Dictionary ( http://education.yahoo.com/reference/dictionary/entry/pediment)... Instead of showing the "dot" symbol to identify a syllabus break, it shows question marks as ped?i?ment, and instead of showing c with a little squiggle underneath for a word facade (within the definition of pediment), it shows fa?ade.
    I did not have this problem on my iMac on Tuesday (01/29/08) morning and I have not installed anything other than the Apple updates.
    I tried with FireFox and Safari, and both show the same thing. Then I went and checked my husband's iMac (OS X 10.5.1 Leopard) and it has a same problem, so I know it is not just my iMac.
    Then (yes, there is more...) I started to work on my Word document and tried to insert "symbol" from insert on the Toolbar... many of my symbol fonts (Symbol, Osaka, etc) has question marks replacing special characters.
    Please help! I am unable to complete my work without these special characters!!

    Well, I went to the Yahoo! Dictionary "Pediment" and it said it was set to Default which is set to Western (ISO Latin 1), so I assumed it was Western (ISO Latin 1). But when I click on Western (ISO Latin 1), the pages appear what it's supposed to look like with special characters... What does this mean?
    I reset all the settings; I chose something else as default, quit Safari, opened it and chose Western (ISO Latin 1) as a default to see if it changes anything... No change. It says my default is Western (ISO Latin 1), but it seems not.
    I am officially confused and have no clue what to do...??????!!!!!
    On the bright side, at least I can get my work done, even if I have to change the encoding one page at a time... Thanks.

  • Weblogic 12c Servlet Response - Special characters show up as question mark

    My web app is running on Weblogic 12c (12.1.1) using WebWork + Hibernate. The program streams data (bytes making up a pdf) from a CLOB in an Oracle Database to the AsciiStream of the servlet output response. No exceptions are thrown, but the generated pdf contains blank pages. Comparing the bytes of the generated pdf, special characters are showing up as question marks.
    Some of the bytes read in from the database contain 8 bits (correct data), but the bytes that the servlet return contain only 7 (all bytes with 8 bits become "1111111"). The number of bytes returned from the servlet is correct.
    Code:
    //Response is HttpServletResponse
    response.setContentType("application/pdf");
    response.setHeader("Content-Disposition", "inline; filename=\"test.pdf\"");
    out = response.getOutputStream();
    byte[] buf = new byte[16 * 1024];
    InputStream in = clob.getAsciiStream();
    int size = -1;
    while ((size = in.read(buf)) != -1){
    // buf contains the correct data
    out.write(buf, 0, size);
    // other exception handling code, etc
    out.flush();
    out.close();
    "Correct" pdf byte example:
    10011100
    10011101
    1010111
    1001011
    1101111
    11011011
    Incorrect pdf byte example:
    111111
    111111
    1010111
    1001011
    1101111
    111111
    I have verified that the data read from the CLOB in the database IS correct. My guess is that the Weblogic server has some strange servlet settings that causes the bytes to be written to the servlet output stream incorrectly, or a character encoding issue. Any ideas?
    Edited by: 944705 on Jul 26, 2012 10:17 AM

    Solution found, I'll post the work around to those who might encounter the same problem.
    Somewhere in the layers of technology (webwork or weblogic I'd guess), the servlet response is encoded into UTF-8 regardless. The encoding in the database was ISO-8859-1. Sending ISO encoded bytes by UTF-8 caused the conflicting character codes (anything above 127) to show up as undefined.
    The fix is to decode the input byte array into ISO-8859 string, then encode that string into UTF-8, which can be send by Weblogic.
    isoConvert = new String(buf, "ISO-8859-1");
    out.write(isoConvert.getBytes("UTF-8"), 0, isoConvert.getBytes("UTF-8").length);

  • Special characters are changed to question marks

    I'm using SQL developer 3.107 for unit testing.
    In a particular unit test, the expected result is a sentence (varchar2) with a special character (ë).
    But when I save the result, SQL developer changes the character to 2 question marks.
    And as a result, the unit test fails because the expected result differs from the received result (where the 'ë' remains unchanged).
    I already tried changing the encoding in the SQL Developer preferences from cp1252 to UNICODE and UTF8 but that didn't help.
    Any suggestions?
    Thanks in advance

    Hello:
    I guess that what you observe could be an interaction between the server characterset and the client characterset.
    These are the results with different client characterset settings:
    NLS_LANG=american_america.WE8ISO8859P1
    select 'ë' c, dump('ë') dumped from dual;
    C DUMPED
    ë Typ=96 Len=1: 137
    NLS_LANG=american_america.WE8MSWIN1252
    select 'ë' c, dump('ë') dumped from dual;
    C DUMPED
    + Typ=96 Len=1: 191
    set NLS_LANG=american_america.WE8PC850
    select 'ë' c, dump('ë') dumped from dual;
    C DUMPED
    ë Typ=96 Len=1: 235
    According to the ISO 8859-1, 8-bit single-byte coded graphic character sets document [http://www.open-std.org/JTC1/SC2/WG3/docs/n411.pdf] the encoding of the latin small letter e with diaeresis is 0xEB -> (decimal 235).
    If you set the client to WE8PC850 do you see a correct behaviour?

  • Unicode characters are shown as "question marks" in Eclipse console

    I am trying to retrieve Unicode data from Sybase database using jdbc.
    The data are stored in Sybase with unichar and univarchar datatypes.
    Following is the java code tring to run.
    {color:#808080}public class Test
    public static void main(String [] args) throws Exception
    CoreServiceSoapBindingStub binding_Core = new CoreServiceSoapBindingStub();
    CoreWSServiceLocator locator_Core = new CoreWSServiceLocator();
    binding_Core = (CoreServiceSoapBindingStub) locator_Core.getCoreService();
    Contact[] con = binding_Core.getContact();
    for(int i=0;i< con.length;++i)
    System.out.println(con.getLastName());
    }{color}
    The result of this code in Eclipse console should be as follow (consists of one English and one Japanese name).
    {color:#808080}Suzuki
    &#37428;&#26408;{color}
    However when I run this, I get the following.
    {color:#808080}Suzuki
    {color}
    The alphabetical characters seem to diplay fine in the console, but foreign characters are not....
    The default character set of the database is ISO-8859-1, but I used unichar and univarchar to store data in unicode thus believe no issue at database side...
    Used jconnect 6.05 (com.sybase.jdbc3.jdbc.SybDriver
    ) for the database driver.
    Java files are encoded in UTF-8.
    Console Encoding is UTF-8.
    Is this an issue in database driver?
    Since I set the parameters for character set to UTF-8 in both the database and java files....
    It would be great if someone could give some comments on this issue....
    Thanks a lot.

    It might be better to ask this question on an Eclipse forum. I have a couple of suggestions, but none of them have made the output in my console look entirely correct:
    1. Try to start Eclipse with these parameters: -vmargs -Dfile.encoding=UTF-8
    2. Try switching the font settings for the Console under Preferences in Eclipse.

  • N8- Arabic characters show up as question marks in...

    Hi ..
    I noticed that the subject of the email is shown in the right way ( correct form of arabic characters)..
    but the body of the email which is formatted in some font and color show up as ???????????
    that replaces the arabic characters !!!
    while the exact same email show up correctly in the browser ,, but not the email ..
    ps: my phone is N8 ..
    thanks in advance..

    Dear petrib, what you are saying is 100% right, but its also Nokia issue to fix the problem with Microsoft for their customers to be satisfied... the problem is not when using Windows Live through PCs, I have the problem in my Nokia device and i am trying to contact Nokia to solve it with Microsoft... Am I right ??

  • Search no mapped unicode characters as question mark

    No mapped unicodes characters are replace by question mark.
    I want to search/detect these characters to correct them (i may spend too much time to find the character "?" in a big document, because of real question mark).
    How can i do ?

    Arnis' .mif suggestion is the way I've tackled this problem from time to time, but save a spare copy before you start <g><br /><br />You may not need to, though. I haven't checked, but see if you can find a page with both a normal question-mark and one you know indicates a unicode character. Then copy/paste the unicode question-mark into the Find/replace dialogue and see what happens; I suspect it will have its own value, not the same as a normal question-mark.

  • PDF Preview for a  PO template with polish characters shows question marks

    When I view the PDF using application I see the polish characters(Boilerplate text) as question marks. But when I view it through XML Publisher Dsktop version 5.5.0 it is fine I have download the new version of adobe 7.08 and I can see the polish characters but the preview in XML publisher in the application stll shows "?"

    No I havent configured any fonts on desktop/application .. when I view Document properties on the PDF which works the XML Publisher version 5.5.0 and that of the application is 5.6.0.
    In the fonts tab on the desktop version there is one additional one listed fromt he application one - AlbanyWTJ(Embedded Subset)

  • German special characters not displaying on page instead of that Question marks displayed

    when i submit form,i need to send a mail with filled data.Here Suppose user entered german special characters.I am getting Question marks instead of ä,Ä,ö,Ö,ü,Ü,ß. I have used meta tag like
    <meta http-equiv="content-type" content="text/html;charset=utf-8" />.But i am getting fine characters with IE and Chrome.I am using FireFox 3.6.16 english.

    Hello core team,
    Thanks a lot!!!!
    i m not only the person to use html form.So many non technical persons using the same. how can i tell to all persons regarding changing character encoding. i need some permenant solution,

  • When I write a word with an accent,  a question mark appears

    When I send an email from my iPhone and use words that are "accented" or with a special foreign characters;  on the email sent, these characters are replaced with "question marks"  �.
    Example: 
    Mañana comes up as  Ma�ana
    Cómo comes up as C�mo

    Let Apple know via
    http://www.apple.com/feedback

  • Japanese, Question Marks, Locales, Eclipse, and Windows XP ????

    Hello. I am having some issues localizing JSP to Japanese. I have read a lot of stuff on the topic. I have my .properties file in unicode with native2ascii, etc.
    When I debug under Eclipse 3.0, I see the Japanese characters correctly displayed in my properties file and inside of strings internal to to the program. However, when I try to print them with a System.out.println, I get question marks (??????).
    My reading tells me that the ???? indicate that the characters cannot be displayed. I am somewhat confused because in the same Eclipse context I can clearly see the Japanese characters in the debugging window.
    Thus I am missing the part where I set my standard output to correctly display the characters like Eclipse is displaying them in windows other than the "Console" window.
    My default encoding is CP1252. If I do something like:
    out = new java.io.PrintStream(System.out, true, "UTF-8");
    and print my unicoded resource from the bundle I get the UTF-8 character representation (����&#65533;�����������&#65533;��������&#65533;��&#65533;��&#65533;&#65533;�&#65533;��&#65533;��&#65533;���). With System.out.println I get ?????
    My first reaction would be that the Japanese fonts aren't on my system, but clearly they are as I can see them in other windows.
    When I try to show a Japanese resource on the web page that is the result of the jsp file I get ????. I can display the same characters UTF-8 encoded in a php page.
    Here is another example:
    java.util.Locale[] locales = { new java.util.Locale("en", "US"), new java.util.Locale("ja","JP"), new java.util.Locale("es", "ES"),
    new java.util.Locale("it", "IT") };
    for (int x=0; x< locales.length; ++x) {
    String displayLanguage = locales[x].getDisplayLanguage(locales[x]);
    System.out.println(locales[x].toString() + ": " + displayLanguage);
    displays:
    en_US: English
    ja_JP: ???
    es_ES: espa�ol
    it_IT: italiano
    instead of the correct Japanese characters.
    What's the secret?
    Thanks.
    -- Gary

    What do you want to do exactly? 1. Making a window application? 2. Making a console application? 3. Making a JSP webpage?
    1. If it's window application, there's nothing to worry about if you use Swing. But if you use AWT, it's time to switch to Swing.
    2. If you're making console application, solution does exist as others had pointed out but you'd better forget it because no console in whatever platform supports Unicode (Linux xterm may be an exception? But it probably has font problem). So, even if you could display characters in your computer, the solution isn't universal. You can't ask your every user to switch system locale, install whatever font in order to display a few characters!!
    3. If you're making JSP, I'd advise you to use UTF-8 in webpages. Most browsers nowadays (probably more than 90%) could support UTF-8. All you need is to add the following JSP headers to your every page:
    <%@ page contentType="text/html;charset=utf-8" %>
    <%@ page pageEncoding="iso-8859-1" %>Now, your every out.println(s); will send the correct data to browser without the least effort from you. All conversions are automatic!
    However, just to make things even surer, you could add this HTML meta header:
    <meta http-equiv="Content-Type" content="text/html; charset=utf-8">You use Tomcat, right? I do, and I don't have any problem.
    Last words:
    But, if all you want to do with System.out.println is for debugging, you could use
    JOptionPane.showMessageDialog(null, "your string here");But you'd better have Java 5, or at least 1.4.2, if you want to have everything displayed correctly.

Maybe you are looking for

  • F110 - Customer Callout

    Dear All, When I run the F110 to customers, the system does not select the invoices that are not over due. This is needed here in Brazil, because the bank will be the responsible part to receive those invoice amounts. In this process, the system shou

  • OS X crashing to much...errr

    Hey guys so basically my mac was perfect when my bro gave it to me. its a powermac G4, and i reformatted recentley due to crashing problems. Whenever i have Azureus open, with lets say itunes, or more than one program it crashes. the whole screen dim

  • Query on Message/Output Determination in GR/GI Slip

    Hi, Its regarding Message/Output Determination in Inventory Management: We are using the condition Type "WE03" in our system for Goods Receipt. Currently, Print Parameter for the Output Type "WE03" is maintained as 7(i.e Plant/storage location). Now,

  • How to deactivate serial numbers PEPE10

    I Restored my computer to its factory defaults. when reinstalling  PEPE10, Accidentally installed disc 1, I should have installed disc 2 for windows 7 x64bit.The uninstaller tells me to deactivate and remove serial numbers used when installing on my

  • Materialized Views indexed are not excluded when exporting ddl

    Hello, When exporting ddl of the schema, through the export function of the gui, if I disabled materialized views from the objects to be exported, then the indexes of materliazied view are not excluded of the final DDL. This happens on  Version 4.0.3