Inserting Japanese Characters

I am running Oracle 9.2 on a WIN2k m/c.
I need to insert JAPANESE KANJI characters into my tables.
1) Would like to know what are the setting required for the same.
I would be pulling the data from remote SQL SERVER using OWB.
Createad a SQL Server Transparent Gategway(tg4msql) to connect to the remote SQL Server.
Current NLS Setting
NLS_LANGUAGE     AMERICAN
NLS_TERRITORY     AMERICA
NLS_CURRENCY     $
NLS_ISO_CURRENCY     AMERICA
NLS_NUMERIC_CHARACTERS     .,
NLS_CALENDAR     GREGORIAN
NLS_DATE_FORMAT     DD-MON-RR
NLS_DATE_LANGUAGE     AMERICAN
NLS_CHARACTERSET     AL32UTF8
NLS_SORT     BINARY
NLS_TIME_FORMAT     HH.MI.SSXFF AM
NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
NLS_DUAL_CURRENCY     $
NLS_NCHAR_CHARACTERSET     AL16UTF16
NLS_COMP     BINARY
NLS_LENGTH_SEMANTICS     BYTE
NLS_NCHAR_CONV_EXCP     FALSE
2) How can i see/verify the inserted data using SQL-PLUS.
Need your help in solving the same.
TIA
shankar

Well I would assume that the first thing that is required is to run the db in utf8, which you are doing.
So storing should be no problem. To display in SQL Plus I would suspect you need to set your NLS language to something that would "read" the utf8 character sets.
I have not looked it up but I'm sure the RDBMS documentation will cover most of these topics extensively. Did you take a look at that doc set?
Jean-Pierre

Similar Messages

  • Inserting JAPANESE characters in a database

    Can somebody please let me know how to insert the JAPANESE (Kanji) Characters into the Database.
    Database Version: Oracle 9.2.0.1.0
    Paremeters Setting:
    NLS_CHARACTERSET - UTF8
    NLS_NCHAR_CHARACTERSET - UTF8
    Server OS: Win2K
    Client OS: Win2K

    Not sure what your overall requirements from an application
    support standpoint. But a simple way would be to use UNIST.
    Here is a description:
    UNISTR takes as its argument a string and returns it in the national character set.The national character set of the database can be either AL16UTF16 or UTF8.
    UNISTR provides support for Unicode string literals by letting you specify the Unicode encoding value of characters in the string. This is useful, for example, for
    inserting data into NCHAR columns.
    The Unicode encoding value has the form '\xxxx' where 'xxxx' is the hexadecimal value of a character in UCS-2 encoding format. To include the backslash in
    the string itself, precede it with another backslash (\\).
    For portability and data preservation, Oracle Corporation recommends that in the UNISTR string argument you specify only ASCII characters and the Unicode
    encoding values.
    Examples
    The following example passes both ASCII characters and Unicode encoding values to the UNISTR function, which returns the string in the national character
    set:
    SELECT UNISTR('abc\00e5\00f1\00f6') FROM DUAL;
    UNISTR
    abceqv

  • Inserting Japanese characters to SQL Server via CFMX

    What is necessary in the setup to save non-Latin characters
    to SQL Server via CFMX form? The ColdFusion data source has the
    Unicode option enabled (Enable Unicode for data sources configured
    for non-Latin characters). The target field in the database is an
    nvarchar. What else is necessary to properly insert and then later
    display the non-Latin characters?
    On Microsoft's site they describe the need to convert to and
    from UCS-2 when accessing SQL Server via ASP. Is this type of
    conversion relevant to CFMX?

    jegrubbs wrote:
    > What is necessary in the setup to save non-Latin
    characters to SQL Server via
    > CFMX form? The ColdFusion data source has the Unicode
    option enabled (Enable
    > Unicode for data sources configured for non-Latin
    characters). The target field
    > in the database is an nvarchar. What else is necessary
    to properly insert and
    > then later display the non-Latin characters?
    - define db columns as "N" types
    - ensure cf pages are utf-8 encoding:
    --tag the files w/a BOM
    --use
    <cfprocessingDirective pageencoding="utf-8">
    on each page
    --use
    <cfset setEncoding("form","utf-8")>
    <cfcontent type="text/html; charset=utf-8">
    in application.cfm or .cfc
    - when doing INSERT/UPDATE make sure to use either unicode
    hinting (N'text') or
    cfqueryparam (making sure to turn on the unicode option for
    that DSN in
    cfadmin). cfqueryparam is the best choice.
    also see:
    http://www.sustainablegis.com/unicode/greekTest.cfm
    > On Microsoft's site they describe the need to convert to
    and from UCS-2 when
    > accessing SQL Server via ASP. Is this type of conversion
    relevant to CFMX?
    nope the JDBC driver will handle that gruff. make sure you
    use the JDBC driver
    (name as ms sql server in cfadmin) and NOT the odbc bridge
    thnig.

  • How to Insert  Chinese characters in Japanese Database

    Hi all,
    I am having following characteristics on my computer
    Machine OS --Windows Server 2003
    OS language --Japanese
    Oracle
    Oracle9i Release 9.2.0.1.0 - Production
    NLS_LANGUAGE     JAPANESE
    NLS_CHARACTERSET     JA16SJIS
    Now, i want to insert into database chinese characters. Please guide me how to do the following thing.
    How to insert chinese characters on local machine and if i want to insert on the remote databse (i can not create database link for remote database). I have to send batch file or SQL file and they will execute it on their side.
    if i use this command
    alter session set nls_language = "SIMPLIFIED CHINESE"
    and then insert the records and revert back to japanese character set. Is this correct way....?
    Thanks in advance,
    Pal

    As dombrooks has pointed out, unless all the Chinese characters you are trying to store can be represented in the Shift-JIS character set, which seems unlikely, but I'm not an expert on East Asian languages and I believe there are some glyphs that are shared between various languages, then you're not going to be able to store this data in this database in CHAR or VARCHAR2 columns.
    Depending on the national character set, you may be able to store the data in NCHAR/ NVARCHAR2 columns, though using these data types can substantially increase application complexities since various languages and libraries don't support NCHAR/ NVARCHAR2 columns or require you to jump through some hoops to use them. Your applications would also have to support both character sets, so your applications would all have to be Unicode enabled most likely, which is certainly possible but it may not be a trivial change.
    Justin

  • How to store japanese characters in mysql 5.0

    I want to store japanese characters in mysql 5.0 database through java program and then to retrive the same characters through the program only.Java program means a form containing first name ,last name and address. I am entering corresponding translations in japanese to this fields while inserting to database. In another form i am retrieving those japanese characters.those should display
    in this form.

    How to handle the unicode sir for japanese characters..Pls give me more hints and any reference links to get the answer.

  • Acrobat Pro 9.3.1 does not convert certain Japanese characters

    I have a text document that contains a mix of Roman and Japanese characters - when I do Create PDF From File and read that text document in, there is a sequence of 2 Japanese characters that disappear - the text before them and after them appear in the PDF, but there's a void between.
    The sequence is (don't know if I can insert Japanese in here...)
    before監査証跡after
    When the PDF is generated, the first 2 Japanese characters (after the last 'e' in before) do not appear in the PDF.
    Here is the source text document (UTF-8 encoded with BOM): http://www.scribd.com/doc/28158046
    and here is the resulting PDF: http://www.scribd.com/doc/28158121
    Anyone seen this before?

    If I paste your "before監査証跡after" into Notepad and save it as UTF-8 text, I can print the file to the Acrobat 9.3.1 Pro "Adobe PDF" printer with no problems at all: the 4 kanji appear in a font Acrobat calls "MS-UIGothic".  If I right-click on the saved *.txt file in Windows Exploreer (Vista 64) and select "Convert to Adobe PDF" I still get all the kanji, although the first shows up in Adobe Ming, the 2nd in Adobe Song, and the last 2 in KozGoPr6N.
    I can't explain what's going on here, but perhaps this can help point you down a useful path.
    David

  • Japanese Characters are showing as Question Marks '?'

    Hi Experts,
    We are using Oracle Database with below nls_database_parameters:
    PARAMETER VALUE
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET WE8MSWIN1252
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_CSMIG_SCHEMA_VERSION 3
    NLS_RDBMS_VERSION 11.1.0.7.0
    When we are trying to view the Japanese characters (windows 7) in SQLdeveloper, toad or sqlPlus, we are getting data like '????'.
    Can anybody please explain us the setups required to view the Japanese characters from the local machine and database.
    Thanks in advance.

    user542601 wrote:
    [Note: If I insert the Japanese characters from Sql Developer or Toad, I am unable to see proper results.]For JDBC connections in Oracle SQL Developer, I believe a different parameter setting is required.
    Try running Sql Dveloper with jvm option: -Doracle.jdbc.convertNcharLiterals=true.
    I need to use this data in Oracle 6i Reports now.
    When I am creating reports using the table where I have Japanese characters stored in NVARCHAR2 column, the value is not displaying correctly in Report Regardless of Reports support for nchar columns, 6i is very very old and based on equally ancient database client libraries (8.0.x if memory serves me). Earliest version of Oracle database software that support the N literal replacement feature is 10.2. So, obviously not available for Reports 6i.
    I'm guessing only way to fully support Japanese language symbols is to move to a UTF8 database (if not migrating to a current version of Report Services).
    Please help to provide a workaround for this. Or do I need to post this question in any other forums?There is a Reports forum around here somewhere. Look in the dev tools section or maybe Middleware categories.
    Edit: here it is: {forum:id=84}
    Edited by: orafad on Feb 25, 2012 11:12 PM
    Edited by: orafad on Feb 25, 2012 11:16 PM

  • Handling Multi-byte/Unicode (Japanese) characters in Oracle Database

    Hello,
    How do I handle the Japanase characters with Oracle database?
    I have a Java application which retrieves some values from the database; makes some changes to these [ex: change value of status column, add comments to Varchar2 column, etc] and then performs an UPDATE back to the database.
    Everything works fine for the English. But NOT for Japanese language, which uses Multi-byte/Unicode characters. The Japanese characters are garbled after the performing the database UPDATE.
    I verified that Java by default uses UTF16 encoding. So there shouldn't be any problem with Java/JDBC.
    What do I need to change at #1- Oracle (Database) side or #2- at the OS (Linux) side?
    /* I tried changing the NLS_LANG value from OS and NLS_SESSION_PARAMETERS settings in Database and tried 'test' insert from SQL*plus. But SQL*Plus converts all Japanese characters to a question mark (?). So could not test it via SQL*plus on my XP (English) edition.
    Any help will be really appreciated.
    Thanks

    Hello Sergiusz,
    Here are the values before & after Update:
    --BEFORE update:
    select tar_sid, DUMP(col_name, 1016) from table_name where tar_sid in ('6997593.880');
    /* Output copied from SQL-Developer: */
    6997593.88 Typ=1 Len=144 CharacterSet=UTF8: 54,45,53,54,5f,41,42,53,54,52,41,43,54,e3,81,ab,e3,81,a6,4f,52,41,2d,30,31,34,32,32,e7,99,ba,e7,94,9f,29,a,4d,65,74,61,6c,69,6e,6b,20,e3,81,a7,e7,a2,ba,e8,aa,8d,e3,81,84,e3,81,9f,e3,81,97,e3,81,be,e3,81,97,e3,81,9f,e3,81,8c,e3,80,81,52,31,30,2e,32,2e,30,2e,34,20,a,e3,81,a7,e3,81,af,e4,bf,ae,e6,ad,a3,e6,b8,88,e3,81,bf,e3,81,ae,e4,ba,8b,e4,be,8b,e3,81,97,e3,81,8b,e7,a2,ba,e8,aa,8d,e3,81,a7,e3,81,8d,e3,81,be,e3,81,9b,e3,82,93,2a
    --AFTER Update:
    select tar_sid, DUMP(col_name, 1016) from table_name where tar_sid in ('6997593.880');
    /* Output copied from SQL-Developer: */
    6997593.88 Typ=1 Len=144 CharacterSet=UTF8: 54,45,53,54,5f,41,42,53,54,52,41,43,54,e3,81,ab,e3,81,a6,4f,52,41,2d,30,31,34,32,32,e7,99,ba,e7,94,9f,29,a,4d,45,54,41,4c,49,4e,4b,20,e3,81,a7,e7,a2,ba,e8,aa,8d,e3,81,84,e3,81,9f,e3,81,97,e3,81,be,e3,81,97,e3,81,9f,e3,81,8c,e3,80,81,52,31,30,2e,32,2e,30,2e,34,20,a,e3,81,a7,e3,81,af,e4,bf,ae,e6,ad,a3,e6,b8,88,e3,81,bf,e3,81,ae,e4,ba,8b,e4,be,8b,e3,81,97,e3,81,8b,e7,a2,ba,e8,aa,8d,e3,81,a7,e3,81,8d,e3,81,be,e3,81,9b,e3,82,93,2a
    So the values BEFORE & AFTER Update are the same!
    The problem is that sometimes, the Japanese data in VARCHAR2 (abstract) column gets corrupted. What could be the problem here? Any clues?

  • Unable to insert Chinese characters in Database

    My problem is that I am not able to insert chinese
    (to traditional chinese) characters into my tables in the
    database.
    I have changed the character set to UTF8 while creating the
    database and also tried the alter session command in SQL to
    alter the NLS_LANGUAGE and NLS_TERRITORY (to say traditional chinese).
    But this did not solve my problem.
    Also tried all possibilites like getting Chinese characters
    in my notepad by copy - paste from a Chinese web site
    but while giving the insert into command in my database
    it takes some junk values.
    Someone PLEASE HELP!!!URGENT!!!
    Thanks in advance.
    RKP
    null

    You mentioned in your first note that you have set your database character set to UTF-8? If so, then you are able to store and retrieve multilingual data, including Chinese and Japanese characters. Your issue is not the database. Your client OS must be able to support these languages as well. It is likely that your version of OS supports only Latin and Western European characters. By the way changing your NT regional setting only effects sorting, date formats etc. It doesn't help you change the languages that your keyboard will support.
    1.To determine your Win32 operating system's current ANSI CodePage (ACP). This can be found by bringing up the registry editor (Start --> Run..., type "regedit.exe", and click "OK") and looking at the
    registry entry HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\Nls\CodePage\ACP (there are many registry entries with very similar names, so please make sure that you are looking at the right place in the registry).
    2.To find the character set in the table below based on the ACP you got above.
    ANSI CodePage (ACP) Client character set (3rd part of NLS_LANG) (*1)
    1250 EE8MSWIN1250
    1251 CL8MSWIN1251
    1252 WE8MSWIN1252
    1253 EL8MSWIN1253
    1254 TR8MSWIN1254
    1255 IW8MSWIN1255
    1256 AR8MSWIN1256
    1257 BLT8MSWIN1257
    1258 VN8MSWIN1258
    874 TH8TISASCII
    932 JA16SJIS
    936 ZHS16GBK
    949 KO16MSWIN949
    950 ZHT16MSWIN950
    others UTF8 (*2)
    (*1) The character sets listed here are compatible with Win32's non-Unicode graphical user interface (GUI). Since Win32's MSDOS Box (Command Prompt) uses different character sets, NLS_LANG needs to be manually set in the MSDOS Box (or set in a batch script) in order to handle the difference
    between Win32's GUI and MSDOS Box. (Please see "NLS_LANG Settings in MS-DOS Mode and Batch Mode" in the Oracle8i Installation Guide Release 2 (8.1.6) for Windows NT, part# A73010-01.)
    (*2) If you use UTF8 for the 3rd part of NLS_LANG on Win32, client programs that you can use on this operating system would be limited to the ones that explicitly support this configuration. Recent versions of Oracle Forms' Client/Server mode (Fat-Client) on NT4.0 would be an example of such client
    programs. This is because the user interface of Win32 is not UTF8, therefore the client programs have to perform explicit conversions between UTF8 (used in Oracle side) and UTF16 (used in Win32 side).

  • Inserting Japanese in Oracle

    Hi,
    I want to insert japanese in a column of Oracle table. Session paramters NLS_LANG, NLS_CHARACTERSET and NLS_NCHAR_CHARACTERSET are American, WE8ISO8859P1 and AL16UTF16 respectively.
    I am getting the data through MS Excel, and inserting the data into table.
    Stored data is coming in as ??????????,
    Can someone please help me for this....
    its very urgent......

    Hi user635879,
    The following helped me insert into an nvarchar2 field using:
    drop table mynvarchar;
    create table mynvarchar(id int, mynv nvarchar2(100));
    insert into mynvarchar values (1, N'世界您好');
    commit;
    select * from mynvarchar;
    on a WE8ISO8859P1 database (giving the Chinese output as non ????)
    -Turloch
    from:
    Re: SQL*Plus and chinese characters
    Read      Re: SQL*Plus and chinese characters
    Posted: Dec 13, 2007 2:43 PM in response to: leonid.pavlov in response to: leonid.pavlov      
         Click to reply to this thread      Reply
    To insert Chinese characters into NVARCHAR2 columns of a 10.2 or higher database using N'literals' and SQL Developer, go to sqldeveloper\sqldeveloper\bin directory, open the file sqldeveloper.conf and add another AddVMOption line below other such lines:
    AddVMOption -Doracle.jdbc.convertNcharLiterals=true
    Note, both SQL Developer's JDBC driver and the database must be at least 10.2. Otherwise, you have to use UNISTR or the Table Data Grid.
    -- Sergiusz

  • Showing Japanese characters from Report Builder

    Hi , We've some problem showing the Japanese characters using Oracle Report builder 6i (6.0.8.17.0).
    Our database version is Oracle9i-9.0.1.4.0 and
    we'r using the DB character set as US7ASCII and National charecter set as UTF8 (Tried with AL16UTF16 but it's supported as per Oracle-metalink). We'r using Solaris 9 (OS 5.9) to invoke the Report Builder.
    Japanese characters are able to insert and select from SQL prompt using Japanese Terminal..
    But from the Report builder when we run the Report, we'r seeing some junk characters. Please help if anybody has worked on showing the Japanese characters from Report builder.
    Thanks
    -TN Reddy

    Hi Vinod and Sripathy,
    Thanks much for your feedback. We did some testing and it manage to get the results from txt/html/xml outputs but still have problem with post script format.We'r Checking this now..
    The prblem was UTF-8 was installed at the OS level (solaris 9).
    Anyway, Thank you very much.
    Regards
    -TN Reddy

  • Can display Japanese characters but can't save to db properly

    Hello! I'm having quite a predicament here! When getting Japanese characters in the database(characters are as-is), it successfully displays them. I manually put it in the database. But when I try to save to the database using the browser(textfields, submits, connections, etc), it saves as garbage.
    Here's the code on my first page. It only gets user input then submits to another jsp
    <%@ page contentType="text/html; charset=utf-8" language="java" import="java.sql.*" errorPage="" %>
    <jsp:useBean id="helper" class="com.ats.equipc.DBHelper" scope="request" />
    <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
    <html xmlns="http://www.w3.org/1999/xhtml">
    <head>
    <title>Untitled Document</title>
    </head>
    <body>
         <form name="formOne" action="testOut.jsp">
              in1:  <input type="text" name="in1" />
              <br>
              <input type="submit" value="submit" />
         </form>
         <%! ResultSet rs = null; %>
         <% rs = helper.doGetQuery("select * from t_test;");
              while(rs.next()) {
                   out.print(rs.getString(1) + "<br>");
         %>
    </body>
    </html>Don't mind the Resultset, I just put it there so I can see the saves.
    My second page. This is where I save the entry
    <%@ page contentType="text/html; charset=utf-8" language="java" import="java.sql.*" errorPage="" %>
    <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
    <html xmlns="http://www.w3.org/1999/xhtml">
    <jsp:useBean id="helper" class="com.ats.equipc.DBHelper" scope="request"/>
    <head>
    <title>Untitled Document</title>
    </head>
    <body>
         <%! ResultSet rs = null; %>
         <% request.setCharacterEncoding("utf-8");
              String entry = request.getParameter("in1");
              out.println(entry);
         %>
         <% helper.doSaveQuery("insert into t_test values('" + entry + "');"); %>
    </body>
    </html>Help please...

    Thanks once again for replying
    I got it to work but its still foggy to me why it does work.
    I replaced the code
    <%! ResultSet rs = null; %>
         <% request.setCharacterEncoding("utf-8");
              String entry = request.getParameter("in1");
              out.println(entry);
         %>with this
    <%! ResultSet rs = null; %>
         <% //request.setCharacterEncoding("utf-8");
              String entry = new String(request.getParameter("in1").getBytes("iso-8859-1"), "utf-8");
              out.println(entry);
         %>Take a look at:
    String entry = new String(request.getParameter("in1").getBytes("iso-8859-1"), "utf-8");Why do I need to specify the charset to "iso-8859-1" when getting the bytes then making it "utf-8" when finally making it into a string?

  • Japanese characters in Dreamweaver CS5

    Hi,
    I am trying to update a webpage on my company's site that is in Japanese using Dreamweaver CS5. When I had Dreamweaver CS4, I was able to just copy the Japanese characters into the design view and they would show up fine, but now they are not. When I did it before, the pasted characters shows as hex code in the code view. The characters I just pasted still look like the Japanese characters now. I don't know what changed, and I haven't been able to find a message board that addresses this, as everywhere I've looked says what I did should work. Any suggestions are welcome. Thank you!

    Hi,
    to illustrate that it works I'll show you my draft with two characters:
    http://hansgd.de/AdobTest/***.php
    Feel free to use my source code.
    Hans-Günter
    P.S.
    In my DW I can see thre "***", but the link works by clicking. If you have problems, insert "j a p" (delete spaces) instead of these stars.
    Message modified by hans-g.

  • Converting garbled characters for JAPANESE characters in a custom table

    Hi all,
    I have a custom table that store Japanese characters.
    After my company has upgraded to ECC6.0, this data in the custom table has become garbled and its alot of it garbled.
    Is there any SAP tool that can I use to make the correction on those garbled Japanese characters?
    Thanks,
    William Wilstroth

    Hi Nils,
    I really really really had a field day reading and testing around UC... To my dissappointment, I do not have the authorization to use SUMG and SCP too as well as a few of the TCODES...
    I finally told my higher level technical mgnt. that this table might need some changes...
    Has this problem of mine got anything to do with MDMP since its no longer supported in ECC6 and I found one coding that search for MDMP in RSVTPROT...
    My colleagues suggest that the data be corrected from table DBTABLOG... which i find, in my opinion, is not the right way...
    Thanks,
    William

  • Oracle Report Server Issue with Japanese Characters

    We are trying to setup a Oracle Report Server to print the Japanese characters in the PDF format.
    We have separate Oracle Report servers for printing English, Chinese and Vietnamese characters in PDF formats using Oracle Reports in the production which are running properly with Unix AIX version 5.3. Now we have a requirement to print the Japanese characters. Hence we tried to setup the new server for the same and the configurations are done as same as Chinese/Vietnamese report servers. But we are not able to print the Japanese characters.
    I am providing the details which we followed to configure this new server.
    1.     We have modified the reports.sh to map the proper NLS_LANG (JAPANESE_AMERICA.UTF8) and other Admin folder settings.
    2.     We have configured the new report server via OPMN admin.
    3.     We have copied the arialuni.ttf to Printers folder and we have converted this same .ttf file in AFM format. This AFM file has been copied to $ORACLE_HOME/guicommon/gk/JP_Admin/AFM folder.
    4.     We have modified the uifont.ali (JP_admin folder) file for font subsetting.
    5.     We have put an entry in JP_admin/PPD/datap462.ppd as *Font ArialUnicodeMS: Standard "(Version 1.01)" Standard ROM
    6.     We have modified the Tk2Motif.rgb (JP_admin folder) file for character set mapping (Tk2Motif*fontMapCs: iso8859-1=UTF8) as we have enabled this one for other report servers as well.
    Environment Details:-
    Unix AIX version : 5300-07-05-0831
    Oracle Version : 10.1.0.4.2
    NLS_LANG : JAPANESE_AMERICA.UTF8
    Font Mapping : Font Sub Setting in uifont.ali
    Font Used for Printing : arialuni.ttf (Font Name : Arial Unicode MS)
    The error thrown in the rwEng trace (rwEng-0.trc) file is as below
    [2011/9/7 8:11:4:488] Error 50103 (C Engine): 20:11:04 ERR REP-3000: Internal error starting Oracle Toolkit.
    The error thrown when trying to execute the reports is…
    REP-0177: Error while running in remote server
    Engine rwEng-0 crashed, job Id: 67
    Our investigations and findings…
    1.     We disabled the entry Tk2Motif*fontMapCs: iso8859-1=UTF8 in Tk2Motif.rgb then started the server. We found that no error is thrown in the rwEng trace file and we are able to print the report also in PDF format… (Please see the attached japarial.pdf for your verification) but we are able to see only junk characters. We verified the document settings in the PDF file for ensuring the font sub set. We are able to see the font sub setting is used.
    2.     If we enable the above entry then the rwEng trace throwing the above error (oracle toolkit error) and reports engine is crashed.
    It will be a great help from you if you can assist us to resolve this issue…

    Maybe 7zip or another tool has workarounds for broken file names, you could try that.
    Or you could try to go over the files in the zip archive one-by-one and write it to files out-1, out-2, ..., out-$n without concerning yourself with the file names. You could get file endings back via the mimetype.
    This script might work:
    #include <stdio.h>
    #include <zip.h>
    static const char *template = "./out-%04d.bin";
    int main(int argc, char**argv)
    int err = 0;
    zip_t *arc = zip_open((const char*)argv[1], ZIP_RDONLY, &err);
    if(arc == NULL)
    printf("Failed to open ZIP, error %d\n", err);
    return -1;
    zip_int64_t n = zip_get_num_entries(arc, 0);
    printf("%s: # of packed files: %d\n", argv[1], n);
    for(int i = 0; i < n; i++)
    zip_stat_t stat;
    zip_stat_index(arc, i, ZIP_FL_UNCHANGED, &stat);
    char buf[stat.size];
    char oname[sizeof(template)];
    zip_file_t *f = zip_fopen_index(arc, (zip_int64_t)i, ZIP_FL_UNCHANGED);
    zip_fread(f, (void*)&buf[0], stat.size);
    snprintf(&oname[0], sizeof(template), template, i);
    FILE *of = fopen(oname, "wb");
    fwrite(&buf[0], stat.size, 1, of);
    printf("%s: %s => %lu bytes\n", argv[1], oname, stat.size);
    zip_fclose(f);
    fclose(of);
    zip_close(arc);
    return 0;
    Compile with
    gcc -std=gnu99 -O3 -o unzip unzip.c -lzip
    and run as
    ./unzip $funnyzipfile
    You should get template-named, numbered output files in the current directory.
    Last edited by 2ion (2015-05-21 23:09:29)

Maybe you are looking for

  • Exported slices not working in Dreamweaver

    I'm hoping someone can tell me what I'm doing wrong. I created a simple rollover in fireworks, an image that when you rollover it will reveal text. I followed instructions from a tutorial, created my duplicate frame, everything looked great when I te

  • Is there a big difference?

    I'm thinking about picking up a mac mini. I was thinking about just going with the 1.8Ghz, as opposed to the 2.0Ghz. I doesn't seem like there is a big advantage with going with the 2.0Ghz. Am I missing something? Thanks.

  • Error binding ServerSocket to any of the supplied ports

    Hi, I am using Oracle coherence 3.7 version and below is the my coherence configuration proxy file: <?xml version="1.0"?> <cache-config xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"    xmlns="http://xmlns.oracle.com/coherence/coherence-cache-

  • Typo error in iTunes store categories

    Browsing the iTunes Store using the category interface shows a typo in the subcategory column. Shouldn't that be "Loading Subcategories"? Attention to detail, folks! Apple doesn't do typos.

  • InDesign CS5.5 - Adobe Acrobat export (incorrect master pages and crop marks)

    Hi, I've recently experienced a couple of anomalies when exporting books from InDesign to PDF. The first issue was that, on seemingly random pages of the book, the exporter suddenly seemed to interpret the crop marks incorrectly. The book pages would