BIG5 and HKSCS Character Set Support

Hi,
We're experiencing some problems inserting a string containing both BIG5 and HKSCS characters to a 7.3.4 Oracle DB using JDBC. The underlying character set used by the DB is ZHT16BIG5 (this cannot be changed). The characters can be inserted correctly if we use SQLPlus/WorkSheet.
Take note that the BIG5 character set can be inserted correctly. The problem occurs if we include HKSCS characters in the statement.
We have tried a number of ways already but failed to convert the data properly.
We tried converting the data using ByteToCharConverter.getConverter("Big5") but this cannot handle the HKSCS properly.
We even tried using the CharacterSet.ZHT16BIG5_CHARSET provided by the NLS character set but it cannot convert all HKSCS characters correctly.
Any ideas on how to solve this problem? Or is it because the HKSCS character set is NOT supported by the JDBC driver?
Below is a sample text containg both BIG5 and HKSCS characters:
'i$h%49D$G$Q$T89     Ize     _     ^     S(     R     @     A     Y     q
Any help/suggestion is most welcome.
Thanks,
Cis
null

I got the exact same problem as you.
(The Oracle I using is 8.1.7)
Can any one help??

Similar Messages

  • 9iLite and multibyte character set support

    Does 9iLite support a character set that will allow for accented characters?
    for example: i

    "NLS Character Integrity Issues for Consolidator
    When Mobile Sync synchronizes with an Oracle database which has a
    multibyte character set other than UTF8, the character integrity issue
    occurs. Mobile Sync retrieves data from the server database through Oracle
    8.1.7 OCI JDBC Driver for Oracle9iAS version 1.0.2.2, and 9i for Oracle9iAS
    version 2.0. Character sets are converted from database character sets to
    UTF8 by Oracle Servers NLS functions. In the code conversion, some
    multibyte characters are garbled because of the difference of the character
    mapping. This is not a bug of Mobile Sync.
    For more Information, see "Character Integrity Issues in NLS Environment"
    technical paper on Oracle Technology Network (technet.oracle.com)
    Java/SQLJ & JDBC section in Technologies category."
    from the Readme file with the media (read the manual I guess)

  • Character set supported by post and util.zip

    hi friends
    in my application i was sending a file(in client) using post method and in server getting using servlets.i what that file to be compressed using util.zip in client and uncompresed in server but i was unsucessful .i think the file before post (compressed one) and retrieving one are not same. is it problem different character set supported by http and zip files? or any thing else?
    plz help me.......

    Just ran into this myself yester day.
    There is a Metalink Note on it.
    http://metalink.oracle.com/metalink/plsql/ml2_documents.showFrameDocument?p_database_id=NOT&p_id=190281.1
    Problem Description
    You are using OC4J and trying to connect to a database using JDBC OCI and
    are getting:
    "java.sql.SQLException: Character Set Not Supported !!: DBConversion"
    Solution Description
    Replace the <OC4J_HOME>\jdbc\lib\classes12dms.jar with
    <ORACLE_HOME>\jdbc\lib\classes12.jar and rename it with classes12dms.jar.
    Explanation
    It seems there is a mismatch of classes12.zip supplied with OC4J 9.0.2
    and the Oracle9i client libraries ocijdbc8.dll or ocijdbc8.so.
    OC4J 9.0.2 does not use jdbc\lib\classes12.jar instead it uses
    jdbc\lib\classes12dms.jar. So, in order to use the 9.0.1 client with OC4J, you
    will need to make a copy of classes12.jar and rename it to classes12dms.jar
    References
    [NOTE:108876.1] Creating Connection gives "No ocijdbc8 in java.library.path"
    [NOTE:174808.1] JDev9i and OCI Connections
    I copied and renamed the jar (classes12.jar) as they stated.
    Note: It should be in the directory you set in the JDev.conf, mine is
    AddNativeCodePath D:\OraNT\9iDS\bin
    Didn't try the other reply's suggestion of setting an environment variable.

  • Character Set Support

    Hi,
    I need to open my application for international users and support multi lingual languages, mostly confined to Europe.
    I am using Forms 4.5 with Oracle 8.1.6 on HP Unix.
    My questions are
    Is UTF8 character set supported by Forms 4.5? I know forms 6i can.
    Has anyone tried to achieve it? Someone can share his experience here?
    Thanks in advance
    Sanjay
    null

    Any one...

  • Need suggestion on Multi currency and Unicode character set use in ABAP

    Hi All,
    Need suggestion. In one of the requirement I saw 'multi-currency and Unicode character set experience in FICO'.
    Can you please elaborate me how ABAPers are invlolved in multi currency as I think this is FICO fuctional area.
    And also what is Unicode character set exp.? Please give me some document of you have any.
    Thanks
    Sreedevi
    Moderator message - This isn't the place to prepare for interviews - thread locked
    Edited by: Rob Burbank on Sep 17, 2009 4:45 PM

    Use the default parser.
    By default, WebLogic Server is configured to use the default parser and transformer to parse and transform XML documents. The default parser and transformer are those included in the JDK 5.0.
    The built-in WebLogic Server DOM factory implementation class is com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl.
    The DocumentBuilderFactory.newInstance method returns the built-in parser.
    DocumentBuilderFactory dbf = DocumentBuilderFactory.newInstance();

  • How to find Database, APPL_TOP and IANA character set on 11i?

    Hi,
    Could you please tell How to find out Database character set, APPL_TOP character set and IANA character set on existing 11i environment?
    This is required to pass the input during R12 upgrade.
    Regards,
    AV

    Database:
    SQL> select value
    from V$NLS_PARAMETERS
    where parameters='NLS_CHARACTERSET';
    Application:
    $ echo $NLS_LANG
    IANA:
    Check the value of "s_iana_cset" context variable in the context file or check the value of "ICX:Client IANA Encoding" profile option.
    NLS Frequently Asked Questions [ID 399789.1]
    Oracle Applications 11i Internationalization Guide [ID 333785.1]
    How autconfig determines the value for Iana Charsets s_iana_cset value set in XML context file [ID 1380683.1]
    Thanks,
    Hussein

  • [urgent] oracle character set and national character set !!(dictionary)

    Hi. everyone.
    What is the oracle dictionary that contains information of
    oracle character set and national character set?
    I checked v$database, but there was not the information.
    It seems that there are some differences between "nls_* " init parameters
    and the database character set.
    "Alter database backup controlfile to trace" gave me the character set of db,
    but I would like to know whether there are oracle dictionary regarding them.
    Thanks in advance. Have a nice day.
    Best Regards.

    I found the dictionary which contains the information of character set and
    natiional character set of database.
    select * from nls_database_parameters
    where parameter like '%CHARACTERSET';
    Thanks for reading.
    Have a good day.
    Best Regards.

  • Expdp with network_link and different character sets for source and target?

    DB versions 10.2 and 10.1
    Is there any limitation for implementing expdp with network_link but target and source have different character sets?
    I tried with many combinations, but only had success when target and source have the same character sets. In other combinations there was invalid character set conversion (loose of some characters in VARCHAR2 and CLOB fields).
    I didn't find anything on the internet, forums or Oracle documentation. There was only limitation for transportable tablespace mode export (Database Utilities).
    Thanks

    Hi DBA-One
    This link
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm#CEGFCFFI
    is from Database Utilities 10g Release 2 (10.2) (I read it many times)
    There is only one thing about different character sets.
    I quote:
    Data Pump supports character set conversion for both direct path and external tables. Most of the restrictions that exist for character set conversions in the original Import utility do not apply to Data Pump. The one case in which character set conversions are not supported under the Data Pump is when using transportable tablespaces.
    Parameter VERSION also doesn't play role here, because the behaviour (character set) is the same if I use for target and source only 10.2 or (10.2 and 10.1) respectivly.

  • Lanovo ix4 300d and european character set for samba shares

    Hi,
    I am installing a couple of IX4-300d at a client site, struggling to set and fix the samba character set for the Windows shares. Here in Italy everyone is used to put western european characters into filenames (for example è ò ° €) but the NAS acts wierdly when it finds such caracters in folder names and file names.
    As you know samba gives the possibility to declare dos charset= ISO8859-1 and unix charset= ISO8859-1 in the smb.conf file but the option is not present in the IX4 setup pages.
    Is there an easy way to circumvent the problem and fix this option? As it is now the machine is fearly useless, at least if you need to migrate files and folder from a server or storage with ISO8859 charset already applied.
    thanks a lot
    np
    Solved!
    Go to Solution.

    This issue has been raised on all iomega drives since day 1. although 'documented' it is still shame, that units sold to 'r-o-w' means rest-of-world, outside US, are stuckto Aa-Zz09 basically wheras client systems ( windows, linux, ios ) don't care and take what they are supposed to do.
    Take it or leave it. ( I mean, take another vendor), I do not believe someone ever has taken that seriously or will do it......
    Various PCs / Laptops ( sorry I still really love Dell and Fujitsu ;-))
    Supporting Customers ix2s and ix4s -- Love Networking ( not only technically ).
    I am not a Lenovo Employee.
    If you find a post helpful and it answers your question, please mark it as an "Accepted Solution"!

  • Big5 go ZH Character Set Translation

    Dear all
    Any build-in package inside Oracle can translate Big5 to ZH character encoding
    Thank you
    Kwan
    17 Sept 2003

    Certainly, a bit more information about what you're trying to accomplish would be helpful.
    If you have a client whose NLS settings indicate it's using the Big5 character set and your database uses the ZH character set, Oracle's networking layer will automatically and transparently do the conversion.
    Justin

  • US7ASCII and WE8 character set compatibility

    Hello,
    I want to build two separates environments : one for development, one for production.
    I have currently installed OWB 9.2.0.2.8 on a WE8 character set database version 9.2.0.5. This is the development database that contains the design repository, the runtime repository and the targets schemas.
    The production database on which I'll run the packages is an US7ASCII database wersion 9.2.0.5. so, again, I will install on it the design repository, the runtime repository and the targets schemas.
    My strategy is to export a MDL file for each project, from the development repository and then import it into the production repository. Then, I will change and register the locations and generate/deploy the project to the production target schema.
    my questions are :
    - Is is possible to import into a US7ASCII database (production) a MDL coming from a WE8 (development) database ?
    - Can the generated packages process with some sources tables in an WE8 database even if the target is a US7ASCII table ? if yes, are the french accents still supported ?
    Is my stategy correct or must I work with similar character set databases?
    Thanks

    Hello,
    We finally created a new US7ASCII database. Then, we have installed OWB in this database.
    OWB version must be greater than 9.2.0.4.
    1) from the WE8 repository we export the metadata in US7ASCII characterset to a file
    2) from the US7ASCII characterset repository we import the metadata file just created.
    ==> In the US7ASCII characterset repository , all the characters with accentuation are replaced with others characters. For example é become a "["
    ==> we created a mapping in the US7ASCII characterset repository. The source is a WE8 characterset table. At execution the accentuation characters are translated to a non accentuation character. For example, the "é" character become an "e"
    In conclusion we will not use any accent characters in our repository.
    regards

  • Multi character set support

    we plan to implement Interconnect to get new customer records created in many databases (at distributed locations) to be populated into a central system.
    all the databases are running on Oracle Database 9.x and we plan to use DB adapters to access each of them.
    3 of these databases store data in double-byte format (Traditional Chinese, Simplified Chinese, Japanese). others are in standard US/American English
    my questions are:-
    1) does my central system's database have to be set to use a certain character set too ?
    2) do i have to configure Interconnect to ensure that data from non-English systems are retrieved and populated correctly into my central database
    thanks for any response

    Hi,
    We are also facing similar problem. We are using IC version 9.0.2. We have AQ adapter at source and DB adapter at destination. The characterset at DB application database is UTF-8. Where as in AQ application database is AR8MSWIN1256. We are using request/reply scenario. But the reply message which gets enqueued to the queue by AQ adapter contains '??????' (arabic data)
    We have specified the encoding type in both adapter.ini files but no use.
    We also followed the steps as stated in Metalink bug id 2375248. But nothing works. Can somebody help us???
    -Vimala

  • USER_FILTER and database character set

    Hello,
    I'm currently working on the integration of a tool into Oracle Text for filtering PDFs. My current approach is to call a command line tool via a USER_FILTER preference, and this works fine as long as the database character set is AL32UTF8. The tool is creating the filtered text as UTF-8.
    I'm struggling now with the case that the database character set is not Unicode, for example LATIN1. I had hoped that I can specify a chain of filters for this situation when creating the index, first a USER_FILTER to get the text out of the document and then a CHARSET_FILTER to convert the filtered text from UTF-8 into the database character set. This is my attempt to set this up:
    execute ctx_ddl.create_preference ('my_pdf_datastore', 'file_datastore')
    execute ctx_ddl.create_preference ('my_pdf_filter', 'user_filter')
    execute ctx_ddl.set_attribute ('my_pdf_filter', 'command', 'tetfilter.bat')
    execute ctx_ddl.create_preference('my_cs_filter', 'CHARSET_FILTER');
    execute ctx_ddl.set_attribute('my_cs_filter', 'charset', 'UTF8');
    create index tetindex on pdftable (pdffile) indextype is ctxsys.context parameters ('datastore my_pdf_datastore filter my_pdf_filter filter my_cs_filter');
    These are the error messages I'm getting (sorry, German Windows):
    FEHLER in Zeile 1:
    ORA-29855: Fehler bei Ausf³hrung der Routine ODCIINDEXCREATE
    ORA-20000: Oracle Text-Fehler:
    DRG-11004: Doppelter oder unvereinbarer Wert f³r FILTER
    ORA-06512: in "CTXSYS.DRUE", Zeile 160
    The relevant message is DRG-11004, which translates to "duplicate or incompatible value for FILTER".
    ORA-06512: in "CTXSYS.TEXTINDEXMETHODS", Zeile 364
    So here is my question:
    Do I understand it correctly that with the USER_FILTER the text is always expected in the database encoding, and that it is mandatory to create the filtered text in the database character set, or are there any alternatives?
    Thanks
    Stephan

    The previous experiments were performed with Oracle 10i. I just saw that in Oracle 11.1.0.7 there is this new feature: "USER_FILTER is now sensitive to FORMAT and CHARSET columns for better indexing performance.".
    This seems to be exactly what I was looking for.
    Regards
    Stephan

  • Discoverer 3I and Cyrilic character set

    I have a discoverer 3I EUL set up and I'm trying to pull data out that is stored in the database using the cyrilic character set. All I get, however, when I view the data is ? substituted for the cyrilic character. Is discoverer able to handle cyrilic and if so how do I get it to show the actual russian character?

    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Dave Clark ([email protected]):
    I have a discoverer 3I EUL set up and I'm trying to pull data out that is stored in the database using the cyrilic character set. All I get, however, when I view the data is ? substituted for the cyrilic character. Is discoverer able to handle cyrilic and if so how do I get it to show the actual russian character?<HR></BLOCKQUOTE>
    This is possibly a patch issue. For 3i this requires version 3.3.59.
    null

  • Oc4j and international character sets

    Does anyone know if there is a problem in oc4j concerning international charactersets support?
    My pages contain non-english characters and so i use utf-8 characterset. When the page is created by a servlet the characters are displayed correctly, but when it comes from a jsp this is not the case.
    The problem is that the text is not displayed in utf-8 but in "western european (windows)" instead which is not the correct encoding. This does not happen in the servlet case above. Still if I change the borwser encoding setting
    the information is displayed correctly. But I cannot do that for every page, they are to many...
    If i try to instruct the page about the encoding the outcome is a real mess because no brwser characterset encoding configuration option seems to be suitable. I tried the following:
    - the page directive: <%@page contentType=... pageEncoding...
    - the request: request.setContentType
    - oc4j configuration: default-charset configuration in the global-application.xml config file
    - and the oracle extension <% serwriterEncoding(...) ; %>
    all of the above had no result while servlets perform swell !
    Am I missing something here?
    Does any body knows something that I do not?
    Does the jsp container mess with the encoding and sends to the browser incorrect page encoding?
    Thank you in advance,
    Joe

    HI Joe:
    Are you able to find any solution for this. I am also stuck with the same issue. Could you please share your solution to me. my email id is: [email protected]
    Thanks & Regards
    Sridhar Doki

Maybe you are looking for