Problem with character set UTF-16 LE

Hello.
There were difficulties with character set change, using function convert ()
The matter is that in the list v$nls_valid_values the character set AL16UTF16LE does not appear, during too time the inquiry is successfully carried out
convert ([some-national-characters], ' CL8MSWIN1251 ', ' AL16UTF16LE ').
But at giving on an input of the data stored in CLOB, there is an error "a character set is not supported"
What are possible ways of the decision of a problem?

You can try to use DBMS_LOB. SUBSTR to access LOB data like in the following example:
SQL> select * from v$version;
BANNER
Oracle Database 10g Express Edition Release 10.2.0.1.0 - Product
PL/SQL Release 10.2.0.1.0 - Production
CORE    10.2.0.1.0      Production
TNS for 32-bit Windows: Version 10.2.0.1.0 - Production
NLSRTL Version 10.2.0.1.0 - Production
SQL>
SQL> declare
  2  v_i clob;
  3  v_o clob;
  4  begin
  5  v_i := 'a';
  6  v_o:=convert(dbms_lob.substr(v_i,1,1),'AL16UTF16LE', 'CL8MSWIN1251');
  7  end;
  8  /
PL/SQL procedure successfully completed.

Similar Messages

  • Problem with character set - Reports 11.1.1.4

    Hi!
    I have a problem with Oracle Reports 11g regarding character set configuration. The default character set WE8ISO8859P1 works, so PDF reports have a regular display except for Eastern European (EE) letters which are replaced by "¿" sign.
    So, when I set any other character set in reports.sh, which would be a normal step to get EE letters, I'm always getting Greek Alphabet in PDF reports. Why Greek Alphabet?
    The character sets I tried to use are: EE8ISO8859P2, UTF8 and AL32UTF8.
    I changed uifont.ali and included PDF Subset with all four Arial font variants and, of course, I placed all fonts in fonts folder which is pointed by REPORTS_FONT_DIRECTORY.
    In Reports Builder everything works fine, but when I have to deploy the report to the Reports Services, the problem occurs.
    Also, when I've tried to execute PDF report using In-Process Reports Server (rep_wls_reports_hostnameasinst1) instead of AS Instance Reports Server (RptSvr_hostnameasinst1, which is a regular server), I'm getting Greek Alphabet in PDF reports even if the default character set is WE8ISO8859P1 in reports.sh. What is wrong with it? Where is Greek Alphabet configured?
    The production environment is 64-bit Oracle Linux 5.6 with Weblogic 10.3.4 and Forms&Reports 11.1.1.4. Forms works fine with character set EE8ISO8859P2 defined in default.env file.
    Thanks in advance!
    Regards,
    Dejan

    Thank you, Denis!
    Doc 300416.1 was very useful but Note 356221.1 - A Practical Methodology on Porting Reports from Windows to Unix with Different Font is actually crucial for configuring Reports on Linux.
    Also, there is a bug in 11.1.1.3 and 11.1.1.4, which can be fixed using the patch ( Note 1138405.1 - PDF Reports With Font Subsetting Raises Error "Bad /Bbox" on 64-Bit Linux ).
    Kind regards!

  • Problem with character set (ubuntu linux)

    hello everyone.
    I 'have already installed oracle-xe-universal_10.2.0.1-1.1_i386 in ubuntu.
    The problem is that greek characters from the db appear like ??????.
    How can i set the right nls_lang character set to solve the problem in ubuntu linux?
    Thank you in advance!

    Character code point translation is in the realm of the client, try setting NLS_LANG environment variable.
    If your client programs handle UTF8, as do many linux utilities, that is the best choice, set your <language>_<locale>.<characterset>, i.e.
    export NLS_LANG=AMERICAN_AMERICA.AL32UTF8
    Or french_france, or german_germany ... depends on the locale you want to use. Try a `putty` session to the host, the terminal can be set for many different character sets, i.e. ISO8859-<n> a.k.a. the WE8<n> or Western European, UTF, etc. under Window/Translation.
    Try a few select ... from dual; statements in sqlplus with a literal, and different unicode values- the unistr() and dump() functions can come in quite handy.
    select unistr('\ac20 euro' ) from dual;
    ... € euro ...
    select dump( unistr('\ac20 euro' ) ) from dual;
    ... <ascii codes for the character values> ...
    select dump( unistr('\ac20 euro' ),16 ) from dual; -- this one for hex dump

  • Problem with Character Set in Oracle database 10g

    Hi,
    I tried to import one tablespace into test server. Source server with Oracle 8i and Target server with Oracle database 10g. The error I get is
    Import: Release 10.2.0.1.0 - Production on Thu Aug 3 00:20:49 2006
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Username: sys as sysdba
    Password:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V08.01.07 via conventional path
    About to import transportable tablespace(s) metadata...
    import done in WE8DEC character set and AL16UTF16 NCHAR character set
    export server uses WE8DEC NCHAR character set (possible ncharset conversion)
    . importing SYS's objects into SYS
    . importing SYS's objects into SYS
    IMP-00017: following statement failed with ORACLE error 19736:
    "BEGIN sys.dbms_plugts.beginImport ('8.1.7.4.0',2,'2',NULL,'NULL',67051,25"
    "51,2); END;"
    IMP-00003: ORACLE error 19736 encountered
    ORA-19736: can not plug a tablespace into a database using a different national character set
    ORA-06512: at "SYS.DBMS_PLUGTS", line 2386
    ORA-06512: at "SYS.DBMS_PLUGTS", line 1946
    ORA-06512: at line 1
    IMP-00000: Import terminated unsuccessfully
    PLZ somebody help in geting resolve this. Has anybody seen this error before.

    The solution to this problem is described in MetaLink note #211920.1. But this note is published with LIMITED access as it involves using a hidden parameter.
    You can get access to the note through Oracle Support only.
    The problem itself is solved generically, if the source database is at least 10.1.0.3 and the target database is 10.2
    -- Sergiusz

  • Problem with Character Set after upgrade

    Hello,
    I have a probelm and was wondering if anyone has seen this before.
    I have been running Java 1.4.2_13 for a while now on some Windows servers. About a month ago we upgraded to Java 1.6.0_12.
    We are reading input files that are in the character set Big5 HKSCS. After the upgrade our application started to report certain characters as invalid. Some of the characters are 0xD843, 0xD844.
    These should be valid characters and were in the previous version of Java (1.4.2_13).
    Has anyone seen this?
    Thanks in advance,
    Jerry

    The [Supported Encodings|http://java.sun.com/javase/6/docs/technotes/guides/intl/encoding.doc.html] page for the Java 6 release describes that charset as
    Big5 with Hong Kong extensions, Traditional Chinese (incorporating 2001 revision)You would know more than I do about that 2001 revision (or at least, not less than I do, since I know nothing about it). Perhaps that's the source of your problem?
    Edit: especially since the [Supported Encodings|http://www.j2ee.me/j2se/1.4.2/docs/guide/intl/encoding.doc.html] page for Java 1.4 doesn't mention the 2001 revision.

  • Problem with Character set

    Hi,
    I am facing some problems with Oracle character sets.
    I have set the database character set as UTF8.
    I have set the environment variable NLS_CHAR to Portuguese_Portugal.UTF8.
    I insert a Portuguese text for varchar and clob columns in a table.
    However when I select the text from that table, I get some junk values for the clob.
    Any help is appriciated. Thanx in advance.
    Yash
    null

    Hi Yash,
    Can u tell me how to insert values in a nchar/nvarchar column in
    Indian language scripts?
    I have set database characterset to 'IN8ISCII'? But while
    inserting values in the table I am getting error as characterset
    mismatch'
    I have tried it using UTF8 chracterset also??
    As fa as your problem is concerened, I think u have to set the
    nls_ characterset value$ in props$ table
    and same for nls_nchar_characterset column also..
    Thanks in advance
    Manoj mehta

  • Problem with character set with servlets and Oracle 9i AS

    I am deploying a servlet with Jserv on a Sun machine.
    My servlet is working but some special characters in text fields like "i", "h", "`" (French language) are replaced by "?" when refreshing the HTML pages.
    Is that a Jserv or Oracle 9i AS settings problem or machine settings problem ?
    Thanks for you help
    Lilian

    The solution to this problem is described in MetaLink note #211920.1. But this note is published with LIMITED access as it involves using a hidden parameter.
    You can get access to the note through Oracle Support only.
    The problem itself is solved generically, if the source database is at least 10.1.0.3 and the target database is 10.2
    -- Sergiusz

  • Problems with Http-servlet : UTF-8

    I am having some difficulty with UTF-8 encoded
    chracaters in a Java servlet.
    My servlet accepts an XML file conteining a question and returns an HTM page. The XML has cyrillic characters encoded as utf-8.. The rendering
    servelt copes with this fine, and the HTML produced
    displays OK in the browser (the response type on the
    Java servelet has to be set to "text/html;
    charset=UTF-8" for this to work).
    I have to send cyrillic characters back in the
    response to the question in a text field in the HTML form.The browser is
    correctly sending back the byte stream (which I am
    printing here as hex): d0b3d0bed180d0bed0b4 (this is a
    cyrillic word correctly coded as utf-8).
    However, on collecting the response (using
    request.getParameterValues(fieldname)) the servlet
    returns the byte stream: d0b3d0bed13fd0bed0b4.
    A mistake in the fifth byte.
    Can anyone help with this problem? Is there a known problem with the JAVA UTF-8 converter?
    Regards
    Graham

    I now know the answer to this problem thanks to Bruno Van Haetsdaele .
    Before calling request.getParameterValues(fieldname));
    one should call request.setCharacterEncoding("UTF-8");
    Hope that helps somebody else!

  • Problem with subform set

    Hi all ,
    I am facing a problem with subform set. When i use two subform sets with same set of conditions in a form, only the first subform set works while the second one does nothing. Did any of you faced the same problem ? If so Please let me know the solution you came up with. Thanks.
    SrPrabhu

    Hi scott,
    The problem is with subform set and not subform. In subform set, we can give conditions to make only one subform visible at a time that satisfies the condition. First create a subform set having more than one subform. Now apply the conditions to that subform set. Check if it works. I m sure it wil work. Now just make a duplicate copy of the same subform set. Place it in the same form so that now two subform sets are available. Now preview the form. U wil fine only one subform set working and the other one doesnt.
    We can use any number of subform set within a form. If different conditions are applied to each subform set , it works fine.
    i wonder now wat wud u do in case if we are in need of two subformset having the same condition in a form ?
    Thanks,
    Sr Prabhu

  • Problem with character code NCR (example : cộng h�a x� hội)

    Dear
    I have problem with character code NCR when display this string "c&#7897;ng h�a x� h&#7897;i" on web page using JSF. I print out like this c & # 7897 ; ng h � a x � h & # 7897 ; i
    Thanks for help

    jverd wrote:
    A better approach would be to take a char rather than a String in that method, since it's a better model for what you're converting. A char would also let you use switch statement.I was going to say this as well but you beat me to it.
    @OP, you really should use char for this. You're unnecessarily taking each char, converting it to a string, and working with the single-character string. That's not very efficient.

  • Weird Problem with FSB setting

    Hi all,
    I've just come across a weird problem with FSB setting.
    If I set FSB to 229 my machine runs @ 1.6Ghz only.
    When I set to 230 it runs @ 3.2Ghz (as it should).
    Any ideas what could be causing this?
    Thanks
    Mick

    If FSB @ =
      Strange thing.. What a magical 1MHz  

  • Database character set = UTF-8, but mismatch error on XML file upload

    Dear experts,
    I am having problems trying to upload an XML file into an XMLType table. The Database is 9.2.0.5.0, with the character set details:
    SELECT *
    FROM SYS.PROPS$
    WHERE name like '%CHA%';
    Query results:
    NLS_NCHAR_CHARACTERSET          UTF8     NCHAR Character set
    NLS_SAVED_NCHAR_CS          UTF8
    NLS_NUMERIC_CHARACTERS          .,     Numeric characters
    NLS_CHARACTERSET          UTF8     Character set
    NLS_NCHAR_CONV_EXCP          FALSE     NLS conversion exception
    To upload the XML file into the XMLType table, I am using the command:
    insert into XMLTABLE
    values(xmltype(getClobDocument('ServiceRequest.xml','UTF8')));
    However, I get the error:
    ORA-31011: XML parsing failed
    ORA-19202: Error occurred in XML processing
    LPX-00200: could not convert from encoding UTF-8 to UCS2
    Error at line 1
    ORA-06512: at "SYS.XMLTYPE", line 0
    ORA-06512: at line 1
    Why does it mention UCS2, as can't see that on the Database character set?
    Many thanks for your help,
    Mark

    USC2 is known as AL16UTF16(LE/BE) by Oracle...
    Try using AL32UTF8 as the character set name
    AFAIK The main difference between Oracle's UTF8 and AL32UTF8 character set is that is the UTF8 character set does not support those UTF-8 characteres that require 4 bytes..
    -Mark

  • HTTPService + SimpleEmailForm - Problem with character

    Problem with swedish character å, ä och ö in a
    simple form mail with flex, php and mySQL.
    Have tried to change this in flex - mySimpleFormMail.mxml:
    <?xml version="1.0" encoding="utf-8"?> to
    <?xml version="1.0" encoding="Latin1"?> and to
    <?xml version="1.0" encoding="iso-8859-1"?>
    Have also at the same time changed this in mail.php:
    <meta http-equiv="Content-Type" content="text/html;
    charset=utf-8" /> to
    <meta http-equiv="Content-Type" content="text/html;
    charset=Latin1" /> and to
    <meta http-equiv="Content-Type" content="text/html;
    charset=iso-8859-1" />
    The mySimpleFormail looks like this is this:
    From---------------------------------------
    Name----------------------------------------
    To------------------------------------------
    E-mail---------------------------------------
    Mesage---------------------------------------
    SEND (button)
    The e-mail generated from mySimpleFormail looks like this:
    From: Patrik Hällman
    Email Address: [email protected]
    To:
    Patrik Hällman
    Email Address: [email protected]
    Message:Now I´m testing with the swedish character -
    åä och ö.
    ...as you can see swedish character å, ä och ö
    is wrong displayed.
    Here is the result from <traceTarget>:
    [SWF] C:\AppServ\www\email \bin-debug\email_send4.swf - 910
    229 bytes after decompression
    'AB8F6784-15FC-7BD4-D26D-F73192A7717F' producer set
    destination to 'DefaultHTTP'.
    'direct_http_channel' channel endpoint set to http:
    'AB8F6784-15FC-7BD4-D26D-F73192A7717F' producer sending
    message 'AA84F829-98F0-E93F-D7AC-F732847CF6B2'
    'direct_http_channel' channel sending message:
    (mx.messaging.messages::HTTPRequestMessage)#0
    body = (Object)#1
    emailMessage = "Now I´m testing with the swedish
    character å, ä and ö."
    senderEmail = "[email protected]"
    senderName = "Patrik Hällman"
    sendToEmail = "[email protected]"
    sendToName = "Patrik Hällman"
    clientId = (null)
    contentType = "application/x-www-form-urlencoded"
    destination = "DefaultHTTP"
    headers = (Object)#2
    httpHeaders = (Object)#3
    messageId = "AA84F829-98F0-E93F-D7AC-F732847CF6B2"
    method = "POST"
    recordHeaders = false
    timestamp = 0
    timeToLive = 0
    url = "
    http://www.myhomesite.se/email/email.php"
    'AB8F6784-15FC-7BD4-D26D-F73192A7717F' producer connected.
    'AB8F6784-15FC-7BD4-D26D-F73192A7717F' producer acknowledge
    of 'AA84F829-98F0-E93F-D7AC-F732847CF6B2'.
    Decoding HTTPService response
    Processing HTTPService response message:
    (mx.messaging.messages::AcknowledgeMessage)#0
    body = "<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0
    Transitional//EN" "
    http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
    <html xmlns="
    http://www.w3.org/1999/xhtml">
    <head>
    <meta http-equiv="Content-Type" content="text/html;
    charset=iso-8859-1" />
    <title>Untitled Document</title>
    </head>
    <body>
    </body>
    </html>
    clientId = "DirectHTTPChannel0"
    correlationId = "AA84F829-98F0-E93F-D7AC-F732847CF6B2"
    destination = ""
    headers = (Object)#1
    messageId = "5624FC90-E89E-2B47-92B9-F7328A8CB993"
    timestamp = 0
    timeToLive = 0
    This is catastrophic when you working with cms and databases.
    All help I can get from this forum is really appreciated.

    Well, using UNICODE for upload/download files I think is most suitable way. If in AL11 you don't see it being displayed correctly (odd characters appear), that doesn't necessary mean it has wrong encoding set. Maybe AL11 is using different decoding type by default. Normally storage of file on AS is not its final destination, right? You place it there i.e. for customer's job to read it back and send via email. So just make sure that downloading program uses same decoding you used for encoding (UTF-8). Then the characters in the output stream (i.e. frontend file, mail body) should be displayed correctly
    What I mean is don't worry how it is dispalyed in AL11, just ensure that after you download it, the characters are correctly decoded.
    PS: Make sure that you metion encoding type explicitly both for upload and download programs
    OPEN DATASET dset .... IN TEXT MODE ENCODING/DECODING UTF-8.
    Regards
    Marcin

  • JSF and Character Sets (UTF-8)

    Hi all,
    This question might have been asked before, but I'm going to ask it anyway because I'm completely puzzled by how this works in JSF.
    Let's begin with the basics, I have an application running on an OC4J servlet container, and am using JSF 1.1 (MyFaces). The problems I am having with this setup, is that it seems that the character encodings I want the server/client to use are not coming across correctly. I'm trying to enforce the application to be UTF-8, but after the response is rendered to my client, I've magically been reverted to ISO-8859-1, which is the main character set for the netherlands. However, I'm building the application to support proper internationalization; which means I NEED to use UTF-8.
    I've executed the following steps to reach this goal:
    - All JSP files contain page directives, noting the character set:
    <%@ page pageEncoding="UTF-8" contentType="text/html; charset=UTF-8" %>I've checked the generated source that comes from the JSP's, it looks as expected.
    - I've created a servlet filter to set the character set directly on the request and response objects:
        public void doFilter(ServletRequest req, ServletResponse res, FilterChain chain) throws IOException, ServletException {
            // Set the characterencoding for the request and response streams.
            req.setCharacterEncoding("UTF-8");
            res.setContentType("text/html; charset=UTF-8");       
            // Complete (continue) the processing chain.
            chain.doFilter(req, res); 
        }I've debugged the code, and this works fine, except for where JSF comes in. If I use the above situation, without going through JSF, my pages come back UTF-8. When I go through JSF, my pages come back as ISO-8859-1. I'm baffled as to what is causing this. On several forums, writing a filter was proposed as the solution, however this doesn't do it for me.
    It looks like somewhere internally in JSF the character set is changed to ISO. I've been through the sources, and I've found several pieces of code that support that theory. I've seen portions of code where the character set for the response is set to that of the request. Which in my case coming from a dutch system, will be ISO.
    How can this be prevented? Can anyone give some good insight on the inner workings of JSF with regards to character sets in specific? Could this be a servlet container problem?
    Many thanks in advance for your assistance,
    Jarno

    Jarno,
    I've been investigating JSF and character encodings a bit this weekend. And I have to say it's more than a little confusing. But I may have a little insight as to what's going on here.
    I have a post here:
    http://forum.java.sun.com/thread.jspa?threadID=725929&tstart=45
    where I have a number of open questions regarding JSF 1.2's intended handling of character encodings. Please feel free to comment, as you're clearly struggling with some of the same questions I have.
    In MyFaces JSF 1.1 and JSF-RI 1.2 the handling appears to be dependent on the raw Content-Type header. Looking at the MyFaces implementation here -
    http://svn.apache.org/repos/asf/myfaces/legacy/tags/JSF_1_1_started/src/myfaces/org/apache/myfaces/application/jsp/JspViewHandlerImpl.java
    (which I'm not sure is the correct code, but it's the best I've found) it looks like the raw header Content-Type header is being parsed in handleCharacterEncoding. The resulting value (if not null) is used to set the request character encoding.
    The JSF-RI 1.2 code is similar - calculateCharacterEncoding(FacesContext) in ViewHandler appears to parse the raw header, as opposed to using the CharacterEncoding getter on ServletRequest. This is understandable, as this code should be able to handle PortletRequests as well as ServletRequests. And PortletRequests don't have set/getCharacterEncoding methods.
    My first thought is that calling setCharacterEncoding on the request in the filter may not update the raw Content-Type header. (I haven't checked if this is the case) If it doesn't, then the raw header may be getting reparsed and the request encoding getting reset in the ViewHandler. I'd suggest that you check the state of the Content-Type header before and after your call to req.setCharacterEncoding('UTF-8"). If the header charset value is unset or unchanged after this call, you may want to update it manually in your Filter.
    If that doesn't work, I'd suggest writing a simple ViewHandler which prints out the request's character encoding and the value of the Content-Type header to your logs before and after the calls to the underlying ViewHandler for each major method (i.e. renderView, etc.)
    Not sure if that's helpful, but it's my best advice based on the understanding I've reached to date. And I definitely agree - documentation on this point appears to be lacking. Good luck
    Regards,
    Peter

  • Document character set UTF-8

    Hi, sorry for my bad english. I am Brazilian. So..
    I have problem when i upload the muse site to the hosting... problems with: document character set. They are configured to UTF-8... put characters "á" and "é" didnt work... but it is just on this hosting "lgplasticos.com.br" and i had uploaded the same files in another hosting and its fine "weking.com.br/clientes/lg"... how to solve this? I contacted the hosting support and they said that: "is no problem in our server configuration , probably some wrong configuration file when sending files ..." but in weking.com/clientes/lg server worked ... how to solve ? Please help me ..

    Hi,
    did you check this thread How to change charset from UTF-8 to ISO-8859-1 in Muse?

Maybe you are looking for