How to display " ISO-8859-1 " (eg - " &eacute " ) characters in LabelField .

hai,
How to display  " ISO-8859-1 " (eg - " &eacute " ) characters in LabelField .
Regards
Ratheesh R Kurup

Hi,
Perhaps
with a as
select 'TESTING' text from dual
select level, substr(a.text,level,1)
from a
connect by level <= length(a.text)

Similar Messages

  • How to get ISO-8859 characters from DOM ?

    Hi,
    I have problems to get the ISO-8859 characters from the DOM. I parse a XML file and read the values from the DOM. All "umlauts" are scrambled.
    The details:
    The first line of the XML file is:
    <?xml version="1.0" encoding="ISO-8859-1"?>
    It contains lines like:
    <tag>Dpfel</tag>
    I perform the following steps:
    I'm parsing the XML file ...
    xmlparser.parse( parser, full_fname_xml );
    ... get the document ...
    doc := xmlparser.getDocument( parser );
    ... and read the value from the TEXT_NODE:
    nodevalue := xmlparser.getNodeValue( node );
    And here is the problem: nodevalue does
    not contain 'Dpfel' but '?pfel'.
    Its not a problem of the database. Explizit INSERTs with "umlauts" e.g. are working fine.
    This problem occurs with the XML Parser for PLSQL 1.0.0.1.0 on NT.
    [Pls do not recommend to use the newer version. I did a test with the newer version and know it works there.]
    My questions relates only to this older version:
    Will it never work with this version
    or what am I doing wrong ?
    Any hints welcome !
    Tnx
    Franz
    null

    Tnx for the quick answer.
    Sorry, indeed I can not use the newer
    version in the moment.
    So I urgently need a solution or any
    workaround for this version.
    F.

  • How to display - Message Length more than 100 characters

    Hi,
    I am displaying a message string which gets generated dynamically. Which means I donno theolength of that. It may go upto 200 characters also.
    However, In message display, the number of characters have some limit.
    Could you please let me know, How can I achieve this?
    MESSAGE E000 with lv_string.
    Thanks,
    Sandeep

    Hi Srini,
    I have already 4 place holders. And one of the place holder will be populated dynamically which I mentioned above. And more than 4 place holders, it is not accepting.
    Could you please help me.
    Thanks,
    Sandeep

  • How to display feeds with non-latin utf8 characters in Raggle?

    Has anyone tried to use raggle to read feeds with non-latin utf8 characters?
    If you are successful, how to do it?
    Thanks

    i have this problem too...
    Last edited by vdo (2008-09-02 12:19:31)

  • How convert ISO 8859-1 (Latin-1) to char

    i use java read some web page, it contain ISO 8859-1 (Latin-1) Characters, like this:
    &#233
    &#234(i removed ";" at end)
    how convert it to char?
    thanks.

    Tin,
    I think you're looking for Charset.decode and maybe encode.
    keith.

  • WIN-1252 characters in ISO-8859-1 textfields

    My questing concerns the following situation :
    The Apex-Dads is configured to be WE8ISO8859P1, the database has the same character set.
    When submitting a text-field in an APEX-page, a user can submit win1252 characters : chr:128-159, although the page encoding is
    ISO-8859-1.
    These characters are even displayed on the apex-page(in IE, FF on windows) as the represented win1252 character.
    However when these characters are inserted in the database, I have non-display characters.
    My question is, how can I prevent these characters from being entered in the database ?
    I've tried these
    * I can assume that all input arrives in Win1252 and convert all data in the database. (seems not the optimal solution)
    * I can configure the dad to be UTF8, so there will be a conversion from UTF-8-&gt; ISO-8859-1.
    However some characters aren't translated properly.
    Partly this is acceptable : chr128 (euro) isn't available in ISO-8859-1 but chr146 : right shifted quote is 'translated' to ¿ (inverted quotation mark).
    Perhaps someone had the same experience and can give me some more info or tips.
    Thanks in advance,
    Art
    Edited by: amels on Sep 19, 2008 1:11 PM

    Hello Art,
    >> Do you know why this is / should be ?
    This setting became mandatory in version 2.0, when the AJAX framework was introduced. Since then, the APEX environment itself is massively using this technology. Although the XMLHttpRequest object can support deferent character sets, its default setting is UTF-8. Using AL32UTF8 in the DAD simplify the “behind the scenes” AJAX support. This configuration also simplifies import/export of APEX applications and data, minimizing some client-server character set conversions, and I probably don’t know all the reasons that led to make this configuration a mandatory one.
    The important thing to remember is that it is a mandatory setting, and ignoring it will defiantly cause you problems in functionality, both in the development environment, and the run-time environment.
    Regards,
    Arie.

  • Big5 to ISO-8859-1

    Hi, I want to convert a big5 chinese character to ISO-8859-1 character (&#xxxxx;)
    Here is my code:
    String record = "���~"; //a big5 string
    System.out.println("BIG5: " + record); //Display ok
    byte[] b = record.getBytes("ISO-8859-1");
    String target = new String(b, "ISO-8859-1");
    System.out.println("Target: " + target);
    But I can't get the ISO-8859-1 code which like &#xxxxx; but gives me ???.
    Please advice.

    ISO-8859-1 DO have chinese characters
    In JSP, when I submit a chinese value in a form, I "ll
    receive this chinese characters in ISO-8859-1
    encodings and get: &#22909;&#20154; (In chinese:
    �n�H)
    And when I put this ISO-8859-1 characters in the value
    field of a form, the html gives me chinese character
    in Big5 correctly.
    I just don't know how to convert ISO-8859-1 to Big5 in
    Java and vice versa.
    Please help.ISO-8859-1 doesn't support chinese characters.
    The reason that you can received chinese characters from a html form was because the charset of that html page was set to BIG-5. When a user makes a request, all request parameters and values will be encoded with the charset of the HTML page.

  • Java App on Linux : Unable to read iso-8859-1 encoded file correctly.

    I have a file which is encoded as iso-8859-1, and contains characters such as ô .
    I am reading this file with java code, something like:
    File in = new File("myfile.csv");
    InputStream fr = new FileInputStream(in);
    byte[] buffer = new byte[4096];
    while (true) {
    int byteCount = fr.read(buffer, 0, buffer.length);
    if (byteCount <= 0) {
    break;
    String s = new String(buffer, 0, byteCount,"ISO-8859-1");
    System.out.println(s);
    However the ô character is always garbled, usually printing as a ? .
    I am running this on a Linux machine. It works fine on my XP machine.
    I have verified that I can see the correct characters when I cat the file on the terminal.
    (Interestingly, but I think maybe only by co-incidence, it works when I run with the -Dfile.encoding=UTF16 option, but not with UTF8, although this appears a hack rather than a fix since this option was not intended for developer use by sun - but I thought mentioning it may provide some clues as to what is going on)

    I think your main probelm is with the console. When you send text to the console, it's sent in the system default encoding. On an English-locale system that might be ASCII, ISO-8859-1, windows-1252, UTF-8, MacRoman, and probably several other possibilities. Then the console decodes the the bytes using whatever encoding it feels like using--on my WinXP machine, it uses cp437 by default (just for laughs, as far as I can tell). If the text happens to be pure, seven-bit ASCII, there's no problem, since all those encodings are identical in that range.
    But if you need to output anything other than ASCII characters, avoid the console. Send the output to a file and specify an encoding that you know will be able to handle your characters--UTF-8 can handle anything. Then open the file with an editor that can read that encoding; most of them can handle UTF-8 these days, and many will even detect it automatically. You also need to be using a font that can display your characters.
    However, you're also going about the reading part wrong. Instead of reading the text in as bytes and passing them to a String constructor, you should use an InputStreamReader and read it as text from the beginning: BufferedReader br = new BufferedReader(
      new InputStreamReader(
        new FileInputStream("myfile.csv"), "ISO-8859-1"));I am curious about your statement that "it works" when you run with the -Dfile.encoding=UTF16 option. I wouldn't be surprised to see it output the correct characters (ASCII characters, anyway), but I would expect to see the characters interspersed with blank spaces or rectangles.

  • How change n900 xterm to sv_SE.ISO-8859-1?

    I ssh from n900 to a debian server, which has locale sv_SE.ISO-8859-1. This mostly works fine, except when reading/writing email in mutt. I have had this problem with other terminals and usually solve it by changing locale to ISO-8859-1 (yes, I am not interested in UTF-8), but I can't find any instructions for changing n900 xterm to ISO-8859-1. Anyone know how to do this? Thanks!

    I have discarded the option of converting to UTF-8. Too much time and energy is invested in my old files/emails/environments to change all that just because of the N900 terminal. Besides, it seems that if I start sending out UTF-8 emails, they will display as garbage in my friends inboxes. This however might be the result of me not having a complete UTF-8 environment.
    Anyway, I tried the change to 8859-1 with gconftool. And it worked! But just once... After a reboot, I am back to UTF characters although gconftool -g still shows 8859-1. Setting 8859-1 again doesn't change anything now. I even tried setting back to UTF-8 and then 8859-1, but that didn't work either.
    However, maybe there is a way forward if I rephrase the problem. I don't care what the character encoding is in the N900, the only thing of interest to me is that I get ISO-8859-1 when I ssh to my servers. Would that problem be solvable by doing something in the ssh setup or the environment or something else?
    Another question about this: Even though I have 8859-1 in gconftool, LANG and LC_* are still sv_SE. Is this normal, or should they have changed?
    In my servers I have LANG and LC_* set to sv_SE.ISO-8859-1.
    Thanks very much for any help (I'm getting a bit desperate about this now) !

  • Website not displaying correctly. Firefox is changing the character set to Western (ISO-8859-1) automatically.

    Normally I have set Firefox (or it's set by default) to Character Set Unicode (UTF-8) and everything displays perfectly. I've never had a problem before.
    Now however, whenever I upload my own website, for some bizarre reason on that particular tab (and only that tab) the Character Set is changed over to Western (ISO-8859-1) and then there's a few characters within my site that do not display correctly, namely apostrophes and hypens.
    It definitely isn't my software (Serif WebPlus X4) because the page displays correctly in every other browser. Plus it displays correctly in Firefox if I change the Character set back to Unicode.
    PS The site is a work in progress

    That happens because the server sends a content-type (<b>text/html; charset=ISO-8859-1</b>) via the HTTP response headers and in that case that content type prevails. The page code is saved with an UTF-8 byte order mark () that you see in this case.
    *http://web-sniffer.net/?url=http%3A%2F%2Fwww.valuevisionglasses.co.uk&http=1.1&gzip=yes&type=HEAD&uak=0
    *http://httpd.apache.org/docs/current/mod/mod_mime.html#AddType

  • How to switch DB string storage format from/to UTF-8 to/from ISO 8859-1 ?

    As far as I understand strings in tables are stored by default in an Oracle DB in ISO 8859-1 format.
    How can I switch the storage to UTF-8 format?
    Do I have to change just a parameter (which ?) or do I have to setup/install the whole DB again?
    If just a parameter switching is necessary:
    How can I change already existing strings from one format to another?
    Does it take place automatically or do I have to issue and explicite convert command (which ?).
    Peter

    Please refer to
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14225/ch2charset.htm#sthref157
    And
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14225/ch11charsetmig.htm#sthref1476
    Sybrand Bakker
    Senior Oracle DBA

  • How to store UTF-8 characters in an iso-8859-1 encoded oracle database?

    How can we store UTF-8 characters in an iso-8859-1 encoded oracle database? We can NOT change the database encoding but need to store e.g. Polish or Russian characters besides other European languages.
    Is there any stable sollution with good performance?
    We use Oracle 8.1.6 with iso-8859-1 encoding, Bea WebLogic 7.0, JDK 1.3.1 and the following thin driver: "Oracle JDBC Driver version - 9.0.2.0.0".

    There are a couple of unsupported options, but I wouldn't consider using them on a production database running other critical applications. I would also strongly discourage their use unless you understand in detail how Oracle National Language Support (NLS) works, otherwise you could end up with corrupt data or worse.
    In a sense, you've been asked to do the impossible. The existing databas echaracter sets do not support encoding the data you've been asked to store.
    Can you create a new database with an appropriate database character set and deploy your application there? That's probably the easiest solution.
    If that isn't an option, and you really need to store data in this database, you could use one of the binary data types (RAW and BLOB), but that would mean that it would be exceptionally difficult for applications other than yours to extract the data. You would have to ensure that the data was always encoded in the same character set, otherwise you wouldn't be able to properly decode it later. This would also add a lot of complexity to your application, since you couldn't send or recieve string data from the database.
    Unfortunately, I suspect you will have to choose from a list of bad options.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • How to interface greek fonts (iso-8859-7) with labview for linux 8.5.1

    I am trying to interface an Labview Linux application with greek fonts ,how i can achieve this out ? I want to use iso-8859-7 icharset. I am using a gentoo distro. Can anyone help me  ?
    thanks....
    Ελευθερία σημαίνει ότι μαθαίνεις να έχεις απαιτήσεις μόνο από τον εαυτό σου, όχι από τη ζωή ή τους άλλους

    Hi Sak1s,
    Are you looking to create labels and captions with Greek fonts?  You may be interested in the LabVIEW Help topic on Localizing VIs. 
    Jennifer R.
    National Instruments
    Applications Engineer

  • How to set the Xml Encoding ISO-8859-1 to Transformer or DOMSource

    I have a xml string and it already contains an xml declaration with encoding="ISO-8859-1". (In my real product, since some of the element/attribute value contains a Spanish character, I need to use this encoding instead of UTF-8.) Also, in my program, I need to add more attributes or manipulate the xml string dynamically, so I create a DOM Document object for that. And, then, I use Transformer to convert this Document to a stream.
    My problme is: Firstly, once converted through the Transformer, the xml encoding changed to default UTF-8, Secondly, I wanted to check whether the DOM Document created with the xml string maintains the Xml Encoding of ISO-8859-1 or not. So, I called Document.getXmlEncoding(), but it is throwing a runtime error - unknown method.
    Is there any way I can maintain the original Xml Encoding of ISO-8859-1 when I use either the DOMSource or Transformer? I am using JDK1.5.0-12.
    Following is my sample program you can use.
    I would appreciate any help, because so far, I cannot find any answer to this using the JDK documentation at all.
    Thanks,
    Jin Kim
    import java.io.*;
    import org.w3c.dom.Document;
    import org.w3c.dom.Element;
    import org.w3c.dom.Node;
    import org.w3c.dom.NodeList;
    import org.w3c.dom.Attr;
    import org.xml.sax.InputSource;
    import javax.xml.parsers.DocumentBuilder;
    import javax.xml.parsers.DocumentBuilderFactory;
    import javax.xml.transform.Templates;
    import javax.xml.transform.Transformer;
    import javax.xml.transform.TransformerException;
    import javax.xml.transform.TransformerFactory;
    import javax.xml.transform.TransformerConfigurationException;
    import javax.xml.transform.dom.DOMSource;
    import javax.xml.transform.Source;
    import javax.xml.transform.stream.StreamSource;
    import javax.xml.transform.stream.StreamResult;
    public class XmlEncodingTest
        StringBuffer xmlStrBuf = new StringBuffer();
        TransformerFactory tFactory = null;
        Transformer transformer = null;
        Document document = null;
        public void performTest()
            xmlStrBuf.append("<?xml version=\"1.0\" encoding=\"iso-8859-1\"?>\n")
                     .append("<TESTXML>\n")
                     .append("<ELEM ATT1=\"Yes\" />\n")
                     .append("</TESTXML>\n");
            // the encoding is set to iso-8859-1 in the xml declaration.
            System.out.println("initial xml = \n" + xmlStrBuf.toString());
            try
                //Test1: Use the transformer to ouput the xmlStrBuf.
                // This shows the xml encoding result from the transformer which will change to UTF-8
                tFactory = TransformerFactory.newInstance();
                transformer = tFactory.newTransformer();
                StreamSource ss = new StreamSource( new StringBufferInputStream( xmlStrBuf.toString()));
                System.out.println("Test1 result = ");
                transformer.transform( ss, new StreamResult(System.out));
                //Test2: Create a DOM document object for xmlStrBuf and manipulate it by adding an attribute ATT2="No"
                DocumentBuilderFactory dfactory = DocumentBuilderFactory.newInstance();
                DocumentBuilder builder = dfactory.newDocumentBuilder();
                document = builder.parse( new StringBufferInputStream( xmlStrBuf.toString()));
                // skip adding attribute since it does not affect the test result
                // Use a Transformer to output the DOM document. the encoding becomes UTF-8
                DOMSource source = new DOMSource(document);
                StreamResult result = new StreamResult(System.out);
                System.out.println("\n\nTest2 result = ");
                transformer.transform(source, result);
            catch (Exception e)
                System.out.println("<performTest> Exception caught. " + e.toString());
        public static void main( String arg[])
            XmlEncodingTest xmlTest = new XmlEncodingTest();
            xmlTest.performTest();
    }

    Thanks DrClap for your answer. With your information, I rewrote the sample program as in the following, and it works well now as I intended! About the UTF-8 and Spanish charaters, I think you are right. It looks like there can be many factors involved on this subject though - for example, the real character sets used to create an xml document, and the xml encoding information declared will matter. The special character I had a trouble was u00F3, and somehow, I found out that Sax Parser or even Document Builder parser does not like this character when encoding is set to "UTF-8" in the Xml document. My sample program below may not be a perfect example, but if you replaces ISO-8859-1 with UTF-8, and not setting the encoding property to the transfermer, you may notice that the special character in my example is broken in Test1 and Test2. In my sample, I decided to use ByteArrayInputStream instead of StringBufferInpuptStream because the documentation says StringBufferInputStream may have a problem with converting characters into bytes.
    Thanks again for your help!
    Jin Kim
    import java.io.*;
    import java.util.*;
    import org.w3c.dom.Document;
    import org.w3c.dom.Element;
    import org.w3c.dom.Node;
    import org.w3c.dom.NodeList;
    import org.w3c.dom.Attr;
    import org.xml.sax.InputSource;
    import javax.xml.parsers.DocumentBuilder;
    import javax.xml.parsers.DocumentBuilderFactory;
    import javax.xml.transform.Templates;
    import javax.xml.transform.Transformer;
    import javax.xml.transform.TransformerException;
    import javax.xml.transform.TransformerFactory;
    import javax.xml.transform.TransformerConfigurationException;
    import javax.xml.transform.dom.DOMSource;
    import javax.xml.transform.Source;
    import javax.xml.transform.stream.StreamSource;
    import javax.xml.transform.stream.StreamResult;
    * XML encoding test for Transformer
    public class XmlEncodingTest2
        StringBuffer xmlStrBuf = new StringBuffer();
        TransformerFactory tFactory = null;
        Document document = null;
        public void performTest()
            xmlStrBuf.append("<?xml version=\"1.0\" encoding=\"iso-8859-1\"?>\n")
                     .append("<TESTXML>\n")
                     .append("<ELEM ATT1=\"Resoluci�n\">\n")
                     .append("Special charatered attribute test")
                     .append("\n</ELEM>")
                     .append("\n</TESTXML>\n");
            // the encoding is set to iso-8859-1 in the xml declaration.
            System.out.println("**** Initial xml = \n" + xmlStrBuf.toString());
            try
                //TransformerFactoryImpl transformerFactory = new TransformerFactoryImpl();
                //Test1: Use the transformer to ouput the xmlStrBuf.
                tFactory = TransformerFactory.newInstance();
                Transformer transformer = tFactory.newTransformer();
                byte xmlbytes[] = xmlStrBuf.toString().getBytes("ISO-8859-1");
                StreamSource streamSource = new StreamSource( new ByteArrayInputStream( xmlbytes ));
                ByteArrayOutputStream xmlBaos = new ByteArrayOutputStream();
                Properties transProperties = transformer.getOutputProperties();
                transProperties.list( System.out); // prints out current transformer properties
                System.out.println("**** setting the transformer's encoding property to ISO-8859-1.");
                transformer.setOutputProperty("encoding", "ISO-8859-1");
                transformer.transform( streamSource, new StreamResult( xmlBaos));
                System.out.println("**** Test1 result = ");
                System.out.println(xmlBaos.toString("ISO-8859-1"));
                //Test2: Create a DOM document object for xmlStrBuf to add a new attribute ATT2="No"
                DocumentBuilderFactory dfactory = DocumentBuilderFactory.newInstance();
                DocumentBuilder builder = dfactory.newDocumentBuilder();
                document = builder.parse( new ByteArrayInputStream( xmlbytes));
                // skip adding attribute since it does not affect the test result
                // Use a Transformer to output the DOM document.
                DOMSource source = new DOMSource(document);
                xmlBaos.reset();
                transformer.transform( source, new StreamResult( xmlBaos));
                System.out.println("\n\n****Test2 result = ");
                System.out.println(xmlBaos.toString("ISO-8859-1"));
                //xmlBaos.flush();
                //xmlBaos.close();
            catch (Exception e)
                System.out.println("<performTest> Exception caught. " + e.toString());
            finally
        public static void main( String arg[])
            XmlEncodingTest2 xmlTest = new XmlEncodingTest2();
            xmlTest.performTest();
    }

  • How to set codepage iso-8859-1 in output XML?

    Dear experts,
    I want to create an XML file with codepage "iso-8859-1" from PI 7.11.
    I have already tried to put the filetype to TEXT and ISO-8859-1 in the File Encoding field of my receiving CC (using the FTP adapter), but I still get an xml in codepage UTF-8.
    It should also be possible to realise this via an XSL mapping, but where do I have to do this? Or is there any other way to use codepage ISO-8859-1?
    Thanks in advance,
    William

    Hi
    You can do this change at Visual Administrator level. There are some parameters you need to edit and add IS0-8859, so that Integration engine will set for that encode.
    So, you can ask the Basis guys to perform this.
    This would certainly works out for you.
    Regards
    Pothana

Maybe you are looking for

  • PC hard disk replaced. How do I get music from iPod back onto PC

    My pc hard disk crashed. I replaced it and reinstalled Windows XP. I have lost all the music I had on my laptop. I don't want to reload from CD's again and I don't want to loose the music I downloaded. Can I get the music from my iPod mini back onto

  • Error about CHsxServer.cpp

    When login smartview,show the error,"请求ConnectToProvider"的响应包括意外的标记属性clientExtVersion And hfmlog show the error: Error*11*hfm25*09/13/2011 10:21:46*CHsxServer.cpp*Line 2758*<?xml version="1.0"?> <EStr><Ref>{6342FD28-E9E0-43D7-AD40-B67A369529AD}</Ref>

  • The Status Column, what is it for?

    What data goes in the Status Column? Can I edit it?

  • CS11503 with AVS 3120 appliance

    Hi, I am looking at deploying a pair of CS11503 (probably in ASR redundancy mode) connected to redundant 6509 data center aggregation switches. I am also considering a pair of AVS 3120 accelerator appliances. Is it best to have each AVS 3120 directly

  • IPhone wifi grayed out?

    So I have a lifeproof iPhone case (waterproof). Apparently it leaked ever so slightly while taking under water pictures one day, causing minimal water entry to the top right (at the power button) The only side effects (after leaving the phone in rice