Default Character Encoding stuck on UTF-8 - Firefox 7

I cannot change the Character Encoding - it is stuck on Unicode UTF-8 and I can not change it! When a web page opens I get these little boxes with "FF FD" instead of Quote marks. When I change the character encoding on that page using "View->Character Encoding" and click on the Western (ISO-8859-1), the page displays correctly. Every page opens using Unicode UTF-8 as the default.
View->Character Encoding -- shows Unicode UTF-8 as the default.
View->Character Encoding->Auto-detect -- shows OFF
Tool->Options->Content->Advance->Fonts->Default Character Encoding -- shows Western (ISO-8859-1) as well as the "Allow Pages to choose their own fonts..." IS CHECKED in the check box
THE PAGES ARE NOT UTF-8!!!! The "View Page Source" IS NOT Unicode UTF-8! -- It shows <meta http-equiv="Content-Type" content="text/html; charset=iso-8859-1">.
The "View Page Info" shows MetaTag - Content-Type: text/html; charset=iso-8859-1
Why can I not change the Default Character Encoding?
I would also like to point out that the Unicode UTF-8 seems to be broken because it is indicating that the QUOTE CHARACTER is an UNPRINTABLE character "FF FD"
----- EDIT -----
The UTF-8 is not broken. The problem as pointed out in http://en.wikipedia.org/wiki/Replacement_character#Replacement_character is that my Firefox being STUCK processing UTF-8 encoding cannot read the clearly marked iso-8859-1 data. So the UTF-8 is reinterpreting smart quotes -&ldquo; and &rdquo;- (“ and ”) as replacement (unprintable) characters.
So the real problem is why my Firefox is stuck on Unicode UTF-8

The real problem is that the font that is used doesn't have those characters.
Do you see the special quotes -“ and ” on this forum page?
Does it help if you disable the website fonts and set another font as the default font?
*Tools > Options > Content : Fonts & Colors > Advanced
*http://en.wikipedia.org/wiki/Punctuation
*http://en.wikibooks.org/wiki/Unicode/Character_reference/2000-2FFF

Similar Messages

  • Setting the default character encoding

    How do you set the default character encoding for the portal to UTF-8 so that unicode
    characters work within porlets?
    Any help would be much appreciated,
    Troy

    You can put this tag in the portal.jsp in the header:
    <meta http-equiv='Content-Type' content='text/html;charset=UTF-8'>
    "Troy" <[email protected]> wrote:
    >
    That doesn't seem to work when put into my portlet's content JSP. Is
    there another
    place I could put the page directive that will control the entire portal
    page?
    "Sai S Prasad" <[email protected]> wrote:
    Troy,
    you can try the page directive with encoding as:
    <%@ page contentType="text/html";charset="UTF-8" %>
    "Troy" <[email protected]> wrote:
    How do you set the default character encoding for the portal to UTF-8
    so that unicode
    characters work within porlets?
    Any help would be much appreciated,
    Troy

  • Setting DEFAULT character encoding ???

    To set the default character encoding to UTF-8 for JVM I can do the
    following on Solaris:
    LC_ALL=en_US.UTF-8; java ....
    How do I achieve the same on Windows?
    Is there a java property where I can specify the default character
    encoding or do I have to getBytes( str, "UTF-8") everywhere?
    Thanks,
    Artur...

    Hi Artur,
    there is a way. The property is file.encoding. Example:
    import java.io.*;
    public class ShowEncoding {
      public static void main(String[] args) {
        System.out.println("Default encoding: " +
            new InputStreamReader(System.in).getEncoding());
    }On my system I get the following results:
    java ShowEncoding Default encoding: Cp1252
    java -Dfile.encoding=UTF-8 ShowEncoding Default encoding: UTF8
    java -Dfile.encoding=Latin1 ShowEncoding Default encoding: ISO8859_1I don't know, if setting this property using the -D argument is documented or implementation-independendent. But so far, it was working for me. For setting the locale, you may use the properties user.language and user.region. But this is undocumented and implementation-dependendent.
    I wouldn't use these properties at all. I would rather specify the encoding in all constructors. Probably using an application property file.
    Good luck!
    Marcus.

  • Default character encoding ???

    To set the default character encoding to UTF-8 for JVM I can do the
    following on Solaris:
    LC_ALL=en_US.UTF-8; java ....
    How do I achieve the same on Windows?
    Is there a java property where I can specify the default character
    encoding or do I have to getBytes( str, "UTF-8") everywhere?
    Thanks,
    Artur...

    Hi Artur,
    please see my responses to the same question you've posted in
    Forum Home > Internationalization
    Kind regards,
    Marcus

  • How to set a platform's default character encoding

    Hi,
    Does anybody know how to set how to set a platform's default character encoding in Java? Thank you.
    Yugang
    [email protected]

    You do 0mean for Java, not from Java? (The latter would make absolutely no sense at all.) If so, pass it to the runtime using the -D switch (you've got SUN's java, right?):
    java -Dfile.encoding=the-encoding-i-like the.name.of.YourClass

  • Character encoding (unicode to utf-8) conversion problem

    I have run into a problem that I can't seem to find a solution to.
    my users are copying and pasting from MS-Word. My DB is Oracle with its encoding set to "UTF-8".
    Using Oracle's thin driver it automatically converts to the DB's default character set.
    When Java tries to encode Unicode to UTF-8 and it runs into an unknown character (typically a character that is in the High Ascii range) it substitutes it with '?' or some other wierd character.
    How do I prevent this.

    my users are copying and pasting from MS-Word. My DB
    is Oracle with its encoding set to "UTF-8".Pasting where? Into the database? If they are pasting into the database (however they might do that) and getting bad results then that's nothing to do with Java.
    Using Oracle's thin driver it automatically converts
    to the DB's default character set.Okay, I will assume that is correct.
    When Java tries to encode Unicode to UTF-8 and it
    runs into an unknown character (typically a character
    that is in the High Ascii range) it substitutes it
    with '?' or some other wierd character.This is false. When converting from Unicode to UTF-8 there are no "unknown characters". I don't know what you mean by the "High Ascii range" but if your users are pasting MS stuff into your Java program somehow, then a conversion from something into Unicode is done at that time. If "something" isn't the right encoding then you have the problems already, before you try to write to the DB.
    How do I prevent this.First identify the problem. You have input coming from somewhere, then you are writing to the database. Two different steps. Either of them could have a problem. Test them separately so you know which one of them is the problem.

  • SPA504G SPA514G Default Character Encoding stay in ISO-8859-1

    Hi,
    I have configure like:
      <Dictionary_Server_Script ua="na">serv=http://{{ provisioning.server }}/telecom/language/;d0=English;x0=spa50x_30x_en_v754.xml;d1=French;x1=spa50x_30x_fr_v754.xml;</Dictionary_Server_Script>
      <Language_Selection ua="na">French</Language_Selection>
      <Default_Character_Encoding ua="na">UTF-8</Default_Character_Encoding>
      <Locale ua="na">fr-FR</Locale>
    Dictionary and Provisioning Profile are encoded in UTF-8.
    but when the phone start after provisioning the Default_Character_Encoding set to ISO-8859-1
    and the lines labels are misprinted.
    Ligne 1
    Ligne 2
    Olivier
    Françoise
    instead of
    Ligne 1
    Ligne 2
    Olivier
    Françoise
    Any idea ?

    I got an answer from the developer.
    Pasted here.
    I think the default encoding is set back to ISO8859 after customer download the dictionary.
    Here is the reason: after 7.5.3, SPA 50x will parse the trkLocaleName in dictionary, for French it will set the phone’s default encoding to iso8859-1 since it is good to French.
    French
    =================================
    •1.         If the customer want to use UTF-8 after xml downloading, please modify the trkLocaleName in the French dictionary xml as following:
    croatian
    It is a workaround, but it's strange why French user will use UTF8. Thanks.
    •2.         Another way is that user can manually set the default encoding value to UTF-8 after xml downloading.

  • Character encoding of Unicode (UTF-8) is what seems to be the default for printing. The page looks funny and the print is spread out.

    the layout of the page is different that it appears on the screen. I printed the confirmation of my VISA payment on line and it is not concise. The information is correct, but the layout isn't.
    Should I use another encoding? One of the Western ones? Maybe unicode (UTF-16)?

    There is a known bug involving printing in beta 12 that has been fixed in the Firefox 4 release candidate which is due out soon.

  • How can I define a default character enconding for Firefox 4?

    Dear Sirs,
    I would like to know how can I change the character encoding of Firefox 4 to Western (ISO-8859-1) and keep that character enconding as the default every time I open Firefox 4.
    I have noticed the every time I change the default character encoding from Unicode (UTF-8) to Western(ISO8859-1) and close Firefox 4, when I open Firefox 4 the character encoding returns to the Unicode (UTF-8) character encoding.
    I would like to keep the Western (ISO8859-1) as the default character encoding. How can I do that?
    Best regards,
    IMeMine

    If the server sends an encoding then Firefox will always switch to that encoding.<br />
    The default is only used on pages where the server does not send en encoding via the HTTP response header and the code doesn't specify the encoding via a meta tag.
    There is usually no need to change the encoding from the default setting Western (ISO-8859-1).
    It is probably better to try an Auto-Detect setting (e.g. View > Character Encoding > Auto-Detect > Universal) if you have a problem on specific websites.

  • Can character encoding be predefined for certain pages?

    Certain pages that I visit frequently require me to manually set my character encoding to Western (ISO Latin 1), both when my default character encoding is set as UTF-8 and Western (ISO Latin 1).
    As the pages that show up malformed are embedded in other frames I suspect that the top frame forces a different encoding than is on the embedded page.
    An example page is here (847.is). The topic list of this message board is in order, but when any one of the topics is viewed all accented and special characters are missing, until Western (ISO Latin 1) is manually set as the character encoding. Similarly, opening any of the topics in a tab will result in missing characters.
    Is there some way for me to circumvent having to go through all those menus to set it? Can I somehow define that these pages should be viewed in Western (ISO Latin 1) or can I set a keyboard shortcut for Western (ISO Latin 1)?
    MacBook 2006   Mac OS X (10.4.7)   Safari version 2.0.4
    MacBook 2006   Mac OS X (10.4.7)  

    For instance if I opened
    an entry on Vísindavefur
    HÍ out
    of the parent frame the accented letters would
    show up somehow mangled, but this is no longer a
    problem.
    That page has no charset in the source and thus should only display correctly if you have Latin-1 set as the browser default. With UTF-8 set as the default you should see (in Safari) a ton of black diamonds with question marks inside.
    Firefox never displays ð (eth), þ (thorn) and ý
    (accented y) correctly for me, did not do it on the
    old machine and does not do it on this one either,
    FireFox displays them perfectly for me in both 10.3 and 10.4.
    Actually, if I set Opera encoding to UTF-8 it
    displays the topics on 847.is as Safari does.
    This indicates it is a system issue rather than Safari. Sorry I can't duplicate it and have no good idea what could cause it on a normal system. Have you (or the place where you buy your machines) by chance installed any special software add-ons to enable the use of non-Unicode Icelandic (for apps like Appleworks, WordX, etc)?

  • Character Encoding is changing random

    Hello,
    For a short while I'm having the next problem:
    When I am using FireFox, after a while characters on pages are showed as 'boxes'. With setting backView -> Character Encoding to ISO-8859-1 the characters are shown correctly again.
    I do not understand why the character encoding is changing random to UTF-8, while I have selected the ISO-8859-1 and set automatically recognize to false.
    In the options menu I have also set default 'ISO-8859-1' as default character encoding.
    Hoping someone can tell me why it random changes and why the same page can be show correctly for 10 times, but the 11th time the character set has changed? And of course how can I solve this problem.

    I do know a website can determine the character encoding, but the strange part of the problem I occur is that the same page can be shown 10 times correctly, but an 11th time the characters like é and ë, are shown as 'blocks'/'questionmarks'.
    Is there an explanation for that behaviour?

  • Importance of character encoding?

    I am making a list of errors in someone else's code. What about this:
    outline of existing code
    byte[] binaryMssg = "sync/ack".getBytes(); // client
    --- DataOutputStreamtream --->
    String mssg = new String(byte[] buf); // server
    proposed correction:
    byte[] binaryMssg = "sync/ack".getBytes(Charset.forName("UTF-16")); // client
    --- DataOutputStream --->
    String mssg = new String(byte[] buf, Charset.forName("UTF-16")); // server
    The original developer omitted character encoding.
    I have zero knowledge/experience with character encoding. I only just noticed method overloads that optionally use them. I don't want to look stupid by putting what is really a non-issue on my list. I don't want to look like a jerk and draw attention to something that, in real world usage, is a non-issue. Otherwise, just to be thorough, I would like to put it on my list. I think it might create more robust code? Any comments welcomed.
    Finally, I was so surprised to notice Charset is in the nio package and so not available until v1.4. This hints character encodings where not a prioriry for the Java development team. So, since they did not consider it important, perhaps neither should I. While there may be niches for character encoding usage, the normal-case rule is to not use explicit character encoding.

    dpxqb wrote:
    I am making a list of errors in someone else's code. What about this:
    outline of existing code
    byte[] binaryMssg = "sync/ack".getBytes(); // client
    --- DataOutputStreamtream --->
    String mssg = new String(byte[] buf); // server
    proposed correction:
    byte[] binaryMssg = "sync/ack".getBytes(Charset.forName("UTF-16")); // client
    --- DataOutputStream --->
    String mssg = new String(byte[] buf, Charset.forName("UTF-16")); // serverMuch much much better but I would use utf-8 since it is usually more compact for western languages.
    >
    The original developer omitted character encoding.Criminal.
    >
    I have zero knowledge/experience with character encoding. I only just noticed method overloads that optionally use them. I don't want to look stupid by putting what is really a non-issue on my list. I don't want to look like a jerk and draw attention to something that, in real world usage, is a non-issue. Otherwise, just to be thorough, I would like to put it on my list. I think it might create more robust code? Any comments welcomed.It is correct to define the character encoding to be used if the bytes are being transferred from one computer to another since the two machines may have a different default character encoding.
    >
    Finally, I was so surprised to notice Charset is in the nio package and so not available until v1.4. This hints character encodings where not a prioriry for the Java development team. So, since they did not consider it important, perhaps neither should I. While there may be niches for character encoding usage, the normal-case rule is to not use explicit character encoding.Character encoding has always been important in Java. One has always been able to explicitly set the character encoding by using a String to define it. For example one could use
    byte[] binaryMssg = "sync/ack".getBytes("utf-16"); // client
    --- DataOutputStream --->
    String mssg = new String(byte[] buf, "UTF-16"); // server

  • What's the difference of character encoding between 1.4.0and1.4.2 in Linux

    As i find, the character encoding about chinese in jdk1.4.2 no langer the same of jdk1.4.0.
    In jdk1.4.0, the character encoding used the "file.encoding" system property, we often set the
    property with "gb2312".
    But in jdk1.4.2, i find that the default character encoding no longer used the "file.encoding" system property.
    Who knows the reason?
    Test Program:
    public class B{
    public static void main(String args[]) throws Exception{
    byte [] bytes = new byte[]{(byte)0xD6,(byte)0xD0,(byte)0xCE,(byte)0xC4};
    String s1 = new String(bytes);
    String s2 = new String(bytes,System.getProperty("file.encoding"));
    System.out.println("s1="+s1+" , s2="+s2);
    System.out.println("s1.length=" + s1.length() + " , s2.length="+s2.length());
    run four times and the result list:
    [root@app15 component]# /usr/local/j2sdk1.4.0/bin/java -Dfile.encoding=ISO-8859-1 -cp . B
    s1=&#20013;&#25991; , s2=&#20013;&#25991;
    s1.length=4 , s2.length=4
    [root@app15 component]# /usr/local/j2sdk1.4.0/bin/java -Dfile.encoding=gb2312 -cp . B
    s1=&#20013;&#25991; , s2=&#20013;&#25991;
    s1.length=2 , s2.length=2
    [root@app15 component]# /usr/local/j2sdk1.4.2/bin/java -Dfile.encoding=ISO-8859-1 -cp . B
    s1=&#20013;&#25991; , s2=&#20013;&#25991;
    s1.length=4 , s2.length=4
    [root@app15 component]# /usr/local/j2sdk1.4.2/bin/java -Dfile.encoding=gb2312 -cp . B
    s1=&#20013;&#25991; , s2=??
    s1.length=4 , s2.length=2
    [root@app15 component]#

    I don't know for sure, but:
    -- The API documentation for String says that "new String(byte[])" uses "the platform's default charset".
    -- The API documentation for Charset says "The default charset is determined during virtual-machine startup and typically depends upon the locale and charset being used by the underlying operating system."
    You'll notice that it doesn't say anything about using the file.encoding system value, so presumably (based on your experiments) it doesn't. I did a search for "java default charset" and didn't find anything specific, but this site says "As of Java 1.4.1, the default Charset varies from platform to platform" and suggests you explicitly hard-code your charset. I would agree with that.

  • Problems with Forms and character encoding

    I'm having problems trying to read unicode data inputted into a Form on my JSP page.
    I've used the meta tag <meta http-equiv="Content-Type" content="text/html; charset=utf-8"/> to set the charset of the page to UTF-8. I've inputted some chinese characters inot my form and when I try to read the subsequent request parameter in my servlet using request.getParameter() the string returned is this
    "&#26469;&#28304;" which is the escape sequence required by HTML to display these characters.
    From what I've read on the subject this doesn't seem like the expected value. I've tried other ways of getting the correct string value such as setting the character encoding request.setCharacterEncoding("UTF-8") and then converting the bytes using this encoding value but it doesn't seem to work.
    I could write a method to split up the string using the ; as a token and working out the correct unicode character but this doesn't seem like the right thing to do.
    Any help on how to pass the correct information from the Form in the JSP page to the servlet would be greatly appreciated

    I don't believe that is correct, but if it's returning HTML escapes instead of URL Encoded characters, then it's the browser doing it. This is my test page for playing with Chinese...
    <%@ page language="java" contentType="text/html; charset=UTF-8" %>
    <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
    <html>
    <head>
         <title></title>
         <meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
    </head>
    <body bgcolor="#ffffff" background="" text="#000000" link="#ff0000" vlink="#800000" alink="#ff00ff">
    <%
    request.setCharacterEncoding("UTF-8");
    String str = "\u7528\u6237\u540d";
    String name = request.getParameter("name");
    %>
    req enc: <%= request.getCharacterEncoding() %><br />
    rsp enc: <%= response.getCharacterEncoding() %><br />
    str: <%= str %><br />
    name: <%= name %><br />
    <form method="GET" action="_lang.jsp" encoding="UTF-8">
    Name: <input type="text" name="name" value="" >
    <input type="submit" name="submit" value="GET Submit" />
    </form>
    <form method="POST" action="_lang.jsp" encoding="UTF-8">
    Name: <input type="text" name="name" value="" >
    <input type="submit" name="submit" value="POST Submit" />
    </form>
    </body>
    </html>

  • Changing the default char encoding of the current JVM ?!

    Is there any way that could be used to alter the default character encoding of the Java Virtual Machine at the start of a Java application?

    This seems a little dangerous...considering that file i/o etc depend on correct charset encodings for filenames, etc.
    However, perhaps you could try setting the file.encoding property on the command line:
    java -Dfile.encoding=Big5 YourApplication
    Regards,
    John O'Conner

Maybe you are looking for

  • Problems importing Hi8 to a PC for editing / archiving

    Hey Video Experts, I'm trying to import Hi8 video from my cam-corder through a PCI-E TV Tuner card (s-video) installed in my PC. In PElements, I'm importing the video by "getting media," using the "Webcam or WDM device" option.    The video does impo

  • Keynote reverts back to 3.0.2

    Updated my computer today and - all of the sudden all of my keynote files are now associated with 3.0.2 and want me to get iwork 06. I've got iwork 08 and never had version 3.0.2. What happened? I've already read "Topic : Keynote 3.03 and 4.01" and i

  • Duplicating large size  (7 MB) photos in IPhoto 08

    I am using IPhoto 08 with Leopard OS. I have a new camera and when I download 7MB photos to IPhoto and duplicate the photo, IPhoto reduces the size of the duplicated photo to about half (3.9 MB in this case). When I download and dublicate photos of 4

  • Removing unwanted control characters in exported text files

    I am currently evaluating Crystal Reports 2008 to determine applicability to our requirements. I need to export data files to continuous text to be read by other application software. I have successfully created the files but have what I believe to b

  • HT5318 How to get past security questions when you cannot remember the exact answers needed.?

    How do you get new security questions and remove old security questions and get the system to recognize the new questions?  I have changed the questions but everytime the security questions appear, they are the old ones that I do not remember the exa