Picture encoding issues in IE

I've created a website with iWeb and published it to my .mac account. The only issue I've run into is for people viewing it in IE. Everything views ok except with the blog. When you click an individual blog, the picture that accompanies that specific blog doesn't come up but instead a large gray square. (The picture is visible on the blog frontpage, just not on the individual entry itself). However, if you double click on that and change the encoding to Western or Auto-detect, then the picture appears. All other pictures though the website render fine.
I know the best option would be to have IE users switch to Firefox but that is not an option with a few people (older who don't like change!)
Has anybody else experienced this? And is there a fix? Either in iWeb or in IE so the user doesn't have to change the encoding setting for each blog picture. . .?
Thanks.
Kevin

fixed it myself - please ignore

Similar Messages

  • Premiere and Media Encoder CC encoding issue

    Hi all,
    I am having an encoding issue with PP and ME CC. My video assets are fine, and on the timeline they appear how they should, but when I look at the rendered h264 video there are encoding errors in the video. I have attached two images, the black is how it should look and the white is the error. The video plays fine and then it flickers between the images shown. 
    It has done this on a few different videos I have rendered over the last few days and I don't know why. It also happens to a different machine on CC as well. Does anyone have any suggestions?

    Hi James,
    I've never seen this before. Can you give us more info? Answer all the questions on this FAQ: What information should I provide when asking a question on this forum?
    Thanks,
    Kevin

  • Encoding Issue : JMS and Mapping : utf-8 iso8859-1

    Hi All,
    I am facing some problem with encoding issue.
    Scenario :  JMS -->  SAP PI --> JMS
    Requirment : Input plain text file contain some special characters,"©®" . Based on this condition,In Java Mapping
                       we check the Payload and changed the 'encoding' tag to UTF-8 or   ISO8859-1.                                                     
                   : <?xml version="1.0" encoding="UTF-8"?>     in the target XML output.
    While testing in Operation mapping our Java mapping works fine. as the encodeing tag changes from
                 UTF-8 to ISO8859-1 if the special character exists.But if I test the same in Integration Directory(Test Configuration)
                 or did a end to end  testing. The encoding tag did'nt changes.
    For testing we had to a set of Plain Text files with UTF-8 and ISO8859-1 .
    I tried the options of using beans in Adapter modules in Sender JMS channel.
    MessageTransformBean, TextCodepageConversionBean, XmlAnonymizerBean
    These doc & threads ,was also referred[How to Handle Encoding in PI|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/502991a2-45d9-2910-d99f-8aba5d79fb42]
    Regards,
    Ashutosh R

    Hi
    public static boolean fixSpecialCharforWeb(String text) {
            int i = 0;
            Character c = null;
            char[] ctext = null;
            StringBuffer newText = new StringBuffer("");
            //boolean encodingType = false;
            if ((text == null) || (text.trim().length() == 0)) {
                return encodingType;
            } else {
                try {
                                   for (i = 0; i < text.trim().length(); i++) {
                        ctext = text.trim().substring(i, i + 1).toCharArray();
                        c = new Character(ctext[0]);
                        //Single quote
                        if ((text.trim().substring(i, i + 1).equals("'")) || (c.hashCode() == 8217) || (text.trim().substring(i, i + 1).equals("?")) || (c.hashCode() == 146) || (c.hashCode() == 145)) {
                            //newText.append("'");
                            encodingType = true;
                            return encodingType;
                        //Double quotes
                        if ((c.hashCode() == 8220) || (c.hashCode() == 8221) || (c.hashCode() == 147) || (c.hashCode() == 148)) {
                            //newText.append(""");
                            encodingType = true;
                            return encodingType;
                        // bullet point
                        if ((c.hashCode() == 8226) || (c.hashCode() == 149)){
                            encodingType = true;
                            return encodingType;
                        // tilde
                        if ((c.hashCode() == 732) || (c.hashCode() == 152)){
                            encodingType = true;
                            return encodingType;
                        // Soft Hypen
                        if (c.hashCode() == 173){
                            encodingType = true;
                            return encodingType;
                        // En-Dash
                        if ((c.hashCode() == 8211) || (c.hashCode() == 150)) {
                            encodingType = true;
                            return encodingType;
                        // Em-Dash
                        if ((c.hashCode() == 8212) || (c.hashCode() == 151)) {
                            encodingType = true;
                            return encodingType;
                        // Euro Sign
                        if ((c.hashCode() == 8364) || (c.hashCode() == 128)) {
                            encodingType = true;
                            return encodingType;
                        // Yen Sign
                        if (c.hashCode() == 165) {
                            encodingType = true;
                            return encodingType;
                        // Pound Sign
                        if (c.hashCode() == 163) {
                            encodingType = true;
                            return encodingType;
                        // 1/2 sign
                        if (c.hashCode() == 189) {
                            encodingType = true;
                            return encodingType;
                        // 1/4 sign
                        if (c.hashCode() == 188) {
                            encodingType = true;
                            return encodingType;
                        // 3/4 sign
                        if (c.hashCode() == 190) {
                            encodingType = true;
                            return encodingType;
                        // Sword/dagger
                        if ((c.hashCode() == 8224) || (c.hashCode() == 134)) {
                            encodingType = true;
                            return encodingType;
                        // Trademark
                        if ((c.hashCode() == 8482) || (c.hashCode() == 153)) {
                            encodingType = true;
                            return encodingType;
                        // Ampersand &
                        if ((text.trim().substring(i, i+1).equals("&")) || (c.hashCode() == 38)) {
                            encodingType = true;
                            return encodingType;
                        //Registered mark
                        if ((text.trim().substring(i, i + 1).equals("?")) || (c.hashCode() == 174)) {
                            //newText.append("®");
                            encodingType = true;
                            return encodingType;
                        //Copyright mark
                        if ((text.trim().substring(i, i + 1).equals("?")) || (c.hashCode() == 169)) {
                            encodingType = true;
                            return encodingType;
                        // Question.
                        if (c.hashCode() == 63 && c.toString().equals("?")){
                            //newText.append("?");
                            encodingType = true;
                            return encodingType;
                        //handling symbol ?
                        if ((text.trim().substring(i, i+1).equals("?")) || (c.hashCode() == 233)) {
                            encodingType = true;
                            return encodingType;
                        if ((text.trim().substring(i, i+1).equals("?")) || (c.hashCode() == 232)) {
                            encodingType = true;
                            return encodingType;
                        if (c.hashCode() == 144) {
                            encodingType = true;
                            return encodingType;
                } catch (Exception e) {
                    e.printStackTrace();
            return encodingType;

  • Embedding HTML in XML CDATA and encoding issues

    Hi all,
    I'm embedding HTML code in a CDATA section. My problem is that, depending on the document, the HTML can be encoded in many formats. I borrowed a piece of code that sniffs that format so i can create String in the "right" encoding (or at least the one that was guessed).
    - If I directly injected those in the CDATA section, i guess they'd be encoded in UTF-8 and some character would be misinterpreted?
    - What if i would transcode the HTML from the sniffed format to utf-8?
    -Are there any issues woth doing this?
    Sorry if this is a dumb question but I'm quite new to that kind of encoding issues.
    BTW i'm using DOM.
    Thanks
    lexo

    I don't know if it's a dumb question. I just don't understand it at all. Encoding issues only arise when you write data from a Java program to an external location, or when you read data from an external location into a Java program. And none of the activities you mentioned there have anything to do with that.
    When you write your XML to an external file, or wherever you write it to, it gets encoded at that moment. The whole thing. Elements, attributes, CDATA sections, the whole thing. Doesn't matter what's in it, the whole thing gets encoded in whatever charset was chosen.
    Does that help?

  • Encoding issue in email send from SAP

    Hi,
    I am currently having encoding issue in email send by SRM .
    Can some one help me ?
    My problem :
    When i key in my SC subject/desc through IE7 with encoding unicode (UHT-8). Everything look fine.
    After i order the SC, Email will be sending out to the approver.
    The email received by approver is including strange character.
    Example desciption i putting on SC : Testing Encoding from IE7
    Result : Email which received by approver : Testing Encoding from IE7Â.
    Can some one advice me on this?
    Thanks,
    Vanece

    Hello Vanece,
    >
    Vanece Hee wrote:
    > (...)
    > They look fine in BBP_PD as well.
    >
    > when i trace them in SOST, the word will be look weird.
    > (...)
    Problem could come from internal and external data format.
    Depending on your SRM release, check OSS notes for report you used to generate approval mail ( RSWUWFMLEC, RBBP_NOTIFICATION_APP, etc.).
    >
    Vanece Hee wrote:
    > (...)
    > When i change the encoding setting in outlook to unicode for the email, it work for some email only.
    > (...)
    Ask to your Outlook team to have a look at Outlook settings.
    You can also check OSS note 1319517 - Unicode Collection Note.
    Regards.
    Laurent.

  • Z10 encoding issue in outgoing email

    Hoping someone can help me. I've been an avid bb fan for years and my husband and I both handed in our bonds a week ago to move to the z10. We are both loving it. While my husband has had no issues with his I am very close to having no issues - except for one major one that affects me and my clients greatly. I have 2 email accounts - gmail (working without any issue at all - was up and running within a matter of minutes of receiving my z10). The other however has not been a fun or even manageable experience. My work email is web based and took a few days for me to even get it up and running. With my bold I was able to connect via Exchange but am no longer able to do so. I was told to connect via Active sync however my employer does not have this functionality and it is not supported. Following that I was told to connect using either IMAP or POP. And while I have been finally able to connect and receive my email when I SEND email it is a disaster!!!!! My out going mail seems to be adding in all of the encoding and so it becomes "unreadable" for lack of a better term. I have no idea where to turn on this. I have been constantly surfing and searching and and all forums that I can in hopes to find a resolution but have not yet found one - this is my last resort - otherwise sadly ....... Very, very sadly I will have to return my phone. Any suggestions at all would be much appreciated! Sorry for the "long-winded ness"

    lesedwards wrote:
    I agree. I run my own business in fact two business and I cannot be sending emails out from some half **bleep** address how professional does that look. Everything has worked perfect for 10 years and then I bought this piece of **bleep** called a Z10 and it even worked on it until the OS update.
    Wow lesedwards.. you are on fire. Jacking your issue into every email thread and getting **bleep**ed constantly.
    This thread was about an encoding issue, and from your other posts you are simply having connection issues.
    @Sarealex - I haven't seen this issue with any of my email accounts over IMAP, what software version are you currently using?
    -: DrewRead :-
    Can't find an app on Blackberry World yet? Tweet the developer and ask when they will #GetWithBB10 !!
    "Like it" if you like it.

  • [SOLVED] File name encoding issue

    Hi all,
    I have a large series of files with accented characters, they were all displayed nicely, but at some point, when I copied them to another computer, the characters were replaced by codes, for instance: "ó" --> "ó".
    +Renaming ie. "Pasó" (bad encoding of "Pasó") --> Pasó, while writing it, it shows the correct character, but when pressing enter the name remains ("Pasó")
    +If I rename the file to something else and then to the correct name, it will accept it: Pasó --> Pas --> Pasó will display correctly.
    I don't know if it's a system wide encoding issue because new files are displayed correctly, but I would like to know if I have to change file names manually to make them right.
    PS. When copying bad encoded files to another FS (like a USB drive), nautilus and bash refuse to copy them.
    Last edited by Wasser (2012-09-17 21:10:52)

    My fstab:
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    tmpfs /tmp tmpfs nodev,nosuid 0 0
    # /dev/sda2 LABEL=ROOT
    UUID=d2243d9c-b8e7-442a-8446-5a43a4d9221b / ext4 rw,relatime,data=ordered 0 1
    # /dev/sda5 LABEL=HOME
    UUID=e67f5cfa-3ec3-4c06-9c2c-62c4cc188ffe /home ext4 rw,relatime,data=ordered 0 2
    # /dev/sda3 LABEL=VAR
    UUID=caac4924-2a13-4c97-9926-668ac0595ba3 /var reiserfs rw,relatime 0 2
    # /dev/sda1 LABEL=UEFI
    UUID=1E70-6485 /boot/efi vfat rw,relatime,fmask=0022,dmask=0022,iocharset=iso8859-1,shortname=mixed,errors=remount-ro 0 2
    # /dev/sda4
    UUID=14993c2e-4bc4-42e4-b2e5-9dbc286abb4c none swap defaults 0 0
    Files in question are in /dev/sda5 (HOME)
    Last edited by Wasser (2012-09-16 08:37:52)

  • Character encoding issue in sql server

    Hi Team,
    We have a table with more than 20 columns...In that we have some columns which will have data extracted from the Datawareshouse applicances and its one time load.
    The problem is we have two columns which will may have same set of values from some records and for different set of records
    the below values are the example for same set of records but the values are changed while importing into the sql server database..
    2pk Etiquetas Navide‰as 3000-HG                                
     2pk Etiquetas Navideñas 3000-H                           
    Is there anyway to change the first column values into the second column value ? 
    By looking at the data we can say its the code page issue..(Character encoding issue)..but how to convert it?
    Convertting(2pk Etiquetas Navide‰as 3000-HG)  
    to get   2pk Etiquetas Navideñas 3000-H   in the select query?

    Then it seems that you can do the obvious: replace it.
    DECLARE @Sample TABLE ( Payload NVARCHAR(255) );
    INSERT INTO @Sample
    VALUES ( N'2pk Etiquetas Navide‰as 3000-HG' );
    UPDATE @Sample
    SET Payload = REPLACE(Payload, N'‰', N'ñ');
    SELECT S.Payload
    FROM @Sample S;

  • Java to Excel encoding issue.

    I'm trying to export my data in to Excel file. When I open the excel file the Japanese characters looks garbled.
    iResponse.setHeader("attachment ; filename = \"" + reportFileName() + "\" ", "content-disposition");
    iResponse.setHeader("application/vnd.ms-excel;","content-type");
    iResponse.setHeader("UTF-8","Content-Encoding");
    By Googling I learnt that Excel doesn't like UTF-8 format. Is there any other way I can export the data.
    Thanks.

    Specifying the content encoding for a binary format like Excel is pointless. You are producing native Excel format, aren't you? If you're producing a CSV file and claiming it's Excel, you should have mentioned that.
    Anyway: you're producing native Excel using something like Apache POI, I suppose? I would have thought it handled encoding issues, but check its documentation or its FAQ to see if that's really the case.
    Another option, which may or may not be feasible, is to use the new Office 2007 format. It's based on XML, so all of the encoding issues are automatically handled.

  • Encoding issue for file manager

    I am using the ditto command to duplicate a file. This file has unicode filename and as per http://developer.apple.com/qa/qa2001/qa1173.html I am first normalizing the name to kCFStringNormalizationFormD and then converting it to utf-8 before calling ditto on it. This all works smoothly but when I try to get the FSRef using the original unicode name I get fnfErr. Dosn't the API CFURLGetFSRef convert the string to kCFStringNormalizationFormD? Or is there any alternate for ditto on Tiger.

    no encoding issues if i use xml (xlf or xliff) bundle as xml supports utf-8 encoding.

  • Character encoding issue

    I'm using the below give code to send mail in Trukish language.
    MimeMessage msg = new MimeMessage(session);
    msg.setText(message, "utf-8", "html");
    msg.setFrom(new InternetAddress(from));
    Transport.send(msg);
    But my customer says that he gets sometime unreadable characters in mail. I'm not able to understand how to solve this character encoding issue.
    Should i ask him to change his mail client's character encoding settings?
    If yes which one he should set.

    Send the same characters using a different mailer (e.g., Thunderbird or Outlook).
    If they're received correctly, come the message from that mailer with the message
    from JavaMail. Most likely other mailers are using a Turkish-specific charset instead
    of UTF-8.

  • Media Encoder issue or bug? 720p50 to 720i50

    Can anyone confirm a Media Encoder issue, happening when converting progressive 50-fullframe (or 60 for ntsc) to 50-interlaced frames?
    I tried to convert DVCProHD 720p50 clips to PAL DVD wich uses 576i50 (25 frames devided to 50 interlaced frames). The result was, that every second frame of the source clip was skiped, so I received a PAL clip in 25 full-frames (25p). And it didn't matter which field-order I choosed (upper/lower/progressive) - the result was always the same.
    when converting p50 to i50 I would expect, that AME would convert these 50 full-frames to 50 interlaced-frames with the field order I choosed. For example Canopus Procoder is working this way.
    Btw, I lost the link, where to report Premiere bugs to adobe. Can anyone post it, please?

    See:
    FAQ:How/Where do I report a bug?
    Cheers
    Eddie
    PremiereProPedia   (
    RSS feed)
    - Over 300 frequently answered questions
    - Over 250 free tutorials
    - Maintained by editors like
    you
    Forum FAQ

  • Encoding issue for some languages

    hi,
    i am building multi-lingual adf application.
    i am using resource bundle (in properties file) to store all translation.
    i am running into a strange issue, on 1 of the server (my local one running on windows), languages appear properly.
    on another (dev box running on unix), some languages giving encoding issues.
    where to check? any suggestions?
    thanks.

    no encoding issues if i use xml (xlf or xliff) bundle as xml supports utf-8 encoding.

  • FTP Server encoding issues

    I have an encoding issues with FTP Server of OS X ML
    1. When users connected from NT machine (Windows 2008 R2, Windows 7 etc) to OS X ftp server
    user cannot see cyrilic names, it looks like - Александр Амелин
    2. When user trying to create file or folder with cyrilic name system return an error "While creating Новая папка on FTP server occured error..." in meantime on Server can observe that file actualy creater BUT with strange name - %CD%EE%E2%E0%EF%E0%EF%EA%E0
    Is there way set charset encoding for OS X FTP Server? or any other way to fix this from Server side?

    auto complete is bash or whatever shell you are running.
    ls is not bash. Different programs treating data differently.
    So Vim and nano could be doing the same wrong thing ls is doing.
    Each individual program is its own world and gets to mess up that world the way it wants

  • Creative Suite 5 Premier Pro Encode Issues

    Hello Fellow Experts,
    I am having a problem with my Adobe Creative Suite 5 Premier Pro software. I spent a long time working on a movie and when It tries to Encode in Adobe Media Encoder, it wants to take 10 hours for 2 minutes. The file that I'm trying to encode is not complicated at all. It only consists of 20 pictures (dimensions less then 2500x2000) and one sound track.
    There are transitions and a few effects in the sequence, but not enough to make it take 10-12 hours for just 2 minutes!
    I've already given my 32 bit laptop more virtual memory, and that fixed a low memory error, but it is still wanting to take too long.
    The original file is over 40 minutes long so it would really help if someone had an idea for me. Does anyone know what the problem might be?
    I've encoded 3 minute clips before and it only took 4 hours...
    I don't have access to a better computer either.
    Any ideas?

    Nobody can know. Your info is at best rather broad and generic, at worst vague. In any case, your computer seems seriously underpowered and from the program possibly being too busy with decoding the full 40 minutes first to issues with disk based caches on slow drives to specific CoDec and hardware acceleration issues I could imagine a million things going on. Just because you only use a handful of effects doesn't mean that those otehr things couldn't drag your system down. Extensive file I/O can be a real killer on any project. Well, whatever. What you see is probably perfectly normal within what you can expect on your system. Some housekeeping like defragemnting disks and removing old temp files might improve matters slightly, but I realyl think short of buying a better computer there is not much you can do...
    Mylenium

Maybe you are looking for

  • Xperia Facebook integration doesn't Work

    Hello, When I want to Connect my Xperia Z1 to my facebook account using Xperia Facebook integration a black Screen with loading bars appears and it stays that way indefinetely. I tried cleaning data for Xperia Facebook integration. I uninstalled face

  • From .mov to .avi or .mkv

    Hello together, i want to use Quicktime for AVC/H.264 encoding. The mov container don't support AC3 sound. How can i convert mov to avi- or matroska- (mkv) container format, without new encoding; also only a container-change ? Thank you and best rega

  • Rendering in Sound Bumps

    We've had a weird problem on the last two projects, which is that audio recorded through the voiceover tool develops sound bumps or clicks after being rendered. The files play normally in the viewer with no problems whatsoever. But when we put them i

  • What is the gray horizontal line under the times in the timeline?

    What is the gray horizontal line right under the times across the top of the timeline? It has a little orange beginning and ending tab.

  • Text preview size in PS 2014?

    Hi, Installed PS CC 2014. Now my text preview size is back down to small and I can't seem to figure out how to change it back to a larger preview size. Any suggestions? Thanks! -Rowen