Japanese Character Set - in Safari

I am considering purchasing a new Mac Mini, and need to access my Hotmail account with Japanesse Character Set. My XP Windows system requires the Japanese Language Pack. How will the Mac handle this? in other words, will Safari support the viewing and editing of Japanese Characters, or is there something I need to download? thanks!

found out that the Japanese Language Pack requires W7 Ultimate or Enterpise.
The Windows Japanese Language Pack is for turning the entire OS into Japanese. It has nothing to do with your ability to read Japanese or write Japanese while running your OS in English. I'm sure W7 comes with that installed by default, just like OS X does.
All browsers, Mac and Windows, automatically adjust to the character set provided in the code of the web page they are viewing, and also provide a way for the user to change it in the View > Text Encoding menu. I don't think you will have any trouble reading Japanese webmail with Safari or FireFox or Opera on a Mac.

Similar Messages

  • Problem displaying japanese character set in shopping cart smartform

    Hi All,
    whenever users are entering some text in Japanese character set while creating a shopping cart in SRM, the smartform print output is displaying some junk characters!! even though the system is unicode compatable, did any one have problem ??
    Thanks.

    Hi,
    May be there is some problem with UNICODE conversion.
    See the foll links;
    Note 548016 - Conversion to Unicode
    http://help.sap.com/saphelp_srm50/helpdata/en/9f/fdd13fa69a4921e10000000a1550b0/frameset.htm
    Europe Languages  work  in  Non- Unicode  System
    Re: Multiple Backends
    Re: Language issue
    Standard Code Pages in Non-Unicode System
    Re: Upgrade from EBP 4.0 to SRM 5.0
    http://help.sap.com/saphelp_srm50/helpdata/en/e9/c4cc9b03a422428603643ad3e8a5aa/content.htm
    http://help.sap.com/saphelp_srm50/helpdata/en/11/395542785de64885c4e84023d93d93/content.htm
    BR,
    Disha.
    Do reward points for  useful answers.

  • How to change Japanese character set

    The below are my character set in my DB
    NLS_CHARACTERSET=WE8ISO8859P1
    NLS_NCHAR_CHARACTERSET=UTF8
    Correct Answer (If I use english language the result is correct)
    ==========
    select product(',','AB_BC ,DE') from dual;
    (AB_BC, DE,,,)
    After altering the parameter at session level to get Japanese character set I am getting wrong result
    it is giving wrong result
    ==============
    select product(',','A_BC ,DE') from dual;
    (AB, BC , DE,,,,)
    How to change at session leavel to get Japanese character set

    user446367 wrote:
    Correct Answer (If I use english language the result is correct)What does "use english language" mean in this context?
    After altering the parameter at session level to get Japanese character set I am getting wrong resultThere is no such thing. Show us (copy paste) the commands and the resulting output, please.
    select product(',','A_BC ,DE') from dual;As requested several times already in your other thread on the same subject, it would greatly help forum members to help you if you would post the pl/sql of this function.
    AFAIK, product() is not a built-in standard Oracle function.
    How to change at session leavel to get Japanese character setThat is probably not what's needed, but anyway, here's one simple example:
    export NLS_LANG=.JA16SJIS
    sqlplus u/[email protected]
    sql> ...

  • ORACLE invoices with a Japanese character set

    We are having trouble printing ORACLE invoices with a Japanese character set.
    the printer we are using is a Dell W5300,do I need to configure the printer or is it something that needs to be configure in the software?????please help......

    We are having trouble printing ORACLE invoices with a
    Japanese character set.
    the printer we are using is a Dell W5300,do I need to
    configure the printer or is it something that needs
    to be configure in the software?????please help......What is the "trouble"? Are you seeing the wrong output? It may not be the printer, but the software that is sending the output to the printer.
    If you are using an Oracle Client (SQL*Plus, FOrms, Reports etc), ensure you set the NLS_LANG to JAPANESE_JAPAN.WE8MSWIN1252 or JAPANESE_JAPAN.JA16SJIS

  • HOW can I enter text using Japanese character sets?

    The "Text, Plates, Insets" section of the LOOKOUT(6.01) Help files states:
    "Click the » button to the right of the Text field to expand the field for multiple line entries. You can enter text using international character sets such as Chinese, Korean, and Japanese."
    Can someone please explain HOW to do this? Note, I have NO problem inputting Hirigana, Katakana, and Kanji into MS WORD; the keyboard emulates the Japanese layout and characters (Romaji is default) and the IME works fine converting Romaji, and I can also select charcters directly from the IME Pad. I have tried several different fonts with success and am currently using MS UI Gothic.ttf as default. Again, everything is normal and working in a predictable manner within Word.
    I cannot get these texts into Lookout. I can't cut/paste from HTML pages or from text editors, even though both display properly. Within Lookout with JP selected as language/keyboard, when trying to type directly into the text field, the IME CORRECTLY displays Hirigana until <enter> is pressed, at which point all text reverts to question marks (?? ???? ? ?????). If I use the IME Pad, it does pretty much the same. I managed to get the "Yen" symbol to display, though, if that's relevant. As I said, font selected (in text/plate font options) is MS UI Gothic with Japanese as the selected script. Oddly enough, at this point the "sample" window is showing me the exact Hirigana character I want displayed in Lookout, but it won't. I've also tried staying in English and copying unicode characters from the Windows Character Map. Same results (Yen sign works, Hirigana WON'T).
    Help me!
    JW_Tech

    JW_Tech,
    Have you changed the regional setting to Japanese?
    Doug M
    Applications Engineer
    National Instruments
    For those unfamiliar with NBC's The Office, my icon is NOT a picture of me
    Attachments:
    language.JPG ‏50 KB

  • UTF/Japanese character set and my application

    Blankfellaws...
    a simple query about the internationalization of an enterprise application..
    I have a considerably large application running as 4 layers.. namely..
    1) presentation layer - I have a servlet here
    2) business layer - I have an EJB container here with EJBs
    3) messaging layer - I have either Weblogic JMS here in which case it is an
    application server or I will have MQSeries in which case it will be a
    different machine all together
    4) adapter layer - something like a connector layer with some specific or
    rather customized modules which can talk to enterprise repositories
    The Database has few messages in UTF format.. and they are Japanese
    characters
    My requirement : I need thos messages to be picked up from the database by
    the business layer and passed on to the client screen which is a web browser
    through the presentation layer.
    What are the various points to be noted to get this done?
    Where and all I need to set the character set and what should be the ideal
    character set to be used to support maximum characters?
    Are there anything specifically to be done in my application code regarding
    this?
    Are these just the matter of setting the character sets in the application
    servers / web servers / web browsers?
    Please enlighten me on these areas as am into something similar to this and
    trying to figure out what's wrong in my current application. When the data
    comes to the screen through my application, it looks corrupted. But the asme
    message when read through a simple servlet, displays them without a problem.
    Am confused!!
    Thanks in advance
    Manesh

    Hello Manesh,
    For the database I would recommend using UTF-8.
    As for the character problems, could you elaborate which version of WebLogic
    are you using and what is the nature of the problem.
    If your problem is that of displaying the characters from the db and are
    using JSP, you could try putting
    <%@ page language="java" contentType="text/html; charset=UTF-8"%> on the
    first line,
    or if a servlet .... response.setContentType("text/html; charset=UTF-8");
    Also to automatically select the correct charset by the browser, you will
    have to include
    <meta http-equiv="Content-Type" content="text/html; charset=UTF-8"> in the
    jsp.
    You could replace the "UTF-8" with other charsets you are using.
    I hope this helps...
    David.
    "m a n E s h" <[email protected]> wrote in message
    news:[email protected]
    Blankfellaws...
    a simple query about the internationalization of an enterpriseapplication..
    >
    I have a considerably large application running as 4 layers.. namely..
    1) presentation layer - I have a servlet here
    2) business layer - I have an EJB container here with EJBs
    3) messaging layer - I have either Weblogic JMS here in which case it isan
    application server or I will have MQSeries in which case it will be a
    different machine all together
    4) adapter layer - something like a connector layer with some specific or
    rather customized modules which can talk to enterprise repositories
    The Database has few messages in UTF format.. and they are Japanese
    characters
    My requirement : I need thos messages to be picked up from the database by
    the business layer and passed on to the client screen which is a webbrowser
    through the presentation layer.
    What are the various points to be noted to get this done?
    Where and all I need to set the character set and what should be the ideal
    character set to be used to support maximum characters?
    Are there anything specifically to be done in my application coderegarding
    this?
    Are these just the matter of setting the character sets in the application
    servers / web servers / web browsers?
    Please enlighten me on these areas as am into something similar to thisand
    trying to figure out what's wrong in my current application. When the data
    comes to the screen through my application, it looks corrupted. But theasme
    message when read through a simple servlet, displays them without aproblem.
    Am confused!!
    Thanks in advance
    Manesh

  • Why does Firefox disable some Japanese character sets?

    Firefox 3.5.5 on Mac. Started yesterday at a point. My Japanese character selection (Kotoeri) disables all but Romaji. The only way to get others (like Hiragana and Katakana) back is to restart Firefox. It seems visiting certain web sites may trigger it, but don't know exactly. Why is this happening?
    It started from one of these and their links:
    http://www.yamatoamerica.com/
    http://www.ocsworld.com/
    http://www.dhl.com/
    I tried to recreate the situation, but couldn't. I will report back if I find one that triggers.

    You can check for issues caused by plugins (plugins are not affected by Safe mode).
    *https://support.mozilla.org/kb/Troubleshooting+plugins
    You can check for problems with current Flash plugin versions and try these:
    *disable a possible RealPlayer Browser Record Plugin extension for Firefox and update the RealPlayer if installed
    *disable protected mode in Flash 11.3 and later
    *disable hardware acceleration in the Flash plugin
    *http://kb.mozillazine.org/Flash#Troubleshooting

  • Japanese Character set

    Hi All,
    My DB Version:Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64bi
    I've one procedure which updates comment column(Varchar2(4000BYTE)) of a table,When I am passing a Japanese sentence to this procedure the underlying table column is getting updated with junk values like '??????, ??????????, ????????'.
    My Tbale structure is:
    Table Name:test_landing_commit
    Name Type Nullable Default Comments
    EXIT_COMMENT VARCHAR2(4000) Y
    CREATED_BY VARCHAR2(4000) Y
    Pprocedure is
    CREATE OR REPLACE PROCEDURE TEST_PROC_NM(VAR1 IN VARCHAR2) IS
    BEGIN
    UPDATE TEST_LANDING_COMMIT
    SET EXIT_COMMENT = VAR1
    WHERE CREATED_BY = 'XXX';
    END;
    and NLS_CHARACTERSET is set to UTF8.
    Please provide some advices to resolve this issue.

    The database is not being updated with junk ... you have not globalized your system.
    Go to http://tahiti.oracle.com and google and learn about globalization for your operating system and database version.
    PS: By "Japanese" do you mean Kangi? Hiragana? Katakana? Romanji?

  • Using Document Filters with the Japanese character sets

    Not sure if this belongs here or on the Swing Topic but here goes:
    I have been requested to restrict entry in a JTextField to English alphaNumeric and Full-width Katakana.
    The East Asian language support also allows Hiragana and Half-width Katakana.
    I have tried to attach a DocumentFilter. The filter employs a ValidateString method which strips all non (Latin) alphaNumerics as well as anything in the Hiragana, or Half-width Katakana ranges. The code is pretty simple (Most of the code below is dedicated to debugging):
    public class KatakanaInputFilter extends DocumentFilter
         private static int LOW_KATAKANA_RANGE = 0x30A0;
         private static int LOW_HALF_KATAKANA_RANGE = 0xFF66;
         private static int HIGH_HALF_KATAKANA_RANGE = 0xFFEE;
         private static int LOW_HIRAGANA_RANGE = 0x3041;
         public KatakanaInputFilter()
              super();
         @Override
         public void replace(FilterBypass fb, int offset, int length, String text,
                   AttributeSet attrs) throws BadLocationException
              super.replace(fb, offset, length, validateString(text, offset), null);
         @Override
         public void remove(FilterBypass fb, int offset, int length)
                   throws BadLocationException
              super.remove(fb, offset, length);
         // @Override
         public void insertString(FilterBypass fb, int offset, String string,
                   AttributeSet attr) throws BadLocationException
              String newString = new String();
              for (int i = 0; i < string.length(); i++)
                   int unicodePoint = string.codePointAt(i);
                   newString += String.format("[%x] ", unicodePoint);
              String oldString = new String();
              int len = fb.getDocument().getLength();
              if (len > 0)
                   String fbText = fb.getDocument().getText(0, len);
                   for (int i = 0; i < len; i++)
                        int unicodePoint = fbText.codePointAt(i);
                        oldString += String.format("[%x] ", unicodePoint);
              System.out.format("insertString %s into %s at location %d\n",
                        newString, oldString, offset);
              super.insertString(fb, offset, validateString(string, offset), attr);
              len = fb.getDocument().getLength();
              if (len > 0)
                   String fbText = fb.getDocument().getText(0, len);
                   for (int i = 0; i < len; i++)
                        int unicodePoint = fbText.codePointAt(i);
                        oldString += String.format("[%x] ", unicodePoint);
              System.out.format("document changed to %s\n\n", oldString);
         public String validateString(String text, int offset)
              if (text == null)
                   return new String();
              String validText = new String();
              for (int i = 0; i < text.length(); i++)
                   int unicodePoint = text.codePointAt(i);
                   boolean acceptChar = false;
                   if (unicodePoint < LOW_KATAKANA_RANGE)
                        if ((unicodePoint < 0x30 || unicodePoint > 0x7a)
                                  || (unicodePoint > 0x3a && unicodePoint < 0x41)
                                  || (unicodePoint > 0x59 && unicodePoint < 0x61))
                             acceptChar = false;
                        else
                             acceptChar = true;
                   else
                        if ((unicodePoint >= LOW_HALF_KATAKANA_RANGE && unicodePoint <= HIGH_HALF_KATAKANA_RANGE)
                                  || (unicodePoint >= LOW_HIRAGANA_RANGE && unicodePoint <= LOW_HIRAGANA_RANGE))
                             acceptChar = false;
                        else
                             acceptChar = true;
                   if (acceptChar == true)
                        System.out.format("     Accepted code point = %x\n",
                                  unicodePoint);
                        validText += text.charAt(i);
                   else
                        System.out.format("     Rejected code point = %x\n",
                                  unicodePoint);
              String newString = "";
              for (int i = 0; i < validText.length(); i++)
                   int unicodePoint = validText.codePointAt(i);
                   newString += String.format("[%x] ", unicodePoint);
              System.out.format("ValidatedString = %s\n", newString);
              return validText;
          * @param args
         public static void main(String[] args)
              Runnable runner = new Runnable()
                   public void run()
                        JFrame frame = new JFrame("Katakana Input Filter");
                        frame.setDefaultCloseOperation(JFrame.EXIT_ON_CLOSE);
                        frame.setLayout(new GridLayout(2, 2));
                        frame.add(new JLabel("Text"));
                        JTextField textFieldOne = new JTextField();
                        Document textDocOne = textFieldOne.getDocument();
                        DocumentFilter filterOne = new KatakanaInputFilter();
                        ((AbstractDocument) textDocOne).setDocumentFilter(filterOne);
                        textFieldOne.setDocument(textDocOne);
                        frame.add(textFieldOne);
                        frame.setSize(250, 90);
                        frame.setVisible(true);
              EventQueue.invokeLater(runner);
    }I run this code, use the language bar to switch to Full-width Katakana and type "y" followed by "u" which forms a valid Katakana character. I then used the language bar to switch to Hiragana and retyped the "Y" followed by "u". When the code sees the Hiragana codepoint generated by this key combination it rejects it. My debugging statements show that the document is properly updated. However, when I type the next character, I find that the previously rejected codePoint is being sent back to my insert method. It appears that the text somehow got cached in the composedTextContent of the JTextField.
    Here is the output of the program when I follow the steps I just outlined:
    insertString [ff59] into at location 0 <== typed y (Katakana)
    Accepted code point = ff59
    ValidatedString = [ff59]
    document changed to [ff59]
    insertString [30e6] into at location 0 <== typed u (Katakana)
    Accepted code point = 30e6
    ValidatedString = [30e6]
    document changed to [30e6]
    insertString [30e6] [ff59] into at location 0 <== typed y (Hiragna)
    Accepted code point = 30e6
    Accepted code point = ff59
    ValidatedString = [30e6] [ff59]
    document changed to [30e6] [ff59]
    insertString [30e6] [3086] into at location 0 <== typed u (Hiragana)
    Accepted code point = 30e6
    Rejected code point = 3086
    ValidatedString = [30e6]
    document changed to [30e6]
    insertString [30e6] [3086] [ff59] into at location 0 <== typed u (Hiragana)
    Accepted code point = 30e6
    Rejected code point = 3086
    Accepted code point = ff59
    ValidatedString = [30e6] [ff59]
    document changed to [30e6] [ff59]
    As far as I can tell, the data in the document looks fine. But the JTextField does not have the same data as the document. At this point it is not displaying the ff59 codePoint as a "y" (as it does when first entering the Hiragana character). but it has somehow combined it with another codePoint to form a complete Hiragana character.
    Can anyone see what it is that I am doing wrong? Any help would be appreciated as I am baffled at this point.

    You have a procedure called "remove" but I don't see you calling it from anywhere in your program. When the validation failed, call remove to remove the bad character.
    V.V.

  • Problem with Japanese character Codes

    I have been working for a company which got a product to support the sms activities. We are developing a new feature which supports Japanese language as of now along with english. We developed a new character set called PDC which is a combination of the following Japanese character sets JIS + SHIFT- JIS+ KATAKANA + ASCII.
    I am using Java API Charsets.jar to convert the bytes using the respective character sets to a STRING and presenting the converted STRING on a JSP.
    The Problem is I am able to see the Jap Chars which belong to JIS + ASCII but not for KATAKANA and SHIFT-JIS.
    I am using the below page tag in JSP.
    <%@ page contentType="text/html; charset = JIS " %>
    I couldn't find any other tag specifically which supports all the character sets. Could some one please help me?
    Pradeep.

    Okay... you developed a new character set... that's great. How's the browser supposed to know anything abou it?
    Maybe just use UTF-8 which should support all of that and other languages as well and stop boxing yourself into a fixed charset.

  • Japanese Character Support

    Hi,
    Is IdM 8.0 able to handle Japanese characters. Not only in the input fields, but also in all menus, drop downs etc. Or does this depend on the application server setup used to host IdM? If so im using IBM Websphere 6.1, so i guess the same question applies......
    Thoughts?
    Thanks,
    Ant.

    Yep apparently it does support the full Japanese character set.

  • Apps displyaing Chinese/japanese character

    Hi
    For some odd reason, Yahoo IM and Sofware Update Window are being displayed using non-English charracter. I believe it is Japanese character set.
    have anyone else experience this problem and got any idea on how to fix?
    Thanks

    For some odd reason, Yahoo IM and Sofware Update
    Window are being displayed using non-English
    charracter. I believe it is Japanese character set.
    Most likely it's not Japanese but fractions. Get rid of the font Helvetica Fractions if you have it. Send a screenshot if you want verification (tom at bluesky dot org).

  • Multiple Character set for NLS

    Hi,
    I'm using Oracle 8i database. Is it possible to set the different character set for the database? The requirement is to support the two different character set data, one (main) Japanese character set and other Simplified Japanese. Or is there any other way in which i can store these data (Japanese & Chinese)?
    Thanks & Regards,
    Jayesh

    Please don't get me wrong. Currently it is set in the windows database. I did not set nls_lang at the command prompt before import into windows. However nls_lang is already set and it is character set WE8ISO8859P1 the same as the value I specified in creation script, besides the other two values AMERICAN, AMERICA. They are now same in both solaris and windows. Only the character sets are different because I specified a different one. So, is it ok or do I now need another fresh import this time with nls_lang set to AMERICAN_AMERICA.UTF8 ?

  • Session level character set

    The below are my character set in my DB
    NLS_CHARACTERSET=WE8ISO8859P1
    NLS_NCHAR_CHARACTERSET=UTF8
    Correct Answer (If I use english language the result is correct)
    ==========
    select product(',','AB_BC ,DE') from dual;
    (AB_BC, DE,,,)
    After altering the parameter at session level to get Japanese character set I am getting wrong result
    ALTER SESSION NLS_SORT=JAPANESE_M_AI
    ALTER SESSION NLS_COMP=LINGUISTICS
    it is giving me wrong result (I should get the above result)
    =================
    select product(',','A_BC ,DE') from dual;
    (AB, BC , DE,,,,)
    How to change at session leavel to get Japanese character set

    Ok,
    Let's provide the broad picture, as your setup and your commands are incorrect.
    You set the characterset of the database to a characterset the O/S supports.
    Whether or not you have characterset conversion on the client side is determined by NLS_LANG.
    You set NLS_LANG to a characterset the client O/S supports. Ie if you are running on Windows (as always you provide no details at all), the regional settings of the O/S must have been set to Japanese.
    WE8ISO8859P1 is the Latin-1 alphabet and doesn't support Kanji.
    Also the command you specify deal with sorting of the data, not with the characterset itself.
    Also no one can tell what
    select product(',','AB_BC ,DE') from dual;
    constitutes.
    In summary: you don't seem to be reading documentation, or you only look at it.
    None of what you have posted makes any sense, and clearly shows you didn't try to understand the NLS concept.
    Yet again: you don't provide platform and version info
    Yet again: you don't specify any background.
    If you want help, you need to provide as much info as possible.
    You should not require anyone here to tear the information out of you.
    After all: everyone here is a volunteer and doesn't get paid to help you out, but is spending his/her time.
    If you want to continue to post in this fashion, maybe you should find a forum of mindreaders.
    Sybrand Bakker
    Senior Oracle DBA

  • Crystal XI R2 exporting issues with double-byte character sets

    NOTE: I have also posted this in the Business Objects General section with no resolution, so I figured I would try this forum as well.
    We are using Crystal Reports XI Release 2 (version 11.5.0.313).
    We have an application that can be run using multiple cultures/languages, chosen at login time. We have discovered an issue when exporting a Crystal report from our application while using a double-byte character set (Korean, Japanese).
    The original text when viewed through our application in the Crystal preview window looks correct:
    性能 著概要
    When exported to Microsoft Word, it also looks correct. However, when we export to PDF or even RPT, the characters are not being converted. The double-byte characters are rendered as boxes instead. It seems that the PDF and RPT exports are somehow not making use of the linked fonts Windows provides for double-byte character sets. This same behavior is exhibited when exporting a PDF from the Crystal report designer environment. We are using Tahoma, a TrueType font, in our report.
    I did discover some new behavior that may or may not have any bearing on this issue. When a text field containing double-byte characters is just sitting on the report in the report designer, the box characters are displayed where the Korean characters should be. However, when I double click on the text field to edit the text, the Korean characters suddenly appear, replacing the boxes. And when I exit edit mode of the text field, the boxes are back. And they remain this way when exported, whether from inside the design environment or outside it.
    Has anyone seen this behavior? Is SAP/Business Objects/Crystal aware of this? Is there a fix available? Any insights would be welcomed.
    Thanks,
    Jeff

    Hi Jef
    I searched on the forums and got the following information:
    1) If font linking is enabled on your device, you can examine the registry by enumerating the subkeys of the registry key at HKEY_LOCAL_MACHINEu2013\SOFTWARE\Microsoft\Windows NT\CurrentVersion\FontLink\SystemLink to determine the mappings of linked fonts to base fonts. You can add links by using Regedit to create additional subkeys. Once you have located the registry key that has just been mentioned, from the Edit menu, Highlight the font face name of the font you want to link to and then from the Edit menu, click Modify. On a new line in the dialog field "Value data" of the Edit Multi-String dialog box, enter "path and file to link to," "face name of the font to link".u201D
    2) "Fonts in general, especially TrueType and OpenType, are u201CUnicodeu201D.
    Since you are using a 'true type' font, it may be an Unicode type already.However,if Bud's suggestion works then nothing better than that.
    Also, could you please check the output from crystal designer with different version of pdf than the current one?
    Meanwhile, I will look out for any additional/suitable information on this issue.

Maybe you are looking for

  • Class visibility problems in deployed OC4J application

    I have an entity, autogenerated from a database table by JDeveloper, which has the form: Entity.java: package a.b.c; import ... @Entity @NamedQuery( ... ) @Table( ... ) public class Entity implements Serializable {     @Column( ... )     : etc. }Then

  • Photo are not showing in view area when looking in folder for picture

    When i have a photo folder open the pictures are not showing where i can see them without having to open the file, they used to after i updated my touch smart the photo are is gone

  • RAM Preview issue

    Hi, I'm wondering if anyone has seen this or can provide some direction for further investigation. This is on a new Mac Pro on which I have just installed FCS In some projects the RAM preview has no effect on playback performance. After generating a

  • How to programmatically make pages of a book visible or invisible

    My portal page has three placeholders: Left, center and Right. In the right there are simple static pages giving some information and news that are irrelevant to the whole quesry so dont bother about them. In the center placeholder, I have put a book

  • Mtdchar.ko missing 3.18.6-1

    Hi Everyone, Kernel module mtdchar.ko is not present in 3.18.6-1.  Is this known, intentional, unintentional? I'm new to Arch Linux; how would/should I go about obtaining mtdchar.ko?  My kernel information is: Linux dkarchlinux64 3.18.6-1-ARCH #1 SMP