Latin 2 Cyrilic

Hi,
i need help. This is my problem.
1. I find text in cyrillic (my language is macedonian), on web or pdf document
2. I copy the text and i paste it to word or pages
3. the pasted text is in latin letters
How can i change the text from latin letters into cyrillic letters. I have tried to select all, change the keyboard and change the fond into times cyrillic or similar, but nothing happens. I am forced to type the text in cyrillic again personally.
Any suggestions (maybe settings or so...)
Thx allot

If you could find of a copy of the non-standard font that was used to make this document, or a similar transliteration font that replaces Latin by Macedonian, then you could probably view the text and print it in other apps if you wanted.
I would recommend that no one should be using such fonts to put Macedonian on the internet, since it makes the text unusable for anything other than viewing/printing the original pdf.

Similar Messages

  • Why doesn't Cyrillic (Russian, Bulgarian, Serbian etc.) OO-based spell check work?

    Hi!
    The spell-check in OS X has official support for Russian, but it doesn't for Bulgarian, Serbian, Macedonian etc. One can of course use dictionaries from Open Office, install them, and in theory everything should be fine.
    I find it that none of the OO-based dictionaries for Cyrillic languages work. No words get underlined -- whether they are spelled correctly or incorrectly. I also installed the Open Office dicitonary for German and it worked without a hitch.
    If you are users of Cyrillic languages and if you have added your own .dic and .aff files to your Mac, could you please confirm that everything is working fine for you?
    (And, yes, my Cyrillic dictionaries show up in the language preferences and they are all enabled.)
    Any tips will be greatly appreciated.
    All best
    Tench

    @tenchpsych
    Hi there,
    The problem with Open Office Dictionaries for Slavic, cyrilic languages, is the encoding CP1251 which OS X does not like and will not combine with its native Western-latin encoding.  However OS X likes encoding UTF-8 and if you follow the steps below you will have a working cyrilic spell check-text-as you type, in OS X system wide. 
    I just noticed that the sr-SR package is already UTF-8 so you can use it out of the box.
    However the Macedonian, Bulgarian and Russian package are non UTF-8.
    To make OO non-UTF8 dictionaries work, download them from the Open Office website (dictionary Extentions), change the .oxt extention into .zip.  Unpack them, and browse the unpacked folder. You will only need the following 2 files (depanding which language you download, e.g. mk_MK.dic and mk_MK.aff (or bg_bg, ru_ru dic and aff)
    Now here is the catch: To convert CP1251 you need a converter, there are a couple of them online (google Convert cp-1251 into UTF-8), otherwise if you have Windows, open new Notepad document - copy-paste the text and save as. in the dropdown menu as UTF-8)
    If you are doing it in OS X open both files in TextEdit, select all, copy the unreadable text and paste it in the online converter.  Hit convert and copy and paste it back in the .dic and .aff files, respectively.  Don't forget to change the encoding command at the top of the .aff file it should read at the top * SET UTF-8 * .  Save and close both files.  Copy both files into /System/Library/Spelling and restart OS X.
    Now you have a system wide spellcheck in cyrilic, however it will not work if you choose the Multilanguage option, i.e. in TextEdit to check the spelling (for e.g. MK, SR or BG) you must go to Edit/SpellingandGrammar/ShowSpellingAndGrammar and Switch the language mannualy.  I repeat Automatic by Language will not work for Latin-Cyrilic combination.
    These are not very sophisticated spellchecers and may easily miss some mistakes, thus double-check your spelling. 

  • When I open an html file saved in UTF-8 format the 3-byte file header (usually hex EF BB BF) is shown as text at top-left. How can I avoid this - all other browsers interpret the header correctly and don't show it.

    I write VB apps that create web pages for international sports scoring systems, and these must be able to show Latin, Cyrillic, eastern names etc. that require 2-bytes for some characters in the text. This means saving in UTF-8 to avoid 8-bit ASCII files being created. UTF-8 files typically have a 3-byte header = hex EB BB BF, and with all browsers except Firefox this header is interpreted and not shown. Firefox however always displays these characters as  at the top-left screen position. What do I have to do the make Firefox recognize UTF-8 files, or do to my UTF-8 files to work properly with Firefox?
    See http://www.iacusn.org/USN2010/results/ for a typical recent example.

    You see that UTF-8 BOM (Byte Order Mark) because of the content type specification in the page code.
    <pre><nowiki>content="text/html; charset=windows-1252"</nowiki></pre>
    If you remove that or change it to utf-8 then Firefox won't show that BOM as text. There is however no need to insert that BOM at the start of the file. You can also configure the server to send the file (all files) as utf-8.

  • Xml Parse throws SaxParseException.Encoding is UTF-8 insteadof ISO-8859-1 ?

    Hi All,
    I'm having some korean characters in my xml. when i tried to parse the xml i'm getting SaxParseException .
    <?xml version="1.0" encoding="UTF-8"?> --- Throwing Exception
    <?xml version="1.0" encoding="ISO-8859-1"?> --- No Exception, successfully parsed
    I'm not sure why UTF-8 is failing and ISO is passing. But I'm always getting xml with UTF-8 format? Can anyone know the reason?
    I also like to know the differences between UTF-8 and ISO, i don't find any good article/document for this.
    Thanks,
    J.Kathir

    When SAX throws an exception when the encoding is set to UTF-8, then the XML contains something that is not a valid UTF-8 code (i.e. your source file is not encoded using UTF-8). Also: whenever you ask about an exception you should definitely post the entire exception, including message and stack trace.
    If it doesn't throw an exception when it is set to ISO-8859-1, then it does not mean that this is the correct choice. ISO-8859-1 is defined from 0 up to 255, so any byte stream is correct in that encoding ('though not necessarily meaninful).
    You absolutely have to find out which encoding the file really is, before you can parse it. If it should contain Korean characters then it is definitely not ISO-8859-1 (or any other encoding from the ISO-8859 family), as those only support latin, cyrillic and similar scripts.

  • Unicode support

    It seems there is no way to display UTF-8 encoded ePub files in DE 1.5 correctly. If uncompressed, HTML content looks fine in a browser. In DE all common accented characters are substituted with ? character.
    There are two similar topics on this forum:
    http://www.adobeforums.com/webx/.59b6343e
    http://www.adobeforums.com/webx/.59b60d2d
    My ePub file was generated via DocBook XSL Stylesheets and there is no problem to display it in FBReader.
    I suppose it is caused by default font, which contains probably very limited range of characters. Unfortunately there is no way to select the better one (as it is possible in FBReader).
    Something about changing the default font was "discussed" here: http://www.adobeforums.com/webx/.59b5e8b7
    If I met this issue 10 years ago, I would understand. If anybody in these days offer software for reading text with no support for non-roman languages, it is hard to believe.
    What I would suggest is to deliver DE with any very complete font, for example MinionPro (latin, cyrillic, CE, greek etc.). If this is impossible, please add a feature for selecting the default font.
    If I am totally wrong, clarification of this issue is welcomed.
    OS - Win XP Pro SP2
    Browser - Firefox 3.0.1

    Hi Jim--
    Thanks very much for the explanation on this...
    You know, Adobe could save us all a lot of time if the default faux fonts would include more unicode support....
    (I've spent probably a day overall trying to figure out what the issue is, and now I will spend the rest of the day learning/testing embedding a custom font in an epub, just for Digital Editions...)
    Even Mobipocket and Kindle are getting these entities right in the three books I'm working on, with their default fonts. (We're not talking about super-obscure characters. Try "Milosevic" with the proper characters, and you get a ? in DE.)
    Is there any chance Adobe can get the default fonts smarter in 1.8..?
    Thanks for any help/thoughts on this front...I'm sure I don't understand all the issues...it just seems like such an obvious straightforward thing to fix.
    --Kate

  • Dynamic Embedded Character Sub-Setting ???

    Hi All,
    I have instruction movies with 30 plus languages that I will be uploading to YouTube. I have Flash CS5 - AS3 - Player 10 req.
    I need to add only the font groupings (Latin, Cyrillic, Chinese, etc.) needed for the language specified and ideally only those characters that are being used by the selected language when the file is processed/complied.
    1. Can I programmatically add the characters being used to the internal list of embedded characters? Even if I have to run it twice to get it into the Library Font Object?
    2. Can I programmatically set the language grouping? I know I can turn layers that call different fonts off and off.
    3. Does the embedded character subsetting have a character limit? I can always make a utility to read the XML tags the have my displayed text and make a file or clipboard that I can paste it into the library. But, this is slow and tedious when you have 30 plus languages and 30 plus movies. I am trying to automate this whole process and limit inspection and non-automated testing.
    When I use this on a users CD, all fonts will be in a shared SWF. The fonts are over 8mb in size.
    I have a small issue with the #2 method - I have the potential need have Unicode characters in a text string that is in a ASCII based programming language. (FYI - the programs parser is smart enough to process the text string embedded in the code). I would hate to burden the movie with all the Chinese Characters or my newbie users/programmers (cough, cough) of my movie template with the task of ensuring that all characters a truly emdded and supported.
    Thanks,
    Jim

    A few points:
    (1) The point of embedding fonts in a PDF file is to make it possible for the PDF file to be properly displayed and printed by recipients of the PDF file that don't have the same exact fonts you do. Adobe recommends subset embedding of fonts whenever possible for highest quality display and printing of PDF file text. The subsetting includes only the glyphs you use and thus can save considerable file size. (For example, the Arial font in Windows is over 700K in size including all the glyphs for not only Western Latin languages, but also Cyrillic, Greek, Arabic, and Hebrew alphabets. Subsetting usually reduces the amount of the font embedded to 25K or less for typical documents.)
    (2) You don't own the fonts. You own a license to use the fonts. That is a very important difference!
    (3) Remember that PDF is a “final form file format” with very limited editability. If you need to do any extensive editing, you should have the source file, whether .AI (for Adobe Illustrator), .INDD (for Adobe InDesign), .DOCX (for Microsoft Word), etc. to you. You are responsible to getting a license for the fonts used in these documents if you don't already have one.
              - Dov

  • ORA-12741: Express Instant Client: unsupported server character set LV8PC8L

    How Can I connect to database???
    I've setup a 10g XE on trustix linux 3.0.5
    Compiled oci8 and pdo_oci
    I'm trying to connect to windows database running an old Oracle 8.1 database
    Any ideas how i can read data from that database - WITHOUT MODIFYING IT ?

    I'm not clear how you're connecting but this error could mean that you don't have transalations for Latvian Version IBM-PC Code Page 866 8-bit Latin/Cyrillic (LV8PC8LR). If that's the case it masquerades as the error when the underlying error can't be rendered in the language. This is my best guess.

  • Russian fonts problem

    "Hi
    We get wrong fonts (character set) from calling jolt connection to tuxedo
    The problem the charset that define for conversation to Unicode is good for windows
    (Power Builder environment) and not for Unix environment that run the Weblogic server
    Unix
    CL8ISO8859P5      ISO 8859-5 Latin/Cyrillic      SB, ASCII      
    Windows
    CL8MSWIN1251      MS Windows Code Page 1251 8-bit Latin/Cyrillic      SB, ASCII, EURO      
    Tuxedo server side define
    env | grep LANGNLS_LANG=AMERICAN_AMERICA.CL8MSWIN1251
    We need for weblogic server that run in Unix environment define:
    NLS_LANG=AMERICAN_AMERICA.CL8ISO8859P5
    NLS_LANG = LANGUAGE_TERRITORY.CHARACTER_SET
    Where:
    LANGUAGE     Specifies the language and conventions for displaying messages, day name, and month name.      
    TERRITORY     Specifies the territory and conventions for calculating week and day numbers.      
    CHARACTER_SET     Controls the character set used for displaying messages.      
    For exmple:
    Russian      RUSSIAN_CIS.CL8MSWIN1251      
    English      AMERICAN_AMERICA.WE8ISO8859P1      
    The result test with character set CL8MSWIN1251 define in Tuxedo
    Test 1
    testpet!VIP:/vmpuser1/vmp/vmp/api/Test/joltService> env | grep LANG
    LANG=ru
    System.getProperty- user.language ru
    System.getProperty- file.encoding ISO8859-5
    System.get

    Hi,
    could you please post this to
    weblogic.developer.interest.internationalization.
    Thanks
    Maria Salzberger
    Developer Relations Engineer
    BEA Customer Support
    yaakov lifshitz schrieb in Nachricht <3c33638f$[email protected]>...
    "Hi
    We get wrong fonts (character set) from calling jolt connection to tuxedo
    The problem the charset that define for conversation to Unicode is good forwindows
    (Power Builder environment) and not for Unix environment that run theWeblogic server
    >
    Unix
    CL8ISO8859P5 ISO 8859-5 Latin/Cyrillic SB, ASCII
    Windows
    CL8MSWIN1251 MS Windows Code Page 1251 8-bit Latin/Cyrillic SB, ASCII,EURO
    >
    >
    Tuxedo server side define
    env | grep LANGNLS_LANG=AMERICAN_AMERICA.CL8MSWIN1251
    We need for weblogic server that run in Unix environment define:
    NLS_LANG=AMERICAN_AMERICA.CL8ISO8859P5
    NLS_LANG = LANGUAGE_TERRITORY.CHARACTER_SET
    Where:
    LANGUAGE Specifies the language and conventions for displaying messages,day name, and month name.
    TERRITORY Specifies the territory and conventions for calculating week andday numbers.
    CHARACTER_SET Controls the character set used for displaying messages.
    For exmple:
    Russian RUSSIAN_CIS.CL8MSWIN1251
    English AMERICAN_AMERICA.WE8ISO8859P1
    The result test with character set CL8MSWIN1251 define in Tuxedo
    Test 1
    testpet!VIP:/vmpuser1/vmp/vmp/api/Test/joltService> env | grep LANG
    LANG=ru
    System.getProperty- user.language ru
    System.getProperty- file.encoding ISO8859-5
    System.get

  • AIR crashes on loading swf with 'Imported for runtime sharing' fonts

    Hello everybody,
    I have a trouble with an AIR application developped with Flex but the error seems to come from the Flash/AIR, after some searches a staff member (Flex harUI) said me that I could find more help for my trouble here.
    I'm developping an AIR application which loads an external swf. This swf contains 'Imported for runtime sharing' fonts from another swf to reduce its size.
    When the AIR application has finished to load (see message below) the whole application crashes without giving any error message except the usual "Process terminated unexpectedly" error message.
    End Loading: [SWF] C:\Users\dev02\myProject\index.swf - 83,930 bytes after decompression
    Error Message:
    Process terminated unexpectedly.
    Launch command details:  "C:\Program Files (x86)\Adobe\Adobe Flash Builder 4\sdks\4.6.0\bin\adl.exe" -runtime "C:\Program Files (x86)\Adobe\Adobe Flash Builder 4\sdks\4.6.0\runtimes\air\win" "C:\Users\dev02\Adobe Flash Builder 4\Languages_Editor\bin-debug\Language_Editor-app.xml" "C:\Users\dev02\Adobe Flash Builder 4\Languages_Editor\bin-debug"
    I use Flash Builder 4 (build 272416) with sdk 4.6.
    Full discussion here.
    I tried to load the external swf with the most basics codes, it works with a Flash web app but not with a Flex/AIR app or a Flash/AIR app.
    Thanks for the help !

    Thanks for the anwser.
    In fact, I cannot specify a swc because my main application is an editor of sub-applications where each one has its own library.
    The bug remains when sharing a simple MovieClip too.
    Maybe, if I explain my project it will be a little more clear.
    I'm developping a multi-languages module / application.
    For each module, I have an external xml for texts. So in order to add a new language we don't need to recompile, just to change texts in the xml. But, I had to externalize my fonts, (imagine if each swf contains the whole characters set like latin, cyrillic, etc.).
    And then, I have my main text editor application above the module, which loads texts xmls and a preview of the page where we need to change the text. (This is where it crashes, during the process of loading the preview).
    My editor works great for translating, changing texts, modify images but without the preview, what is really frustrating.
    (not so ergonomic, I know...)
    I hope it will help you to understand my trouble.
    regards

  • Language on Vista

    My laptop doesn't want to read russian latters.Its shows some symbols. What I have to do? please help!!!

    So far as I know, all versions of Windows Vista display Cyrillic characters. Perhaps it's the application.
    In particular, these fonts include Latin, Cyrillic, and Greek characters.
    -Jerry

  • Generic_M

    Can Generic_M be used for sorting Japanese language. As per my reading it looks like Generic_M can only sort Latin based languages but wanted to confirm.

    GENERIC_M sorts linguistically alphabetic European scripts: Latin, Cyrillic, Armenian and Georgian. Japanese characters, as any other non-European characters, are sorted by their Unicode code points.
    -- Sergiusz

  • Adjuset Oracle Charset

    Character Set Name Description Hexadecimal Code Value of the Euro Symbol
    D8EBCDIC1141 EBCDIC Code Page 1141 8-bit Austrian German 9F
    DK8EBCDIC1142 EBCDIC Code Page 1142 8-bit Danish 5A
    S8EBCDIC1143 EBCDIC Code Page 1143 8-bit Swedish 5A
    I8EBCDIC1144 EBCDIC Code Page 1144 8-bit Italian 9F
    F8EBCDIC1147 EBCDIC Code Page 1147 8-bit French 9F
    WE8PC858 IBM-PC Code Page 858 8-bit West European DF
    WE8ISO8859P15 ISO 8859-15 West European A4
    EE8MSWIN1250 MS Windows Code Page 1250 8-bit East European 80
    CL8MSWIN1251 MS Windows Code Page 1251 8-bit Latin/Cyrillic 88
    WE8MSWIN1252 MS Windows Code Page 1252 8-bit West European 80
    EL8MSWIN1253 MS Windows Code Page 1253 8-bit Latin/Greek 80
    WE8EBCDIC1047E Latin 1/Open Systems 1047 9F
    WE8EBCDIC1140 EBCDIC Code Page 1140 8-bit West European 9F
    WE8EBCDIC1140C EBCDIC Code Page 1140 Client 8-bit West European 9F
    WE8EBCDIC1145 EBCDIC Code Page 1145 8-bit West European 9F
    WE8EBCDIC1146 EBCDIC Code Page 1146 8-bit West European 9F
    WE8EBCDIC1148 EBCDIC Code Page 1148 8-bit West European 9F
    WE8EBCDIC1148C EBCDIC Code Page 1148 Client 8-bit West European 9F
    EL8ISO8859P7 ISO 8859-7 Latin/Greek A4
    IW8MSWIN1255 MS Windows Code Page 1255 8-bit Latin/Hebrew 80
    AR8MSWIN1256 MS Windows Code Page 1256 8-Bit Latin/Arabic 80
    TR8MSWIN1254 MS Windows Code Page 1254 8-bit Turkish 80
    BLT8MSWIN1257 MS Windows Code Page 1257 Baltic 80
    VN8MSWIN1258 MS Windows Code Page 1258 8-bit Vietnamese 80
    TH8TISASCII Thai Industrial 620-2533 - ASCII 8-bit 80
    AL32UTF8 Unicode 3.2 UTF-8 Universal character set E282AC
    UTF8 CESU-8 E282AC
    AL16UTF16 Unicode 3.2 UTF-16 Universal character set 20AC
    UTFE UTF-EBCDIC encoding of Unicode 3.0 CA4653
    ZHT16HKSCS MS Windows Code Page 950 with Hong Kong Supplementary Character Set A3E1
    ZHS32GB18030 GB18030-2000 A2E3
    WE8BS2000E Siemens EBCDIC.DF.04 8-bit West European 9F

    If you can insert Chinese characters, your client NLS_LANG must match the database's character set. In that case, what happens is basically
    Client tool (i.e. SQL*Plus) - Here's a string. It's encoded in the WE8ISO8859P1 character set. This is a lie-- the data is really Big5, but the client represents that the data is WE8ISO8859P1.
    Database - OK. I'll trust you and treat the data like a binary stream and just stuff it in.
    When this particular client tool queries the data, it gets back the same binary stream of data that it inserted, which the host OS displays since it, presumably, is expecting Big5 data, despite the fact that both sides of the Oracle connection claim that the data is WE8ISO8859P1.
    The problem comes when any tool decides not to lie to the database or to request the data in a different character set. In that case, Oracle tries to do a character set conversion and discovers that the binary stream it stored isn't WE8ISO8859P1 after all and the conversion either mangles the data or generates an error.
    In order to properly store Chinese, your database character set has to support Chinese (i.e. either Unicode or one of the Chinese language character sets) and the client NLS_LANG has to properly identify the character set the host operating system is using. It's perfectly valid for these to be different (i.e. a UTF-8 database and a Big5 client) so long as both character sets support the same character.
    Justin

  • Cyrillic Decoding To Latin

    I have to admit, I'm completely at a loss on this. I thought I understood what I was doing, but I don't.
    I have a text (htm) file downloaded off the web. When viewed in a browser, it appears with Russian characters. According to the meta-data on the page, it uses Windows-1251 to display.
    I also have a list of the characters and what their Latin equivalents are. I had initially thought I could simply parse through the file, and replace the Russian characters with the Latin equivalents, but that does not work. Usually the characters just get replaced with a ?.
    I was relying on Java being Unicode compliant for it to interpret the non-Ascii characters correctly, but I've since found that the whole characters sets and Unicode stuff is amazingly huge and complex.
    I've tried different approaches, but can't seem to get my head around what is actually happening. I keep bouncing between settings in my IDE (Eclipse), the Console, the Input Stream, the Output Stream, etc. I've tried the following charsets: UTF-8, UTF-16, ISO_8859-5, Windows-1251.
    1) Can conversion from Cyrillic to Latin be done purely using Character Set encoding and decoding, or does a find and replace methodology still stand?
    2) What character set should I be using to pick up the source file? (I viewed the source using Word/Notepad and got Scandinavian characters instead of Russian, so I can't even tell what character set the file itself is actually storing this in).
    3) Do I need to specify a particular character set for the output, either to the file or the console?
    The code below is the last attempt I made at this. Any help is greatly appreciated.
    Rob.
    import java.io.*;
    import java.util.HashMap;
    import java.util.Set;
    import java.util.Iterator;
    public class textrep
         //Snip
         public static void main(String[] args)
                                            //Snip
                        try
                             BufferedReader in = new BufferedReader(new InputStreamReader(new FileInputStream(source), "UTF-16"));
                             StringBuilder s = new StringBuilder();
                             int temp;
                             while ((temp = in.read()) != -1)
                                  s.append(russian((char) temp));
                             in.close();
                             BufferedWriter out = new BufferedWriter(new FileWriter(dest));
                             out.write(s.toString());
                             out.flush();
                             out.close();
                        catch (Exception e)
                                                           //Snip
         //Snip
         public static char russian(char c)
              update(c);
              switch (c)
                   case '�':
                   case '�':
                   case '�':
                   case '�':
                        return 'a';
                                                   //Snip - Carries on for all other known characters.
              return c;
    }Edited by: peridian on Feb 28, 2008 4:24 AM - Stripped out some code that was unneccessary (made it look weird on page)

    peridian wrote:
    I have to admit, I'm completely at a loss on this. I thought I understood what I was doing, but I don't.
    I have a text (htm) file downloaded off the web. When viewed in a browser, it appears with Russian characters. According to the meta-data on the page, it uses Windows-1251 to display.
    I also have a list of the characters and what their Latin equivalents are. I had initially thought I could simply parse through the file, and replace the Russian characters with the Latin equivalents, but that does not work. Usually the characters just get replaced with a ?.That suggests that when you read them in, you're using the wrong encoding. If you know what's in the file, your first step should be to read it in and see if your InputStreamReader is decoding to the correct values.
    I was relying on Java being Unicode compliant for it to interpret the non-Ascii characters correctly, but I've since found that the whole characters sets and Unicode stuff is amazingly huge and complex.There's a lot of detail, but basically it isn't complex at all. A charset converts between Unicode characters (inside your Java code) and bytes (outside your Java code). There's just a large number of charsets that all implement that.
    I've tried different approaches, but can't seem to get my head around what is actually happening. I keep bouncing between settings in my IDE (Eclipse), the Console, the Input Stream, the Output Stream, etc. I've tried the following charsets: UTF-8, UTF-16, ISO_8859-5, Windows-1251.
    1) Can conversion from Cyrillic to Latin be done purely using Character Set encoding and decoding, or does a find and replace methodology still stand?You need find-and-replace. Character encodings are strictly for converting Unicode characters to bytes. Expecting to find an encoding that replaces "&#1097;" by "shch" is absurd.
    2) What character set should I be using to pick up the source file? (I viewed the source using Word/Notepad and got Scandinavian characters instead of Russian, so I can't even tell what character set the file itself is actually storing this in).It's HTML, right? Does it have a <META> tag that states a charset? If it doesn't, then the browser is getting the charset from one of the response headers. You can use Firefox's Live HTTP Headers add-on to see the headers.
    However if you're seeing lots of &Aring; in your text editor that suggests to me that the real encoding is UTF-8. That's just a guess though. Another thing you could try is to load your text file into a hex editor and (looking at what the browser shows) see if you can see the equivalent bytes for the Windows-1251 encoding.
    3) Do I need to specify a particular character set for the output, either to the file or the console?If your output is all Latin, then any of the Latin-based charsets will do. Almost certainly your system's default encoding will be Latin-based.
    The code below is the last attempt I made at this. Any help is greatly appreciated.In your switch statement I don't see anything that uses Cyrillic characters. The case you left in considers four possible characters, none of which is a Cyrillic character. What's that supposed to be for?

  • IChat, ICQ and non-unicode ICQ clients (cyrillic=latin accented letters)

    I would like to use iChat to connect to my ICQ account
    If somebody with an older ICQ client sends me cyrillic I get garbage (latin accented letters).
    If they're using newer ICQ clients I'm getting the message in correct cyrillic.
    This imho has something to do with using non-unicode encoding for the messages.
    In other ICQ clients (e.g. Adium) I've used there was a setting specifying the encoding to use when there's no explicit encoding. Setting this to windows-1251 was solving the problem for me.
    But I see no similar setting for iChat.
    Can this be handled somehow ? Is there some preference somewhere ?

    Tom Gewecke wrote:
    Can this be handled somehow ?
    PS There is one thing you could try, namely setting the font used by iChat to one encoded as Win-1251. You can get some here, for example:
    http://web.ku.edu/~herron/fonts/download_fonts.html
    Whether having these on your system will mess up anything else I don't know.
    This didn't worked unfortunately. I'm still getting the text as accented latin letters. Apparently the network layer of iChat is translating them to unicode somehow.
    And no : having these fonts installed haven't affected anything else on my system from what I can tell.

  • Text lengs in Cyrillic characters vs in Latin characters

    Greetings!
    I've faced a problem, i have a member with a text parameter to input a text data (names), but in Cyrillic Essbase allows to input only 42 characters otherwise in Latin - all 255 as declared, how to solve that problem so a text in Cyrillic could be 255 characters long?
    Correct me if i'm wrong, there should be some table with parameters of values length types (i mean 255 byte, 255 char and so on), where can it be found?
    Thx in advance!

    I guess you mean Planning when you said Essbase. Is your planning repository schema unicode? I think what happens is that when the db is not unicode, text is stored with unicode mapping codes (i.e. For a non latin character something like: /u12345). Try creating another application with unicode db and migrating the artifacts with lcm.
    Cheers,
    Alp

Maybe you are looking for

  • Get name and type of data in an object?

    Hello, I would like to know if there is a way of getting the names and types of data encapsulated by a labview object? For example, lets say I make a simple object, called myObject to be unoriginal, with private data: myObjectName (string), myObjectV

  • Using specification for GDSN attributes?

    We are planning to participate in Global Data Synchronization Network (GDSN).  For that additional master data must be defined and managed. SAP does not provide standard solutions for this requirement. We use EHSM for GLM and other functions.  We the

  • Save and display the content

    Hello Experts, I have created an XML form to capture the data from the End User and created a HTML file to display the form and integrated this to an iView. Now, I have to capture the data from the XML form and display it in another page. Can you ple

  • I accidently put iphoto in Time machine. Iphoto won't open on desktop. Help?

    I must have accidently transfer my iphotos to my time machine. At least that's suggested by the error box. Help? Also, this all started by trynig to backup time capsuke with a seagate backup plus for mac.

  • Looking to simulate a DB for tests

    In order to test applications which uses DB (with JDBC), I would like to know if there's any package or tool to simulate a DB (if it's a file on the local system, or just a virtual memory used by the test application), so the test would be faster, an