Intermedia drawback with UTF-8

Hi ,
There seems to be a limitation in Intermedia . It cannot index characters that are 2 bytes.
For example , Iam indexing an xml which has encoding set as UTF-8 .The indexer is unable to recognise the characters like i g ,etc
Any ideas
Regards,
Syed Niaz

hello
i am using intermedia since 8.1.5 and now we upgrade to 9i
i indexing columns containing multilingual text
(german, english, french, swedish etc)
and i have NOT experience any problems ...
Is there any new info about your problem ????
Thanks in advance

Similar Messages

  • Problem with UTF-8 japanese on Win XP

    Hi,
    I have a Client/server application with a server coded in C++ and runing on HP workstation and the GUI in java runing in a web context via java webstart mechanism on PC machine with win XP sp2 OS.
    when logging into the HP workstation in japanese SHIFT_JIS and downloading the application japanese Charcters are displayed correctly on my PC.
    However when logging into the HP workstation in japanese with UTF-8 encodage and kaunching the application characters aren't displayed correctly only rectangle charcaters are displayed, when launching the GUI on HP workstation it displayed correctly.
    Can any one help me to resolve this problem.
    Regards,

    Both SHIFT_JIS and UTF-8 are character encodings. SHIFT_JIS encodes a superset of JISX-0208 to double-byte codepoints and JISX-0201 to single byte codepoints. UTF-8 encodes all of Unicode.
    The problem here is that you are relying on "default" encoding for a client-server application. This only works when the default encoding is the same on both ends.
    Go through your code, looking for everywhere where you convert to and from Strings, including reading and writing to IO Streams, and replace the methods you are using with equivalent methods that let you specify an encoding.

  • Message Mapping Problem with UTF-16LE Encoded XML

    Hello,
    we have the following scenario:
    IDoc > BPM > HTTP Sync Call > BPM > IDoc
    Resonse message of the HTTP call is a XML file with UTF-16LE processing instruction. This response should then be mapped to a SYSTAT IDoc. However the message mapping fails "...XML Parser: No data allowed here ...".
    So obviously the XML is not considered as well-formed.
    When taking a look at SXMB_MONI the following message appears: "Switch from current encoding to specific encoding not supported.....".
    Strange thing however is if I save the response file as XML and use the same XML file in the test tab message mapping is executed successfully.
    I also tried to use a Java Mapping to switch encodings before executing message mapping, but the error remains.
    Could the problem be, that the codepage UTF-16LE is not installed on the PI system ? Any idea on that ?
    Thank you!
    Edited by: Florian Guppenberger on Feb 2, 2010 2:29 PM
    Edited by: Florian Guppenberger on Feb 2, 2010 2:29 PM

    Hi,
    thank your for your answer.
    This is what I have tried to achieve. I apply the java conversion mapping when receiving the response message - i tried to convert the response to UTF-16, UTF-8 but none of them has helped to solve the problem.
    I guess that using adapter modules is not an option either as it would modify the request message, but not the response, right?

  • How to write csv or txt file through utl_file with UTF-8 Encoding

    Hi All,
    I need your help to write the data from DB to csv or txt file with UTF-8 encoding through utl_file.
    Database character set:AL32UTF8
    Database version:10G
    All the columns in the DB are of varchar2 type.
    Please let me know if there is any way of doing it.

    What was wrong with the info provided in the link(s) given?
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14200/functions027.htm#SQLRF00620]

  • [SOLVED] Problems opening folders with UTF-8 encoded characters

    Hello everyone, I'm having an issue when I acess folders in all my programs ( except Dolphin File Manager). Every time I open the folder navigation window in my programs, folders with UTF-8 encoded characters ( such as "ç", "á ", "ó", "í", etc ) are not shown or the folder name not show these characters, therefore, I can not open documents inside these folders.
    However, as you saw, I can type these characters normally. Here's my "locale.conf" :
    LANG="en_US.UTF-8:ISO-8859-1"
    LC_TIME="pt_BR.UTF-8:ISO-8859-1"
    And here's the output of the command "locale -a" :
    C
    en_US.utf8
    POSIX
    Last edited by regmoraes (2015-04-17 12:55:19)

    Thing is, when I run locale -a, I get
    $ locale -a
    C
    de_DE@euro
    de_DE.iso885915@euro
    de_DE.utf8
    en_US
    en_US.iso88591
    en_US.utf8
    ja_JP
    ja_JP.eucjp
    ja_JP.ujis
    ja_JP.utf8
    japanese
    japanese.euc
    POSIX
    So an entry for every locale I have uncommented in my locale.conf. Just making sure, by "following the steps in the beginner's guide", you also mean running locale-gen?
    Are those folders on a linux filesystem like ext4 or on a windows (ntfs?)

  • [svn:fx-trunk] 7661: Change from charset=iso-8859-1" to charset=utf-8" and save file with utf-8 encoding.

    Revision: 7661
    Author:   [email protected]
    Date:     2009-06-08 17:50:12 -0700 (Mon, 08 Jun 2009)
    Log Message:
    Change from charset=iso-8859-1" to charset=utf-8" and save file with utf-8 encoding.
    QA Notes:
    Doc Notes:
    Bugs: SDK-21636
    Reviewers: Corey
    Ticket Links:
        http://bugs.adobe.com/jira/browse/iso-8859
        http://bugs.adobe.com/jira/browse/utf-8
        http://bugs.adobe.com/jira/browse/utf-8
        http://bugs.adobe.com/jira/browse/SDK-21636
    Modified Paths:
        flex/sdk/trunk/templates/swfobject/index.template.html

    same problem here with wl8.1
    have you sold it and if yes, how?
    thanks

  • Java: Loading Resource Bundle File with UTF-8

    I have found out all my previous problems now, thank you all for your assitance with this...
    During the time we have worked so solve our problem with UTF-8 charset we have found new problem with this?
    I don't know if this is the right forum to put this up, if not please let us know where to go with this?
    I have prepared both Eclipse (JSP Files) and GlassFish to run all my files in charset UTF-8. Then i got my
    resource bundle files(property files) to use UTF-8 encoding. Everything looks fine in my text editor. Then
    when i upload these pages every character retrieved from my resource bundles look garbled. This is because
    Java loads property files using ISO 8859-1 character encoding.
    This is a problem from Sun i have read on the internet!
    Note: that in JDK 1.6, you can use PropertyResourceBundle constructor to take a Reader which uses UTF-8 encoding.
    please have a look at these links for more info:
    [http://www.kai-mai.com/node/128]
    [http://java.sun.com/javase/6/docs/api/java/util/PropertyResourceBundle.html]
    Note: PropertyResourceBundle can be constructed either from an InputStream or a Reader, which represents a property file.
    Constructing a PropertyResourceBundle instance from an InputStream requires that the input stream be encoded in ISO-8859-1.
    In that case, characters that cannot be represented in ISO-8859-1 encoding must be represented by Unicode Escapes,
    whereas the other constructor which takes a Reader does not have that limitation.
    What i have read and understand is that the real solution to the problem is: Get the fmt lib to use a Reader instead of an
    InputStream. Alternativly if it works get a ResourceBundle instead of creating one that is wrong all the time? I have to create
    my own ResourceBundle instead of using my own fmt lib wich i cannot trust in a way any more. Is this right or am i comletly wrong?
    My WebTexts files, en, sv and ry are my resource property bundles. To be able to read them i will use resource property bundle reader.
    Fmt lib gives me this but it seems that it inciates this in an InputStream (wich is looked to iso-8859-1) instead of a reader who can
    read UTF-8 without no problem. So can i get my fmt lib to change this or not, anybody have an idea about this???

    Torleif wrote:
    Ok, i know there are a few ways of doing this already. The problem is that i have no idea how to use these different problem solving issues. First we have the;
    1. native2ascii command line utility.
    2. use the XML properties format.
    3. other ways of converting from iso-8859-1 to utf-8.
    The last one i have tried already but i never get this to work. So after knowing of 2 standardized solutions you choose to implement your own non-standard soltion? That's usually not a good idea until you've verified that the standardized solution don't help.
    This is also not the best solution to the problem i have to say. This way i will have both properties files and source files to work with and that is to much i think. Either i will use the native2ascii command or use XML properties format like you say here. I need some more insight in how these two work only because i have no idea my self.Uhm .. read [the documentation|http://java.sun.com/javase/6/docs/api/java/util/Properties.html]? loadFromXML/storeToXML work pretty much exactly the same as load/store, except they handle XML. Try it, look at what it produces, learn.
    Could you please help and explain how to use either this native2ascii command line work or how this XML format properties would work for me. Please i need guidance here with this!!![native2ascii is documented as well|http://java.sun.com/javase/6/docs/technotes/tools/windows/native2ascii.html].
    If you have a specific question after reading those, feel free to ask here.
    You really have to work on your Google-fu, it seems to be too weak.

  • Working with UTF-8 text files as if they were ANSI?

    I have a text file saved in UTF-8, containing several characters from different languages. It's created in Notepad.
    I want to read that file into my Java application, which before only was used with regular ANSI .txt files.
    When showing the contents of the file on the screen (Swing GUI), the text is totally distorted. The application obviously doesn't understand from the beginning that the text file is encoded with UTF-8.
    Somehow I must convert the information in the file so that
    1) I can process the UTF-8 text in all String functions as if it were common ANSI text, and
    2) I can show the contents of those Strings on the screen in my Java application, and making it look like it should, whether it's English or Russian.
    Does anyone here have any tips or code samples that I can use?
    Thanks

    That is kind of what I'm trying to find out, how to read the file...
    When I read a text file that is saved as UTF-8, the letters don't show properly on the screen. Instead of, for example, a Pi symbol I would get a ?� or something.
    Though if I hardcode a String object that contains "\u0030" or a similar Unicode character, it shows properly.
    How do you read in the file???
    When you read in the file, you have to take the
    encoding into account.
    Take a look at InputStreamReader. This class allows
    you to specify the character encoding you will use.

  • [Solved] Automount Generic MP3 Player with UTF-8 encoding

    Hello, everybody!
    Ubuntu refugee here.
    So far I'm going fine with Arch, I just have a couple of problems related to my Generic USB Mp3 player:
    1) I want HAL to mount the player with UTF-8 encoding. Right now, it shows Arabic characters as ??????.
    2) In Ubuntu, it used to recognize the player as an MP3 player, give it a nice icon, and add it as a music source in Rhythmbox. In Arch, on the other hand, the player is mounted as a generic USB flash drive. How can I make Arch recognize it as an MP3 player?
    Thanks in advance.
    Last edited by farghal (2008-05-10 20:17:09)

    GOT IT WORKING!! Yay!
    I got the solution from here:
    http://blog.pcode.nl/2006/08/24/introdu … io-player/
    The trick is to make HAL identify your Digital Audio Player (DAP) by adding a rule to /usr/share/hal/fdi/information/10freedesktop/10-usb-music-players.fdi  --  and since, in my case, Ubuntu already had a 10-usb-music-players.fdi file that recognized my player, all I had to do was boot up from an Ubuntu live cd and copy Ubuntu's 10-usb-music-players.fdi over Arch's.
    Now my issues with Arch are down to only one: http://bbs.archlinux.org/viewtopic.php?pid=360647
    Thanks everybody.
    Last edited by farghal (2008-05-10 20:16:49)

  • Trouble with UTF-8 and PHP-OCI

    Hi there!
    I'm having some serious trouble with UTF-8. I just tried to insert a lambda into an NCLOB column in one of my databases and it was converted to an inverted question mark. I have verified that the string reaches my PHP script correctly encoded. Also selecting RAWTOHEX(column) in SQL Developer shows, that the inverted question mark is already stored in the column. So the problem must be somewhere between PHP-OCI and the database. Inserting a lambda via SQL developer works. I can also correctly fetch it via PHP.
    I'm using the latest PHP-OCI (2.0.8). v$version says "Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production".
    The database is fairly old and uses WE8MSWIN1252 as the character set and AL16UTF16 as the NCHAR character set. Hence I'm using NCLOB instead of CLOB. However, I connect to the database with AL32UTF8, since my app is running in UTF-8. I was under the impression that this would result in an automatic conversion from UTF-8 to UTF-16 when inserting into NCLOB columns, but apperently this is not the case. It looks like there is some sort of double conversion, first UTF-8 to WE8MSWIN1252 and then to UTF-16, because some non-ASCII characters like ä (a umlaut) get correctly converted from UTF-8 to UTF-16.
    Any ideas? I'm at a loss here. Thanks in advance.

    PHP OCI8 doesn't support NCLOB or NVARCHAR2.
    See NCLOB support in OCI8

  • MySql with UTF-8

    Hi All.
    I'm stuck with this so much time...
    I can't store and retrive UTF-8 strings in the right way. (I get "????" instead of the original string).
    I sow in the forums here that for mySql to work with UTF-8 I only need to set the connection string like this:
    jdbc:mysql://localhost/myDatabase?useUnicode=true&characterEncoding=UTF-8.
    I read my UTF-8 from a UTF-8 text file and I write the result to another text file.
    My java source files are encoded in UTF-8 and I use -Dfile.encoding="UTF-8" for the JVM (though I'm not sure it's necessary).
    Here is my code:
    //here I get db connection
    Class.forName("com.mysql.jdbc.Driver");
    Connection con = DriverManager.getConnection("jdbc:mysql://localhost/myDatabase?useUnicode=true&characterEncoding=UTF-8", "myUserName", "myPass");
    //get string from a utf-8 encoded text file and insert it to myTable
    PreparedStatement ps=con.prepareStatement("INSERT INTO myTable (id) VALUES(?)");
    BufferedReader reader = new BufferedReader( new InputStreamReader (new FileInputStream(new File("in.txt")), "UTF-8"));
    String s = reader.readLine();
    reader.close();ps.setString(1,s);
    ps.executeUpdate();
    //store the utf-8 string that I get from myTable to text file
    ps=con.prepareStatement("SELECT * FROM myTable");
    ResultSet rs=ps.executeQuery();
    if (rs.next()){
    PrintWriter writer = new PrintWriter( new OutputStreamWriter(new FileOutputStream(new File("out.txt")), "UTF-8"));
    writer.println(rs.getString(1));
    writer.close();
    rs.close();
    I use mysql-4.1.1-alpha and j2sdk1.4.1_03.
    That's it.
    I hope you can help me, 'cause I'm totaly despaired with this.
    Thanks.

    !!! Come on !!!

  • Using InterMedia Text with FrameMaker, StarOffice...

    I'm planning to use InterMedia text with various kinds of documents, but inso filter does not support them. In particular, it does not support:
    - Framemaker
    - PostScript (I know it works on encapsulated PostScript)
    - Applix
    Where can I find the right filters?
    Is there a way to use InterMedia text with these documents?
    Thank you
    Paolo Bonanomi

    I'd suggest searching the web - there are a number of little freeware filters out there that could be used as user-filters.>Have anybody found a postscript filter and can post it here ?
    Regards,
    Mihai

  • Insert/Update VO with UTF-8 charset

    Hi,
    I worked so far only with iso-8859-1 charset and everything went fine, but now with UTF-8 I am experiencing strange problems. Every unicode char is converted to html equivalent(e.g. &#xxxx;) and saved in that form to DB. Is there any work around for that issue?
    Thanks in advance.
    Alex

    Hi,
    I worked so far only with iso-8859-1 charset and everything went fine, but now with UTF-8 I am experiencing strange problems. Every unicode char is converted to html equivalent(e.g. &#xxxx;) and saved in that form to DB. Is there any work around for that issue?
    Thanks in advance.
    Alex

  • Sending url not with UTF-8

    Hi,
    In internet explorer, when I try to download some files via a non english ( or non UTF-8? I am not really sure ) http link, I can unselect the "send url only with UTF-8" in perference of IE.
    Can I do the same thing in tiger and safari? is this the case?

    What problem are you having with your browser operations? Can you provide a url where this problem occurs?
    I can unselect the "send url only with UTF-8" in perference of IE.
    Where is this in the preferences exactly?

  • From ASCII txt file to MS SQL with UTF-8 Korean

    I am trying to convert korean text
    ([Àç¹Ì] ¿·Áý
    ºÎÀÎÀÌ
    ¾Æ³»º¸´Ù
    ÁÁÀºÀÌÀ¯? ) which is saved
    in a tab delimited text file into MS SQL with UTF-8/Unicode
    characters ([생각] 인생을
    즐길 수 있는 좋은
    나이는? ).
    I have tried ascii_utf8 and CF_charsetConvert 1.1 and it
    didn't work.
    Is there any other CF tags that can do the jobs?

    atomic wrote:
    > I have tried it but im getting CFMX error when using
    > <cffile action = "read" file = "myPath\myPath.txt"
    variable = "myVar" charset
    > = "euc-kr">
    not sure about cf6, get the charset cfc from here:
    http://www.sustainablegis.com/projects/i18n/charsetTB.cfm
    and see what your system supports. also try one of the others
    like
    x-windows-949, etc.
    >
    > sun.io.ByteToCharEUC_KR.getIndex1()[S
    >
    > Did I miss out on any ColdFusion Administration
    settings? Maybe JVM?
    no, either that JRE doesn't support EUC-KR (kind of unlikely)
    or the JRE isn't
    the international one with the "extra" charsets (my guess).
    if you can update
    the JRE (for cf6-7, it has to be 1.4.x family), try that.
    make sure you get the
    server version & international one (though i think that's
    all they d/l these days).

Maybe you are looking for