Problem UTF8 in WebToolKit

Hi,
I have Oracle 8i, database codepage is CL8ISO8859P5. I would like to download Cyrillic symbols from the database into a UTF8 file using Web Toolkit, but when I try to do this I realize that Cyrillic symbols are corrupted in the file. I suppose this problem is because of parameters of the HTP.P procedure - it accepts VARCHAR2 type only, not NVARCHAR2. Could you give me any suggestion how to resolve this problem?
Code procedure:
CREATE OR REPLACE PROCEDURE to_xml
IS
rec t_xml%ROWTYPE;
s_str varchar2(1);
BEGIN
owa_util.mime_header(ccontent_type => 'file',ccharset => 'utf-8');
htp.p('<?xml version="1.0" encoding="utf-8" standalone="no"?>');
htp.p('<records>');
s_str:=chr(193);
htp.comment(convert(s_str,'UTF8'));
htp.p('</records>');
owa_util.http_header_close;
END to_xml;

I am still seeing garbage when I call this against the string:
Here is breakdown:
URL= http://server/file.jsp?k=%E8%88%92%E4%B8%BD%E9%9B%85%E6%B2%99%E5%8F%91&p=1
keyWord = ??????????????? (this system doesn't handle ML either so these look like ???s to you)
URLDecode = ???????????????
URLEncode = %E8%3F%3F%E4%B8%BD%E9%3F%3F%E6%B2%3F%E5%3F%3F
The output from URLEncode is obviously closest to what I want but at this point the data is corrupted from where I initially set the variable. This seems really easy, but something is missing. Any encoding pros out there?
Thanks.

Similar Messages

  • Mpd and python-mpd problem utf8

    Hi,
    i've just update the new mpd and python mpd package, so there is a problem with utf8 directory. Now it can not display my covers.
    The error :
    Exception in thread Thread-28:
    Traceback (most recent call last):
    File "/usr/lib/python2.5/threading.py", line 486, in __bootstrap_inner
    self.run()
    File "/usr/lib/python2.5/threading.py", line 446, in run
    self.__target(*self.__args, **self.__kwargs)
    File "/usr/lib/python2.5/site-packages/sonata/main.py", line 3971, in _artwork_update
    imgfound = self.artwork_check_for_local()
    File "/usr/lib/python2.5/site-packages/sonata/main.py", line 3993, in artwork_check_for_local
    type, filename = self.artwork_get_local_image()
    File "/usr/lib/python2.5/site-packages/sonata/main.py", line 4038, in artwork_get_local_image
    testfile = img.single_image_in_dir(self.musicdir[self.profile_num] + songpath)
    File "/usr/lib/python2.5/site-packages/sonata/img.py", line 62, in single_image_in_dir
    dir = gobject.filename_from_utf8(dir)
    GError: Séquence d'octets non valide en entrée du convertisseur

    I'm having the same output whenever I try to view an artists's albums in the collection and the album title contains non strictly ascii characters (which happens pretty often, as I am not a native english speaker). The problem is that the albums do not appear in the collection ! I can still access them by the filesystem view, but then the songs are sorted alphabetically, not by track number.
    This problem appeared with Sonata 1.5.

  • ORA-01401: inserted value too large for column from 9i to 8i

    Hi All,
    Am trying to get the data from 9.2.0.6.0 to 8.1.7.0.0.
    The character sets in both of them are as follows
    9i
    NLS_NCHAR_CHARACTERSET : AL16UTF16
    NLS_CHARACTERSET : AL32UTF8
    8i
    NLS_NCHAR_CHARACTERSET : UTF8
    NLS_CHARACTERSET : UTF8
    And the structure of the Table in 9i which am trying to pull is as follows.
    SQL> desc xyz
    Name Null? Type
    PANEL_SITE_ID NOT NULL NUMBER(15)
    PANELIST_ID NUMBER
    CHECKSUM VARCHAR2(150)
    CONTACT_PHONE VARCHAR2(100)
    HH_STATUS NUMBER
    HH_STATUS_DT DATE
    HH_RECRUITMENT_PHONE VARCHAR2(100)
    HH_RECRUITMENT_DT DATE
    FIRST_NET_USAGE_DT DATE
    INSTALL_DT DATE
    FNAME VARCHAR2(4000)
    LNAME VARCHAR2(4000)
    EMAIL_ADDRESS VARCHAR2(200)
    EMAIL_VALID NUMBER
    PASSWORD VARCHAR2(4000)
    And by connecting to one of the 8i schema am running the following script
    CREATE TABLE GPMI.GPM_HOUSEHOLDBASE_FRMP AS
    SELECT PANEL_SITE_ID,
    PANELIST_ID,
    LTRIM(RTRIM(CHECKSUM)) CHECKSUM,
    LTRIM(RTRIM(CONTACT_PHONE)) CONTACT_PHONE,
    HH_STATUS, HH_STATUS_DT,
    LTRIM(RTRIM(HH_RECRUITMENT_PHONE)) HH_RECRUITMENT_PHONE,
    HH_RECRUITMENT_DT,
    FIRST_NET_USAGE_DT,
    INSTALL_DT, LTRIM(RTRIM(FNAME)) FNAME,
    LTRIM(RTRIM(LNAME)) LNAME,
    LTRIM(RTRIM(EMAIL_ADDRESS)) EMAIL_ADDRESS,
    EMAIL_VALID,
    PASSWORD
    FROM [email protected];
    Am gettinh the following error.
    Can anyone of you fix this one.
    PASSWORD
    ERROR at line 14:
    ORA-01401: inserted value too large for column
    Thanks in Advance
    Sudarshan

    Additionally I found this matrix, which explains your problem:
    UTF8 (1 to 3 bytes) AL32UTF8 (1 to 4 bytes)
    MIN MAX MIN MAX
    CHAR 2000 666 2000 500
    VARCHAR2 4000 1333 4000 1000 */
    For column PASSWORD the maximum length is used (4000). UTF8 uses maximal 3 bytes for a character, while AL32UTF8 may use up to 4 characters. So a column defined in AL32UTF8 may contain characters, which do not fit in a corresponding UTF8 character.

  • Problem in Database convertion from US7ASCII to UTF8

    Hi,
    We are facing the following problem while converting the database from US7ASCII to UTF8:
    We have recently changed the database character set from US7ASCII to UTF8 for the internationalization
    purpose. We ran the Character set scanner utility and it did report that some data may pose problems.
    We followed the the below mentioned process to convert into UTF8 -
    1) alter database character set utf8
    2) alter database national character set utf8.
    Now we find some problem while working with the old data in our application which is java based.
    We are getting the following error "java.sql.SQLException: Fail to convert between UTF8 and UCS2: failUTF8Conv".
    We further analyzed our data and found some interesting things :
    e.g.
    DB - UTF8.
    NL_LANG also set to UTF8.
    Select name from t1 where name like 'Gen%';
    NAME
    Genhve
    But when we find out the length of the same data it show like this
    NAME LENGTH(NAME) VSIZE(NAME)
    Genhve 4 6
    The question is why is it showing length as 4 only and when we try to use a substr function
    its extracting like the following :-
    select name,substr(name,4,1) from t1 where name like 'Gen%';
    NAME SUB
    Genhve hve
    We have execute the above queries on US7ASCII DB and it is working fine, length it shows 6
    and using SUBSTR it extracts just 'h' as well.
    We also used dump function on the UTF8 Db for the above query,,this is the result :-
    select name,length(name),vsize(name),dump(name) from t1 where name like 'Gen%';
    NAME LENGTH(NAME) VSIZE(NAME) DUMP(NAME)
    Genhve 4 6 Typ=1 Len=6: 71,101,110,232,118,101
    We checked a lot with the data and it seems 'h' (accented e) is posing the problem.
    We want to know where is the problem and how to overcome this.
    Further, we tried all of the following :
    1)
    Export Server: US7ASCII
    Export Client: did not set NLS_LANG / NLS_CHAR, so presumably US7ASCII as well
    Import Client: did not set NLS_LANG / NLS_CHAR, so presumably US7ASCII as well
    Import Server: UTF8
    RESULT: Acute e became h
    2)
    Export Server: US7ASCII
    Export Client: did not set NLS_LANG / NLS_CHAR, so presumably US7ASCII as well
    Import Client: NLS_LANG=AMERICAN_AMERICA.UTF8 and NLS_CHAR=UTF8
    Import Server: UTF8
    RESULT: IMP 00016 error
    3)
    Export Server: US7ASCII
    Export Client: NLS_LANG=AMERICAN_AMERICA.UTF8 and NLS_CHAR=UTF8
    Import Client: did not set NLS_LANG / NLS_CHAR, so presumably US7ASCII as well
    Import Server: UTF8
    RESULT: Acute E became h
    4)
    Export Server: US7ASCII
    Export Client: NLS_LANG=AMERICAN_AMERICA.UTF8 and NLS_CHAR=UTF8
    Import Client: NLS_LANG=AMERICAN_AMERICA.UTF8 and NLS_CHAR=UTF8
    Import Server: UTF8
    RESULT: Acute e became h
    5)
    Tried using Update sys.props$
    set value$='UTF8'
    where name='NLS_CHARACTERSET'
    RESULT: Acute e shows properly but it gives problem in the application
    "java.sql.SQLException: Fail to convert between UTF8 and UCS2: failUTF8Conv"
    Looking further it was observed the following:
    when you try this command on a column 'city' in a table which contains 'Genhva' (note the acute e after n), it shows
    command: select length(city), vsize(city),substr(city,4,1),city from cities
    Result: 4 6 hva Genhva
    if you see the value of substr(city,4,1) , you will see the problem. Also note that the length shows 4 and size shows 6. Moreover, when these records are selected through JDBC Driver, it starts giving problems.
    6)
    Actually the above (point no. 5) is similar to changing the character set of the database with 'ALTER DATABASE CHARACTER SET UTF8'. Same problem is observed then too.
    7)
    We have also tried to with another method, that is by changing the third byte of the export file which specifies the character set, to the UTF8 code by editing the export file with a Hexdecimal editor. After import the same problem has been observed as defined in (5) and (6) above.
    We have no more ideas how to migrate without corrupting the data. Of course we have known the records where these characters occur through the Oracle's cssacn utility but we do not want to manually rectify each and every record by replacing it with an ASCII character. Any other idea as to how this can be accomplised?
    Thanx
    Ashok

    The problem you have is that although your original database is defined as US7ASCII, the data it contains is more than is covered by this code page (as the reply on Sept 4 to the previous posting already said).
    This has probably happened because the client was also defined as US7ASCII, and when the DB and client are defined as having the same character set no conversion (or checdking) takes place when data is passed between them. However if you are using a Windows client then it will in fact be using Windows code page 1252 (Latin-1) or similar, and this allows many more characters, including h (accented e). So a user can enter all these characters and store them in the database, and similarly read them from the database, because data transfer is transparent.
    When you did ALTER DATABASE CHARACTER SET UTF8 this will only change the label on the database, but not affect the contents. However only part of the contents are valid UTF8, any character above 7F (like h) is invalid. If your original client now uses the database, code page transformation will take place because the client and DB have different character sets defined. The invalid codes can then cause problems.
    Without being able to explain what has happened in detail, it may help to see what your h (dec 232, x'E8') looks like. The actual data has not changed (you can see this as it is reported as 232). However the binary code there (11101000) is invalid UTF8. UTF8 encodes a character in 1 to 4 bytes, and the first bits in a UTF8 character tell how many bytes it uses. 0xxx tell it is one byte (same as the corresponding USASCII character), 110x that it uses 2 bytes, 1110 that it uses 3 bytes etc. So if you interpret what is there as UTF8 it looks like the first byte of a 3-byte character, which explains why the substringing is giving you the other 2 bytes as well.
    Can you fix this without losing data? I believe yes. First you should check what other characters are being flagged by the scan. See if these are contained in another standard character set. If they are only Western European accentet characters then WE8ISO8859P1 is probably ok, but watch out for the euro sign which Windows has at x'80', an undefined character in ISO8859-1.
    You can see the contents of the Microsoft Windows Codepage 1252 at: http://www.microsoft.com/globaldev/reference/sbcs/1252.htm
    For a listing of the US-ASCII defined characters see http://czyborra.com/charsets/iso646.html and for ISO 8859-1 see http://czyborra.com/charsets/iso8859.html#ISO-8859-1
    If all is well, you can first ALTER DATABASE CHARACTER SET to WE8ISO8859P1. This will not change any data, but ensure that all the data gets exported. Then export the DB and import it to a UTF8 DB. This will convert the non-US-ASCII characters to Unicode. You will also have to change the clients character set to something other than USASCII or they will just see ? for the other characters.
    Good Luck!

  • Problems with Accented Characters after migrating to UTF8

    Hi,
    I am puzzled about the following problem.
    We have recently changed the database character set from US7ASCII to UTF8 for the internationalization
    purpose. We ran the Character set scanner utility and it did report that some data may pose problems.
    We followed the the below mentioned process to convert into UTF8 -
    1) alter database character set utf8
    2) alter database national character set utf8.
    Now we find some problem while working with the old data in our application which is java based.
    We are getting the following error "java.sql.SQLException: Fail to convert between UTF8 and UCS2: failUTF8Conv".
    We further analyzed our data and found some interesting things :
    e.g.
    DB - UTF8.
    NL_LANG also set to UTF8.
    Select name from t1 where name like 'Gen%';
    NAME
    Genhve
    But when we find out the length of the same data it show like this
    NAME          LENGTH(NAME)          VSIZE(NAME)
    Genhve 4 6
    My question is why is it showing length as 4 only and when we try to use a substr function
    its extracting like the following :-
    select name,substr(name,4,1) from t1 where name like 'Gen%';
    NAME               SUB
    Genhve          hve
    We have execute the above queries on US7ASCII DB and it is working fine, length it shows 6
    and using SUBSTR it extracts just 'h' as well.
    I also used dump function on the UTF8 Db for the above query,,this is the result :-
    select name,length(name),vsize(name),dump(name) from t1 where name like 'Gen%';
    NAME          LENGTH(NAME)          VSIZE(NAME)               DUMP(NAME)
    Genhve 4 6               Typ=1 Len=6: 71,101,110,232,118,101
    I checked a lot with the data and it seems 'h' (accented e) is posing the problem.
    I want to know where is the problem and how to overcome this.
    Regards
    Siba

    We ran the Character set scanner utility and it did report that some data may pose problems. Did you check out the documentation with the scanner ? It explains why and how to deal with problem data.
    The problems are :-
    1. You had invalid data inside your original US7ASCII database . h is not a valid 7 bit ASCII character. All the invalid US7ASCII characters, ( ie. all those that will be lost after doing ALTER DATABASE CHARACTER SET ) would have been flagged by the scanner.
    2. All non 7 bit ASCII characters need to be converted to their UTF8 counterpart values, and typically these characters will have size expansions too. for example ,from 1 byte to 2 bytes.
    3. ALTER DATABASE CHARACTER SET UTF8 changes the character set tag in the database header only. No data conversions are performed to your existing data .
    Hi,
    I am puzzled about the following problem.
    We have recently changed the database character set from US7ASCII to UTF8 for the internationalization
    purpose. We ran the Character set scanner utility and it did report that some data may pose problems.
    We followed the the below mentioned process to convert into UTF8 -
    1) alter database character set utf8
    2) alter database national character set utf8.
    Now we find some problem while working with the old data in our application which is java based.
    We are getting the following error "java.sql.SQLException: Fail to convert between UTF8 and UCS2: failUTF8Conv".
    We further analyzed our data and found some interesting things :
    e.g.
    DB - UTF8.
    NL_LANG also set to UTF8.
    Select name from t1 where name like 'Gen%';
    NAME
    Genhve
    But when we find out the length of the same data it show like this
    NAME          LENGTH(NAME)          VSIZE(NAME)
    Genhve 4 6
    My question is why is it showing length as 4 only and when we try to use a substr function
    its extracting like the following :-
    select name,substr(name,4,1) from t1 where name like 'Gen%';
    NAME               SUB
    Genhve          hve
    We have execute the above queries on US7ASCII DB and it is working fine, length it shows 6
    and using SUBSTR it extracts just 'h' as well.
    I also used dump function on the UTF8 Db for the above query,,this is the result :-
    select name,length(name),vsize(name),dump(name) from t1 where name like 'Gen%';
    NAME          LENGTH(NAME)          VSIZE(NAME)               DUMP(NAME)
    Genhve 4 6               Typ=1 Len=6: 71,101,110,232,118,101
    I checked a lot with the data and it seems 'h' (accented e) is posing the problem.
    I want to know where is the problem and how to overcome this.
    Regards
    Siba

  • LoadBundle with utf8 problem

    hi everyone , i have problem with f:loadBundle , i saved properties file with utf8 encoding but can not load string in it correctly (aisa characters) , help me please !

    At a guess - loadBundle uses property file syntax.
    Have a look at the API doco for java.util.Properties .
    It states that non8859 characters in property files have to be converted to unicode escapes, and that the native2ascii tool that comes with the JDK will do it for you:
    http://java.sun.com/javase/6/docs/technotes/tools/windows/native2ascii.html
    If you are using ANT to build your project, it may make sense to keep your original properties file in UTF-8 and edit it with your favourite editor, and to convert it as part of the build process.

  • [SOLVED] xfce4 automounting problem on usb dev with utf8 encoded files

    I have problem of auto mounting usb external drive contains utf8 encoded files under xfce4.
    Using hal and dbus for usb device detection.
    Heard that I need to add iocharset=utf8 under fstab or hal fdi configure files and tried them all but didn't work
    usb devices with vfat always got mounted with default charset.
    I found out that exo-mount is the one doing the actual mounting and that it is the one that cannot take all extra mounting options such as iocharset etc.
    Have tried gnome to mount this hard drive and works find with nautilus option that let me set various mount options for each harddrive.
    Is there any way to replace exo-mount with some other alternative mounting tools that work with thunar?
    Any workaround will be helpful as long as I can mount my usb drives with prefer charset so that I don't have to guess all ???marked files.
    Heard that xubuntu replaced that exo-mount with gnome-mount to handle drives with extra mount options such as iocharset.
    but I am pretty new to configuring things and couldn't find the step by step instruction for this so please provide the complete solution.
    Link to the pages that explain workaround would be nice as well.
    Thanks in advance!
    Last edited by axca (2007-10-17 18:24:50)

    Found the patch for exo-mount so that it takes additional mount options HAL suggests.
    https://bugzilla.altlinux.org/attachmen … ction=diff
    Now my usb drive can be mounted with the proper charset.
    Hope this patch helps others with similar situation as well!
    Last edited by axca (2007-10-15 15:41:47)

  • Problem with "locked records" on Oracle with UTF8

    Hello,
    Can somebody help me with "record is locked by another user" problem on Oracle 10.1.0.2 ?
    Oracle is installed with UTF8 character set, and is running on Windows Server 2003. Whenever I try to edit numeric column I get the message "record is locked by another user". Selecting locks from v_$locked_object view, gets empty resultset - nothing is locked.
    The problem occurs only when I edit numeric columns using client applications like TOAD or my own applications. When I do the same column update by sql query, everything works fine.
    Is the problem connected with client configuration or maybe with UTF8 character set, or Windows Server2003?
    We also have second Oracle 10.1 server on Linux, but with WIN1250 character set, and there's no problems like that.
    Thanks in advance for any help,
    remg

    About the only thing you can do is automatically commit or rollback when the user presses the Enter Query key/button. In your key-entqry trigger, you would need to code this:
    Clear_form(no_validate,full_rollback);
    Enter_Query;

  • UTF8 and Traditional chinese display problem

    I have having problems displaying chinese because I am trying to pull data from 2 difference sources into a 3rd source.
    Question 1
    I am using Oracle 8.17 on Sun Solaris - (in China).
    The NLS_LANG on that machine (Machine name is CNCIM) is American. The NLS_CHARACTERSET is ZHS16GBK.
    We are using Java to write data into this server. The data has been stored as UTF8.
    When I use TOAD, to look at the chinese characters, I cannot see the actual characters.
    What I see in TOAD is something like this
    &#19978;&#28023;&#24066;&#23665;&#19996;&#20013;&#36335;145&#21495;
    Convert function does not seem to work. Can someone advice me how I can see the raw data in chinese?
    Question 2
    Currently, I am also trying to import the above UTF8 data into my Local Oracle Machine (APDM)
    The NLS_LANG on that machine is American. The NLS_CHARACTERSET is WE8ISO8859P1.
    And at the sametime, I am trying to import information from SQL2000 into Local Oracle Machine (APDM).
    I was told that the data in SQL2000 has been setup as tradtional Chinese.
    What type of settings should I have on my local machine(APDM) to allow me to view Chinese data (From CNCIM - UTF8) and to see the SQL2000(traditional chinese) data at the same time? Should this settings
    be configured for my client too?

    My hunch is that Toad either does not support the display of Chinese characters, at least not without setting your Windows regional settings to Chinese. Can you retrieve the data via your Java application?
    If your database character set is ZHS16GBK, that should be the character set that the data is stored in the database. If your Java application is sending over UTF-8 encoded data, you need to ensure that the NLS_LANG settings on the client machine identify the data as UTF-8 so that Oracle can do the necessary character set conversion.
    If your local database is using the ISO 8859-1 character set, it cannot store Chinese data. You would need to create a new database with a character set that supports Chinese if you wanted to store Chinese data there.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • UTF8 problem with XMLGEN

    Hi All,
    I'm having problems with the XML generation. The problem is that when we try to generate XML using a query, such as:
    select xmlgen.getXML('select content_id, locale_id, content from bc_content_local where delete_p = ''0''',1) from dual;
    It works fine so long as of the information retrieved by the query is ASCII. But if there are UTF8 characters in there, the
    generation of the XML fails.
    I wasn't actually getting an error message. It would work and return XML but as soon as the first UTF-8 character was encountered,
    the XML just stopped.
    Two questions:
    1) Does the XML generation support UTF-8? 2)
    If it does (I seem to remember hearing that it did) what has to be done so that this
    works correctly?
    We are using a UTF8 characterset we set during database creation. I have also tried setting up NLS_LANG in the client registrty to american_america.utf8 - but in both client and server I get teh same message.
    Database: Oracle 8.1.7 Standard Edition on Sun-Solaris 2.8
    Any help greatly appreciated,
    Nilendu Misra
    [email protected]

    Ok, I found we have DBMS_XMLQUERY installed on the schema other than SYS too. But when I issue the same statement using this package I get a different error:
    SQL> select dbms_xmlquery.getXML('select content_id, locale_id, content
    2 from bc_content_local where delete_p = ''0''',1) from dual;
    select dbms_xmlquery.getXML('select content_id, locale_id, content
    ERROR at line 1:
    ORA-29532: Java call terminated by uncaught Java exception:
    java.lang.NullPointerException
    ORA-06512: at "SYSTEM.DBMS_XMLQUERY", line 212
    ORA-06512: at "SYSTEM.DBMS_XMLQUERY", line 220
    ORA-06512: at "SYSTEM.DBMS_XMLQUERY", line 210
    ORA-06512: at "SYSTEM.DBMS_XMLQUERY", line 228
    ORA-06512: at line 1
    I tried searching the XML forum and Metalink but I found most of the reports on this nature of problem are still unanswered.
    Thanks in appreciation,
    Nilendu
    null

  • JSF 1.1 - problem with overriding default messages with UTF8 ones

    Hi there,
    I'm stuck with JSF 1.1 for now (creating application that is to run on JBoss 4.0.x) and I've encountered a problem... My web application uses UTF8 encoding - so that's also encoding I need in my message bundle. I've created one - and here goes a problem. I need to put Polish special characters into a properties file of a bundle - so I do it with UTF escape notation:
    javax.faces.component.UIInput.REQUIRED=Warto\u009C\u00E6 dla pola {0} jest wymagana
    some.custom.message.CLOSED=zamkni\u0119taThe thing is - while custom messages are working properly, overriden javax.faces.component.UIInput.REQUIRED message is rendered on a page as four "strange" characters, suggesting that it was first interpreted as UTF, then parsed as some kind of an ASCII (ISO) encoding and then back to an UTF8 page...
    How to fix this?
    TIA,
    Wojtek

    IDE fooled me - it was smarter then I expected and properties editor was showing already decoded values, not raw content of file. It's working properly now

  • Problem with use utf8 words in xml

    hi . today many components like Drop down menus and galleries and ...  use dynamic data importin ( xml ) files ...
    after i edit xml and put some utf8 words an save , when i test my swf i see nothing show in rows that i replace english data with other langauge ...
    on some others i see unknown characters !!
    whats the problem ?
    the Xml file have :
    <?xml version="1.0" encoding="UTF-8"?>
    in header so why Flash cant act like Dreamweaver or webpage and ...
    what setting i must change in my xml or Flash to enable use type unicode characters and langauges in xml files ?
    thanks ...

    thanks for replay ...
    what do you mean exactly ?  copy my xml code here ?  or use copy paste for xml ?!
    For example this is one of my Xml :
    <?xml version="1.0" encoding="utf-8"?>
    <menu>
        <menuItem title="منوی 1">
            <subItem title="منو 2 " />
            <subItem title="sub 2" >
                <subSubItem title="sub sub 1"  file="flv.flv" type="video"/>
                <subSubItem title="sub sub 2" />
                <subSubItem title="sub sub 3" />
            </subItem>
        <subItem title="sub 3" />
        </menuItem>
        <menuItem title="درباره ما">
        <subItem title="Sub 1" >
        <subSubItem title="sub sub 1"  file="flv.flv" type="video"/>
                <subSubItem title="sub sub 2" />
                <subSubItem title="sub sub 3" />
        </subItem>
            <subItem title="sub 2" >
                <subSubItem title="sub sub 1"  file="flv.flv" type="video"/>
                <subSubItem title="sub sub 2" />
                <subSubItem title="sub sub 3" />
            </subItem>
        <subItem title="sub 3" />
        </menuItem>
            <menuItem title="Menu 3">
        <subItem title="Sub 1" >
        <subSubItem title="sub sub 1"  file="flv.flv" type="video"/>
                <subSubItem title="sub sub 2" />
                <subSubItem title="sub sub 3" />
        </subItem>
            <subItem title="sub 2" >
                <subSubItem title="sub sub 1"  file="flv.flv" type="video"/>
                <subSubItem title="sub sub 2" />
                <subSubItem title="sub sub 3" />
            </subItem>
        <subItem title="sub 3">
        <subSubItem title="sub sub 1"  file="flv.flv" type="video"/>
                <subSubItem title="sub sub 2" />
                <subSubItem title="sub sub 3" />
        </subItem>
        </menuItem>
    </menu>
    for first and secound menu i can only see the numbers after save xml file

  • JDeveloper3.1.1.2 problem:convert from UTF8 to UCS2 failed

    oracleTeam:
    i test JDeveloper3.1.1.2,it has problem in runtime:convertion from UTF8 to UCS2 failed ,AttributeLoadException.( our language is chinese)
    I found that oracle\jbo\server\QueryCollection.class in dacf.zip maybe has problem,i use this class of JDeveloper3.1 to repalce same_name class in JDeveloper3.1.1.2,above problem disappeared,
    but because this class is not suit of JDeveloper3.1.1.2,other problems appeared.
    so you should work out this problem ,i hope
    it runs correctly.

    I searched this forum and the SQLJ/JDBC forum, and found a few occurrences of this problem. Among the things people suggested:
    * Changing JDBC drivers (experience varied as to which one fixed the problem)
    * Adding nls_charset1x.zip to your CLASSPATH
    * Ensuring you're using the same character set on the client and server.
    I suggest you take a look at the following discussion threads: http://technet.oracle.com:89/ubb/Forum8/HTML/001810.html http://technet.oracle.com:89/ubb/Forum8/HTML/000065.html http://technet.oracle.com:89/ubb/Forum2/HTML/000820.html
    Blaise
    null

  • UTF8 character problems

    I am working in a pan-european organisation and have encountered a problem with printing a special character from the Swedish alphabet. It shows up in the OPP log with an Invalid UTF8 encoding message. Does anyone have any suggestions on how I could get round this?

    Hi Tim,
    thanks for the reply.
    The output does not even get generated - it fails before that stage.
    I am using reports 6i to extract the data - basically I am replacing the layout.
    top of the XML is:
    <?xml version="1.0"?>
    <!-- Generated by Oracle Reports version 6.0.8.27.0 -->
    <CBASCINV_V5>
    <LIST_G_F_DEL_DETAIL_ID>
    and the offending piece if text with the character in it is:
    Supplier: KAR Sj÷ding
    The offending character is whatever is represented by the division sign.
    If I try to load the file into Word via the data import, it throws out an error message saying 'an invalid character was found in texy content' and highlighting this character.
    This does not apply to all Swedish characters - this <C_FROM_ADDRESS_2>Konstruktionsvägen 14</C_FROM_ADDRESS_2>
    prints without any problems!
    Thanks again
    Mike

  • UTF8 characterset problem

    I have a database that has been installed with the UTF8 characterset.
    Therefore the £ sign is ASCII 49827 in this characterset and not ASCII 163 as used in ISO8859P1. I believe my problem stems from me using UTF complaint clients to enter the £ sign into the database. Because the Clients are UTF8 compaint they are passing ASCII 49827 for the £ sign into the database. However when non UTF8 compliants try to view or manipulate the data from this database, instead of them seeing the £ sign they see £.
    I presume the solution should be to ensure all viewing clients are UTF8 compliant ?
    My problem is that I have a 3rd party program that is taking the output from the database in the form of a text file and is then converting the output into a PDF. Hence in the PDF the £ sign is appearing as £, since this is what is effectively stored in the database / text file output
    I presume it never makes sense to try and convert the Database using CSALTER to WE8ISO8859P1 from UTF8 ? ( as WE8ISO8859P1 is a sub set of UTF8 ? )
    The routine that is creating the test file from the database is an E-Business Suite routine, so I cannot interfer with it and put in a translate command.
    Any suggestions ?
    thanks,
    Jim

    what is NLS_CHARACTERSET value of your database?
    I always use AR8MISO8859P6 and it's perfect for Arabic language.
    For your Application Server, in the registry, check
    NLS_LANG = AMERICAN_AMERICA.AR8MSWIN1256
    REPORT_ARABIC_NUMERAL = ARABIC
    REPORTS_PATH = (Make sure your font path is included here)
    Add these entries in your uifont.ali file under [ PDF:Subset]
    [ PDF:Subset ]
    Arial..Italic.Bold.. = "Arialbi.ttf"
    Arial..Italic... = "Ariali.ttf"
    Arial...Bold.. = "Arialbd.ttf"
    Arial..... = "Arial.ttf"
    "Arabic Transparent"..Italic.Bold.. = "ARIALBI.TTF".....
    "Arabic Transparent"..Italic... = "ARIALI.TTF".....
    "Arabic Transparent"...Bold.. = "Arialbd.ttf"
    "Arabic Transparent"..... = "artro.ttf"
    Tahoma = "tahoma.ttf".....ar8mswin1256
    Arial = "Tahoma"
    "Tahoma Bold" = "tahomabd.ttf"
    Make sure Regional and Language option in control panel is properly configured for Arabic.
    restart the Windows Server 2003 and arabic should work properly on Server 2003 also.
    I'm using Oracle Application Server 10g on Windows 2003 Server and Arabic Language is working perfectly for forms as well as PDF Reports.
    I'm using Tahoma font for Arabic.
    Hope this will help and solve your problem.
    Note: Make sure that the font which you are using for Arabic in your reports is included in the uifont.ali
    regards,
    Saadat Ahmad
    Message was edited by:
    saadatahmad

Maybe you are looking for

  • Can the Terminal add a user(with all options) on a single line?

    Howdy All, Can the Mac Terminal add a user, with all options desired, on a single line?    If so, can I get an example of this single terminal line to work from? Thanks

  • Why aren't my changes to a .pdf not saving?

    I am running Acrobat Pro 9 for Mac (9.5.4).  I received an unlocked, not signed .pdf brocure.  I made some edits (rotating pages, inserting arrow and comments, etc), "save[d] As" giving it a new name, and closed the file.  Later, I re-opened it, and

  • PeopleTools 8.49 PeopleSoft Database Cannot Connect with MS SQL Server 2005

    Folks, Hello. I am working on PeopleSoft PIA. My system is Windows Server 2003, MS SQL Server Express 2005 Standard, PeopleTools 8.49. I am setting up PeopleSoft Database \PsMpDbInstall\setup.exe and got the errors as follows: Error: Setup fails to c

  • How do I fix "missing plugin'?

    I seem to be getting a lot of these lately and one today in an email.  How do I find my missing plugin? 

  • Adobe Reader upgrade

    We upgraded Adobe Reader to the latest version, but now it is prompting to install a new verison of flash player, which we are not yet prepared to do.  Is this because of the format of the PDF?  Most other PDFs open fine.