File System Character Set Converter

We are currently trying to understand the impact to our custom code in changing character sets. There is a utility(option under adadmin(Patch 1508473), that allows you to convert files in APPL_TOP. Does anyone know if this utility can be used on any directory?
Thanks,
Tim

Thank you...
What is the database version?We are using v8i
What is the name of the utility and more details on where it can be found?"It is called the File System Character Set Converter"
You can it discribed and where it is used in a paper titled:
"Migrating an Applications Installation to a New Character Set" dated 11/28/2001 on Meta Link
Is this iFS?No, from what I have found it is a utility under adadmin. It is installed via a patch #1508473.
My question is does it automatically default to scanning files under APPL_TOP? or can I point it at another directory to scan.
regards,
Tim

Similar Messages

  • JDBC and character set convert issue

    Hi All,
    For some special Chinese characters, we will have the "ORA-00933: SQL command not properly ended" error when we try to insert the data into a Oracle ASCII DB. I have been tried to use the "HasQuot='no'" in the tag and the "N'xxxx'" in the data field but we still got the error.
    Do anyone have the idea about it?
    Best Regard

    It is possilbe to store Big5 code in ASCII DB since it is one kind of ASCII code too.
    If the locale or code page on client is Big5 or 950(MS windows) then the data can be shown properly.
    We tried many ways to store Big5 code in ASCII DB by using PI in these days but failed.
    The following are the things that I have tried.
    1. enable "hasQuot" with "no" or "No" constant in attribute; set the data as format N'xxxx' : still got ORA-00933.
    2. enable AF_Modules/XMLAnonymizerBean befer and after CallSapAdapter module: for some special characters it still has )RA-00933; even we insert the data into DB successfully, we can not process the response properly.
    3. enable AF_Modules/TextCodepageConversionBean before CallSapAdapter module: we can insert the special characters into DB if we set the parameter of Conversion.charset to ASCII. However, the Chinese character will become to "?" in DB.
    I will keep trying,
    Thanks all of your help.

  • Import error converting character set

    Hi there, the first time i ever post something, so i hope i give enough information so my question can be answered.
    I have an oracle 8.1.6 database on Windows NT4. On the same machine i have installed Designer 6.
    Everything is working just fine i just can't import a dump file made from repository 6. It's not a full dump, so i first created a whole new repository.
    Now, when i want to import the dump file in command prompt, using the imp.exe (present in C:\<oracle_home>\bin), the next error occurs:
    IMP-00069: Could not convert to environment national character set's handle
    IMP-00000: Import terminated unsuccessfully.
    As the import is due to start, it gives the following message:
    import done in WE8iSO8859P1 character set and WE8IS08859P1 NCHAR character set.
    Then the error occurs.
    Now i wonder if this can be solved. I cannot see in the dump file what character set was used there. I did not make the export myself. Is there a way that i can convert the character set in the dumpfile or something? Or should i solve this problem by making changes in the database?
    Thx for your help, i am trying to solve this problem myself for over a week now and i'm getting tired of it.

    Hi friend,
    Same Problem i have faced, and i found that this is due to only the Version mismatch, u attempting to import data from higher version to lower one.
    Just check at ur end also.
    Bye
    Tehzeeb Ahmed

  • Character set conversion UTF-8 -- ISO-8859-1 generates question mark (?)

    I'm trying to convert an XML-file in UTF-8 format to another file with character set ISO-8859-1.
    My problem is that the ISO-8859-1 file generates a question mark (?) and puts it as a prefix in the file.
    ?<?xml version="1.0" encoding="UTF-8"?>
    <ns0:messagetype xmlns:ns0="urn:olof">
    <underkat>testv���rde</underkat>
    </ns0:messagetype>
    Is there a way to do the conversion without getting the question mark?
    My code looks as follows:
    public class ConvertEncoding {
         public static void main(String[] args) {
              String from = "UTF-8", to = "ISO-8859-1";
              String infile = "C:\\temp\\infile.xml", outfile = "C:\\temp\\outfile.xml";
              try {
                   convert(infile, outfile, from, to);
              } catch (Exception e) {
                   System.out.println(e.getMessage());
                   System.exit(1);
         private static void convert(String infile, String outfile,
                                            String from, String to)
                             throws IOException, UnsupportedEncodingException
              //Set up byte streams
              InputStream in = null;
              OutputStream out = null;
              if(infile != null) {
                   in = new FileInputStream(infile);
              if(outfile != null) {
                   out = new FileOutputStream(outfile);
              //Set up character streams
              Reader r = new BufferedReader(new InputStreamReader(in, from));
              Writer w = new BufferedWriter(new OutputStreamWriter(out, to));
              /*Copy characters from input to output.
               * The InputSreamreader converts
               * from Unicode to the output encoding.
               * Characters that cannot be represented in
               * the output encoding are output as '?'
              char[] buffer = new char[4096];
              int len;
              while((len = r.read(buffer))!= -1) { //Read a block of output
                   w.write(buffer, 0, len);
              r.close();
              w.flush();
              w.close();
    }

    Yes the next character is the '<'
    The file that I read from is generated by an integration platform. I send a plain file to it (supposedly in UTF-8 encoding) and it returns another file (in between I call my java class that converts the characterset from UTF-8 to ISO-8859-1). The file that I get back contains the '���' if the conversion doesn't work and '?' if the conversion worked.
    My solution so far is to skip the first "junk-characters" when reading from the inputstream. Something like:
    private static final char UTF_BOM = '\uFEFF'; //UTF-BOM = ?
    String from = "UTF-8", to = "ISO-8859-1";
    if (from != null && from.toLowerCase().startsWith("utf-")) { //Are we reading an UTF encoded file?
    /*Read first character of the UTF-Encoded file
    It will return '?' in the first position if we are dealing with UTF encoding If ? is returned we skip this character in the read
    try {
    r.mark(1); //Only allow to read one char for the reset function to work
    char c;
    int i = r.read();
    c = (char) i;
    if (String.valueOf(UTF_BOM).equalsIgnoreCase(String.valueOf(c))) {
    r.reset(); //reset to start position
    r.skip(1); //Skip first character when reading from the stream
    else {
    r.reset();
    } catch (IOException e) {
    e.getMessage();
    //return null;
    }

  • APPLICATION TOP Character set conversion

    Hi,
    I am changing the characterset of application from UTF8 to AR8MSWIN1256.
    After running adadmin, i am getting the following errors in
    manifest_lossy.lst file
    Source Character Set: UTF8 Destination Character Set: AR8MSWIN1256
    Original Source Buffer is:+---------------------------------------------------------------------------+
    Application Object Library: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    module:
    Current system time is 12-APR-2011 17:52:10
    APP-FND-01630: Cannot open file pE<87> for reading
    Cause: wffpath encountered an error when attempting to open file pE<87> for reading.
    Action: Verify that the filename is correct and that the environment variables controlling that filename are correct.
    Action: If the file is opened in read mode, check that the file exists. Check that you have privileges to read the file in the file directory. Contact your system administrator to obtain read privileges.
    Action: If the file is opened in write or append mode, check that you have privileges to create and write files in the file directory. Contact your system administrator to obtain create and write privileges.
    Concurrent request completed successfully
    Current system time is 12-APR-2011 17:52:18
    Round-trip Source Buffer is:+---------------------------------------------------------------------------+
    Application Object Library: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    thanks

    Please post the details of the application release, database version and OS.
    I am changing the characterset of application from UTF8 to AR8MSWIN1256.How did you change it? Please post the steps.
    Any reason for this conversion?
    APP-FND-01630: Cannot open file pE<87> for readingDo you get this error with all concurrent requests or specific ones only?
    Was the concurrent request output file generated successfully?
    APP-FND-1630 Error For All Concurrent Requests that access the database. [ID 554377.1]
    Thanks,
    Hussein

  • Db character set, client character set, OS character set

    Hi,
    I'm confused about all these character set: DB character set, National character set, Client character set, Operating system character set, NLS_LANGUAGE parameter in init.ora file, NLS_LANG parameter in registry.
    Can anyone explain the differences of these to me?
    Many thanks.

    Hi 101514,
    The answers to your questions are listed here in the Globalization Support FAQ on OTN.
    http://technet.oracle.com/tech/globalization/content.html
    Regards
    Nat

  • SOLVED: kernel loads, but doesn't have a root file system

    Hi,
    The system is an Asus X202E. It does UEFI and has a GPT partition system. I've gotten through that part. And it is clear to me that the kernel loads.
    It's the next step that's giving me grief. I've tried this with two bootloaders: gummiboot and rEFInd.
    With gummiboot, the kernel panics because it can't mount the root file system. With rEFInd, it gets to the intial ramdisk and then drops me to a shell, apparently because the root file system is set to null, and it obviously can't mount that as "real root".
    Here is what I posted on the Arch mailing list, documenting that I have indeed specified the correct root (I'm copying this from the email, eliding the unfortunate line wraps):
    bridge-live# cat /boot/loader/entries/arch.conf
    Title Arch Linux
    linux /vmlinuz-linux
    initrc /initramfs-linux.img
    options root=PARTUUID=d5bb2ad1-9e7d-4c75-b9b6-04865dd77782
    bridge-live# ls -l /dev/disk/by-partuuid
    total 0
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 0ab4d458-cd09-4bfb-a447-5f5fa66332e2 -> ../../sda6
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 3e12caeb-1424-451c-898e-a4ff05eab48d -> ../../sda7
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 432a977b-f26d-4e75-b9ee-bf610ee6f4a4 -> ../../sda3
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 95a1d2c2-393a-4150-bbd2-d8e7179e7f8a -> ../../sda2
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 a4b797d9-0868-4bd1-a92d-f244639039f5 -> ../../sda4
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 d5bb2ad1-9e7d-4c75-b9b6-04865dd77782 -> ../../sda8
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 ed04135b-bd79-4c7c-b3b5-b0f9c2fe6826 -> ../../sda1
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 f64f82a7-8f2b-4748-88b1-7b0c61e71c70 -> ../../sda5
    The root partition is supposed to be /dev/sda8, that is:
    lrwxrwxrwx 1 root root 10 Apr 15 19:26 d5bb2ad1-9e7d-4c75-b9b6-04865dd77782 -> ../../sda8
    So the correct PARTUUID followed by the one I have specified in
    arch.conf is:
    d5bb2ad1-9e7d-4c75-b9b6-04865dd77782
    d5bb2ad1-9e7d-4c75-b9b6-04865dd77782
    I'm guessing that this is really the same problem with both gummiboot and with rEFInd, but don't really know. It's clear to me that the initrd is not being correctly constructed. So I removed /etc/mkinitcpio.conf and did, as per the Arch wiki,
    pacman -Syyu mkinitcpio linux udev
    No joy.
    I don't even know which way to go at this point. If I even knew how to tell it where the real disk is in the initial ram disk shell, that would help. Better of course, would be actually solving the problem.
    Thanks!
    Last edited by n4rky (2013-04-17 21:41:36)

    I have made extremely limited progress on this issue.
    My previous attempt to specify the root partition in mkinitcpio.conf was insufficient. Furthermore, this is no place--despite the documentation--for the orthodoxy about using UUIDs rather than the straight /dev/sdx. In my case:
    root=/dev/sda8
    and run
    mkinitcpio -p linux
    It still drops me into the shell at boot. I can do
    mount /dev/sda8 /new_root/
    and exit the shell. It still won't believe it has the root device and drops me back in. I just exit.
    At this point, for a very brief moment, things look promising. It appears to be starting normally. Then, gdm.service, NetworkManager.service, and dbus.service all fail to start. There may be others but the screen goes by too quickly. At this point, it hangs trying to initialize the pacman keyring and all I can do is CTRL-ALT-DEL.
    It occurred to me that this might extend to the rEFInd configuration and so I modified it to also use /dev/sda8 rather than the UUID, but this made no difference. Trying to boot via gummiboot still yields the previously specified kernel panic.

  • Flat File Load Issue - Cannot convert character sets for one or more charac

    We have recently upgraded our production BW system to SPS 17 (BW SP19)
    and we have issues loading flat files (contain Chinese, Japanese and
    Korean characters) which worked fine before the upgrade. The Asian
    languages appear as invalid characters (garbled) as we can see in PSA. The Character Set
    Settings was previously set as Default Setting and it worked fine until
    the upgrade abd we referred to note 1130965 and with the code page 1100
    the load went through (without the Cannot convert character sets for one or more characters error) however the Asian language characters will appear
    as invalid characters. We tried all the code pages suggested in the note e.g.
    4102, 4103 etc on the info packages but it did not work. Note that the
    flat files are encoded in UTF-8.

    I checked lower case option for all IO
    When i checked the PSA failed log no of records processed is "0 of 0" my question is with out processing single record system is througing this error message.
    When i use same file loading from local workstation no error message
    I am thinking when  I FTP the file from AS/400 to BI Application Server(Linex) some invalid characters are adding but how can we track that invalid char?
    Gurus please share your thoughts on this I will assign full points.
    Thanks,

  • Convert custom character set of nonunicode system to standard character set

    hi all,
    I'm a BI consultant and have the following problem:
    we want to load attribute data (products, customers,...) from R3 into BI via standard extractors.
    R3 is a non-unicode system, and some products (or customers) use a custom character set Z2.
    When we extract this data to BI (a unicode system) we get an error (tRFC) saying he can't confert certain values. SAP says that he can't convert character set Z2 to the standard character set 4102.
    Converting the R3 system to a unicode system would be the solution,but is not an option... unfortunatly...
    Please advice, give gidelines,... because this problem is really a showstopper for our project...
    Thx!!!
    Joke

    hi joke,
    i am providing some  links pls chk them . I think they will be help full for you.
    www.sap-press.de/katalog/buecher/htmlleseproben/gp/htmlprobID-149 - 69k
    www.sap-press.de/download/dateien/1240/sappress_unicode_in_sap_systems.pdf
    http://209.85.207.104/search?q=cache:y__bQrnsix0J:https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f5130af2-0a01-0010-c092-a6d44eadd153convertcustomcharactersetofnonunicodesystemtostandardcharacterset%2B+abap&hl=en&ct=clnk&cd=7&gl=us
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/efd1dd90-0201-0010-e1b0-8437c998cd79
    www.redbooks.ibm.com/redpapers/pdfs/redp4189.pdf
    thanks
    karthik

  • Changing the character set in RoboHelp 7 file converted from 5

    The character set of the output files for WinHelp defaults to
    utf-8. How can the default be set to windows-1252 per project or
    for all projects in RoboHelp 7?

    This is related to the posting - "RoboHelp 7 (with patch) IE6
    does not display popup field definitions". Somewhere in our
    environment the character set is getting changed to western
    european (ISO) for the pages that do not load initially in IE6/XP
    from the OAS even though the code generated by RoboHelp is utf-8.
    The WebHelp generated in RoboHelp v5 has a characterset of
    windows-1252. None of the pages with the character set of
    windows-1252 have a problem loading (or changing the character
    set). The problem may be on the OAS or the single signon
    layer.

  • Cannot convert character sets for one or more characters

    Hello,
    Issue:
    Source File:   Test.csv
    Source file dropped on application server as csv file via FTP
    Total records in csv file 102396
    I can able to load only 38,000 records after that I am getting error message  " convert character sets for one or more characters"  while loading into PSA .
    If i load same csv file from local workstation I am not getting any error message I can able to load total 102396 records.
    I am guessing while FTP the file into Application server some invalid codes adding?
    But I am FTP 7 files,  out of 7 I am unable to load only 1 file.
    Anybody have faced this kind of problem please share with me I will assign full points
    Thanks

    I checked lower case option for all IO
    When i checked the PSA failed log no of records processed is "0 of 0" my question is with out processing single record system is througing this error message.
    When i use same file loading from local workstation no error message
    I am thinking when  I FTP the file from AS/400 to BI Application Server(Linex) some invalid characters are adding but how can we track that invalid char?
    Gurus please share your thoughts on this I will assign full points.
    Thanks,

  • MP-00038: Could not convert to environment character set's handle

    I wan't import an Oracle 10g database dump into Oracle Express:
    source DB runs on Linux
    NLS_LANG=AMERICAN_AMERICA.WE8DEC
    destination
    Windows:
    NLS_LANG=AMERICAN_AMERICA.WE8MSWIN1252
    X:\>imp system/manager file=edb.dmp full=y
    Import: Release 10.2.0.1.0 - Production on Wed Nov 14 09:25:40 2007
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Productio
    n
    IMP-00038: Could not convert to environment character set's handle
    IMP-00000: Import terminated unsuccessfully
    What can I do ???

    Hi sivaram,
    how exactly did you export from 11g, and how did you try to import? (Commands with parameters would be great, probably also the first lines of each log file)
    And which exact database release did you export from? Which character set does it use?
    -Udo

  • Database character set = UTF-8, but mismatch error on XML file upload

    Dear experts,
    I am having problems trying to upload an XML file into an XMLType table. The Database is 9.2.0.5.0, with the character set details:
    SELECT *
    FROM SYS.PROPS$
    WHERE name like '%CHA%';
    Query results:
    NLS_NCHAR_CHARACTERSET          UTF8     NCHAR Character set
    NLS_SAVED_NCHAR_CS          UTF8
    NLS_NUMERIC_CHARACTERS          .,     Numeric characters
    NLS_CHARACTERSET          UTF8     Character set
    NLS_NCHAR_CONV_EXCP          FALSE     NLS conversion exception
    To upload the XML file into the XMLType table, I am using the command:
    insert into XMLTABLE
    values(xmltype(getClobDocument('ServiceRequest.xml','UTF8')));
    However, I get the error:
    ORA-31011: XML parsing failed
    ORA-19202: Error occurred in XML processing
    LPX-00200: could not convert from encoding UTF-8 to UCS2
    Error at line 1
    ORA-06512: at "SYS.XMLTYPE", line 0
    ORA-06512: at line 1
    Why does it mention UCS2, as can't see that on the Database character set?
    Many thanks for your help,
    Mark

    USC2 is known as AL16UTF16(LE/BE) by Oracle...
    Try using AL32UTF8 as the character set name
    AFAIK The main difference between Oracle's UTF8 and AL32UTF8 character set is that is the UTF8 character set does not support those UTF-8 characteres that require 4 bytes..
    -Mark

  • UTF8 as system default character set

    Hi,
    I'm new to OSX and I moved to a new company few days ago. We work with Tiger workstations and servers, several Linux internet servers and Windows Laptops.
    One of the problems I have is to find a way to set UTF8 as the default character set for the system, not just user level preferences. That would solve many of my other problems.
    There is probably an easy way to do that?!!
    Any tip or trick would be greatly appreciated.
    Thanks.

    One of the problems I have is to find a way to set UTF8 as the default character set for the system, not just user level preferences. That would solve many of my other problems.
    What are your problems exactly? If you give some specifics, someone can probably help.
    Generally speaking Unicode UTF-8 is what OS X uses by default for a lot of things. The exception would be the file system, which uses UTF-16 NFD. There is certainly no way to change that.

  • System wide default character set

    Hi,
    I moved to a new company few days ago where we use Tiger workstations and servers.
    I've been trying to figure a way to set UTF8 as the default character set for a whole system, not just a user preference.
    I suppose there is a way to do that, but I'm afraid I won't have more time to find it.
    Any help is welcome.
    thx fred

    I've been trying to figure a way to set UTF8 as the default character set for a whole system, not just a user preference.
    If you could explain exactly why you think you need to do that, someone might be able to help.
    Generally speaking Unicode UTF-8 is what OS X uses by default for a lot of things. The exception would be the file system, which uses UTF-16 NFD. There is no way to change that.

Maybe you are looking for

  • Slow Performance

    My Macbook Pro runs very slow. I know this could be for many reasons. However I've done all the usual to see if that fixes the problem. HD has 130Gb of space, permissions repaired, cache emptied, software updated, scripts run etc etc. I do notice in

  • MAIN_SHDRUN/PARDIST_SHD were negative in Preprocessing phase of Upgrade

    Hello All, We are doing an upgrade in our landscape and getting in to errors in the preprocessing phase Landscape Details : R/3 4.6C to ECC 6.0 EHP4 with DB2 database in AIX. The error at the phase is MAIN_SHDRUN/PARDIST_SHD Checks after phase MAIN_S

  • Notification not going.

    Hi All, When the review_flag changes to red. i am firing a trigger..trigger getting fired correctly when a status changes. and the trigger will call the workflow process(REVIEW_FLAG_VALIDATE). the process diagram is start-->function-->based on the re

  • HT201210 updating i.o.s.

    I updated my i.o.s. software to the latest and at the end my iphone show the "conect to itunes" image ... now the good part is that my iphone is not starting and can't restore from itunes ... it says can not connect to apple servers. Anybody have a c

  • Possible to "Borrow" QT 1/2 Playback speed CODE ?

    Hello Forum readers - With a QT movie that will be posted on the web - is it possible to: Have a button that when clicked, plays the movie at 1/2 the speed ? (borrows the 1/2 playback speed code from the "AV Controls") Since it's incorporated in the