Server uses WE8ISO8859P15 character set (possible charset conversion)

Hi,
when EXP in 9i I receive :
Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
Export done in WE8PC850 character set and AL16UTF16 NCHAR character set
server uses WE8ISO8859P15 character set (possible charset conversion)What is the problem ?
Thank you.
I exported just a table, how to see if it is exported ?

Dear user522961,
You have not defined or misdefined the NLS_LANG environmental variable before trying to run the export command.
Here is a little illustration;
*$ echo $NLS_LANG*
*AMERICAN_AMERICA.WE8ISO8859P9*
$ exp system/password@opttest file=ogan.dmp owner=OGAN
Export: Release 10.2.0.4.0 - Production on Mon Jul 12 18:10:47 2010
Copyright (c) 1982, 2007, Oracle.  All rights reserved.
Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
*Export done in WE8ISO8859P9 character set and AL16UTF16 NCHAR character set*
About to export specified users ...
. exporting pre-schema procedural objects and actions
. exporting foreign function library names for user OGAN
. exporting PUBLIC type synonyms
. exporting private type synonyms
. exporting object type definitions for user OGAN
About to export OGAN's objects ...
. exporting database links
. exporting sequence numbers
. exporting cluster definitions
. about to export OGAN's tables via Conventional Path ...
. exporting synonyms
. exporting views
. exporting stored procedures
. exporting operators
. exporting referential integrity constraints
. exporting triggers
. exporting indextypes
. exporting bitmap, functional and extensible indexes
. exporting posttables actions
. exporting materialized views
. exporting snapshot logs
. exporting job queues
. exporting refresh groups and children
. exporting dimensions
. exporting post-schema procedural objects and actions
. exporting statistics
Export terminated successfully without warnings.
*$ export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P15*
$ exp system/password@opttest file=ogan.dmp owner=OGAN
Export: Release 10.2.0.4.0 - Production on Mon Jul 12 18:12:41 2010
Copyright (c) 1982, 2007, Oracle.  All rights reserved.
Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
*Export done in WE8ISO8859P15 character set and AL16UTF16 NCHAR character set*
*server uses WE8ISO8859P9 character set (possible charset conversion)*
About to export specified users ...
. exporting pre-schema procedural objects and actions
. exporting foreign function library names for user OGAN
. exporting PUBLIC type synonyms
. exporting private type synonyms
. exporting object type definitions for user OGAN
About to export OGAN's objects ...
. exporting database links
. exporting sequence numbers
. exporting cluster definitions
. about to export OGAN's tables via Conventional Path ...
. exporting synonyms
. exporting views
. exporting stored procedures
. exporting operators
. exporting referential integrity constraints
. exporting triggers
. exporting indextypes
. exporting bitmap, functional and extensible indexes
. exporting posttables actions
. exporting materialized views
. exporting snapshot logs
. exporting job queues
. exporting refresh groups and children
. exporting dimensions
. exporting post-schema procedural objects and actions
. exporting statistics
Export terminated successfully without warnings.Hope it Helps,
Ogan

Similar Messages

  • Server uses AL32UTF8 character set

    Hello All,
    OS: Linux (SLES 8)
    Recently I have applied 10g Relase 1 (10.1.0.5) patch set for AIX 64 -- Patch #
    4505133.
    My earler version of database was 10.1.0.3.0.
    I got message patch successfully installed.
    I can login to the database.
    One of our weekly routine is to run our internal shell script interface
    programs.
    Very first step of this program is to created .dmp (exp) file. We are doing this
    from years.
    Now after applying this patch set,
    I am getting error in this program as below:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.5.0 -
    Production
    With the Partitioning, OLAP and Data Mining options
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses AL32UTF8 character set (possible charset conversion)
    About to export specified users ...
    . exporting pre-schema procedural objects and actions
    . exporting foreign function library names for user abc
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions for user abc
    About to export abc's objects ...
    . exporting database links
    . exporting sequence numbers
    . exporting cluster definitions
    EXP-00056: ORACLE error 932 encountered
    ORA-00932: inconsistent datatypes: expected BLOB, CLOB got CHAR
    EXP-00000: Export terminated unsuccessfully
    DN

    Pierre,
    I do not have any invalid objects.
    But I run
    SQL> @?/rdbms/admin/catmetx.sql (Per Metalink note 339938.1).
    But now I am getting errorsbelow:
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_COM_USER_PRIVILEGES 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_COLLECTION_DEFINITION 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_COLLECTION_DEF_REPOS 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_COLLECTION_REPOSITORY 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_ERRORS 0 rows exported
    . . exporting table QUEST_SL_EXPLAIN 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_EXPLAIN_PICK 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_QUERY_DEFINITIONS 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identifier
    . . exporting table QUEST_SL_QUERY_DEF_REPOSITORY 0 rows exported
    EXP-00056: ORACLE error 904 encountered
    ORA-00904: "SYS"."DBMS_EXPORT_EXTENSION"."FUNC_INDEX_DEFAULT": invalid identif
    Now I am running @catproc.sql.
    After that I will run utlrp.sql script if there is any invalid objects..
    DN

  • Possible charset conversion

    Hi
    My character set is :WE8ISO8859P1
    Even I set NLS_LANG prior to export and import, (export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P1 )
    I still get below warning during import. Whats the reason for this? (I didnt get any warning during export)
    import done in WE8ISO8859P1 character set and AL16UTF16 NCHAR character set
    export client uses WE8MSWIN1252 character set (possible charset conversion)

    ok thanks.
    Regarding the standard export and import,
    Suppose I take the export of a full schema: exp ../... owner=app file=....dmp
    During import, user is not created automatically.
    I have to create the user prior to import.
    Is this same in datapump?

  • Message uses a character set that is not supported by the internet service

    Does any one have any advice on how to fix this problem?
    E-mails sent from my iphone 3G periodically arrive in an unreadable form at the recipient. The body of the e-mail has been replaced with the message "This message uses a character set that is not supported by the internet service...." The problem e-mails also include an attachment that contains an unformatted text file containing the original message surrounded by what appears to be lots of formatting data that is displayed as gibberish.
    This occurs sometimes, but not always, even with the same recipients. I am sending e-mail through a G-mail account that is configured on the iphone using IMAP. I have tried the gmail account to use the two available formatting options for mail, but neither fixes the problem.
    I have also upgraded to 2.01 and restored a few times without impact.

    Hi,
    I got somewhat similar problem with special charecters(German umlaud �,�,�..).
    I create a file with java having special charecters in it. Now if I open this file I am able to view the special charecters in it.But If I attach this file send it using following code then receiver can not see the umlaud charecters in it.They get replaced by _ or ?
    MimeBodyPart mbp2 = new MimeBodyPart();
    FileDataSource fds = new FileDataSource(fileName);
    mbp2.setDataHandler(new DataHandler(fds));
    mbp2.setFileName(output.getName());
    Multipart mp = new MimeMultipart();
    mp.addBodyPart(mbp2);
    msg.setContent(mp);
    Transport.send(msg);
    From you message it looks like you are able to send the mail attachment correctly(by preserving special charecters).
    Can you tell me what might be wrong in my code.
    I appriciate your efforts in advance.
    Prasad

  • HOW can I enter text using Japanese character sets?

    The "Text, Plates, Insets" section of the LOOKOUT(6.01) Help files states:
    "Click the » button to the right of the Text field to expand the field for multiple line entries. You can enter text using international character sets such as Chinese, Korean, and Japanese."
    Can someone please explain HOW to do this? Note, I have NO problem inputting Hirigana, Katakana, and Kanji into MS WORD; the keyboard emulates the Japanese layout and characters (Romaji is default) and the IME works fine converting Romaji, and I can also select charcters directly from the IME Pad. I have tried several different fonts with success and am currently using MS UI Gothic.ttf as default. Again, everything is normal and working in a predictable manner within Word.
    I cannot get these texts into Lookout. I can't cut/paste from HTML pages or from text editors, even though both display properly. Within Lookout with JP selected as language/keyboard, when trying to type directly into the text field, the IME CORRECTLY displays Hirigana until <enter> is pressed, at which point all text reverts to question marks (?? ???? ? ?????). If I use the IME Pad, it does pretty much the same. I managed to get the "Yen" symbol to display, though, if that's relevant. As I said, font selected (in text/plate font options) is MS UI Gothic with Japanese as the selected script. Oddly enough, at this point the "sample" window is showing me the exact Hirigana character I want displayed in Lookout, but it won't. I've also tried staying in English and copying unicode characters from the Windows Character Map. Same results (Yen sign works, Hirigana WON'T).
    Help me!
    JW_Tech

    JW_Tech,
    Have you changed the regional setting to Japanese?
    Doug M
    Applications Engineer
    National Instruments
    For those unfamiliar with NBC's The Office, my icon is NOT a picture of me
    Attachments:
    language.JPG ‏50 KB

  • SUPPORT FOR euro WE8ISO8859P15 CHARACTER SET

    Hi,
    I cannot run Portal home page after 9iAS release 2 installation on Sun Solaris.
    On both nodes (infrastructure and middle tier) and in portal DAD I have NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P15.
    Is it possible to run portal with this character set?
    This is the error message:
    Error: The servlet produced the following error stack. java.io.IOException: Unsupported character encoding: "ISO-8859-15"
         at oracle.webdb.page.BaseContentRequest.getResponseJavaEncoding(Unknown Source)
         at oracle.webdb.page.BaseContentRequest.getReader(Unknown Source)
         at oracle.webdb.page.PageBuilder.getMetaData(Unknown Source)
         at oracle.webdb.page.PageBuilder.process(Unknown Source)
         at oracle.webdb.page.ParallelServlet.doGet(Unknown Source)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:244)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:336)
         at com.evermind.server.http.ResourceFilterChain.doFilter(ResourceFilterChain.java:59)
         at oracle.security.jazn.oc4j.JAZNFilter.doFilter(JAZNFilter.java:283)
         at com.evermind.server.http.ServletRequestDispatcher.invoke(ServletRequestDispatcher.java:523)
         at com.evermind.server.http.ServletRequestDispatcher.forwardInternal(ServletRequestDispatcher.java:269)
         at com.evermind.server.http.HttpRequestHandler.processRequest(HttpRequestHandler.java:735)
         at com.evermind.server.http.AJPRequestHandler.run(AJPRequestHandler.java:151)
         at com.evermind.util.ThreadPoolThread.run(ThreadPoolThread.java:64)
    Best regards,
    Zoran

    Hi
    You dont need anything special, however the JVM you use must support the encoding you want (unless you are going with a unicode flavor like UTF-8 which all JVM's support). The rest of the i18n stuff like ResourceBundles for messages, date currency formats are handled in your code anyway.
    regards
    deepak

  • USING THE CHARACTER SET SCANNER

    How can I find and download the Character Set Scanner utility mentioned in MetaLink document # 66320.1.
    Thanks
    Pratap

    I would like to know if there exists a database character set
    scanner for ORACLE816. If not, can I use the ORACLE817 scanner
    to scan my ORACLE816 database?Check out the Globalization Support Home page
    http://technet.oracle.com/tech/globalization/content.html and go
    to the scanner download page for more info.
    Is it important to use a scanner if I am converting my database
    from, say, WE8ISO8859P1 character set to UTF8? Yes definitely. The scanner will identify
    1. If you have any invalid WE8ISO8859P1 data in your database ,
    these will be lost during the character set conversion process.
    2. If there are any data cell expansions, so that column widths
    can be expanded to cater for the new sizing.
    3. Area of possible character set conversion and truncation of
    data.
    Can I assume that a database UTF8 will automatically store all
    my original characters correctly?UTF8 can store all the characters from WE8ISO8859P1 ,and it can
    be stored correctly if your NLS configuration is correct . The
    key is your client NLS_LANG character set setting .
    I have a client with NLS_LANG set to UFT8. When I exported my
    database (in WE8ISO8859P1) from my client, the export was done
    in the client's setting i.e. UTF8. Can I assume that the
    conversion was done at the time of export itself?Yes, your assumption is correct.

  • Use 2 character sets in the same sapscript form (1100 & 4030)

    Is this possible or I need to make a new character set and then make the changes that I need.
    For example:
    The 'Field group' character is usually hex 1D. The problem is
    that hex 1D is not part of the printer character set 1100 . So
    a lot of changes are necessary:
    - Copy your device type in the customer naming room (transaction SPAD,
    Utitities - For device types - Copy device type).
    - Check, which character set your device type uses (transaction SPAD,
    device type).
    - Copy the character set to the customer naming room (SPAD - Character
    set ; customer naming room: 9xxx)
    - Call SPAD - Device Type and put the new character set into all three
    fields at 'Printer character sets'.
    - Now go into your new character set (SPAD - character set , Button 'Edit
    character set '). Unfortunately the character 'Field group'  doesn't
    exist yet. So you should select a character which you don't need and
    which you can 'misuse' as 'Field group' . Let's say, you don't need
    the character 'Thorn'. Then you
    have to add/change the entry of character 354,
    if you want to use the Thorn character for the 'Field group' .
    Here you must insert the sequence '1D' to get the 'Field group'  character .
    Now if you write <354> in the Sapscript layout set in the old
    editor or if you press 'Insert command' in the new editor
    and insert SAP character 354, a 'Field group'  should be inserted
    at that place in the print data.

    The Function Module CLOSE_FORM has an optional TABLES parameter called OTFDATA that can be filled with the contents of the document instead of printing.
    To do this, you must set the field TDGETOTF = 'X' in the OPTIONS structure that can be supplied as an optional parameter to Function Module OPEN_FORM when the document is opened for output.
    This table contains the OTF format data that describes your SAPscript document. Normally the SAPscript compositer would print this to an external printer or fax or email, but once captured in your program you can do anything you like with this information.
    For instance, you can convert OTF data to a PDF file using SX_OBJECT_CONVERT_OTF_PDF. See the many threads on this topic in the ABAP Development forum if you are interested.
    I hope this hint is helpfull to you. Good luck.

  • Export using UTF8 character set

    Hi,
    My client is having a production database with default character set.
    While exporting I need to export the database to create a dump with UTF8 character set.
    Please let me know how to export with UTF8 option.
    Thanks....

    Hi, I am not sure if I got you correst. Here is what I think I have understood:
    - your client has a db which uses UTF8 as character set.
    - you want to export and make sure that there is no conversion taking place.
    for this you must set NLS_LANG variable in the shell from where you call exp, resp. expdp properly.
    NLS_LANG is composit and consists of:
    NLS_LANGUAGE_NLS_TERRITORY.NLS_CHARACTERSET
    In a bash this would look like this:
    $ export NLS_LANG=american_america.UTF8
    In other shells you might need to first define the varaible and then export:
    $ NLS_LANG=american_america.UTF8
    $ export NLS_LANG
    What you also need to know is that you influence the behavior for decimal separator and grand separator in numeric values and a few more settings by specifying NLS_TERRITORY.
    For instance NLS_TERRITORY uses a "." as decimal separator and a "," as grand separator for numeric values.
    This can be a pitfall! If you have the wrong territory specified then it might destroy all you numerics!
    Hope this helps,
    Lutz

  • What is needed to read/use Cyrillic character sets?

    Hello all,
    I'm working with a friend in Russia who is sending me documents typed in Russian, but my computers here don't show anything but jibberish and any non-Cyrillic words included in the text.
    What do I need to do to view this on my computer? Do I need a character set, or just a Cyrillic font?
    Thanks in advance for your help!
    Sincerely,
    wordman

    Ajust the page encoding in the head section. that's all you need. Any browser should correctly render the page based on that using Unicode.
    Mylenium

  • MD5 Migration issues using different  character set

    Hi,
    DBname = test1 , db charcter set WE8ISO8859P1 ,os charcter set WE8ISO8859P1
    DBname = test2 , db charcter set US7ASCII ,os charcter set UTF8
    In both DB table1 having encrypted data using md5 encryption techniques . (dbms_obfuscation_toolkit.md5(input_string => v_input) )
    Now i want to migrate data from test2 DB to test1 DB .How to migrate any one please help.
    Thanks in advance
    Naveen

    I'm not sure that I follow what you're suggesting but it sounds like you're on the right track.
    I'm still not sure whether you need the MD5 hash to match data in the database or not. If you change the character set of the underlying data, the MD5 hash may change so just keeping the hash intact may not be sufficient.
    Justin

  • Data load uses wrong character set, where to correct? APEX bug/omission?

    Hi,
    I created a set of Data Load pages in my application, so the users can upload a CSV file.
    But unlike the Load spreadsheet data (under SQL Workshop\Utilities\Data Workshop), where you can set the 'File Character Set', I didn't see where to set the Character set for Data Load pages in my application.
    Now there is a character set mismatch, "m³/h" and "°C" become "m�/h" and "�C"
    Where to set?
    Seems like an APEX bug or at least omission, IMHO the Data Load page should ask for the character set, as clients with different character sets could be uploading CSV.
    Apex 4.1 (testing on the apex.oracle.com website)

    Hello JP,
    Please give us some more details about your database version and its character set, and the character set of the CSV file.
    >> …But unlike the Load spreadsheet data (under SQL Workshop\Utilities\Data Workshop), where you can set the 'File Character Set', I didn't see where to set the Character set for Data Load pages in my application.
    It seems that you are right. I was not able to find any reference to the (expected/default) character set of the uploaded file in the current APEX documentation.
    >> If it's an APEX omission, where could I report that?
    Usually, an entry on this forum is enough as some of the development team members are frequent participants. Just to be sure, I’ll draw the attention of one of them to the thread.
    Regards,
    Arie.
    &diams; Please remember to mark appropriate posts as correct/helpful. For the long run, it will benefit us all.
    &diams; Author of Oracle Application Express 3.2 – The Essentials and More

  • Conversions between character sets when using exp and imp utilities

    I use EE8ISO8859P2 character set on my server,
    when exporting database with NLS_LANG not set
    then conversion should be done between
    EE8ISO8859P2 and US7ASCII charsets, so some
    characters not present in US7ASCII should not be
    successfully converted.
    But when I import such a dump, all characters not
    present in US7ASCII charset are imported to the database.
    I thought that some characters should be lost when
    doing such a conversions, can someone tell me why is it not so?

    Not exactly. If the import is done with the same DB character set, then no matter how it has been exported. Conversion (corruption) may happen if the destination DB has a different character set. See this example :
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:01 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> create table test(col1 varchar2(1));
    Table created.
    TEST@db102 SQL> insert into test values(chr(166));
    1 row created.
    TEST@db102 SQL> select * from test;
    C
    ¦
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:55 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ©
    Typ=1 Len=1: 166
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ exp test/test file=test.dmp tables=test
    Export: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:47 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P15 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                           TEST          1 rows exported
    Export terminated successfully without warnings.
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:56 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> drop table test purge;
    Table dropped.
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ imp test/test file=test.dmp
    Import: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:15 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P15 character set (possible charset conversion)
    . importing TEST's objects into TEST
    . importing TEST's objects into TEST
    . . importing table                         "TEST"          1 rows imported
    Import terminated successfully without warnings.
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:34 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ¦
    Typ=1 Len=1: 166
    TEST@db102 SQL>

  • Import error character set conversion

    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses AL32UTF8 character set (possible charset conversion)
    export client uses WE8ISO8859P1 character set (possible charset conversion)
    Import terminated successfully without warnings.
    how to resolve this
    importinng into linux machine

    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses AL32UTF8 character set (possible charset conversion)
    export client uses WE8ISO8859P1 character set (possible charset conversion)
    Import terminated successfully without warnings.Your import was successful. However, there are a few things you need to note from the top of the import log. It is advisable to use correct NLS settings when doing export/import to avoid data corruption/loss. That idea is to ensure you set the correct NLS setting (NLS_LANG) at the client before running exp or imp. In some cases, the setting does not matter because the different character sets can go together.
    In this case, I would not worry about it.

  • Checking server character set

    I think this is a simple question, but for some reason I am not able to find the result.
    How can I check the server's default character set?
    We have HP-UX (unix) servers, and most databases are set to AL32UTF8/AL16UTF16.
    I thought it would be something simple like typing 'locale', but on the HP-UX box(es), it doesn't show any setting for LANG.
    I also executed 'env' and looked at all settings, but nothing seems relevant.
    Unless I explicitly set the NLS_LANG=AMERICAN_AMERICA.AL32UTF8, when I do an export, it shows:
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses AL32UTF8 character set (possible charset conversion)
    Where is it getting the US7ASCII?
    If I select * from nls_database_parameters; there is no setting like US7ASCII.
    When I do a 'man -k character', I get numerous possible commands, but none of them appear to be something that displays the default character set.
    But, as I noted above, when I run an export of our database (without setting the NLS_LANG), it shows me that the default character set is US7ASCII.
    So, how do I show this, and if you might also have any ideas how to change this to UTF8.
    Thanks.
    removed references in this posting to my linux boxes...

    The default O/S character set for Unix is 7-bit ASCII. This is element of so-called "C locale". This locale is used by all applications that do not declare themselves sensitive to user locale (i.e. they do not call setlocale() at application startup) and by all applications that have no locale parameters (LANG, or LC_xxxx) set in their environment.
    If you call 'locale' in a Unix session that has no locale environment set, the "C locale" should be reported.
    "C locale" uses US locale formatting conventions and binary collation.
    ## "Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    ## server uses AL32UTF8 character set (possible charset conversion)"
    Here, US7ASCII is the default character set of an Oracle Client. It is not directly related to the default O/S character set. US7ASCII is always the default Oracle client character set on all non-EBCDIC platforms, used if NLS_LANG is not explicitly set.
    -- Sergiusz

Maybe you are looking for