AL32UTF8 / US7ASCII db characterset

Hi Guys,
My DB is currently on 10.2.0.5 using US7ASCII as character set. Definately for now and future, my DB will only be storing english language.
We are intending to migrate to 11g database. Have some queries to ask experts out there (pardon me, i have not started testing. very soon i will)
1. when creating db in 11g, can we choose the character set as us7ascii?
2. oracle recommended to use unicode in 11g (eg: al32utf8). For my case here, does it bring any benefits to me by changing us7ascii to al32utf8 since i'm quite sure we will only be using english? Not sure if it's work the risk or efforts to migrate as my application is very huge.
3. if the db is converted from us7ascii to al32utf8 successfully (without any loose of data using csscan), do i have to change anything on the application end? Will it impact the application in anyway if the db character set is changed?
thanks

dbaing wrote:
1. when creating db in 11g, can we choose the character set as us7ascii?Yes.
2. oracle recommended to use unicode in 11g (eg: al32utf8). For my case here, does it bring any benefits to me by changing us7ascii to al32utf8 since i'm quite sure we will only be using english? Not sure if it's work the risk or efforts to migrate as my application is very huge.If you're only storing US7ASCII data, there is effectively no difference between a US7ASCII database and an AL32UTF8 (assuming the clients have been properly configured all along). Every character is still going to require 1 byte of storage. But if you use AL32UTF8, you have the potential to store far more characters in the future. Even if you're just going to store English, is there any chance that you'd need to store a Euro character or a British Pound symbol? Microsoft smart-quotes?
3. if the db is converted from us7ascii to al32utf8 successfully (without any loose of data using csscan), do i have to change anything on the application end? Will it impact the application in anyway if the db character set is changed?If the application is configured correctly, no. If the application is configured incorrectly (particularly if it relies on bypassing character set conversion), then you'll have a much bigger task.
Justin

Similar Messages

  • Using AL32UTF8 or ZHS16GBK Characterset

    Hi,
    Currently, our database is using AL32UTF8 as the characterset but we need to store some Simplified Chinese data in tables.
    I know what AL32UTF8 characterset is able to store Simplified Chinese data but not being able to sort the Chinese character by Han yu pin yin (By default). The default is by Binary.
    I also know that ZHS16GBK characterset is able to store Simplified Chinese and the default sorting is by Han yu pin yin, which is what I want.
    1. So I just wonder what type of characterset should I use, AL32UTF8 or ZHS16GBK?
    2. AL32UTF8 is an unicode characterset which can support multi-language but ZHS16GBK characterset can only store Simplified Chinese? Am I right? What about English words, can ZHS16GBK characterset being able to store English words?
    3.Which is the recommended Characterset to use for my case?
    4. If I choose to set my characterset as AL32UTF8 but I want all the Simplified Chinese data to be sort by Han yu pin yin, how should I do?
    5. I read one of the documentation and it state that I can set the NLS_SORT in the nls_Session_parameters to SCHINESE_PINYIN_M. After setting the session parameters, I want into SQLPLUS and execute a select statement (e.g. select short_name from client order by short_name) and the Simplified Chinese short name retrieve is sorted by Han yu pin yin. But when my application (NOT through SQLPLUS) execute the same select statement, the result is not sorted by Han yu pin yin. Why is this so? Did I miss out anything?
    6. Lastly, I tried to set the NLS_SORT in the instance parameters as well by changing the spfile but the application using the select statement still cannot sort the Simplified Chinese data by Han yu pin yin. What should I do?

    Hi,
    Actually, I change the nls_sort for the session paramaters through the setting of the nls_sort in the registry instead of using the alter session command line.
    This is so as using alter session command line will only change the nls_sort for that particular session and once I exit the session, the nls_sort will be reset back to the default which is in Binary. I will have to re-alter the nls_sort in the session parameters again.
    Thats why I set the nls_sort in the registry so that the nls_sort in the session parameters will remain in the client machine.
    After setting the nls_sort in the registry, I am able to sort the Simplified Chinese characters by Han yu pin yin using the SQL*PLUS but not in my application.
    My applications are running on Windows environment (Microsoft Windows 2000 5.00.2195 SP3).
    The applications are split into front-end and back-end. The backend is using java and the front-end is using a software that is also using java to control.
    Basically, the front-end application will execute an electronic data exchange service and it will establish a jdbc oracle thin connection to the database.
    There is configuration file that is attached with each Electronic data exchange service. This configuration file contains the select statement (select short_name from client order by short_name) and each configuration file is linked with a datasource which will establish a jdbc oracle thin connection to the database.
    But seems like the result of the select statement in the configuration file are not sorted by Han yu pin yin. But executing the same select statement through SQL*PLUS will return the data sorted by Han yu pin yin.
    Thanks.

  • Error in displayed french characters

    Hello,
    I'm using a DB 11g configured with NLS_LANG=FRENCH_FRANCE.US7ASCII
    I have a WLS_FORMS server, the .env file has NLS_LANG=FRENCH_FRANCE.US7ASCII
    FMX files are compiled in an environment having NLS_LANG=FRENCH_FRANCE.US7ASCII in its registry.
    When I open the application, all specific characters can't be properly displayed (I have an ugly square instead).
    Where do I miss something ?

    christian erlinger wrote:
    So your french characters are stored in a database with US7ASCII database characterset? Somehow I doubt that. Or did you just set the NLS_LANG environment variable on your database server (which doesn't influence the database characterset at all) to US7ASCII?The database was created with the US7ASCII characterset, so yes the characters are stored with this characterset.
    In any case: US7ASCII is not the correct characterset for you, as it won't allow you to store your french characters. Take a look:
    http://docs.oracle.com/cd/B19306_01/server.102/b14225/applocaledata.htm#sthref1958
    Thanks for the link
    >
    So if you really have US7ASCII as database characterset you'd need to change that to something that fits your needs. The ultimate characterset where you shouldn't have problems storing any characters would be AL32UTF8, and as it is a superset of US7ASCII you should be able to change it with csalter
    cheersSo I have to migrate the DB character set, right ?

  • Problem Importing Sample Application - Err 7621

    Hi,
    Experiencing a problem exporting/importing the sample application into the same workspace. This is a new workspace we've created, the second one in the same DB.
    Export:
    Application: Sample Application
    File Format: UNIX
    Build Status: Run and Build Application
    Import:
    File Type: Application/Page Export
    File Character Set: Unicode UTF-8
    Error:
    ERR-7621 Could not determine workspace for application (:) on application accept.
    Expecting p_company or wwv_flow_company cookie to contain security group id of application owner.
    Env:
    User: GREGRO
    Workspace LPM_CUSTOM>About
    HTML DB
    Product Build: 1.6.0.00.87
    Schema Compatibility: 2004.07.04
    Last DDL Time: 02/21/2005 11:17:04 AM
    Host Schema: GREGRO
    Application Owner: FLOWS_010600
    Workspace ID: 874618965604689
    Workspace Name: LPM_CUSTOM
    Current User: GREGRO
    Language Preference: en
    Current Time (on server): 05/13/2005 11:15:18 AM
    Database version
    Database Banner
    Oracle9i Enterprise Edition Release 9.2.0.5.0 - 64bit Production
    PL/SQL Release 9.2.0.5.0 - Production
    CORE 9.2.0.6.0 Production
    TNS for Solaris: Version 9.2.0.5.0 - Production
    NLSRTL Version 9.2.0.5.0 - Production
    NLS_CHARACTERSET US7ASCII
    DAD CHARACTERSET US-ASCII
    Any insight to the problem would be appreciated
    Cheers Greg

    Hi Scott,
    Here is the last few lines of the log.
    Fri May 13 09:22:11 2005] [error] [client 130.194.145.15] [ecid: 1115940131:130.194.82.14
    5:6163:0:22,0] mod_plsql: /pls/CDUT/htmldb/wwv_flow.accept HTTP-404 ORA-00942: table or vi
    ew does not exist
    [Fri May 13 09:23:04 2005] [error] [client 130.194.145.15] [ecid: 1115940184:130.194.82.14
    5:10341:0:578,0] mod_plsql: /pls/CDUT/htmldb/wwv_flow.accept HTTP-404 ORA-00942: table or
    view does not exist
    [Fri May 13 09:33:37 2005] [error] [client 130.194.146.86] [ecid: 1115940817:130.194.82.14
    5:10341:0:596,0] mod_plsql: /pls/CDUT/htmldb/htmldb HTTP-401 ORA-1017
    [Fri May 13 09:48:19 2005] [error] [client 130.194.144.30] [ecid: 1115941699:130.194.82.14
    5:10341:0:611,0] mod_plsql: /pls/CDUT/htmldb/htmldb HTTP-401 ORA-1017
    [Fri May 13 09:51:14 2005] [error] [client 130.194.144.30] [ecid: 1115941874:130.194.82.14
    5:10355:0:612,0] File does not exist: /u10/CDUT/htmldb/doc/favicon.ico
    [Fri May 13 10:00:02 2005] [error] [client 130.194.144.30] [ecid: 1115942402:130.194.82.14
    5:10340:0:633,0] mod_plsql: /pls/CDUT/htmldb/f HTTP-401 ORA-1017
    [Fri May 13 10:00:12 2005] [error] [client 130.194.144.30] [ecid: 1115942412:130.194.82.14
    5:10342:0:633,0] mod_plsql: /pls/CDUT/htmldb/f HTTP-401 ORA-1017
    [Fri May 13 10:01:17 2005] [error] [client 130.194.147.207] [ecid: 1115942471:130.194.82.1
    45:10340:0:644,0] mod_plsql: /pls/CDUT/htmldb/wwv_flow.accept HTTP-404 ORA-00942: table or
    view does not exist
    Tried the different char sets to no avail.
    If we run the import script from TOAD directly it works ok.
    Cheers Greg

  • SQL Loader and foreign characters in the data file problem

    Hello,
    I have run into an issue which I can't find an answer for. When I run SQL Loader, one of my control files is used to get file content (LOBFILE) and one of the fields in the data file has a path to that file. The control file looks like:
    LOAD DATA
    INFILE 'PLACE_HOLDER.dat'
    INTO TABLE iceberg.rpt_document_core APPEND
    FIELDS TERMINATED BY ','
    doc_core_id "iceberg.seq_rpt_document_core.nextval",
    -- created_date POSITION(1) date "yyyy-mm-dd:hh24:mi:ss",
    created_date date "yyyy-mm-dd:hh24:mi:ss",
    document_size,
    hash,
    body_format,
    is_generic_doc,
    is_legacy_doc,
    external_filename FILLER char(275) ENCLOSED by '"',
    body LOBFILE(external_filename) terminated by EOF
    A sample data file looks like:
    0,2012-10-22:10:09:35,21,BB51344DD2127002118E286A197ECD4A,text,N,N,"E:\tmp\misc_files\index_testers\foreign\شیمیایی.txt"
    0,2012-10-22:10:09:35,17,CF85BE76B1E20704180534E19D363CF8,text,N,N,"E:\tmp\misc_files\index_testers\foreign\ลอบวางระเบิด.txt"
    0,2012-10-22:10:09:35,23552,47DB382558D69F170227AA18179FD0F0,binary,N,N,"E:\tmp\misc_files\index_testers\foreign\leesburgis_á_ñ_é_í_ó_ú_¿_¡_ü_99.doc"
    0,2012-10-22:10:09:35,17,83FCA0377445B60CE422DE8994900A79,binary,N,N,"E:\tmp\misc_files\index_testers\foreign\làm thế nào bạn làm ngày hôm nay"
    The problem is that whan I run this, SQL Loader throws an error that it can't find the file. It appears that it can't interpret the foreign characters in a way that allows it to find that path. I have tried adding a CHARACTERSET (using AL32UTF8 or UTF8) value in the control file but that only has some success with Western languages, not the ones listed above. Also, there is no set of defined languages that could be found in the data file. It essentaially could be any language.
    Does anyone know if there is a way to somehow get SQL Loader to "understand" the file system paths when a folder and/or file name could be in some other langauge?
    Thanks for any thoughts - Peter

    Thanks for the reply Harry. If I try to open the file in various text editors like Wordpad, Notepad, GVIM, andTextpad, they all display the foreign characters differently. Only Notepad comes close to displaying the characters properly. I have a C# app that will read the file and display the contents and it renders it fine. If you look at the directory of files in Windows Explorer, they all are displayed properly. So it seems things like .Net and Windows have some mechanism to understand the characters in order to render them properly. Other applications, again like Wordpad, do not know how to render them properly. It would seem that whatever SQL Loader is using to "read" the data files also is not rendering the characters properly which prevents it from finding the directory path to the file. If I add "CHARACTERSET AL32UTF8" in the control file, all is fine when dealing with Western langauges (ex, German, Spanish) but not for the Eastern languages (ex. Thai, Chinese). So .... telling SQL Loader to use a characterset seems to work, but not in all cases. The AL32UTF8 is the characterset that the Oracle database was created with. I have not had any luck if I try to set the CHARACTERSET to whatever the Thai character set is, for example. There problem there though is that even if that did work, I can't target specific lagauages because the data could come from anywhere. It's like I need some sort of global "super set" characterset to use. It seems like the CHARACTERSET is the right track to follow but I am not sure, and even if it is, is there a way to handle all languages.
    Thanks - Peter

  • How to encrypt column of some table with the single method ?

    How to encrypt column of some table with the single method ?

    How to encrypt column of some table with the single
    method ?How to encrypt column of some table with the single
    method ?
    using dbms_crypto package
    Assumption: TE is a user in oracle 10g
    we have a table need encrypt a column, this column SYSDBA can not look at, it's credit card number.
    tha table is
    SQL> desc TE.temp_sales
    Name Null? Type
    CUST_CREDIT_ID NOT NULL NUMBER
    CARD_TYPE VARCHAR2(10)
    CARD_NUMBER NUMBER
    EXPIRY_DATE DATE
    CUST_ID NUMBER
    1. grant execute on dbms_crypto to te;
    2. Create a table with a encrypted columns
    SQL> CREATE TABLE te.customer_credit_info(
    2 cust_credit_id number
    3      CONSTRAINT pk_te_cust_cred PRIMARY KEY
    4      USING INDEX TABLESPACE indx
    5      enable validate,
    6 card_type varchar2(10)
    7      constraint te_cust_cred_type_chk check ( upper(card_type) in ('DINERS','AMEX','VISA','MC') ),
    8 card_number blob,
    9 expiry_date date,
    10 cust_id number
    11      constraint fk_te_cust_credit_to_cust references te.customer(cust_id) deferrable
    12 )
    13 storage (initial 50k next 50k pctincrease 0 minextents 1 maxextents 50)
    14 tablespace userdata_Lm;
    Table created.
    SQL> CREATE SEQUENCE te.customers_cred_info_id
    2 START WITH 1
    3 INCREMENT BY 1
    4 NOCACHE
    5 NOCYCLE;
    Sequence created.
    Note: Credit card number is blob data type. It will be encrypted.
    3. Loading data encrypt the credit card number
    truncate table TE.customer_credit_info;
    DECLARE
    input_string VARCHAR2(16) := '';
    raw_input RAW(128) := UTL_RAW.CAST_TO_RAW(CONVERT(input_string,'AL32UTF8','US7ASCII'));
    key_string VARCHAR2(8) := 'AsDf!2#4';
    raw_key RAW(128) := UTL_RAW.CAST_TO_RAW(CONVERT(key_string,'AL32UTF8','US7ASCII'));
    encrypted_raw RAW(2048);
    encrypted_string VARCHAR2(2048);
    BEGIN
    for cred_record in (select upper(CREDIT_CARD) as CREDIT_CARD,
    CREDIT_CARD_EXP_DATE,
    to_char(CREDIT_CARD_NUMBER) as CREDIT_CARD_NUMBER,
    CUST_ID
    from TE.temp_sales) loop
    dbms_output.put_line('type:' || cred_record.credit_card || 'exp_date:' || cred_record.CREDIT_CARD_EXP_DATE);
    dbms_output.put_line('number:' || cred_record.CREDIT_CARD_NUMBER);
    input_string := cred_record.CREDIT_CARD_NUMBER;
    raw_input := UTL_RAW.CAST_TO_RAW(CONVERT(input_string,'AL32UTF8','US7ASCII'));
    dbms_output.put_line('> Input String: ' || CONVERT(UTL_RAW.CAST_TO_VARCHAR2(raw_input),'US7ASCII','AL32UTF8'));
    encrypted_raw := dbms_crypto.Encrypt(
    src => raw_input,
    typ => DBMS_CRYPTO.DES_CBC_PKCS5,
    key => raw_key);
    encrypted_string := rawtohex(UTL_RAW.CAST_TO_RAW(encrypted_raw)) ;
    dbms_output.put_line('> Encrypted hex value : ' || encrypted_string );
    insert into TE.customer_credit_info values
    (TE.customers_cred_info_id.nextval,
    cred_record.credit_card,
    encrypted_raw,
    cred_record.CREDIT_CARD_EXP_DATE,
    cred_record.CUST_ID);
    end loop;
    commit;
    end;
    4. Check credit card number script
    DECLARE
    input_string VARCHAR2(16) := '';
    raw_input RAW(128) := UTL_RAW.CAST_TO_RAW(CONVERT(input_string,'AL32UTF8','US7ASCII'));
    key_string VARCHAR2(8) := 'AsDf!2#4';
    raw_key RAW(128) := UTL_RAW.CAST_TO_RAW(CONVERT(key_string,'AL32UTF8','US7ASCII'));
    encrypted_raw RAW(2048);
    encrypted_string VARCHAR2(2048);
    decrypted_raw RAW(2048);
    decrypted_string VARCHAR2(2048);
    cursor cursor_cust_cred is select CUST_CREDIT_ID, CARD_TYPE, CARD_NUMBER, EXPIRY_DATE, CUST_ID
    from TE.customer_credit_info order by CUST_CREDIT_ID;
    v_id customer_credit_info.CUST_CREDIT_ID%type;
    v_type customer_credit_info.CARD_TYPE%type;
    v_EXPIRY_DATE customer_credit_info.EXPIRY_DATE%type;
    v_CUST_ID customer_credit_info.CUST_ID%type;
    BEGIN
    dbms_output.put_line('ID Type Number Expiry_date cust_id');
    dbms_output.put_line('-----------------------------------------------------');
    open cursor_cust_cred;
    loop
         fetch cursor_cust_cred into v_id, v_type, encrypted_raw, v_expiry_date, v_cust_id;
    exit when cursor_cust_cred%notfound;
    decrypted_raw := dbms_crypto.Decrypt(
    src => encrypted_raw,
    typ => DBMS_CRYPTO.DES_CBC_PKCS5,
    key => raw_key);
    decrypted_string := CONVERT(UTL_RAW.CAST_TO_VARCHAR2(decrypted_raw),'US7ASCII','AL32UTF8');
    dbms_output.put_line(V_ID ||' ' ||
    V_TYPE ||' ' ||
    decrypted_string || ' ' ||
    v_EXPIRY_DATE || ' ' ||
    v_CUST_ID);
    end loop;
    close cursor_cust_cred;
    commit;
    end;
    /

  • [Oracle][ODBC SQL Server Driver]String data, right truncation {01004}

    While importing data from SQL Server 2005 to Oracle gateway 11g Release2, i am getting following error:
    insert into CSDescr select * from CSDescr@sqlserver
    ERROR at line 1:
    ORA-28500: connection from ORACLE to a non-Oracle system returned this message:
    +[Oracle][ODBC SQL Server Driver]String data, right truncation {01004}+
    ORA-02063: preceding 2 lines from SQLSERVER
    Oracle database characterset is AL32UTF8
    SQLServer database characterset is SQL_Latin1_General_CP1_CI_AS
    Following is the configuration in the parameter file for gateway:
    HS_KEEP_REMOTE_COLUMN_SIZE=LOCAL
    HS_NLS_LENGTH_SEMANTICS=CHAR
    I think setting the HS_LANGUAGE parameter shall fix the error but I  want to know what should be the value of this parameter?

    HS_LANGUAGE needs to be set to a character set that is used by the foreign database
    So try: HS_LANGUAGE=american_america.WE8MSWIN1252
    Also specify HS_NLS_NCHAR=UCS2 as the nvarchars of a SQl Server are stored in UCS2 character set

  • Nvarchar2 data type

    Hi
    I have emp_ino table in oracle database 10g R2.
    emp_id Number(5);
    emp_name nvarchar2(50);
    I insert data through sql script its ok.
    In emp_name I used Arabic character.
    I creat A form module in oracle developer suite 10g R2 when I query record in forms 10g then it does not display in arabic format.
    please tell me what i have to do ?
    thanks

    emp_name nvarchar2(50);First of all ist the table column you are inserting your data also of type nvarchar2 ? The NVARCHAR2 uses the National Characterset whereas VARCHAR2 uses the database characterset which means you could have a characterset conversion with possible data loss at hand when trying to insert NVARCHAR2 into VARCHAR2 or vice versa.
    The national characterset itself is simply there to support 2 different charactersets in your database; meaning when having a West-European Database Characterset you could switch the national characterset to something else if you'd need to support something else quickly.
    Personally I wouldn't go for it. The idea of having different charactersets with possible characterset conversions and data loss in one database sounds complicated enough; you'd have to worry everytime which datatype you have to use and how to convert it from characterset A to characterset B. The horrorstory continues when you plan to use another database characterset, as - you guessed it - the national characterset will make this also more complicated.
    The by far easier way is to use AL32UTF8 as Database characterset and forget about that there even is a National Chararacterset.
    cheers

  • DBMS_Crypto.Encrypt

    Reference to the site:
    http://www.oracle.com/technology/oramag/oracle/05-jan/o15security.html
    I have created get_enc_val function in the database.
    function get_enc_val
    p_in in varchar2,
    p_key in raw
    return raw is
    l_enc_val raw (2000);
    l_mod number := dbms_crypto.ENCRYPT_AES128
    + dbms_crypto.CHAIN_CBC
    + dbms_crypto.PAD_PKCS5;
    begin
    l_enc_val := dbms_crypto.encrypt
    UTL_I18N.STRING_TO_RAW
    (p_in, 'AL32UTF8'),
    l_mod,
    p_key
    return l_enc_val;
    end;
    When i run the following:
    create table test(res_id varchar2(19), res_salary raw(2000));
    insert into test
    (res_id, res_salary)
    values
    ('001',
    get_enc_val (
    '2000', dbms_crypto.randombytes (128))
    System shows error:
    ORA-28239: no key provided
    ORA-06512: at "SYS.DBMS_CRYPTO_FFI", line 3
    ORA-06512: at "SYS.DBMS_CRYPTO", line 10
    ORA-06512: at "BMS.GET_ENC_VAL", line 12
    Can anybody help? Thanks a lot!

    DECLARE
    input_string VARCHAR2(16) := 'tigertigertigert';
    raw_input RAW(128) :=
    UTL_RAW.CAST_TO_RAW(CONVERT(input_string,'AL32UTF8','US7ASCII'));
    key_string VARCHAR2(8) := 'scottsco';
    raw_key RAW(128) :=
    UTL_RAW.CAST_TO_RAW(CONVERT(key_string,'AL32UTF8','US7ASCII'));
    encrypted_raw RAW(2048);
    encrypted_string VARCHAR2(2048);
    decrypted_raw RAW(2048);
    decrypted_string VARCHAR2(2048);
    -- 1. Begin testing Encryption BEGIN
    dbms_output.put_line('> Input String : ' ||
    CONVERT(UTL_RAW.CAST_TO_VARCHAR2(raw_input),'US7ASCII','AL32UTF8'));
    dbms_output.put_line('> ========= BEGIN TEST Encrypt =========');
    encrypted_raw := dbms_crypto.Encrypt(
    src => raw_input,
    typ => DBMS_CRYPTO.DES_CBC_PKCS5,
    key => raw_key);
    dbms_output.put_line('> Encrypted hex value : ' ||
    rawtohex(UTL_RAW.CAST_TO_RAW(encrypted_raw)));
    decrypted_raw := dbms_crypto.Decrypt(
    src => encrypted_raw,
    typ => DBMS_CRYPTO.DES_CBC_PKCS5,
    key => raw_key);
    decrypted_string :=
    CONVERT(UTL_RAW.CAST_TO_VARCHAR2(decrypted_raw),'US7ASCII','AL32UTF8');
    dbms_output.put_line('> Decrypted string output : ' ||
    decrypted_string);
    if input_string = decrypted_string THEN
    dbms_output.put_line('> String DES Encyption and Decryption successful');
    END if; dbms_output.put_line(''); dbms_output.put_line('> ========= BEGIN TEST Hash =========');
    encrypted_raw := dbms_crypto.Hash(
    src => raw_input,
    typ => DBMS_CRYPTO.HASH_SH1);
    dbms_output.put_line('> Hash value of input string : ' ||
    rawtohex(UTL_RAW.CAST_TO_RAW(encrypted_raw)));
    dbms_output.put_line('> ========= BEGIN TEST Mac =========');
    encrypted_raw := dbms_crypto.Mac(
    src => raw_input,
    typ => DBMS_CRYPTO.HMAC_MD5,
    key => raw_key);
    dbms_output.put_line('> Message Authentication Code : ' ||
    rawtohex(UTL_RAW.CAST_TO_RAW(encrypted_raw)));
    dbms_output.put_line(''); dbms_output.put_line('> End of DBMS_CRYPTO tests '); END; /
    error:
    dbms_output.put_line('> Input String : ' ||
    ERROR at line 17:
    ORA-06550: line 17, column 12:
    PLS-00103: Encountered the symbol "." when expecting one of the following:
    constant exception <an identifier>
    <a double-quoted delimited-identifier> table LONG_ double ref
    char time timestamp interval date binary national character
    nchar
    The symbol "<an identifier>" was substituted for "." to continue.
    ORA-06550: line 19, column 12:
    PLS-00103: Encountered the symbol "." when expecting one of the following:
    constant exception <an identifier>
    <a double-quoted delimited-identifier> table LONG_ double ref
    char time timestamp interval date binary national chara
    ORA-06550: line 20, column 15:
    PLS-00103: Encountered the symbol "=" when expecting one of the following:
    constant exception <an identifier>
    <a double-quoted delimited-identifier> table LONG_ double ref
    char time timestamp interval date binary national chara
    ORA-06550: line 26, column 15:
    PLS-00103: Encountered the symbol "." when expecting one of the following:
    constant exception <an iden
    Please help quickly.

  • Export From 3.0 and Import to 3.1

    Hi guys,
    I'm having error at below while i'm trying to import my application which is exported from apex 3.0. Right now i'm working on Application Express 3.1.0.00.32
    Can anybody help me please?
    Thanks for your interest
    ORA-20001: GET_BLOCK Error. ORA-20001: Execution of the statement was unsuccessful. ORA-06550: line 17, column 71: PLS-00103: Encountered the symbol &amp;quot;,&amp;quot; when expecting one of the following: * &amp;amp; = - + ; &amp;lt; / &amp;gt; at in is mod remainder not rem &amp;lt;an exponent (**)&amp;gt; &amp;lt;&amp;gt; or != or ~= &amp;gt;= &amp;lt;= &amp;lt;&amp;gt; and or like LIKE2_ LIKE4_ LIKEC_ between || member SUBMULTISET_ &lt;pre&gt;begin declare p varchar2(32767) := null;

    All DADs used for Application Express must use AL32UTF8 as the characterset. If they are something else you need to change them, restart the HTTP servers and export again. Then try to import/install. I would test the export and import/install on 3.0 first. When that works error-free, take the export files and import/install into 3.1.
    Scott

  • Converting charaterset.

    Hi All,
    My DB is in 11.1.0.7.0 and NLS_Characterset is AL32UTF8.
    One of my table column is in varchar2 and it contains different characterset (� ). My BO reports are failing because of this character set (BO character set is in AL32UTF8)
    Using covert function one set/pattern of data i converted to the original characterset. [ Select column1, convert(column2, 'AL32UTF8', 'WE8ISO8859P1') FROM xyz ]
    I also create a new table with same filed as nvarchar2 and inserted the data from the coverted records. From the new table BO reports are fine.
    Now i wanted to move complete table data (around 20 GB, not sure how many rows and and what all types of charater set having non-UTF8 character) to the new table with nvarchar2. Problem is that Since it is worldwide application there are many types of characterset involved in the database. Converting each one like previous query won't be possible and accurate.
    Is there any way i can perform this task
    1. Identify all the rows that not UTF8 format.
    2. Convert to a correct character set and move to a new table with nvarchar2 that by default will take care of this (I hope so). Pelase note that i can't remove the charaters that have non-utf8 format
    So looks like diffrent character set data is in UTF8 format and we can't really see the extact charater set since it is in AL32UTF8. we need to convert back to the original character set and save it in nvarchar2 format.
    Please share your suggestions. Thanks in advance.
    Regards,
    Anto.

    Hi Anto,
    I would doubt that, as NVARCHAR2 columns are in the NCHAR characterset.
    The original idea was: you can use characters in the characterset in both columns and identifiers, so you would use a more or less restrictive characterset for that.
    For the NCHAR characterset, used in colum values only, you would typically use an unicode characterset. You had to designate a literal as NCHAR by prefixing N before the literal.
    Now Oracle allows AL32UTF8 for the normal characterset about the only NCHAR characterset is AL16UTF16. The N prefix is removed. If the change to NVARCHAR2 would force you to change you anything in your code, I wouldn't do that. You should be able to resolve it using AL32UTF8.
    I once, several years ago, read Oracle was going to desupport single byte characters in the next major release of Oracle. You may best use Oracle's recommendations which is to use AL32UTF8 as database characterset when you need unicode.
    Sybrand Bakker
    Senior Oracle DBA

  • Junk text issue

    Hi all,
    I have a remote client that I need to support for our application. They have a problem where they are seeing junk characters instead of some french characters in our application and SQL Developer. They have tried using two databases in AL32UTF8 and WE8MSWIN1252 characterset, but the issue persists. I need to try and drill down the issue and fix it. I have their database but I can see the characters fine without any issues in databases with AL32UTF8 and WEMSWIN1252 characterset settings.
    I am at the kind of lost as to what could be causing the problem.
    Can anyone suggest what would the issue be?
    What parameters can be causing this?
    What should be the key things that I need to look at?
    Thanks.
    Message was edited by:
    N

    Thanks for your response Roelie Viviers,
    Yes - I cannot agree more with you on this - its certainly getting to be a very complex issue. Especially when I am not sure what settings can I look or what more information should I try to fetch from the client.
    The most sensible way to go around this seemed to me was - to replicate the problem inhouse and see I can somehow see the problem. I have had some luck with it.
    I am now able to see garbage characters in the SQL*Plus Worksheet, but the application still shows it correctly. So what could it be in the client's environment that could be causing the issue!!
    I think, I might be able to safely elimate the possibility of nls_lang parameter in the registry causing this. Heres, what I tried.
    select * from nls_database_parameters;PARAMETER VALUE
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET WE8MSWIN1252
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_RDBMS_VERSION 10.1.0.2.0
    20 rows selected.
    Now tried querying the offending columns from two clients
    i) nls_lang set to ENGLISH_CANADA.AL32UTF8
    ii)nls_lang set to NA
    I still get same results. For eg. A test record in the database show something like this in the application Copie de à, è, etc Cédille (ç), àâäïÿùÔ; in SQL*Plus worksheet however it shows something like this:
    Copie de à , è, etc Cédille (ç), à âäïÿùÔ
    Surely, there is a problem somewhere. Is there anything in the application that make it can behave like SQL*Plus?
    Any help is appreciated.
    Thanks!

  • CSSCAN in 11g - Characterset not changing from WE8IMSWIN1252 to AL32UTF8

    All,
    We have installed a 11g database in Linux box and once after that we wanted to change the character set to AL32UTF8 from default WE8MSWIN1252.
    We took the cs-alter approach and ran cs-scan utility, upon going through csscan.txt files generated by csscan utility we found that there are no lossy data but convertible data was found in data dictionary. Below is the output from csscan.txt
    This is the Scan Summary
    *[Scan Summary]*
    All character type data in the data dictionary are convertible to the new character set
    All character type application data are convertible to the new character set
    Database Scan Summary Report
    Time Started  : 2012-10-17 21:42:17
    Time Completed: 2012-10-17 21:42:47
    Process ID         Time Started       Time Completed
             1  2012-10-17 21:42:18  2012-10-17 21:42:46
             2  2012-10-17 21:42:18  2012-10-17 21:42:46
             3  2012-10-17 21:42:18  2012-10-17 21:42:46
    [Database Size]
    Tablespace                           Used            Free           Total       Expansion
    SYSTEM                            709.75M         256.00K         710.00M           2.42M
    SYSAUX                            645.63M          34.38M         680.00M          12.52M
    UNDOTBS1                           13.13M          16.88M          30.00M            .00K
    TEMP                                 .00K            .00K            .00K            .00K
    USERS                               1.31M           3.69M           5.00M            .00K
    HYPE_DATA                       1,024.00K      19,999.00M      20,000.00M            .00K
    HYPE_INDX                       1,024.00K      19,999.00M      20,000.00M            .00K
    Total                           1,371.81M      40,053.19M      41,425.00M          14.94M
    The size of the largest CLOB is 1625114 bytes
    [Database Scan Parameters]
    Parameter                      Value
    CSSCAN Version                 v2.1
    Instance Name                  dvhp081
    Database Version               11.2.0.3.0
    Scan type                      Full database
    Scan CHAR data?                YES
    Database character set         WE8MSWIN1252
    FROMCHAR                       WE8MSWIN1252
    TOCHAR                         al32utf8
    Scan NCHAR data?               NO
    Array fetch buffer size        10240
    Number of processes            3
    Capture convertible data?      NO
    [Scan Summary]
    All character type data in the data dictionary are convertible to the new character set
    All character type application data are convertible to the new character set
    [Data Dictionary Conversion Summary]
    Data Dictionary Tables:
    Datatype                    Changeless      Convertible       Truncation            Lossy
    VARCHAR2                     5,408,302                0                0                0
    CHAR                             4,261                0                0                0
    LONG                           249,018                0                0                0
    CLOB                            67,652            3,794                0                0
    VARRAY                          49,807                0                0                0
    Total                        5,779,040            3,794                0                0
    Total in percentage             99.934%           0.066%           0.000%           0.000%
    The data dictionary can be safely migrated using the CSALTER script
    XML CSX Dictionary Tables:
    Datatype                    Changeless      Convertible       Truncation            Lossy
    VARCHAR2                           702                0                0                0
    CHAR                                 0                0                0                0
    LONG                                 0                0                0                0
    CLOB                                 0                0                0                0
    VARRAY                               0                0                0                0
    Total                              702                0                0                0
    Total in percentage            100.000%           0.000%           0.000%           0.000%
    [Application Data Conversion Summary]
    Datatype                    Changeless      Convertible       Truncation            Lossy
    VARCHAR2                     2,550,581                0                0                0
    CHAR                                 0                0                0                0
    LONG                                 0                0                0                0
    CLOB                            22,187            8,287                0                0
    VARRAY                               0                0                0                0
    Total                        2,572,768            8,287                0                0
    Total in percentage             99.679%           0.321%           0.000%           0.000%
    [Distribution of Convertible, Truncated and Lossy Data by Table]
    Data Dictionary Tables:
    USER.TABLE                                              Convertible       Truncation            Lossy
    MDSYS.SDO_COORD_OP_PARAM_VALS                                   200                0                0
    MDSYS.SDO_GEOR_XMLSCHEMA_TABLE                                    1                0                0
    MDSYS.SDO_STYLES_TABLE                                           78                0                0
    MDSYS.SDO_XML_SCHEMAS                                             5                0                0
    SYS.METASTYLESHEET                                              179                0                0
    SYS.RULE$                                                         1                0                0
    SYS.SCHEDULER$_EVENT_LOG                                        356                0                0
    SYS.WRH$_SQLTEXT                                                537                0                0
    SYS.WRH$_SQL_PLAN                                               514                0                0
    SYS.WRI$_ADV_DIRECTIVE_META                                       5                0                0
    SYS.WRI$_ADV_OBJECTS                                             28                0                0
    SYS.WRI$_ADV_SQLT_PLANS                                           2                0                0
    SYS.WRI$_ADV_SQLT_PLAN_STATS                                      2                0                0
    SYS.WRI$_DBU_FEATURE_METADATA                                   193                0                0
    SYS.WRI$_DBU_FEATURE_USAGE                                        9                0                0
    SYS.WRI$_DBU_HWM_METADATA                                        21                0                0
    SYS.WRI$_REPT_FILES                                              27                0                0
    SYSMAN.MGMT_IP_ELEM_DEFAULT_PARAMS                              130                0                0
    SYSMAN.MGMT_IP_REPORT_ELEM_PARAMS                             1,475                0                0
    SYSMAN.MGMT_IP_SQL_STATEMENTS                                    31                0                0
    XML CSX Dictionary Tables:
    USER.TABLE                                              Convertible       Truncation            Lossy
    Application Data:
    USER.TABLE                                              Convertible       Truncation            Lossy
    APEX_030200.WWV_FLOW_BANNER                                      10                0                0
    APEX_030200.WWV_FLOW_BUTTON_TEMPLATES                            12                0                0
    APEX_030200.WWV_FLOW_CUSTOM_AUTH_SETUPS                          19                0                0
    APEX_030200.WWV_FLOW_FLASH_CHART_SERIES                           5                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES                             298                0                0
    APEX_030200.WWV_FLOW_PAGE_GENERIC_ATTR                           44                0                0
    APEX_030200.WWV_FLOW_PAGE_PLUGS                               3,240                0                0
    APEX_030200.WWV_FLOW_PAGE_PLUG_TEMPLATES                        254                0                0
    APEX_030200.WWV_FLOW_PROCESSING                                  45                0                0
    APEX_030200.WWV_FLOW_ROW_TEMPLATES                               66                0                0
    APEX_030200.WWV_FLOW_SHORTCUTS                                   39                0                0
    APEX_030200.WWV_FLOW_STEPS                                    1,795                0                0
    APEX_030200.WWV_FLOW_STEP_PROCESSING                          2,238                0                0
    APEX_030200.WWV_FLOW_TEMPLATES                                  192                0                0
    APEX_030200.WWV_FLOW_WORKSHEETS                                  30                0                0
    [Distribution of Convertible, Truncated and Lossy Data by Column]
    Data Dictionary Tables:
    USER.TABLE|COLUMN                                       Convertible       Truncation            Lossy
    MDSYS.SDO_COORD_OP_PARAM_VALS|PARAM_VALUE_FILE                  200                0                0
    MDSYS.SDO_GEOR_XMLSCHEMA_TABLE|XMLSCHEMA                          1                0                0
    MDSYS.SDO_STYLES_TABLE|DEFINITION                                78                0                0
    MDSYS.SDO_XML_SCHEMAS|XMLSCHEMA                                   5                0                0
    SYS.METASTYLESHEET|STYLESHEET                                   179                0                0
    SYS.RULE$|CONDITION                                               1                0                0
    SYS.SCHEDULER$_EVENT_LOG|ADDITIONAL_INFO                        356                0                0
    SYS.WRH$_SQLTEXT|SQL_TEXT                                       537                0                0
    SYS.WRH$_SQL_PLAN|OTHER_XML                                     514                0                0
    SYS.WRI$_ADV_DIRECTIVE_META|DATA                                  5                0                0
    SYS.WRI$_ADV_OBJECTS|ATTR4                                       28                0                0
    SYS.WRI$_ADV_SQLT_PLANS|OTHER_XML                                 2                0                0
    SYS.WRI$_ADV_SQLT_PLAN_STATS|OTHER                                2                0                0
    SYS.WRI$_DBU_FEATURE_METADATA|INST_CHK_LOGIC                     22                0                0
    SYS.WRI$_DBU_FEATURE_METADATA|USG_DET_LOGIC                     171                0                0
    SYS.WRI$_DBU_FEATURE_USAGE|FEATURE_INFO                           9                0                0
    SYS.WRI$_DBU_HWM_METADATA|LOGIC                                  21                0                0
    SYS.WRI$_REPT_FILES|SYS_NC00005$                                 27                0                0
    SYSMAN.MGMT_IP_ELEM_DEFAULT_PARAMS|VALUE                        130                0                0
    SYSMAN.MGMT_IP_REPORT_ELEM_PARAMS|VALUE                       1,475                0                0
    SYSMAN.MGMT_IP_SQL_STATEMENTS|SQL_STATEMENT                      31                0                0
    XML CSX Dictionary Tables:
    USER.TABLE|COLUMN                                       Convertible       Truncation            Lossy
    Application Data:
    USER.TABLE|COLUMN                                       Convertible       Truncation            Lossy
    APEX_030200.WWV_FLOW_BANNER|BANNER                               10                0                0
    APEX_030200.WWV_FLOW_BUTTON_TEMPLATES|TEMPLATE                   12                0                0
    APEX_030200.WWV_FLOW_CUSTOM_AUTH_SETUPS|AUTH_FUNC                 8                0                0
    APEX_030200.WWV_FLOW_CUSTOM_AUTH_SETUPS|PAGE_SENT                10                0                0
    APEX_030200.WWV_FLOW_CUSTOM_AUTH_SETUPS|POST_AUTH                 1                0                0
    APEX_030200.WWV_FLOW_FLASH_CHART_SERIES|SERIES_QU                 5                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|ITEM_TEMPLATE                20                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|ITEM_TEMPLATE                20                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|LIST_TEMPLATE               105                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|LIST_TEMPLATE               105                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|SUB_LIST_ITEM                12                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|SUB_LIST_ITEM                12                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|SUB_TEMPLATE_                12                0                0
    APEX_030200.WWV_FLOW_LIST_TEMPLATES|SUB_TEMPLATE_                12                0                0
    APEX_030200.WWV_FLOW_PAGE_GENERIC_ATTR|ATTRIBUTE_                44                0                0
    APEX_030200.WWV_FLOW_PAGE_PLUGS|PLUG_SOURCE                   3,240                0                0
    APEX_030200.WWV_FLOW_PAGE_PLUG_TEMPLATES|TEMPLATE               166                0                0
    APEX_030200.WWV_FLOW_PAGE_PLUG_TEMPLATES|TEMPLATE                88                0                0
    APEX_030200.WWV_FLOW_PROCESSING|PROCESS_SQL_CLOB                 45                0                0
    APEX_030200.WWV_FLOW_ROW_TEMPLATES|ROW_TEMPLATE1                 54                0                0
    APEX_030200.WWV_FLOW_ROW_TEMPLATES|ROW_TEMPLATE2                 10                0                0
    APEX_030200.WWV_FLOW_ROW_TEMPLATES|ROW_TEMPLATE3                  2                0                0
    APEX_030200.WWV_FLOW_SHORTCUTS|SHORTCUT                          39                0                0
    APEX_030200.WWV_FLOW_STEPS|HELP_TEXT                          1,513                0                0
    APEX_030200.WWV_FLOW_STEPS|HTML_PAGE_HEADER                     282                0                0
    APEX_030200.WWV_FLOW_STEP_PROCESSING|PROCESS_SQL_             2,238                0                0
    APEX_030200.WWV_FLOW_TEMPLATES|BOX                               64                0                0
    APEX_030200.WWV_FLOW_TEMPLATES|FOOTER_TEMPLATE                   64                0                0
    APEX_030200.WWV_FLOW_TEMPLATES|HEADER_TEMPLATE                   64                0                0
    APEX_030200.WWV_FLOW_WORKSHEETS|SQL_QUERY                        30                0                0
    [Indexes to be Rebuilt]
    USER.INDEX on USER.TABLE(COLUMN)
    APEX_030200.WWV_FLOW_WORKSHEETS_UNQ_IDX on APEX_030200.WWV_FLOW_WORKSHEETS(SYS_NC00078$)
    APEX_030200.WWV_FLOW_WORKSHEETS_UNQ_IDX on APEX_030200.WWV_FLOW_WORKSHEETS(SYS_NC00079$)
    APEX_030200.WWV_FLOW_WORKSHEETS_UNQ_IDX on APEX_030200.WWV_FLOW_WORKSHEETS(SYS_NC00080$)
    APEX_030200.WWV_FLOW_WORKSHEETS_UNQ_IDX on APEX_030200.WWV_FLOW_WORKSHEETS(SYS_NC00081$)
    APEX_030200.WWV_FLOW_WS_UNQ_ALIAS_IDX on APEX_030200.WWV_FLOW_WORKSHEETS(SYS_NC00082$)
    APEX_030200.WWV_FLOW_WS_UNQ_ALIAS_IDX on APEX_030200.WWV_FLOW_WORKSHEETS(ALIAS)
    ----------------------------------------------------------------------------------We followed few metalink documents *Solving Convertible or Lossy data in Data Dictionary objects reported by Csscan when changing the NLS_CHARACTERSET [ID 258904.1]* and found that we are good to go as convertible was found only in data dictionary and that too CLOB data. But while running csalter.plb csalter came out without changing the characterset. We ran the following query given the said document and it returned no rows which again confirms there is no problem and go ahead with running csalter.
    SELECT DISTINCT z.owner_name
    || '.'
    || z.table_name
    || '('
    || z.column_name
    || ') - '
    || z.column_type
    || ' - '
    || z.error_type
    || ' ' NotHandledDataDictColumns
    FROM csmig.csmv$errors z
    WHERE z.owner_name IN
    (SELECT DISTINCT username FROM csmig.csm$dictusers
    ) minus
    SELECT DISTINCT z.owner_name
    || '.'
    || z.table_name
    || '('
    || z.column_name
    || ') - '
    || z.column_type
    || ' - '
    || z.error_type
    || ' ' DataDictConvCLob
    FROM csmig.csmv$errors z
    WHERE z.error_type ='CONVERTIBLE'
    AND z.column_type = 'CLOB'
    AND z.owner_name IN
    (SELECT DISTINCT username FROM csmig.csm$dictusers
    ORDER BY NotHandledDataDictColumns
    /Sorry to have made the thread so big but to make sure and give a complete picture of the issue pasted the csscan contents. Request the PRO's to help us in this issue.

    You have convertible data in the application tables. CLOB or not, such data prevents csalter.plb from changing the character set.
    You are on 11.2.0.3, so use the DMU (http://www.oracle.com/technetwork/products/globalization/dmu/overview/index.html). It can cope with such data.
    -- Sergiusz

  • Character set Conversion (US7ASCII to AL32UTF8) -- ORA-31011 problem

    Hello,
    We've run into some problems as part of our character set conversion from US7ASCII to AL32UTF8. The latest problem is that we have a query that works in US7ASCII, but after converting to AL32UTF8 it no longer works and generates an ORA-31011 error. This is very concerning to us as this error indicates an XML parsing problem and we are doing no XML whatsoever in our DB. We do not have XML columns (nor even CLOBs or BLOBs) nor XML tables and it's not XMLDB.
    For reference, we're running 11.2.0.2.0 over Solaris.
    Has anyone seen this kind of problem before?
    If need be, I'll find a way to post table definitions. However, it's safe to assume that we are only using DATE, VARCHAR2 and NUMBER column types in these tables. All of the tables are local to the DB.
    Thanks

    We converted using the database using scripts I developed. I'm not quite sure how we converted is relevant, other than saying that we did not use the Oracle conversion utility (not csscan, but the GUI Java tool).
    A summary:
    1) We replaced the lossy characters by parsing a csscan output file
    2) After re-scanning with csscan and coming up clean, our DBA converted the database to AL32UTF8 (changed the parameter file, changing the character set, switched the semantics to char, etc).
    3) Final step was changing existing tables to use char semantics by changing the table schema for VARCHAR2 columns
    Any specific steps I cannot easily answer, I worked with a DBA at our company to do this work. I handled the character replacement / DDL changes and the DBA ran csscan & performed the database config changes.
    Our actual error message:
    ORA-31011: XML parsing failed
    ORA-19202: Error occurred in XML processing
    LPX-00210: expected '<' instead of '�Error at line 1
    31011. 00000 - "XML parsing failed"
    *Cause:    XML parser returned an error while trying to parse the document.
    *Action:   Check if the document to be parsed is valid.
    Error at Line: 24 Column: 15
    This seems to match the the document ID referenced below. I will ask our DBA to pull it up and review it.
    Please advise if more information is needed from my end.

  • Changing database character set from US7ASCII to AL32UTF8

    Our database is running on Oracle database 10.1.0.4.0 (AIX) The following are its parameters:
    SQL> select value from NLS_DATABASE_PARAMETERS where parameter='NLS_CHARACTERSET';
    VALUE
    US7ASCII
    We would like to change the database character set to AL32UTF8. After following Metalink notes: 260192.1 (which helped us resolve "Lossy" and "Truncated" data, the final output of the CSSCAN utility is:
    [Scan Summary]
    All character type data in the data dictionary are convertible to the new character set
    All character type application data are convertible to the new character set
    [Data Dictionary Conversion Summary]
    The data dictionary can be safely migrated using the CSALTER script
    We have no (0) Truncation and Lossy entries on the .txt file. We only have Changeless and Convertible. Now accdg to the documentation, we can do a FULL EXP and FULL IMP. But it did not detail how to do the conversion on the same database. The discussion on the document tells how to do it from one database to another database. But how about on the same database?
    We cannot use CSALTER as stated on the document.
    (Step 6
    Step 12
    12.c) When using Csalter/Alter database to go to AL32UTF8 and there was NO "Truncation" data, only "Convertible" and "Changeless" in the csscan done in point 4:)
    After performing a FULL export of the database, how can we change its character set? What do we need to do the the existing database to change its character set to AL32UTF8 before we import back our dump file into the same database?
    Please help.

    There you are! Thanks! Seems like I am right in my understanding about the Oracle Official Documentation. Thanks!
    Hmmmmm...when you say:
    *"you can do selective export of only convertible tables, truncate the tables, use CSALTER, and re-import."*
    This means that:
    1. After running csscan on database PROD, i will take note of the convertible tables in the .txt output file.
    2. Perform selective EXPORT on PROD (EXP the convertible tables)
    3. Truncate the convertible tables on PROD database
    4. Use CSALTER on PROD database
    5. Re-import the tables into PROD database
    6. Housekeeping.
    Will you tell me if these steps are the correct one? Based on our scenario: This is what i have understood referring to the Official Doc.
    Am i correct?
    I really appreciate your help Sergiusz.

Maybe you are looking for