US7ASCII and WE8 character set compatibility

Hello,
I want to build two separates environments : one for development, one for production.
I have currently installed OWB 9.2.0.2.8 on a WE8 character set database version 9.2.0.5. This is the development database that contains the design repository, the runtime repository and the targets schemas.
The production database on which I'll run the packages is an US7ASCII database wersion 9.2.0.5. so, again, I will install on it the design repository, the runtime repository and the targets schemas.
My strategy is to export a MDL file for each project, from the development repository and then import it into the production repository. Then, I will change and register the locations and generate/deploy the project to the production target schema.
my questions are :
- Is is possible to import into a US7ASCII database (production) a MDL coming from a WE8 (development) database ?
- Can the generated packages process with some sources tables in an WE8 database even if the target is a US7ASCII table ? if yes, are the french accents still supported ?
Is my stategy correct or must I work with similar character set databases?
Thanks

Hello,
We finally created a new US7ASCII database. Then, we have installed OWB in this database.
OWB version must be greater than 9.2.0.4.
1) from the WE8 repository we export the metadata in US7ASCII characterset to a file
2) from the US7ASCII characterset repository we import the metadata file just created.
==> In the US7ASCII characterset repository , all the characters with accentuation are replaced with others characters. For example é become a "["
==> we created a mapping in the US7ASCII characterset repository. The source is a WE8 characterset table. At execution the accentuation characters are translated to a non accentuation character. For example, the "é" character become an "e"
In conclusion we will not use any accent characters in our repository.
regards

Similar Messages

  • Need suggestion on Multi currency and Unicode character set use in ABAP

    Hi All,
    Need suggestion. In one of the requirement I saw 'multi-currency and Unicode character set experience in FICO'.
    Can you please elaborate me how ABAPers are invlolved in multi currency as I think this is FICO fuctional area.
    And also what is Unicode character set exp.? Please give me some document of you have any.
    Thanks
    Sreedevi
    Moderator message - This isn't the place to prepare for interviews - thread locked
    Edited by: Rob Burbank on Sep 17, 2009 4:45 PM

    Use the default parser.
    By default, WebLogic Server is configured to use the default parser and transformer to parse and transform XML documents. The default parser and transformer are those included in the JDK 5.0.
    The built-in WebLogic Server DOM factory implementation class is com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl.
    The DocumentBuilderFactory.newInstance method returns the built-in parser.
    DocumentBuilderFactory dbf = DocumentBuilderFactory.newInstance();

  • How to find Database, APPL_TOP and IANA character set on 11i?

    Hi,
    Could you please tell How to find out Database character set, APPL_TOP character set and IANA character set on existing 11i environment?
    This is required to pass the input during R12 upgrade.
    Regards,
    AV

    Database:
    SQL> select value
    from V$NLS_PARAMETERS
    where parameters='NLS_CHARACTERSET';
    Application:
    $ echo $NLS_LANG
    IANA:
    Check the value of "s_iana_cset" context variable in the context file or check the value of "ICX:Client IANA Encoding" profile option.
    NLS Frequently Asked Questions [ID 399789.1]
    Oracle Applications 11i Internationalization Guide [ID 333785.1]
    How autconfig determines the value for Iana Charsets s_iana_cset value set in XML context file [ID 1380683.1]
    Thanks,
    Hussein

  • [urgent] oracle character set and national character set !!(dictionary)

    Hi. everyone.
    What is the oracle dictionary that contains information of
    oracle character set and national character set?
    I checked v$database, but there was not the information.
    It seems that there are some differences between "nls_* " init parameters
    and the database character set.
    "Alter database backup controlfile to trace" gave me the character set of db,
    but I would like to know whether there are oracle dictionary regarding them.
    Thanks in advance. Have a nice day.
    Best Regards.

    I found the dictionary which contains the information of character set and
    natiional character set of database.
    select * from nls_database_parameters
    where parameter like '%CHARACTERSET';
    Thanks for reading.
    Have a good day.
    Best Regards.

  • Expdp with network_link and different character sets for source and target?

    DB versions 10.2 and 10.1
    Is there any limitation for implementing expdp with network_link but target and source have different character sets?
    I tried with many combinations, but only had success when target and source have the same character sets. In other combinations there was invalid character set conversion (loose of some characters in VARCHAR2 and CLOB fields).
    I didn't find anything on the internet, forums or Oracle documentation. There was only limitation for transportable tablespace mode export (Database Utilities).
    Thanks

    Hi DBA-One
    This link
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm#CEGFCFFI
    is from Database Utilities 10g Release 2 (10.2) (I read it many times)
    There is only one thing about different character sets.
    I quote:
    Data Pump supports character set conversion for both direct path and external tables. Most of the restrictions that exist for character set conversions in the original Import utility do not apply to Data Pump. The one case in which character set conversions are not supported under the Data Pump is when using transportable tablespaces.
    Parameter VERSION also doesn't play role here, because the behaviour (character set) is the same if I use for target and source only 10.2 or (10.2 and 10.1) respectivly.

  • BIG5 and HKSCS Character Set Support

    Hi,
    We're experiencing some problems inserting a string containing both BIG5 and HKSCS characters to a 7.3.4 Oracle DB using JDBC. The underlying character set used by the DB is ZHT16BIG5 (this cannot be changed). The characters can be inserted correctly if we use SQLPlus/WorkSheet.
    Take note that the BIG5 character set can be inserted correctly. The problem occurs if we include HKSCS characters in the statement.
    We have tried a number of ways already but failed to convert the data properly.
    We tried converting the data using ByteToCharConverter.getConverter("Big5") but this cannot handle the HKSCS properly.
    We even tried using the CharacterSet.ZHT16BIG5_CHARSET provided by the NLS character set but it cannot convert all HKSCS characters correctly.
    Any ideas on how to solve this problem? Or is it because the HKSCS character set is NOT supported by the JDBC driver?
    Below is a sample text containg both BIG5 and HKSCS characters:
    'i$h%49D$G$Q$T89     Ize     _     ^     S(     R     @     A     Y     q
    Any help/suggestion is most welcome.
    Thanks,
    Cis
    null

    I got the exact same problem as you.
    (The Oracle I using is 8.1.7)
    Can any one help??

  • USER_FILTER and database character set

    Hello,
    I'm currently working on the integration of a tool into Oracle Text for filtering PDFs. My current approach is to call a command line tool via a USER_FILTER preference, and this works fine as long as the database character set is AL32UTF8. The tool is creating the filtered text as UTF-8.
    I'm struggling now with the case that the database character set is not Unicode, for example LATIN1. I had hoped that I can specify a chain of filters for this situation when creating the index, first a USER_FILTER to get the text out of the document and then a CHARSET_FILTER to convert the filtered text from UTF-8 into the database character set. This is my attempt to set this up:
    execute ctx_ddl.create_preference ('my_pdf_datastore', 'file_datastore')
    execute ctx_ddl.create_preference ('my_pdf_filter', 'user_filter')
    execute ctx_ddl.set_attribute ('my_pdf_filter', 'command', 'tetfilter.bat')
    execute ctx_ddl.create_preference('my_cs_filter', 'CHARSET_FILTER');
    execute ctx_ddl.set_attribute('my_cs_filter', 'charset', 'UTF8');
    create index tetindex on pdftable (pdffile) indextype is ctxsys.context parameters ('datastore my_pdf_datastore filter my_pdf_filter filter my_cs_filter');
    These are the error messages I'm getting (sorry, German Windows):
    FEHLER in Zeile 1:
    ORA-29855: Fehler bei Ausf³hrung der Routine ODCIINDEXCREATE
    ORA-20000: Oracle Text-Fehler:
    DRG-11004: Doppelter oder unvereinbarer Wert f³r FILTER
    ORA-06512: in "CTXSYS.DRUE", Zeile 160
    The relevant message is DRG-11004, which translates to "duplicate or incompatible value for FILTER".
    ORA-06512: in "CTXSYS.TEXTINDEXMETHODS", Zeile 364
    So here is my question:
    Do I understand it correctly that with the USER_FILTER the text is always expected in the database encoding, and that it is mandatory to create the filtered text in the database character set, or are there any alternatives?
    Thanks
    Stephan

    The previous experiments were performed with Oracle 10i. I just saw that in Oracle 11.1.0.7 there is this new feature: "USER_FILTER is now sensitive to FORMAT and CHARSET columns for better indexing performance.".
    This seems to be exactly what I was looking for.
    Regards
    Stephan

  • Lanovo ix4 300d and european character set for samba shares

    Hi,
    I am installing a couple of IX4-300d at a client site, struggling to set and fix the samba character set for the Windows shares. Here in Italy everyone is used to put western european characters into filenames (for example è ò ° €) but the NAS acts wierdly when it finds such caracters in folder names and file names.
    As you know samba gives the possibility to declare dos charset= ISO8859-1 and unix charset= ISO8859-1 in the smb.conf file but the option is not present in the IX4 setup pages.
    Is there an easy way to circumvent the problem and fix this option? As it is now the machine is fearly useless, at least if you need to migrate files and folder from a server or storage with ISO8859 charset already applied.
    thanks a lot
    np
    Solved!
    Go to Solution.

    This issue has been raised on all iomega drives since day 1. although 'documented' it is still shame, that units sold to 'r-o-w' means rest-of-world, outside US, are stuckto Aa-Zz09 basically wheras client systems ( windows, linux, ios ) don't care and take what they are supposed to do.
    Take it or leave it. ( I mean, take another vendor), I do not believe someone ever has taken that seriously or will do it......
    Various PCs / Laptops ( sorry I still really love Dell and Fujitsu ;-))
    Supporting Customers ix2s and ix4s -- Love Networking ( not only technically ).
    I am not a Lenovo Employee.
    If you find a post helpful and it answers your question, please mark it as an "Accepted Solution"!

  • 9iLite and multibyte character set support

    Does 9iLite support a character set that will allow for accented characters?
    for example: i

    "NLS Character Integrity Issues for Consolidator
    When Mobile Sync synchronizes with an Oracle database which has a
    multibyte character set other than UTF8, the character integrity issue
    occurs. Mobile Sync retrieves data from the server database through Oracle
    8.1.7 OCI JDBC Driver for Oracle9iAS version 1.0.2.2, and 9i for Oracle9iAS
    version 2.0. Character sets are converted from database character sets to
    UTF8 by Oracle Servers NLS functions. In the code conversion, some
    multibyte characters are garbled because of the difference of the character
    mapping. This is not a bug of Mobile Sync.
    For more Information, see "Character Integrity Issues in NLS Environment"
    technical paper on Oracle Technology Network (technet.oracle.com)
    Java/SQLJ & JDBC section in Technologies category."
    from the Readme file with the media (read the manual I guess)

  • Discoverer 3I and Cyrilic character set

    I have a discoverer 3I EUL set up and I'm trying to pull data out that is stored in the database using the cyrilic character set. All I get, however, when I view the data is ? substituted for the cyrilic character. Is discoverer able to handle cyrilic and if so how do I get it to show the actual russian character?

    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Dave Clark ([email protected]):
    I have a discoverer 3I EUL set up and I'm trying to pull data out that is stored in the database using the cyrilic character set. All I get, however, when I view the data is ? substituted for the cyrilic character. Is discoverer able to handle cyrilic and if so how do I get it to show the actual russian character?<HR></BLOCKQUOTE>
    This is possibly a patch issue. For 3i this requires version 3.3.59.
    null

  • Oc4j and international character sets

    Does anyone know if there is a problem in oc4j concerning international charactersets support?
    My pages contain non-english characters and so i use utf-8 characterset. When the page is created by a servlet the characters are displayed correctly, but when it comes from a jsp this is not the case.
    The problem is that the text is not displayed in utf-8 but in "western european (windows)" instead which is not the correct encoding. This does not happen in the servlet case above. Still if I change the borwser encoding setting
    the information is displayed correctly. But I cannot do that for every page, they are to many...
    If i try to instruct the page about the encoding the outcome is a real mess because no brwser characterset encoding configuration option seems to be suitable. I tried the following:
    - the page directive: <%@page contentType=... pageEncoding...
    - the request: request.setContentType
    - oc4j configuration: default-charset configuration in the global-application.xml config file
    - and the oracle extension <% serwriterEncoding(...) ; %>
    all of the above had no result while servlets perform swell !
    Am I missing something here?
    Does any body knows something that I do not?
    Does the jsp container mess with the encoding and sends to the browser incorrect page encoding?
    Thank you in advance,
    Joe

    HI Joe:
    Are you able to find any solution for this. I am also stuck with the same issue. Could you please share your solution to me. my email id is: [email protected]
    Thanks & Regards
    Sridhar Doki

  • Use of UTF8 and AL32UTF8 for database character set

    I will be implementing Unicode on a 10g database, and am considering using AL32UTF8 as the database character set, as opposed to AL16UTF16 as the national character set, primarily to economize storage requirements for primarily English-based string data.
    Is anyone aware of any issues, or tradeoffs, for implementing AL32UTF8 as the database character set, as opposed to using the national character set for storing Unicode data? I am aware of the fact that UTF-8 may require 3 bytes where UTF-16 would only require 2, so my question is more specific to the use of the database character set vs. the national character set, as opposed to differences between the encoding itself. (I realize that I could use UTF8 as the national character set, but don't want to lose the ability to store supplementary characters, which UTF8 does not support, as this Oracle character set supports up to Unicode 3.0 only.)
    Thanks in advance for any counsel.

    I don't have a lot of experience with SQL Server, but my belief is that a fair number of tools that handle SQL Server NCHAR/ NVARCHAR2 columns do not handle Oracle NCHAR/ NVARCHAR2 columns. I'm not sure if that's because of differences in the provided drivers, because of architectural differences, or because I don't have enough data points on the SQL Server side.
    I've not run into any barriers, no. The two most common speedbumps I've seen are
    - I generally prefer in Unicode databases to set NLS_LENGTH_SEMANTICS to CHAR so that a VARCHAR2(100) holds 100 characters rather than 100 bytes (the default). You could also declare the fields as VARCHAR2(100 CHAR), but I'm generally lazy.
    - Making sure that the client NLS_LANG properly identifies the character set of the data going in to the database (and the character set of the data that the client wants to come out) so that Oracle's character set conversion libraries will work. If this is set incorrectly, all manner of grief can befall you. If your client NLS_LANG matches your database character set, for example, Oracle doesn't do a character set conversion, so if you have an application that is passing in Windows-1252 data, Oracle will store it using the same binary representation. If another application thinks that data is really UTF-8, the character set conversion will fail, causing it to display garbage, and then you get to go through the database to figure out which rows in which tables are affected and do a major cleanup. If you have multiple character sets inadvertently stored in the database (i.e. a few rows of Windows-1252, a few of Shift-JIS, and a few of UTF8), you'll have a gigantic mess to clean up. This is a concern whether you're using CHAR/ VARCHAR2 or NCHAR/ NVARCHAR2, and it's actually slightly harder with the N data types, but it's something to be very aware of.
    Justin

  • Transport tablespaces between databases with different character sets

    Hi everyone:
    I have two 10R2 databases on the same hp-ux 64bit server, 1st one with NLS_CHARACTERSET=US7ASCII, 2nd one with
    NLS_CHARACTERSET=AL32UTF8.
    NLS_NCHAR_CHARACTERSET on both databases is AL16UTF16.
    Can I transfer tablespaces from the 1st one to the 2nd. The data could be in English, French & Spanish.
    If not what are my options?
    Thanks in advance.

    First off, if you are storing French and Spanish data in database 1 where the character set is US7ASCII, you've got some serious problems. US7ASCII doesn't support non-English characters (accents, tildes, etc). If you're storing data this way, you've introduced data corruption that you'd have to resolve before copying the data data over to another machine.
    Second, technically, the source and target character set have to be identical. Since US7ASCII is a strict binary superset of AL32UTF8, you could theoretically transport a US7ASCII tablespace to an AL32UTF8 database. In your case, though, since the data is not really US7ASCII, you'd end up with corruption.
    Any of the Oracle built-in replication options is going to require that you resolve the corruption issue. Assuming that you can figure out what character set the source database really is, you could potentially dump the data to flat files (taking care not to allow character set conversion to take place) and SQL*Loader them into the destination system by identifying the proper character set in your control file. That's obviously going to be a rather laborious process, though.
    Justin

  • Character set mismatch during import

    I have exported a tablespace from a database using transportable tablespace option
    with CHARACTERSET UTF8. And I tried to import into another db where the CHARACTERSET is US7ASCII. Now it says that it cannot import because the charactersets are different. Is there any way to get around this problem?
    Regards,
    Kallur Manoj.

    Manoj,
    One of the limitations for transportable tablespace is that the source and target databases should use the same character set and national character set.
    You may need to either export the tablespace with us7ascii, or import to a db with utf8 characterset.
    AGK

  • Us7ascii and we8iso8859p1 encoding problem.

    Hi frnds,
    I have a oracle 8i database with character set US7ASCII. In one of the tables in the database tif images are stored in a LONG column. I want to fetch those images using oracle odbc drivers and visual basic.
    I have oracle 9i client and client character set is WE8ISO8859P1.When i fetch data from the database, i m not able to fetch the images correctly. Special characters appear as blocks or '?'.
    I suspect this is due to character set problem. What do i do to solve this problem?

    Migrate the data to LONG RAW or BLOB. Images should not be stored in LONG.
    Oracle ODBC Drivers starting from 8.1.5.5 are Unicode drivers and internal conversion to Unicode will most probably corrupt your data.
    Do not try to solve the problem by cheating. It may help for a moment but will cause problems later on, e.g. when you start using JDBC.
    -- Sergiusz

Maybe you are looking for

  • Update according to a condition

    how do i update a table according to a condition (i.e. dynamically update). I have a table which i need to insert a new record for each customer if they have purchased a specific item before. my purchase table has the following columns (id, cust id,

  • Committing Database Transactions

    The workflow user guide advises not to issue commits in procedures called from the workflow since it will commit workflow-specific data as well. Does the workflow engine issue a commit after each activity, e.g. if PROCEDURE1 inserts a record into a t

  • Can I use the iPod Nano cable and AC power supply with the iPhone?

    Just purchased my iPhone 3G. Phone was in my pocket and the AT&T clerk neglected to put the box with the computer cable in the bag with the other accessories. Until I can pick it up, can I use my iPod Nano cable? The plug seems the same size but I do

  • Event Trigger for Plan data

    Hi , I have a scenario where my user wants to see the plan numbers instantly in the report after he saves it in the Layout. The report is not directly on the planning cube there are 2 levels of staging on top of which is the actual report. He is the

  • HT4623 There is no option to download to the latest iOS software

    I go to settings> general and there are no buttons to download the latest version of iOS. I am unable to update or download apps because I need ios5 or higher