Import dump from a satabase with a different character set

My database has this character set:
select * from database_properties:
NLS_CHARACTERSET     AL32UTF8     Character set
NLS_NCHAR_CHARACTERSET     AL16UTF16     NCHAR Character set
I need to import a dump from a database with WE8MSWIN1252 character set.
After the import I have seet that some character in the table are wrog:
I see this simbol " " insted of this "à".
How can I solve the problem?
The nls_lang variable on my os is: NLS_LANG=ITALIAN_ITALY.AL32UTF8
I work with oracle 10.0.4 on linux
Message was edited by:
user613483

I have read thos doc on metalink: Note:227332.1
I also tried to set nls_lang = NLS_LANG=ITALIAN_ITALY.WE8MSWIN1252
and then I execute the import command.
BUt didn't work.

Similar Messages

  • Migrating from 8.1.7 to 9.2 with different character set

    Are there going to be issues when I try to migrate from 8.1.7 to 9.2 where the character set in the 8i database was the default western european to utf8 in 9i?
    I ran the 8i scan utility to see if there would be issues importing into an 8i db with a different char set and it said that there would.

    Hi Jim,
    As your first step, I would recommend reading the white paper Character Set Migration Best Practices for Oracle9i on the Globalization Support Home page - http://otn.oracle.com/tech/globalization/index.html
    Since you mentioned that CSSCAN has already reported there are potential issues (I guess you have exceptional/Lossy data ?), I would strongly recommend splitting up the UTF8 character set change and the database version upgrade into 2 distant phrases. Probably moving to UTF8 in 8.1.7 first, then upgrading to 9.2 later.
    BTW, there is a separate Globlization Support and NLS discussion forum on OTN for further discussion on character set related issues.
    Nat

  • Importing from a different character set

    Oracle 8.1.7 / Windows NT
    I'm trying to import a dump file which was created with character set WE8ISO8859P9. My database uses character set UTF8. Some of the records can't be inserted because of error "ORA-1401: Value too large for column". Is this because of the different character sets? If I switch my session to WE8ISO8859P9, imp says "character set conversion from x to y not supported."
    How can I get these last records inserted? Here's an excerpt from the log:
    Verbunden mit: Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    <
    Export-Datei wurde von EXPORT:V08.00.05 |ber konventionellen Pfad erstellt
    Warnung: Die Objekte wurden von NOC_ADMIN exportiert, nicht von Ihnen.
    Importvorgang mit Zeichensatz WE8ISO8859P9 und Zeichensatz UTF8 NCHAR durchgef|hrt
    Import-Server verwendet Zeichensatz UTF8 (mvgliche Zeichensatzkonvertierung)
    Export-Server verwendet Zeichensatz WE8ISO8859P9 NCHAR (mvgliche Zeichensatzkonvertierung)
    . Import NOC_ADMIN's Objekte in NOC_ADMIN
    . . Import der Tabelle "ACCESSROUTERIFS_" 782 Zeilen importiert
    . . Import der Tabelle "ITEM_"
    IMP-00019: Zeile zur|ckgewiesen aufgrund von Oracle-Fehler 1401
    IMP-00003: Oracle-Fehler 1401 gefunden
    ORA-01401: Eingef|gter Wert zu gro_ f|r Spalte
    Spalte 1 33886
    Spalte 2
    Spalte 3
    Spalte 4 1323
    Spalte 5
    Spalte 6 11
    Spalte 7 18600
    Spalte 8 18600
    Spalte 9 20-NOV-2000:00:00:00
    Spalte 10 processing
    Spalte 11 inactive
    Spalte 12
    Spalte 13
    Spalte 14 35682.0
    Spalte 15
    Spalte 16
    Spalte 17
    Spalte 18 05.12.00: KD weiss noch nix neues, er wird uns inf...
    Spalte 19
    Spalte 20 kschmid
    Spalte 21 09-FEB-2001:15:50:21
    Spalte 22
    Spalte 23 12
    Spalte 24
    Spalte 25 06-NOV-2000:00:00:00
    null

    Please try ORacle RDBMS support. this issues is to do with Oracle Import.

  • How do I import directly from Canon XA10 with Final Cut X?

    How do I import directly from Canon XA10 with Final Cut X?

    I have an XA10, and importing is quite simple.  But you do have to make sure your camera is set to the proper settings to allow you to import the files into FCPX.
    First, make sure the camera is powered on.  Connect it through the USB connection.  Then open the camera's viewfinder and press the button on the side of the camera that allows you to see the videos you've shot in the viewfinder.  (This toggle button toggles between seeing what the camera is pointed at live and seeing the thumbnails of the video you've already shot that are stored on the internal memory card.)  Now, the camera will present you with a menu asking which memory card you want to look at (there's an internal flash memory card, plus you can insert SD cards as well).  Choose the memory card on which your videos are stored within the camera, and then you'll be good to go.
    Now you can go into your FCPX event and do the Import from Camera.  Now the Import from Camera interface will open up and you'll be able to select the XA10, where it will present you with all the videos that are on the camera's memory card that you selected.
    From here,  you can select individual videos and import them, or you can set in and out points on videos and just import the parts you want.

  • Expdp with network_link and different character sets for source and target?

    DB versions 10.2 and 10.1
    Is there any limitation for implementing expdp with network_link but target and source have different character sets?
    I tried with many combinations, but only had success when target and source have the same character sets. In other combinations there was invalid character set conversion (loose of some characters in VARCHAR2 and CLOB fields).
    I didn't find anything on the internet, forums or Oracle documentation. There was only limitation for transportable tablespace mode export (Database Utilities).
    Thanks

    Hi DBA-One
    This link
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm#CEGFCFFI
    is from Database Utilities 10g Release 2 (10.2) (I read it many times)
    There is only one thing about different character sets.
    I quote:
    Data Pump supports character set conversion for both direct path and external tables. Most of the restrictions that exist for character set conversions in the original Import utility do not apply to Data Pump. The one case in which character set conversions are not supported under the Data Pump is when using transportable tablespaces.
    Parameter VERSION also doesn't play role here, because the behaviour (character set) is the same if I use for target and source only 10.2 or (10.2 and 10.1) respectivly.

  • Query distributed database with different character sets.

    Hello experts, this is my situation:
    I have two databases A and B, the same version 11.1.0.7, the same OS Suse Linux Enterprise 10 but with different character sets, A has WE8MSWIN1252 while B has AL32UTF8. The database B is my XML DB repository so there I have some XML type tables. I need to query this tables from the database A using a dblink and in fact I have done that but the XML content is trasformed due to the different character sets between the databases. Some time there are data loss and some time there are data missmatch.
    Is there any way to query the tables stored in the database B without problems? I do not know if the following is correct: Maybe I can set the character set for the session in the database A during the time it query the database B. That is, change the character set in fly at session level.
    Do you have any special suggestion?
    I hope you can help me, thank you in advance.

    The Globalization Support Guide for 11.1.0.7 has a chapter on character set migration that should be helpful. AL32UTF8 is a superset of WE8MSWIN1252 but it is not a strict superset. That is, it doesn't meet the second prong of the test in the documentation
    The new character set is a strict superset of the current character set if:
    Each and every character in the current character set is available in the new character set.
    Each and every character in the current character set has the same code point value in the new character set. For example, many character sets are strict supersets of US7ASCII.Exporting the data from the A, changing the character set (or creating a new database with the AL32UTF8 character set), and then importing the data may be the easiest approach in your case.
    Justin
    Edited by: Justin Cave on Jan 13, 2011 12:08 PM

  • NLS settings for a database link between DBs with different character sets

    I am using a database link to move data from one database to another and I am seeing some strange data problems. The databases have different character sets and different NLS settings. I wonder if this could be causing my problem.
    Here are the NLS parameters for the database where the database link exists. (the SOURCE database)
    1     NLS_CALENDAR     GREGORIAN
    2     NLS_CHARACTERSET     WE8MSWIN1252
    3     NLS_COMP     BINARY
    4     NLS_CURRENCY     $
    5     NLS_DATE_FORMAT     DD-MON-RR
    6     NLS_DATE_LANGUAGE     AMERICAN
    7     NLS_DUAL_CURRENCY     $
    8     NLS_ISO_CURRENCY     AMERICA
    9     NLS_LANGUAGE     AMERICAN
    10     NLS_LENGTH_SEMANTICS     BYTE
    11     NLS_NCHAR_CHARACTERSET     AL16UTF16
    12     NLS_NCHAR_CONV_EXCP     FALSE
    13     NLS_NUMERIC_CHARACTERS     .,
    14     NLS_SORT     BINARY
    15     NLS_TERRITORY     AMERICA
    16     NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    17     NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    18     NLS_TIME_FORMAT     HH.MI.SSXFF AM
    19     NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    Here are the NLS parameters for the database that the database link connects to. (the TARGET database)
    1     NLS_CALENDAR     GREGORIAN
    2     NLS_CHARACTERSET     AL32UTF8
    3     NLS_COMP     BINARY
    4     NLS_CURRENCY     $
    5     NLS_DATE_FORMAT     DD-MON-RR
    6     NLS_DATE_LANGUAGE     AMERICAN
    7     NLS_DUAL_CURRENCY     $
    8     NLS_ISO_CURRENCY     AMERICA
    9     NLS_LANGUAGE     AMERICAN
    10     NLS_LENGTH_SEMANTICS     BYTE
    11     NLS_NCHAR_CHARACTERSET     AL16UTF16
    12     NLS_NCHAR_CONV_EXCP     FALSE
    13     NLS_NUMERIC_CHARACTERS     .,
    14     NLS_SORT     BINARY
    15     NLS_TERRITORY     AMERICA
    16     NLS_TIMESTAMP_FORMAT     DD-MON-RR HH.MI.SSXFF AM
    17     NLS_TIMESTAMP_TZ_FORMAT     DD-MON-RR HH.MI.SSXFF AM TZR
    18     NLS_TIME_FORMAT     HH.MI.SSXFF AM
    19     NLS_TIME_TZ_FORMAT     HH.MI.SSXFF AM TZR
    The SOURCE database version is 10g Release 10.2.0.3.0 - Production
    The TARGET database version is 11g Release 11.1.0.6.0 - 64bit Production
    Do I need to modify the NLS settings in the SOURCE database before executing a script to insert data into the TARGET database?
    Thanks, Jack

    The difference in settings is not a problem by itself, especially that only the NLS_CHARACTERSET matters. Of course, this difference may lead to certain issues if not taken into consideration.
    Please, describe symptoms of your problems.
    -- Sergiusz

  • Searching with different character sets

    Hello,
    we have a problem with Intermedia 8.1.6. running on Solaris.
    The table contains the text with different character sets and that's the problem. User submits the query in his char.set and the IM sometimes doesn't find the data.
    Idea is to create the index using the flat ascii chars and to search in ascii ... but how?
    Can anybody help me?
    Thanks.
    Zozzi
    null

    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Zozzi ([email protected]):
    sorry, wrong email in the prev msg ...
    this one is correct<HR></BLOCKQUOTE>
    Hello,
    Did you solve it ?
    If yes, how to do it. I am interested in knowing it.
    Many Thanks
    null

  • Different Character sets (Codepages) in the same Login

    Hi,
    I need to enter different character sets (Russian and English) in the same login and save the different characters to SAP database tables.
    The codepage for Russian is different from English codepage. I have Russian codepage language settings on both Windows and SAP. Can anybody let me know how to proceed with this.
    Any pointers on this would be really helpful.
    Thanks and Regards,
    Sirisha

    Hi,
    I need to enter different character sets (Russian and English) in the same login and save the different characters to SAP database tables.
    The codepage for Russian is different from English codepage. I have Russian codepage language settings on both Windows and SAP. Can anybody let me know how to proceed with this.
    Any pointers on this would be really helpful.
    Thanks and Regards,
    Sirisha

  • How to include different Character Sets in Flex AIR project

    Hi All,
    Flex, AIR(Desktop) Project.
    PROBLEM Scenario:
    My Application needs a path from Windows registry, So I am launching a .cmd file using Native Process API.
    As a result I get a text file with the path.
    So, everything works fine(reading/writing), but if the path contains different languages like Korean, Chinese then the READING OPERATION IS FAILED only when I install the application. It works fine while debugging.
    (By failing I mean its not getting the characters, its printing some junk values.)
    I am using readMultibyte & writeMultibyte APIs, and I am assiging the character set dynamically as per the OS language. (Which is Enslish in current scenario)
    FINDING:
    I installed Windows with Korean Language on a system, so now if I build the .exe on this machine then it works fine. So, i guess its not able to get the Korean Character set from the English OS, thats why it failed in earlier case.
    Is there any way to include different character sets in a Flex project?
    Regards,
    Mayank

    Some loaded SWFs may not be compatible with Flex.  Try making a simpler SWF (like one that just draws a circle where you click) and see if it can be loaded and interacted with.
    -Alex

  • Different Character sets?

    Oracle version: 11.1.0.7.0
    There are any problems if I configure an Oracle Streams Replication between to database with different character set? I have configured a propagation between two database A and B. The A character set is WE8MSWIN1252 and the B character set is WEISO8859P1 but when the propagation passes about 30 minutes configured and ora-07445 begin. Also when I do the same configuration between two database with the same character set the problems does not appear.
    Here I write the error in the alert log:
    ORA-07445: se ha encontrado una excepción: volcado de memoria [kohrsmc()+1484] [SIGSEGV] [ADDR:0x18] [PC:0x7CCC92E] [Address not mapped to object] []
    Incident details in: /opt/oracle/diag/rdbms/bdmdic/bdmdic2/incident/incdir_110004/bdmdic2_j000_18185_i110004.trc
    Fri Aug 06 16:57:31 2010
    Trace dumping is performing id=[cdmp_20100806165731]
    Exception [type: SIGILL, Illegal operand] [ADDR:0x2ACE05EEFE60] [PC:0x2ACE05EEFE60, {empty}]
    Errors in file /opt/oracle/diag/rdbms/bdmdic/bdmdic2/trace/bdmdic2_j000_18185.trc (incident=110005):
    ORA-07445: se ha encontrado una excepción: volcado de memoria [PC:0x2ACE05EEFE60] [SIGILL] [ADDR:0x2ACE05EEFE60] [PC:0x2ACE05EEFE60] [Illegal operand] []
    ORA-07445: se ha encontrado una excepción: volcado de memoria [kohrsmc()+1484] [SIGSEGV] [ADDR:0x18] [PC:0x7CCC92E] [Address not mapped to object] []
    Incident details in: /opt/oracle/diag/rdbms/bdmdic/bdmdic2/incident/incdir_110005/bdmdic2_j000_18185_i110005.trc

    Where did I said that charset is interfering in propogataion? It will not interfere. Char set difference will cause the target database populated with data which it may not understand and will show you as garbage data. As you replication is working find without an issue for sometime, so it looks like you are hitting some bug when you hit replication for some specific data. To confirm this you may have to work with Oracle to find the exact cause or which bug you may be hitting.
    All my suggestions till this point in time is without knowing your proper environment (Server A,B OS and database version, also the error description except english words are all foreign to me). As suggested you will be better off working with Oracle support to find the resolution. Please let us know what fixed your issue when you work with Oracle to help all forum members.
    Regards

  • Different character sets on database links

    Does someone know, how to select data from different machines with different character sets via database link transferring the umlauts correctly ?
    The destination instance uses we8iso8859p1, the source machines have us7ascii (SAP) or we8iso8859p1. It is NOT allowed to change the character sets on the source databases!

    This is not so easy to solve. The correct procedure is to fix your non-ASCII database to reflect the correct character set encoding of your data. Have you checked with SAP to see whether it is really not possible/supported to change the application db character set ?
    The only workaround that I can think of ,is to create another database with 'WE8ISO8859P1' as the db character set. Populate it with the data you need from your SAP database using either SQL*LOADER or from insert statements generated by SPOOLing out the data using SQL*PLUS

  • ORACLE invoices with a Japanese character set

    We are having trouble printing ORACLE invoices with a Japanese character set.
    the printer we are using is a Dell W5300,do I need to configure the printer or is it something that needs to be configure in the software?????please help......

    We are having trouble printing ORACLE invoices with a
    Japanese character set.
    the printer we are using is a Dell W5300,do I need to
    configure the printer or is it something that needs
    to be configure in the software?????please help......What is the "trouble"? Are you seeing the wrong output? It may not be the printer, but the software that is sending the output to the printer.
    If you are using an Oracle Client (SQL*Plus, FOrms, Reports etc), ensure you set the NLS_LANG to JAPANESE_JAPAN.WE8MSWIN1252 or JAPANESE_JAPAN.JA16SJIS

  • How to import data from CSV file with columns separated by semicolon?

    I migrate database from MS SQL 2008 to ORACLE 11g
    I export data to CSV file from MS SQL
    then I try to import it to Oracle
    several tables goes fine using Import data option in the SQL Developer
    Standard CSV files with data separated by comma were imported.
    chars, date (with format string), and integer data are imported via import wizard without problems
    the problems were when I try to import table with noninteger numbers with modal part separated by comma not by dot
    comma is the standard separator for columns in CSV file
    so I must change the standard separator to semicolon
    then the import wizard have problem to correct recognize the columns data because it use only standard CSV comma separator :-/
    In SQL Developer 1.5.3 Tools -> Preferences -> Migration -> Data Move Options
    I change "End of Column Delimiter" to ; but it doens't work
    Is this possible to change the standard column separator for import data wizzard in SQL Developer 1.5.3?
    Or maybe someone know how to import data in SQL Developer 1.5.3 from CSV when CSV colunn separator is set to semicolon?

    A new preference has been added to customize the import delimiter in main code line. This should be available as part of future release.

  • FR Studio - Is there a workaround for the issue with referencing data from a grid with a different DB connection and different # of dimensions?

    Hi,
    I am trying to reference a column in a different grid with a different database connection. I did some research in the forums, and it seems like this is a known issue with FR Studio. The issue is that if the 2 different database connections have different # of dimensions, then the reference will throw an error.
    For example, what I'm trying to do is below:
    In grid2.column[A], I have the following formula:
    grid1.column[A]
    The result is an error.
    The ultimate goal is to have a basic variance formula in the grid between grid1 and grid2, which have different database connections with a different # of dimensions. Is there a workaround at all to this bug? The example of the basic formula is below:
    grid1.column[A] - grid2.column[A] 
    Any help is much appreciated!
    Thanks,
    Tiffany

    Thanks for the detailed response Mike.
    The issue with trying it in Grid1 vs Grid2 is that Grid1's DB connection has more dimensions than Grid2's DB connection. I can't change the DB connection to a database with more dimensions because you have to be able to map all dimensions from the existing DB connection  (Grid1) to the new DB connection (Grid2). This isn't possible, so instead I have to map Grid2 (less dimensions) to Grid1 (more dimensions). It just ignores the extra dimensions from Grid1. Does that make sense?
    Thanks again for your responses. It's very much appreciated.
    Tiffany

Maybe you are looking for

  • What is the best way to connect my previous PowerG4 to the mac mini?

    I want to utilize my old G4's HD, but I don't know how to do it so that everything runs smooth. Right now it is hooked-up to the mini in "target mode" via firewire but it seems to slow everything down. Is an enclosure the best option and if so, which

  • Compiling code in a program to a java file?

    is there a way to in a program write code to a java file and compile that file? is there a way with a simple .compile or .run or something like that?

  • Displaying an HTML Data Set on button click

    I'm using the code from the HTML DATA Set Example I have the code in a function and I want to have it display in a div on the page when a user clicks a specific button. I get this error message: "spry:region or spry:detailregion attribute has no data

  • Can't delete contacts in Kopete.

    Hi, I am not able to permanently delete Yahoo! contacts in Kopete. When I restart the program the deleted contacts are back. I have tried many times using both "Delete" and "Remove" No go. Anyone else noticed this? I am running Kopete Version 0.80.2

  • Web Service  implementation and consumer in different version

    Currently we have a Web Service which is written in J2EE 5 and this web service are consumed by Web Application written in 1.4 We are using SOAP for connection and JAXB API for parsing. We could like to know if there is any issue which could arise be