Export and Import language setting

Dear all
I export the database by using NLS_LANGUAGE='TRADITIONAL CHINESE' in Sun UNIX box with Oracle 8.0.5 (characterset and n characterset are both 'ZHT16BIG5').
When I tried to import to a WinXP with Oracle 8.1.7, even though I had set the environment variable as 'TRADITIONAL CHINESE'. (i.e. set NLS_LANGUAGE='TRADITIONAL CHINESE') When I tried to import, there is a warning showed that the current import characterset is 'AMERICAN', and character conversion will be occurred. How can I prevent the character conversion during the import. Is there any mistake I had made? It worked in both unix environment before.
Thanks in advanced
Kwan
16 June 2005

Hi,
Well, afaik, when exporting/importing in order to have the language options right, the use of NLS_LANG is recomanded..
Maybe you should try over with something like $ #before export
$ export NLS_LANG=CHINESE_CHINA.ZHT16BIG5
$ exp ...
C:\> REM Before import
C:\> set NLS_LANG=CHINESE_CHINA.ZHT16BIG5
C:\> imp ...I usualy (not using any Chinese charset though) do it this way and have no problem.
HTH,
Yoann.

Similar Messages

  • Is it possible to export and import a set of variables?

    Hi,
    I'm using RH8 under WinXP. I use variables for our product names (=single software modules). This makes it easier to change the name, if a product name has changed. For each of these products there is a single RH-project. Now I want to import a set of variables, that I've created in one of those projects, to another project. Up to now, I haven't found an export/import option in RH for these set of variables. Is there a possibility to import/export these variables?
    Kind regards
    Stefan

    Set up a test project close it, copy the rhvariable.apj file from a project with the variables into it. Reopen the test project.
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • Language Problem while exporting and importing data

    hi,
    I have Oracle version 8.1.7.0.0 installed on one server and 9.2.0.1.0 installed on new server.
    I'm copying and pasting my version info from SQL*Plus:
    SQL*Plus: Release 8.1.7.0.0 - Production on Mon Aug 22 10:46:31 2005
    (c) Copyright 2000 Oracle Corporation. All rights reserved.
    Connected to:
    Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    SQL>
    SQL*Plus: Release 9.2.0.1.0 - Production on Mon Aug 22 12:30:06 2005
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Connected to:
    Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    SQL>
    I created new user on my new server from enterprise manager.
    Exported user from the old server and imported in the new server.
    i.e: from Oracle8i Enterprise Edition Release 8.1.7.0.0, I did
    c:\>exp system/manager file=abc.dmp owner=abc
    Then on the new server Release 9.2.0.1.0, I did
    c:\>imp system/manager file=abc.dmp fromuser=abc touser=abc
    I'm using Arabic Language on my both servers. NLS_LANG parameter on both the servers is AMERICAN_AMERICA.WE8MSWIN1252.
    On both the servers I'm able to insert and select data in arabic.
    However, after I export the data from old server to the new server, the arabic data comes in question marks.
    If I create new table and insert arabic data on new server's user abc it is displaying well. Only the data which I exported and imported is not showing arabic.
    On both old and new servers operating system is Windows XP.
    I'm stuck with this problem. Anybody having any idea about how to solve this problem please help.
    Thank you all in advance.
    Regards

    Let me be clear here. Storing Arabic data in a WE8MSWIN1252 database is not supported by Oracle and will lead to problems. You are incorrectly using the NLS_LANG to prevent proper conversion and your data appears to be okay when you use utilties like SQL*PLUS to view your data. When you write applications that don't rely on the NLS_LANG like JDBC thin driver for instance you will realize your data is in fact invalid. To learn more about the NLS_LANG you can take a look at this FAQ: http://www.oracle.com/technology/tech/globalization/htdocs/nls_lang%20faq.htm
    To migrate your database to a proper character set you can refer to this paper:
    http://www.oracle.com/technology/tech/globalization/pdf/mwp.pdf
    But please do not ask for help in supporting your current configuration in this forum.

  • Export and Import in Oracle

    I need to export users with tables and data from Oracle 8.0.4 and import into Oracle 8.1.7.
    But the condition is that we need to retain the Japanese characters.
    I did this using simple export and import functions but Japanese characters were imported as Junk.
    Can somebody tell me what is the clean way of exporting and importing such that Japanese charaters are not converted to Junk.

    You must set the correct characterset before exporting and importing (I don't know which characterset is correct for you, but choosing the same characterset as the characterset of your database should work).
    Use NLS_LANG environment variable to do this,
    NLS_LANG=<LANGUAGE>_<TERRITORY>.<CHARACTERSET>. On UNIX if NLS_LANG is not set explicitly, it defaults to AMERICAN_AMERICA.US7ASCII, that means all special characters cannot be translated.
    On Windows NLS_LANG is set at installation time, it corresponds to the settings of Windows itself.

  • Export and import problems

    i have 2 major problems importing my application from htmldb.oracle.com to my serevr at home.
    1.
    i export my application and thems from htmldb.oracle.com.
    when i opened the sql export file i saw that all hebrew chracters was "jibrish".
    how can i export my application with hebrew fonts ?
    2. i tried to export and import xml data (my tables data).
    the export xml file showing me the hewbrew data but when i import it to my tables it gets ????????? instead of hebrew.
    i think it's the nls lang but i don't know what nls land i need and where to change it
    (database ,enviroment....)
    can someone help me with this problems ?
    thanks
    p.s
    my db is 10.1.0.3
    red hat linux as 4 on vmware 4.5
    my nls parameters:
    nls_language: american
    nls_territory: america
    nls_characterset: we8iso8859p1
    nls_nchar_charcterset: al16utf16
    thanks

    Zvika,
    I have two options that may work for you. First, define the fields that have Hebrew characters using the CHAR qualifier. Example, varchar2(30 char). That allows you to store double byte characters from other languages, and overrides the nls_length_semantics (byte default) parameter.
    Or, you could set the nls_length_semantics to default to char. It definitely uses more space when the char qualifier is used, so it's probably best to override the byte default only when necessary.
    Hope that helps!
    Also, I posted the 'Blogging Tool for HTMLDB' question with the username 'aufan89', which I had to change because my email changed.
    So my new username is now 'aufan1989'. I look forward to your email when you get your app fixed and posted onto HTMLDB Studio!
    Thanks!

  • Exporting and Importing of Schema

    The Export and Import of Schema is not working consistently in SP4.
    When I create a new repository from the schema, all the subtables get created. The main table is has only one field in it. It reads Category.
    The other fields in the main table are missing and are not imported during the Import Schema operation.
    Has anyone encountered the same issue?
    If so, please let me know how you overcame this errorneous behaviour?

    Hi Adhappan
    I have tried importing without success. The import is not consistent. The same schema creates different sets of tables & fields each time. Sometimes if you reimport some more tables which were not set up the first time gets selected for import.
    I did a simple exercise to test the import / export functionality.
    1. I unarchived a repository from standard content provided by SAP
    2. I exported the schema
    3. I tried importing the schema without any modifications to schema
    4. The results were highly inconsistent.
    Arvind
    Pls reward useful answers

  • Exporting and importing just table definitions

    Hi,
    I have this production database that has a huge amount of data in it. I was asked to set up a test database based on the exact same schema as the live database. When I tried to do an export (from live) and import (to test), with the parameters rows=N and compress=y, the target (test database) data file will still grow enormously, presumably because of the huge number of extents already allocated to the table in the live database. My test database of course, has a limited hard-disk space.
    Is there a way to export and import the table definitions without having the target database experiencing a huge growth in the size of the tablespace?
    Thanks,
    Chris.

    If an export with compress=n is still creating initial extents that a too large, you can still build with the import file but it will take a little work.
    run imp with indexfile=somefile.sql
    when imp is finished, edit somefile.sql by:
    1. remove all the REM statements.
    2. remove all the storage clauses (tables and indexes)
    Make sure your tablespaces have a small (say 1k) default initial extent.
    run imp again with rows=n
    All your tables and indexes will be created with the default tablespace initial extent.

  • How to export and import dependent tables from 2 different schema

    I have a setup where schema1 has table 1 and schema 2 has table2. And table 1 from schema1 depends upon table 2 from schema 2.
    I would like to export and import these tables only and not any other tables from these 2 schemas with all information like grants,constraints.
    Also will there be same method for Oracle 10g R1,R2 and Oracle 11G.
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10825/dp_export.htm#i1007514
    Looking at this For table mode it says
    Also, as in schema exports, cross-schema references are not exported
    Not sure what this means.
    As I am interested in only 2 tables I think I need to use table mode. But if I try to run export with both tables names, it says table mode support only one schema at a time. Not sure then How would the constraints would get exported in that case.
    -Rohit

    worked for my 1st time I tried
    exp file=table2.dmp tables="dbadmin.temp1,scott.emp"
    Export: Release 10.2.0.1.0 - Production on Mon Mar 1 16:32:07 2010
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Username: / as sysdba
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P1 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    Current user changed to DBADMIN
    . . exporting table                          TEMP1         10 rows exported
    EXP-00091: Exporting questionable statistics.
    Current user changed to SCOTT
    . . exporting table                            EMP         14 rows exported
    EXP-00091: Exporting questionable statistics.
    EXP-00091: Exporting questionable statistics.
    Export terminated successfully with warnings.

  • Shell Script For Export And Import Of Table Records

    Hello,
    We have production and test instances and for constant testing we need to copy data from production to test or development environment.
    At the moment what we do is manually doing export and import table records. At times this could be very tedious as we may need
    to do this exercise a couple of times in a day.
    Is it a good idea to do this exercise using shell script? If so how could I do this? If this is not a good idea what are the best alternatives?
    Any input is highly appreciated.
    Thanks

    Ah I see, your company prefers stupidity over efficiency. It would be possible to do it in a controlled environment, wouldn't it? Also the test database would be allowed to select only.
    So the non-allowance is just plain stupid.
    To the second question: do you use hard-coded passwords in shell scripts?
    Don't you think that poses a security risk?
    Don't you think that is a bigger risk than a database link, properly set up?
    In my book it is!
    Sybrand Bakker
    Senior Oracle DBA

  • Export and Import of Configuration Data

    Hi All,
    I have created a number of custom properties, structures, groups, renderers, layout sets etc and would like to move them from Dev to Test system.
    I have found the following in the help doco:
    "Export and Import of Configuration Data"
    http://help.sap.com/saphelp_nw04/helpdata/en/e1/029c414c79b25fe10000000a1550b0/content.htm
    However I am unable to see its functionality in our portal systems (namely the Actions->Export and Actions->Import functions).
    Our version of Portal is: 6.0.9.6.0
    KM: 6.0.9.3.0 (NW04 SPS09 Patch3)
    Therefore we should be able to see it? no?
    Are there any special settings to see these buttons?
    Cheers,
    Vic

    Thanks for your replies.
    OK, so in otherwords, it is not available in SP09 and only in SP12+.
    So, to "transport" this information I will have to manually reapply thes settings in the Test system!
    Message was edited by: Victor Yeoh

  • Export and Import of mappings/process flows etc

    Hi,
    We have a single repository with multiple projects for DEV/UAT and PROD of the same logical project. This is a nightmare for controlling releases to PROD and in fact we have a corrupt repository as a result I suspect. I plan to split the repository into 3 separate databases so that we have a design repos for DEV/UAT and PROD. Controlling code migrations between these I plan to use the metadata export and subsequent import into UAT and then PROD once tested. I have used this successfully before on a project but am worried about inherent bugs with metadata export/imports (been bitten before with Oracle Portal). So can anyone advise what pitfalls there may be with this approach, and in particular if anyone has experienced loss of metadata between export and import. We have a complex warehouse with hundreds of mappings, process flows, sqlldr flatfile loads etc. I have experienced process flow imports that seem to lose their links to the mappings they encapsulate.
    Thanks for any comments,
    Brandon

    This should do the trick for you as it looks for "PARALLEL" therefore it only removes the APPEND PARALLEL Hint and leaves other Hints as is....
    #set current location
    set path "C:/TMP"
    # Project parameters
    set root "/MY_PROJECT"
    set one_Module "MY_MODULE"
    set object "MAPPINGS"
    set path "C:/TMP
    # OMBPLUS and tcl related parameters
    set action "remove_parallel"
    set datetime [clock format [clock seconds] -format %Y%m%d_%H%M%S]
    set timestamp [clock format [clock seconds] -format %Y%m%d-%H:%M:%S]
    set ext ".log"
    set sep "_"
    set ombplus "OMBplus"
    set omblogname $path/$one_Module$sep$object$sep$datetime$sep$ombplus$ext
    set OMBLOG $omblogname
    set logname $path/$one_Module$sep$object$sep$datetime$ext
    set log_file [open $logname w]
    set word "PARALLEL"
    set i 0
    #Connect to OWB Repository
    OMBCONNECT .... your connect tring
    #Ignores errors that occur in any command that is part of a script and moves to the next command in the script.
    set OMBCONTINUE_ON_ERROR ON
    OMBCC "'$root/$one_Module'";      
    #Searching Mappings for Parallel in source View operators
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    foreach mapName [OMBLIST MAPPINGS] {
    foreach opName [OMBRETRIEVE MAPPING '$mapName' GET TABLE OPERATORS] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT)] {
    if { [ regexp $word $prop1] == 1 } {
    incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET DIMENSION OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET CUBE OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET VIEW OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    if { $i == 0 } {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
         } else {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
    close $log_file;
    Enjoy!
    Michel

  • Export and Import of Access DB tables in Apex

    Hi,
    i need help.
    I wanna export an MSACCESS DB Table into an ORACLE DB 10g2.
    but every time i get an error course my ODBC driver which isnt compatible to my tns service name.
    what could i do?
    i only want to export a table..not the whole database.
    if i want to export and import a Table over a XE Version, there are no errors and i could see the table in APEX.
    ...thx for help :)

    Hi Timo,
    If you open your table in Access, and click the "select all" (the empty grey box at the left of the header row), you can copy the entire table and paste it into Apex as though you were copying from Excel - the structure of the copied data is the same. The only thing is that you will be limited to 32,000 characters.
    Obviously, though, getting hold of, and setting up, the correct ODBC drivers is the more proper solution but the above may solve your immediate problem.
    Regards
    Andy

  • Export and Import tab in BusinessView Manager

    Post Author: palm
    CA Forum: Crystal Reports
    Hello All, I want to know the use of Export and Import tabs in BVM I tried to export a BV using the Export tab , it created an XML file and now , what can we do with that XML file Appreciate any replies on this
    Thanks!!

    HI
    Export is to take a copy of rule set from CC and export it to a flat file. Same way you can import the to any other environemnt of CC
    My second question regarding creating a new rule set for Training environemnt, that you have to create Functions with different names probably you can put "z" or "T"in front of the functions do denote it is for training system and than create it
    Parveen

  • Export and Import of Portal Users (UME database only)

    Hi
    What is the most efficient way to export all Portal user that are stored in Portal UME database so that it can be imported to another Portal installation and be used..
    Thanks in Advance
    Ananda

    Hi,
    SAP User Administration is the standard way to export and import users in the portal. Using the export functionality, all the user information will get exported, including the roles and groups assigned:
    [User]
    uid=userid
    last_name=Lastname
    first_name=Tobias
    language=en
    accessibility=0
    role=pcd:portal_content/com.sap.portal.xxx;
    group=Administrators;
    If you only want to export the user data without the roles and groups, you'll have to delete these lines.
    SAP Help: http://help.sap.com/saphelp_nw04/helpdata/en/70/9be23d44d48e5be10000000a114084/frameset.htm
    Export Format: http://help.sap.com/saphelp_nw04/helpdata/en/ae/7cdf3dffadd95ee10000000a114084/frameset.htm
    SDN Article: https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f02e3e37-5ee3-2910-129d-967aef3fa386
    br,
    Tobias

  • Export and import XMLType table

    Hi ,
    I want to export one table which contain xmltype column form oracle 11.2.0.1.0 and import into 11.2.0.2.0 version.
    I got following errors when i export the table , when i tried with exp and imp utility
    EXP-00107: Feature (BINARY XML) of column ZZZZ in table XXXX.YYYY is not supported. The table will not be exported.
    then i tried export and import pump.Exporting pump is working ,following is the log
    Export: Release 11.2.0.1.0 - Production on Wed Oct 17 17:53:41 2012
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ;;; Legacy Mode Active due to the following parameters:
    ;;; Legacy Mode Parameter: "log=<xxxxx>Oct17.log" Location: Command Line, Replaced with: "logfile=T<xxxxx>_Oct17.log"
    ;;; Legacy Mode has set reuse_dumpfiles=true parameter.
    Starting "<xxxxx>"."SYS_EXPORT_TABLE_01": <xxxxx>/******** DUMPFILE=<xxxxx>Oct172.dmp TABLES=<xxxxx>.<xxxxx> logfile=<xxxxx>Oct17.log reusedumpfiles=true
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 13.23 GB
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "<xxxxx>"."<xxxxx>" 13.38 GB 223955 rows
    Master table "<xxxxx>"."SYS_EXPORT_TABLE_01" successfully loaded/unloaded
    Dump file set for <xxxxx>.SYS_EXPORT_TABLE_01 is:
    E:\ORACLEDB\ADMIN\LOCALORA11G\DPDUMP\<xxxxx>OCT172.DMP
    Job "<xxxxx>"."SYS_EXPORT_TABLE_01" successfully completed at 20:30:14
    h4. I got error when i import the pump using following command
    +impdp sys_dba/***** dumpfile=XYZ_OCT17_2.DMP logfile=import_vmdb_XYZ_Oct17_2.log FROMUSER=XXXX TOUSER=YYYY CONTENT=DATA_ONLY TRANSFORM=oid:n TABLE_EXISTS_ACTION=append;+
    error is :
    h3. KUP-11007: conversion error loading table "CC_DBA"."XXXX"*
    h3. ORA-01403: no data found*
    h3. ORA-31693: Table data object "XXX_DBA"."XXXX" failed to load/unload and is being skipped due to error:*
    Please help me to get solution for this.

    CREATE UNIQUE INDEX "XXXXX"."XXXX_XBRL_XMLINDEX_IX" ON "CCCC"."XXXX_XBRL" (EXTRACTVALUE(SYS_MAKEXML(128,"SYS_NC00014$"),'/xbrl'))above index is created by us because we are storing file which like
    <?xml version="1.0" encoding="UTF-8"?>
    <xbrl xmlns="http://www.xbrl.org/2003/instance" xmlns:AAAAAA="http://www.AAAAAA.COM/AAAAAA" xmlns:ddd4217="http://www.xbrl.org/2003/ddd4217" xmlns:link="http://www.xbrl.org/2003/linkbase" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
    <link:schemaRef xlink:href="http://www.fsm.AAAAAA.COM/AAAAAA-xbrl/v1/AAAAAA-taxonomy-2009-v2.11.xsd" xlink:type="simple" />
    <context id="Company_Current_ForPeriod"> ...I tried to export pump with and without using DATA_OPTIONS=XML_CLOBS too.Both time exporting was success but import get same KUP-11007: conversion error loading table "tab_owner"."bbbbb_XBRL" error.
    I tried the import in different ways
    1. Create table first then import data_only (CONTENT=DATA_ONLY)
    2. Import all table
    In both way table ddl is created successfully ,but it fail on import data.Following is the log when i importing
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "aaaaa"."SYS_IMPORT_TABLE_02" successfully loaded/unloaded
    Starting "aaaaa"."SYS_IMPORT_TABLE_02":  aaaaa/********@vm_dba TABLES=tab_owner.bbbbb_XBRL dumpfile=bbbbb_XBRL_OCT17_2.DMP logfile=import_vmdb
    bbbbb_XBRL_Oct29.log DATA_OPTIONS=SKIP_CONSTRAINT_ERRORS
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    *KUP-11007: conversion error loading table "tab_owner"."bbbbb_XBRL"*
    *ORA-01403: no data found*
    ORA-31693: Table data object "tab_owner"."bbbbb_XBRL" failed to load/unload and is being skipped due to error:
    ORA-29913: error in executing ODCIEXTTABLEFETCH callout
    ORA-26062: Can not continue from previous errors.
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "aaaaa"."SYS_IMPORT_TABLE_02" completed with 1 error(s) at 18:42:26

Maybe you are looking for

  • FINAL CUT PRO KEEPS QUITTING WHEN TRYING TO OPEN PROJECTS-PLEASE HELP!

    I am finishing up two projects and ready to export them to quicktime and final cut pro will not let me open the projects as it quit on me and these projects will not re-open. Can someone please advise?1?! Thanks

  • Error while using XML

    One of my users while trying to send a a batch of information to the main database by XML had the following java execption code ORA-01401:inserted value too large for column. But I can't find which column is the one affected. Thanks

  • Help on update

    Hi All , I am trying to INSERT updated_by and Updated_date with ctut_id and sysdate When user trying to update any record. IF USER scott is logged into system and trying to update table Then I need to get his id number(CTUT_ID) into updated_by And sy

  • Tab scroll arrows are missing when they should be displayed.

    I just got upgraded to FF32 because I'd failed to remember to disable updates following my latest clean install. Now I don't have tab scroll arrows when there are enough tabs to need them. This happens in a default profile as well as with my normal p

  • How to find out URL for ArchiveLink objects?

    Hi All, SCENARIO: I'm working on "Archive Link", where a document (like word, JPEG photo, etc) is stored on a "Content Server" (specially meant for Database Management System) through a txn code(OAWD). This is successfully storing the document on the