Exporting and importing set-ups

Is anyone aware of a neat way to export all set-up information (e.g. ID, roles, even passwords, flex-fields, etc.) from one Oracle instance to another, so that if we want to perform a fresh Oracle Apps install, we do not have to manually re-enter the set-up information but will end up with the new instance having exactly the same set-up as our existing environment?
- we have GL, AP, PO, Grants/11.5.3 on RDBMS 8.1.6, Sun Solaris
Thanks!

I have the same question but under MS Windows.
Thanks in anticipation.

Similar Messages

  • Language Problem while exporting and importing data

    hi,
    I have Oracle version 8.1.7.0.0 installed on one server and 9.2.0.1.0 installed on new server.
    I'm copying and pasting my version info from SQL*Plus:
    SQL*Plus: Release 8.1.7.0.0 - Production on Mon Aug 22 10:46:31 2005
    (c) Copyright 2000 Oracle Corporation. All rights reserved.
    Connected to:
    Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    SQL>
    SQL*Plus: Release 9.2.0.1.0 - Production on Mon Aug 22 12:30:06 2005
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Connected to:
    Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    SQL>
    I created new user on my new server from enterprise manager.
    Exported user from the old server and imported in the new server.
    i.e: from Oracle8i Enterprise Edition Release 8.1.7.0.0, I did
    c:\>exp system/manager file=abc.dmp owner=abc
    Then on the new server Release 9.2.0.1.0, I did
    c:\>imp system/manager file=abc.dmp fromuser=abc touser=abc
    I'm using Arabic Language on my both servers. NLS_LANG parameter on both the servers is AMERICAN_AMERICA.WE8MSWIN1252.
    On both the servers I'm able to insert and select data in arabic.
    However, after I export the data from old server to the new server, the arabic data comes in question marks.
    If I create new table and insert arabic data on new server's user abc it is displaying well. Only the data which I exported and imported is not showing arabic.
    On both old and new servers operating system is Windows XP.
    I'm stuck with this problem. Anybody having any idea about how to solve this problem please help.
    Thank you all in advance.
    Regards

    Let me be clear here. Storing Arabic data in a WE8MSWIN1252 database is not supported by Oracle and will lead to problems. You are incorrectly using the NLS_LANG to prevent proper conversion and your data appears to be okay when you use utilties like SQL*PLUS to view your data. When you write applications that don't rely on the NLS_LANG like JDBC thin driver for instance you will realize your data is in fact invalid. To learn more about the NLS_LANG you can take a look at this FAQ: http://www.oracle.com/technology/tech/globalization/htdocs/nls_lang%20faq.htm
    To migrate your database to a proper character set you can refer to this paper:
    http://www.oracle.com/technology/tech/globalization/pdf/mwp.pdf
    But please do not ask for help in supporting your current configuration in this forum.

  • Exporting and Importing of Schema

    The Export and Import of Schema is not working consistently in SP4.
    When I create a new repository from the schema, all the subtables get created. The main table is has only one field in it. It reads Category.
    The other fields in the main table are missing and are not imported during the Import Schema operation.
    Has anyone encountered the same issue?
    If so, please let me know how you overcame this errorneous behaviour?

    Hi Adhappan
    I have tried importing without success. The import is not consistent. The same schema creates different sets of tables & fields each time. Sometimes if you reimport some more tables which were not set up the first time gets selected for import.
    I did a simple exercise to test the import / export functionality.
    1. I unarchived a repository from standard content provided by SAP
    2. I exported the schema
    3. I tried importing the schema without any modifications to schema
    4. The results were highly inconsistent.
    Arvind
    Pls reward useful answers

  • Exporting and importing just table definitions

    Hi,
    I have this production database that has a huge amount of data in it. I was asked to set up a test database based on the exact same schema as the live database. When I tried to do an export (from live) and import (to test), with the parameters rows=N and compress=y, the target (test database) data file will still grow enormously, presumably because of the huge number of extents already allocated to the table in the live database. My test database of course, has a limited hard-disk space.
    Is there a way to export and import the table definitions without having the target database experiencing a huge growth in the size of the tablespace?
    Thanks,
    Chris.

    If an export with compress=n is still creating initial extents that a too large, you can still build with the import file but it will take a little work.
    run imp with indexfile=somefile.sql
    when imp is finished, edit somefile.sql by:
    1. remove all the REM statements.
    2. remove all the storage clauses (tables and indexes)
    Make sure your tablespaces have a small (say 1k) default initial extent.
    run imp again with rows=n
    All your tables and indexes will be created with the default tablespace initial extent.

  • How to export and import dependent tables from 2 different schema

    I have a setup where schema1 has table 1 and schema 2 has table2. And table 1 from schema1 depends upon table 2 from schema 2.
    I would like to export and import these tables only and not any other tables from these 2 schemas with all information like grants,constraints.
    Also will there be same method for Oracle 10g R1,R2 and Oracle 11G.
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10825/dp_export.htm#i1007514
    Looking at this For table mode it says
    Also, as in schema exports, cross-schema references are not exported
    Not sure what this means.
    As I am interested in only 2 tables I think I need to use table mode. But if I try to run export with both tables names, it says table mode support only one schema at a time. Not sure then How would the constraints would get exported in that case.
    -Rohit

    worked for my 1st time I tried
    exp file=table2.dmp tables="dbadmin.temp1,scott.emp"
    Export: Release 10.2.0.1.0 - Production on Mon Mar 1 16:32:07 2010
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Username: / as sysdba
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P1 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    Current user changed to DBADMIN
    . . exporting table                          TEMP1         10 rows exported
    EXP-00091: Exporting questionable statistics.
    Current user changed to SCOTT
    . . exporting table                            EMP         14 rows exported
    EXP-00091: Exporting questionable statistics.
    EXP-00091: Exporting questionable statistics.
    Export terminated successfully with warnings.

  • Shell Script For Export And Import Of Table Records

    Hello,
    We have production and test instances and for constant testing we need to copy data from production to test or development environment.
    At the moment what we do is manually doing export and import table records. At times this could be very tedious as we may need
    to do this exercise a couple of times in a day.
    Is it a good idea to do this exercise using shell script? If so how could I do this? If this is not a good idea what are the best alternatives?
    Any input is highly appreciated.
    Thanks

    Ah I see, your company prefers stupidity over efficiency. It would be possible to do it in a controlled environment, wouldn't it? Also the test database would be allowed to select only.
    So the non-allowance is just plain stupid.
    To the second question: do you use hard-coded passwords in shell scripts?
    Don't you think that poses a security risk?
    Don't you think that is a bigger risk than a database link, properly set up?
    In my book it is!
    Sybrand Bakker
    Senior Oracle DBA

  • Export and Import of Configuration Data

    Hi All,
    I have created a number of custom properties, structures, groups, renderers, layout sets etc and would like to move them from Dev to Test system.
    I have found the following in the help doco:
    "Export and Import of Configuration Data"
    http://help.sap.com/saphelp_nw04/helpdata/en/e1/029c414c79b25fe10000000a1550b0/content.htm
    However I am unable to see its functionality in our portal systems (namely the Actions->Export and Actions->Import functions).
    Our version of Portal is: 6.0.9.6.0
    KM: 6.0.9.3.0 (NW04 SPS09 Patch3)
    Therefore we should be able to see it? no?
    Are there any special settings to see these buttons?
    Cheers,
    Vic

    Thanks for your replies.
    OK, so in otherwords, it is not available in SP09 and only in SP12+.
    So, to "transport" this information I will have to manually reapply thes settings in the Test system!
    Message was edited by: Victor Yeoh

  • Export and Import of mappings/process flows etc

    Hi,
    We have a single repository with multiple projects for DEV/UAT and PROD of the same logical project. This is a nightmare for controlling releases to PROD and in fact we have a corrupt repository as a result I suspect. I plan to split the repository into 3 separate databases so that we have a design repos for DEV/UAT and PROD. Controlling code migrations between these I plan to use the metadata export and subsequent import into UAT and then PROD once tested. I have used this successfully before on a project but am worried about inherent bugs with metadata export/imports (been bitten before with Oracle Portal). So can anyone advise what pitfalls there may be with this approach, and in particular if anyone has experienced loss of metadata between export and import. We have a complex warehouse with hundreds of mappings, process flows, sqlldr flatfile loads etc. I have experienced process flow imports that seem to lose their links to the mappings they encapsulate.
    Thanks for any comments,
    Brandon

    This should do the trick for you as it looks for "PARALLEL" therefore it only removes the APPEND PARALLEL Hint and leaves other Hints as is....
    #set current location
    set path "C:/TMP"
    # Project parameters
    set root "/MY_PROJECT"
    set one_Module "MY_MODULE"
    set object "MAPPINGS"
    set path "C:/TMP
    # OMBPLUS and tcl related parameters
    set action "remove_parallel"
    set datetime [clock format [clock seconds] -format %Y%m%d_%H%M%S]
    set timestamp [clock format [clock seconds] -format %Y%m%d-%H:%M:%S]
    set ext ".log"
    set sep "_"
    set ombplus "OMBplus"
    set omblogname $path/$one_Module$sep$object$sep$datetime$sep$ombplus$ext
    set OMBLOG $omblogname
    set logname $path/$one_Module$sep$object$sep$datetime$ext
    set log_file [open $logname w]
    set word "PARALLEL"
    set i 0
    #Connect to OWB Repository
    OMBCONNECT .... your connect tring
    #Ignores errors that occur in any command that is part of a script and moves to the next command in the script.
    set OMBCONTINUE_ON_ERROR ON
    OMBCC "'$root/$one_Module'";      
    #Searching Mappings for Parallel in source View operators
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    foreach mapName [OMBLIST MAPPINGS] {
    foreach opName [OMBRETRIEVE MAPPING '$mapName' GET TABLE OPERATORS] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT)] {
    if { [ regexp $word $prop1] == 1 } {
    incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET DIMENSION OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET CUBE OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET VIEW OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    if { $i == 0 } {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
         } else {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
    close $log_file;
    Enjoy!
    Michel

  • Export and Import of Access DB tables in Apex

    Hi,
    i need help.
    I wanna export an MSACCESS DB Table into an ORACLE DB 10g2.
    but every time i get an error course my ODBC driver which isnt compatible to my tns service name.
    what could i do?
    i only want to export a table..not the whole database.
    if i want to export and import a Table over a XE Version, there are no errors and i could see the table in APEX.
    ...thx for help :)

    Hi Timo,
    If you open your table in Access, and click the "select all" (the empty grey box at the left of the header row), you can copy the entire table and paste it into Apex as though you were copying from Excel - the structure of the copied data is the same. The only thing is that you will be limited to 32,000 characters.
    Obviously, though, getting hold of, and setting up, the correct ODBC drivers is the more proper solution but the above may solve your immediate problem.
    Regards
    Andy

  • Export and Import in Oracle

    I need to export users with tables and data from Oracle 8.0.4 and import into Oracle 8.1.7.
    But the condition is that we need to retain the Japanese characters.
    I did this using simple export and import functions but Japanese characters were imported as Junk.
    Can somebody tell me what is the clean way of exporting and importing such that Japanese charaters are not converted to Junk.

    You must set the correct characterset before exporting and importing (I don't know which characterset is correct for you, but choosing the same characterset as the characterset of your database should work).
    Use NLS_LANG environment variable to do this,
    NLS_LANG=<LANGUAGE>_<TERRITORY>.<CHARACTERSET>. On UNIX if NLS_LANG is not set explicitly, it defaults to AMERICAN_AMERICA.US7ASCII, that means all special characters cannot be translated.
    On Windows NLS_LANG is set at installation time, it corresponds to the settings of Windows itself.

  • Export and Import tab in BusinessView Manager

    Post Author: palm
    CA Forum: Crystal Reports
    Hello All, I want to know the use of Export and Import tabs in BVM I tried to export a BV using the Export tab , it created an XML file and now , what can we do with that XML file Appreciate any replies on this
    Thanks!!

    HI
    Export is to take a copy of rule set from CC and export it to a flat file. Same way you can import the to any other environemnt of CC
    My second question regarding creating a new rule set for Training environemnt, that you have to create Functions with different names probably you can put "z" or "T"in front of the functions do denote it is for training system and than create it
    Parveen

  • Export and import XMLType table

    Hi ,
    I want to export one table which contain xmltype column form oracle 11.2.0.1.0 and import into 11.2.0.2.0 version.
    I got following errors when i export the table , when i tried with exp and imp utility
    EXP-00107: Feature (BINARY XML) of column ZZZZ in table XXXX.YYYY is not supported. The table will not be exported.
    then i tried export and import pump.Exporting pump is working ,following is the log
    Export: Release 11.2.0.1.0 - Production on Wed Oct 17 17:53:41 2012
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ;;; Legacy Mode Active due to the following parameters:
    ;;; Legacy Mode Parameter: "log=<xxxxx>Oct17.log" Location: Command Line, Replaced with: "logfile=T<xxxxx>_Oct17.log"
    ;;; Legacy Mode has set reuse_dumpfiles=true parameter.
    Starting "<xxxxx>"."SYS_EXPORT_TABLE_01": <xxxxx>/******** DUMPFILE=<xxxxx>Oct172.dmp TABLES=<xxxxx>.<xxxxx> logfile=<xxxxx>Oct17.log reusedumpfiles=true
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 13.23 GB
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "<xxxxx>"."<xxxxx>" 13.38 GB 223955 rows
    Master table "<xxxxx>"."SYS_EXPORT_TABLE_01" successfully loaded/unloaded
    Dump file set for <xxxxx>.SYS_EXPORT_TABLE_01 is:
    E:\ORACLEDB\ADMIN\LOCALORA11G\DPDUMP\<xxxxx>OCT172.DMP
    Job "<xxxxx>"."SYS_EXPORT_TABLE_01" successfully completed at 20:30:14
    h4. I got error when i import the pump using following command
    +impdp sys_dba/***** dumpfile=XYZ_OCT17_2.DMP logfile=import_vmdb_XYZ_Oct17_2.log FROMUSER=XXXX TOUSER=YYYY CONTENT=DATA_ONLY TRANSFORM=oid:n TABLE_EXISTS_ACTION=append;+
    error is :
    h3. KUP-11007: conversion error loading table "CC_DBA"."XXXX"*
    h3. ORA-01403: no data found*
    h3. ORA-31693: Table data object "XXX_DBA"."XXXX" failed to load/unload and is being skipped due to error:*
    Please help me to get solution for this.

    CREATE UNIQUE INDEX "XXXXX"."XXXX_XBRL_XMLINDEX_IX" ON "CCCC"."XXXX_XBRL" (EXTRACTVALUE(SYS_MAKEXML(128,"SYS_NC00014$"),'/xbrl'))above index is created by us because we are storing file which like
    <?xml version="1.0" encoding="UTF-8"?>
    <xbrl xmlns="http://www.xbrl.org/2003/instance" xmlns:AAAAAA="http://www.AAAAAA.COM/AAAAAA" xmlns:ddd4217="http://www.xbrl.org/2003/ddd4217" xmlns:link="http://www.xbrl.org/2003/linkbase" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
    <link:schemaRef xlink:href="http://www.fsm.AAAAAA.COM/AAAAAA-xbrl/v1/AAAAAA-taxonomy-2009-v2.11.xsd" xlink:type="simple" />
    <context id="Company_Current_ForPeriod"> ...I tried to export pump with and without using DATA_OPTIONS=XML_CLOBS too.Both time exporting was success but import get same KUP-11007: conversion error loading table "tab_owner"."bbbbb_XBRL" error.
    I tried the import in different ways
    1. Create table first then import data_only (CONTENT=DATA_ONLY)
    2. Import all table
    In both way table ddl is created successfully ,but it fail on import data.Following is the log when i importing
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "aaaaa"."SYS_IMPORT_TABLE_02" successfully loaded/unloaded
    Starting "aaaaa"."SYS_IMPORT_TABLE_02":  aaaaa/********@vm_dba TABLES=tab_owner.bbbbb_XBRL dumpfile=bbbbb_XBRL_OCT17_2.DMP logfile=import_vmdb
    bbbbb_XBRL_Oct29.log DATA_OPTIONS=SKIP_CONSTRAINT_ERRORS
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    *KUP-11007: conversion error loading table "tab_owner"."bbbbb_XBRL"*
    *ORA-01403: no data found*
    ORA-31693: Table data object "tab_owner"."bbbbb_XBRL" failed to load/unload and is being skipped due to error:
    ORA-29913: error in executing ODCIEXTTABLEFETCH callout
    ORA-26062: Can not continue from previous errors.
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "aaaaa"."SYS_IMPORT_TABLE_02" completed with 1 error(s) at 18:42:26

  • Export and import problem

    hi i am having a test database 9i(9.2.0.1) on windows xp.i want to export a table from test database to development database which is 9i(9.2.0.6) on IBM AIX 5.2. till now i am able to export and import but suddenly i am getting 'TABLESPACE <tablespace name? doesn't exist' error.
    the following is what i am getting
    D:\>imp file=news.dmp fromuser='COUNTER' touser='IBT' tables='NEWS_MASTER'
    Import: Release 9.2.0.1.0 - Production on Sat Oct 6 17:06:00 2007
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Username: system@dev
    Password:
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.6.0 - 64bit Production
    With the Partitioning, Real Application Clusters, OLAP and Oracle Data Mining o
    tions
    JServer Release 9.2.0.6.0 - Production
    Export file created by EXPORT:V09.02.00 via conventional path
    Warning: the objects were exported by COUNTER, not by you
    import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    import server uses US7ASCII character set (possible charset conversion)
    . importing COUNTER's objects into IBT
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE TABLE "NEWS_MASTER" ("NEWS_DATE" DATE NOT NULL ENABLE, "SUBJECT" VAR"
    "CHAR2(75) NOT NULL ENABLE, "DESCRIPTION" VARCHAR2(300) NOT NULL ENABLE, "PH"
    "OTO" BLOB, "REG_NO" NUMBER(5, 0), "SYSTEM_IP" VARCHAR2(15) NOT NULL ENABLE)"
    " PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 STORAGE(INITIAL 65536 FREEL"
    "ISTS 1 FREELIST GROUPS 1) TABLESPACE "TESTTBS" LOGGING NOCOMPRESS LOB ("PHO"
    "TO") STORE AS (TABLESPACE "TESTTBS" ENABLE STORAGE IN ROW CHUNK 8192 PCTVE"
    "RSION 10 NOCACHE STORAGE(INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1))"
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'TESTTBS' does not exist
    Import terminated successfully with warnings.
    help me i want to know what happend actually?and what may be the solution?
    thanks in advance

    tablespace TESTTBS doesn't exist in the database where your are importing.
    Either create tablespace in target database or
    use show=y option while importing to get table structure and create manually in target database then import and give ignore=y.

  • Please provide me unix shell script (export and import of database schema)

    please i am new in unix
    \please give me sample unix shell script (export and import of database schema)

    please i am new in unix
    \please give me sample unix shell script (export and import of database schema)Instead of providing you the readymade unix shell script for your requirement, I will give you the hints to prepare the same at your own.
    Create a file with .sh extension.
    # Specify and set all required Environment variables.
    ORACLE_HOME, PATH, NLS_LANG, etc.,
    # Use the export command with all clauses
    export scott/tiger@orcl file=scott.dmp log=scott_emp.log
    # Compress it.
    compress scott.dmp
    Refer any unix/linux documents for scripting basics.
    http://www.bijoos.com/ora7/oracle_unix.htm
    Regards,
    Sabdar Syed.

  • Export and import Oracle 10G database

    Hello all
    Can anyone tell me once for all how can i backup oracle 10G database and restore it in other computer (export and import)
    thanks alot
    Ron

    Do you can access to metalink?
    If not, I copy an paste this doc's....
    Doc ID: Note:10767.1
    Subject: How to perform FULL System Export/Imports
    Type: BULLETIN
    Status: PUBLISHED
    Content Type: TEXT/X-HTML
    Creation Date: 18-APR-1994
    Last Revision Date: 30-AUG-2002
    Purpose
    This article highlights some important points to consider when performing
    a Full System Export and Import.
    Scope and Application
    It is intended to assist users performing FULL System Exports and Imports.
    A Full System Export and Import is a useful way to replicate or clean up
    a database. Please note the following guidelines when performing a Full
    System Export and Import:
    1. When performing the import, ensure you are pointing at the correct
    instance. Always check values for the SID and two task driver before
    proceeding.
    *Note:   On Unix systems, just entering a sub-shell, such as using
    C-shell, can change the database that the import would work
    against.
    2. When possible, it is advisable to take a physical copy of the
    exported database and the database you intend to import into before
    starting the import. This ensures any mistakes are reversible.
    3. NEVER run a FULL SYSTEM IMPORT on a machine that has more than one
    database UNLESS you are 100% sure all tablespaces have been
    pre-created. A full import will create any undefined tablespaces
    using the same datafile names as the exported database. There are
    situations where this is very dangerous:
    -> If the datafiles belong to ANY other database, they will be
    CORRUPTED. This is especially true if the exported database is
    on the same machine, because its datafiles will be re-used by
    the database you are importing into.
    -> If the datafiles have names that clash with existing operating
    system files.
    4. When exporting, it is advisable to also produce a report showing
    information you may need before / after the import:
    -> List the tablespaces and datafiles
    -> List the rollback segments
    -> List a count, by user, of each object type such as tables,
    indexes, etc.
    This enables you to ensure that tablespaces have been pre-created
    and to check the success of the import after it is complete.
    5. If you are creating a completely new database from an export, remember
    to create an extra rollback segment in SYSTEM, and make this available
    in the init.ora file before proceeding with the import.
    6. There are some versions of export that do not retain sequence numbers
    correctly. It is advisable to select next_value for each sequence
    after the export so that these can be checked after import.
    Following these guidelines will help you to avoid major problems.
    Doc ID: Note:174226.1
    Subject: How To Copy an Oracle Database To Another Machine
    Type: HOWTO
    Status: PUBLISHED
    Content Type: TEXT/X-HTML
    Creation Date: 12-OCT-2001
    Last Revision Date: 01-OCT-2003
    goal: How to copy an Oracle database to another machine
    fact: Oracle Server - Enterprise Edition
    fix:
    Pre-requisites: The copy is between 2 machines, both have to be on the same OS
    and have to have exactly the same database version installed.
    1. Make sure the database you want to copy was closed with a SHUTDOWN IMMEDIATE,
    SHUTDOWN NORMAL or SHUTDOWN TRANSACTIONAL.
    2. Copy init.ora and control files to create instance and be able to go in
    mount mode. Check the init.ora for the locations where the controlfiles have
    to be, if those locations are not valid on the machine put the control files on
    different places and adjust the init.ora accordingly.
    3.a. Copy the datafiles (all of them).
    b. Copy the redo-logfiles (all of them).
    4.a. (Unix only) Set the environment variables:
    ORACLE_SID - set to the database name you wish to create
    ORACLE_HOME - set to full pathname of the Oracle system home directory
    PATH - needs to include $ORACLE_HOME/bin
    b. (NT/2000 only) Do 'set ORACLE_SID=<SID>'
    Use oradim to create the service for the instance. For more information
    on oradim please refer to (the part that refers to creating a new instance):
    Note:68720.1 Creating a new 7.3, 8.0, 8.1 Instance/Database
    5. Use servermanager (check the name to use for your version of oracle) or
    sqlplus (version 9i and above) to startup the database in mount mode.
    Do
    CONNECT INTERNAL/<PASSWORD>
    then
    STARTUP MOUNT
    Then do a rename of the copied datafiles if they are not in the same path as on
    the other machine. For all the files that are in the result of the query:
    SELECT NAME FROM V$DATAFILE;
    do
    ALTER DATABASE RENAME FILE '<oldfullpath>\<filename>' to <
    newfullpath>\<filename>';
    6. Query the datadictionary for the old location of the redolog files using:
    SELECT MEMBER FROM V$LOGFILE;
    If the new place is not the same as the old do:
    ALTER DATABASE RENAME FILE '<oldfullpath>\<redologfilename>' to &
    lt;newfullpath>\<redologfilename>';
    7. Now open the database:
    ALTER DATBASE OPEN;
    ******************************************************************

Maybe you are looking for

  • Laserjet 8000DN issue with Snow Leopard 10.6.2

    Hello, I am having a connectivity issue with a LaserJet 8000DN. The previous iMac was connected to the printer via a D-Link switch, and I assume it was using AppleTalk via Ethernet to read the printer. The OS installed on that machine was OS X Leopar

  • Podcasts are not listed in Podcast menu

    I resynched my iPod and all my podcast files are mixed in with the music titles. When I select the podcast menu item, none of the podcasts are displayed. However, when I select Genre on the main menu, then my podcast files can be found within several

  • Using and Sharing a CRT monitor w/a PC

    Hello all, I'm curious if I can use a CRT monitor as a second monitor with my Imac(Intel 24in. 2.4 GHz, for an extended desktop, etc. I'd also like to know if I could share the monitor with a PC with a KVM-type switch... If doing these is possible, w

  • Noise on optical output from Airport Express

    I have an almost new Airport Express (actually as an extension of my old AE) which I use for streaming music to my HiFi. At first I used the analogue output for my amplifier, without any problem. But now I have installed a HiFi D/A converter (Rega) a

  • Adobe CC installation fails (updates don't work either)

    Hi everyone, When I try to install Adobe CC apps, it says the installation failed (with every app). However, the apps seem to work properly, except for Audition (crashes when started). I tried updating the apps, but updating fails too. I'm running it