Export and Import of master repository - error snpuser not found

We had installed ODI on a windows server and created physicla and logical topologies etc and developed some models and projects.
But we had sone all of this while still pointing to the demo master repository using the HypersonicSQL. This means we always have to have the 'Start Demo Agent' running etc.
So we wanted to move to an Oracle based master repository. I have used the 'Create Master Repository' function to create a new repository using ORacle thinnet driver and this is all OK and I can sign in to the SUPERVISOR user.
But when I export our existing repository from the 'HypersonocSQL' demo one to a ZIP file and then use the 'Import Master Repository' to create a master repository in the Oracle using thinnet driver etc - firstly it removes the repository details that I originally created - fair enough - then creates it all again but from the ZIP file.
The SNP_ tables all exist in the oracle database user and I can see that in SNP_USER there is the record for SUPERVISOR etc.
But when I try to sign in to this through topology manager or security manager - as SUPERVISOR I immediately get the error 'SNPUser does not exist'
Am I doing something wrong with exporting my 'demo' master repository and trying to import it in to the new Oracle repository ?

thanks - but that is effectivley what I did - used the repository import from my exported ZIP file - and it all imported OK - but the error on login to topology manager afterwards is that the snp user does not exist.
Although I initially used the 'Master Repository Creation' I then adaondoned this - dropped and recreated my Oracle user 'snpm' that owns the master repository and used the import repository approach.
I am trying to sign in as the standard SUPERVISOR user (that existed OK in the Hypersonic master repository that was exported to the ZIP file) and also trying other users that I know existed in the Hypersonic reporitory that I exported - same problem. I can see in Oracle SQL*Plus that in the SNP_USERS table the SUPERVISOR user exists.
If I do instead carry out a standard 'Create Master Repository' and do NOT try to import anything then it works OK and I can sign in to the Topology Manager as SUPERVISOR.

Similar Messages

  • Export and import user master records

    Hi,
    We are planning for system refresh activity. So I am planning Production server data to copy in Quality server. I want retain Quality server users like that only.  Is there any way I can export user master records and import it later?
    Please advice us.
    Our environment is 4.6 C, Oracle 10g, windows 2003 server.
    Regards
    Satya

    Hello Satyanarayana,
    please be aware, that with SAP_USER standard change logs are not copied. So if you have such in your QAS, they will be propably overwritten by the PRD content. You will loose then change logs for role changes, which are kept in the standard change documents (cdpos,cdhdr).
    User change docs are copied together with the users with provfile SAP_USER.
    That is just for your information. In general there will not be change documents for roles in QAS.....
    b.rgds, Bernhard

  • Guide me for exporting and importing data from oracle 10g to oracle 9i

    I have been trying export and import options in oracle but m not very clear about the procedure..
    Also when i am logging in as Scott/tiger i am facing an error
    ORA 1542 - tablespace 'USER' is offline and cannot allocate space in it.how to resolve it

    Hello,
    I see two questions.
    ORA 1542 - tablespace 'USER' is offline and cannot allocate space in it.how to resolve it About the error ORA-01542 you have to bring the Tablespace USERS online:
    http://www.error-code.org.uk/view.asp?e=ORACLE-ORA-01542
    If you cannot bring it online you may have to Recover it first.
    I have been trying export and import options in oracle but m not very clear about the procedure..About the Export in 10g and Import in 9i (9.2 or 9.0.1 ?), you have to export with the Export utility of the Target Database (the lowest Database version here: 9.0.1 or 9.2) then, you have to Import to the Target Database using its Import utility (in 9.0.1 or 9.2).
    The following Note of MOS may guide you:
    Compatibility Matrix for Export And Import Between Different Oracle Versions [Video] [ID 132904.1]
    NB: You cannot use here the DATAPUMP.
    Hope this help.
    Best regards,
    Jean-Valentin

  • Export and Import in Oracle

    I need to export users with tables and data from Oracle 8.0.4 and import into Oracle 8.1.7.
    But the condition is that we need to retain the Japanese characters.
    I did this using simple export and import functions but Japanese characters were imported as Junk.
    Can somebody tell me what is the clean way of exporting and importing such that Japanese charaters are not converted to Junk.

    You must set the correct characterset before exporting and importing (I don't know which characterset is correct for you, but choosing the same characterset as the characterset of your database should work).
    Use NLS_LANG environment variable to do this,
    NLS_LANG=<LANGUAGE>_<TERRITORY>.<CHARACTERSET>. On UNIX if NLS_LANG is not set explicitly, it defaults to AMERICAN_AMERICA.US7ASCII, that means all special characters cannot be translated.
    On Windows NLS_LANG is set at installation time, it corresponds to the settings of Windows itself.

  • Error "Account not found" in Catalog

    Hi All!
    I have a problem with the catalog of OBI EE.
    When connecting to multiple users showed the error "Account not found".
    I've located the "Catalog Manager" accounts whose permits were <delete Account>. I deleted those accounts and have been fully rebuilt. These accounts are already functioning.
    But to connect to the "Presentation" with the manager and access the option "Manage users and groups" by clicking the link "Show users and groups," again gives the error "Account not found". The log shows:
    An error occurred during execution of "connect". No connection could be made Actively Because the target machine refused it.
    [Socket: 868]
    Type: Error
    Severity: 42
    Time: Mon Oct 26 13:31:01 2009
    File: project / WebSecurity / securitymanager.cpp Line: 84
    Properties: ThreadID-6808; HttpCommand-getAccounts; remoteIp-10.12.127.129; User-Administrator
    Location:
    saw.httpserver.request
    saw.rpc.server.responder
    saw.rpc.server
    saw.rpc.server.handleConnection
    saw.rpc.server.dispatch
    saw.threadPool
    saw.threads
    Account not found.
    Any ideas?
    Regards.

    When you log in into presentation services as administrator, you can still see these users in the user list? You have to delete them not only via the catalog manager, but also overthere.
    And restart your services once...
    Hope it helps.
    Kr,
    A

  • Yahoo loads log in click mail 404 error page not found why?

    I upload yahoo site, log in, click 'mail' and am given a '404' error; page not found... This does NOT happen on Chrome or IE, but the pages of yahoo on IE suck and Chrome won't let me have a single toolbar even though there are all sorts of claim you can have one... (sorry) WHY doesn't FIREFOX let yahoo mail load? I LOVE FIREFOX and ALL THINGS MOZILLA, but after I ran into crappy-sappy babylon... it put a toolbar on IE, CHROME and MOZILLA... I had to delete everything associated with them all, get toolbar sweepers and better uninstall programs to weed their junk out and now MOZILLA, FIREFOX, rather warns no one and makes me submit this rather than help

    Clear the cache and the cookies from sites that cause problems.
    * "Clear the Cache": Tools > Options > Advanced > Network > Offline Storage (Cache): "Clear Now"
    * "Remove the Cookies" from sites causing problems: Tools > Options > Privacy > Cookies: "Show Cookies"

  • Model Export and Import - Error 40144

    I'm trying to transport a model from one system to another and I get the following error when I try the Model export and import functionality in IBP 4.0
    "errorCode":"40144","errorMsg":"Repository: delivery unit not found; deliveryUnit: 'SOP_MODEL_EXPORT_DU', vendor:sap.com'"
    Steps:
    1. Model Export -> Create Export
    2. Once the Export is created, I clicked 'Download Deployment Unit' and that's when I get the above error.
    Am I doing anything wrong? If so, what is the procedure to execute the model export and import between systems?

    It turns out a transportation path has to be set up (one time) from the backend (by SAP) to enable this feature.

  • Exporting and Importing Destination Controls Error

    When I export table within destination control and import the table to another Ironport again within destination control I receive error - "Wrong format of the destination config file: ip_sort_pref is required for the global settings."
    We are updating the ASYNC on each of our de clustered Ironport.  So when one is not in use we need the other to handle all the traffic, including domain controls.
    The output displays
    [ABCDE.com]
    max_host_concurrency = 500
    limit_apply = system
    limit_type = host
    max_messages_per_connection = 50 ......
    How can I import this without losing or altering the data within.
    ^^^^^^^^^^^^^^^^^^^^^^
    Solution
    We upgrade one of our SMTP ASYNC to 8.0.1 which creates a additional line within default of the upgraded Ironport exported control destination table that is named... wait for it.... ip_sort_pref. Just add it.

    The applications and drivers have been imported into the Dev SCCM server however I cannot find the actual content so I am not sure this has worked properly. 
    When you exported did you choose to "Export all content for selected task sequence and dependencies".
    80070003 = The system cannot find the path specified.
    See here for exporting and importing task sequences
    http://blogs.technet.com/b/inside_osd/archive/2012/01/19/configuration-manager-2012-task-sequence-export-and-import.aspx
    Gerry Hampson | Blog:
    www.gerryhampsoncm.blogspot.ie | LinkedIn:
    Gerry Hampson | Twitter:
    @gerryhampson

  • Error when exporting and importing TS

    Hi I am running SCCM 2012 SP1 and am trying to export a Task Sequence including with the Windows 7 wim, the boot wim, some applications and drivers.  When I exported the TS I selected to export dependencies and content.
    I have setup dev SCCM environment (same spec as production) and am trying to import the .zip that was created when I completed the export.  When doing the import I am getting the below error on both the Windows 7 wim and the boot wim :
    Error Imported Operating System Image (1):
    Create new Windows 7 Enterprise x64
    Generic failure
    instance of SMS_ExtendedStatus
     Description = "Failed to get the image property from the source WIM file due
    to error 80070003".
     ErrorCode= 2147942403;
     File="e:\\nts_sccm_release\\sms\\siteserver\\sdk_provider\\smsprov\
    \sspimagepackage.cpp";
     Line = 483;
     Operation = "Putinstance";
     ParameterInfo = "";
     ProviderName = "ExtnProv";
     StatusCode = 2147749889;
    The applications and drivers have been imported into the Dev SCCM server however I cannot find the actual content so I am not sure this has worked properly.  I am wondering if I would be better off using the Migration wizard to take the TS and
    content across.  Would appreciate any advice on either a) how to fix the issues with the importing of the TS or b) whether I should just be using the migration wizard.
    Thanks
    G

    The applications and drivers have been imported into the Dev SCCM server however I cannot find the actual content so I am not sure this has worked properly. 
    When you exported did you choose to "Export all content for selected task sequence and dependencies".
    80070003 = The system cannot find the path specified.
    See here for exporting and importing task sequences
    http://blogs.technet.com/b/inside_osd/archive/2012/01/19/configuration-manager-2012-task-sequence-export-and-import.aspx
    Gerry Hampson | Blog:
    www.gerryhampsoncm.blogspot.ie | LinkedIn:
    Gerry Hampson | Twitter:
    @gerryhampson

  • Export and import change document for user master data

    Dear Gurus,
    I have two queries on change document for user master data:
    1. Are there any approaches to export and import change document for user master data?
    We often do system copy from PRD to QAS for UAT and troubleshooting. Before system copy we export the user master data from QAS and then import after the copy process. We would like to keep the change document for user master data on QAS from being refreshed from PRD for security reason.
    2. Change document for Role change in QAS
    When the role is created or modified in DEV and then transported to QAS, the role change document doen't include this change log. The role change document in QAS only records those role changes directly made in QAS.
    Could you advise this is by SAP design or are there any approaches to record this transported role change  in the role change document in QAS?
    Thanks
    YBY

    1. Perhaps you want to consider a system copy to a "virtual system" for UAT?
    2. Changes in QAS (as with PROD as well) will give you the delta. They should ideally be clean... You need to check the source system.
    Another option is to generate the profiles in the target system. But for that your config has to be sqeaky clean and in sync, including very well maintained and sync'ed Su24 data.
    Cheers,
    Julius

  • Portal Export and Import Steps

    Hi all,
    I need to export all page groups from Production portal instance and import those in brand new portal instance (target Portal)*. Source and target portal instance is as follows
    Production Oracle Database Version: 10.1.0.5
    Target Oracle Database Version: 10.2.0.4
    Production Oracle Infrastructure Version : 10.1.2.0.2
    Target Oracle Infrastructure Version: 10.1.2.0.2
    Production Portal Version: 10.1.2.0.2
    Target Portal Version: 10.1.2.0.2
    What i have done till now is as follows
    1. Install Database server 10.2.0.1
    2. Upgrade Database Server to 10.2.0.4
    3. Install 10.2.0.1 products from Companion CD (ultra search Schemas)
    4. Create Metadata Repository using Repository Creation assistance
    5. Install IAS infrastructure (10.1.2.0.2)
    6. Install IAS Middle Tier (10.1.2.02.)
    7. Create Transport sets in production instance
    Please tell me what i am supposed to do now.
    I am following below link for reference but not able to follow all
    http://www.scott-tiger.dk/portalHelp2/ohw/state?navSetId=_&navId=0
    Regards
    Anindya

    For Portal Export Import (transport), the following are the foremost links on Metalink/ My Oracle Support.
    Note 306785.1 Overview of the Portal Export-Import Process
    Note 263995.1 - Master Note for OracleAS Portal Export / Import Issues
    Note 333867.1 Portal Export and Import Utility Supportability Scenarios
    Regarding your question, please do go through all of these notes.
    - Overview will tell you how to prepare your instances/environments for portal transport. you will see how to ensure they are ready, and that some patches are needed.
    - Master note will tell you which patches are needed, how to get them, what are the diagnostic tools which can help you check and clean your portal repositories, and what common errors you may get and how to fix them.
    - Supportability Scenario tells you what actually you can transport and what you cannot. i.e., what is supported and what is not. Do read it for your case.
    I have seen several Portal Admins who keep banging their heads with problems they create in their systems because they do not read this note and try to force a case which does not follow a supportable scenario.
    hope that helps!
    AMN

  • SQL Developer 2.1: Problem exporting and importing unit tests

    Hi,
    I have created several unit tests on functions that are within packages. I wanted to export these from one unit test repository into another repository on a different database. The export and import work fine, but when running the tests on the imported version, there are lots of ORA-06550 errors. When debugging this, the function name is missing in the call, i.e. it is attempting <SCHEMA>.<PACKAGE> (parameters) instead of <SCHEMA>.<PACKAGE>.<FUNCTION> (parameters).
    Looking in the unit test repository itself, it appears that the OBJECT_CALL column in the UT_TEST table is null - if I populate this with the name of the function, then everything works fine. Therefore, this seems to be a bug with export and import, and it is not including this in the XML. The same problem happens whether I export a single unit test or a suite of tests. Can you please confirm whether this is a bug or whether I am doing something wrong?
    Thanks,
    Pierre.

    Hi Pierre,
    Thanks for pointing this out. Unfortunately, it is a bug on our side and you have found the (ugly) "work-around".
    Bug 9236694 - 2.1: OTN: UT_TEST.OBJECT_CALL COLUMN NOT EXPORTED/IMPORTED
    Brian Jeffries
    SQL Developer Team

  • Export and Import of mappings/process flows etc

    Hi,
    We have a single repository with multiple projects for DEV/UAT and PROD of the same logical project. This is a nightmare for controlling releases to PROD and in fact we have a corrupt repository as a result I suspect. I plan to split the repository into 3 separate databases so that we have a design repos for DEV/UAT and PROD. Controlling code migrations between these I plan to use the metadata export and subsequent import into UAT and then PROD once tested. I have used this successfully before on a project but am worried about inherent bugs with metadata export/imports (been bitten before with Oracle Portal). So can anyone advise what pitfalls there may be with this approach, and in particular if anyone has experienced loss of metadata between export and import. We have a complex warehouse with hundreds of mappings, process flows, sqlldr flatfile loads etc. I have experienced process flow imports that seem to lose their links to the mappings they encapsulate.
    Thanks for any comments,
    Brandon

    This should do the trick for you as it looks for "PARALLEL" therefore it only removes the APPEND PARALLEL Hint and leaves other Hints as is....
    #set current location
    set path "C:/TMP"
    # Project parameters
    set root "/MY_PROJECT"
    set one_Module "MY_MODULE"
    set object "MAPPINGS"
    set path "C:/TMP
    # OMBPLUS and tcl related parameters
    set action "remove_parallel"
    set datetime [clock format [clock seconds] -format %Y%m%d_%H%M%S]
    set timestamp [clock format [clock seconds] -format %Y%m%d-%H:%M:%S]
    set ext ".log"
    set sep "_"
    set ombplus "OMBplus"
    set omblogname $path/$one_Module$sep$object$sep$datetime$sep$ombplus$ext
    set OMBLOG $omblogname
    set logname $path/$one_Module$sep$object$sep$datetime$ext
    set log_file [open $logname w]
    set word "PARALLEL"
    set i 0
    #Connect to OWB Repository
    OMBCONNECT .... your connect tring
    #Ignores errors that occur in any command that is part of a script and moves to the next command in the script.
    set OMBCONTINUE_ON_ERROR ON
    OMBCC "'$root/$one_Module'";      
    #Searching Mappings for Parallel in source View operators
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    foreach mapName [OMBLIST MAPPINGS] {
    foreach opName [OMBRETRIEVE MAPPING '$mapName' GET TABLE OPERATORS] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT)] {
    if { [ regexp $word $prop1] == 1 } {
    incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET DIMENSION OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET CUBE OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET VIEW OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    if { $i == 0 } {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
         } else {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
    close $log_file;
    Enjoy!
    Michel

  • Export and import XMLType table

    Hi ,
    I want to export one table which contain xmltype column form oracle 11.2.0.1.0 and import into 11.2.0.2.0 version.
    I got following errors when i export the table , when i tried with exp and imp utility
    EXP-00107: Feature (BINARY XML) of column ZZZZ in table XXXX.YYYY is not supported. The table will not be exported.
    then i tried export and import pump.Exporting pump is working ,following is the log
    Export: Release 11.2.0.1.0 - Production on Wed Oct 17 17:53:41 2012
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ;;; Legacy Mode Active due to the following parameters:
    ;;; Legacy Mode Parameter: "log=<xxxxx>Oct17.log" Location: Command Line, Replaced with: "logfile=T<xxxxx>_Oct17.log"
    ;;; Legacy Mode has set reuse_dumpfiles=true parameter.
    Starting "<xxxxx>"."SYS_EXPORT_TABLE_01": <xxxxx>/******** DUMPFILE=<xxxxx>Oct172.dmp TABLES=<xxxxx>.<xxxxx> logfile=<xxxxx>Oct17.log reusedumpfiles=true
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 13.23 GB
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "<xxxxx>"."<xxxxx>" 13.38 GB 223955 rows
    Master table "<xxxxx>"."SYS_EXPORT_TABLE_01" successfully loaded/unloaded
    Dump file set for <xxxxx>.SYS_EXPORT_TABLE_01 is:
    E:\ORACLEDB\ADMIN\LOCALORA11G\DPDUMP\<xxxxx>OCT172.DMP
    Job "<xxxxx>"."SYS_EXPORT_TABLE_01" successfully completed at 20:30:14
    h4. I got error when i import the pump using following command
    +impdp sys_dba/***** dumpfile=XYZ_OCT17_2.DMP logfile=import_vmdb_XYZ_Oct17_2.log FROMUSER=XXXX TOUSER=YYYY CONTENT=DATA_ONLY TRANSFORM=oid:n TABLE_EXISTS_ACTION=append;+
    error is :
    h3. KUP-11007: conversion error loading table "CC_DBA"."XXXX"*
    h3. ORA-01403: no data found*
    h3. ORA-31693: Table data object "XXX_DBA"."XXXX" failed to load/unload and is being skipped due to error:*
    Please help me to get solution for this.

    CREATE UNIQUE INDEX "XXXXX"."XXXX_XBRL_XMLINDEX_IX" ON "CCCC"."XXXX_XBRL" (EXTRACTVALUE(SYS_MAKEXML(128,"SYS_NC00014$"),'/xbrl'))above index is created by us because we are storing file which like
    <?xml version="1.0" encoding="UTF-8"?>
    <xbrl xmlns="http://www.xbrl.org/2003/instance" xmlns:AAAAAA="http://www.AAAAAA.COM/AAAAAA" xmlns:ddd4217="http://www.xbrl.org/2003/ddd4217" xmlns:link="http://www.xbrl.org/2003/linkbase" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
    <link:schemaRef xlink:href="http://www.fsm.AAAAAA.COM/AAAAAA-xbrl/v1/AAAAAA-taxonomy-2009-v2.11.xsd" xlink:type="simple" />
    <context id="Company_Current_ForPeriod"> ...I tried to export pump with and without using DATA_OPTIONS=XML_CLOBS too.Both time exporting was success but import get same KUP-11007: conversion error loading table "tab_owner"."bbbbb_XBRL" error.
    I tried the import in different ways
    1. Create table first then import data_only (CONTENT=DATA_ONLY)
    2. Import all table
    In both way table ddl is created successfully ,but it fail on import data.Following is the log when i importing
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "aaaaa"."SYS_IMPORT_TABLE_02" successfully loaded/unloaded
    Starting "aaaaa"."SYS_IMPORT_TABLE_02":  aaaaa/********@vm_dba TABLES=tab_owner.bbbbb_XBRL dumpfile=bbbbb_XBRL_OCT17_2.DMP logfile=import_vmdb
    bbbbb_XBRL_Oct29.log DATA_OPTIONS=SKIP_CONSTRAINT_ERRORS
    Processing object type TABLE_EXPORT/TABLE/TABLE
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    *KUP-11007: conversion error loading table "tab_owner"."bbbbb_XBRL"*
    *ORA-01403: no data found*
    ORA-31693: Table data object "tab_owner"."bbbbb_XBRL" failed to load/unload and is being skipped due to error:
    ORA-29913: error in executing ODCIEXTTABLEFETCH callout
    ORA-26062: Can not continue from previous errors.
    Processing object type TABLE_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type TABLE_EXPORT/TABLE/INDEX/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "aaaaa"."SYS_IMPORT_TABLE_02" completed with 1 error(s) at 18:42:26

  • Client Export and Import (Speacially Import Procedures)

    Dear All,
    I have requirement to copy my production client's all data to Quality System's client.
    What procedure should I follow?
    If I am going to export and Import. I am not getting followings's Steps 2.3 to 2.6.
    1. Client Export:
         1.1 Run SCC8
         1.2. Select Profile for desired copy type (Usually All SAP_ALL or user master only SAP_USER. You will need direction from  
                 the requester as to the correct selection here. Use Profile -> Display Profile to display profile details.)
         1.3. Select target System (or group)
         1.4. De- Select u201CTest Runu201D (If selected)
         1.5. Run Export
                - Up to 3 requests are created, depending on the data selected and available:
                I. u201CSIDKO00353u2033 for transporting client-independent data, if you have selected this
                II. u201CSIDKT00353u2033 for transporting client-specific data
                III. u201CSIDKX00353u2033 for transporting client-specific texts, provided texts are available in this client
         1.6.Monitor TP logs for errors and export files for growth
    2. Client Import:
        2.1. Create client (scc4)
        2.2. Login to client (sap* u2013 pass)
        2.3. Manually add u201COu201D transport then u201CXu201D then u201CTu201D to TMS buffer
        2.4. Highlight #1 and use u201CRequest -> Importu201D to launch import tool
        2.5. Monitor u201CIu201D file in OS u201C/usr/sap/trans/tmpu201D dir for progress info
        2.6. After Import is complete perform u201Cpost processing stepsu201D from client tool (SCC7)           So if you have 3 transport requests and the data has been exported successfully then go for an import in quality system
    Please let me know how will I  Manually add u201COu201D transport then u201CXu201D then u201CTu201D to TMS buffer, Highlight #1 and use u201CRequest -> Importu201D to launch import tool, Monitor u201CIu201D file in OS u201C/usr/sap/trans/tmpu201D dir for progress info and After Import is complete perform u201Cpost processing stepsu201D from client tool (SCC7) .
    Thanks for sharing knowledge.
    With Warm Regards
    Dharmendra

    Once u login to new client,
    to Manually add u201COu201D transport then u201CXu201D then u201CTu201D to TMS buffer
    go to STMS----> select the system
    At top menu select
    Extras --> Other Requests --> Add(here give the above created transport requests, 1 @ each time)
    once u add those 3 trasnport requests u need to import them.
    Highlight #1 and use u201CRequest -> Importu201D
    for that from top menu select  Request -> Import(Ctrl + F11)
    coming to SCC7. its a tcode. execute the tcode and see.
    Regards,
    Ravi

Maybe you are looking for

  • Get values from selected row in a Table?

    Hello. I'm on VC 7.1 (the trial version downloaded from SDN). I'm trying to figure out a way to retrieve some values from the currently selected row in a Table element through the output connector. I have a web-service which returns results to the Ta

  • Two iTunes accounts on the same PC computer

    I am using iPhone and my wife is going to buy a iPod touch. Can we have separate iTunes account on the same PC computer? Meaning I have my own account & syn data and she has her own. Or she better use another computer.

  • Services for a New Object

    Helo community: We have created a new object ("8C - job family") , this object is a copy of object "C" job,  we have done the customizing, for example in table: T778O and we have filled it with: 8C     Familia de Funciones     PDOTYPE_C     ICON_ISO_

  • Invalid Column Name message in the subquery

    Hi All, I am receiving the invalid column name message in the subquery. Please help me if there is any work around or change in the query: Sample query: update table1 set table1.column1 = (Select column1 from (Select column1 from table2 where table1.

  • I don't think I input a wrong code but ...

    Why I cannot unlock the ipad? I dont think I input a wrong code. And the number I type did not show on the 4 squares...I saw a lock sign in the middle of the top bar. I also do not think someone change my passwork. How can I reset the machine?? Pls.