Exporting object from 734 to 10g

Hi,
I am going to migrate my Oracle 7.3.4 database to Oracle 10g. I am running only one instance which is having 3 schemas.
I want to know do I take an export of a particular schema so that entire objects of that schema's can imported to Oracle 10g.
Regards,
Indraneel

when i run the expdp it gives me following error
Export: Release 10.1.0.4.2 - Production on Wednesday, 27 April, 2011 11:43
Copyright (c) 2003, Oracle. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit
Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
ORA-39002: invalid operation
ORA-39070: Unable to open the log file.
ORA-29283: invalid file operation
ORA-06512: at "SYS.UTL_FILE", line 536
ORA-29283: invalid file operation
what can be wrong
Regards
Abbas

Similar Messages

  • Exporting schema from 11g to 10g

    how can i export full schema(suppost scott) from 11g to 10g both are on different machines and i dont know much about db side, if any body can suggest some queries.
    thanks and regards
    Abbas

    when i run the expdp it gives me following error
    Export: Release 10.1.0.4.2 - Production on Wednesday, 27 April, 2011 11:43
    Copyright (c) 2003, Oracle. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit
    Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-39002: invalid operation
    ORA-39070: Unable to open the log file.
    ORA-29283: invalid file operation
    ORA-06512: at "SYS.UTL_FILE", line 536
    ORA-29283: invalid file operation
    what can be wrong
    Regards
    Abbas

  • Export/import from 9i to 10G

    Dear all
    i have oracle 9i database on windows
    i want to migrate to 10G on windows
    i know it is done by export/import utility but how it can be done please let me know as i need step by step documentation of it
    please let me know its urgent
    Regards

    To do a proper upgrade, I would go through the Upgrade Guide make sure my server hardware and configuration met pre-requirement, there's no compatibility issue between my applications and newly upgraded server etc.
    The Guide is very detailed and thorough.
    I really don't think there's a document that is both simple and in details.
    The upgrade guide even has a section specially talk about export/import option
    http://download-west.oracle.com/docs/cd/B19306_01/server.102/b14238/expimp.htm#i262247

  • Export Process from BPM Stuido 10g

    The process can be exported from Studio manually.
    Does anyone know how to write the code (PBL or Java) to export the process?
    Or the API of connecting BPM Studio.
    Sincerely

    I'm not too sure why you would need to do all that, but it would involve quite a bit of custom work.....
    For the export itself, just zip the directory (without build and system directories) and rename to .exp I've not tried it this way, but it should work.
    Several of the components aren't going to work when renamed. They store both a 'name' and an 'id'. Process Ids cannot be changed easily (I don't recommend it).
    From what you're describing, why don't you just use Studio and the Process Admin... thats what it was intended to do?
    If its just the deployment, you can use the Ant scripts provided, and use them to deploy and publish a project to standalone.
    Hope this helps...

  • Migrate objects from ALUI 6.0 to WCI 10g?

    Is it possible to use the migration tool to export objects from an ALUI 6.0 portal and import them into a WCI 10g portal?

    No, migration needs to happen between same versions. You'll need to upgrade your first 6.0 instance to 10g then migrate objects from there to your other 10g instance.

  • Error while Transport objects from DEv to QA-File transport

    Hai Experts!
    While exporting objects from Dev to Quality using File Transport. it taking long time and no erro msg is thown and it is not getting end .Did any face this problem. Can any one help me to solve this.
    Regard's
    Preethi.

    Hai!
    Issue is solved.
    I try to export the Individual object -->Message mapping.
    While exporting i unchecked the Option With software component version Defenition.
    Now it is success. Thanks all replies.
    Regard's
    Preethi.

  • Help Importing objects from 1 user to another user and problem with trigger

    Hello community, i am having a little difficulty with the exporting objects from one user to another specifically the exporting of the trigger.
    Here is the situation, because of SOX purpose whenever a update is sent to the client the dba have to execute the script as himself (priviliged user) and is not allowed to log into the schema to make changes. Therefore we perpend the object definitions with &user_schema.. and they define user_schema in sqlplus and execute the update script.
    Here is a small example which requires two users (user1 and user2) with the following grants (connect, create table, create trigger, create view, create sequence). Please forgive the naming of the objects, just trying to be as simple as possible.
    I start out by logging in as system user via sqlplus and execute the following.
    ------------------Begin sqlplus----------------------
    define user_schema=user1;
    create table &user_schema..abc01 (
      col1 number,
      col2 varchar2(20),
      col3 number,
      constraint pk_abc01_col1 primary key (col1)
    create table &user_schema..xyz01 (
      col1 number,
      col2 varchar2(20),
      col3 number,
      constraint pk_xyz01_col1 primary key (col1)
    create or replace view &user_schema..view1 as
    select x.col1, x.col2, x.col3, a.col1 as acol1, a.col2 as acol2, a.col3 as acol3
    from xyz01 x
    inner join abc01 a on a.col1 = x.col1;
    create sequence &user_schema..seq_xyz01 start with 1 increment by 1;
    create or replace trigger &user_schema..trig01
    before insert on &user_schema..xyz01 for each row
    begin
      if (nvl(:new.col1, -1) = -1) then
        select seq_xyz01.nextval into :new.col1 from dual;
      end if;
    end;
    /--------------------End sqlplus----------------------
    I would then proceed to export using the exp utility via the command line
    exp system/systempassword file=user1.dmp owner=user1
    Then import user1 objects into user2
    imp system/systempassword file=user1.dmp fromuser=user1 touser=user2
    Now the problem:
    When i take a look at the sql for user2 trigger (trig01) i see the following (viewed via sqldeveloper)
    create or replace TRIGGER "USER2".trig01
    before insert on user1.xyz01 for each row
    begin
      if (nvl(:new.col1, -1) = -1) then
        select seq_xyz01.nextval into :new.col1 from dual;
      end if;
    end;its referring to user1.xyz01 table, however i want it to point to is user2.xyz01 table. Can someone please help me out or offer another solution to go about this because i need the ability to import the objects into a different user without the import failing and having to recompile the object.
    I've also tried executing this while connected as system user via sqlplus:
    define user_schema=user1
    create or replace trigger &user_schema..trig01
    before insert on xyz01 for each row
    begin
      if (nvl(:new.col1, -1) = -1) then
        select seq_xyz01.nextval into :new.col1 from dual;
      end if;
    end;
    /but that fails stating that table or view does not exist, please help
    however that fails because i
    Edited by: user3868150 on Nov 6, 2009 6:05 PM

    When performing an update in their system, the same script will be run with different values, thats not the problem.
    The client currently have just that one schema in their environment, however they want to have another instance of the application set up in their environment (same database) and have that go off on its own track separate from the original application.
    Now when we do an exp of the schema and the imp it into another user it get imported, however its incorrect. As stated before the trigger will be acting on the table in the original schema when it should be acting on the table in the newly imported schema.
    I suppose there is no other way around this when a trigger is created the way in which i outlined above. I guess after the data gets imported into a different user the trigger would have to be recompiled to point to the correct table.
    If you have an alternate solution to go about this i am open to suggestions. However like i mentioned in the original post because of SOX purpose the dba is not allowed to log in and execute update scripts as the schema user. The scripts should only be executed as that privileged user (dba).
    Also if i hard code the user when the trigger is created
    create or replace trigger user1.trig01
    before insert on user1.xyz01 for each row
    begin
    if (nvl(:new.col1, -1) = -1) then
    select seq_xyz01.nextval into :new.col1 from dual;
    end if;
    end;
    /it still doesn't get imported the way that i want it to be imported in user2
    create or replace TRIGGER "USER2".trig01
    before insert on user1.xyz01 for each row
    begin
      if (nvl(:new.col1, -1) = -1) then
        select seq_xyz01.nextval into :new.col1 from dual;
      end if;
    end;now when i do an insert to test
    insert into user1.xyz01 (col2, col3) values ('abc', 123); -- Works fine, no problems here
    insert into user2.xyz01 (col2, col3) values ('abc', 123);Results in an error ORA-01400: cannot insert NULL into ("USER2".XYZ01"."COL1") because the trigger doesn't exist on user2.xyz01 table.
    Just try and create the schema like how i outlined above in your environment to see what i'm talking about. It seems that the only way to get the trigger to be imported the way i want it to is to actually log in as that user and create the trigger.
    sqlplus user1/user1
    {code}
    create or replace trigger trig01
    before insert on xyz01 for each row
    begin
    if (nvl(:new.col1, -1) = -1) then
    select seq_xyz01.nextval into :new.col1 from dual;
    end if;
    end;
    {code}
    Edited by: user3868150 on Nov 6, 2009 6:10 PM

  • Exporting Data From Oracle 10g to Access

    Hi Experts,
    I need to export data from oracle apps shcema(each and every object in it) to MS ACCESS File.
    Please let me know how to do it
    Thanks and Regards,
    Andy

    Since you did not provide details of what you are planning to do, a very simple, however, depending upon the number of your tables, very labour intensive solution:
    Although being not very familiar with access, I think access offers a means to import csv files.
    So e.g. using SQL*Plus you can execute the following steps.
    set pagesize <nnn>
    set linesize <nnn>
    set colsep ',' -- You may choose other special characters to separate columns of your output
    spool <csv-file>
    select * from <tab>;
    spool offWherever you see <nnn> exchange by appropriate numeric values to get just one page as result and to configure your line size to hold a complete record in one output line.
    Replace <csv-file> and <tab> by the name of your output file to be read in to access and by the name of your table, respectively.

  • Moving/Exporting Seeburger configurations/objects from QA to Production

    Hi,
    I need some expert advise in moving Seeburger configuration objects like splitter adapter,AS2 adapter configurations from QA to Prod.
    Our scenario is like this , in the Seeburger message splitter we have defined the splitter adapter based on the AS2ID for different vendors/suppliers/customers. now when we go live on couple of customers every week, we are manually creating/configuring the seeburger channels in production. Is there a way to do it automatically like Exporting the objects in XI from QA to Prod???
    Any help is much appreciated.
    Thanks and Regards,
    Preeti

    Hi Preeti,
    For moving the congiguration objects from quality to production, you need to configure transport groups in the SLD so that when you are moving configuration objects from quality to production it will take Production R/3 system.
    For Transport Groups check the below links
    /people/sap.india5/blog/2005/11/03/xi-software-logistics-1-sld-preparation
    http://help.sap.com/saphelp_nw04/helpdata/en/ef/a21e3e0987760be10000000a114084/content.htm
    But you need to manually change AS2ID and Partner Number according to production environment in the Party which you configured for each customer in the ID.
    Regards,
    Prakash

  • Export error from Rep. Object Navigator

    Hi
    I got an error CDI-22582 when I tried to export container or workarea from Rep. Object Navigator. Errlog.err file contains this:
    EXP-00056: ORACLE error 12154 encountered
    ORA-12154: TNS:could not resolve service name
    EXP-00000: Export terminated unsuccessfully
    Everything is OK with my TNS. I can successfully connect to Rep. DB from Designer or Enterprise Manager. I'm using:
    Designer6i v. 6.5.40.20.0 c.4.012
    Thanks.

    Nenad,
    There are various causes to this issue. The most common one is the following:
    When you connect to your repository, you use the TNSNAMES.ORA from the Designer Oracle Home.
    When you export from Designer, it can be (depending on the registry variables EXECUTE_EXPORT) that you use the export executable from another Oracle Home, what means it uses a different TNSNAMES.ORA
    Could you please check if it's the case?
    If it is, make the file tnsnames.ora in both Oracle Homes identical.
    If you have access to MetaLink, I would recommend you to have a look to Note 160378.1: Common causes of CDI-22582 during export/import in Designer 6i and 9i
    Hope this helps,
    Didier.

  • Exporting database from oracle 11g and then importing it in oracle 10g

    Hi! there every one...
    please tell me the way to export the data and my data structures i.e the complete database from Oracle 11g to older version(oracle 10g) and how to import it there..
    i'm moving the database from one machine to another without lan, preferably via storage device...

    EXPORT FROM 11G to 10G.
    1) log on to 11G box.
    2) cd /<>/product/app/oracle/product/10.2/network/admin/
    3) Get the connection string from tnsnames.ora for the respective instances
    4) export ORACLE_HOME=/<>/app/oracle/product/10.2.0
    5) export LD_LIBRARY_PATH=$ORACLE_HOME/lib
    6) cd /<>/app/oracle/product/10.2.0/bin
    7) #>./exp file= test.dmp ......
    Username: user1@<connection_string>
    Password:*******
    8) Import as usual

  • How to export ESR objects from NWDS

    Hi Team
    Can you please let me know how can we export the ESR objects from NWDS.
    Right now I can not see options when I right click on the namespace
    I have already check this and but nothing is there for ESR objects.
    Transporting PI objects with NWDS (using CTS+) for Beginners

    Hi Indrajit,
    Please check the below discussion
    Import ESR content with eclipse development tool
    Import from a File
      In the main menu, choose Process Integration Transport Import PI Objects .
      In the Import from File dialog, select the .tpz file from the folder location.
      Choose Open
    regards,
    Harish

  • Data pump import from 11g to 10g

    I have 2 database: first is 11.2.0.2.0 and second is 10.2.0.1.0
    In 10g i created database link on 11g
    CREATE DATABASE LINK "TEST.LINK"
    CONNECT TO "monservice" IDENTIFIED BY "monservice"
    USING '(DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = host)(PORT = port))) (CONNECT_DATA = (SID = sid)))';
    And execute this query for test dbLink which work fine:
    select * from v$[email protected];
    After it i try to call open function:
    declare
    h number;
    begin
    h := dbms_datapump.open('IMPORT', 'TABLE', 'TEST.LINK', null, '10.2');
    end;
    and get exception: 39001. 00000 - "invalid argument value"
    if i remove 'TEST.LINK' from the arguments it works fine
    Edited by: 990594 on 26.02.2013 23:41

    Hi Richard Harrison,
    result for import from 11g to 10g:
    impdp user/pass@dburl schemas=SCHEMANAME network_link=TEST_LINK version=10.2
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    ORA-39001: invalid argument value
    ORA-39169: Local version of 10.2.0.1.0 cannot work with remote version of 11.2.0.2.0
    result for export from 11g to 10g:
    expdp user/pass@dburl schemas=SCHEMANAME network_link=TEST_LINK version=10.2
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64 bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing option
    ORA-39006: internal error
    ORA-39065: unexpected master process exception in DISPATCH
    ORA-04052: error occurred when looking up remote object SYS.KUPM$MCP@TEST_LINK
    ORA-00604: error occurred at recursive SQL level 3
    ORA-06544: PL/SQL: internal, error, arguments: [55916], [], [], [], [], [], [], []
    ORA-06553: PLS-801: internal error [55916]
    ORA-02063: preceding 2 lines from TEST_LINK
    ORA_39097: Data Pump job encountered unexpected error -4052

  • Is it possible to export tables from diffrent schema using expdp?

    Hi,
    We can export tables from different schema using exp. Ex: exp user/pass file=sample.dmp log=sample.log tables=scott.dept,system.sales ...But
    Is it possible in expdp?
    Thanks in advance ..
    Thanks,

    Hi,
    you have to use "schemas=user1,user2 include=table:"in('table1,table2')" use parfileexpdp scott/tiger@db10g schemas=SCOTT include=TABLE:"IN ('EMP', 'DEPT')" directory=TEST_DIR dumpfile=SCOTT.dmp logfile=expdpSCOTT.log{quote}
    I am not able to perform it using parfile also.Using parfile it shows "UDE-00010: multiple job modes requested, schema and tables."
    When trying the below, i get error
    {code}
    bash-3.00$ expdp directory=EXP_DUMP dumpfile=test.dmp logfile=test.log SCHEMAS=(\'MM\',\'MMM\') include=TABLE:\"IN\(\'EA_EET_TMP\',\'WS_DT\'\)\"
    Export: Release 10.2.0.4.0 - 64bit Production on Friday, 15 October, 2010 18:34:32
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Username: / as sysdba
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYS"."SYS_EXPORT_SCHEMA_01": /******** AS SYSDBA directory=EXP_DUMP dumpfile=test.dmp logfile=test.log SCHEMAS=('MM','MMM') include=TABLE:"IN('EA_EET_TMP','WS_DT')"
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "MM"."EA_EET_TMP" 0 KB 0 rows
    ORA-39165: Schema MMM was not found.
    Master table "SYS"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    Dump file set for SYS.SYS_EXPORT_SCHEMA_01 is:
    /export/home/nucleus/dump/test.dmp
    Job "SYS"."SYS_EXPORT_SCHEMA_01" completed with 1 error(s) at 18:35:19
    {code}
    When checking expdp help=y shows :-
    {code}TABLES Identifies a list of tables to export - one schema only.{code}
    As per few testing,tables from different schemas are not possible to export using expdp in a single command.
    Anand

  • Exporting & Importing Contexts in ODI 10g

    Hi,
    I am using ODI 10g.
    I want to export Contexts from topology manager from my ODI test environment and import those contexts in ODI production environment.
    My requirement is that I want to export & import the contexts using some scripts and not manually.
    What steps can I follow for this?
    Thanks,
    Divya

    Hi Divya,
    Personally i feel rather than exporting individual components/objects, its suggested to export master repository as such.
    You can make use of ODI utility OdiExportMaster (under <ur package>->Tools->Oracle Data Integrator Objects) for exporting and Import Master Repository wizard (All Programs->Oracle-> Oracle Data Integrator->Repository Management-> Master Repository Import )for importing.
    Thanks,
    Guru.

Maybe you are looking for