Expdp remap_data

Hi,
Is there any way to use remap_data option of 11g's expdp in 10g, or
Is there any way to use 11g's expdp executable in 10g
Kind Regards..

No.

Similar Messages

  • K-anonymity algorithm in sql

    Hi,
    does anyone have information on how to implemend k-anonymity algorithm with oracle or soemthing relevant?
    I would be thankful if someone could help me.

    I do not know if this helps, but you may want to try to use datapump
    expdp remap_data=table_name.column_name:pkg.function ...
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28319/dp_export.htm#BABDCEDA
    well, it is not possible in 10.2

  • Pass more parameter to the packaged function with REMAP_DATA in EXPDP

    1) can we pass more than one parameter to the packaged function specified in REMAP_DATA during DATA pump export?
    2) if yes to point#1 then, how can we pass rownum & some other columns value in the same record passed into the function?

    when all else fails, Read The Fine Manual
    http://download.oracle.com/docs/cd/E11882_01/server.112/e16536/dp_export.htm#sthref115

  • Remap_data feature in DBMS_DATAPUMP

    Hi,
    I can't see how to use the REMAP_DATA expdp feature from the DBMS_DATAPUMP API.
    Did I miss something?
    Thanks.

    Hi,
    Anything available through the client is available through the api. Through the impdp client, you specify it like:
    remap_data=hr.employees.employee_id:hr.remap_doc.minus10
    What this means is
    data in the hr.employee table and column employee_id will be remapped.
    The value in that column will be sent to the function hr.remap_doc.minus10 and the value returned from that function will be stored in the row in the table.
    So, let's assume that the minus10 function takes in a number and simply subtracts 10 from it. If your table had this
    hr.employees column employee_id
    ======================
    20
    21
    22
    23
    24
    25
    After using the remap_data the table would be:
    hr.employees column employee_id
    ======================
    10
    11
    12
    13
    14
    15
    Here is the definition that I find in dbmsdp.sql
    -- handle - Identifies the particular job or operation (from OPEN)
    -- name - The type of data remapping to be performed
    -- table_name - Table name for applying the remap.
    -- column - Column name for where the data needs to be remapped
    -- function - Function used to remap the column data.
    -- schema - Name of the schema owning the table to apply the data
    -- remap. Null implies all schemas in the job.
    Hope this helps.
    Dean

  • Incremental EXPDP !!!

    Dear all,
    I have a production database on windows 2003 server 64 bit with oracle 10.2.0.3 installed. This db has a schema which expdp file is of 33gb, every day I am taking the schema level export and importing the same to my test server.
    Moreover, we also had configured flashback database.
    Its taking a lot of time to import everyday despite there is little changes in the schema. Our intention is to refresh the test schema every day.
    Is there a way to take the incremental expdp/flashback to record only the changes and then import it into the test server.
    Thanks.

    C:\>expdp help=y
    Export: Release 11.1.0.6.0 - Production on Friday, 16 October, 2009 10:41:50
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    The Data Pump export utility provides a mechanism for transferring data objects
    between Oracle databases. The utility is invoked with the following command:
       Example: expdp scott/tiger DIRECTORY=dmpdir DUMPFILE=scott.dmp
    You can control how Export runs by entering the 'expdp' command followed
    by various parameters. To specify parameters, you use keywords:
       Format:  expdp KEYWORD=value or KEYWORD=(value1,value2,...,valueN)
       Example: expdp scott/tiger DUMPFILE=scott.dmp DIRECTORY=dmpdir SCHEMAS=scott
                   or TABLES=(T1:P1,T1:P2), if T1 is partitioned table
    USERID must be the first parameter on the command line.
    Keyword               Description (Default)
    ATTACH                Attach to existing job, e.g. ATTACH [=job name].
    COMPRESSION           Reduce size of dumpfile contents where valid keyword.
                          values are: ALL, (METADATA_ONLY), DATA_ONLY and NONE.
    CONTENT               Specifies data to unload where the valid keyword
                          values are: (ALL), DATA_ONLY, and METADATA_ONLY.
    DATA_OPTIONS          Data layer flags where the only valid value is:
                          XML_CLOBS-write XML datatype in CLOB format
    DIRECTORY             Directory object to be used for dumpfiles and logfiles.
    DUMPFILE              List of destination dump files (expdat.dmp),
                          e.g. DUMPFILE=scott1.dmp, scott2.dmp, dmpdir:scott3.dmp.
    ENCRYPTION            Encrypt part or all of the dump file where valid keyword
                          values are: ALL, DATA_ONLY, METADATA_ONLY,
                          ENCRYPTED_COLUMNS_ONLY, or NONE.
    ENCRYPTION_ALGORITHM  Specify how encryption should be done where valid
                          keyword values are: (AES128), AES192, and AES256.
    ENCRYPTION_MODE       Method of generating encryption key where valid keyword
                          values are: DUAL, PASSWORD, and (TRANSPARENT).
    ENCRYPTION_PASSWORD   Password key for creating encrypted column data.
    ESTIMATE              Calculate job estimates where the valid keyword
                          values are: (BLOCKS) and STATISTICS.
    ESTIMATE_ONLY         Calculate job estimates without performing the export.
    EXCLUDE               Exclude specific object types, e.g. EXCLUDE=TABLE:EMP.
    FILESIZE              Specify the size of each dumpfile in units of bytes.
    FLASHBACK_SCN         SCN used to set session snapshot back to.
    FLASHBACK_TIME        Time used to get the SCN closest to the specified time.
    FULL                  Export entire database (N).
    HELP                  Display Help messages (N).
    INCLUDE               Include specific object types, e.g. INCLUDE=TABLE_DATA.
    JOB_NAME              Name of export job to create.
    LOGFILE               Log file name (export.log).
    NETWORK_LINK          Name of remote database link to the source system.
    NOLOGFILE             Do not write logfile (N).
    PARALLEL              Change the number of active workers for current job.
    PARFILE               Specify parameter file.
    QUERY                 Predicate clause used to export a subset of a table.
    REMAP_DATA            Specify a data conversion function,
                          e.g. REMAP_DATA=EMP.EMPNO:REMAPPKG.EMPNO.
    REUSE_DUMPFILES       Overwrite destination dump file if it exists (N).
    SAMPLE                Percentage of data to be exported;
    SCHEMAS               List of schemas to export (login schema).
    STATUS                Frequency (secs) job status is to be monitored where
                          the default (0) will show new status when available.
    TABLES                Identifies a list of tables to export - one schema only.
    TABLESPACES           Identifies a list of tablespaces to export.
    TRANSPORTABLE         Specify whether transportable method can be used where
                          valid keyword values are: ALWAYS, (NEVER).
    TRANSPORT_FULL_CHECK  Verify storage segments of all tables (N).
    TRANSPORT_TABLESPACES List of tablespaces from which metadata will be unloaded.
    VERSION               Version of objects to export where valid keywords are:
                          (COMPATIBLE), LATEST, or any valid database version.
    The following commands are valid while in interactive mode.
    Note: abbreviations are allowed
    Command               Description
    ADD_FILE              Add dumpfile to dumpfile set.
    CONTINUE_CLIENT       Return to logging mode. Job will be re-started if idle.
    EXIT_CLIENT           Quit client session and leave job running.
    FILESIZE              Default filesize (bytes) for subsequent ADD_FILE commands.
    HELP                  Summarize interactive commands.
    KILL_JOB              Detach and delete job.
    PARALLEL              Change the number of active workers for current job.
                          PARALLEL=<number of workers>.
    REUSE_DUMPFILES       Overwrite destination dump file if it exists (N).
    START_JOB             Start/resume current job.
    STATUS                Frequency (secs) job status is to be monitored where
                          the default (0) will show new status when available.
                          STATUS[=interval]
    STOP_JOB              Orderly shutdown of job execution and exits the client.
                          STOP_JOB=IMMEDIATE performs an immediate shutdown of the
                          Data Pump job.Nothing for incrementally updating the clone database.
    Once you create a database using logical backup, it's altogether new database i.e. no links with the original database. So it's hard to co-relate the databases.
    Regards,
    S.K.

  • Expdp /impdp  Win 2003 Oracle 11g

    Hi Guys,
    I did the export from production and I want to restore on dev. What is the sysntax for impdp.
    expdp userid=system/system dumpfile=livelink.dmp schemas=livelink logfile=livelink.log
    impdp userid=system/system dumpfile=e:/oradata/livelink.dmp ........ Is there fromuser=livelink touser=livelink
    Thanx.

    Handle:      NPD
    Status Level:      Newbie
    Registered:      Mar 15, 2007
    Total Posts:      359
    Total Questions:      82 (73 unresolved)
    so many questions & so few answers.
    Is there fromuser=livelink touser=livelinkNO
    check out SCHEMAS
    bcm@bcm-laptop:~$ impdp help=yes
    Import: Release 11.2.0.1.0 - Production on Thu Oct 7 08:44:28 2010
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    The Data Pump Import utility provides a mechanism for transferring data objects
    between Oracle databases. The utility is invoked with the following command:
         Example: impdp scott/tiger DIRECTORY=dmpdir DUMPFILE=scott.dmp
    You can control how Import runs by entering the 'impdp' command followed
    by various parameters. To specify parameters, you use keywords:
         Format:  impdp KEYWORD=value or KEYWORD=(value1,value2,...,valueN)
         Example: impdp scott/tiger DIRECTORY=dmpdir DUMPFILE=scott.dmp
    USERID must be the first parameter on the command line.
    The available keywords and their descriptions follow. Default values are listed within square brackets.
    ATTACH
    Attach to an existing job.
    For example, ATTACH=job_name.
    CONTENT
    Specifies data to load.
    Valid keywords are: [ALL], DATA_ONLY and METADATA_ONLY.
    DATA_OPTIONS
    Data layer option flags.
    Valid keywords are: SKIP_CONSTRAINT_ERRORS.
    DIRECTORY
    Directory object to be used for dump, log and sql files.
    DUMPFILE
    List of dumpfiles to import from [expdat.dmp].
    For example, DUMPFILE=scott1.dmp, scott2.dmp, dmpdir:scott3.dmp.
    ENCRYPTION_PASSWORD
    Password key for accessing encrypted data within a dump file.
    Not valid for network import jobs.
    ESTIMATE
    Calculate job estimates.
    Valid keywords are: [BLOCKS] and STATISTICS.
    EXCLUDE
    Exclude specific object types.
    For example, EXCLUDE=SCHEMA:"='HR'".
    FLASHBACK_SCN
    SCN used to reset session snapshot.
    FLASHBACK_TIME
    Time used to find the closest corresponding SCN value.
    FULL
    Import everything from source [Y].
    HELP
    Display help messages [N].
    INCLUDE
    Include specific object types.
    For example, INCLUDE=TABLE_DATA.
    JOB_NAME
    Name of import job to create.
    LOGFILE
    Log file name [import.log].
    NETWORK_LINK
    Name of remote database link to the source system.
    NOLOGFILE
    Do not write log file [N].
    PARALLEL
    Change the number of active workers for current job.
    PARFILE
    Specify parameter file.
    PARTITION_OPTIONS
    Specify how partitions should be transformed.
    Valid keywords are: DEPARTITION, MERGE and [NONE].
    QUERY
    Predicate clause used to import a subset of a table.
    For example, QUERY=employees:"WHERE department_id > 10".
    REMAP_DATA
    Specify a data conversion function.
    For example, REMAP_DATA=EMP.EMPNO:REMAPPKG.EMPNO.
    REMAP_DATAFILE
    Redefine datafile references in all DDL statements.
    REMAP_SCHEMA
    Objects from one schema are loaded into another schema.
    REMAP_TABLE
    Table names are remapped to another table.
    For example, REMAP_TABLE=EMP.EMPNO:REMAPPKG.EMPNO.
    REMAP_TABLESPACE
    Tablespace object are remapped to another tablespace.
    REUSE_DATAFILES
    Tablespace will be initialized if it already exists [N].
    SCHEMAS
    List of schemas to import.
    SKIP_UNUSABLE_INDEXES
    Skip indexes that were set to the Index Unusable state.
    SOURCE_EDITION
    Edition to be used for extracting metadata.
    SQLFILE
    Write all the SQL DDL to a specified file.
    STATUS
    Frequency (secs) job status is to be monitored where
    the default [0] will show new status when available.
    STREAMS_CONFIGURATION
    Enable the loading of Streams metadata
    TABLE_EXISTS_ACTION
    Action to take if imported object already exists.
    Valid keywords are: APPEND, REPLACE, [SKIP] and TRUNCATE.
    TABLES
    Identifies a list of tables to import.
    For example, TABLES=HR.EMPLOYEES,SH.SALES:SALES_1995.
    TABLESPACES
    Identifies a list of tablespaces to import.
    TARGET_EDITION
    Edition to be used for loading metadata.
    TRANSFORM
    Metadata transform to apply to applicable objects.
    Valid keywords are: OID, PCTSPACE, SEGMENT_ATTRIBUTES and STORAGE.
    TRANSPORTABLE
    Options for choosing transportable data movement.
    Valid keywords are: ALWAYS and [NEVER].
    Only valid in NETWORK_LINK mode import operations.
    TRANSPORT_DATAFILES
    List of datafiles to be imported by transportable mode.
    TRANSPORT_FULL_CHECK
    Verify storage segments of all tables [N].
    TRANSPORT_TABLESPACES
    List of tablespaces from which metadata will be loaded.
    Only valid in NETWORK_LINK mode import operations.
    VERSION
    Version of objects to import.
    Valid keywords are: [COMPATIBLE], LATEST or any valid database version.
    Only valid for NETWORK_LINK and SQLFILE.
    The following commands are valid while in interactive mode.
    Note: abbreviations are allowed.
    CONTINUE_CLIENT
    Return to logging mode. Job will be restarted if idle.
    EXIT_CLIENT
    Quit client session and leave job running.
    HELP
    Summarize interactive commands.
    KILL_JOB
    Detach and delete job.
    PARALLEL
    Change the number of active workers for current job.
    START_JOB
    Start or resume current job.
    Valid keywords are: SKIP_CURRENT.
    STATUS
    Frequency (secs) job status is to be monitored where
    the default [0] will show new status when available.
    STOP_JOB
    Orderly shutdown of job execution and exits the client.
    Valid keywords are: IMMEDIATE.
    bcm@bcm-laptop:~$ Edited by: sb92075 on Oct 7, 2010 8:53 AM

  • Fromuser/touser equivanet in expdp/impdp ??

    hi .
    i got a dump file from the following cmmand.
    expdp system/123456@SCHDB dumpfile=studentinfo.dmp logfile=studentinfo.log tables=school.studentmaster,school.studentmarks, school.studentleave directory=mydir
    now i want to import these 2 tables into a different schema (old_school) using impdp
    whats the fromuser / touser equivalent in impdp/expdp ???

    REMAP_DATA
    Specify a data conversion function.
    For example, REMAP_DATA=EMP.EMPNO:REMAPPKG.EMPNO.
    REMAP_DATAFILE
    Redefine data file references in all DDL statements.
    REMAP_SCHEMA
    Objects from one schema are loaded into another schema.
    REMAP_TABLE
    Table names are remapped to another table.
    For example, REMAP_TABLE=HR.EMPLOYEES:EMPS.
    REMAP_TABLESPACE
    Tablespace objects are remapped to another tablespace.
    REUSE_DATAFILES
    Tablespace will be initialized if it already exists [N].

  • How to run expdp from client ?

    Hi All,
    I tried searching google and forums for my issue but to no avail > How to run expdp from client side ....like in my laptop.
    Because currently our PROD database server has no space for expdp dump file. So I want it directed to my laptop which has an extenal USB of 1TB harddisk...via client EXPDP
    import data using impdp command
    Posted on: 08-May-2012 11:36, by user: 841731 -- Relevance: 53% -- Show all results within this thread
    below command is correct or not? if it is not correct could you please send me the correct command. impdp user/pass@databasename schemas=sourceschemaname remap_schema=sourceschemaname:destinationschemaname ...
    System generated Index names different on target database after expdp/impdp
    Posted on: 30-May-2012 11:58, by user: 895124 -- Relevance: 43% -- Show all results within this thread
    After performing expdp/impdp to move data from one database (A) to another (B), the system name generated indexes has different ...
    [ETL] TTS vs expdp/impdp vs ctas (dblink
    Posted on: 08-May-2012 21:10, by user: 869578 -- Relevance: 39% -- Show all results within this thread
    (table : 500 giga, index : 500 giga), how much faster is TTS (transportable tablespace) over expdp/impdp, and over ctas (dblink) ? As you know, the speed of etl depends on the hardware capacity. (io ...
    Oracle Client
    Posted on: 21-Jun-2012 22:47, by user: Sh**** -- Relevance: 32% -- Show all results within this thread
    Hi Guys, Please can you guys elaborate the difference between Oracle Client and Oracle Instant Client. Also, please can you advise from where I can download the Oracle normal ...
    Oracle 10g Client
    Posted on: 05-Jun-2012 10:11, by user: dzm -- Relevance: 26% -- Show all results within this thread
    to search at oracle site and this forum, but i wasn't able to find a link to download the oracle 10g client. I really need especificaly the 10g version. Anybody know the link or another way to download ...
    9i client to access 11g database
    Posted on: 22-Jun-2012 07:31, by user: kkrm333 -- Relevance: 24% -- Show all results within this thread
    Hi, Can i access a 11g database using 9i client? Thanks,
    SQLplus in Oracle Client
    Posted on: 14-Jun-2012 00:36, by user: Tim B. -- Relevance: 24% -- Show all results within this thread
    Hi, I tried to install an 11g oracle client in linux. As I've compared the files with the files when you install using the oracle instant ...
    Re: Information on Oracle Client 11202-1.1.4-6
    Posted on: 05-Jun-2012 03:33, by user: 898763 -- Relevance: 23% -- Show all results within this thread
    Actually thats the client requirement
    Analysing the performance of a single client
    Posted on: 28-Mar-2012 02:05, by user: 880172 -- Relevance: 23% -- Show all results within this thread
    timeouts even on some of the simplest queries. I want to try and get some data about how just this one client is performing and what it’s doing, but everything Google has thrown up so far is orientated around ...
    to make client connection as sys
    Posted on: 12-Jun-2012 22:04, by user: user11221081 -- Relevance: 23% -- Show all results within this thread
    Dear gurus can i connect to my server from my client machine with sysdba without giving sys password i have connected in different ways as sys@abc ...Thanks a lot.

    Though you can initiate the binary from your client side but for the file creation, there is no other way but to store it on the server side. So your best bet would be to get some space free on the server side of yours.
    Aman....

  • Data Pump - expdp and slow performance on specific tables

    Hi there
    I have af data pump export af a schema. Most of the 700 tables is exported very quickly (direct path) but a couple of them seems to be extremenly slow.
    I have chekced:
    - no lobs
    - no long/raw
    - no VPD
    - no partitions
    - no bitmapped index
    - just date, number, varchar2's
    I'm runing with trace 400300
    But I'm having trouble reading the output from it. It seems that some of the slow performning tables is runinng with method 4??? Can anyone find an explanation for the method in the trace:
    1 > direct path (i think)
    2 > external table (i think)
    4 > ?
    others?
    I have done some stats using v$filestat/v$session_wait (history) - and it seems that we always wait for DB seq file read - and doing lots and lots of SINGLEBLKRDS. Not undo is read
    I have a table 2.5 GB -> 3 minutes
    and then this (in my eyes) similar table 2.4 GB > 1½ hrs.
    There are 367.000 blks (8 K) and avg rowlen = 71
    I'm on Oracle 11.2 on a Linux box with plenty of RAM and CPU power.
    Trace file /opt/oracle112/diag/rdbms/prod/prod/trace/prod_dw00_24268.trc
    Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /opt/oracle112/product/11.2.0.2/dbhome_1
    System name:  Linux
    Node name:  tiaprod.thi.somethingamt.dk
    Release:  2.6.18-194.el5
    Version:  #1 SMP Mon Mar 29 22:10:29 EDT 2010
    Machine:  x86_64
    VM name:  Xen Version: 3.4 (HVM)
    Instance name: prod
    Redo thread mounted by this instance: 1
    Oracle process number: 222
    Unix process pid: 24268, image: [email protected] (DW00)
    *** 2011-09-20 09:39:39.671
    *** SESSION ID:(401.8395) 2011-09-20 09:39:39.671
    *** CLIENT ID:() 2011-09-20 09:39:39.671
    *** SERVICE NAME:(SYS$BACKGROUND) 2011-09-20 09:39:39.671
    *** MODULE NAME:() 2011-09-20 09:39:39.671
    *** ACTION NAME:() 2011-09-20 09:39:39.671
    KUPP:09:39:39.670: Current trace/debug flags: 00400300 = 4195072
    *** MODULE NAME:(Data Pump Worker) 2011-09-20 09:39:39.672
    *** ACTION NAME:(SYS_EXPORT_SCHEMA_09) 2011-09-20 09:39:39.672
    KUPW:09:39:39.672: 0: ALTER SESSION ENABLE PARALLEL DML called.
    KUPW:09:39:39.672: 0: ALTER SESSION ENABLE PARALLEL DML returned.
    KUPC:09:39:39.693: Setting remote flag for this process to FALSE
    prvtaqis - Enter
    prvtaqis subtab_name upd
    prvtaqis sys table upd
    KUPW:09:39:39.819: 0: KUPP$PROC.WHATS_MY_ID called.
    KUPW:09:39:39.819: 1: KUPP$PROC.WHATS_MY_ID returned.
    KUPW:09:39:39.820: 1: worker max message number: 1000
    KUPW:09:39:39.822: 1: Full cluster access allowed
    KUPW:09:39:39.823: 1: Original job start time: 11-SEP-20 09:39:38 AM
    KUPW:09:39:39.862: 1: KUPP$PROC.WHATS_MY_NAME called.
    KUPW:09:39:39.862: 1: KUPP$PROC.WHATS_MY_NAME returned. Process name: DW00
    KUPW:09:39:39.862: 1: KUPV$FT_INT.GET_INSTANCE_ID called.
    KUPW:09:39:39.866: 1: KUPV$FT_INT.GET_INSTANCE_ID returned. Instance name: prod
    KUPW:09:39:39.870: 1: ALTER SESSION ENABLE RESUMABLE called.
    KUPW:09:39:39.870: 1: ALTER SESSION ENABLE RESUMABLE returned.
    KUPW:09:39:39.871: 1: KUPF$FILE.INIT called.
    KUPW:09:39:39.996: 1: KUPF$FILE.INIT returned.
    KUPW:09:39:39.998: 1: KUPF$FILE.GET_MAX_CSWIDTH called.
    KUPW:09:39:39.998: 1: KUPF$FILE.GET_MAX_CSWIDTH returned.
    KUPW:09:39:39.998: 1: Max character width: 1
    KUPW:09:39:39.998: 1: Max clob fetch: 32757
    KUPW:09:39:39.998: 1: Max varchar2a size: 32757
    KUPW:09:39:39.998: 1: Max varchar2 size: 7990
    KUPW:09:39:39.998: 1: In procedure GET_PARAMETERS
    KUPW:09:39:40.000: 1: In procedure GET_METADATA_FILTERS
    KUPW:09:39:40.001: 1: In procedure GET_METADATA_TRANSFORMS
    KUPW:09:39:40.002: 1: In procedure GET_DATA_FILTERS
    KUPW:09:39:40.004: 1: In procedure GET_DATA_REMAPS
    KUPW:09:39:40.005: 1: In procedure PRINT_MT_PARAMS
    KUPW:09:39:40.005: 1: Master table             : "SYSTEM"."SYS_EXPORT_SCHEMA_09"
    KUPW:09:39:40.005: 1: Metadata job mode        : SCHEMA_EXPORT
    KUPW:09:39:40.005: 1: Debug enable             : TRUE
    KUPW:09:39:40.005: 1: Profile enable           : FALSE
    KUPW:09:39:40.005: 1: Transportable enable     : FALSE
    KUPW:09:39:40.005: 1: Metrics enable           : FALSE
    KUPW:09:39:40.005: 1: db version               : 11.2.0.2.0
    KUPW:09:39:40.005: 1: job version              : 11.2.0.0.0
    KUPW:09:39:40.005: 1: service name             :
    KUPW:09:39:40.005: 1: Current Edition          : ORA$BASE
    KUPW:09:39:40.005: 1: Job Edition              :
    KUPW:09:39:40.005: 1: Abort Step               : 0
    KUPW:09:39:40.005: 1: Access Method            : AUTOMATIC
    KUPW:09:39:40.005: 1: Data Options             : 0
    KUPW:09:39:40.006: 1: Dumper directory         :
    KUPW:09:39:40.006: 1: Master only              : FALSE
    KUPW:09:39:40.006: 1: Data Only                : FALSE
    KUPW:09:39:40.006: 1: Metadata Only            : FALSE
    KUPW:09:39:40.006: 1: Estimate                 : BLOCKS
    KUPW:09:39:40.006: 1: Data error logging table :
    KUPW:09:39:40.006: 1: Remote Link              :
    KUPW:09:39:40.006: 1: Dumpfile present         : TRUE
    KUPW:09:39:40.006: 1: Table Exists Action      :
    KUPW:09:39:40.006: 1: Partition Options        : NONE
    KUPW:09:39:40.006: 1: Tablespace Datafile Count: 0
    KUPW:09:39:40.006: 1: Metadata Filter    Index : 1    Count : 10
    KUPW:09:39:40.006: 1:         1           Name - INCLUDE_USER
    KUPW:09:39:40.006: 1:                    Value - TRUE
    KUPW:09:39:40.006: 1:              Object Name - SCHEMA_EXPORT
    KUPW:09:39:40.006: 1:         2           Name - SCHEMA_EXPR
    KUPW:09:39:40.006: 1:                    Value -  IN ('TIA')
    KUPW:09:39:40.006: 1:         3           Name - NAME_EXPR
    KUPW:09:39:40.006: 1:                    Value -  ='ACC_PAYMENT_SPECIFICATION'
    KUPW:09:39:40.006: 1:                   Object - TABLE
    KUPW:09:39:40.006: 1:         4           Name - INCLUDE_PATH_EXPR
    KUPW:09:39:40.006: 1:                    Value -  IN ('TABLE')
    KUPW:09:39:40.006: 1:         5           Name - ORDERED
    KUPW:09:39:40.006: 1:                    Value - FALSE
    KUPW:09:39:40.006: 1:                   Object - TABLE_DATA
    KUPW:09:39:40.006: 1:         6           Name - NO_XML
    KUPW:09:39:40.006: 1:                    Value - TRUE
    KUPW:09:39:40.006: 1:                   Object - XMLSCHEMA/EXP_XMLSCHEMA
    KUPW:09:39:40.006: 1:         7           Name - XML_OUTOFLINE
    KUPW:09:39:40.006: 1:                    Value - FALSE
    KUPW:09:39:40.006: 1:                   Object - TABLE/TABLE_DATA
    KUPW:09:39:40.006: 1:         8           Name - XDB_GENERATED
    KUPW:09:39:40.006: 1:                    Value - FALSE
    KUPW:09:39:40.006: 1:                   Object - TABLE/TRIGGER
    KUPW:09:39:40.007: 1:         9           Name - XDB_GENERATED
    KUPW:09:39:40.007: 1:                    Value - FALSE
    KUPW:09:39:40.007: 1:                   Object - TABLE/RLS_POLICY
    KUPW:09:39:40.007: 1:         10           Name - PRIVILEGED_USER
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1: MD remap schema    Index : 4    Count : 0
    KUPW:09:39:40.007: 1: MD remap other     Index : 5    Count : 0
    KUPW:09:39:40.007: 1: MD Transform ddl   Index : 2    Count : 11
    KUPW:09:39:40.007: 1:         1           Name - DBA
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1:                   Object - JOB
    KUPW:09:39:40.007: 1:         2           Name - EXPORT
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1:         3           Name - PRETTY
    KUPW:09:39:40.007: 1:                    Value - FALSE
    KUPW:09:39:40.007: 1:         4           Name - SQLTERMINATOR
    KUPW:09:39:40.007: 1:                    Value - FALSE
    KUPW:09:39:40.007: 1:         5           Name - CONSTRAINTS
    KUPW:09:39:40.007: 1:                    Value - FALSE
    KUPW:09:39:40.007: 1:                   Object - TABLE
    KUPW:09:39:40.007: 1:         6           Name - REF_CONSTRAINTS
    KUPW:09:39:40.007: 1:                    Value - FALSE
    KUPW:09:39:40.007: 1:                   Object - TABLE
    KUPW:09:39:40.007: 1:         7           Name - OID
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1:                   Object - TABLE
    KUPW:09:39:40.007: 1:         8           Name - RESET_PARALLEL
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1:                   Object - INDEX
    KUPW:09:39:40.007: 1:         9           Name - OID
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1:                   Object - TYPE
    KUPW:09:39:40.007: 1:         10           Name - OID
    KUPW:09:39:40.007: 1:                    Value - TRUE
    KUPW:09:39:40.007: 1:                   Object - INC_TYPE
    KUPW:09:39:40.007: 1:         11           Name - REVOKE_FROM
    KUPW:09:39:40.008: 1:                    Value - SYSTEM
    KUPW:09:39:40.008: 1:                   Object - ROLE
    KUPW:09:39:40.008: 1: Data Filter        Index : 6    Count : 0
    KUPW:09:39:40.008: 1: Data Remap         Index : 7    Count : 0
    KUPW:09:39:40.008: 1: MD remap name      Index : 8    Count : 0
    KUPW:09:39:40.008: 1: In procedure DISPATCH_WORK_ITEMS
    KUPW:09:39:40.009: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:39:40.009: 1: KUPC$QUEUE.TRANSCEIVE called.
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  7
    kwqberlst ascn -90145310 lascn 22
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    KUPW:09:39:40.036: 1: KUPC$QUEUE.TRANSCEIVE returned. Received 2011
    KUPW:09:39:40.036: 1: DBMS_LOB.CREATETEMPORARY called.
    KUPW:09:39:40.037: 1: DBMS_LOB.CREATETEMPORARY returned.
    KUPW:09:39:40.038: 1: Flags: 18
    KUPW:09:39:40.038: 1: Start sequence number:
    KUPW:09:39:40.038: 1: End sequence number:
    KUPW:09:39:40.038: 1: Metadata Parallel: 1
    KUPW:09:39:40.038: 1: Primary worker id: 1
    KUPW:09:39:40.041: 1: In procedure GET_TABLE_DATA_OBJECTS
    KUPW:09:39:40.041: 1: In procedure CREATE_MSG
    KUPW:09:39:40.041: 1: KUPV$FT.MESSAGE_TEXT called.
    KUPW:09:39:40.041: 1: KUPV$FT.MESSAGE_TEXT returned.
    KUPW:09:39:40.041: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:39:40.041: 1: KUPC$QUEUE_INT.SEND called.
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  7
    kwqberlst ascn -90145310 lascn 22
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    KUPW:09:39:40.044: 1: KUPC$QUEUE_INT.SEND returned.
    KUPW:09:39:40.044: 1: Estimate in progress using BLOCKS method...
    KUPW:09:39:40.044: 1: In procedure UPDATE_TYPE_COMPLETION_ROW
    KUPW:09:39:40.044: 1: Old Seqno: 0 New Path: SCHEMA_EXPORT/TABLE/TABLE_DATA PO Num: -5 New Seqno: 62
    KUPW:09:39:40.046: 1: Created type completion for duplicate 62
    KUPW:09:39:40.046: 1: In procedure CREATE_MSG
    KUPW:09:39:40.046: 1: KUPV$FT.MESSAGE_TEXT called.
    KUPW:09:39:40.046: 1: KUPV$FT.MESSAGE_TEXT returned.
    KUPW:09:39:40.046: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:39:40.046: 1: KUPC$QUEUE_INT.SEND called.
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  7
    kwqberlst ascn -90145310 lascn 22
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    KUPW:09:39:40.047: 1: KUPC$QUEUE_INT.SEND returned.
    KUPW:09:39:40.047: 1: Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    KUPW:09:39:40.048: 1: In procedure CONFIGURE_METADATA_UNLOAD
    KUPW:09:39:40.048: 1: Phase: ESTIMATE_PHASE Filter Name:  Filter Value:
    KUPW:09:39:40.048: 1: DBMS_METADATA.OPEN11.2.0.0.0 called.
    KUPW:09:39:40.182: 1: DBMS_METADATA.OPEN11.2.0.0.0 returned. Source handle: 100001
    KUPW:09:39:40.182: 1: DBMS_METADATA.SET_FILTER called. metadata_phase: ESTIMATE_PHASE
    KUPW:09:39:40.182: 1: DBMS_METADATA.SET_FILTER returned. In function GET_NOEXP_TABLE
    KUPW:09:39:40.194: 1: DBMS_METADATA.SET_PARSE_ITEM called.
    *** 2011-09-20 09:39:40.325
    KUPW:09:39:40.325: 1: DBMS_METADATA.SET_PARSE_ITEM returned.
    KUPW:09:39:40.325: 1: DBMS_METADATA.SET_COUNT called.
    KUPW:09:39:40.328: 1: DBMS_METADATA.SET_COUNT returned.
    KUPW:09:39:40.328: 1: DBMS_METADATA.FETCH_XML_CLOB called.
    *** 2011-09-20 09:39:42.603
    KUPW:09:39:42.603: 1: DBMS_METADATA.FETCH_XML_CLOB returned.
    KUPW:09:39:42.603: 1: In procedure CREATE_TABLE_DATA_OBJECT_ROWS
    KUPW:09:39:42.603: 1: In function GATHER_PARSE_ITEMS
    KUPW:09:39:42.603: 1: In function CHECK_FOR_REMAP_NETWORK
    KUPW:09:39:42.603: 1: Nothing to remap
    KUPW:09:39:42.603: 1: In procedure BUILD_OBJECT_STRINGS
    KUPW:09:39:42.604: 1: In procedure LOCATE_DATA_FILTERS
    KUPW:09:39:42.604: 1: In function NEXT_PO_NUMBER
    KUPW:09:39:42.620: 1: In procedure DETERMINE_METHOD_PARALLEL
    KUPW:09:39:42.620: 1: flags mask: 0
    KUPW:09:39:42.620: 1: dapi_possible_meth: 1
    KUPW:09:39:42.620: 1: data_size: 3019898880
    KUPW:09:39:42.620: 1: et_parallel: TRUE
    KUPW:09:39:42.620: 1: object: TABLE_DATA:"TIA"."ACC_PAYMENT_SPECIFICATION"                               <<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<
    KUPW:09:39:42.648: 1: l_dapi_bit_mask: 7
    KUPW:09:39:42.648: 1: l_client_bit_mask: 7
    KUPW:09:39:42.648: 1: TABLE_DATA:"TIA"."ACC_PAYMENT_SPECIFICATION" either, parallel: 12               <<<<< Here is says either (I thought that was method ?)  <<<<<<<<<<<<<<<<
    KUPW:09:39:42.648: 1: FORALL BULK INSERT called.
    KUPW:09:39:42.658: 1: FORALL BULK INSERT returned.
    KUPW:09:39:42.660: 1: DBMS_LOB.TRIM called. v_md_xml_clob
    KUPW:09:39:42.660: 1: DBMS_LOB.TRIM returned.
    KUPW:09:39:42.660: 1: DBMS_METADATA.FETCH_XML_CLOB called.
    KUPW:09:39:42.678: 1: DBMS_METADATA.FETCH_XML_CLOB returned.
    KUPW:09:39:42.678: 1: DBMS_LOB.CREATETEMPORARY called.
    KUPW:09:39:42.678: 1: DBMS_LOB.CREATETEMPORARY returned.
    KUPW:09:39:42.678: 1: In procedure UPDATE_TD_ROW_EXP with seqno: 62
    KUPW:09:39:42.680: 1: 1 rows fetched
    KUPW:09:39:42.680: 1: In function NEXT_PO_NUMBER
    KUPW:09:39:42.680: 1: Next table data array entry: 1 Parallel: 12 Size: 3019898880 Method: 4Creation_level: 0               <<<<<<<<<<<<<<<<  HERE IT SAYS METHOD = 4  and PARALLEL=12 (I'm not using the parallel parameter ???)  <<<<<<<<<<<<<<<<<<
    KUPW:09:39:42.681: 1: In procedure UPDATE_TD_BASE_PO_INFO
    KUPW:09:39:42.683: 1: Updated 1 td objects with bpo between 1 and 1
    KUPW:09:39:42.684: 1: Send table_data_varray called.  Count: 1
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  7
    kwqberlst ascn -90145310 lascn 22
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    KUPW:09:39:42.695: 1: Send table_data_varray returned.
    KUPW:09:39:42.695: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:39:42.695: 1: In procedure UPDATE_TYPE_COMPLETION_ROW
    KUPW:09:39:42.695: 1: Old Seqno: 62 New Path:  PO Num: -5 New Seqno: 0
    KUPW:09:39:42.695: 1: Object count: 1
    KUPW:09:39:42.697: 1: 1 completed for 62
    KUPW:09:39:42.697: 1: DBMS_METADATA.CLOSE called. Handle: 100001
    KUPW:09:39:42.697: 1: DBMS_METADATA.CLOSE returned.
    KUPW:09:39:42.697: 1: In procedure CREATE_MSG
    KUPW:09:39:42.697: 1: KUPV$FT.MESSAGE_TEXT called.
    KUPW:09:39:42.698: 1: KUPV$FT.MESSAGE_TEXT returned.
    KUPW:09:39:42.698: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:39:42.698: 1: KUPC$QUEUE_INT.SEND called.
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  7
    kwqberlst ascn -90145310 lascn 22
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    KUPW:09:39:42.699: 1: KUPC$QUEUE_INT.SEND returned.
    KUPW:09:39:42.699: 1: Total estimation using BLOCKS method: 2.812 GB
    KUPW:09:39:42.699: 1: In procedure CONFIGURE_METADATA_UNLOAD
    KUPW:09:39:42.699: 1: Phase: WORK_PHASE Filter Name: BEGIN_WITH Filter Value:
    KUPW:09:39:42.699: 1: DBMS_METADATA.OPEN11.2.0.0.0 called.
    KUPW:09:39:42.837: 1: DBMS_METADATA.OPEN11.2.0.0.0 returned. Source handle: 200001
    KUPW:09:39:42.837: 1: DBMS_METADATA.SET_FILTER called. metadata_phase: WORK_PHASE
    KUPW:09:39:42.837: 1: DBMS_METADATA.SET_FILTER returned. In function GET_NOEXP_TABLE
    KUPW:09:39:42.847: 1: DBMS_METADATA.SET_PARSE_ITEM called.
    KUPW:09:39:42.964: 1: DBMS_METADATA.SET_PARSE_ITEM returned.
    KUPW:09:39:42.964: 1: DBMS_METADATA.SET_COUNT called.
    KUPW:09:39:42.967: 1: DBMS_METADATA.SET_COUNT returned.
    KUPW:09:39:42.967: 1: KUPF$FILE.OPEN_CONTEXT called.
    KUPW:09:39:42.967: 1: KUPF$FILE.OPEN_CONTEXT returned.
    KUPW:09:39:42.968: 1: DBMS_METADATA.FETCH_XML_CLOB called. Handle: 200001
    *** 2011-09-20 09:40:01.798
    KUPW:09:40:01.798: 1: DBMS_METADATA.FETCH_XML_CLOB returned.
    KUPW:09:40:01.798: 1: Object seqno fetched:
    KUPW:09:40:01.799: 1: Object path fetched:
    KUPW:09:40:01.799: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:40:01.799: 1: In procedure COMPLETE_EXP_OBJECT
    KUPW:09:40:01.799: 1: KUPF$FILE.FLUSH_LOB called.
    KUPW:09:40:01.815: 1: KUPF$FILE.FLUSH_LOB returned.
    KUPW:09:40:01.815: 1: In procedure UPDATE_TYPE_COMPLETION_ROW
    KUPW:09:40:01.815: 1: Old Seqno: 226 New Path:  PO Num: -5 New Seqno: 0
    KUPW:09:40:01.815: 1: Object count: 1
    KUPW:09:40:01.815: 1: 1 completed for 226
    KUPW:09:40:01.815: 1: DBMS_METADATA.CLOSE called.  Handle: 200001
    KUPW:09:40:01.816: 1: DBMS_METADATA.CLOSE returned.
    KUPW:09:40:01.816: 1: KUPF$FILE.CLOSE_CONTEXT called.
    KUPW:09:40:01.820: 1: KUPF$FILE.CLOSE_CONTEXT returned.
    KUPW:09:40:01.821: 1: In procedure SEND_MSG. Fatal=0
    KUPW:09:40:01.821: 1: KUPC$QUEUE.TRANSCEIVE called.
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  7
    kwqberlst ascn -90145310 lascn 22
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  7
    KUPW:09:40:01.827: 1: KUPC$QUEUE.TRANSCEIVE returned. Received 2012
    KUPW:09:40:01.827: 1: DBMS_LOB.CREATETEMPORARY called.
    KUPW:09:40:01.828: 1: DBMS_LOB.CREATETEMPORARY returned.
    KUPW:09:40:01.828: 1: Process order range: 1..1
    KUPW:09:40:01.828: 1: Method: 1
    KUPW:09:40:01.828: 1: Parallel: 1
    KUPW:09:40:01.828: 1: Creation level: 0
    KUPW:09:40:01.830: 1: BULK COLLECT called.
    KUPW:09:40:01.830: 1: BULK COLLECT returned.
    KUPW:09:40:01.830: 1: In procedure BUILD_OBJECT_STRINGS
    KUPW:09:40:01.836: 1: In procedure MOVE_DATA UNLOADing process_order 1 TABLE_DATA:"TIA"."ACC_PAYMENT_SPECIFICATION"            <<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<
    KUPW:09:40:01.839: 1: KUPD$DATA.OPEN called.
    KUPW:09:40:01.840: 1: KUPD$DATA.OPEN returned.
    KUPW:09:40:01.840: 1: KUPD$DATA.SET_PARAMETER - common called.
    KUPW:09:40:01.843: 1: KUPD$DATA.SET_PARAMETER - common returned.
    KUPW:09:40:01.843: 1: KUPD$DATA.SET_PARAMETER - flags called.
    KUPW:09:40:01.843: 1: KUPD$DATA.SET_PARAMETER - flags returned.
    KUPW:09:40:01.843: 1: KUPD$DATA.START_JOB called.
    KUPW:09:40:01.918: 1: KUPD$DATA.START_JOB returned. In procedure GET_JOB_VERSIONThis is how I called expdp:
    expdp system/xxxxxxxxx schemas=tia directory=expdp INCLUDE=TABLE:\" =\'ACC_PAYMENT_SPECIFICATION\'\" REUSE_DUMPFILES=Y LOGFILE=expdp:$LOGFILE TRACE=400300

    Hi there ...
    I have read the note - thats where I found the link to the trace note 286496.1 - on now to setup a trace
    But I still need an explanation for the methods (1,2,4 etc)
    regards
    Mette

  • Error using expdp on Oracle XE

    Hi,
    When I try and do an export in Oracle XE I get the following error:
    D:\>expdp SYSTEM/vodafone SCHEMAS=XE_DATA DIRECTORY=dmpdir DUMPFILE=xedata.dmp
    Export: Release 10.2.0.1.0 - Production on Thursday, 29 June, 2006 10:47:10
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Producti
    n
    ORA-39002: invalid operation
    ORA-39070: Unable to open the log file.
    ORA-29283: invalid file operation
    ORA-06512: at "SYS.UTL_FILE", line 475
    ORA-29283: invalid file operation
    Any ideas??
    /John

    Hi guys,
    The problem was that Oracle didn't have access to the directory I created. So I created a directory in the Oracle home directory and it worked..
    Thanks,
    John

  • Error while doing an expdp on a large datafile

    Hello,
    I tried an export using expdp in oracle 10g express edition. It was working perfectly until when the db size reached 2.1 gb. I got the following error message:
    ---------------- Start of error message ----------------
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    Starting "SERVICE_2_8"."SYS_EXPORT_SCHEMA_05": service_2_8/******** LOGFILE=3_export.log DIRECTORY=db_pump DUMPFILE=service_2_8.dmp CONTENT=all
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    ORA-39125: Worker unexpected fatal error in KUPW$WORKER.GET_TABLE_DATA_OBJECTS while calling DBMS_METADATA.FETCH_XML_CLOB []
    ORA-01116: error in opening database file 5
    ORA-01110: data file 5: '/usr/lib/oracle/xe/oradata/service_3_0.dbf'
    ORA-27041: unable to open file
    Linux Error: 13: Permission denied
    Additional information: 3
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPW$WORKER", line 6235
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    0x3b3ce18c 14916 package body SYS.KUPW$WORKER
    0x3b3ce18c 6300 package body SYS.KUPW$WORKER
    0x3b3ce18c 9120 package body SYS.KUPW$WORKER
    0x3b3ce18c 1880 package body SYS.KUPW$WORKER
    0x3b3ce18c 6861 package body SYS.KUPW$WORKER
    0x3b3ce18c 1262 package body SYS.KUPW$WORKER
    0x3b0f9758 2 anonymous block
    Job "SERVICE_2_8"."SYS_EXPORT_SCHEMA_05" stopped due to fatal error at 03:04:34
    ---------------- End of error message ----------------
    Selinux was disabled completely and I have put the permission of 0777 to the appropriate datafile.
    Still, it is not working.
    Can you please tell me how to solve this problem or do you have any ideas or suggestions regarding this?

    Hello rgeier,
    I cannot access this tablespace which is service_3_0 (2.1 gb) through a php web-application or through sqlplus. I can access a small tablespace which is service_2_8 through the web-application or through sqlplus. When I tried to access service_3_0 through sqlplus, the following error message was returned:
    ---------------- Start of error message ----------------
    ERROR at line 1:
    ORA-01116: error in opening database file 5
    ORA-01110: data file 5: '/usr/lib/oracle/xe/oradata/service_3_0.dbf'
    ORA-27041: unable to open file
    Linux Error: 13: Permission denied
    Additional information: 3
    ---------------- End of error message ----------------
    The following are the last eset of entries in the alert_XE.log file in the bdump folder:
    ---------------- Start of alert log ----------------
    db_recovery_file_dest_size of 40960 MB is 9.96% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    Wed Aug 20 05:13:59 2008
    Completed: alter database open
    Wed Aug 20 05:19:58 2008
    Shutting down archive processes
    Wed Aug 20 05:20:03 2008
    ARCH shutting down
    ARC2: Archival stopped
    The value (30) of MAXTRANS parameter ignored.
    kupprdp: master process DM00 started with pid=27, OS id=7463 to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_SCHEMA_06', 'SERVICE_2_8', 'KUPC$C_1_20080820054031', 'KUPC$S_1_20080820054031', 0);
    kupprdp: worker process DW01 started with worker id=1, pid=28, OS id=7466 to execute - SYS.KUPW$WORKER.MAIN('SYS_EXPORT_SCHEMA_06', 'SERVICE_2_8');Wed Aug 20 05:40:48 2008
    The value (30) of MAXTRANS parameter ignored.
    The value (30) of MAXTRANS parameter ignored.
    The value (30) of MAXTRANS parameter ignored.
    The value (30) of MAXTRANS parameter ignored.
    The value (30) of MAXTRANS parameter ignored.
    ---------------- End of alert log ----------------

  • Date-wise backup using expdp

    Dear Friends ,
    I am using Oracle database 10g . Every week I take full export backup using 'expdp' . Beetween the days of the week, I dont take any backup . Now I need backup of two days ago . In this moment , is it possible to take backup using datapump (expdp) before two days ago ?
    i.e., Is there any facility in oracle 10g datapump to take export backup using date-wise ?
    My another question ,
    Is there any incremental backup procedure available in oracle 10g datapump ?
    Waiting for kind reply ... ...

    Hi,
    Dear Friends ,
    I am using Oracle database 10g . Every week I take full export backup using 'expdp' . Beetween the days of the week, I dont take any backup . Now I need backup of two days ago . In this moment , is it possible to take backup using datapump (expdp) before two days ago ?
    i.e., Is there any facility in oracle 10g datapump to take export backup using date-wise ?
    My another question ,
    Is there any incremental backup procedure available in oracle 10g datapump ?
    Waiting for kind reply ... ...You must not use or treat Expdp or logical backup as the Backup which can be used for future purpose. Please search forums their are so many links you lan find out and so many people told many time.. even ORACLE suggest that.
    Go for Custom backups and RMAn backups.
    - Pavan Kumar N
    - Pavan Kumar N

  • Expdp failed job disappear

    Hi all,
    I am testing the expdp on attach job in 10.2R3
    1. expdp job and ^C to stop it
    C:\TEMP>expdp system/oracle directory=testdir dumpfile=test5.dmp logfile=test5.log full=y job_name=expjob
    . . exported "SCOTT"."DEPT" 5.656 KB 4 rows
    . . exported "SCOTT"."EMP" 7.820 KB 14 rows
    . . exported "SCOTT"."SALGRADE" 5.585 KB 5 rows
    . . exported "SYSMAN"."AQ$_MGMT_NOTIFY_QTABLE_S" 7.734 KB 1 rows
    ^C
    2. in sqlplus check job
    select * from dba_datapump_jobs;
    OWNER_NAME JOB_NAME OPERATION JOB_MODE
    STATE DEGREE ATTACHED_SESSIONS DATAPUMP_SESSIONS
    SYSTEM EXPJOB EXPORT FULL
    EXECUTING 1 1 3
    3. then after 1-2 mins, check the job status again, it disappears????
    select * from dba_datapump_jobs;
    no rows selected
    Please advise what is happening.....
    thanks
    andrew

    ^C doesn't stop the job it should take you to a prompt that looks like
    Export>
    then you can stop the job using
    Export>STOP_JOB
    the job is finished that's why your not seeing it when you do a
    select * from dba_datapump_jobs

  • Check the status expdp datapump job

    All,
    I have started expdp (datapump) on size of 250GB schema.I want to know when the job will be completed.
    I tried to find from views dba_datapump_sessions and v$session_longops,v$session.But but all in vain.
    Is it possible to find completion time of expdp job?
    Your help is really appreciated.

    Hi,
    Have you started the Job in "Interactie mode".. ??
    If yes then you can get the status of Exectuion with "STATUS"- defauly is zero (If you setting to not null value then based on that it will show the status - seconds)
    Second, thing check "dba_datapump_jobs"
    - Pavan Kumar N

  • How to schedule a expdp job on dbconsole 10g

    Hi all,
    is there any way to create a schedule expdp job to execute automatically all days on enterprise mananger 10g (dbconsole) ???
    Thanks
    Wander (Brazil)

    Hi,
    on dbconsol i get to schedule a job only to execute immediately or posteriorly, just one time, but I want that it run all day in a specific time.
    Wander(Brazil)

Maybe you are looking for

  • Problem in Restoring the database

    Hi, I want to restore the database RMAN> startup force mount; RMAN-00571: =========================================================== RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS =============== RMAN-00571: ================================

  • Can i use C Code at Compact Vision or Compact Rio?

    hi, i have a very important question!! can i use the C code that the Vision Assistant generates, in the compact vision system, even in DLL form??? if yes, how can i download it to the compact vison? can i use C code at the compact rio? greets, gerald

  • Linkage for Template Monster Fla Files

    Alright I have looked about everywhere, except the right place. I know 0 about flash. Just got Flash CS3. Everything I find is for older versions of flash including the info on template monster. I need to know how to assign links for the buttons in t

  • Cisco Network Assistant errors

    Hi. I'm hoping someone can help me understand some of the errors I've been receiving with CNA version 5.8.8.1. It seems to be working fine with the 2960 Catalyst switch. But I see this  Java error "The launching request for Monitor > Event Notificati

  • Java ME emulator crashes

    I recently installed Fedora 9 with openjdk 6. I also installed the Netbeans 6.1 mobility pack. Initially I was able to do my development without any hassles. But after a short while, I haven't been able to run the emulator. Whenever I try to run any