Logical export backup

LQ = > I exported hr schema with three different option of keyword
CONTENT = ALL,
CONTENT = METADATA_ONLY
AND CONTENT = NONE
1.with keyword CONTENTS= ALL
the size of dmp file was 416KB
2. with keyword CONTENTS=DATA_ONLY
I got size = 216
3.Last with keyword CONTENTS=METADATA_ONLY
Igot size = 316
here , My question is that with the keyword = ALL and keyword DATA_ONLY I have clearly understood
that with "ALL" the dumpfile contents metadata + data of schema such as table etc so size was 416KB.
with data_only the dumpfile only contains object's data so size was 216 KB
but , why with the option of METADATA_ONLY the size was 316 KB..As in metadata it should be
416 - 216 = 200 KB ( ALL - DATA_ONLY = METADATA_ONLY)..This is my assumption..If am wrong that
kindly show me the right concept...
And One more thing I want to say that if there is any internal things is happening please let me know..hope I will be
receiving the answer and the concept behind the space that was used by the schemas with various option...

Hi,
can you please provide me the hint..that how can I find out the size of master table etc..The size of the master table is going to be different for every job. It is a table that Data Pump creates and inserts data into as the export and import jobs are running. If you want to know the final size, then add
keep_master=y
on your expdp or impdp command. Then what I did, was to export just that table and see how big it is.
expdp system/manager schemas=hr ... keep_master=y job_name=foo
expdp system/manager tables=system.foo ...
And one more concept I am eagerly waiting to know from you that during datapump when I used to export the whole database >, there I found the warning / error
ORA-39139: Data Pump does not support XMLSchema objects. TABLE_DATA:"OE"."PURCHASEORDER" will be skipped.
I think that XML schema is not supported by our newly feature of export( datapump)
I have done this by giving the command
expdp directory=test_dir logfile=full_database.log dumpfile=full_database.dmp full=y
Note: but , when the same database was exported through the old feature (exp) , the database full was exported successfully >without any error...
can you please throw some light in this regard...You don't say what version you are on, but I believe that support for xml types was not in 10.1 Data Pump but was in old exp/imp. I also think support for this was added in a later version for Data Pump, but I can't recall the exact versions. If you have documentation, you should look at that.
Dean

Similar Messages

  • Apply RMAN archivelogs on export backups

    Hello Everyone
    We are using the below script for RMAN
    run
    allocate channel c1 type 'sbt_tape';
    backup database plus archivelog;
    release channel c1;
    So whether we shall be able to apply archivelogs on export backups ? If yes how?
    Thanks.
    Achyot

    It's impossible to apply archived redo logs (physical backup) after importing a export file (logical backup).
    You can only apply archived redo logs to a physical database backup or to a physical or logical standby database (Data Guard feature).
    See logical and physical backup definitions:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14220/glossary.htm#sthref4165
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14220/glossary.htm#i997232

  • Export Backup Size is different

    Hi Gurus,
    We are using Oracle 10G (10.2.0.1.0) on Solaris 10 and we used to take export backup daily basis. While taking export backup using 'exp' then .dmp file size is becoming 30.3 GB where as when we use 'expdp' .dmp file size is 26.1 GB. I've checked both the log files and found no. of tables along with their records are same. so I'm confused why this size difference? and as a result we are not in a position to implement 'expdp'. Can anybody tell me why this size difference and whether can we rely on expdp or not?

    user606947 wrote:
    Hi Gurus,
    We are using Oracle 10G (10.2.0.1.0) on Solaris 10 and we used to take export backup daily basis. While taking export backup using 'exp' then .dmp file size is becoming 30.3 GB where as when we use 'expdp' .dmp file size is 26.1 GB. I've checked both the log files and found no. of tables along with their records are same. so I'm confused why this size difference? and as a result we are not in a position to implement 'expdp'. Can anybody tell me why this size difference and whether can we rely on expdp or not?
    Oracle recommends that you use the new Data Pump Export and Import utilities because they support all Oracle Database 10g features
    Look at the following documentation:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm#sthref14

  • Daily export backup

    Hi,
    I have production & DR database of 350gb version 10.2.0.4 on aix 5.3 platform.
    In morning time after completion of daily batch, we are taking export backup (by stopping application services till application is not available for users) by using datapump utility which is taking 2 to 2.30 hrs to completes.
    After that we have starting the application services and then only application available for users.
    This export backup dump we are daily using for restoration on another development server.
    Kindly help me, is there any option to avoid export backup time so that it will not affect buisness time.
    and also i can do daily restoration on development server.
    Please help.
    Regards,

    In morning time after completion of daily batch, we are taking export backup (by stopping application services till application is not available for users) by using datapump utility which is taking 2 to 2.30 hrs to completes.
    After that we have starting the application services and then only application available for users.
    This export backup dump we are daily using for restoration on another development server.
    Kindly help me, is there any option to avoid export backup time so that it will not affect buisness time.
    and also i can do daily restoration on development server.if you want complete database refresh?
    you can use STREAMS.
    If certain schema/objects you can use MViews.

  • Export backup error while using where clause

    I am heaving Oracle9i on solaris platform. Whem i am trying to take the export backup of a table it is giving below error:
    exp swtiob tables=NDC_ATMPROOF_HIST file=NDC_ATMPROOF_HIST.dmp log=NDC_ATMPROOF_HIST.log query="where PROOF_DATE >= '01-july-2010'" statistics=none
    LRM-00112: multiple values not allowed for parameter 'query'
    EXP-00019: failed to process parameters, type 'EXP HELP=Y' for help
    EXP-00000: Export terminated unsuccessfully

    You need to escape the stuff like this:
    $ exp scott/tiger tables=emp file=emp.dmp log=emp.log query=\"where HIREDATE\>\'09-JUN-1981\'\"
    Export: Release 9.2.0.8.0 - Production on Wed Jul 7 12:54:48 2010
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.8.0 - Production
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P1 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                            EMP          6 rows exported
    EXP-00091: Exporting questionable statistics.
    Export terminated successfully with warnings.
    $

  • How to ZIP Oracle Datapump export backup file

    Hello All,
    My customer is asking to give him the production data dump to the following path \\138.90.17.56\OMNISAFE.
    I really don't understand his requirement and he also wants me to zip the export backup file. How do I do that, Do you know any unix command to zip backup files.
    thanks and regards
    cherry

    1013498 wrote:
    Well Thanks for your reply.....my oracle version is 11.2.0.3.b and if we have the compression option can you please elaborate how to do that......
    It's in the documentation.  See Data Pump Export
    let us say my expdp file is abc.dmp...should I give the command gzip abc.dmp or any different.
    Let me google that for you
    One more question what does teh customer mean by production data dump to the following path \\138.90.17.56\OMNISAFE. please explain
    How do we know what the customer means?  Why don't you ask him?
    That said, it looks like a url to an ip address and a defined folder at that ip address.  Again, if the customer wants you to send them a file, you need to be working with said customer on the mechanics of accessing their system.
    All that said ....
    Learning how to look things up in the documentation is time well spent investing in your career.  To that end, you should drop everything else you are doing and do the following:
    Go to tahiti.oracle.com.
    Locate the link for your Oracle product and version, and click on it.
    You are now at the entire documentation set for your selected Oracle product and version.
    BOOKMARK THAT LOCATION
    Spend a few minutes just getting familiar with what is available here. Take special note of the "books" and "search" tabs. Under the "books" tab (for 10.x) or the "Master Book List" link (for 11.x) you will find the complete documentation library.
    Spend a few minutes just getting familiar with what kind  of documentation is available there by simply browsing the titles under the "Books" tab.
    Open the Reference Manual and spend a few minutes looking through the table of contents to get familiar with what kind of information is available there.
    Do the same with the SQL Reference Manual.
    Do the same with the Utilities manual.
    You don't have to read the above in depth.  They are reference manuals.  Just get familiar with what is there to be referenced. Ninety percent of the questions asked on this forum can be answered in less than 5 minutes by simply searching one of the above manuals.
    Then set yourself a plan to dig deeper.
    - *Read a chapter a day from the Concepts Manual*.
    - Take a look in your alert log.  One of the first things listed at startup is the initialization parms with non-default values. Read up on each one of them (listed in your alert log) in the Reference Manual.
    - Take a look at your listener.ora, tnsnames.ora, and sqlnet.ora files. Go to the Network Administrators manual and read up on everything you see in those files.
    - *When you have finished reading the Concepts Manual, do it again*.
    Give a man a fish and he eats for a day. Teach a man to fish and he eats for a lifetime.

  • Can't take export backup

    Hello,
    I can't take export backup as it ends up with a message column name invalid.How to solve the problem. My database is on UNIXWARE 7.1.1 & Oracle8i.
    D ROY
    SHILLONG- INDIA

    content of the log file:
    Connected to: Oracle8i Enterprise Edition Release 8.1.7.4.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.4.0 - Production
    Export done in US7ASCII character set and US7ASCII NCHAR character set
    About to export the entire database ...
    . exporting tablespace definitions
    . exporting profiles
    . exporting user definitions
    . exporting roles
    . exporting resource costs
    . exporting rollback segment definitions
    . exporting database links
    . exporting sequence numbers
    EXP-00008: ORACLE error 904 encountered
    ORA-00904: invalid column name
    EXP-00000: Export terminated unsuccessfully
    Last line of the dmp files
    CONNECT SYSTEM
    EXECFULL DBMS_DISTRIBUTED_TRUST_ADMIN.allow_all;
    CREATE SEQUENCE "AQ$_QIDSEQ" MINVALUE 1 MAXVALUE 999999999999999999999999999 INCREMENT BY 1 START WITH 21 CACHE 20 NOORDER NOCYCLE
    ___________________________________

  • Export backup failing with ORA-1555snapshot too old: rollback segment ERROR

    DB version =8.1.7.4
    OS=Solaris =5.10.
    I know this is outdated. But still we are supporting this version of database.
    We are facing ORA-1555 error daily when we are running the export backup of database. Consistent parameter is set to N .
    I needed to know why export uses rollback segments during export operation. even though CONSISTENT parameter is N.
    Thanks,
    Edited by: ahshivap on Dec 12, 2012 5:19 AM

    Hi,
    Consistent=y just mean the tables are consistent with each other. Oracle still has to guarantee that each individual table is at least consistent with itself. That's your issue.
    Regards,
    Harry

  • Export backups failing - Need advice

    Hello All,
    We have a 10.2.0.3 database for which we have to take export backups everyday in maintainence window, when only batch jobs and other non essentials SQL's are running on the database.
    The export command we have used in our script is as under:
    $ORACLE_HOME/bin/exp / file=$ORACLE_DUMP_FILE compress=y full=y direct=n buffer=20480000 consistent=y statistics=none log=$ORACLE_LOGAtleast 3-4 times in a week, the export backup us failing with ORA-01555 errors like the one below:
    EXP-00008: ORACLE error 1555 encountered
    ORA-01555: snapshot too old: rollback segment number 10 with name "_SYSSMU10$" too small
    EXP-00000: Export terminated unsuccessfullyMy DB is configured as:
    SQL> show parameter undo_
    NAME                                 TYPE        VALUE
    undo_management                      string      AUTO
    undo_retention                       integer     54000
    undo_tablespace                      string      UNDOTBSAlso:
    SQL> select max(tuned_undoretention) from v$undostat;
    MAX(TUNED_UNDORETENTION)
                       75598The Max value is reached somewhere in the evening, but our job runs at
    SQL> select begin_time from v$undostat where TUNED_UNDORETENTION = 75598;
    BEGIN_TIME
    29/05/2012 17:39:46At the time (as per my log file) when my export job failed, the above value was:
    SQL> select TUNED_UNDORETENTION from v$undostat where BEGIN_TIME like '29/05/2012 16:2%';
    TUNED_UNDORETENTION
                  71318Please advice how the above issue can be addressed/resolved.
    For further information, please let me know. Thank you.
    Suddhasatwa
    India.
    Edited by: SBhaumik_DBA on May 30, 2012 9:50 PM

    Hello. Thanks for the inputs on CONSISTENT option. There are pre-laid change mgmt processes in every organisation which needs to be maintained/taken in account before such changes are made, and that too, with client approvals, etc. Since it was not a feasible option to get EXPDP or RMAN implemented at this point, so was this question raised.
    There is a small issue, however, if we use CONSISTENT=N. Below is an extract from asktom.oracle.com:
    >
    imagine if you started an export at 9am of the EMP and DEPT tables.
    EMP started exporting at 9am and DEPT at 9:15am.
    Now, the EMP data would be "as of 9am", but the DEPT data would be as of 9:15am. What if
    you import that data now -- will it work? maybe, maybe not -- perhaps at 9:10am, someone
    fired the last employee in department 50 and deleted department 50. Your EMP export
    would have them in this deptno, your DEPT export would not have this deptno. The import
    would fail.
    If you use consistent=y, Oracle will export all of the data "as of 9am", so deptno=50
    will be included in all tables and the import will succeed.
    >
    That's one of the risks involved.
    Is there any other solution we can have in place for this issue? Thanks.

  • EXPORT Backup failing due to character set problem

    Hi ,
    The export Backup failing due to character set problem
    . . exporting table ravidlx
    EXP-00008: ORACLE error 6552 encountered
    ORA-06552: PL/SQL: Compilation unit analysis terminated
    ORA-06553: PLS-553: character set name is not recognized
    P
    Please suggest how to set character set
    Regards,
    kk
    Edited by: kk001 on Aug 29, 2011 7:22 PM

    kk001 wrote:
    Hi ,
    The export Backup failing due to character set problem
    . . exporting table ravidlx
    EXP-00008: ORACLE error 6552 encountered
    ORA-06552: PL/SQL: Compilation unit analysis terminated
    ORA-06553: PLS-553: character set name is not recognized
    P
    Please suggest how to set character set
    I don't know what you have.
    I don't know what you do.
    I don't know what you see.
    It is really, Really, REALLY difficult to fix a problem that can not be seen.
    use COPY & PASTE so we can see what you do & how Oracle responds.
    do as below so we can know complete Oracle version & OS name.
    Post via COPY & PASTE complete results of
    SELECT * from v$version;

  • Export backup terminated due to ORA-00600: internal error code, arguments:

    Hi,
    exp system/system full=y file=exp_bkp.dmp log=exp_bkp.log statistics=none
    My export backup terminated unsuccessfully and database also down due to ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    os version=RHEL4
    DB Version=10.2.0.1.0
    export log
    exp_bkp.log
    About to export the entire database ...
    . exporting tablespace definitions
    . exporting profiles
    . exporting user definitions
    . exporting roles
    . exporting resource costs
    . exporting rollback segment definitions
    . exporting database links
    . exporting sequence numbers
    . exporting directory aliases
    . exporting context namespaces
    . exporting foreign function library names
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions
    . exporting system procedural objects and actions
    . exporting pre-schema procedural objects and actions
    . exporting cluster definitions
    . about to export SYSTEM's tables via Conventional Path ...
    . . exporting table DEF$_AQCALL 0 rows exported
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Alert Log
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:43:11 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:43:12 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:43:14 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:44:14 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_pmon_3648.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:44:16 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_pmon_3648.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:44:16 2009
    PMON: terminating instance due to error 472
    Mon Jun 8 13:44:16 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_psp0_3650.trc:
    ORA-00472: PMON process terminated with error
    Instance terminated by PMON, pid = 3648

    Hi,
    Trace files
    */u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_pmon_3648.trc*
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u02/app/oracle/product/10.2.0/db_1
    System name:     Linux
    Node name:     test1.domain.com
    Release:     2.6.9-5.EL
    Version:     #1 Wed Jan 5 19:22:18 EST 2005
    Machine:     i686
    Instance name: PRIM
    Redo thread mounted by this instance: 1
    Oracle process number: 2
    Unix process pid: 3648, image: [email protected] (PMON)
    *** 2009-06-08 13:44:14.697
    *** SERVICE NAME:(SYS$BACKGROUND) 2009-06-08 13:44:14.695
    *** SESSION ID:(170.1) 2009-06-08 13:44:14.695
    ********** Internal heap ERROR 17112 addr=0x4152199c *********
    ***** Dump of memory around addr 0x4152199c:
    41520990 00001001 [....]
    415209A0 4151F99C 2001CFBC 00000000 415219A8 [..QA... ......RA]
    415209B0 D0000FED 00000000 415219B8 40EFEB10 [..........RA...@]
    415209C0 415219BC 00000025 00000000 0C7470F0 [..RA%........pt.]
    415209D0 00000000 46440007 454C544C 0000004E [......DFLTLEN...]
    415209E0 000A000C 0E000001 00000015 415209C4 [..............RA]
    415209F0 0C7470D8 41520A70 41520A1C 00000015 [.pt.p.RA..RA....]
    41520A00 415209E8 0C7470D8 41520A84 41520A1C [..RA.pt...RA..RA]
    41520A10 00000055 415209FC 0C747378 00000201 [U.....RAxst.....]
    41520A20 00000000 00000059 00008100 00030040 [....Y.......@...]
    41520A30 00000000 00000016 41548140 00000000 [[email protected]....]
    41520A40 04000008 00000000 40FFBEFC 00000000 [...........@....]
    41520A50 00000000 40F34AF0 40F56490 00000000 [[email protected].@....]
    */u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3677.trc*
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u02/app/oracle/product/10.2.0/db_1
    System name:     Linux
    Node name:     test1.domain.com
    Release:     2.6.9-5.EL
    Version:     #1 Wed Jan 5 19:22:18 EST 2005
    Machine:     i686
    Instance name: PRIM
    Redo thread mounted by this instance: 1
    Oracle process number: 15
    Unix process pid: 3677, image: [email protected] (TNS V1-V3)
    *** SERVICE NAME:() 2009-06-08 13:41:55.582
    *** SESSION ID:(159.3) 2009-06-08 13:41:55.582
    Thread 1 checkpoint: logseq 183, block 2, scn 473582
    cache-low rba: logseq 183, block 3
    on-disk rba: logseq 183, block 101, scn 473671
    start recovery at logseq 183, block 3, scn 0
    ----- Redo read statistics for thread 1 -----
    Read rate (ASYNC): 49Kb in 0.55s => 0.09 Mb/sec
    Total physical reads: 4096Kb
    Longest record: 0Kb, moves: 0/152 (0%)
    Longest LWN: 6Kb, moves: 0/24 (0%), moved: 0Mb
    Last redo scn: 0x0000.00073a46 (473670)
    ----- Recovery Hash Table Statistics ---------
    Hash table buckets = 32768
    Longest hash chain = 1
    Average hash chain = 41/41 = 1.0
    Max compares per lookup = 1
    Avg compares per lookup = 223/264 = 0.8
    *** 2009-06-08 13:41:56.180
    KCRA: start recovery claims for 41 data blocks
    *** 2009-06-08 13:41:57.051
    KCRA: blocks processed = 41/41, claimed = 41, eliminated = 0
    *** 2009-06-08 13:41:57.053
    Recovery of Online Redo Log: Thread 1 Group 1 Seq 183 Reading mem 0
    ----- Recovery Hash Table Statistics ---------
    Hash table buckets = 32768
    Longest hash chain = 1
    Average hash chain = 41/41 = 1.0
    Max compares per lookup = 1
    Avg compares per lookup = 264/264 = 1.0
    tkcrrsarc: (WARN) Failed to find ARCH for message (message:0x1)
    tkcrrpa: (WARN) Failed initial attempt to send ARCH message (message:0x1)

  • Export / Backup all the production data

    Dear Experts,
    Thanks for your attention!
    Due to we have some upgrading issue in our production system. We plan to re-install the GRC AC SP10 in PRD.
    We have go-live for half a year, there are around 500 requests in PRD, all these transaction data need to be export / backup before re-installation.
    Could any one give us any advice?
    Thanks a lot!

    Hi,
    This will be handled by BASIS team, you need not worry. Once the data was restored you need to amke sure all the production data was avilable with configuration settings, audit trial, all the requests data.
    If this was first time you are performing, better check out restore on non-production server and identify any issues and resolve and then proceed onto production.
    Backup and restore will be done only by basis and documents avilable in market place, also that was dependent on your DB type.
    regards.

  • Export backup degrades performance?

    Hi all,
    Please suggest me if i take export backup of the database does it degrades performance of the database when the applications are connected?
    Can this export be done while application is using the database?
    Is there any risk concerning performance of the application during DB export?
    Thank you.

    An export is a query against the database. Generally, a query to extract the whole schema or whole table (or whole database). It has the same performance impact as any query doing FullTableScans against the tables being exported.
    Note :If you do an export using 'exp', ensure that you use "CONSISTENT=Y" or "FLASHBACK_SCN" or "FLASBACK_TIME" to get consistency across tables, else exp does Table-level consistency.
    With 'expdp' you can use FLASHBACK_SCN or FLASHBACK_TIME

  • Export/Backup Thesaurus

    Hi,
    Does anybody know how to export/backup and re-import an oracle theaurus?
    Thanks and best regards,
    Martin

    Use ctxload with -thesdump to export and -thes to import as in the examples from the online documentation below.
    http://download.oracle.com/docs/cd/B28359_01/text.111/b28304/cexec.htm#CCREF2187
    14.1.3.1 Thesaurus Import Example
    The following example imports a thesaurus named tech_doc from an import file named tech_thesaurus.txt:
    ctxload -user jsmith/123abc -thes -name tech_doc -file tech_thesaurus.txt
    14.1.3.2 Thesaurus Export Example
    The following example dumps the contents of a thesaurus named tech_doc into a file named tech_thesaurus.out:
    ctxload -user jsmith/123abc -thesdump -name tech_doc -file tech_thesaurus.out

  • Export Backup takes 90% CPU Usage

    hi Gurus,
    We have export backup scheduled on PROD database. We have noticed that export backup consumes alot of CPU usage.
    Database: Release 9.2.0.6
    CPU usage: Around 90% TO 95% during backup activity
    Backup last for around 20min
    Script file_
    @echo off
    for /F "tokens=2-4 delims=/ " %%f in ('date /t') do (
    set mm=%%f
    set dd=%%g
    set yyyy=%%h
    set DT=%mm%%dd%%yyyy%
    set ORACLE_SID=PRODUCT1
    echo Oracle SID set to PRODUCT1
    exp SYSTEM/DIAL1NG file=E:\ExportBackup\PRODUCT1\PRODUCT1_EXP_FULL_%mm%%dd%%yyyy%.dmp log=E:\ExportBackup\PRODUCT1\PRODUCT1_EXP_FULL_%mm%%dd%%yyyy%.log full=y grants=y indexes=y statistics=none direct=y consistent=y buffer=5000000
    goto END
    : END
    please advise us whether this can be fine tuned?
    Regards
    Shreeshail

    Google "restrict cpu usage windows"
    Lots of stuff out there.

Maybe you are looking for

  • Acrobat XI Std register issue on non-english os

    Hi, I'm deploying Acrobat XI via SCCM in unattended mode. MST has been created with all settings EULA, license key, etc and following language options: in "Installation Options" (silent mode): Application Language English (US). No other options marke

  • After iOS 6, iMessage Microphone Gray

    Okay, I'ved searched everywhere and can't find an answer.  After I updated to iOS 6 in iMessage ONLY, my microphone to dictate is grayed out.  Siri Dictation works everywhere else (email, notes, etc) with out a problem.  Any help would be greatly app

  • Thread program: Strange output

    Hi, I have written a simple thread program. Previously I was trying to print hello world in the run method, the I changed it to print the value of a static data member, but strangely its still printing hello World. However if I change the print state

  • HT4859 how to manage photos in icloud?

    How to view my photos storage in icloud? also, how to delete there photos in icloud?

  • My External Hard Drive and Pen Drives doesn't show up in desktop but shows in finder.

    Hi, I've been wondering why my Pen Drives and Hard Drives won't show up in desktop when I plug it in. Its showing up in finder but I cant seem to solve this problem. help!