Daily export backup

Hi,
I have production & DR database of 350gb version 10.2.0.4 on aix 5.3 platform.
In morning time after completion of daily batch, we are taking export backup (by stopping application services till application is not available for users) by using datapump utility which is taking 2 to 2.30 hrs to completes.
After that we have starting the application services and then only application available for users.
This export backup dump we are daily using for restoration on another development server.
Kindly help me, is there any option to avoid export backup time so that it will not affect buisness time.
and also i can do daily restoration on development server.
Please help.
Regards,

In morning time after completion of daily batch, we are taking export backup (by stopping application services till application is not available for users) by using datapump utility which is taking 2 to 2.30 hrs to completes.
After that we have starting the application services and then only application available for users.
This export backup dump we are daily using for restoration on another development server.
Kindly help me, is there any option to avoid export backup time so that it will not affect buisness time.
and also i can do daily restoration on development server.if you want complete database refresh?
you can use STREAMS.
If certain schema/objects you can use MViews.

Similar Messages

  • Daily export backup (via datapump) of a 600GB production database

    Hi Guys,
    I have a 600GB DB.
    10G Database.
    Currently I have daily rman backup to tape.
    Based on your experience, should i bother to perform a daily export of production database of such size?
    Do u think it's useful?
    thanks

    Fran wrote:
    All depends, what do you want to do? Do you have enough space to save a backup and export daily?
    In my opinion with one backup is enough to save your database. I'm sorry, I can't agree with that. I've been in situations where the first, second and THIRD backup sources have been unavailable in the event of a Production restore. I had to go to Plan D and cross my fingers (and sweat a lot). That might be overkill, but you should never rely on just one backup method/source for Production data.
    Not only that, but, obviously, you can also use an export of the data and import into another database or a schema inside the database in case you need to address any logical corruption. We also use the exports to refresh test databases.
    We have many similar sizes of databases and we take exports weekly, instead of daily. An export will be an incomplete backup, obviously, though you can ensure consistency by giving it an SCN to use as a datapoint. You can also run Data Pump in parallel which does speed things up.
    If you use it as part of your backup strategy, I'd make sure I had regular backups of the parameter file and the controlfiles (both binary and trace). I'd also make sure I had regular txt files showing the necessary information for the tablespaces and the datafiles which you can use to recreate the files in case you needed to.
    Mark

  • Export Backup Size is different

    Hi Gurus,
    We are using Oracle 10G (10.2.0.1.0) on Solaris 10 and we used to take export backup daily basis. While taking export backup using 'exp' then .dmp file size is becoming 30.3 GB where as when we use 'expdp' .dmp file size is 26.1 GB. I've checked both the log files and found no. of tables along with their records are same. so I'm confused why this size difference? and as a result we are not in a position to implement 'expdp'. Can anybody tell me why this size difference and whether can we rely on expdp or not?

    user606947 wrote:
    Hi Gurus,
    We are using Oracle 10G (10.2.0.1.0) on Solaris 10 and we used to take export backup daily basis. While taking export backup using 'exp' then .dmp file size is becoming 30.3 GB where as when we use 'expdp' .dmp file size is 26.1 GB. I've checked both the log files and found no. of tables along with their records are same. so I'm confused why this size difference? and as a result we are not in a position to implement 'expdp'. Can anybody tell me why this size difference and whether can we rely on expdp or not?
    Oracle recommends that you use the new Data Pump Export and Import utilities because they support all Oracle Database 10g features
    Look at the following documentation:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm#sthref14

  • Export backup failing with ORA-1555snapshot too old: rollback segment ERROR

    DB version =8.1.7.4
    OS=Solaris =5.10.
    I know this is outdated. But still we are supporting this version of database.
    We are facing ORA-1555 error daily when we are running the export backup of database. Consistent parameter is set to N .
    I needed to know why export uses rollback segments during export operation. even though CONSISTENT parameter is N.
    Thanks,
    Edited by: ahshivap on Dec 12, 2012 5:19 AM

    Hi,
    Consistent=y just mean the tables are consistent with each other. Oracle still has to guarantee that each individual table is at least consistent with itself. That's your issue.
    Regards,
    Harry

  • Daily Export

    I want to schedule a daily export using the EM on all my databases, but this seems to be impossible?
    When you define an export using the EM, you can schedule it, but there is no repeat option? Also in the job tab you can create new jobs, but there is no specific Export job.
    What I tried was scheduling an export, and then do an edit of this job. Then you can put it on repeat and safe the job in the library. But when you try to submit the job, you get en 'One or more required job paramters are missing.' error.
    Does anyone has a solution for this problem?
    Regards,
    Stevne

    I have the same problem.
    So i looked around and found this answer on Metalink by Oracle:
    "However please note that Edit is not supported for this job type, only general information about the job can be updated. Therefore you cannot schedule repeated executions for this job type. This is an enhancement request. "
    It is ridiculous that they took something out that was always in there. Because they want to push export/import back, they just kinda quit supporting it in this way.
    I'm a big fan op RMAN, which they support totally, but with database that are not running in archivelog mode, i'm much more flexibele with a dump file so i can correct little errors (like a dropped table).
    A big thanks for nothing goes out to Oracle!

  • Daily database backup runnig on RAC 10.2.0.3 with two Solaris nodes & ASM

    Hello,
    I'm confused about making daily incremental RMAN backup of a database runnig on Oracle RAC 10.2.0.3 on two Solaris nodes and ASM.
    Databases cannot be stopped for backup and point in time recovery is needed.
    Please correct me if I am doing wrong:
    1) Backup must be Weekly database Level 0 and Daily database Level 1 is appropriate.
    2) Is best to use a Flash Recovery Area (FRA) - I've setup one on a separate ASM volume. I have enough space to do the backup on FRA.
    3) I am confused about the number channels I need to use. Do I need to use separate RMAN backup channels for each node or I can just leave the default:
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    4) Below are the RMAN scripts I plan to use for
    - Weekly database backup Level 0:
    #!/usr/bin/ksh
    echo BACKUP FULL: $ORACLE_HOME/bin/rman target / nocatalog log ${LOGFILE} append
    $ORACLE_HOME/bin/rman target / nocatalog log ${LOGFILE} append << eof
    run {
    backup incremental level 0 database TAG bkp_level0_200906010;
    SQL 'alter system archive log current';
    backup archivelog all;
    delete noprompt obsolete;
    exit;
    eof
    - Daily database backup Level 1:
    #!/usr/bin/ksh
    echo BACKUP INCREMENTAL: $ORACLE_HOME/bin/rman target / nocatalog log ${LOGFILE} append
    $ORACLE_HOME/bin/rman target / nocatalog log ${LOGFILE} append << eof
    run {
    backup incremental level 1 database TAG bkp_level1_20090611;
    SQL 'alter system archive log current';
    backup archivelog all;
    delete noprompt obsolete;
    exit;
    eof
    I'm I doing things wrong?

    Show also the parameter I have set for RMAN:
    node1:oracle$ rman target /
    Recovery Manager: Release 10.2.0.3.0 - Production on Tue Jun 9 23:46:05 2009
    connected to target database: ORCL (DBID=1198388159)
    RMAN> show all;
    RMAN configuration parameters are:
    CONFIGURE RETENTION POLICY TO REDUNDANCY 7;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '%F'; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '/u01/oracle/product/10gr2/db/dbs/snapcf_orcl1.f'; # default
    Showing some instance parameters on node1:
    SQL> show parameter recovery;
    NAME                                 TYPE        VALUE
    db_recovery_file_dest                string      DGFRA+
    db_recovery_file_dest_size           big integer 300G
    recovery_parallelism                 integer     0
    SQL> show parameter archive;
    NAME                                 TYPE        VALUE
    archive_lag_target                   integer     0
    log_archive_config                   string
    log_archive_dest                     string
    log_archive_dest_1                   string
    log_archive_dest_10                  string      location=use_db_recovery_file_
    dest
    log_archive_dest_2                   string
    log_archive_dest_3                   string
    log_archive_dest_4                   string
    log_archive_dest_5                   string
    log_archive_dest_6                   string
    NAME                                 TYPE        VALUE
    log_archive_dest_7                   string
    log_archive_dest_8                   string
    log_archive_dest_9                   string
    log_archive_dest_state_1             string      enable
    log_archive_dest_state_10            string      enable
    log_archive_dest_state_2             string      enable
    log_archive_dest_state_3             string      enable
    log_archive_dest_state_4             string      enable
    log_archive_dest_state_5             string      enable
    log_archive_dest_state_6             string      enable
    log_archive_dest_state_7             string      enable
    NAME                                 TYPE        VALUE
    log_archive_dest_state_8             string      enable
    log_archive_dest_state_9             string      enable
    log_archive_duplex_dest              string
    log_archive_format                   string      %t_%s_%r.dbf
    log_archive_local_first              boolean     TRUE
    log_archive_max_processes            integer     2
    log_archive_min_succeed_dest         integer     1
    log_archive_start                    boolean     FALSE
    log_archive_trace                    integer     0
    remote_archive_enable                string      true
    standby_archive_dest                 string      ?/dbs/arch

  • Export backup error while using where clause

    I am heaving Oracle9i on solaris platform. Whem i am trying to take the export backup of a table it is giving below error:
    exp swtiob tables=NDC_ATMPROOF_HIST file=NDC_ATMPROOF_HIST.dmp log=NDC_ATMPROOF_HIST.log query="where PROOF_DATE >= '01-july-2010'" statistics=none
    LRM-00112: multiple values not allowed for parameter 'query'
    EXP-00019: failed to process parameters, type 'EXP HELP=Y' for help
    EXP-00000: Export terminated unsuccessfully

    You need to escape the stuff like this:
    $ exp scott/tiger tables=emp file=emp.dmp log=emp.log query=\"where HIREDATE\>\'09-JUN-1981\'\"
    Export: Release 9.2.0.8.0 - Production on Wed Jul 7 12:54:48 2010
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.8.0 - Production
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P1 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                            EMP          6 rows exported
    EXP-00091: Exporting questionable statistics.
    Export terminated successfully with warnings.
    $

  • How to ZIP Oracle Datapump export backup file

    Hello All,
    My customer is asking to give him the production data dump to the following path \\138.90.17.56\OMNISAFE.
    I really don't understand his requirement and he also wants me to zip the export backup file. How do I do that, Do you know any unix command to zip backup files.
    thanks and regards
    cherry

    1013498 wrote:
    Well Thanks for your reply.....my oracle version is 11.2.0.3.b and if we have the compression option can you please elaborate how to do that......
    It's in the documentation.  See Data Pump Export
    let us say my expdp file is abc.dmp...should I give the command gzip abc.dmp or any different.
    Let me google that for you
    One more question what does teh customer mean by production data dump to the following path \\138.90.17.56\OMNISAFE. please explain
    How do we know what the customer means?  Why don't you ask him?
    That said, it looks like a url to an ip address and a defined folder at that ip address.  Again, if the customer wants you to send them a file, you need to be working with said customer on the mechanics of accessing their system.
    All that said ....
    Learning how to look things up in the documentation is time well spent investing in your career.  To that end, you should drop everything else you are doing and do the following:
    Go to tahiti.oracle.com.
    Locate the link for your Oracle product and version, and click on it.
    You are now at the entire documentation set for your selected Oracle product and version.
    BOOKMARK THAT LOCATION
    Spend a few minutes just getting familiar with what is available here. Take special note of the "books" and "search" tabs. Under the "books" tab (for 10.x) or the "Master Book List" link (for 11.x) you will find the complete documentation library.
    Spend a few minutes just getting familiar with what kind  of documentation is available there by simply browsing the titles under the "Books" tab.
    Open the Reference Manual and spend a few minutes looking through the table of contents to get familiar with what kind of information is available there.
    Do the same with the SQL Reference Manual.
    Do the same with the Utilities manual.
    You don't have to read the above in depth.  They are reference manuals.  Just get familiar with what is there to be referenced. Ninety percent of the questions asked on this forum can be answered in less than 5 minutes by simply searching one of the above manuals.
    Then set yourself a plan to dig deeper.
    - *Read a chapter a day from the Concepts Manual*.
    - Take a look in your alert log.  One of the first things listed at startup is the initialization parms with non-default values. Read up on each one of them (listed in your alert log) in the Reference Manual.
    - Take a look at your listener.ora, tnsnames.ora, and sqlnet.ora files. Go to the Network Administrators manual and read up on everything you see in those files.
    - *When you have finished reading the Concepts Manual, do it again*.
    Give a man a fish and he eats for a day. Teach a man to fish and he eats for a lifetime.

  • Can't take export backup

    Hello,
    I can't take export backup as it ends up with a message column name invalid.How to solve the problem. My database is on UNIXWARE 7.1.1 & Oracle8i.
    D ROY
    SHILLONG- INDIA

    content of the log file:
    Connected to: Oracle8i Enterprise Edition Release 8.1.7.4.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.4.0 - Production
    Export done in US7ASCII character set and US7ASCII NCHAR character set
    About to export the entire database ...
    . exporting tablespace definitions
    . exporting profiles
    . exporting user definitions
    . exporting roles
    . exporting resource costs
    . exporting rollback segment definitions
    . exporting database links
    . exporting sequence numbers
    EXP-00008: ORACLE error 904 encountered
    ORA-00904: invalid column name
    EXP-00000: Export terminated unsuccessfully
    Last line of the dmp files
    CONNECT SYSTEM
    EXECFULL DBMS_DISTRIBUTED_TRUST_ADMIN.allow_all;
    CREATE SEQUENCE "AQ$_QIDSEQ" MINVALUE 1 MAXVALUE 999999999999999999999999999 INCREMENT BY 1 START WITH 21 CACHE 20 NOORDER NOCYCLE
    ___________________________________

  • Export backups failing - Need advice

    Hello All,
    We have a 10.2.0.3 database for which we have to take export backups everyday in maintainence window, when only batch jobs and other non essentials SQL's are running on the database.
    The export command we have used in our script is as under:
    $ORACLE_HOME/bin/exp / file=$ORACLE_DUMP_FILE compress=y full=y direct=n buffer=20480000 consistent=y statistics=none log=$ORACLE_LOGAtleast 3-4 times in a week, the export backup us failing with ORA-01555 errors like the one below:
    EXP-00008: ORACLE error 1555 encountered
    ORA-01555: snapshot too old: rollback segment number 10 with name "_SYSSMU10$" too small
    EXP-00000: Export terminated unsuccessfullyMy DB is configured as:
    SQL> show parameter undo_
    NAME                                 TYPE        VALUE
    undo_management                      string      AUTO
    undo_retention                       integer     54000
    undo_tablespace                      string      UNDOTBSAlso:
    SQL> select max(tuned_undoretention) from v$undostat;
    MAX(TUNED_UNDORETENTION)
                       75598The Max value is reached somewhere in the evening, but our job runs at
    SQL> select begin_time from v$undostat where TUNED_UNDORETENTION = 75598;
    BEGIN_TIME
    29/05/2012 17:39:46At the time (as per my log file) when my export job failed, the above value was:
    SQL> select TUNED_UNDORETENTION from v$undostat where BEGIN_TIME like '29/05/2012 16:2%';
    TUNED_UNDORETENTION
                  71318Please advice how the above issue can be addressed/resolved.
    For further information, please let me know. Thank you.
    Suddhasatwa
    India.
    Edited by: SBhaumik_DBA on May 30, 2012 9:50 PM

    Hello. Thanks for the inputs on CONSISTENT option. There are pre-laid change mgmt processes in every organisation which needs to be maintained/taken in account before such changes are made, and that too, with client approvals, etc. Since it was not a feasible option to get EXPDP or RMAN implemented at this point, so was this question raised.
    There is a small issue, however, if we use CONSISTENT=N. Below is an extract from asktom.oracle.com:
    >
    imagine if you started an export at 9am of the EMP and DEPT tables.
    EMP started exporting at 9am and DEPT at 9:15am.
    Now, the EMP data would be "as of 9am", but the DEPT data would be as of 9:15am. What if
    you import that data now -- will it work? maybe, maybe not -- perhaps at 9:10am, someone
    fired the last employee in department 50 and deleted department 50. Your EMP export
    would have them in this deptno, your DEPT export would not have this deptno. The import
    would fail.
    If you use consistent=y, Oracle will export all of the data "as of 9am", so deptno=50
    will be included in all tables and the import will succeed.
    >
    That's one of the risks involved.
    Is there any other solution we can have in place for this issue? Thanks.

  • EXPORT Backup failing due to character set problem

    Hi ,
    The export Backup failing due to character set problem
    . . exporting table ravidlx
    EXP-00008: ORACLE error 6552 encountered
    ORA-06552: PL/SQL: Compilation unit analysis terminated
    ORA-06553: PLS-553: character set name is not recognized
    P
    Please suggest how to set character set
    Regards,
    kk
    Edited by: kk001 on Aug 29, 2011 7:22 PM

    kk001 wrote:
    Hi ,
    The export Backup failing due to character set problem
    . . exporting table ravidlx
    EXP-00008: ORACLE error 6552 encountered
    ORA-06552: PL/SQL: Compilation unit analysis terminated
    ORA-06553: PLS-553: character set name is not recognized
    P
    Please suggest how to set character set
    I don't know what you have.
    I don't know what you do.
    I don't know what you see.
    It is really, Really, REALLY difficult to fix a problem that can not be seen.
    use COPY & PASTE so we can see what you do & how Oracle responds.
    do as below so we can know complete Oracle version & OS name.
    Post via COPY & PASTE complete results of
    SELECT * from v$version;

  • Apply RMAN archivelogs on export backups

    Hello Everyone
    We are using the below script for RMAN
    run
    allocate channel c1 type 'sbt_tape';
    backup database plus archivelog;
    release channel c1;
    So whether we shall be able to apply archivelogs on export backups ? If yes how?
    Thanks.
    Achyot

    It's impossible to apply archived redo logs (physical backup) after importing a export file (logical backup).
    You can only apply archived redo logs to a physical database backup or to a physical or logical standby database (Data Guard feature).
    See logical and physical backup definitions:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14220/glossary.htm#sthref4165
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14220/glossary.htm#i997232

  • Export backup terminated due to ORA-00600: internal error code, arguments:

    Hi,
    exp system/system full=y file=exp_bkp.dmp log=exp_bkp.log statistics=none
    My export backup terminated unsuccessfully and database also down due to ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    os version=RHEL4
    DB Version=10.2.0.1.0
    export log
    exp_bkp.log
    About to export the entire database ...
    . exporting tablespace definitions
    . exporting profiles
    . exporting user definitions
    . exporting roles
    . exporting resource costs
    . exporting rollback segment definitions
    . exporting database links
    . exporting sequence numbers
    . exporting directory aliases
    . exporting context namespaces
    . exporting foreign function library names
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions
    . exporting system procedural objects and actions
    . exporting pre-schema procedural objects and actions
    . exporting cluster definitions
    . about to export SYSTEM's tables via Conventional Path ...
    . . exporting table DEF$_AQCALL 0 rows exported
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Alert Log
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:43:11 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:43:12 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:43:14 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3704.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    ORA-00600: internal error code, arguments: [17147], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:44:14 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_pmon_3648.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:44:16 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_pmon_3648.trc:
    ORA-00600: internal error code, arguments: [17112], [0x4152199C], [], [], [], [], [], []
    Mon Jun 8 13:44:16 2009
    PMON: terminating instance due to error 472
    Mon Jun 8 13:44:16 2009
    Errors in file /u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_psp0_3650.trc:
    ORA-00472: PMON process terminated with error
    Instance terminated by PMON, pid = 3648

    Hi,
    Trace files
    */u02/app/oracle/product/10.2.0/db_1/admin/PRIM/bdump/prim_pmon_3648.trc*
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u02/app/oracle/product/10.2.0/db_1
    System name:     Linux
    Node name:     test1.domain.com
    Release:     2.6.9-5.EL
    Version:     #1 Wed Jan 5 19:22:18 EST 2005
    Machine:     i686
    Instance name: PRIM
    Redo thread mounted by this instance: 1
    Oracle process number: 2
    Unix process pid: 3648, image: [email protected] (PMON)
    *** 2009-06-08 13:44:14.697
    *** SERVICE NAME:(SYS$BACKGROUND) 2009-06-08 13:44:14.695
    *** SESSION ID:(170.1) 2009-06-08 13:44:14.695
    ********** Internal heap ERROR 17112 addr=0x4152199c *********
    ***** Dump of memory around addr 0x4152199c:
    41520990 00001001 [....]
    415209A0 4151F99C 2001CFBC 00000000 415219A8 [..QA... ......RA]
    415209B0 D0000FED 00000000 415219B8 40EFEB10 [..........RA...@]
    415209C0 415219BC 00000025 00000000 0C7470F0 [..RA%........pt.]
    415209D0 00000000 46440007 454C544C 0000004E [......DFLTLEN...]
    415209E0 000A000C 0E000001 00000015 415209C4 [..............RA]
    415209F0 0C7470D8 41520A70 41520A1C 00000015 [.pt.p.RA..RA....]
    41520A00 415209E8 0C7470D8 41520A84 41520A1C [..RA.pt...RA..RA]
    41520A10 00000055 415209FC 0C747378 00000201 [U.....RAxst.....]
    41520A20 00000000 00000059 00008100 00030040 [....Y.......@...]
    41520A30 00000000 00000016 41548140 00000000 [[email protected]....]
    41520A40 04000008 00000000 40FFBEFC 00000000 [...........@....]
    41520A50 00000000 40F34AF0 40F56490 00000000 [[email protected].@....]
    */u02/app/oracle/product/10.2.0/db_1/admin/PRIM/udump/prim_ora_3677.trc*
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u02/app/oracle/product/10.2.0/db_1
    System name:     Linux
    Node name:     test1.domain.com
    Release:     2.6.9-5.EL
    Version:     #1 Wed Jan 5 19:22:18 EST 2005
    Machine:     i686
    Instance name: PRIM
    Redo thread mounted by this instance: 1
    Oracle process number: 15
    Unix process pid: 3677, image: [email protected] (TNS V1-V3)
    *** SERVICE NAME:() 2009-06-08 13:41:55.582
    *** SESSION ID:(159.3) 2009-06-08 13:41:55.582
    Thread 1 checkpoint: logseq 183, block 2, scn 473582
    cache-low rba: logseq 183, block 3
    on-disk rba: logseq 183, block 101, scn 473671
    start recovery at logseq 183, block 3, scn 0
    ----- Redo read statistics for thread 1 -----
    Read rate (ASYNC): 49Kb in 0.55s => 0.09 Mb/sec
    Total physical reads: 4096Kb
    Longest record: 0Kb, moves: 0/152 (0%)
    Longest LWN: 6Kb, moves: 0/24 (0%), moved: 0Mb
    Last redo scn: 0x0000.00073a46 (473670)
    ----- Recovery Hash Table Statistics ---------
    Hash table buckets = 32768
    Longest hash chain = 1
    Average hash chain = 41/41 = 1.0
    Max compares per lookup = 1
    Avg compares per lookup = 223/264 = 0.8
    *** 2009-06-08 13:41:56.180
    KCRA: start recovery claims for 41 data blocks
    *** 2009-06-08 13:41:57.051
    KCRA: blocks processed = 41/41, claimed = 41, eliminated = 0
    *** 2009-06-08 13:41:57.053
    Recovery of Online Redo Log: Thread 1 Group 1 Seq 183 Reading mem 0
    ----- Recovery Hash Table Statistics ---------
    Hash table buckets = 32768
    Longest hash chain = 1
    Average hash chain = 41/41 = 1.0
    Max compares per lookup = 1
    Avg compares per lookup = 264/264 = 1.0
    tkcrrsarc: (WARN) Failed to find ARCH for message (message:0x1)
    tkcrrpa: (WARN) Failed initial attempt to send ARCH message (message:0x1)

  • Export / Backup all the production data

    Dear Experts,
    Thanks for your attention!
    Due to we have some upgrading issue in our production system. We plan to re-install the GRC AC SP10 in PRD.
    We have go-live for half a year, there are around 500 requests in PRD, all these transaction data need to be export / backup before re-installation.
    Could any one give us any advice?
    Thanks a lot!

    Hi,
    This will be handled by BASIS team, you need not worry. Once the data was restored you need to amke sure all the production data was avilable with configuration settings, audit trial, all the requests data.
    If this was first time you are performing, better check out restore on non-production server and identify any issues and resolve and then proceed onto production.
    Backup and restore will be done only by basis and documents avilable in market place, also that was dependent on your DB type.
    regards.

  • Export backup degrades performance?

    Hi all,
    Please suggest me if i take export backup of the database does it degrades performance of the database when the applications are connected?
    Can this export be done while application is using the database?
    Is there any risk concerning performance of the application during DB export?
    Thank you.

    An export is a query against the database. Generally, a query to extract the whole schema or whole table (or whole database). It has the same performance impact as any query doing FullTableScans against the tables being exported.
    Note :If you do an export using 'exp', ensure that you use "CONSISTENT=Y" or "FLASHBACK_SCN" or "FLASBACK_TIME" to get consistency across tables, else exp does Table-level consistency.
    With 'expdp' you can use FLASHBACK_SCN or FLASHBACK_TIME

Maybe you are looking for

  • Loading sound settings in ga

    I am running Windoes Xp with Audigty 2 zs. When I play games like Halo I don't get all the sound settings like environmental sound. How can I have all the sound settings turned on so I can use the card I have? Is there some settings I need to do from

  • Include files in Oracle9i

    Hi, I'm trying to compile my C program using gcc in Oracle9i. But came out error. Here's the error below:- test.c:154:19: sqlca.h: No such file or directory test.c:155:19: sqlda.h: No such file or directory test.c:156:20: sqlcpr.h: No such file or di

  • Cancel payment - void cheque & reverse posting

    Hi All, I have a request to cancel payment using FCH8 for a range of cheques. (Mass cancellation  eg.10 cheques at one go) Is there another t-code that can function the same way as FCH8 but for mass cancellation? Help much appreciated. Thanks.

  • SAP Work Flow related

    Hi Experts, I am new to Webdynpro programming. I need to find the program, from where the data is getting populated in one of the column at front end portal. Is it possible to track program name from the below structure: Webdynpro componant !->View  

  • Premiere Pro CS4 graphic problem?

    When adding a video clip or audio and image to the timeline I get this ghost effect on the timeline and its hard to see whats going on.what could be the problem and how do i fix this?