Full database recovery without redolog availability

Dear DBAs,
I performed full database backup using RMAN in catalog mode, then I physically removed the database, then using parameter file I started the database in nomount mode and recovered the control file using RMAN, after that I altered the database to mount mode and restored all the datafiles, now definetly I can not open the database untill it has not redo logs available in it which I removed early, the resetlogs option is also not working and I am stuck, tell me what I should do, the database is in archivelog mode and all archive files are available.
Kindly help,
Regards,
Asif

It sounds like, after you completed the RMAN backup, you didn't archive the current log file. That means that the last bit of redo you need to get your database consistent is in the online redo that's no longer available. I think you're screwed....
Yes, there are ways to open the database as is, but, before we go there, what situation are you in? Is this a database you need? were you practicing backup and recovery?
If you absolutely need to recover this database, let us know.
Otherwise, I recommend retrying the procedure, and this time, archiving the current log file after the RMAN backup completes, and before removing the online redo. If you do that, and make sure that last bit of redo gets archived, I think the procedure should work.
Hope that helps,
-Mark

Similar Messages

  • Database recovery (online redolog ?)

    hi all,
    Been awhile since i touch on oracle db, i have been reading around and the emphasis for recovery is always on the backup and archivelog, but i think its wrong.
    can i check ->
    q1) for full database recovery, do i need the online redo log as well ?
    q2) if the answer to q1) is yes, how do i duplicate online redo log to standby site ? (i don't think rsync will work as it cannot sure consistency in the redolog)
    will oracle dataguard sync online redolog as well ?
    q3) for archivelog, beside manual rsyncing, LOG_ARCHIVE_DEST_2 = 'SERVICE=standby1'
    do i need the enterprise edition for the above ?
    Regards,
    Alan

    q1) For a complete recovery, yes you need online redolog as well. Without online redolog,its still considered incomplete recovery since u lose data resides in online redolog
    q2) You no need to synch online redo log manually. Once the backup is restored to the DR dataguard site and MRP process initiated, Oracle will synch online redolog/archivelog automatically based on protection mode specified
    q3) Oracle dataguard applies to Enterprise Edition only. Without Enterprise Edition, we can configure log shipping (manual way).
    Regards,
    Ilan

  • Import all users and their objects without doing full database import

    Hi Guys,
    I have a task to that involves Importing all existing users and their objects in production to the test database without doing a full database import - Please how do i do this?
    Pls i need your help urgently.

    SQL> select * from v$version;
    BANNER
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Product
    PL/SQL Release 10.2.0.1.0 - Production
    CORE 10.2.0.1.0 Production
    TNS for Linux: Version 10.2.0.1.0 - Production
    NLSRTL Version 10.2.0.1.0 - Production
    I tried to import objects and data from a user from a FULL dump file. File was created with the following command:
    server is: SQL*Plus: Release 10.2.0.1.0 - Production on Wed May 26 15:34:05 2010
    exp full=y FILE="full.dmp" log=full.log
    Now I imported:
    imp file=full.dmp log=full.log INCTYPE=SYSTEM
    imp fromuser=user1 file=full.dmp
    Results: not all the user procedures have been imported:
    SQL> select count(*) from user_procedures;
    the Original
    COUNT(*)
    134
    the current:
    select count(*) from user_procedures;
    COUNT(*)
    18
    I also tried these alternatives:
    exp tablespaces="user1_data" FILE="user1.dmp" log=user1.log
    exp LOG=user1.log TABLESPACES=user1_data FILE=user1_data.dmp
    exp LOG=user1owner.log owner=user1 FILE=user1owner.dmp
    expdp DIRECTORY=dpump_dir1 dumpfile=servdata_pump version=compatible SCHEMAS=user1
    impdp directory=data_pump_dir dumpfile=servdata_pump.dmp :
    ORA-39213: Metadata processing is not available
    SQL> execute dbms_metadata_util.load_stylesheets
    BEGIN dbms_metadata_util.load_stylesheets; END;
    ERROR at line 1:
    ORA-31609: error loading file "kualter.xsl" from file system directory
    "/usr/lib/oracle/xe/app/oracle/product/10.2.0/server/rdbms/xml/xsl"
    ORA-06512: at "SYS.DBMS_METADATA_UTIL", line 1793
    ORA-06512: at line 1
    file kualter.xsl does not exist in XE !!
    imp owner=user1 rows=n ignore=y
    imp full=y file=user1_data.dmp
    imp full=y file=full.dmp tablespaces=user1_data,user1_index rows=n ignore=y
    So, I do not understand why user1 objects are not imported:
    see this part of the first import log:
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P1 character set (possible charset conversion)
    . importing SYS's objects into SYS
    . importing USER1's objects into USER1
    . . importing table .........................
    why only 18 rows?
    if you have an suggestion, you are welcome, as I do not have any other idea...
    ren
    Edited by: ronpetitpatapon on May 26, 2010 12:38 PM
    Edited by: ronpetitpatapon on May 26, 2010 12:41 PM
    Edited by: ronpetitpatapon on May 26, 2010 1:03 PM

  • Full database restore and recovery

    Hi all,
    All my datafiles are corrupted.
    I have a incremental level 0 backup
    can i restore and recover the db with this backup?
    regards,
    prem

    we went for a point in time recovery
    while restoring we received the following error.
    Full Database restoration failed with error
    ORACLE error from target database:
    ORA-00600: internal error code, arguments: [2103], [0], [0], [1], [900], [], [], []
    ORA-06512: at "SYS.X$DBMS_BACKUP_RESTORE", line 5157
    ORA-06512: at line 1
    released channel: GMUP1
    released channel: GMUP2
    released channel: GMUP3
    Please help us in this regard

  • Restore a single table from full database backup

    Hi,
    I have a full database (10.2.0.1) backup with RMAN.
    Suddenly a table got truncated. How to get back that table without restoring and recovering the full database (Point in time recovery) since if I restore the database to an earlier time then the other transactions will hamper.
    Also creating a standby database and import export the table will solve my problem. But is there any other way to restore the lost table.
    Thanks in advance.

    No other way to restore the data if you don't want do point in time recovery. If you drop the table then you can restore it from the recyclebin but for the truncate operation it won't help. If you have enabled the flash back table for that particular table then you can restore the data for only that table.
    Regards
    Asif Kabir

  • Is backup full database , is backuing up controlfile ?

    Hi
    i am on 8174
    is that enough to cold backup the database?
    shutdown immediate;
    startup mount;
    run {
    allocate channel t1 type 'SBT_TAPE';
    send 'NB_ORA_POLICY=....';
    backup full format 'bk_%U_%t' (database);
    or do i need
    backup current controlfile; ?
    i can see here http://download-west.oracle.com/docs/cd/A87860_01/doc/server.817/a76990/rmanconc.htm#436070
    this statment "A full backup is different from a whole database backup, which is a backup of all datafiles and the current control file."
    is that means
    backup full database , is backuing up controlfile ?(i think yes) and is restoring from it requires any archivelogs?(i think no)
    Thanks very much
    Edited by: user9145417 on 25-Aug-2010 16:42

    The reason to have the current controlfile is that it keeps all the information of the database. If you lose the entire machine, it will be much more difficult to recover to another machine without that controlfile. That is one reason that control files are multiplexed, so that you do not lose all of them.
    With 8174 I don't believe there was a method to always included it automatically like with 10g. So yes...always backup your control file with an exact statement.
    Will you need archive logs? If you want to do a complete recovery you will need all the archivelogs that occur after the start of the backup untill the time the database crashes.
    Do you need to backup the archivelogs within this script? Well that depends if you can think of any circumstance that would require you to recover to a point in time using the backup taken before this one.
    Regards
    Tim

  • Object Level Recovery or Whole Database recovery

    I'm hoping someone may know how to advise me on the following;
    On a datawarehouse db (10.2.0.1.0) a team member removed records from three tables, and I have since attempted flashback recovery without success. The database is in Archivelog mode, with Flashback enabled, but no Flashback logging enabled. The rows were removed on Friday afternoon (it is now Monday). I attempted to get flashback logging enabled by tagging the "Enable Flashback Database" tag in the Flash Recovery region of Recovery settings, and restarting the database. The database when restarted went into mount state, and subsequently on restarting (from mount - I did not dismount the db), it still has Flashback logging disabled. I attempted flashback again but the team member states the record stil arn't there. EM however had given the message 'The select tables...X X....have been flashed back'. However I can see also that Em says flashback logging is still disabled.
    I now consider I might be better off to perform a 'Whole Database Recovery', as I simply want to get the tables recovered. I'm not sure if this will mean re-keying though. Can anyone advise? Thanks in advance. DW
    Message was edited by:
    David_W

    The first step you should try is flashback query. Because using flashback query your database will remain intact, you don't lose anything from your database. Of course most likely it's too late for you now. Just for future reference.
    Flashback database is only available after you configured Flash Recovery Area and turn it on. Sound like it's does apply here as well. Remember even you could successful flashback your database to the point before deletion, you will lost all data changes after that point. Flashback database only buy you sometime, because you don't need to restore datafile from backup.
    The third option would be restore from your last backup, ( the latest one before deletion happen) and do incomplete restore to the point in time right before the incident.

  • Questions About Database Recovery (-30975)

    Hello,
    In Berkeley 4.5.20, we are seeing the following error sporadically, but more frequently than we'd like (which is, to say, not at all): "BerkeleyDbErrno=-30975 - DbEnv::open: DB_RUNRECOVERY: Fatal error, run database recovery"
    This exception is being thrown mostly, if not exclusively, during the environment open call. Still investigating.
    I will post my environment below, but first some questions.
    1. How often should a database become become corrupt?
    2. What are the causes of this corruption? Can they be caused by "chance?" (I.e. app is properly coded.) Can they be caused by improper coding? If so, is there a list of common things to check?
    3. Does Oracle expect application developers to create their own recovery handlers, especially for apps that require 100% uptime? E.g. using DB_ENV->set_event_notify or filtering on DB_RUNRECOVERY.
    Our environment:
    Windows Server 2003 SP2
    Berkeley DB 4.5.20
    set_verbose(DB_VERB_WAITSFOR, 1);
    set_cachesize(0, 65536 * 1024, 1);
    set_lg_max(10000000);
    set_lk_detect(DB_LOCK_YOUNGEST);
    set_timeout(60000000, DB_SET_LOCK_TIMEOUT);
    set_timeout(60000000, DB_SET_TXN_TIMEOUT);
    set_tx_max(100000);
    set_flags(DB_TXN_NOSYNC, 1);
    set_flags(DB_LOG_AUTOREMOVE, 1);
    set_lk_max_lockers(10000);
    set_lk_max_locks(10000);
    set_lk_max_objects(10000);
    open(sPath, DB_CREATE | DB_INIT_LOCK | DB_INIT_LOG | DB_INIT_MPOOL | DB_THREAD | DB_INIT_TXN | DB_RECOVER, 0);
    set_pagesize     (4096);
    u_int32_t dbOpenFlags = DB_CREATE | DB_AUTO_COMMIT;
    pDbPrimary->open(NULL, strFile, NULL, DB_HASH, dbOpenFlags, 0);
    We also have a number of secondary databases.
    One additional piece of information that might be relevant is that the databases where this happens (we have 8 in total managed by our process,) seem to be the two specific databases that at times aren't opened until well after the process is up and running due to the nature of their data. This is to say that 6 of the other databases are normally opened during startup of our service. We are still investigating this to see if this is consistently true.

    Here is the output from the error logs (we didn't have this properly set up until now) when this error opening the environment happens:
    12/17/2007 17:18:12 (e64/518) 1024: Berkeley Error: CDbBerkeley MapViewOfFile: Not enough storage is available to process this command.
    12/17/2007 17:18:12 (e64/518) 1024: Berkeley Error: CDbBerkeley PANIC: Not enough space
    12/17/2007 17:18:12 (e64/518) 1024: Berkeley Error: CDbBerkeley DeleteFile: C:\xxxxxxxx\Database\xxxJOB_OAT\__db.003: Access is denied.
    12/17/2007 17:18:12 (e64/518) 1024: Berkeley Error: CDbBerkeley MapViewOfFile: Not enough storage is available to process this command.
    12/17/2007 17:18:12 (e64/518) 1024: Berkeley Error: CDbBerkeley PANIC: Not enough space
    12/17/2007 17:18:12 (e64/518) 1024: Berkeley Error: CDbBerkeley PANIC: DB_RUNRECOVERY: Fatal error, run database recovery
    12/17/2007 17:18:30 (e64/518) 1024: Berkeley Error: CDbBerkeley unable to join the environment
    12/17/2007 17:18:30 (e64/518) 1024: Berkeley Error: CDbBerkeley DeleteFile: C:\xxxxxxxx\Database\xxxJOB_OAT\__db.003.del.0547204268: Access is denied.
    12/17/2007 17:18:30 (e64/518) 1024: Berkeley Error: CDbBerkeley DeleteFile: C:\xxxxxxxx\Database\xxxJOB_OAT\__db.003: Access is denied.
    12/17/2007 17:19:18 (e64/518) 1024: Database EInitialize failed. (C:\xxxxxxxx\Database\xxxJOB_OAT: BerkeleyDbErrno=-30975 - DbEnv::open: DB_RUNRECOVERY: Fatal error, run database recovery)
    The last line is generated by a DbException and was all we were seeing up until now.
    I also set_verbose(DB_VERB_RECOVERY, 1) and set_msgcall to the same log file. We get verbose messages on the 1st 7 database files that open successfully, but none from the last one, I assume because they output to set_errcall instead.
    There is 67GB of free space on this disk by the way, so not sure what "Not enough space" means.
    Thanks again for your help.

  • Problem in performing multiple Point-In-Time Database Recovery using RMAN

    Hello Experts,
    I am getting an error while performing database point in time recovery multiple times using RMAN. Details are as follows :-
    Environment:
    Oracle 11g, ASM,
    Database DiskGroups : DG_DATA (Data files), DG_ARCH(Archive logs), DG_REDO(Redo logs Control file).
    Snapshot DiskGroups :
    Snapshot1 (taken at 9 am): SNAP1_DATA, SNAP1_ARCH, +SNAP1_REDO
    Snapshot2 (taken at 10 am): SNAP2_DATA, SNAP2_ARCH, +SNAP2_REDO
    Steps performed for point in time recovery:
    1. Restore control file from snapshot 2.
         RMAN> RESTORE CONTROLFILE from '+SNAP2_REDO/orcl/CONTROLFILE/Current.256.777398261';
    2. For 2nd recovery, reset incarnation of database to snapshot 2 incarnation (Say 2).
    3. Catalog data files from snapshot 1.
    4. Catalog archive logs from snapshot 2.
    5. Perform point in time recovery till given time.
         STARTUP MOUNT;
         RUN {
              SQL "ALTER SESSION SET NLS_DATE_FORMAT = ''dd-mon-yyyy hh24:mi:ss''";
              SET UNTIL TIME "06-mar-2013 09:30:00";
              RESTORE DATABASE;
              RECOVER DATABASE;
              ALTER DATABASE OPEN RESETLOGS;
    Results:
    Recovery 1: At 10.30 am, I performed first point in time recovery till 9:30 am, it was successful. Database incarnation was raised from *2* to *3*.
    Recovery 2: At 11:10 am, I performed another point in time recovery till 9:45 am, while doing it I reset the incarnation of DB to *2*, it failed with following error :-
    Starting recover at 28-FEB-13
    using channel ORA_DISK_1
    starting media recovery
    media recovery failed
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of recover command at 03/06/2013 11:10:57
    ORA-00283: recovery session canceled due to errors
    RMAN-11003: failure during parse/execution of SQL statement: alter database recover if needed
    start until time 'MAR 06 2013 09:45:00'
    ORA-00283: recovery session canceled due to errors
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '+DG_REDO/orcl/onlinelog/group_1.257.807150859'
    ORA-17503: ksfdopn:2 Failed to open file +DG_REDO/orcl/onlinelog/group_1.257.807150859
    ORA-15012: ASM file '+DG_REDO/orcl/onlinelog/group_1.257.807150859' does not exist
    Doubts:
    1. Why did recovery failed 2nd time, but not 1st time and why is RMAN looking for online redo log group_1.257.807150859 in 2nd recovery ?
    3. I tried restoring control file from AutoBackup, in that case both 1st and 2nd recovery succeeded.
    However for this to work, I always need to keep the AutoBackup feature enabled.
    How reliable is control file AutoBackup ? Is there any alternative to using AutoBackup, can I restore control file from snapshot backup only ?
    4. If I restore control file from AutoBackup, then from what point of time/SCN does RMAN restores the control file ?
    Please help me out in this issue.
    Thanks.

    992748 wrote:
    Hello experts,
    I'm little newbie to RMAN recovery. Please help me in these doubts:
    1. If I have datafiles, archive logs & control files backup, but current online REDO logs are lost, then can I perform incomplete database recovery ?yes, if you have backups of everything else
    2. Till what maximum time/scn can incomplete database recovery be performed ??Assuming the only thing lost is the redo logs, you can recover to the last scn in the last archivelog.
    3. What is role of online REDO logs in incomplete database recovery ? They provide the final redo changes - the ones that have not been written to archivelogs
    Are they required for incomplete recovery ?It depends on how much incomplete recovery you need to do.
    Think of all of your changes as a constant stream of redo information. As a redolog fills, it is copied to archive, then (eventually) reused. over time, your redo stream is in archivelog_1, continuing into archvivelog_2, then to 3, and eventually, when you get to the last archivelog, into the online redo. A recovery will start with the oldest necessary point in the redo stream and continue forward. Whether or not you need the online redo for a PIT recovery depends on how far forward you need to recover.
    But you should take every precaution to prevent loss of online redo logs .. starting with having multiple members in each redo group ... and keeping those multiple members on physically separate disks.

  • How to get a true "full screen" view without the edit tools?  The help manual promises such a view.

    How do I get a true "full screen" view in the new version of Iphoto without the edit tools showing?  Weirdly enough, the help manual promises such a "full" view, which was always available in my earlier version of Iphoto.
    Thanks, in advance.

    Click on the Edit button to get into Edit mode. Click on it again to come out of edit mode.
    Regards
    TD

  • Full Database Export/Import

    I was attempting to export a full database from one server (DB files on the N: drive) and import it into a freshly created new database with the same instance name on a new server (DB files on the D: drive).
    I was told to:
    1) Export the full database as sys using full=y
    2) Pre-create all of the tablespaces on the target database that do not exist on the target database
    3) Import the full database as sys using full=y
    =========================================================
    Here are the steps I had to follow:
    1) Export the full database as sys using full=y
    2) Pre-create all of the tablespaces on the target database
    3) Import the full database as sys using full=y - Result: over 1900 invalid objects and thousands of errors
    4) Run Utlrp.sql - Result: 8 to 10 invalid objects
    5) Manually attempt to compile or rebuild objects. Several grants were required to get most of these to work. - Result: After quite a bit of hand crafting, two objects are still invalid.
    6) After spending almost two hours in a Collaborative Conference with an analyst who is working very hard to sold this, there are still two objects that are invalid.
    Note: From my perspective this is not a straight forward process
    =========================================================
    Questions:
    1) Has anyone successfully moved a database as described without all of these annoying steps?
    2) Is there a document that details the full database export/import?
    Thanks, RJZ

    Hello RJZ,
    There is not sufficient information to reply accurately to your queries,
    1. Could you post the error message that you're getting when you compile these two programs? Do they compile okay on the old server?
    2. What was the reason you choose exp and imp, to move your db to the new server and different disk? Were you migrating to a different version, or was it because of the db-reorg? It would have been possible to simple copy the db-files to the new location, and then rename the datafiles in mount-state. It would be more simpler I believe unless as stated above
    Did you check the import logs? While importing were there any errors? Actually it should have been pretty straight forward (no reason why you need documentation for exp / imp). Grants should have automatically come, unless you have select grants=n option
    I think because of the changes you have done for step-5, you might have changed some stuff somewhere. Please post the error message, and tell us if you have migrated the db to a higher version
    Regards
    Sudhanshu

  • Full Database Backup in DB13 for MS SQL Server databases

    Hello,
    We have some SAP systems using the MS SQL Server database. I want to know if it is possible to setup the Full Database Backup option in transaction DB13 to store the files in a hard disk space.
    I already did this type of configuration in SAP systems in Oracle databases and for this I used the init<SID>.sap where we can put all the configuration, for instance the disk path where we want to save the backup files! But for MS SQL Server databases, in DB13, I don´t know how to do that, when I add the "Full Database Backup" option in DB13, this show me some options but only for tape, but I don´t want to store the backup files in tapes, I want to store/save the files on the server disk itself or else to an external disk but I don´t see where I can set this way of storage!
    This is what you see in DB13 (when using a MS SQL Server database) in "Full Database Backup" and as you see I only have options for TAPE not for a disk or other type of storage! How can I do that? (if that is possible):
    Can you help me please?
    Kind regards,
    samid raif

    Hello Raja
    Sorry for the delay of my answer! Many thanks for that tip, it helps a lot and it solve the last error/problem that I reported here, in fact, the problem was we had named the device with two words containing a space. So we removed that space from the backup device (in SQL management studio) with one word only and the job runs successfully without any errors!
    It works in Development and Quality systems but in our Production system (AM1) the Full backup option in DB13 stopped with the following error, in fact the error is on verification of backup, it means that the Full backup finished with success, but when it does the backup verification, the job in DB13 stops with an error, the following one:
    ***************************** SQL Job information   ******************************
    Jobname:    SAP CCMS Full DB Backup of AM1 [20140417103726-4-103726]
    Type:       TSQL
    DB-Name:    AM1
    For Run:    20140417  10:37:26
    **************************** Job history information  *****************************
    Stepname:   CCMS-step 1
    Command:    declare @exeStmt nvarchar(2000) exec am1.sap_backup_databases @dbList=
                "AM1",@r3Db="AM1",@bDev="fullprd",@expDays= 27,@jobName= "SAP CCMS Ful
                l DB Backup of AM1 [20140417103726-4-103726]",@bkupChecksum="Y",@bkupT
                ype="F",@nativeBkup="N",@exeDate = "20140417103726",@bkupSim = "N",@fo
                rmat = 0,@init = 0,@bkupDb = "AM1",@unload = 0,@exeStmt = @exeStmt OUT
                PUT
    Status:     (success)
    Message:    3014
    Severity:                                                                                                                                                                                                                                                    0
    Duration:   4  hours(s)  41 min(s) 22 sec(s)
    Last msg:   Executed as user: am1. Processed 7434328 pages for database 'AM1', fil
                e 'A01DATA1' on file 1. [SQLSTATE 01000] (Message 4035)  Processed 315
                1504 pages for database 'AM1', file 'A01DATA2' on file 1. [SQLSTATE 01
                000] (Message 4035)  Processed 4574152 pages for database 'AM1', file
                'A01DATA3' on file 1. [SQLSTATE 01000] (Message 4035)  Processed 44363
                92 pages for database 'AM1', file 'A01DATA4' on file 1. [SQLSTATE 0100
                0] (Message 4035)  Processed 25598 pages for database 'AM1', file 'A01
                LOG1' on file 1. [SQLSTATE 01000] (Message 4035)  BACKUP DATABASE succ
                essfully processed 19621974 pages in 16881.638 seconds (9.521 MB/sec).
                 [SQLSTATE 01000] (Message 3014).  The step succeeded.
                        <------------- End of Job Step History --------------->
    Stepname:   CCMS-step 2
    Command:    declare @exeStmt nvarchar(2000) exec am1.sap_verify_backups @nativeBku
                p = "N",@bkupSim = "N",@bDev = "fullprd",@bkupChecksum = "Y",@exeDate
                = "20140417103726",@unload = 1,@dbCnt =1,@exeStmt = @exeStmt OUTPUT
    Status:     (failure)
    Message:    3201
    Severity:                                                                                                                                                                                                                                                   16
    Duration:   0  hours(s)  8  min(s) 30 sec(s)
    Last msg:   Executed as user: am1. Cannot open backup device 'fullprd(\\10.0.0.45\
                backupsap\prd\prdfullqua.bak)'. Operating system error 1265(error not
                found). [SQLSTATE 42000] (Error 3201).  The step failed.
                        <------------- End of Job Step History --------------->
    **************************** Job history information  *****************************
    Can you help me please with this one!? Can you tell me why in verification it stops with that error?
    Kind regards,
    samid raif

  • How to replicate full database in 10g

    Hi,
    We have 9i database but soon we will be migrating to 10gR2. I want to see are there any other options in 10g to replicate full database other than data guard. I want to use that replicated database when ever primary database is not available due to network or other issues.
    Thanks!

    I'm not sure I follow. A database is either in ARCHIVELOG mode or it isn't. That's not a table-level setting. Are you saying that you do unlogged operations into a subset of tables on a regular basis?
    If so, does that mean that this is a data warehouse/ data mart and not an OLTP system? Could you rebuild the standby once a day, say, after the batch load and fail over to that point-in-time system? Or are there ongoing transactions that need to be replicated?
    Justin

  • "Fatal error, run database recovery " when there are no txns to recover.

    Hi, all.
    I have a DB file containing multiple databases. Without using DBEnvironments, I can open it to get the dbnames. I can open the databases RDONLY,
    and see that their contents are correct. I can open them RW, and everything works.
    But when I try to create a new one, I get this:
    D = bsddb3.db.DB()
    D.open('test.db',dbname='test',dbtype=B.DB_BTREE,flags=B.DB_CREATE)Traceback (most recent call last):
    File "<stdin>", line 1, in <module>
    bsddb3.db.DBRunRecoveryError: (-30974, 'DB_RUNRECOVERY: Fatal error, run database recovery -- PANIC: fatal region error detected; run recovery')
    Note that this is in the non-transactional case. There is no Env, and there are no logfiles or __db files. So the error code mystifies me.
    Strace shows that the file is opened RW, and read through.
    B.DB_VERSION_STRING'Berkeley DB 4.8.24: (August 14, 2009)'
    >>>
    So, where to proceed? Many thanks for any and all help.

    Hmm. Other thing to note:
    [tradedesk@vader 2010-05-06.test]$ /usr/local/BerkeleyDB.4.8/bin/db_verify foo.db
    db_verify: Subdatabase entry references page 266 of invalid type 13
    db_verify: Page 0: non-invalid page 40 on free list
    db_verify: trading.db: DB_VERIFY_BAD: Database verification failed
    Not sure how that came about or how to prevent it, but it might have to do wit this issue.

  • Database recovery in a replication environment

    Hi everybody,
    I've configured a replication environment with Oracle 11g (Multimaster synchronous replication). Apparently it works fine, when replication is started any change in a table that belongs in a node is replicated in the rest of the nodes.
    One of the tests I'm making is the following: Try to see what happens if one of the nodes drops. Meanwhile one of the nodes is no avalaible, any modification in an avalaible node is not allowed (as I expected). I'm trying to see if the database service can be avalaible meanwhile a node is missing, although the replications is not running.
    What I've experienced is replication cannot be suspended until the the node was down is restarted.
    Is this the usual behaviour or can I set any parameters to avoid this? In other words, is it possible the database service is running meanwhile one of the nodes is down? The application I'm working on needs this type of replication, buy also needs the service is running always?
    What can I do to make possible the service is running even when a node falls?
    Thanks in advance

    Yes, I've considered an asynchronous replication, with a short period of time (it should be almost synchronous).
    I've tested that, if one of the nodes drops, database service is not available. User intervention is required to stop replication and let the system go on without replication.
    I could be a great idea that, althoug replication is not working, database service is still running and, when replication is restarted, system continue as in the beginning.
    When I test the asynchronous choice I'll answer this post.
    Best wishes,
    Payo

Maybe you are looking for

  • WEI could not measure video performanc​e , brand new w530 I7 quad

    WEI could not measure video performance , brand new w530 I7 quad I keep getting this error still have no score at all in the WEI. here is log 60388 (8976) - winsat\logging.cpp:0815: --- START 2013\12\8 19:52:57 PM --- 960404 (8976) - winsat\main.cpp:

  • A Global Catalog Server could not be located - All GC's are down SBS 2011

    I have been searching through these forums and manage to find similar errors but am struggling to find an answer that applies to this me. I seem to be having a number of issues with our SBS. I believe this was originally domain was previously on a SB

  • Raw files from canon 550D TO Elements7

    Elements 7 appears to download the raw files from my camera ( I can see them as they are selected) but when I choose to import them, the message I receive is that the files are damaged or that the format cannot be imported to Elements 7. Any help wou

  • In search for an 5.1. set with an digital optical out for my PS3

    i have bought an Playstation 3 but now i want to put an 5.1. on it with the optical out the Ps3 has but i really don't know if creative makes them any more and what the prices are of it so if any body got any ideas or information about it that would

  • Embedding a browser shortcut in swf

    i am using javascript to send my animation from one swf to a fullscreen swf file. it all works fine except that both ie and firefox can't cover the start menu unless the user hits f11. its not really a problem but it would be convinient if i could fi