Adding redo logs for dataguard primary server

Dear all:
i have physical dataguard servers
i want to add new redologs for primary server with big size ..
please advice what will be the action with stand by server.
Thanks ,,

Dear all:
i have physical dataguard servers
i want to add new redologs for primary server with
big size ..
please advice what will be the action with stand by
server.
Thanks ,,Most of these information are kept in the control file . You have to recreate your stand by control file after you added the new online redo logs. Transfer this to the standby server , with the the new online redo redo while you standby is down . Your standby system should be able to recognize it when it goes back to stand by mode.

Similar Messages

  • Resizing online and standby redo log in dataguard setup.

    In 10gr2 dataguard i would like to increase redo logsize from 50M to 100M.
    on primary
    standby_file_management=manual
    added online redo groups with 100M
    log switched
    drop old one and readded with 100m
    deleted log added in step2.
    same for standby redo logs.
    On standby
    was able to resize standby redo logs.
    but cannot resize online redologs status is clearing or clearing_current.
    please comment. thanks.

    I assume you just had to wait until the Primary switched out of that online log so it became inactive at the standby as well? We track where the Primary is by marking the online redo log files at the standby as clearing_current so you can tell where the primary was at any given moment.
    Make sure you create new standby redo log files at the Primary and Standby to match the new online redo log file size.
    Larry

  • Use of standby redo log files in primary database

    Hi All,
    What is the exact use of setting up standby redo log files in the primary database on a data guard setup?
    any good documents?

    A standby redo log is required for the maximum protection and maximum availability modes and the LGWR ASYNC transport mode is recommended for all databases. Data Guard can recover and apply more redo data from a standby redo log than from archived redo log files alone.
    You should plan the standby redo log configuration and create all required log groups and group members when you create the standby database. For increased availability, consider multiplexing the standby redo log files, similar to the way that online redo log files are multiplexed.
    refer the link,and Perform the following steps to configure the standby redo log.:-
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/create_ps.htm#i1225703
    If the real-time apply feature is enabled, log apply services can apply redo data as it is received, without waiting for the current standby redo log file to be archived. This results in faster switchover and failover times because the standby redo log files have been applied already to the standby database by the time the failover or switchover begins.
    refer the link
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/log_apply.htm#i1023371

  • Differetn redo log names on primary db and standby db.

    we are using oracle 10.2.0.2 to run sap. we plan to use data guard.the os is suse linux 10.3.
    the archived redo log name's prefix is '1_' like '1_167109_678247864.dbf' on primary db.
    if i recover standy database mannuly with cmd 'ALTER DATABASE RECOVER automatic standby database' to fix redo log gap, the required archived redo log name's prefix is 'PRDarch1_' like 'PRDarch1_167260_678247864.dbf'.
    i checked the oracle parameter 'LOG_ARCHIVE_FORMAT' on both side were set to '%t_%s_%r.dbf'.
    my question is the different name have any effect on datagurad? Or we just can ignore it, because it doesn't matter.
    thanks in advance.

    I'm not interested in a debate so much as finding the answer. Thanks for the info. And welcome to the forum by the way.
    That said I would think B14239-05 would answer your question.
    In section 5.7.1 it shows:
    LOG_ARCHIVE_FORMAT=log%t_%s_%r.arc
    You may be correct on the default, but they only answer I can give you is I have never setup Data Guard that way. I hope this saves you some time.
    Where did you find this? (log_archive_dest_1 = "LOCATION=/oracle/PRD/oraarch/PRDarch MANDATORY REOPEN",)
    I would consider avoiding the use of MANDATORY. This attribute will cause you a world of trouble if your Standby is unreachable.
    Also if you are using a Flash Recovery Area you do not need to setup a local archiving destination.
    If you are not using Flash Recovery with Data Guard you should. Try testing a failover without it.
    Best Regards
    mseberg

  • Unable to copy redo-logs for cloning

    Hello
    We are using R12 with 10.2.0.2 db
    we take db tier backups every night and apps backup twice a month(15 days interval)
    We normally run autocfg.sh and adprecolne.pl on dbTier.After this we stop all services, shutdown the database and copy all the datafiles,controlfiles and logfiles)
    This backup is helpful in restoration in case any failure on PROD server and for cloning purpose on an another node for testing purpose
    Now my question is, are we doing the clone procedure properly or is there a better way for cloning for suppose RMAN..or maybe any other
    Secondly, Last night all the datafiles got copied after a clean shutdown except for logfiles.
    Will i be able to restore the data later if i dont have the log files? can i build new logfiles?
    Will this data be succesful for cloning without the redo logs
    Please Note: This is a cold backup we take i.e Consistent backup
    Thanks

    I would go against the oft-repeated and common "Expert" and "Documentation" advice here.
    If you are doing a COLD Backup and are are planning to use the Backup to restore to another server, there is no harm in taking the online redo logs as well.
    Where the risk with online redo log backups arises is when the SysAdmin/DBA , when doing a FullRestore also restores the online redo logs back to the target database overwriting any present online redo logs.
    For this to occcur :
    a. The SysAdmin/DBA happens to restore all the files in the backup set without selectively excluding the online redo logs
    b. The databases's online redo logs are present and good on disk
    c. The DBA isn't very confident of being able to rollforward through his redo logs
    It is for those reasons that it is easier to advice everyone not to backup online redo logs.
    However, as I've noted above, if you are doing a Cold Backup and restoring to another server (with no database present at the target site), there is no harm in taking the online redo logs. Such a backup and restore makes your scripting very easy and doesn't even need a DBA.
    Hemant K Chitale
    http://hemantoracledba.blogpsot.com
    Here are some of my comments on the issue of backups of online redo logs :
    http://hemantoracledba.blogspot.com/2008/03/backup-online-redo-logs.html

  • How to damage online redo log for simulation

    Dear All,
    Kindly please help to inform me, is there any way to damage online redo log from database level (not from OS command like dd) ?
    I need to do it for test case to enable db_block_checking and db_block_checksum (set both parameter to TRUE).
    Is those parameter will help from redo log corruption ? that's why, i want to prove it.
    Thanks
    Anthony

    user12215770 wrote:
    My purpose is i want to verify that the db_block_checking and db_block_checksum can avoid redo corruption (the corruption caused by process in the database).Redo corruption could also occur due to other issues as http://docs.oracle.com/cd/E11882_01/server.112/e25513/initparams049.htm#REFRN10030 says:
    >
    Checksums allow Oracle to detect corruption caused by underlying disks, storage systems, or I/O systems. If set to FULL, DB_BLOCK_CHECKSUM also catches in-memory corruptions and stops them from making it to the disk.
    >
    You could try to use ORADEBUG POKE command to write directly in the SGA if you know how to find log buffer blocks ... About ORADEBUG please read http://www.juliandyke.com/Diagnostics/Tools/ORADEBUG/ORADEBUG.html.

  • Is There a Way to Run a Redo log for a Single Tablespace?

    I'm still fairly new to Oracle. I've been reading up on the architecture and I am getting the hang of it. Actually, I have 2 questions.
    1) My first question is..."Is there a way to run the redo log file...but to specify something so that it only applies to a single tablespace and it's related files?"
    So, in a situation where, for some reason, only a single dbf file has become corrupted, I only have to worry about replaying the log for those transactions that affect the tablespace associated with that file.
    2) Also, I would like to know if there is a query I can run from iSQLPlus that would allow me to view the datafiles that are associated with a tablespace.
    Thanks

    1) My first question is..."Is there a way to run the
    redo log file...but to specify something so that it
    only applies to a single tablespace and it's related
    files?"
    No You can't specify a redolog file to record the transaction entries for a particular tablespace.
    In cas if a file gets corrupted.you need to apply all the archivelogs since the last backup plus the redologs to bring back the DB to consistent state.
    >
    2) Also, I would like to know if there is a query I
    can run from iSQLPlus that would allow me to view the
    datafiles that are associated with a tablespace.Select file_name,tablespace_name from dba_data_files will give you the
    The above will give you the number of datafiles that a tablespace is made of.
    In your case you have created the tablespace iwth one datafile.
    Message was edited by:
    Maran.E

  • Why not use Redo log for consistent read

    Oracle 11.1.0.7:
    This might be a stupid question.
    As I understand if a select was issued at 7:00 AM and the data that select is going to read has changed at 7:10 AM even then Oracle will return the data that existed at 7:00 AM. And for this Oracle needs the data in Undo segments.
    My question is since redo also has past and current information why can't redo logs be used to retreive that information? Why is undo required when redo already has all that information.

    user628400 wrote:
    Thanks. I get that piece but isn't it the same problem with UNDO? It's written as it expires and there is no guranteee until we specifically ask oracle to gurantee the UNDO retention? I guess I am trying to understand that UNDO was created for effeciency purposes so that there is less performance overhead as compared to reading and writing from redo.And this also you said,
    >
    If data was changed to 100 to 200 wouldn't both the values be there in redo logs. As I understand:
    1. Insert row with value 100 at 7:00 AM and commit. 100 will be writen to redo log
    2. update row to 200 at 8:00 AM and commit. 200 will be written to redo log
    So in essence 100 and 200 both are there in the redo logs and if select was issued at 7:00 data can be read from redo log too. Please correct me if I am understanding it incorrectly.I guess you didnt understand the explaination that I did. Its not the old data that is kept. Its the changed vector of Undo that is kept which is useful to "recover" it when its gone but not useful as such for a select statement. Whereas in an Undo block, the actual value is kept. You must remember that its still a block only which can contain data just like your normal block which may contain a table like EMP. So its not 100,200 but the change vectors of these things which is useful to recover the transaction based on their SCN numbers and would be read in that order as well. And to read the data from Undo, its quite simple for oracle to do so using an Undo block as the transaction table which holds the entry for the transaction, knows where the old data is kept in the Undo Segment. You may have seen XIDSEQ, XIDUSN, XIDSLOT in the tranaction id which are nothing but the information that where the undo data is kept. And to read it, unlke redo, undo plays a good role.
    About the expiry of Undo, you must know that only INACTIVE Undo extents are marked for expiry. The Active Extents which are having an ongoing tranaction records, are never marked for it. You can come back after a lifetime and if undo is there, your old data would be kept safe by oracle since its useful for the multiversioning. Undo Retention is to keep the old data after commit, something which you need not to do if you are on 11g and using Total Recall feature!
    HTH
    Aman....

  • Adding STAND BY REDO LOG in the Primary side........

    Hi All,
    I have set STANDBY_FILE_MANAGEMENT=AUTO in standby side and also LOG_FILE_NAME_CONVERT is pointing to an exiting location at OS level...
    When i added a datafile to an existing tablespace in the primary side and performed a log switch on the same . The added datafile got
    relflected in STANDBY side ...
    But when I am trying to add a standby logfile (new group) is not getting added .........
    Please help me on this...........
    For this will I need to perform the below steps .
    1-Add a standby redolog in the Primary side......
    2-then Cancel the managed recovery process from the standby side...
    3- Add the same standby redolog (same name and size) in the standby side
    4-put the standby in recovery mode ......
    Thanks

    Thanks for the reply ......
    I have gone thgrough it....but i donot find my answer .....
    My question is "Why *STANDBY_FILE_MANAGEMENT=AUTO* functionality is not adding a standby redolog in STANDBY SIDE automatically
    when i am adding a standby redolog in the PRIMARY SIDE.....
    Thanks.......

  • Log for "Transfer SQL Server Objects" (SSIS 2008)

    We are transferring several large tables (Data only) and the Transfer SQL Server Objects Control has remained yellow "state" for six hours. Is there a log file or other approach we can use tell what is happening? (For example, I found the
    process  spid and ran dbcc inputbuffer(<spid>) but the results are not too helpful.  Repeatedly running "sp_who2" show the process as suspended but the CPU\DiskIO numbers are constantly increasing. Go figure)
    TIA,
    edm2

    Use Adam's great procedure
    Who Is Active? v10.00 (2010-10-21)
    (C) 2007-2010, Adam Machanic
    Feedback: mailto:[email protected]
    Updates: http://sqlblog.com/blogs/adam_machanic/archive/tags/who+is+active/default.aspx
    "Beta" Builds: http://sqlblog.com/files/folders/beta/tags/who+is+active/default.aspx
    License: 
    Who is Active? is free to download and use for personal, educational, and internal 
    corporate purposes, provided that this header is preserved. Redistribution or sale 
    of Who is Active?, in whole or in part, is prohibited without the author's express 
    written consent.
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • ARD Reporting failing and causing thousands of entries in my system.log for my OSX Server clients

    I have a small four system network - all running latest version of Mavericks with latest (3.7.1 ARD)
    1) OSX Server running File Sharing, DNS, OpenDirectory, Web
    2) OSX Server running DNS (secondary), OpenDirectory (secondary)
    3) OSX Server - installed but not running any services
    4) OSX Client
    When using ARD from any of the systems, any attempt to generage an ARD report targeting system #1 or #2 - the ARD report task stops at "Collecting Report Data" and I start seeing:  sysinfocachegen[10506]: FSPathMakeRef error = -43 in the system.log that I'm targeting.  Currently over 11000 entries in my last log file (rolled over) and am over 8000 already in the log file after my last test. 
    If I target systems 3 or 4, reporting works fine. 
    Any ideas what is going on here?

    I'm getting this error too.
    20/5/2014 4:07:56.192 pm sysinfocachegen[27828]: FSPathMakeRef error = -43
    20/5/2014 4:07:56.192 pm sysinfocachegen[44825]: FSPathMakeRef error = -43
    20/5/2014 4:07:56.192 pm sysinfocachegen[63375]: FSPathMakeRef error = -43
    etc
    My OS is 10.9.3, ARD is 3.7.2 (372A19)
    I'm not trying to collect any report data as far as I know.
    I have a few ARD windows open.
    James.

  • SCSM 2012 - Adding Action Log to Microsoft SQL Server Report - Your Urgent help would be really appreciated

    Hopefully a straight forward yes or no question.
    Is it possible to add action logs to a report? either a report from a cube or a report from SQL Report Builder?
    My understanding is that the action log information is not transferred to the database, therefore I presume you cannot get this into a cube report of SQL report - is my understnading correct?
    tamrep

    Hi,
    Rob Ford made a contribution
    here for a similar task.
    Cheers,
    Marat
    Site: www.scutils.com  Twitter:
    LinkedIn:
    Graveyard:

  • Why do we need standby redo log on Primary database.

    Hi Gurus,
    I was going through the document in OBE,
    http://www.oracle.com/technology/obe/11gr1_db/ha/dataguard/physstby/physstdby.htm
    I have two queries:
    1) I noticed the statement -
    "Configure the primary database to receive redo data, by adding the standby logfiles to the primary. "
    Why do we have to create standby redo log on a primary database?
    2) There is another statement --
    "It is highly recommended that you have one more standby redo log group than you have online redo log groups as the primary database. The files must be the same size or larger than the primary database’s online redo logs. "
    Why do we need one additional standby redo log group than in Primary database.
    Could anyone please explain to me in simple words.
    Thanks
    Cherrish Vaidiyan

    Hi,
    1. Standby redo logs are used only when the database_role is standby, it is recommended to be added in primary also so that they can be used on role reversal, however during normal working standby redo logs will not be used at all on primary.
    2. In case of 3 online redo log groups, it is recommended to use 4 standby redo log group this is in case if log switching is happening frequently on primary and all 3 standby redo logs are still not completely archived on the standby and 4th can be used here as there will be some delay on standby due to network or slowness of arch on standby.
    Use of the standby redo log groups depends on the redo generation rate, you can see only 2 standby redo logs are getting used while you have 4 standby redo log groups, when the redo generation rate is less.
    So it is recommended to have one more standby redo log group when redo generation rate is high and all of the existing standby redo log group are getting used.
    Regards
    Anudeep

  • When changing redo-logs on Primary, do I need to change my standby also?

    Hi, I need to resize my redo logs on a primary db which has a physical standby attached. Do I need to modify this standby or will the change propagate. It seems unlikely to me.
    thanx

    Only clue that I have found is at
    http://download-east.oracle.com/docs/cd/B12037_01/server.101/b10726/configbp.htm
    Section "Use Multiplexed Standby Redo Logs and Configure Size Appropriately"
    >>
    The remote file server (RFS) process for the standby database writes only to an SRL whose size is identical to the size of an online redo log for the production database. If it cannot find an appropriately sized SRL, then RFS creates an archived redo log file directly instead and logs the following message in the alert log:
    No standby redo log files of size <#> blocks available.
    >>

  • Oracle DataGuard - Standy Redo log on physical Standby db

    Hello Guys,
    A quick question on my 10.2.0.2 db with Windows 2003 x64 OSs.
    I have 2 machines - One for primary & the other for physical standby
    I have successfully setup DG with Real Time Apply and also tested switchover & failover scenarios and they work well and as expected...
    But I have a query which popped up when I was trying to set this up at home.
    I create the physical standby db by shutting the primary db and copying all the datafiles, tempfiles,online redo logs and then creating a standby control file which will be used for the physical standby db to start off with.
    After mounting the standby db, I tried to create standby redo logs files starting with Group 4 but I got error that standby group 4 already exists.
    Upon querying v$logfile view, I noticed that the standby redo logs that I created on primary are also showing up on the standby db which I understand is from the standby control file.
    So here is my quesiton: What is correct method of creating these standby redo logs on the standby database?
    I know that I could drop those 4 standby redo logs from the standby db and recreate them but all the DG docs online and the documents that I have referred say that I should create the standby redo logs on the standby as I did on the primary but how can this duplication be avoided i.e. from the standby control file?
    I know that I could use another method to create a hot backup such as RMAN etc...but I wanted to follow this way of shutting down the primary and copy the relevant database files.
    Any help appreciated...and thanks in advance guys!
    -Bharath

    so with the setup that I used, i.e. create a standby control file which contains info about the standby redo logs at the primary site, should I also copy over the standby redo logs from the primary to standby as I only copied the datafiles, tempfile(s) & online redo logs? But then the filenames of the standby redo logs will be the exact same as that of the primary. Will that cause any issue during a switchover/failover?
    I don't remember exactly how I did it when I got the whole thing working but I have a vague remembrance that I drop the std. redo log files entries from the std database and then recreate new ones with different filenames (as compared to primary).
    Thanks

Maybe you are looking for

  • Is it possible to replace a processor in an ipod touch 5g?

    I was wondering, if the next ipod processor came out, like an A7 or higher? I know you cant upgrade your memory, but i have 32GB of it, so im good. I ahve an A5 and i just dont feel like it has enough power, so i want to upgrade. Help?

  • Exception -not-null property references a null or transient value

    HI, I am trying to save a relational table and it is throwing the following exception in Hibernate .Pls help me Caused by: org.hibernate.PropertyValueException: not-null property references a null or transient value: com.aexp.acqsys.sim.dm.DmPackageC

  • Recovery a password protect excel file.

    I have a password protected excel file and now it's corrupted and i can't open it. I tried to open_and_repair feature of excel2010 and different other stuffs, without success. I was wondering if there's a way to remove the password without opening th

  • Parsing Date and converting it toString()

    Hi, When I parse my date to a specific timezone, why does it not display that same timezone when I convert it to a string? Instead it seems to be using my system's current timezone, and I don't want that. This is how I'm doing it... SimpleDateFormat

  • App store in 10.6.8 beach ball when trying to install Mountain Lion

    Hi I'm trying to upgrade from 10.6.8, on my Mac book pro (early 2010) to Mountain lion. Problem is whenever I try, via App store, the beach ball appears and the App store becomes non responsive. Any suggestions? Thanks