Backup validation best practice  11GR2 on Windows

Hi all
I am just reading through some guides on checking for various types of corruption on my database. It seems that having DB_BLOCK_CHECKSUM set to TYPICAL takes care of much of the physical corruption and will alert you to the fact any has occurred. Furthermore RMAN by default does its own physical block checking. Logical corruption on the other hand does not seem to be checked automatically unless the CHECK LOGICAL is added to the RMAN command. There are also various VALIDATE commands that could be run on various objects.
My question is really, what is best practice for checking for block corruption. Do people even bother regularly checking this and just allow Oracle to manage itself, or is it best practice to have the CHECK LOGICAL command in RMAN (even though its not added by default when configuring backup jobs through OEM) or do people schedule jobs and output reports from a VALIDATE command on a regular basis?
Many thanks

To use CHECK LOGICAL clause is considered best practice at least by Oracle Support according to
NOTE:388422.1  Top 10 Backup and Recovery best practices
(referenced in http://blogs.oracle.com/db/entry/master_note_for_oracle_recovery_manager_rman_doc_id_11164841).

Similar Messages

  • Validation Best Practice for an javafx project

    Hello,
    I am investigating how to validate a JavaFX project and specifically where to display the validation results with appropriate icons and text. Do you have any thoughts on best practice for this? Are there any features in JavaFX to do this? Are there any third-party library that does this?
    Thanks

    See:
    Re: Forms and validations - here's some of my ideas, what are yours? "Thread: Forms and validations - here's some of my ideas, what are yours?"
    And messages in this mailing list thread:
    http://mail.openjdk.java.net/pipermail/openjfx-dev/2012-February/000717.html "Validate Me"

  • Best Practices question re Windows XP & Parallels 4.0 installation

    To Apple Gurus here:
    I am a new convert from Windows to Mac. Just bought a Macbook Pro (4G/320G, 2.4Gz) and a copy of Parallels 4.0. I have an OEM copy of Windows XP Pro & Photoshop CS4 for Windows. The question before me is what sequence should I go about installing Windows & Parallels. Logically, I think I should install:
    1) Windows XP using Boot Camp first,
    2) then install PhotoShop CS3 for Windows in the Windows partition
    3) then install MS Office
    4) and finally install Parallels 4.0.
    Is this the right sequence or indeed a "Best Practices" scenario?
    Any tips for a 'Best Practice' installation will be highly appreciated.
    Also, is anyone here using the SAP GUI for Mac OS-X & Citrix Presentation Server Client for Mac OS 10.0 (now renamed XenApp)?

    First, my creds. I don't consider myself an Apple guru. I have been running a MB since last December and at that time, I installed Parallels 3.0. If I remember correctly, after installing Parallels, I installed Windows Vista, and then Office and while I was impressed to be able to run MS Office on a MB, it took what I considered to be TOO long to load and then the performance was not that great. So, mostly I've stayed on the Mac side of the operation and only loaded Parallels if I had to run some MS program.
    About a week ago I got an offer from Parallels to buy 4.0 at an upgrade price of $40. I went with the box version since it was the same price as the download version. Tonight I got my courage up to do the upgrade. I was leery because I thought I might have to reinstall all my MS stuff (Office Pro, etc.) When I put the disk in to install the program, I receive a message saying there was a later edition available with the option to download it or install the box edition. After a few minutes of thought, I decided to do the download version. I would still recommend getting the box version since you get a manual with it although the download version comes with a PDF manual.
    When I finished, I then clicked on upgrade/install and the installation proceeded without much input from me. Lo and behold, the installation finished and it booted up to my previous Vista installation with all my programs intact.
    So far, I must stay I'm VERY impressed with this upgrade Parallels edition. It seems to load much faster, the programs are more responsive, Vista so far seems very stable and the ability to switch back and forth from Windows to OS X is totally better. From what I've seen so far, I would highly recommend anyone using Parallels 3.0 get this upgrade. While I've only been using it a few hours, it seems like the best upgrade for ANY program/system (Windows 95-->Vista) that I've ever done.
    A few months ago I saw a piece on an upgraded version of Fusion which stated that it moved Fusion ahead of Parallels. If that were so, I think the ball must be back in Parallels court with 4.0.

  • Validation best practices?

    I'm still working on learning Flex, having previously worked
    in VB, ASP and PHP. Today I was working on a small form that needed
    to be validated. My specs are that all fields are required, and the
    submit button is to remain disabled until all fields are valid.
    In the past I have used a simple summing function to see how
    many fields in a form were valid; if less than the total field
    count then validation failed.
    Here is what I wrote in my Flex project. Is there a more
    efficient method that I should be using? Execution seems to be
    instant, but if this does not conform to "best practices" I'd like
    to know the alternative. Thank you for your input.
    Paul

    To me, you might be able to use binding to just have the
    button enabled property bound to a Boolean variable.
    Also have an ArrayCollection named validAC.
    private function
    handlerValidation(event:ValidationResultEvent,
    validatorName:String):void{
    if(!validAC.contains(event.target)){
    validAC.addItem(event.target);
    submitBtn.enabled = numControls == (validAC.length);

  • Table Backup and Best Practice

    Hi Guys,
    We have 2 tables, master and child with millions of records in it.These tables gets populated from flat files that we receive from external systems.The major operation on this table are insert/delete(update). The way we do update is, if the record from flat file already exist in table we delete the matching record from master table & child table and re-insert the data from flat files.
    Business has decided to delete (archive) old and inactive data from these 2 tables. This process is run every year. I want to take a backup of this table by creating a similar table in database and start the delete process.
    What would be the best approach to take backup of table,because we run the archive process every year. Archive process here refers to physically deleting a record from a table.
    Any help is greatly appreciated.

    922855 wrote:
    Hi Guys,
    We have 2 tables, master and child with millions of records in it.These tables gets populated from flat files that we receive from external systems.The major operation on this table are insert/delete(update). The way we do update is, if the record from flat file already exist in table we delete the matching record from master table & child table and re-insert the data from flat files.
    Business has decided to delete (archive) old and inactive data from these 2 tables. This process is run every year. I want to take a backup of this table by creating a similar table in database and start the delete process.
    What would be the best approach to take backup of table,because we run the archive process every year. Archive process here refers to physically deleting a record from a table.
    Any help is greatly appreciated.expdp

  • Best practices for upgrading Windows Server Essentials 2012 to R2?

    Hi,
    I need to upgrade my Windows Server Essentials 2012 on an HP microserver to Essentials R2 (specifically the R2 role on Standard rather than as a VM). After a reasonable amount of searching, I haven't found anything online that specifically addresses this in-place
    upgrade scenario. Since this is already an Essentials server, it's the PDC in its own domain, etc. Has anyone that's done this upgrade have any experiences or warnings to share?
    Thanks,
    Sean

    Hello there,
    There is no upgrade path from Server 2012 Essentials to R2.  R2 is not an upgrade, it is a new build.  Essentials does not upgrade easily, you will want to perform a migration of the server.  This will require additional hardware.  You
    can possibly backup the 2012 Essentials, build a temp server on workstation hardware just to do the migration from,  restore the backup to the new server and then perform the migration to R2. 

  • Best Practices for Configuring Windows Server 2012 R2 for VM Performance

    Good afternoon,
    I've been told that VMWare has written a document or article on how to get the most performance out of an instance of Server 2012 R2 as a virtual machine. I was told this magical article has things like a checklist of features or services that can be turned off (such as tablet features) that will increase the performance/low overhead on the VM.
    Can anyone help me find such an article?
    Thanks,
    Johnathan

    Hello there,
    There is no upgrade path from Server 2012 Essentials to R2.  R2 is not an upgrade, it is a new build.  Essentials does not upgrade easily, you will want to perform a migration of the server.  This will require additional hardware.  You
    can possibly backup the 2012 Essentials, build a temp server on workstation hardware just to do the migration from,  restore the backup to the new server and then perform the migration to R2. 

  • _msdcs subdomain best practice with NS records?

    I have the _msdcs subfolder under my domain (the grey folder). example below
    It has only one DC inside of it for a NS server. This DC is old and no longer exists. I checked my test environment and it has the same scenario (an old DC that does that not exist). example below
    I'm just wondering:
    1) Is this normal, should this folder update itself with other servers?
    2) should I be adding one of my other DC's? and removing the original?
    I have a single forest, single domain setup 2008 functional level. My normal
    _msdcs Zone does behave as expected and removes and add the appropriate records. Thanks.

    I apologize for the late response. I see you've gone further than what I've recommended.
    No, you shouldn't have deleted the _msdc.parent.local zone!!!!!! I'm not sure why you did that. Are you working with someone else on this that recommended to do that? If not,
    you're over-thinking it. I provide specifics to fix it by simply  updating the NS records, that's it. If you only found the _msdcs folder had the wrong record, then that's all you had to change.
    In cases where DCs are removed, replaced, upgraded, etc, it's also best practice to check a few things to make sure things are in order, and one of them is check the NS records on all zones and delegations. Delegation's NS records won't update
    automatically with changes, but zone NS records will if DCs are properly demoted.
    The _msdcs delegated zone is required by Active Directory. And yes, based on your thread subject, it's best practice. When Windows 2000 came out, and IF you had created the initial domain with it, it did not have it this way, but all domains initially created
    with Windows 2003 and newer are designed this way. If you had upgraded from 2000 to 2003, then one of the steps that we must perform is to create the _msdcs delegation.
    Please re-create it in this order:
    In the DNS console, right-click Forward Lookup Zones, and then click
    New Zone. Click Next
    On the Zone Type page in the New Zone Wizard, click
    Primary zone, and then click to select the Store the zone in Active Directory check box. Click
    Next
    On the Active Directory Zone Replication Scope page, click "To all DNS servers in the Active Directory forest parent.local.
    On the Zone Name page, in the Zone Name box, type
    _msdcs.parent.local
    Complete the wizard by accepting all the default options.
    After you've done that:
    Delete the _msdcs subfolder under parent.local.
    Right-click parent.local, choose New Delegation.
    Type in _msdcs
    In the Nameserver page, type in the name of your server, and its IP address.
    Complete the wizard
    You should now see a grayed out _msdcs folder under parent.local.
    Go to c:\windows\system32\config\ folder
    Find netlogon.dns and rename it to netlogon.dns.old
    Find netlogon.dnb and rename it to netlogon.dnb.old
    Open a command prompt
    Run ipconfig /registerdns
    Run net stop netlogon
    Run net start netlogon
    Wait a few minutes, then click on the _msdcs.parent.local zone, and click the F5 button to refresh it.
    You should see the data populate.
    Ace Fekay
    MVP, MCT, MCITP/EA, MCTS Windows 2008/R2 & Exchange 2007, Exchange 2010 EA, MCSE & MCSA 2003/2000, MCSA Messaging 2003
    Microsoft Certified Trainer
    Microsoft MVP - Directory Services
    Technical Blogs & Videos: http://www.delawarecountycomputerconsulting.com/
    This post is provided AS-IS with no warranties or guarantees and confers no rights.

  • Best practice deploying additional updates

    Hello what is the best practice concerning monthy windows updates. We are currently adding additional windows updates to the existing 1 package and updating the content on the DP's. However this seems to work with inconsistant results.
    DPs are not finalising content .
    Other places I have worked we would create a seperate package each month for additional updates and never had an issue. Any thoughts?
    SCCM Deployment Technician

    The documented best practices are all related to the maximum number of patches that are part of one deployment. That number should not pas the 1000,
    Remember this is a hard limit of 1000 updates per Software Update Group (not deployment package). It's quite legitimate to use a single deployment package.
    I usually create static historical Software Updates Groups at a point in time (eg November 2014). In this case it is not possible to have a single SUG for all products (Windows 7 has over 600 updates for example). You have to split them. I deploy these
    updates (to pilot and production) and leave the deployments in place. Then I create an ADR which creates a new SUG each month and deploy (to pilot and production).
    You can use a single deployment package for all the above.
    Gerry Hampson | Blog:
    www.gerryhampsoncm.blogspot.ie | LinkedIn:
    Gerry Hampson | Twitter:
    @gerryhampson

  • Is there a list of best practices for Azure Cloud Services?

    Hi all;
    I was talking with a Sql Server expert today and learned that Azure Sql Server can take up to a minute to respond to a query that normally takes a fraction of a second. This is one of those things where it's really valuable to learn it when architecting as
    opposed to when we go live.
    Cloud Services are not Sql Server (obviously) but that led to the question - Is there a list of best practices for Azure Cloud Services? If so, what are they?
    We will be placing the cloud services in multiple datacenters and using traffic manager to point people to the right one. The cloud service will set between an IMAP client & server, pretending to be the mail client to the server, and the server to the client.
    Mostly it will pass all requests & responses across from one to the other.
    thanks - dave
    What we did for the last 6 months -
    Made the world's coolest reporting & docgen system even more amazing

    hi dave,
    >>Cloud Services are not Sql Server (obviously) but that led to the question - Is there a list of best practices for Azure Cloud Services? If so, what are they?
    For this issue, I have collected some blogs and document about best practices for azure cloud service, you can view them, but I am not sure they are your need.
    http://msdn.microsoft.com/en-us/library/azure/xx130451.aspx
    http://gauravmantri.com/2013/01/11/some-best-practices-for-building-windows-azure-cloud-applications/
    http://www.hanselman.com/blog/CloudPowerHowToScaleAzureWebsitesGloballyWithTrafficManager.aspx
    http://msdn.microsoft.com/en-us/library/azure/jj717232.aspxhttp://azure.microsoft.com/en-us/documentation/articles/best-practices-performance/
    >>The cloud service will set between an IMAP client & server, pretending to be the mail client to the server, and the server to the client. Mostly it will pass all requests & responses across from one to the other.
    For your scenarioes, If you'd like to communicate with each instances, I recommend you refer to this document (
    http://msdn.microsoft.com/en-us/library/azure/hh180158.aspx ). And generally, if we want connect the client to server on Azure, the service bus is a good choice (http://azure.microsoft.com/en-us/documentation/articles/cloud-services-dotnet-multi-tier-app-using-service-bus-queues/
    If I misunderstood, please let me know.
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Windows 2012 R2 File Server Cluster Storage best practice

    Hi Team,
    I am designing  Solution for 1700 VDi user's . I will use Microsoft Windows 2012 R2 Fileserver Cluster to host their Profile data by using Group Policy for Folder redirection.
    I am looking best practice to define Storage disk size for User profile data . I am looking to have Single disk size of 30 TB to host user Profile data .Single disk which will spread across two Disk enclosure .
    Please let me know if if single disk of 30 Tb can become any bottle neck to hold user active profile data .
    I have SSD Writable disk in storage with FC connectivity.
    Thanks
    Ravi

    Check this
    TechEd session,
    the
    Windows Server 2012 VDI deployment Guide (pages 8,9), and 
    this article
    General considerations during volume size planning:
    Consider how long it will take if you ever have to run chkdsk. Chkdsk has gone significant improvements in 2012 R2, but it will still take a long time to run against a 30TB volume.  That's down time..
    Consider how will volume size affect your RPO, RTO, DR, and SLA. It will take a long time to backup/restore a 30 TB volume. 
    Any operation on a 30TB volume like snapshot will pose performance and additional disk space challenges.
    For these reasons many IT pros choose to keep volume size under 2TB. In your case, you can use 15x 2TB volumes instead of a single 30 TB volume. 
    Sam Boutros, Senior Consultant, Software Logic, KOP, PA http://superwidgets.wordpress.com (Please take a moment to Vote as Helpful and/or Mark as Answer, where applicable) _________________________________________________________________________________
    Powershell: Learn it before it's an emergency http://technet.microsoft.com/en-us/scriptcenter/powershell.aspx http://technet.microsoft.com/en-us/scriptcenter/dd793612.aspx

  • E-business backup best practices

    What is E-business backup best practices, tools or techniques?
    For example what we do now is taking copy from the oracle folder in D partition every two days.
    but it take long time, and i believe there is a better way.
    We are on windows server 2003 and E-business 11i

    user11969921 wrote:
    What is E-business backup best practices, tools or techniques?
    For example what we do now is taking copy from the oracle folder in D partition every two days.
    but it take long time, and i believe there is a better way.
    We are on windows server 2003 and E-business 11i
    Please see previous threads for the same topic/discussion -- https://forums.oracle.com/search.jspa?view=content&resultTypes=&dateRange=all&q=backup+EBS&rankBy=relevance
    Please also see RMAN manuals (incremental backup, hot backup, ..etc) -- http://www.oracle.com/pls/db112/portal.portal_db?selected=4&frame=#backup_and_recovery
    Thanks,
    Hussein

  • Best Practices for patch/rollback on Windows?

    All,
    I have been working on BO XI with UNIX for some time now and while I am pretty comfortable with managing it on UNIX, I am not too sure about the "best practices" when it comes to Windows.
    I have a few specific questions:
    1) What is the best way to apply a patch or Service Pack to BO XI R2 on a Windows envt without risking a system corruption?
    - It is relatively easier on UNIX because you don't have to worry about registry entries and you can even perform multiple installations on the same box as long as you use different locations and ports.
    2) What should be the ideal "rollback" strategy in case an upgrade/patch install fails and corrupts the system?
    I am sure I will have some follow up questions, but if someone can get the discussion rolling with these for now, I would really appreciate!
    Is there any documentation available around these topics on the boards some place?
    Cheers,
    Sarang

    This is unofficial but usually if you run into a disabled system as a result of a patch and the removal/rollback does NOT work (in other words you are still down).
    You should have made complete backups of your FRS, CMS DB, and any customizations in your environment.
    Remove the base product and any seperate products that share registry keys (i.e. crystal reports)
    Remove the left over directories (XIR2 this is boinstall\business objects\*)
    Remove the primary registry keys (hkeylocalmachine\software\businessobjects\* & hkeycurrentuser\software\businessobjects\* )
    Remove any legacy keys (i.e. crystal*)
    Remove any patches from the registry (look in control panel and search for the full patch name)
    Then reinstall the product (test)
    add back any customizations
    reinstall either the latest(patch prior to update) or newest patch(if needed)
    and restore the FRS and CMS DB.
    There are a few modifications to these steps and you should leave room to add more (if they improve your odds at success).
    Regards,
    Tim

  • Unity Backup Best Practices

    We are looking at our Disaster Recovery when it comes to our Unity Voicemail servers and I am looking for some Best Practices when it comes to backing up Unity. Does anyone have any whitepapers/suggestions? Thanks.

    if you mean unity backup as in backing up the data, then the suggested method is four fold...
    1) using a backupExec or the likes, with a SQL agent, backup the SQL database(s) on unity.
    2) using a backupExec or the likes, with an Exchange agent, backup the Exchange database(s) used by unity.
    3) using a backupExec or the likes, perform a Windows 2000/2003 full backup including SystemState. (open file option is also available for backupExec)
    4) for DR recovery purposes, using DiRT, backup the unity system.
    see this link for more unity backup info:
    http://www.cisco.com/en/US/products/sw/voicesw/ps2237/products_white_paper09186a00801f4ba7.shtml
    if you mean backing up unity as far as failover, then you can use the link that follows to create a failover environment for your unity:
    http://www.cisco.com/en/US/products/sw/voicesw/ps2237/prod_installation_guides_list.html

  • (Request for:) Best practices for setting up a new Windows Server 2012 r2 Hyper-V Virtualized AD DC

    Could you please share your best practices for setting up a new Windows Server 2012 r2 Hyper-V Virtualized AD DC, that will be running on a new WinSrv 2012 r2 host server.   (This
    will be for a brand new network setup, new forest, domain, etc.)
    Specifically, your best practices regarding:
    the sizing of non virtual and virtual volumes/partitions/drives,  
    the use of sysvol, logs, & data volumes/drives on hosts & guests,
    RAID levels for the host and the guest(s),  
    IDE vs SCSI and drivers both non virtual and virtual and the booting there of,  
    disk caching settings on both host and guests.  
    Thanks so much for any information you can share.

    A bit of non essential additional info:
    We are small to midrange school district who, after close to 20 years on Novell networks, have decided to design and create a new Microsoft network and migrate all of our data and services
    over to the new infrastructure .   We are planning on rolling out 2012 r2 servers with as much Hyper-v virtualization as possible.
    During the last few weeks we have been able to find most of the information we need to undergo this project, and most of the information was pretty solid with little ambiguity, except for
    information regarding virtualizing the DCs, which as been a bit inconsistent.
    Yes, we have read all the documents that most of these posts tend point to, but found some, if not most are still are referring to performing this under Srvr 2008 r2, and haven’t really
    seen all that much on Srvr2012 r2.
    We have read these and others:
    Introduction to Active Directory Domain Services (AD DS) Virtualization (Level 100), 
    Virtualized Domain Controller Technical Reference (Level 300),
    Virtualized Domain Controller Cloning Test Guidance for Application Vendors,
    Support for using Hyper-V Replica for virtualized domain controllers.
    Again, thanks for any information, best practices, cookie cutter or otherwise that you can share.
    Chas.

Maybe you are looking for