Backup blocks all check logical validate database

Hello,
From Metalink : How To Use RMAN To Check For Logical & Physical Database Corruption [ID 283053.1]
The following example shows how to validate all datafiles:
run {
allocate channel d1 type disk;
backup blocks all check logical validate database;
release channel d1;
Does this following command just check for corruption , or is it actually trying to backup the datafiles to the disks ?
Thanks

Hello,
The BACKUP VALIDATE statement is used for checking Block Corruption not for running a Backup.
You'll have more detail on the following links:
http://download.oracle.com/docs/cd/B28359_01/backup.111/b28270/rcmvalid.htm#CHDECGBJ
http://download.oracle.com/docs/cd/B28359_01/backup.111/b28270/rcmvalid.htm#CHDCEHFD
Hope this help.
Best regards,
Jean-Valentin

Similar Messages

  • How to delete the foreign archivelogs in a Logical Standby database

    How do I remove the foreign archive logs that are being sent to my logical standby database. I have files in the FRA of ASM going back weeks ago. I thought RMAN would delete them.
    I am doing hot backups of the databases to FRA for both databases. Using ASM, FRA, in a Data Guard environment.
    I am not backing up anything to tape yet.
    The ASM FRA foreign_archivelog directory on the logical standby FRA keeps growing and nothing is get deleted when
    I run the following command every day.
    delete expired backup;
    delete noprompt force obsolete;
    Primary database RMAN settings (Not all of them)
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 9 DAYS;
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE DB_UNIQUE_NAME 'WMRTPRD' CONNECT IDENTIFIER 'WMRTPRD_CWY';
    CONFIGURE DB_UNIQUE_NAME 'WMRTPRD2' CONNECT IDENTIFIER 'WMRTPRD2_CWY';
    CONFIGURE DB_UNIQUE_NAME 'WMRTPRD3' CONNECT IDENTIFIER 'WMRTPRD3_DG';
    CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON ALL STANDBY;
    Logical standby database RMAN setting (not all of them)
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 9 DAYS;
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
    How do I cleanup/delete the old ASM foreign_archivelog files?

    OK, the default is TRUE which is what it is now
    from DBA_LOGSTDBY_PARAMETERS
    LOG_AUTO_DELETE     TRUE          SYSTEM     YES
    I am not talking about deleting the Archive logs files for the Logical database that it is creating, but the Standby archive log files being sent to the Logical Database after they have been applied.
    They are in the alert log as follows under RFS LogMiner: Registered logfile
    RFS[1]: Selected log 4 for thread 1 sequence 159 dbid -86802306 branch 763744382
    Thu Jan 12 15:44:57 2012
    *RFS LogMiner: Registered logfile [+FRA/wmrtprd2/foreign_archivelog/wmrtprd/2012_01_12/thread_1_seq_158.322.772386297] to LogM*
    iner session id [1]
    Thu Jan 12 15:44:58 2012
    LOGMINER: Alternate logfile found. Transition to mining archived logfile for session 1 thread 1 sequence 158, +FRA/wmrtprd2/
    foreign_archivelog/wmrtprd/2012_01_12/thread_1_seq_158.322.772386297
    LOGMINER: End mining logfile for session 1 thread 1 sequence 158, +FRA/wmrtprd2/foreign_archivelog/wmrtprd/2012_01_12/threa
    d_1_seq_158.322.772386297
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 159, +DG1/wmrtprd2/onlinelog/group_4.284.771760923                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Backup validate check logical database

    What is exactly the following RMAN command do? I want to know Is it doing a full backup or not?
    rman> backup validate check logical database' . Kindly help me.

    RMAN does not physically backup the database with this command. But it reads all blocks and checks for corruptions.
    If it finds corrupted blocks it will place the information about the corruption into a view:
    v$database_block_corruption;
    Now we can tell RMAN to recover all the blocks which it has found as being corrupt:
    RMAN> blockrecover corruption list; # (all blocks from v$database_block_corruption)
    Use below link for reference.
    http://luhartma.blogspot.com/2006/04/how-to-check-for-and-repair-block.html
    -Bharath

  • BACKUP VALIDATE vs VALIDATE in checking logical/physical corruption

    Hello all,
    I am checking if our 10gR2 database has any physical or logical corruption. I have read in some places where they state that VALIDATE command is enough to check database for physical corruption. Our database was never backed up by RMAN specifically before. Are any configuration settings needed for running BACKUP VALIDATE command? The reason I am asking is because just the VALIDATE command returns an error and BACKUP VALIDATE command runs without error but it is not showing the
    "File Status Marked Corrupt Empty Blocks Blocks Examined High SCN" lines.
    I used the command in two different formats and both do not show individual data file statuses:
    RMAN> run {
    CONFIGURE DEFAULT DEVICE TYPE TO DISK;
    CONFIGURE DEVICE TYPE DISK PARALLELISM 10 BACKUP TYPE TO BACKUPSET;
    BACKUP VALIDATE CHECK LOGICAL DATABASE FILESPERSET=10;
    RMAN> BACKUP VALIDATE CHECK LOGICAL DATABASE
    RMAN> VALIDATE DATABASE;
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-00558: error encountered while parsing input commands
    RMAN-01009: syntax error: found "database": expecting one of: "backupset"
    RMAN-01007: at line 1 column 10 file: standard input
    However on a different database already being backed up by RMAN daily, BACKUP VALIDATE output shows list of datafiles and STATUS = OK as below:
    List of Datafiles
    =================
    File Status Marked Corrupt Empty Blocks Blocks Examined High SCN
    How can we check every individual datafile status. Appreciate your responses. Thanks.

    Hi,
    After you have run:
    BACKUP VALIDATE CHECK LOGICAL DATABASE You can use sqlplus and run:
    select * from v$database_block_corruption.The output will tell you which block in which datafile is corrupt.
    Regards,
    Tycho
    Edited by: tychos on 8-sep-2011 18:34

  • Backup validate check logical

    hi,
    An mssql dba asked me if oracle could do logical/physical corruption checking when performing backups. Im aware of BACKUP VALIDATE CHECK LOGICAL.
    However, it doesn't create any backup. It just checks for corruption. If I remember, you can run BACKUP CHECK LOGICAL which will actually create the backup and check logical at the same time, but my question is...
    Why would you not want one or both of these as standard? I get it that there's overhead, more so for VALIDATE, but would you really want backups that contain corruption? Even if you keep 30 days worth of backups, you could realise after 40 days that you had corruption, and not be able to restore.
    Or is the reasoning behind not making it default, that you've got 'enough' backups that even if a datafile gets logical/physical corruption, you're bound to discover it before your backups become obsolete, and that it's then not worth the overhead of making it default?

    You can have the settings with SET MAXCORRUPT FOR DATAFILE.
    By default a checksum is calculated for every block read from a datafile and stored in the backup or image copy. If you use the NOCHECKSUM option, then checksums are not calculated. If the block already contains a checksum, however, then the checksum is validated and stored in the backup. If the validation fails, then the block is marked corrupt in the backup.
    The SET MAXCORRUPT FOR DATAFILE command sets how many corrupt blocks in a datafile that BACKUP will tolerate. If a datafile has more corrupt blocks than specified by the MAXCORRUPT parameter, the command terminates. If you specify the CHECK LOGICAL option, RMAN checks for logical and physical corruption.
    By default, the BACKUP command terminates when it cannot access a datafile. You can specify parameters to prevent termination, as listed in the following table.
    If you specify the option ... Then RMAN skips...
    SKIP INACCESSIBLE Inaccessible datafiles. A datafile is only considered inaccessible if it cannot be read. Some offline datafiles can still be read because they exist on disk. Others have been deleted or moved and so cannot be read, making them inaccessible.
    SKIP OFFLINE Offline datafiles.
    SKIP READONLY Datafiles in read-only tablespaces.

  • Restore Validate Database vs Backup Validate Database

    Hi,
    I want to add the following sentence after taking the backup using following command (backup as compressed backupset database plus archivelog;)
    restore database validate;
    or
    backup validate database;
    which one is best?

    It depends on what you are trying to accomplish, as per documentation:
    http://download.oracle.com/docs/cd/B19306_01/backup.102/b14194/rcmsynta009.htm#sthref141BACKUP .. VALIDATE: Causes RMAN to scan the specified files and verify their contents, testing whether this file can be backed up. RMAN creates no output files. Use this command periodically to check for physical and logical errors in database files.
    http://download.oracle.com/docs/cd/B19306_01/backup.102/b14194/rcmsynta051.htm#sthref796
    RESTORE... VALIDATE causes RMAN to select existing backups that it would use to perform a RESTORE operation, and scan them all to ensure that they are present and free of corruption.>
    Guessing what you're trying to do, I recommend:
    backup check logical as compressed backupset database plus archivelog;
    restore validate database;
    CHECK LOGICAL. Tests data and index blocks that pass physical corruption checks for logical corruption, for example, corruption of a row piece or index entry. If RMAN finds logical corruption, then it logs the block in the alert.log and server session trace file. HTH
    Enrique
    PS If you have access to Metalink see Note:388422.1 Top 10 Backup and Recovery best practices.
    Edited by: Enrique Orbegozo on Dec 17, 2008 4:21 PM

  • Oracle EBS R12 - 11.2.0.2 - Logical Standby Database & RMAN backups

    Hi all,
    I have Oracle EBusiness Suite R12 running on 11.2.0.2. Our organization used to have most of the business in North America, South America and Europe; but now we are also in Australia, China and Indonesia. Previously running large reporting batch jobs and RMAN backups at night time in USA was feasible but now that is beginning to show up as performance issue for some countries.
    I was thinking would it make sense to have a Active Data Guard logical standby database on a separate server in the same US based data center where we have our main servers and use that logical standy database for reporting and RMAN backups. I am especially concerned about RMAN backups. Would I be able to make RMAN backups from the logical standy database and somehow use those to recover our primary database if some issue happens?
    Any thoughts/suggestions?
    Thanks all,

    Active Data Guard only deals with physical standbys. If you are using a logical standby you can always open the database without using Active Data Guard option.
    http://docs.oracle.com/cd/E11882_01/server.112/e25608/rman.htm#sthref846 says that:
    >
    Backups of logical standby databases are not usable at the primary database.

  • (Cisco Historical Reporting / HRC ) All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054

    Hi All,
    I am getting an error message "All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054"  when trying to log into HRC (This user has the reporting capabilities) . I checked the log files this is what i found out 
    The log file stated that there were ongoing connections of HRC with the CCX  (I am sure there isn't any active login to HRC)
    || When you tried to login the following error was being displayed because the maximum number of connections were reached for the server .  We can see that a total number of 5 connections have been configured . ||
    1: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Current number of connections (5) from historical Clients/Scheduler to 'CRA_DATABASE' database exceeded the maximum number of possible connections (5).Check with your administrator about changing this limit on server (wfengine.properties), however this might impact server performance.
    || Below we can see all 5 connections being used up . ||
    2: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:[DB Connections From Clients (count=5)]|[(#1) 'username'='uccxhrc','hostname'='3SK5FS1.ucsfmedicalcenter.org']|[(#2) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#3) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#4) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#5) 'username'='uccxhrc','hostname'='47BMMM1.ucsfmedicalcenter.org']
    || Once the maximum number of connection was reached it threw an error . ||
    3: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Number of max connection to 'CRA_DATABASE' database was reached! Connection could not be established.
    4: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Database connection to 'CRA_DATABASE' failed due to (All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054.)
    Current exact UCCX Version 9.0.2.11001-24
    Current CUCM Version 8.6.2.23900-10
    Business impact  Not Critical
    Exact error message  All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054
    What is the OS version of the PC you are running  and is it physical machine or virtual machine that is running the HRC client ..
    OS Version Windows 7 Home Premium  64 bit and it’s a physical machine.
    . The Max DB Connections for Report Client Sessions is set to 5 for each servers (There are two servers). The no of HR Sessions is set to 10.
    I wanted to know if there is a way to find the HRC sessions active now and terminate the one or more or all of that sessions from the server end ? 

    We have had this "PRX5" problem with Exchange 2013 since the RTM version.  We recently applied CU3, and it did not correct the problem.  We have seen this problem on every Exchange 2013 we manage.  They are all installations where all roles
    are installed on the same Windows server, and in our case, they are all Windows virtual machines using Windows 2012 Hyper-V.
    We have tried all the "this fixed it for me" solutions regarding DNS, network cards, host file entries and so forth.  None of those "solutions" made any difference whatsoever.  The occurrence of the temporary error PRX5 seems totally random. 
    About 2 out of 20 incoming mail test by Microsoft Connectivity Analyzer fail with this PRX5 error.
    Most people don't ever notice the issue because remote mail servers retry the connection later.  However, telephone voice mail systems that forward voice message files to email, or other such applications such as your scanner, often don't retry and
    simply fail.  Our phone system actually disables all further attempts to send voice mail to a particular user if the PRX5 error is returned when the email is sent by the phone system.
    Is Microsoft totally oblivious to this problem?
    PRX5 is a serious issue that needs an Exchange team resolution, or at least an acknowledgement that the problem actually does exist and has negative consequences for proper mail flow.
    JSB

  • I did a backup from all my files saved in old external hd to time capsule. After that double checked and recover one file from TC to mac air and it worked just perfect. Today went through TC to recover another file and none of them were there?

    I did a backup from all my files saved in old external hd to time capsule. After that double checked and recover one file from TC to mac air and it worked just perfect. Today went through TC to recover another file and none of them were there?Anybody has a clue what happened?

    renatocremonese wrote:
    I want to use it for backing up my Mac.
    It's good for that . . . 
    But also I don't keep all my stuff in my Mac.
    But not for that. 
    This older and not day-by-day usage files I want to store in the time machine.
    You can do that (see below), but how are you going to back them up?   If your only copies are on the TC, when (not if) it fails, you risk losing your only copy of the data.
    Is it possible to split TC in two partitions.
    No, but there are some workarounds, including making a fixed-size disk image on it to "reserve" some space.  See #Q3 in Using Time Machine with a Time Capsule.
    But you still won't have backups of the stuff you put there.
    You don't say what kind of Mac you have.  If it's a desktop model, just keep the external HD connected to it, and let Time Machine back it up along with your internal HD.
    However, it sounds like you may have a laptop, where that's not going to be convenient.  In that case, your best bet might be to copy the external HD to a disk image on the TC as above, then keep the HD in a safe place.
    To finish, when i enter the TC and go through the Time line how can I get a file from there and move it back to Mac hd.
    Via the "Star Wars" display, per #15 in Time Machine - Frequently Asked Questions.
    You might also want to review the Time Machine Tutorial, and perhaps browse the rest of the FAQ.

  • Do I need put VALIDATE DATABASE in RMAN? I have crosscheck backup

    Does the Crosscheck backup tell if a database has corruption? do I need run command validate database?

    Hello,
    This is the Berkeley DB forum. One of the forums found via the "Forum Home" link in the upper left of this page might be more helpful.
    Thanks,
    Sandra

  • MV Logs not getting purged in a Logical Standby Database

    We are trying to replicate a few tables in a logical standby database to another database. Both the source ( The Logical Standby) and the target database are in Oracle 11g R1.
    The materialized views are refreshed using FAST REFRESH.
    The Materialized View Logs created on the source ( the Logical Standby Database) are not getting purged when the MV in the target database is refreshed.
    We checked the entries in the following Tables: SYS.SNAP$, SYS.SLOG$, SYS.MLOG$
    When a materialized view is created on the target database, a record is not inserted into the SYS.SLOG$ table and it seems like that's why the MV Logs are not getting purged.
    Why are we using a Logical Standby Database instead of the Primary ? Because, the load on the Primary Database is too much and the machine doesn't have enough resources to support MV based replication. The CPU usage is 95% all the time. The appplication owner won't allow us to go against the Primary database.
    Do we have to do anything different in terms of Configuration/Privileges etc. because we are using a Logical Standby Database as a source ?
    Thanks in Advance.

    We have a 11g RAC database in solaris OS where there is huge gap in archive log apply.
    Thread Last Sequence Received Last Sequence Applied Difference
    1 132581 129916 2665
    2 108253 106229 2024
    3 107452 104975 2477
    The MRP0 process seems not to be working also.Almost 7000+ archives lag in standby if compared with primary database.
    i suggest you to go with Incremental rollforward backups to make it SYNC, use this below link for step by step procedure.
    http://www.oracle-ckpt.com/rman-incremental-backups-to-roll-forward-a-physical-standby-database-2/
    Here questions.
    1) Whether those archives are transported & just not applied?
    2) Is in production do you have archives or backup of archives?
    3) What you have found errors in alert log file?
    post
    SQL> select severity,message,error_code,timestamp from v$dataguard_status where dest_id=2;
    4) What errors in primary database alert log file?
    Also post
    select     ds.dest_id id
    ,     ad.status
    ,     ds.database_mode db_mode
    ,     ad.archiver type
    ,     ds.recovery_mode
    ,     ds.protection_mode
    ,     ds.standby_logfile_count "SRLs"
    ,     ds.standby_logfile_active active
    ,     ds.archived_seq#
    from     v$archive_dest_status     ds
    ,     v$archive_dest          ad
    where     ds.dest_id = ad.dest_id
    and     ad.status != 'INACTIVE'
    order by
         ds.dest_id
    /Also check errors from standby database.

  • Backup-GPO -All, can this be filtered?

    Hi guys,
    I'm using the following to backup all the GPO's in the domain:
    $Backups = Backup-GPO -All -Path $SubBackupFolder -Domain $DomainFQDN -Comment "Add comment here"
    However, I do not want to backup all the GPO's. I want to backup only about 20 Gpo's found inside 2 OU's. Is this achievable via backup-gpo? Rather than the -All parameter can I instead say -All that are inside these 2 OU's. All the GPO's I want to backup
    contain the words "win7" somewhere within the GPO name if that helps.
    I'm using Ian's script "Comprehensive Group Policy Backup Script" that performs a backup and there's a script that imports the exported data as well. Keeps links, gpo settings, delegation, etc etc..
    Thank you so much for your help.

    Hi Mike,
    Thanks for the response. I think I understand what you're saying, I now need to find out how to do it. I'm a complete scripting beginner. I can read scripts but cant write, I'll read up on filtering with Where and piping. BTW here is the complete script
    I'm working with:
    <#
    .SYNOPSIS
    Backs up all GPOs from a specified domain and includes additional GPO information.
    .DESCRIPTION
    The script backs up all GPOs in a target domain and captures additional GPO management information, such
    as Scope of Management, Block Inheritance, Link Enabled, Link Order, Link Enforced and WMI Filters.
    The backup can then be used by a partner script to mirror GPOs in a test domain.
    Details:
    * Creates a XML file containing PSCustomObjects used by partner import script
    * Creates a XML file WMI filter details used by partner import script
    * Creates a CSV file of additional information for readability
    * Additional backup information includes SOM (Scope of Management) Path, Block Inheritance, Link Enabled,
    Link Order', Link Enforced and WMI Filter data
    * Each CSV SOM entry is made up of "DistinguishedName:BlockInheritance:LinkEnabled:LinkOrder:LinkEnforced"
    * Option to create a Migration Table (to then be manually updated)
    Requirements:
    * PowerShell GroupPolicy Module
    * PowerShell ActiveDirectory Module
    * Group Policy Management Console
    .EXAMPLE
    .\BackUp_GPOs.ps1 -Domain wintiptoys.com -BackupFolder "\\wingdc01\backups\"
    This will backup all GPOs in the domain wingtiptoys.com and store them in a date and time stamped folder
    under \\wingdc01\backups\.
    .EXAMPLE
    .\BackUp_GPOs.ps1 -Domain contoso.com -BackupFolder "c:\backups" -MigTable
    This will backup all GPOs in the domain contoso.com and store them in a date and time stamped folder
    under c:\backups\. A migration table, MigrationTable.migtable, will also be created for manual editing.
    .OUTPUTS
    Backup folder name in the format Year_Month_Day_HourMinuteSecond
    GpoDetails.xml
    WmiFilters.xml
    GpoInformation.csv
    MigrationTable.migtable (optional)
    EXIT CODES: 1 - GPMC not found
    .NOTES
    THIS CODE-SAMPLE IS PROVIDED "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED
    OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE IMPLIED WARRANTIES OF MERCHANTABILITY AND/OR
    FITNESS FOR A PARTICULAR PURPOSE.
    This sample is not supported under any Microsoft standard support program or service.
    The script is provided AS IS without warranty of any kind. Microsoft further disclaims all
    implied warranties including, without limitation, any implied warranties of merchantability
    or of fitness for a particular purpose. The entire risk arising out of the use or performance
    of the sample and documentation remains with you. In no event shall Microsoft, its authors,
    or anyone else involved in the creation, production, or delivery of the script be liable for
    any damages whatsoever (including, without limitation, damages for loss of business profits,
    business interruption, loss of business information, or other pecuniary loss) arising out of
    the use of or inability to use the sample or documentation, even if Microsoft has been advised
    of the possibility of such damages, rising out of the use of or inability to use the sample script,
    even if Microsoft has been advised of the possibility of such damages.
    #>
    ## Script Options and Parameters
    #Requires -version 3
    #Requires -modules ActiveDirectory,GroupPolicy
    #Define and validate parameters
    [CmdletBinding()]
    Param(
    #The target domain
    [parameter(Mandatory=$True,Position=1)]
    [ValidateScript({Get-ADDomain $_})]
    [String]$Domain,
    #The backup folder
    [parameter(Mandatory=$True,Position=2)]
    [ValidateScript({Test-Path $_})]
    [String]$BackupFolder,
    #Whether to create a migration table
    [Switch]
    $MigTable
    #Set strict mode to identify typographical errors (uncomment whilst editing script)
    #Set-StrictMode -version Latest
    ## Main
    ##BACKUP FOLDER DETAILS
    #Create a variable to represent a new backup folder
    #(constructing the report name from date details and the supplied backup folder)
    $Date = Get-Date
    $SubBackupFolder = "$BackupFolder\" + `
    "$($Date.Year)_" + `
    "$("{0:D2}" -f $Date.Month)_" + `
    "$("{0:D2}" -f $Date.Day)_" + `
    "$("{0:D2}" -f $Date.Hour)" + `
    "$("{0:D2}" -f $Date.Minute)" + `
    "$("{0:D2}" -f $Date.Second)"
    ##BACKUP ALL GPOs
    #Create the backup folder
    New-Item -ItemType Directory -Path $SubBackupFolder | Out-Null
    #Make sure the backup folder has been created
    If (Test-Path -Path $SubBackupFolder) {
    #Connect to the supplied domain
    $TargetDomain = Get-ADDomain $Domain
    #Obtain the domain FQDN
    $DomainFQDN = $TargetDomain.DNSRoot
    #Obtain the domain DN
    $DomainDN = $TargetDomain.DistinguishedName
    #Backup all the GPOs found in the domain
    $Backups = Backup-GPO -All -Path $SubBackupFolder -Domain $DomainFQDN -Comment "Scripted backup created by $env:userdomain\$env:username on $(Get-Date -format d)"
    #Instantiate an object for Group Policy Management (GPMC required)
    Try {
    $GPM = New-Object -ComObject GPMgmt.GPM
    } #End of Try...
    Catch {
    #Display exit message to console
    $Message = "ERROR: Unable to connect to GPMC. Please check that it is installed."
    Write-Host
    Write-Error $Message
    #Exit the script
    Exit 1
    } #End of Catch...
    #Import the GPM API constants
    $Constants = $GPM.getConstants()
    #Connect to the supplied domain
    $GpmDomain = $GPM.GetDomain($DomainFQDN,$Null,$Constants.UseAnyDc)
    ##COLLECT SPECIFIC GPO INFORMATION
    #Loop through each backed-up GPO
    ForEach ($Backup in $Backups) {
    #Get the GPO GUID for our target GPO
    $GpoGuid = $Backup.GpoId
    #Get the backup GUID for our target GPO
    $BackupGuid = $Backup.Id
    #Instantiate an object for the relevant GPO using GPM
    $GPO = $GpmDomain.GetGPO("{$GpoGuid}")
    #Get the GPO DisplayName property
    $GpoName = $GPO.DisplayName
    ##Retrieve SOM Information
    #Create a GPM search criteria object
    $GpmSearchCriteria = $GPM.CreateSearchCriteria()
    #Configure search critera for SOM links against a GPO
    $GpmSearchCriteria.Add($Constants.SearchPropertySOMLinks,$Constants.SearchOpContains,$GPO)
    #Perform the search
    $SOMs = $GpmDomain.SearchSOMs($GpmSearchCriteria)
    #Empty the SomPath variable
    $SomInfo = $Null
    #Loop through any SOMs returned and write them to a variable
    ForEach ($SOM in $SOMs) {
    #Capture the SOM Distinguished Name
    $SomDN = $SOM.Path
    #Capture Block Inheritance state
    $SomInheritance = $SOM.GPOInheritanceBlocked
    #Get GPO Link information for the SOM
    $GpoLinks = (Get-GPInheritance -Target $SomDN).GpoLinks
    #Loop through the GPO Link information and match info that relates to our current GPO
    ForEach ($GpoLink in $GpoLinks) {
    If ($GpoLink.DisplayName -eq $GpoName) {
    #Capture the GPO link status
    $LinkEnabled = $GpoLink.Enabled
    #Capture the GPO precedence order
    $LinkOrder = $GpoLink.Order
    #Capture Enforced state
    $LinkEnforced = $GpoLink.Enforced
    } #End of If ($GpoLink.DisplayName -eq $GpoName)
    } #End of ForEach ($GpoLink in $GpoLinks)
    #Append the SOM DN, link status, link order and Block Inheritance info to $SomInfo
    [Array]$SomInfo += "$SomDN`:$SomInheritance`:$LinkEnabled`:$LinkOrder`:$LinkEnforced"
    } #End of ForEach ($SOM in $SOMs)...
    ##Obtain WMI Filter path using Get-GPO
    $WmiFilter = (Get-GPO -Guid $GpoGuid).WMiFilter.Path
    #Split the value down and use the ID portion of the array
    $WMiFilter = ($WmiFilter -split "`"")[1]
    #Add selected GPO properties to a custom GPO object
    $GpoInfo = [PSCustomObject]@{
    BackupGuid = $BackupGuid
    Name = $GpoName
    GpoGuid = $GpoGuid
    SOMs = $SomInfo
    DomainDN = $DomainDN
    WmiFilter = $WmiFilter
    } #End of $Properties...
    #Add our new object to an array
    [Array]$TotalGPOs += $GpoInfo
    } #End of ForEach ($Backup in $Backups)...
    ##BACKUP WMI FILTERS
    #Connect to the Active Directory to get details of the WMI filters
    $WmiFilters = Get-ADObject -Filter 'objectClass -eq "msWMI-Som"' `
    -Properties msWMI-Author, msWMI-ID, msWMI-Name, msWMI-Parm1, msWMI-Parm2 `
    -ErrorAction SilentlyContinue
    ##CREATE REPORT FILES
    ##XML reports
    #Create a variable for the XML file representing custom information about the backed up GPOs
    $CustomGpoXML = "$SubBackupFolder\GpoDetails.xml"
    #Export our array of custom GPO objects to XML so they can be easily re-imported as objects
    $TotalGPOs | Export-Clixml -Path $CustomGpoXML
    #If $WMIFilters contains objects write these to an XML file
    If ($WmiFilters) {
    #Create a variable for the XML file representing the WMI filters
    $WmiXML = "$SubBackupFolder\WmiFilters.xml"
    #Export our array of WMI filters to XML so they can be easily re-imported as objects
    $WmiFilters | Export-Clixml -Path $WmiXML
    } #End of If ($WmiFilters)
    ##CSV report
    #Create a variable for the CSV file that will contain the SOM (Scope of Management) information for each backed-up GPO
    $SOMReportCSV = "$SubBackupFolder\GpoInformation.csv"
    #Now, let's create the CSV report
    ForEach ($CustomGPO in $TotalGPOs) {
    #Start constructing the CSV file line entry for the current GPO
    $CSVLine = "`"$($CustomGPO.Name)`",`"{$($CustomGPO.GPOGuid)}`","
    #Expand the SOMs property of the current object
    $CustomSOMs = $CustomGPO.SOMs
    #Loop through any SOMs returned
    ForEach ($CustomSOM in $CustomSOMs) {
    #Append the SOM path to our CSV line
    $CSVLine += "`"$CustomSOM`","
    } #End of ForEach ($CustomSOM in $CustomSOMs)...
    #Write the newly constructed CSV line to the report
    Add-Content -Path $SOMReportCSV -Value $CSVLine
    } #End of ForEach ($CustomGPO in $TotalGPOs)...
    ##MIGTABLE
    #Check whether a migration table should be created
    If ($MigTable) {
    #Create a variable for the migration table
    $MigrationFile = "$SubBackupFolder\MigrationTable.migtable"
    #Create a migration table
    $MigrationTable = $GPM.CreateMigrationTable()
    #Connect to the backup directory
    $GpmBackupDir = $GPM.GetBackUpDir($SubBackupFolder)
    #Reset the GPM search criterea
    $GpmSearchCriteria = $GPM.CreateSearchCriteria()
    #Configure search critera for the most recent backup
    $GpmSearchCriteria.Add($Constants.SearchPropertyBackupMostRecent,$Constants.SearchOpEquals,$True)
    #Get GPO information
    $BackedUpGPOs = $GpmBackupDir.SearchBackups($GpmSearchCriteria)
    #Add the information to our migration table
    ForEach ($BackedUpGPO in $BackedUpGPOs) {
    $MigrationTable.Add($Constants.ProcessSecurity,$BackedUpGPO)
    } #End of ForEach ($BackedUpGPO in $BackedUpGPOs)...
    #Save the migration table
    $MigrationTable.Save($MigrationFile)
    } #End of If ($MigTable)...
    } #End of If (Test-Path $SubBackupFolder)...

  • Parameter ' check logical ' question

    Hello,
    in the rman backup and recovery reference 10gR2, it says for the check logical parameter for backup:
    If the initialization parameter DB_BLOCK_CHECKSUM=TRUE, and if MAXCORRUPT and NOCHECKSUM are not set, then specifying CHECK LOGICAL detects all types of corruption that are possible to detect.
    But there is no description for the alternative of DB_BLOCK_CHECKSUM=FALSE.
    Now Ct. wants to know: What does check logical check if this parameter is not set to true?
    Ct. wants to check all kinds of corruption which is checkable by rman.
    regards
    Ulli

    Without the CHECK LOGICAL clause, RMAN will perform 2 types of validation in the cache layer of the datablock: block checksum and head/tail verification. When you add CHECK LOGCIAL clause, we perform additional checks in the transaction and data layer of the block which will add to the the processing time.
    However, a delay of 3.2 times seems a bit of higher side. You might want to check for the CPU/IO utilization on the machine. If you have idle CPU, you may consider adding more channels. Look out for v$backup_async_io.LONG_WAITS, if the number is high, IO is slow.
    It is also possible that due to the additional checks, RMAN is not able to fill the output buffer as fast as earlier. In that case, consider increasing multiplexing (FILESPERSET / MAXOPENFILE)

  • Creation of Logical Standby Database Using RMAN ACTIVE DATABASE COMMAND

    Hi All,
    I am in confusion how to create logical standby database from primary database using rman active database command.
    What i did:-
    Create primary database on machine 1 on RHEL 5 with Oracle 11gR2
    Create standby database on machine 2 on RHEL 5 With Oracle 11gR2 from primary using RMAN active database command
    Trying to create logical standby database on machine 3 on RHEL 5 with Oracle 11gR2 using RMAN active database command from primary.
    The point which confuse me is to start the logical standby in nomount mode on machine 3 with which pfile like i create the pfile for standby database do i need to create the pfile for logical standby db.
    I done the creation of logical standby database by converting physical standby to logical standby database
    I am following the below mentioned doc for the same:
    Creating a physical and a logical standby database in a DR environment | Chen Guang&amp;#039;s Blog
    Kindly guide me how to work over the same or please provide me the steps of the same.
    Thanks in advance.

    Thanks for your reply
    I already started the logical standby database with pfile in nomount mode. And successfully completed the duplication of database. by mentioning the DB_FILE_NAME_CONVERT and LOG_FILE_NAME_CONVERT parameter.
    But i am not able to receive the logs on the above mentioned blog i run the sql command to check the logs but getting "no rows selected"
    My primary database pfile is:
    pc01prmy.__db_cache_size=83886080
    pc01prmy.__java_pool_size=12582912
    pc01prmy.__large_pool_size=4194304
    pc01prmy.__oracle_base='/u01/app/oracle'#ORACLE_BASE set from environment
    pc01prmy.__pga_aggregate_target=79691776
    pc01prmy.__sga_target=239075328
    pc01prmy.__shared_io_pool_size=0
    pc01prmy.__shared_pool_size=134217728
    pc01prmy.__streams_pool_size=0
    *.audit_file_dest='/u01/app/oracle/admin/pc01prmy/adump'
    *.audit_trail='db'
    *.compatible='11.1.0.0.0'
    *.control_files='/u01/app/oracle/oradata/PC01PRMY/controlfile/o1_mf_91g3mdtr_.ctl','/u01/app/oracle/flash_recovery_area/PC01PRMY/controlfile/o1_mf_91g3mf6v_.ctl'
    *.db_block_size=8192
    *.db_create_file_dest='/u01/app/oracle/oradata'
    *.db_domain=''
    *.db_file_name_convert='/u01/app/oracle/oradata/PC01SBY/datafile','/u01/app/oracle/oradata/PC01PRMY/datafile'
    *.db_name='pc01prmy'
    *.db_recovery_file_dest='/u01/app/oracle/flash_recovery_area'
    *.db_recovery_file_dest_size=2147483648
    *.diagnostic_dest='/u01/app/oracle'
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=pc01prmyXDB)'
    *.fal_client='PC01PRMY'
    *.fal_server='PC01SBY'
    *.log_archive_config='DG_CONFIG=(pc01prmy,pc01sby,pc01ls)'
    *.log_archive_dest_1='LOCATION=/u01/app/oracle/flash_recovery_area/PC01PRMY/ VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=pc01prmy'
    *.log_archive_dest_2='SERVICE=pc01sby LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=pc01sby'
    *.log_archive_dest_3='SERVICE=pc01ls LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=pc01ls'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='DEFER'
    *.log_archive_dest_state_3='DEFER'
    *.log_archive_max_processes=30
    *.log_file_name_convert='/u01/app/oracle/oradata/PC01SBY/onlinelog','/u01/app/oracle/oradata/PC01PRMY/onlinelog'
    *.open_cursors=300
    *.pga_aggregate_target=78643200
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=236978176
    *.undo_tablespace='UNDOTBS1'
    My logical standby pfile is:-
    pc01ls.__db_cache_size=92274688
    pc01ls.__java_pool_size=12582912
    pc01ls.__large_pool_size=4194304
    pc01ls.__oracle_base='/u01/app/oracle'#ORACLE_BASE set from environment
    pc01ls.__pga_aggregate_target=79691776
    pc01ls.__sga_target=239075328
    pc01ls.__shared_io_pool_size=0
    pc01ls.__shared_pool_size=125829120
    pc01ls.__streams_pool_size=0
    *.audit_file_dest='/u01/app/oracle/admin/pc01ls/adump'
    *.audit_trail='db'
    *.compatible='11.1.0.0.0'
    *.control_files='/u01/app/oracle/oradata/PC01LS/controlfile/o1_mf_91g3mdtr_.ctl','/u01/app/oracle/flash_recovery_area/PC01LS/controlfile/o1_mf_91g3mf6v_.ctl'
    *.db_block_size=8192
    *.db_create_file_dest='/u01/app/oracle/oradata'
    *.db_domain=''
    *.db_file_name_convert='/u01/app/oracle/oradata/PC01SBY/datafile','/u01/app/oracle/oradata/PC01PRMY/datafile'
    *.db_name='pc01prmy'
    *.db_unique_name='pc01ls'
    *.db_recovery_file_dest='/u01/app/oracle/flash_recovery_area'
    *.db_recovery_file_dest_size=2147483648
    *.diagnostic_dest='/u01/app/oracle'
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=pc01prmyXDB)'
    *.log_archive_config='DG_CONFIG=(pc01prmy,pc01sby,pc01ls)'
    *.log_archive_dest_1='LOCATION=/u01/app/oracle/flash_recovery_area/PC01PRMY/ VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=pc01prmy'
    *.log_archive_dest_2='LOCATION=/u01/app/oracle/flash_recovery_area/PC01LS/ VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=pc01ls'
    *.log_archive_dest_3='SERVICE=pc01ls LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=pc01ls'
    *.log_archive_dest_state_1='ENABLE'
    *.log_archive_dest_state_2='ENABLE'
    *.log_archive_max_processes=30
    *.log_file_name_convert='/u01/app/oracle/oradata/PC01SBY/onlinelog','/u01/app/oracle/oradata/PC01PRMY/onlinelog'
    *.open_cursors=300
    *.pga_aggregate_target=78643200
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=236978176
    *.undo_tablespace='UNDOTBS1'
    Kindly advice over the same

  • Windows Server 2012 - Backup failing with Exchange - The application will not be available for recovery from this backup. the consistency check failed

    Hi
    We have a Windows 2012 server with Exchange 2013, all is working fine except now i am getting issues with the backup.
    'Exchange - The application will not be available for recovery from this backup. the consistency check failed for the component Microsoft Exchange Server'
    I have checked the database all is fine, i have created a new db and move all mailbox;s over and then removed the old db, i have enabled circular logging and then disabled it, it seems no matter what i do i cannot get a full backup!
    i did have to restore the server once and the backups still worked for about 4 days after that and then stopped, i have also tried to remove and re add the backup role!
    i am stumped, any advice would be great!

    Hi
    Ok, i created a test db and tried to back it up right away, it failed, i did not add any mailbox's to it either. i got quite a few events in the windows logs, as well as the same event above i got the following:
    Log Name:      Application
    Source:        MSExchangeRepl
    Date:          21/01/2013 10:16:30
    Event ID:      2038
    Task Category: Exchange VSS Writer
    Level:         Warning
    Keywords:      Classic
    User:          N/A
    Computer:      NERDS-DC01.nerds.local
    Description:
    Microsoft Exchange VSS Writer backup failed. No log files were truncated. Instance 75754d0d-8dfe-4909-8beb-5a4f824254a9. Database 4843b37c-7b3c-42b2-8b57-1393615c2c15.
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="MSExchangeRepl" />
        <EventID Qualifiers="32772">2038</EventID>
        <Level>3</Level>
        <Task>2</Task>
        <Keywords>0x80000000000000</Keywords>
        <TimeCreated SystemTime="2013-01-21T10:16:30.000000000Z" />
        <EventRecordID>261645</EventRecordID>
        <Channel>Application</Channel>
        <Computer>NERDS-DC01.nerds.local</Computer>
        <Security />
      </System>
      <EventData>
        <Data>75754d0d-8dfe-4909-8beb-5a4f824254a9</Data>
        <Data>4843b37c-7b3c-42b2-8b57-1393615c2c15</Data>
      </EventData>
    </Event>
    AND
    Log Name:      Application
    Source:        MSExchangeRepl
    Date:          21/01/2013 10:16:30
    Event ID:      2038
    Task Category: Exchange VSS Writer
    Level:         Warning
    Keywords:      Classic
    User:          N/A
    Computer:      NERDS-DC01.nerds.local
    Description:
    Microsoft Exchange VSS Writer backup failed. No log files were truncated. Instance 75754d0d-8dfe-4909-8beb-5a4f824254a9. Database db5826f3-1029-4219-ad80-441a0e94537a.
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="MSExchangeRepl" />
        <EventID Qualifiers="32772">2038</EventID>
        <Level>3</Level>
        <Task>2</Task>
        <Keywords>0x80000000000000</Keywords>
        <TimeCreated SystemTime="2013-01-21T10:16:30.000000000Z" />
        <EventRecordID>261646</EventRecordID>
        <Channel>Application</Channel>
        <Computer>NERDS-DC01.nerds.local</Computer>
        <Security />
      </System>
      <EventData>
        <Data>75754d0d-8dfe-4909-8beb-5a4f824254a9</Data>
        <Data>db5826f3-1029-4219-ad80-441a0e94537a</Data>
      </EventData>
    </Event>
    and
    Log Name:      Application
    Source:        MSExchangeRepl
    Date:          21/01/2013 10:16:30
    Event ID:      2034
    Task Category: Exchange VSS Writer
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:      NERDS-DC01.nerds.local
    Description:
    The Microsoft Exchange Replication service VSS Writer (Instance 75754d0d-8dfe-4909-8beb-5a4f824254a9) failed with error FFFFFFFC when processing the backup completion event.
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="MSExchangeRepl" />
        <EventID Qualifiers="49156">2034</EventID>
        <Level>2</Level>
        <Task>2</Task>
        <Keywords>0x80000000000000</Keywords>
        <TimeCreated SystemTime="2013-01-21T10:16:30.000000000Z" />
        <EventRecordID>261649</EventRecordID>
        <Channel>Application</Channel>
        <Computer>NERDS-DC01.nerds.local</Computer>
        <Security />
      </System>
      <EventData>
        <Data>75754d0d-8dfe-4909-8beb-5a4f824254a9</Data>
        <Data>FFFFFFFC</Data>
      </EventData>
    </Event>
    and
    Log Name:      Application
    Source:        SPP
    Date:          21/01/2013 10:16:30
    Event ID:      16389
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:      NERDS-DC01.nerds.local
    Description:
    Writer Microsoft Exchange Writer experienced retryable error during shadow copy creation. Retrying...  More info: .
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="SPP" />
        <EventID Qualifiers="0">16389</EventID>
        <Level>2</Level>
        <Task>0</Task>
        <Keywords>0x80000000000000</Keywords>
        <TimeCreated SystemTime="2013-01-21T10:16:30.000000000Z" />
        <EventRecordID>261650</EventRecordID>
        <Channel>Application</Channel>
        <Computer>NERDS-DC01.nerds.local</Computer>
        <Security />
      </System>
      <EventData>
        <Data>Microsoft Exchange Writer</Data>
        <Data>
        </Data>
        <Data>The writer experienced a transient error.  If the backup process is retried, the error may not reoccur. (0x800423F3)</Data>
        <Data>
        </Data>
        <Binary>00000000A5120000981200000000000042BEB7C511CAC619E59C92030000000000000000</Binary>
      </EventData>
    </Event>

Maybe you are looking for