RMAN difference bewteen full and incremental?

Hi,
Recently we had configured one recovery catalog database for our production database backup. We have made scripts it use to run through cron jobs (i.e) from sun-fri incremental backup and sat night full backup Since we have already configured in such manner only in another machine that db version was 10.2.0.3.0 After db upgrade to 10.2.0.4.0 So we upgraded that recovery catalog database also to 10.2.0.4.0 In this last night it started backup and as scheduled backup as per the RMAN Backup logs it got completed incremental successfully.But what i want to know Since this is the first time how rman will directly takes incremental without any full backup from where it compares if required? pls let me know.
Thanks.

hi,
Everything documented, just need to study in docs.
http://download.oracle.com/docs/cd/B19306_01/backup.102/b14192/bkup004.htm
The following command performs a level 1 differential incremental backup of the database:
RMAN> BACKUP INCREMENTAL LEVEL 1 DATABASE;
If no level 0 backup is available, then the behavior depends upon the compatibility mode setting. If compatibility is >=10.0.0, RMAN copies all blocks changed since the file was created, and stores the results as a level 1 backup. In other words, the SCN at the time the incremental backup is taken is the file creation SCN. If compatibility <10.0.0, RMAN generates a level 0 backup of the file contents at the time of the backup, to be consistent with the behavior in previous releases.
- Pavan Kumar N
Oracle 9i/10g - OCP
http://oracleinternals.blogspot.com/

Similar Messages

  • Understanding Windows Small Business Server 2008 Backup - Full and Incremental with 2 Backup Destinations

    Hi,
    I am re-posting a question from the following page here as I have the same problem now and I can't seem to find the answer even though the question has been marked as solved: http://social.technet.microsoft.com/Forums/en/windowsbackup/thread/daff108b-effa-4dad-807a-d604345975dd
    Below is a copy of the question:
    I have 2 backup drives, (Backup Drive 1 and Backup Drive 2)
    Lets assume I have never performed a backup, so Day 1 is the first backup of the server.
    Backup Drive 1 connected on Day 1 it performs a FULL Backup - this is expected.
    Backup Drive 1 connected on Day 2 it performs an INCREMENTAL Backup - this is expected.
    Backup Drive 2 connected on Day 3 it performs a FULL Backup - this is expected.
    Backup Drive 1 connected on Day 4 it performs a FULL Backup - WHY? my understanding is that is should perform an Incremental Backup from Day 2
    Backup Drive 2 connected on Day 5 it performs a FULL Backup - again why is it not performing an Incremental Backup from Day 3?
    This means that in normal operation (Backup Drives alternate every day) We are performing FULL Backups everyday. In a few months this wont be an acceptable backup strategy; it will take too long.
    I've used 'Configure Performance Settings' in the Windows Server Backup mmc to specify 'Always perform incremental backup' - it makes no difference.
    If I look at the Backup Drive Disk Usage details it confuses me even more. It may be performing Full Backups everyday but it's definitely not storing Full Backup copies on the Backup Drives.
    It seems to be that even when a Full Backup is performed, only the deltas are written to the Backup Drive so even though it takes longer it has the same effect as an incremental  (so why doesn't it just perform an incremental?)
    I don't understand Microsoft's use of Full and Incremental, it seems obtuse to provide a choice that appears to have no effect on the actual data written to the Backup Drive.
    My real-world experience is at odds with that statements made in
    The Official SBS Blog  it states "every backup is incremental from a storage point of view" as well as "Because the wizard schedules differential-based backups' (nowhere in the Backup mmc have I seen any reference or options for differential),
    "Backup runs quickly" and "...works the same for multiple disk rotation." (This is simply not the case with a 2-disk rotation. It runs a lengthy FULL Backup every time.)
    The backup has been configured using SBS Console, runs once a day at 16:00. 2 Backup Drives, alternating daily.
    Can anyone clarify Windows Backup operation for me?
    I'd appreciate any feedback at all, thanks.

    Optimizing Backup and Server Performance
    (Windows Server 2008 R2)
    http://technet.microsoft.com/en-us/library/dd759145.aspx
    Even if you choose the option to create incremental backups, Windows Server Backup will create a full backup once every 14 days, or after 14 incremental backup operations, to reduce the risk from disk corruptions.
    Of course, this is for R2.
    Merv  Porter   [SBS-MVP]
    ============================
    "Merv Porter [SBS-MVP]" wrote in message
    news:a1ca618e-ad66-4770-8c39-21285a08f671...
    Interesting post...
    WBAdmin to remote backup server
    http://social.technet.microsoft.com/Forums/en-US/windowsbackup/thread/764fe9a4-960e-4e90-b8fb-8e7581752a9d#520f38fe-149c-4424-9c0b-54695297e919
    In Windows Server 2008, there are several reasons which may cause the backup to be full instead of incremental
    1. Backup on the target is deleted/not present.
    2. Source volume snapshot is deleted, from which the last backup was taken.
    3. 7 days have passed or 6 incremental backups have happened since the last full backup.
    4. Churn on the backup source is high (more than 50%)
    Abhinav Mathur[MSFT] Microsoft
    Merv  Porter   [SBS-MVP]
    ============================
    "Les Connor [SBS-MVP]" wrote in message
    news:0053cd83-75d1-4dbc-8182-ae67cadf4780...
    I believe it's how backup is designed, as I see the same thing. Why it works
    the way it does is another story though, I don't know the answer to that.
    Les Connor [SBS MVP]
    "Kurni" wrote in message
    news:[email protected]...
    > Hi Les,
    >
    > Thank you for your reply.
    >
    > Are you saying that that's how the backup is designed? What I (and the
    > original poster of the question) have in mind is actually different (each
    > disk should have their own base backup).
    >
    > Quoting from the original question:
    >
    > Backup Drive 1 connected on Day 1 it performs a FULL Backup - this is
    > expected.
    > Backup Drive 1 connected on Day 2 it performs an INCREMENTAL Backup - this
    > is expected.
    > Backup Drive 2 connected on Day 3 it performs a FULL Backup - this is
    > expected.
    > Backup Drive 1 connected on Day 4 it performs a FULL Backup - WHY? my
    > understanding is that is should perform an Incremental Backup from Day 2
    > Backup Drive 2 connected on Day 5 it performs a FULL Backup - again why is
    > it not performing an Incremental Backup from Day 3 ?
    > Please let me know if I actually misunderstand how windows server backup
    > work.
    >
    MVP - SBS

  • How do full and incremental Informatica workflow differ?

    Hi:
    I've read that a custom Informatica workflow should have full and incremental versions. I've compared the incremental and full versions of several seeded OBIA workflows, but I cannot find how they differ. For example, when I look at the session objects for each workflow they both have a source filter that limits a date field by LAST_EXTRACT_DATE.
    Can someone shed some light on this?
    Thanks.

    To answer your question, they differ in various ways for various mappings. For most FACT tables which hold high volume transactional data, there may be a SQL override in the FULL session that uses INITIALEXTRACT_DATE and a different override in the INCREMENTAL (does not have Full Suffix) that uses LASTEXTRACT_DATE. For dimension tables, its not always required that you have a different logic for FULL versus incremental.
    Also, all FULL sessions (even if they have the same SQL) will have a BULK option turned on in the Session properties that allow a faster load since the table is usually truncated on FULL versus Incremental. As a best practice, for FACTS, it is best to have a FULL versus INCREMENTAL session. For Dimensions, depending on the volume of transactions or the change capture date field available, you may or may not have different logic. If you do a FULL load, however, it is better to use BULK to speed up the load.
    if this is helpful, please mark as correct or helpful.

  • Difference between full and target quantity in iteme category

    What is the difference between Target quantity and Full quantity in Completion rule in Item Catergory customization?

    hi
    full quantity is used for quotations
    target quantity for contracts
    Completion rules
    An item may be considered fully referenced if one of the following criteria have been met
    An item is complete as soon as it is used as a reference. In this case, the item is fully referenced, even if only a part of the quantity is copied.
    An item is only complete once the full quantity has been copied into a subsequent document. If the item is partially referenced, it is possible to enter several sales documents for the item until the quantity is fully accounted for.
    An item is only complete once the target quantity (in a contract, for example) has been copied into a subsequent document. If the item is partially referenced, it is possible to enter several sales documents for the item until the quantity is fully accounted for.
    The standard version of the system defines an inquiry item as complete even if only one part of the inquiry quantity has been copied into a quotation. On the other hand, when you copy a quotation into a sales order, the quotation is not complete until the quotation quantity has been fully referenced.
    Contracts, like quotations, are only considered complete once the full target quantity of the contract has been released.

  • Restore User Archives from Full and Incremental backups

    Ok Groupwise 8. User's keep their archives under the main users folder on a Novell Server. Groupwise is running from Linux server, so the main post office and domain are backed up nightly on the Linux server, the user archives are backed up nightly on that server.
    I have had to restore current messages a user accidentally deleted from their mailbox, Reload handled that as those were current. But now I have one user who didn't know what the folder GWArchive was and blew it away. It was his archives. He needs them back. So I have restored them from the Novell server backup, first from the last full backup and then the two incremental backups that cover up to the point he deleted the folder. The incrementals were restored to separate folders because I didn't want those to overwrite the main files as these are incremental.
    How do you "merge" all these together to rebuild that archive index? Is it just a matter of moving all the .000, .001, .002, etc. files over, then the *.db files, and the index files? Or something different?
    Thanks in advance - I wish there was someplace that describes all these file types, what they are, etc. a really technical level explanation so I understand this a lot better.
    Kind regards,

    * iliadmin,
    when you do incremental backups, it's very useful to have an option to restore a certain point in time and let the backup app collect the data which was in place at that point in time. Else copy back the base, then *overwrite* with the first incremental, then the second etc., moving up to the latest backup you have.
    Which file is which is pretty well explained in the manual. The files are the same as in the live PO, just for only one user when you look at an archive.
    HTH
    Uwe
    Novell Knowledge Partner (NKP)
    Please don't send me support related e-mail unless I ask you to do so.

  • What is the difference between full checkpoint and incremental checkpoint?

    What is the difference between full checkpoint and incremental checkpoint?
    And what is checkpoint queue?
    Can someone clarify these concepts?
    Thanks!

    Hi,
    there are different types of checkpoints:
    - Full checkpoint:
    => DBWR writes all dirty buffers from the Buffer cache to the datafiles and CKPT retrieves a new Checkpoint Change Number from a sys owned sequence and writes this number to all file headers and the controlfile.
    -- can be triggered by different events, like a logswitch, a manual checkpoint (alter system ..), a shutdown and so on
    This is the setup point for SMON for a crash recovery.
    - Partial checkpoint:
    => DBWR writes all dirty buffers of a singel tablespace from the Buffer cache to the datafiles and CKPT retrieves a new Checkpoint Change Number from a sys owned sequence and writes this number to all file headers and the controlfile.
    -- can e triggered by an ALTER TABLESPACE OFFLINE, ALTER TABLESPACE READ ONLY, ALTER TABLESPACE BEGIN BACKUP statement.
    Incremental checkpoint:
    number to all file headers and the controlfile.
    -- can be triggered by different events, like a logswitch, a manual checkpoint (alter system ..), a shutdown and so on
    - Partial checkpoint:
    => DBWR writes all dirty buffers of a single tablespace from the Buffer cache to the datafiles and CKPT retrieves a new Checkpoint Change Number from a sys owned sequence and writes this number to all file headers and the controlfile.
    -- frequency is determined by FAST_START_MTTR_TARGET parameter (new feature to Oracle 9i), with wich you can specify a time in seconds which SMON is allowed to take maximal for a Crash Recovery until the database must be open again.
    Dapending on the system you have SMON must calculate the maximum number of Redolog-blocks which it can manage to recover in the specified number of seconds. It will then create so called incremental checkpoints which will be tracked in the so called checkpoint queue in memory.
    FAST_START_MTTR_TARGET is auto-tuned in Oracle 10g and Oracle tries to manage (incremental) checkpoint in a fashion that a minimum of I/Os are caused and a minimum time for crash recovery is needed.
    If you set LOG_CHECKPOINTS_TO_ALERT to TRUE you will find checkpoint information in the alertSID.log file. You will see FULL and INCREMENTAL checkpoints then.
    Hope this clarifies your question,
    Lutz Hartmann

  • Full vs Incremental Refresh?

    What is the main difference between full an incremental (auto)refreshes?
    Documentation suggests that full refresh is recommended on the groups when data is changing very frequently while Incremental refresh is advised to use on the tables when data is more or less static.
    1. Is there any way to justify/quantify what is the frequent data change (50% of records in 24 hours, etc..)?
    2. In the documentation I also found that using Incremental Data Refresh TT will hook up triggers on the tables in the Oracle DB. Is that correct?
    3. For Full DataRefresh TT will be using logs from Oracle DB to refresh the data in TT. In some cases logs in the Oracle DB are kept only for a specific number of days (i.e. 5 days). What will happen if those logs are not available?
    4. Using Full refresh TT will be reloading all the data from the Oracle DB. Is there any impact for the data availability in TT by using Full refresh?
    5. I believe that Full data refresh will be taking much more time compared to the Incremental data refresh. In this case if I set autorefresh to every 5 seconds, then basically I will end up with the infinite refresh. What is recommended frequency for Full and Incremental refreshes?
    6. Full vs Incremental refresh. Which one is using less resources in Oracle DB and TT and why?
    Thank you.

    Full refresh works by discarding all the data in the cache group tables and then reloading it all from Oracle on every refresh cycle. it is best when (a) the tables are very small and (b) the refresh interval is not too short.
    Incremental autorefresh works by installing a trigger and a tracking table on each of the base tables in Oracle. As changes occur in Oracle they are tracked and when the next autorefrsh cycle comes around only the changes are propagated to the cache tables. Incremental is recommended when (a) the tables involved are of any substantial size and/or (b) when a short refresh interval is required. To try and answer your specific questions:
    1. Is there any way to justify/quantify what is the frequent data change (50% of records in 24 hours, etc..)?
    CJ>> Not really. This comes down to application requirements, how much load you can tolerate on Oracle etc.
    2. In the documentation I also found that using Incremental Data Refresh TT will hook up triggers on the tables in the Oracle DB. Is that correct?
    CJ>> Yes, a single trigger and a single tracking will be instantiated on each base table that is cached in cache group that uses incremental autorefresh.
    3. For Full DataRefresh TT will be using logs from Oracle DB to refresh the data in TT. In some cases logs in the Oracle DB are kept only for a specific number of days (i.e. 5 days). What will happen if those logs are not available?
    CJ>> No. Neither incremental nor full refresh uses Oracle logs. Full refresh simply queries the data in the table and incremental uses a combination of the data in the table and the tracking table.
    4. Using Full refresh TT will be reloading all the data from the Oracle DB. Is there any impact for the data availability in TT by using Full refresh?
    CJ>> Yes. Using full refresh TimesTen starts by emptying the cache table and so there is a significant impact on data availability. Incremental refresh does not have this issue.
    5. I believe that Full data refresh will be taking much more time compared to the Incremental data refresh. In this case if I set autorefresh to every 5 seconds, then basically I will end up with the infinite refresh. What is recommended frequency for Full and Incremental refreshes?
    CJ>> regardless of the refresh method chosen you should ensure that the refresh interval is set >> than the time it takes to perform the refresh. Setting it to a shorter value simply resuls in almost cointihnuous refresh and a much heavier load on Oracle. This is especially problematic with full refresh.
    6. Full vs Incremental refresh. Which one is using less resources in Oracle DB and TT and why?
    CJ>> Again it depends but in general incremental is the preferred choice.
    In my experience I have rarely seen anyone use full autorefresh.
    Chris

  • Full and delta Data loads in SAP BI 7.0

    Hi,
    Kindly explain me the difference between full and delta data loads in cube and DSO.
    When do we select change log or active table(with or without archive) with full/delta in DTP for DSO.
    Please explain the differences in data load for above different combinations.
    With thanks and regards,
    Mukul

    Please search the forums. This topic has already been discussed a lot of times.

  • Appraisal and Increment

    Hi Pals,
          I would like to get the exact information and difference between Appraisal and 
          increment. if u know please share and also give me the technical detail - how
          to do the both.
               i'm in urgent need. Advance Thanks for ur work.
    Regards,
    Manian.

    Hi Balaman,
    I Can say this is this process is combination of Appraisal and Compensation. Generally once we perform the Appraisal. We will get the Appraisal results in the IT 0025, Similary We will create the compensation plans as merit increase and We will give the Matrix and Martix dimensions. We will assign the Default Percentage, Minimum and Maximum increase percentage for each of this Appraisal Rating (i.e., Segments),
    Now, based on your appraisal results you can default the increaments. we have standard report to update the increased salary in IT 0008. But this we will good if your country is having ANSAL concepot.
    Regards
    <u><i><b>Sam</b></i></u>
    <i><b>Reward the Points if my answers help you out</b></i>

  • Difference between Full backup and lncremental level 0 backup

    Hi,
    What is the difference between Full backup and lncremental level 0 backup?
    what I understand is:
    Actually incremental backup backups all the block containing data and skips the blocks that are unused.
    Similarly full backup backups all the blocks containing data.(does it skip unused blocks ?????).
    The only difference that I feel is that the level 0 backup is recorded as an incremental backup in the RMAN repository, so it can be used as the parent for a level 1 backup where Full database backup is not recorded in the Rman respository and cannot be used as parent for level 1 backup.
    (if you still feel you have some more points to share regarding the difference between backups, I request to share with me)
    so my question here is
    Can we use Incremental level 0 backup for full database recovery. Else is there any compulsion like we must only use full database backup for full database recovery?
    can anyone throw light on this
    with regards;
    Boo

    nowhere in this thread we talked about the image copies as we are only talking about the incremental backup.
    The word "always" has a meaning. If you now wish to say, "well, RMAN will mostly use a level 0 or full backup, except when there are image copies around" then the word "always" doesn't fit your newly-revised description. That would make it an inappropriate word to use. And indeed, it's a factually wrong word to use.
    If you want to get really picky about it, incidentally, the words 'datafile copy' appear in a post from me at Aug 4, 2007 6:03 PM whereas your post (in which the "always" word was used) only appears Aug 4, 2007 8:32 PM -a full 2 and a half hours (and and two posts) AFTER the concept of image copies had been mentioned.
    But regardless of times: you can't just pretend image copies don't exist just because they haven't been mentioned so far in a thread. If you post "this is the way something ALWAYS works", you'd better be completely right about it, because "ALWAYS" is a big word. And there's no grey about this: either "always" is right, or it's wrong. And in this case, it's wrong: there are many times when RMAN will not do what you baldly asserted it would do. Image copies aren't some rare, exotic beast, either... so the occasions when your 'rule' breaks down are many and common.
    until now I was simply beliving in what Oracle docs and metalink notes says
    Try not to do that. Because you will fall into error after error if you do. Metalink is not the Holy Gospel of Oracle. Testing is, and it's because I've been testing this stuff since 1999 I will, on this occasion, decline your kind invitation to do "a small test". Been there, done that.
    All of which leaves us discussing things that are of no relevance to the original poster, and I have no desire to take the thread off into silly areas of dispute. So I will leave it there, and I have nothing further to say on the matter. RMAN does what it does, and what it does is not what you said it would do. End of story.

  • What's the difference between "Full Name" and "Account Name"

    Just re-install a new OS X Lion. On the "Create Your Computer Account" screen, what's the difference between "Full Name" and "Account Name"?

    Your computer name by default would be "John David Appleseed's Macbook Pro"
    But you can always change that in System Preferences->Sharing
    And yes, the first account you set up will be an administrator.

  • DIFFERENCE BETWEEN THE FULL AND BULK-LOGGED RECOVERY MODEL

    DIFFERENCE  BETWEEN THE  FULL AND BULK-LOGGED  RECOVERY MODEL 

    In bulk logged recovery mode certain bulk operations are minimally logged. In FULL recovery mode these are fully logged. These bulk operations are as mentioned below
    1. SELECT INTO
    2. BULK IMport operations including BULK INSERT and BCP
    3. INSERT INTO SELECT command using the OPENROWSET(BULK) function
    4. Partial updates to columns having large value data type
    5. Using WRITE clause in UPDATE statements
    6. Index operations e.g CREATE INDEX, ALTER INDEX REBUILD , DROP INDEX
    In Bulk Logged Recovery Mode when you execute these operations SQL Server only logs the fact that these operation occured and information about space allocation. The actual change in the data is maintained in the BCM (Bulk
    Changed Map)
    Since the actual changes are not recorded in the log file, the log file size in the reletively less in size but this tradeoff comes with the price of increased backup time. This is so because during the log backup its not just the log being backup, but also
    the extents that are marked by the Bulk Changed Map as changed.
    SQLEnthusiast
    http://sqlsimplified.com/
    Please click the Mark as Answer button if a post solves your problem!

  • Whats the difference bewteen the latest 30gb and the older 1?

    whats the difference bewteen the latest 30gb and the older 1?

    I think you will find the battery life only changed
    for the 60/new 80 and not the 30 either for the 1st
    5th gen or 2nd.
    Actually this is not accurate. Playback time for music wasn't changed, but video playback went from 2 hours to 3.5 hours. Much of this was due to some hardware changes between the newer and older 5g iPod's. Also battery life leaving the display on a low setting has very little impact on the newer iPod compared to the older one. So, they are different, no doubt about it.

  • EXACT DIFFERENCE BETWEEN  FULL LOAD AND REPAIRFULL LOAD?

    HI Champ,
    Can anyone explain me the exact difference between  full load and Repairfull load?Give e some senario where we can go for this?Please.....
    10zin

    Hi,
    Full repair can be said as a Full with selections. But the main use or advantage of Full repair load is that it wont affect delta loads in the system. If you load a full to a target with deltas running you again will have to initialize them for deltas to continue. But if you do full repair it wont affect deltas.
    This is normally done we when we lose some data or there is data mismatch between source system and BW.
    Check the OSS Note 739863 'Repairing data in BW' for all the details
    Symptom
    Some data is incorrect or missing in the PSA table or in the ODS object (Enterprise Data Warehouse layer).
    There may be a number of reasons for this problem: Errors in the relevant application, errors in the user exit, errors in the DeltaQueue, handling errors in the customers posting procedure (for example, a change in the extract structure during production operation if the DeltaQueue was not yet empty; postings before the Delta Init was completed, and so on), extractor errors, unplanned system terminations in BW and in R/3, and so on.
    Solution
    Read this note in full BEFORE you start actions that may repair your data in BW. Contact SAP Support for help with troubleshooting before you start to repair data.
    BW offers you the option of a full upload in the form of a repair request (as of BW 3.0B). If you want to use this function, we recommend that you use the ODS object layer.
    Note that you should only use this procedure if you have a small number of incorrect or missing records. Otherwise, we always recommend a reinitialization (possibly after a previous selective deletion, followed by a restriction of the Delta-Init selection to exclude areas that were not changed in the meantime).
    1. Repair request: Definition
    If you flag a request as a repair request with full update as the update mode, it can be updated to all data targets, even if these already contain data from delta initialization runs for this DataSource/source system combination. This means that a repair request can be updated into all ODS objects at any time without a check being performed. The system supports loading by repair request into an ODS object without a check being performed for overlapping data or for the sequence of the requests. This action may therefore result in duplicate data and must thus be prepared very carefully.
    The repair request (of the "Full Upload" type) can be loaded into the same ODS object in which the 'normal' delta requests run. You will find this request under the "Repair Request" option in the InfoPackage (Maintenance) menu.
    2. Prerequisites for using the "Repair Request" function
    2.1. Troubleshooting
    Before you start the repair action, you should carry out a thorough analysis of the possible cause of the error to make sure that the error cannot recur when you execute the repair action. For example, if a key figure has already been updated incorrectly in the OLTP system, it will not change after a reload into BW. Use transaction RSA3 (Extractor Checker) in the source system for help with troubleshooting. Another possible source of the problem may be your user exit. To ensure that the user exit is correct, first load a user exit with a Probe-Full request into the PSA table and check whether the data is correct. If it is not correct: Search for the error in the exit user. If you do not find it, we recommend that you deactivate the user exit for testing purposes and request a new Full Upload. It If the data arrives correctly, it is highly probable that the error is indeed in the user exit.
    We always recommend that you load the data into the PSA table in the first step and check the result there.
    2.2. Analyze the effects on the downstream targets
    Before you start the Repair request into the ODS object, make sure that the incorrect data records are selectively deleted from the ODS object. However, before you decide on selective deletion, you should read the Info Help for the "Selective Deletion" function, which you can access by pressing the extra button on the relevant dialog box. The activation queue and the ChangeLog remain unchanged during the selective deletion of the data from the ODS object, which means that the incorrect data is still in the change log afterwards. After the selective deletion, you therefore must not reconstruct the ODS object if it is reconstructed from the ChangeLog. (Reconstruction is usually from the PSA table but, if the data source is the ODS object itself, the ODS object is reconstructed from its ChangeLog). You MUST read the recommendations and warnings about this (press the "Info" button).
    You MUST also take into account the fact that the delta for the downstream data targets is created from the changelog. If you perform selective deletion and then reload data into the deleted area, this may result in data inconsistencies in the downstream data targets.
    If you only use MOVE and do not use ADD for updates in the ODS object, selective deletion may not be required in some cases (for example, if incorrect records only have to be changed, rather than deleted). In this case, the DataMart delta also remains intact.
    2.3. Analysis of the selections
    You must be very precise when you perform selective deletion: Some applications do not provide the option of selecting individual documents for the load process. Therefore, you must first ensure that you can load the same range of documents into BW as you would delete from the ODS object. This note provides some application-specific recommendations to help you "repair" the incorrect data records.
    If you updated the data from the ODS object into the InfoCube, you can also delete it there using the "Selective deletion" function. However, if it is compressed at document level there and deletion is no longer possible, you must delete the InfoCube content and fill the data in the ODS object again after repair.
    You can only perform this action after a thorough analysis of all effects of selective data deletion. We naturally recommend that you test this first in the test system.
    The procedure generally applies for all SAP applications/extractors. The application determines the selections. For example, if you cannot use the document number for selection but you can select documents for an entire period, then you are forced to delete and then update documents for the entire period in the data target. Therefore, it is important to look first at the selections in the InfoPackage exactly before you delete data from the data target.
    Some applications have additional special features:
    Logistics cockpit: As preparation for the repair request, delete the SetUp table (if you have not already done so) and fill it selectively with concrete document numbers (or other possible groups of documents determined by the selection). Execute the Repair request.
    Caution: You can currently use the transactions that fill SetUp tables with reconstruction data to select individual documents or entire ranges of documents (at present, it is not possible to select several individual documents if they are not numbered in sequence).
    FI: The Repair request for the Full Upload is not required here. The following efficient alternatives are provided: In the FI area, you can select documents that must be reloaded into BW again, make a small change to them (for example, insert a period into the assignment text) and save them -> as a result, the document is placed in the delta queue again and the previously loaded document under the same number in the BW ODS object is overwritten. FI also has an option for sending the documents selectively from the OLTP system to the BW system using correction programs (see note 616331).
    3. Repair request execution
    How do you proceed if you want to load a repair request into the data target? Go to the maintenance screen of the InfoPackage (Scheduler), set the type of data upload to "Full", and select the "Scheduler" option in the menu -> Full Request Repair -> Flag request as repair request -> Confirm. Update the data into the PSA and then check that it is correct. If the data is correct, continue to update into the data targets."
    Refer.
    Repair full request
    Re: Repair full request
    Steps to perform repair full request
    full repair request
    repair full request
    Re: Repair Full Request
    Thanks,
    JituK

  • RMAN backup files differences in numbers and sizes

    hi,
    I have cloned one database and after doing i can see the target (cloned database has more number of files and in different sizes).
    i can see the scripts are the same and environments are the same..
    where do i find out why there is difference between source and target ..
    ulimit -a > is same
    RMAN backup scripts > is same
    any clues i can find out >?
    Thanks.

    user13051945 wrote:
    hi,
    Let me make it clear here
    Cloning done well no issues..
    now two databases A and B are seperate now.. functioning in two different serveres
    The observation (not an issue) here is
    Server 1 > the backup files generated by Database A is only 10 backup files (10 * 5 = 50 )
    Server 2 > the backup files generated by Database B is 25 files.... (25 * 2 = 50 )
    so how the differences came in both backup scripts are the same....
    hope you got my concern...10 * 5 and 25 * 2
    *5 and 2* are backup piece and *10 and 25* are backupset, yes?
    From each server, post:
    RMAN>show all;

Maybe you are looking for