RMAN - CTRL-file - Sync

hi all,
is there a possiblity to transfer the information about backup from the control file to the rman catalog?
I have first created rman-backups without the rman - catalog and created today the rman catalog. If I exeute the command "list backup of database" I only got the information when I am not logged in on the rman-catalog
is there a possiblity to sync the information from the controle file to the rman catalog?
best regards
Stefan

Issue RESYNC CATALOG on the rman prompt being connected to the recovery catalog.
http://download.oracle.com/docs/cd/B19306_01/backup.102/b14191/rcmcatdb003.htm
Resynchronization, compares the recovery catalog to either the current control file of the target database/backup control file and updates the recovery catalog with information that is missing or changed.
Adith

Similar Messages

  • RMAN causing "log file sync"

    Hi,
    Maybe someone can help me on this.
    We have a RAC database in production that (for some) applications need a response of 0,5 seconds. In general that is working.
    Outside of production hours we make a weekly full backup and daily incremental backup so that is not bothering us. However as soon as we make an archive backup or a backup of the control file during production hours we have a problem as the application have to wait for more then 0,5 seconds for a respons caused by the event "log file sync" with wait class "Commit".
    I already adjusted the script for RMAN so that we use only have 1 files per set and also use one channel. However that didn't work.
    Increasing the logbuffer was also not a success.
    Increasing Large pool is in our case not an option.
    We have 8 redolog groups with each 2 members ( each 250 Mb) and an average during the day of 12 logswitches per hour which is not very alarming. Even during the backup the I/O doesn't show very high activity. The increase of I/O at that moment is minor but (maybe) apperantly enough to cause the "log file sync".
    Oracle has no documentation that gives me more possible causes.
    Strange thing is that before the first of October we didn't have this problem and there were no changes made.
    Has anyone an idea where to look further or did anyone experience a thing like this and was able to solve it?
    Kind regards

    The only possible contention I can see is between the log writer and the archiver. 'Backup archivelog' in RMAN means implicitly 'ALTER SYSTEM ARCHIVE LOG CURRENT' (log switch and archiving the online log).
    You should alternate redo logs on different disks to minimize the effect of the archiver on the log writer.
    Werner

  • Log file sync  during RMAN archive backup

    Hi,
    I have a small question. I hope someone can answer it.
    Our database(cluster) needs to have a response within 0.5 seconds. Most of the time it works, except when the RMAN backup is running.
    During the week we run one time a full backup, every weekday one incremental backup, every hour a controlfile backup and every 15 minutes an archival backup.
    During a backup reponse time can be much longer then this 0.5 seconds.
    Below an typical example of responsetime.
    EVENT: log file sync
    WAIT_CLASS: Commit
    TIME_WAITED: 10,774
    It is obvious that it takes very long to get a commit. This is in seconds. As you can see this is long. It is clearly related to the RMAN backup since this kind of responsetime comes up when the backup is running.
    I would like to ask why response times are so high, even if I only backup the archivelog files? We didn't have this problem before but suddenly since 2 weeks we have this problem and I can't find the problem.
    - We use a 11.2G RAC database on ASM. Redo logs and database files are on the same disks.
    - Autobackup of controlfile is off.
    - Dataguard: LogXptMode = 'arch'
    Greetings,

    Hi,
    Thank you. I am new here and so I was wondering how I can put things into the right category. It is very obvious I am in the wrong one so I thank the people who are still responding.
    -Actually the example that I gave is one of the many hundreds a day. The respone times during the archive backup is most of the time between 2 and 11 seconds. When we backup the controlfile with it, it is for sure that these will be the response times.
    -The autobackup of the controfile is put off since we already have also a backup of the controlfile every hour. As we have a backup of archivefiles every 15 minutes it is not necessary to also backup the controlfile every 15 minutes, specially if that even causes more delay. Controlfile is a lifeline but if you have properly backupped your archivefiles, a full restore with max 15 minutes of data loss is still possible. We put autobackup off since it is severely in the way of performance at the moment.
    As already mentioned for specific applications the DB has to respond in 0,5 seconds. When it doesn’t happen then an entry will be written in a table used by that application. So I can compare the time of failure with the time of something happening. The times from the archivelog backup and the failure match in 95% of the cases. It also show that log file sync at that moment is also part of this performance issue. I actually built a script that I used for myself to determine out of the application what the cause is of the problem;
    select ASH.INST_ID INST,
    ASH.EVENT EVENT,
    ASH.P2TEXT,
    ASH.WAIT_CLASS,
    DE.OWNER OWNER,
    DE.OBJECT_NAME OBJECT_NAME,
    DE.OBJECT_TYPE OBJECT_TYPE,
    ASH.TIJD,
    ASH.TIME_WAITED TIME_WAITED
    from (SELECT INST_ID,
    EVENT,
    CURRENT_OBJ#,
    ROUND(TIME_WAITED / 1000000,3) TIME_WAITED,
    TO_CHAR(SAMPLE_TIME, 'DD-MON-YYYY HH24:MI:SS') TIJD,
    WAIT_CLASS,
    P2TEXT
    FROM gv$active_session_history
    WHERE PROGRAM IN ('yyyyy', 'xxxxx')) ASH,
    (SELECT OWNER, OBJECT_NAME, OBJECT_TYPE, OBJECT_ID FROM DBA_OBJECTS) DE
    WHERE DE.OBJECT_id = ASH.CURRENT_OBJ#
    AND ASH.TIME_WAITED > 2
    ORDER BY 8,6
    - Our logfiles are 250M and we have 8 groups of 2 members.
    - Large pool is not set since we use memory_max_target and memory_target . I know that Oracle maybe doesn’t use memory well with this parameter so it is truly a thing that I should look into.
    - I looked for the size of the logbuffer. Actually our logbuffer is 28M which in my opinion is very large so maybe I should put it even smaller. It is very well possible that the logbuffer is causing this problem. Thank you for the tip.
    - I will also definitely look into the I/O. Eventhough we work with ASM on raid 10 I don’t think it is wise to put redo logs and datafiles on the same disks. Then again, it is not installed by me. So, you are right, I have to investigate.
    Thank you all very much for still responding even if I put this in the totally wrong category.
    Greetings,

  • Performance Issue: Wait event "log file sync" and "Execute to Parse %"

    In one of our test environments users are complaining about slow response.
    In statspack report folowing are the top-5 wait events
    Event Waits Time (cs) Wt Time
    log file parallel write 1,046 988 37.71
    log file sync 775 774 29.54
    db file scattered read 4,946 248 9.47
    db file parallel write 66 248 9.47
    control file parallel write 188 152 5.80
    And after runing the same application 4 times, we are geting Execute to Parse % = 0.10. Cursor sharing is forced and query rewrite is enabled
    When I view v$sql, following command is parsed frequently
    EXECUTIONS PARSE_CALLS
    SQL_TEXT
    93380 93380
    select SEQ_ORDO_PRC.nextval from DUAL
    Please suggest what should be the method to troubleshoot this and if I need to check some more information
    Regards,
    Sudhanshu Bhandari

    Well, of course, you probably can't eliminate this sort of thing entirely: a setup such as yours is inevitably a compromise. What you can do is make sure your log buffer is a good size (say 10MB or so); that your redo logs are large (at least 100MB each, and preferably large enough to hold one hour or so of redo produced at the busiest time for your database without filling up); and finally set ARCHIVE_LAG_TARGET to something like 1800 seconds or more to ensure a regular, routine, predictable log switch.
    It won't cure every ill, but that sort of setup often means the redo subsystem ceases to be a regular driver of foreground waits.

  • Statspack: High log file sync timeouts and waits

    Hi all,
    Please see an extract from our statpack report:
    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~ % Total
    Event Waits Time (s) Ela Time
    log file sync 349,713 215,674 74.13
    db file sequential read 16,955,622 31,342 10.77
    CPU time 21,787 7.49
    direct path read (lob) 92,762 8,910 3.06
    db file scattered read 4,335,034 4,439 1.53
    Avg
    Total Wait wait Waits
    Event Waits Timeouts Time (s) (ms) /txn
    log file sync 349,713 150,785 215,674 617 1.8
    db file sequential read 16,955,622 0 31,342 2 85.9
    I hope the above is readable. I'm concerned with the very high number of Waits and Timeouts, particulary around the log file sync event. From reading around I suspect that the disk our redo log sits on isn't fast enough.
    1) Is this conclusion correct, are these timeouts excessively high (70% seems high...)?
    2) I see high waits on almost every other event (but not timeouts), is this pointing towards an incorrect database database setup (give our very high loads of 160 executes second?
    Any help would be much appreciated.
    Jonathan

    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~ % Total
    Event Waits Time (s) Ela Time
    log file sync 349,713 215,674 74.13
    db file sequential read 16,955,622 31,342 10.77
    CPU time 21,787 7.49
    direct path read (lob) 92,762 8,910 3.06
    db file scattered read 4,335,034 4,439 1.53
    Avg
    Total Wait wait Waits
    Event Waits Timeouts Time (s) (ms) /txn
    log file sync 349,713 150,785 215,674 617 1.8
    db file sequential read 16,955,622 0 31,342 2 85.9What's the time frame of this report on?
    It looks like your disk storage can't keep up with the volume of I/O requests from your database.
    The first few thing need to look at, what're IO intensive SQLs in your database. Are these SQLs doing unnecessary full table scan?
    Find out the hot blocks and the objects they belong.
    Check v$session_wait view.
    Is there any other suspicious activity going on in your Server ? Like other program other than Oracle doing high IO activities? Are there any core dump going on?

  • I recently purchased a movie (Prometheus) on my computer through iTunes and every time I try to put it on my iPhone, it fails. How can I successfully sync this video? Mind you, every other file syncs flawlessly.

    I recently purchased a movie (Prometheus) on my computer through iTunes and every time I try to put it on my iPhone, it fails. How can I successfully sync this video? Mind you, every other file syncs flawlessly.

    Can you connect to a wifi source on the device? Enable wifi in settings on the device itself. If you don't have wifi at home, there may be a local hotspot you can use for this.
    The reason this happens is carriers don't want people downloading huge files that would slow their network as well as eat up your data plan.

  • Offline Files sync gives Access Denied on Windows 8.1 Enterprise

    A small number of our staff have now been issued with Windows 8.1 Enterprise hybrid tablet computers, however there is a problem with using Offline Files on them - when synchronising, it responds "Access Denied".
    The tablets have Windows 8.1 Enterprise with all the latest updates on them. Staff users have a home folder on the network under \\server\staff\homes\departmentname\username which gets mapped to U: and their My Documents is redirected there. The server is currently
    Windows Server 2003 R2 SP2.
    We have tried:
    Resetting the Offline Files cache using the FormatDatabase registry key
    Using Group Policy Objects to force Offline Files synchronisation at logon and logoff
    Clearing the local cached copy of the user's profile from the machine and getting them to log back on to recreate it
    Setting up Offline Files event logging to the event viewer - this provides no useful information as it only logs disconnect/reconnect and logoff/logon events
    Forcing Group Policy update using gpupdate /force
    Forcing synchronisation using PowerShell and https://msdn.microsoft.com/en-us/library/windows/desktop/bb309189%28v=vs.85%29.aspx?f=255&MSPPError=-2147217396
    As suggested by http://support.microsoft.com/kb/275461 we gave the All Staff security group Read permissions on F:\Staff (which is the one that is shared as \\server\staff) and then blocked inheritance for folders below that
    We also checked the following:
    The CSC cache has not been relocated
    No error 7023 or event 7023 errors relating to Offline Files are present in the event logs
    The Offline Files service is running
    The OS is already Windows 8.1 Enterprise, so installing the Pro Pack is not applicable
    In HKLM\SOFTWARE\Microsoft\Windows\CurrentVersion\UserState\UserStateTechnologies\ConfigurationControls  all the values are set to 0 and not 1
    We do not use System Center Configuration Manager
    No errors were found in the Folder Redirection event logs
    None of these solved the problem, does anyone have any suggestions?
    Here is the error we are seeing:
    Thanks,
    Dan Jackson (Lead ITServices Technician)
    Long Road Sixth Form College
    Cambridge, UK

    Hi,
    Generally speaking, this problem is most probably occurs at File Server Client. 
    Firstly, please check the sharing file Sync Settings.
    Shared file properties\Sharing\Advanced Sharing\Caching 
    Also check shared file user list, make sure these problematic user account have full permission.
    On the other hand, could you able to access to the shared file directly in Windows Explorer?
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact [email protected]
    Yes, the user can access the shared folder in Windows Explorer. The user has the following permissions:
    Traverse Folder/Execute File
    List Folder/Read Data
    Read Attributes
    Read Extended Attributes
    Create Files/Write Data
    Create Folders/Append Data
    Write Attributes
    Write Extended Attributes
    Delete
    Read Permissions
    Here is a screenshot of how the Caching settings are set up on the top-level Staff share.

  • RMAN backup files are still exist since long time, how to delete?

    Dear sir;
    I'm using the below script to do daily backup, however there are many rman backup files are still exist and consumes HD size, how could I delete these files in daily bases? some files dated in FEB, MAR, APR,
    ============Daily RMAN script=========
    rman target /<<!
    backup incremental level=0 as compressed backupset database format '/u15/rman/full_backup_%U.rman';
    backup archivelog all not backed up 2 times format '/u15/rman/arc_backup_%U.rman';
    backup current controlfile format '/u15/rman/control_%U.rman';
    delete archivelog all backed up 2 times to device type disk completed before 'sysdate-7';
    delete noprompt obsolete;
    ================================END
    Thanks and best regards
    Ali

    Hi;
    Our backup policy should have 7 days; however we have here some files from JAN, FEB,MAR, APR /2012 WHICH ARE BEYOND THE RETENTION DATE and these files should be deleted by executing " delete noprompt obsolete; ".
    All files are exist in /u15/rman/
    -rw-r----- 1 oracle oinstall 1151763968 Jan 21 01:36 arc_backup_7kn19h4a_1_1.rman
    -rw-r----- 1 oracle oinstall 1136882176 Jan 21 01:36 arc_backup_7ln19h4q_1_1.rman
    -rw-r----- 1 oracle oinstall 1135984640 Jan 21 01:36 arc_backup_7mn19h5a_1_1.rman
    -rw-r----- 1 oracle oinstall 1126627328 Jan 21 01:37 arc_backup_7nn19h5q_1_1.rman
    -rw-r----- 1 oracle oinstall 880606720 Mar 12 02:53 arc_backup_7nn5ldhp_1_1.rman
    -rw-r----- 1 oracle oinstall 1093043712 Jan 21 01:37 arc_backup_7on19h6a_1_1.rman
    -rw-r----- 1 oracle oinstall 9797632 Dec 15 01:04 control_04mu7tcp_1_1.rman
    -rw-r----- 1 oracle oinstall 36896768 Mar 3 02:55 control_4cn4tm9k_1_1.rman
    -rw-r----- 1 oracle oinstall 36896768 Mar 4 02:53 control_4on50ahm_1_1.rman
    -rw-r----- 1 oracle oinstall 36896768 Mar 5 02:55 control_56n52v1j_1_1.rman
    -rw-r----- 1 oracle oinstall 16252928 Jan 23 01:40 control_8tn1eq3t_1_1.rman
    -rw-r----- 1 oracle oinstall 16252928 Jan 24 01:40 control_9cn1heg0_1_1.rman
    -rw-r----- 1 oracle oinstall 202940416 Dec 15 01:04 full_backup_01mu7t50_1_1.rman
    -rw-r----- 1 oracle oinstall 1097728 Dec 15 01:04 full_backup_02mu7tcc_1_1.rman
    -rw-r----- 1 oracle oinstall 201285632 Dec 14 01:04 full_backup_0nmu58ou_1_1.rman
    -rw-r----- 1 oracle oinstall 5957304320 Feb 29 02:46 full_backup_2ln4g9l1_1_1.rman
    -rw-r----- 1 oracle oinstall 4128768 Feb 29 02:47 full_backup_2mn4gft8_1_1.rman
    -rw-r----- 1 oracle oinstall 6027075584 Mar 1 02:49 full_backup_32n4o6ov_1_1.rman
    -rw-r----- 1 oracle oinstall 4128768 Mar 1 02:49 full_backup_33n4od66_1_1.rman
    -rw-r----- 1 oracle oinstall 6187171840 Mar 2 02:51 full_backup_3gn4qr50_1_1.rman
    -rw-r----- 1 oracle oinstall 4145152 Mar 2 02:51 full_backup_3hn4r1kn_1_1.rman
    -rw-r----- 1 oracle oinstall 6115786752 Mar 3 02:51 full_backup_40n4tfgu_1_1.rman
    above is a short list of contents.
    to do our daily backup we perform the following script (in daily)
    ==================
    backup incremental level=0 as compressed backupset database format '/u15/rman/full_backup_%U.rman';
    backup archivelog all not backed up 2 times format '/u15/rman/arc_backup_%U.rman';
    backup current controlfile format '/u15/rman/control_%U.rman';
    delete archivelog all backed up 2 times to device type disk completed before 'sysdate-7';
    delete noprompt obsolete;
    ==================
    Thanks and best regards
    Ali

  • If I turn on Adobe File Sync will that duplicate files in my Dropbox? (Unnecessary and I can't afford the bandwidth).

    I have gigabytes of Adobe CC files arranged in appropriate client folders in Dropbox (and mirrored on my local hard drive).
    It looks like if I turn File Sync on (needed to access Assets) some or all of this will be duplicated and separated from non-Adobe data in each client file.
    Messy, unnecessary and 100GB of bandwidth I don't want to pay for! (I'm on 4G so it's expensive.)

    MichaelGli2,
    The Creative Cloud desktop application will only sync files that are in your "Creative Cloud Files" folder on your local machine.
    By default, the "Creative Cloud Files" folder is in your user's home folder (/Users/<yourusername>/Creative Cloud Files/ on Mac, or C:\Users\<yourusername>\Creative Cloud Files\ on Windows).
    Unless you have deliberately changed the location of your Creative Cloud Files folder yourself so that it is inside your local Dropbox folder, there will be no duplication.

  • Log file sync question

    Metalink note 34592.1 has been mentioned several times in this forum as well as elsewhere, notably here
    http://christianbilien.wordpress.com/2008/02/12/the-%E2%80%9Clog-file-sync%E2%80%9D-wait-event-is-not-always-spent-waiting-for-an-io/
    The question I have relates to the stated breakdown of 'log file sync' wait event:
    1. Wakeup LGWR if idle
    2. LGWR gathers the redo to be written and issue the I/O
    3. Time for the log write I/O to complete
    4. LGWR I/O post processing
    5. LGWR posting the foreground/user session that the write has completed
    6. Foreground/user session wakeup
    Since the note says that the system 'read write' statistic includes steps 2 and 3, the suggestion is that the difference between it and 'log file sync' is due to CPU related work on steps 1, 4, 5 and 6 (or on waiting on the CPU run queue).
    Christian's article, quoted above, theorises about 'CPU storms' and the Metalink note also suggests that steps 5 and 6 could be costly.
    However, my understanding of how LGWR works is that if it is already in the process of writing out one set of blocks (let us say associated with a commit of transaction 'X' amongst others) at the time a another transaction (call it transaction 'Y') commits, then LGWR will not commence the write of the commit for transaction 'Y' until the I/Os associated with the commit of transaction 'X' complete.
    So, if I have an average 'redo write' time of, say, 12ms and a 'log file sync' time of, say 34ms (yes, of course these are real numbers :-)) then I would have thought that this 22ms delay was due at least partly to LGWR 'falling behind' in it's work.
    Nonetheless, it seems to me that this extra delay could only be a maximum of 12ms so this still leaves 10ms (34 - 12 -12) that can only be accounted for by CPU usage.
    Clearly, my analsys contains a lot of conjecture, hence this note.
    Can anybody point me in the direction of some facts?

    Tony Hasler wrote:
    Metalink note 34592.1 has been mentioned several times in this forum as well as elsewhere, notably here
    http://christianbilien.wordpress.com/2008/02/12/the-%E2%80%9Clog-file-sync%E2%80%9D-wait-event-is-not-always-spent-waiting-for-an-io/
    The question I have relates to the stated breakdown of 'log file sync' wait event:
    1. Wakeup LGWR if idle
    2. LGWR gathers the redo to be written and issue the I/O
    3. Time for the log write I/O to complete
    4. LGWR I/O post processing
    5. LGWR posting the foreground/user session that the write has completed
    6. Foreground/user session wakeup
    Since the note says that the system 'read write' statistic includes steps 2 and 3, the suggestion is that the difference between it and 'log file sync' is due to CPU related work on steps 1, 4, 5 and 6 (or on waiting on the CPU run queue).
    Christian's article, quoted above, theorises about 'CPU storms' and the Metalink note also suggests that steps 5 and 6 could be costly.
    However, my understanding of how LGWR works is that if it is already in the process of writing out one set of blocks (let us say associated with a commit of transaction 'X' amongst others) at the time a another transaction (call it transaction 'Y') commits, then LGWR will not commence the write of the commit for transaction 'Y' until the I/Os associated with the commit of transaction 'X' complete.
    So, if I have an average 'redo write' time of, say, 12ms and a 'log file sync' time of, say 34ms (yes, of course these are real numbers :-)) then I would have thought that this 22ms delay was due at least partly to LGWR 'falling behind' in it's work.
    Nonetheless, it seems to me that this extra delay could only be a maximum of 12ms so this still leaves 10ms (34 - 12 -12) that can only be accounted for by CPU usage.
    Clearly, my analsys contains a lot of conjecture, hence this note.
    Can anybody point me in the direction of some facts?It depends on what you mean by facts - presumably only the people who wrote the code know what really happens, the rest of us have to guess.
    You're right about point 1 in the MOS note: it should include "or wait for current lgwr write and posts to complete".
    This means, of course, that your session could see its "log file sync" taking twice the "redo write time" because it posted lgwr just after lgwr has started to write - so you have to wait two write and post cycles. Generally the statistical effects will reduce this extreme case.
    You've been pointed to the two best bits of advice on the internet: As Kevin points out, if you have lgwr posting a lot of processes in one go it may stall as they wake up, so the batch of waiting processes has to wait extra time; and as Riyaj points out - there's always dtrace (et al.) if you want to see what's really happening. (Tanel has some similar notes, I think, on LFS).
    If you're stuck with Oracle diagnostics only then:
    redo size / redo synch writes for sessions will tell you the typical "commit size"
    redo size + redo wastage / redo writes for lgwr will tell you the typical redo write size
    If you have a significant number of small processes "commit sizes" per write (more than CPU count, say) then you may be looking at Kevin's storm.
    Watch out for a small number of sessions with large commit sizes running in parallel with a large number of sessions with small commit sizes - this could make all the "small" processes run at the speed of the "large" processes.
    It's always worth looking at the event histogram for the critical wait events to see if their patterns offer any insights.
    Regards
    Jonathan Lewis

  • Log file sync top event during performance test -av 36ms

    Hi,
    During the performance test for our product before deployment into product i see "log file sync" on top with Avg wait (ms) being 36 which i feel is too high.
                                                               Avg
                                                              wait   % DB
    Event                                 Waits     Time(s)   (ms)   time Wait Class
    log file sync                       208,327       7,406     36   46.6 Commit
    direct path write                   646,833       3,604      6   22.7 User I/O
    DB CPU                                            1,599          10.1
    direct path read temp             1,321,596         619      0    3.9 User I/O
    log buffer space                      4,161         558    134    3.5 ConfiguratAlthough testers are not complaining about the performance of the appplication , we ,DBAs, are expected to be proactive about the any bad signals from DB.
    I am not able to figure out why "log file sync" is having such slow response.
    Below is the snapshot from the load profile.
                  Snap Id      Snap Time      Sessions Curs/Sess
    Begin Snap:    108127 16-May-13 20:15:22       105       6.5
      End Snap:    108140 16-May-13 23:30:29       156       8.9
       Elapsed:              195.11 (mins)
       DB Time:              265.09 (mins)
    Cache Sizes                       Begin        End
    ~~~~~~~~~~~                  ---------- ----------
                   Buffer Cache:     1,168M     1,136M  Std Block Size:         8K
               Shared Pool Size:     1,120M     1,168M      Log Buffer:    16,640K
    Load Profile              Per Second    Per Transaction   Per Exec   Per Call
    ~~~~~~~~~~~~         ---------------    --------------- ---------- ----------
          DB Time(s):                1.4                0.1       0.02       0.01
           DB CPU(s):                0.1                0.0       0.00       0.00
           Redo size:          607,512.1           33,092.1
       Logical reads:            3,900.4              212.5
       Block changes:            1,381.4               75.3
      Physical reads:              134.5                7.3
    Physical writes:              134.0                7.3
          User calls:              145.5                7.9
              Parses:               24.6                1.3
         Hard parses:                7.9                0.4
    W/A MB processed:          915,418.7           49,864.2
              Logons:                0.1                0.0
            Executes:               85.2                4.6
           Rollbacks:                0.0                0.0
        Transactions:               18.4Some of the top background wait events:
    ^LBackground Wait Events       DB/Inst: Snaps: 108127-108140
    -> ordered by wait time desc, waits desc (idle events last)
    -> Only events with Total Wait Time (s) >= .001 are shown
    -> %Timeouts: value of 0 indicates value was < .5%.  Value of null is truly 0
                                                                 Avg
                                            %Time Total Wait    wait    Waits   % bg
    Event                             Waits -outs   Time (s)    (ms)     /txn   time
    log file parallel write         208,563     0      2,528      12      1.0   66.4
    db file parallel write            4,264     0        785     184      0.0   20.6
    Backup: sbtbackup                     1     0        516  516177      0.0   13.6
    control file parallel writ        4,436     0         97      22      0.0    2.6
    log file sequential read          6,922     0         95      14      0.0    2.5
    Log archive I/O                   6,820     0         48       7      0.0    1.3
    os thread startup                   432     0         26      60      0.0     .7
    Backup: sbtclose2                     1     0         10   10094      0.0     .3
    db file sequential read           2,585     0          8       3      0.0     .2
    db file single write                560     0          3       6      0.0     .1
    log file sync                        28     0          1      53      0.0     .0
    control file sequential re       36,326     0          1       0      0.2     .0
    log file switch completion            4     0          1     207      0.0     .0
    buffer busy waits                     5     0          1     116      0.0     .0
    LGWR wait for redo copy             924     0          1       1      0.0     .0
    log file single write                56     0          1       9      0.0     .0
    Backup: sbtinfo2                      1     0          1     500      0.0     .0During a previous perf test , things didnt look this bad for "log file sync. Few sections from the comparision report(awrddprt.sql)
    {code}
    Workload Comparison
    ~~~~~~~~~~~~~~~~~~~ 1st Per Sec 2nd Per Sec %Diff 1st Per Txn 2nd Per Txn %Diff
    DB time: 0.78 1.36 74.36 0.02 0.07 250.00
    CPU time: 0.18 0.14 -22.22 0.00 0.01 100.00
    Redo size: 573,678.11 607,512.05 5.90 15,101.84 33,092.08 119.13
    Logical reads: 4,374.04 3,900.38 -10.83 115.14 212.46 84.52
    Block changes: 1,593.38 1,381.41 -13.30 41.95 75.25 79.38
    Physical reads: 76.44 134.54 76.01 2.01 7.33 264.68
    Physical writes: 110.43 134.00 21.34 2.91 7.30 150.86
    User calls: 197.62 145.46 -26.39 5.20 7.92 52.31
    Parses: 7.28 24.55 237.23 0.19 1.34 605.26
    Hard parses: 0.00 7.88 100.00 0.00 0.43 100.00
    Sorts: 3.88 4.90 26.29 0.10 0.27 170.00
    Logons: 0.09 0.08 -11.11 0.00 0.00 0.00
    Executes: 126.69 85.19 -32.76 3.34 4.64 38.92
    Transactions: 37.99 18.36 -51.67
    First Second Diff
    1st 2nd
    Event Wait Class Waits Time(s) Avg Time(ms) %DB time Event Wait Class Waits Time(s) Avg Time
    (ms) %DB time
    SQL*Net more data from client Network 2,133,486 1,270.7 0.6 61.24 log file sync Commit 208,355 7,407.6
    35.6 46.57
    CPU time N/A 487.1 N/A 23.48 direct path write User I/O 646,849 3,604.7
    5.6 22.66
    log file sync Commit 99,459 129.5 1.3 6.24 log file parallel write System I/O 208,564 2,528.4
    12.1 15.90
    log file parallel write System I/O 100,732 126.6 1.3 6.10 CPU time N/A 1,599.3
    N/A 10.06
    SQL*Net more data to client Network 451,810 103.1 0.2 4.97 db file parallel write System I/O 4,264 784.7 1
    84.0 4.93
    -direct path write User I/O 121,044 52.5 0.4 2.53 -SQL*Net more data from client Network 7,407,435 279.7
    0.0 1.76
    -db file parallel write System I/O 986 22.8 23.1 1.10 -SQL*Net more data to client Network 2,714,916 64.6
    0.0 0.41
    {code}
    *To sum it sup:
    1. Why is the IO response getting such an hit during the new perf test? Please suggest*
    2. Does the number of DB writer impact "log file sync" wait event? We have only one DB writer as the number of cpu on the host is only 4
    {code}
    select *from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    PL/SQL Release 11.1.0.7.0 - Production
    CORE 11.1.0.7.0 Production
    TNS for HPUX: Version 11.1.0.7.0 - Production
    NLSRTL Version 11.1.0.7.0 - Production
    {code}
    Please let me know if you would like to see any other stats.
    Edited by: Kunwar on May 18, 2013 2:20 PM

    1. A snapshot interval of 3 hours always generates meaningless results
    Below are some details from the 1 hour interval AWR report.
    Platform                         CPUs Cores Sockets Memory(GB)
    HP-UX IA (64-bit)                   4     4       3      31.95
                  Snap Id      Snap Time      Sessions Curs/Sess
    Begin Snap:    108129 16-May-13 20:45:32       140       8.0
      End Snap:    108133 16-May-13 21:45:53       150       8.8
       Elapsed:               60.35 (mins)
       DB Time:              140.49 (mins)
    Cache Sizes                       Begin        End
    ~~~~~~~~~~~                  ---------- ----------
                   Buffer Cache:     1,168M     1,168M  Std Block Size:         8K
               Shared Pool Size:     1,120M     1,120M      Log Buffer:    16,640K
    Load Profile              Per Second    Per Transaction   Per Exec   Per Call
    ~~~~~~~~~~~~         ---------------    --------------- ---------- ----------
          DB Time(s):                2.3                0.1       0.03       0.01
           DB CPU(s):                0.1                0.0       0.00       0.00
           Redo size:          719,553.5           34,374.6
       Logical reads:            4,017.4              191.9
       Block changes:            1,521.1               72.7
      Physical reads:              136.9                6.5
    Physical writes:              158.3                7.6
          User calls:              167.0                8.0
              Parses:               25.8                1.2
         Hard parses:                8.9                0.4
    W/A MB processed:          406,220.0           19,406.0
              Logons:                0.1                0.0
            Executes:               88.4                4.2
           Rollbacks:                0.0                0.0
        Transactions:               20.9
    Top 5 Timed Foreground Events
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
                                                               Avg
                                                              wait   % DB
    Event                                 Waits     Time(s)   (ms)   time Wait Class
    log file sync                        73,761       6,740     91   80.0 Commit
    log buffer space                      3,581         541    151    6.4 Configurat
    DB CPU                                              348           4.1
    direct path write                   238,962         241      1    2.9 User I/O
    direct path read temp               487,874         174      0    2.1 User I/O
    Background Wait Events       DB/Inst: Snaps: 108129-108133
    -> ordered by wait time desc, waits desc (idle events last)
    -> Only events with Total Wait Time (s) >= .001 are shown
    -> %Timeouts: value of 0 indicates value was < .5%.  Value of null is truly 0
                                                                 Avg
                                            %Time Total Wait    wait    Waits   % bg
    Event                             Waits -outs   Time (s)    (ms)     /txn   time
    log file parallel write          61,049     0      1,891      31      0.8   87.8
    db file parallel write            1,590     0        251     158      0.0   11.6
    control file parallel writ        1,372     0         56      41      0.0    2.6
    log file sequential read          2,473     0         50      20      0.0    2.3
    Log archive I/O                   2,436     0         20       8      0.0     .9
    os thread startup                   135     0          8      60      0.0     .4
    db file sequential read             668     0          4       6      0.0     .2
    db file single write                200     0          2       9      0.0     .1
    log file sync                         8     0          1     152      0.0     .1
    log file single write                20     0          0      21      0.0     .0
    control file sequential re       11,218     0          0       0      0.1     .0
    buffer busy waits                     2     0          0     161      0.0     .0
    direct path write                     6     0          0      37      0.0     .0
    LGWR wait for redo copy             380     0          0       0      0.0     .0
    log buffer space                      1     0          0      89      0.0     .0
    latch: cache buffers lru c            3     0          0       1      0.0     .0     2 The log file sync is a result of commit --> you are committing too often, maybe even every individual record.
    Thanks for explanation. +Actually my question is WHY is it so slow (avg wait of 91ms)+3 Your IO subsystem hosting the online redo log files can be a limiting factor.
    We don't know anything about your online redo log configuration
    Below is my redo log configuration.
        GROUP# STATUS  TYPE    MEMBER                                                       IS_
             1         ONLINE  /oradata/fs01/PERFDB1/redo_1a.log                           NO
             1         ONLINE  /oradata/fs02/PERFDB1/redo_1b.log                           NO
             2         ONLINE  /oradata/fs01/PERFDB1/redo_2a.log                           NO
             2         ONLINE  /oradata/fs02/PERFDB1/redo_2b.log                           NO
             3         ONLINE  /oradata/fs01/PERFDB1/redo_3a.log                           NO
             3         ONLINE  /oradata/fs02/PERFDB1/redo_3b.log                           NO
    6 rows selected.
    04:13:14 perf_monitor@PERFDB1> col FIRST_CHANGE# for 999999999999999999
    04:13:26 perf_monitor@PERFDB1> select *from v$log;
        GROUP#    THREAD#  SEQUENCE#      BYTES    MEMBERS ARC STATUS                 FIRST_CHANGE# FIRST_TIME
             1          1      40689  524288000          2 YES INACTIVE              13026185905545 18-MAY-13 01:00
             2          1      40690  524288000          2 YES INACTIVE              13026185931010 18-MAY-13 03:32
             3          1      40691  524288000          2 NO  CURRENT               13026185933550 18-MAY-13 04:00Edited by: Kunwar on May 18, 2013 2:46 PM

  • Unable to launch File sync and to update Adobe Exchange panel in Illustrator CC 2014

    I suppose that these two problems might be related, as I need the file sync to sync the extensions?
    Anyway, I’m unable to launch a file sync, and when I update the Adobe Exchange panel in Illustrator, it asks for a reboot, and after the reboot, it asks again for an update…
    Thanks for the help!

    I've since received information of a workaround - listed in a different thread: Unable to install non-standard updates - Cannot launch/update Adobe Digital Publishing for InDesign CC 2014
    Try the following to enable updates:
    1) Navigate to C:\Program Files (x86)\Common Files\Adobe\AAMUpdaterInventory\1.0 and you will find a .dat file ( AdobeUpdaterAdminPrefs.dat).
    2) Please open the file in question and you will find a line item : <Suppressed>1</Suppressed> .
    3) Please change the Value from 1 to 0 and save the file back in the same location.
    On a Mac, essentially the same, however location: /Library/Application Support/Adobe/AAMUpdaterInventory/1.0/AdobeUpdaterAdminPrefs.dat
    I was given further advice that DPS toolkit updates are not available when you create the deployment package from Adobe Application Manager Enterprise Edition on Mac OS. (I believe this may be the same using Creative Cloud Packager as you've found)
    Deployment DPS Toolkit Update| AAMEE |Mac OS
    The above article suggests that you can download and install the DPS update silently, however, an end user must be logged into computer - not really feasible for enterprise environments.
    So I guess the process for this would be:
    1)Copy the contents of the DPS update package to remote machine
    2)Enable updates by modifying the AdobeUpdaterAdminPrefs.dat file
    3)Remotely launch the DPS update silently (when someone is logged onto machine)
    4)Disable updates again when complete...

  • Why can't I use the creative cloud connection file syncing?

    I hear people raving about it on twitter, so why am I only seeing a 'coming soon' message in the files tab of my creative cloud interface? I would very much like to use the file syncing to work between two computers! What is going on here? Is there some kind of area restriction going on for this service?

    Sign up for early access to File and Font Syncing. It's being rolled out slowly to work out the early glitches:
    http://blogs.adobe.com/creativecloud/file-and-typekit-font-sync-rolling-out/

  • File Sync does not stay on in Adobe Creative Cloud - OSX Mavericks

    To download assets (clip art) from Creative Cloud Market, File Sync in CC must be on. I turn it on, but when I try to download it turns itself off. Anyone else have this issue?

    Please try the workaround posted here:
    Re: Creative cloud doesnt create a cloud folder on my Users Folder?
    This seems to have worked for a few other users.  Hopefully we'll be able to release a fix to this issue soon.
    Thanks,
    Elba

  • How to call sql loader ctrl file with in the pl/sql procedure

    Hi Friends,
    I am doing a project related with the transferring data using queues. In the queue , I will get a tab delimited data in the form of CLOB variable/message. I do want to store that dat in the oracle table.
    When updating data into the table ,
    1. Don't want to write that data into a file.( Want to access directly after dequeueing from the specfic queue).
    2. As the data is in tab delimited form, I want to use sql loader concept.
    How do I call the sql loader ctrl file with in my pl/sql procedure. When I searched , most of the forums recommending external procedure or Java program.
    Please Guide me on this issue. my preferrence is pl sql, But don't know about external procedure . If no other way , I will try Java.
    I am using oracle 9.2.0.8.0.
    Thanks in advance,
    Vimal..

    Neither SQL*Loader nor external tables are designed to read data from a CLOB stored in the database. They both work on files stored on the file system. If you don't want the data to be written to a file, you're going to have to roll your own parsing code. This is certainly possible. But it is going to be far less efficient than either SQL*Loader or external tables. And it's likely to involve quite a bit more code.
    The simplest possible thing that could work would be to use something like Tom Kyte's string tokenization package to read a line from the CLOB, break it into the component pieces, and then store the different tokens in a meaningful collection (i.e. an object type or a record type that corresponds to the table definition). Of course, you'll need to handle things like converting strings to numbers or dates, rejecting rows, writing log files, etc.
    Justin

Maybe you are looking for

  • Difference between path and classpath

    Difference between path and classpath?

  • Topic properties

    I am having a real problem with topic properties not cooperating. For some reason, RoboHelp insists on making some of my topic titles lowercase. For example - the title Copyright. It keeps changing the "c" to lowercase. Also, I want the filenames to

  • What Language is SAP written in?

    Hi All,           Can anyone plaese tell me in what language is SAP written in. Not SAP coding language, but the source code of SAP...

  • Question for more experienced Iphone users!

    Hey guys (and gals), I'm very new to the Iphone (3GS, 32Gb) and would like someone to explain a small issue if possible. Firstly, if this has been asked and answered elsewhere then please forgive me. Secondly, it's a trivial isse (I hope) but its bug

  • Cisco CME / 7925G: Leading zero stripped in Placed Calls list

    Hi, we are using a "0" as a prefix to dialout to the PSTN (ISDN PRI) on our CME router (2921). The leading 0 will be stripped off  on the dial peer via a voice translation rule: voice translation-rule 11  rule 1 /^00\(.*\)/ /0\1/ voice translation-pr