Log files in bdump and udump

Respected Sir
Its very serious because log files in bump and udump are created frequently and they take too much space.
The problem is that i am not able to understand or solve these problem.I am pasting few log here.
BDUMP->
Dump file h:\oracle\product\10.2.0\admin\orcl2\bdump\orcl2_lgwr_1364.trc
Thu Jul 30 09:15:58 2009
ORACLE V10.2.0.1.0 - Production vsnsta=0
vsnsql=14 vsnxtr=3
Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
With the Partitioning, OLAP and Data Mining options
Windows Server 2003 Version V5.2 Service Pack 1
CPU : 4 - type 586, 1 Physical Cores
Process Affinity : 0x00000000
Memory (Avail/Total): Ph:3481M/4094M, Ph+PgF:5024M/5975M, VA:1290M/2047M
Instance name: orcl2
Redo thread mounted by this instance: 1
Oracle process number: 6
Windows thread id: 1364, image: ORACLE.EXE (LGWR)
*** 2009-07-30 09:15:58.687
*** SERVICE NAME:() 2009-07-30 09:15:58.671
*** SESSION ID:(166.1) 2009-07-30 09:15:58.671
Media recovery not enabled or manual archival only 0x10000
Maximum redo generation record size = 156160 bytes
Maximum redo generation change vector size = 150672 bytes
*** 2009-07-30 10:34:04.875
Media recovery not enabled or manual archival only 0x10000
*** 2009-07-30 10:34:28.312
Media recovery not enabled or manual archival only 0x10000
*** 2009-07-30 11:23:36.000
Media recovery not enabled or manual archival only 0x10000
*** 2009-07-30 18:06:53.718
Media recovery not enabled or manual archival only 0x10000
*** 2009-07-30 22:02:26.734
Media recovery not enabled or manual archival only 0x10000
*** 2009-07-31 04:55:48.312
Media recovery not enabled or manual archival only 0x10000
UDUMP
Dump file h:\oracle\product\10.2.0\admin\orcl2\udump\orcl2_ora_192.trc
Wed Jul 29 12:34:17 2009
ORACLE V10.2.0.1.0 - Production vsnsta=0
vsnsql=14 vsnxtr=3
Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
With the Partitioning, OLAP and Data Mining options
Windows Server 2003 Version V5.2 Service Pack 1
CPU : 4 - type 586, 1 Physical Cores
Process Affinity : 0x00000000
Memory (Avail/Total): Ph:3653M/4094M, Ph+PgF:5205M/5975M, VA:1325M/2047M
Instance name: orcl2
Redo thread mounted by this instance: 0 <none>
Oracle process number: 149
Windows thread id: 192, image: ORACLE.EXE (SHAD)
*** SERVICE NAME:() 2009-07-29 12:34:17.703
*** SESSION ID:(159.1) 2009-07-29 12:34:17.703
kccsga_update_ckpt: num_1 = 8, num_2 = 0, num_3 = 0, lbn_2 = 0, lbn_3 = 0
Successfully allocated 3 recovery slaves
Using 364 overflow buffers per recovery slave
Thread 1 checkpoint: logseq 317, block 2, scn 2070858
cache-low rba: logseq 317, block 3
on-disk rba: logseq 317, block 248, scn 2071389
start recovery at logseq 317, block 3, scn 0
----- Redo read statistics for thread 1 -----
Read rate (ASYNC): 122Kb in 0.35s => 0.34 Mb/sec
Total physical reads: 4096Kb
Longest record: 8Kb, moves: 0/201 (0%)
Longest LWN: 10Kb, moves: 0/63 (0%), moved: 0Mb
Last redo scn: 0x0000.001f9b5c (2071388)
----- Recovery Hash Table Statistics ---------
Hash table buckets = 32768
Longest hash chain = 1
Average hash chain = 92/92 = 1.0
Max compares per lookup = 1
Avg compares per lookup = 598/690 = 0.9
*** 2009-07-29 12:34:24.015
KCRA: start recovery claims for 92 data blocks
*** 2009-07-29 12:34:24.343
KCRA: blocks processed = 92/92, claimed = 92, eliminated = 0
*** 2009-07-29 12:34:24.390
Recovery of Online Redo Log: Thread 1 Group 1 Seq 317 Reading mem 0
----- Recovery Hash Table Statistics ---------
Hash table buckets = 32768
Longest hash chain = 1
Average hash chain = 92/92 = 1.0
Max compares per lookup = 1
Avg compares per lookup = 690/690 = 1.0
Thanks

You will still get lots of messages. Which messages depends on many things, some of which are changes you can make in the init files. On others, Oracle only knows to get rid of them if people log support calls. Whether they decide to get rid of them is mysterious, but probably not likely on a version soon to stop development.
Don't get mad at Hemant, he's just an experienced guy trying to help you, for free. You can get mad at me though, because my cares are elsewhere. You can get mad at Oracle too, because this is evidence of sloppiness on their part. But be nice to the support people, it's not their fault, they're being paid to help you. Though you can get mad at them if they don't help you after you've paid them, and you've jumped through hoops to get their attention.

Similar Messages

  • Log file in session and call transaction methods

    Hi Experts,
    How to see log file in session and call transaction methods?

    hi
    If its a session then goto SM35, select ur session and then u can see log button over there at the top... that will give the log
    If its a call transaction then in ur program u need to declare internal table IT_MSGS like this
    DATA:  IT_MSGS LIKE BDCMSGCOLL OCCURS 0 WITH HEADER LINE.
      LOOP AT IT_SALES.
        PERFORM POPULATE_BDC.
        CALL TRANSACTION 'VA01' USING IT_BDCDATA
                                      MODE MODE
                                      UPDATE UPDATE
                                      MESSAGES INTO IT_MSGS.
        IF NOT IT_MSGS[] IS INITIAL.
          LOOP AT IT_MSGS.
            CALL FUNCTION 'FORMAT_MESSAGE'
              EXPORTING
                ID        = IT_MSGS-MSGID
                LANG      = 'EN'
                NO        = IT_MSGS-MSGNR
                V1        = IT_MSGS-MSGV1
                V2        = IT_MSGS-MSGV2
                V3        = IT_MSGS-MSGV3
                V4        = IT_MSGS-MSGV4
              IMPORTING
                MSG       = V_MSG
              EXCEPTIONS
                NOT_FOUND = 1
                OTHERS    = 2.
            IF SY-SUBRC <> 0.
              MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
                      WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
            ENDIF.
            WRITE:/ V_MSG.
          ENDLOOP.
        ENDIF.
      ENDLOOP.
    This will create a log for you, dont forget to award points if found helpful

  • Very high log file sequential read and control file sequential read waits?

    I have a 10.2.0.4 database and have 5 streams capture processes running to replicate data to another database. However I am seeing very high
    log file sequential read and control file sequential read by the capture procesess. This is causing slowness in the database as the databass is wasting so much time on these wait events. From AWR report
    Elapsed: 20.12 (mins)
    DB Time: 67.04 (mins)
    and From top 5 wait events
    Event Waits Time(s) Avg Wait(ms) % Total Call Time Wait Class
    CPU time 1,712 42.6
    log file sequential read 99,909 683 7 17.0 System I/O
    log file sync 49,702 426 9 10.6 Commit
    control file sequential read262,625 384 1 9.6 System I/O
    db file sequential read 41,528 378 9 9.4 User I/O
    Oracle support hasn't been of much help, other than wasting my 10 days and telling me to try this and try that.
    Do you have streams running in your environment, are you experiencing this wait. Have you done anything to resolve these waits..
    Thanks

    Welcome to the forums.
    There is insufficient information in what you have posted to know that your analysis of the situation is correct or anything about your Streams environment.
    We don't know what you are replicating. Not size, not volume, not type of capture, not rules, etc.
    We don't know the distance over which it is being replicated ... 10 ft. or 10 light years.
    We don't have any AWR or ASH data to look at.
    etc. etc. etc. If this is what you provided Oracle Support it is no wonder they were unable to help you.
    To diagnose this problem, if one exists, requires someone on-site or with a very substantial body of data which you have not provided. The first step is to fill in the answers to all of the obvious first level questions. Then we will likely come back with a second level of questioning.
    But when you do ... do not post here. Your questions are not "Database General" they are specific to Streams and there is a Streams forum specifically for them.
    Thank you.

  • Copy System Error - No sapinst workdir retrieved yet. Therefore, no SAPInst log files were written and can be retrieved

    Hi all,
    I am getting a strange error when performing a "Copy System" that I am unable to resolve. The error occurs during step "User Management". The step is supposed to start SAPinst and install the new users.
    "No sapinst workdir retrieved yet. Therefore, no SAPinst log files were written and can be retrieved"
    "LVM Error Host Agent Operation 'xxxxx' failed (OperationExecption. FaultCode: '127', Message: Command execution failed. ')"
    I have checked the C:\temp\VCM_sapinst_PR\sapinst2014_09_10_09_48_28 folder and the only file there is the "inifile.xml".
    The copy is based on vmware, the SAP systems are Windows 2008 R2 with MSSQL 2008. The host agent Version is 190 and the SAP kernel is 600.
    Under "System and AS Provistion Configuration" I have enabled Provistioning type "System Copy" for release 7.X and Windows NT. The path is C:\INST\LVM_WIN_731\SPM and exist on all systems locally. The SWPM Version is SWPM10SP06_3-20009707.SAR.
    The only thing I have found that kind of fits the error Message seems to be related to Linux/Unix:
    http://wiki.scn.sap.com/wiki/display/TechTSG/(ACC)+SAPControl+Fails+with+HTTP+401+Unauthorized+Error
    Does anyone have an solution for this issue?
    BR/
    Stian
    Added ASUI.log file from ./hostctrl/wok folder and a text file with the error.

    Hi all,
    The problem was resolved by the OSS Message.
    The problem was with the SAP Host Agent 720 patch 190. Quote from the OSS Message:
    "Try downgrading to SAP Host Agent patch 189. The problem seems to be with duplicate parameters for the sapinst Call (exitOnError) wich was add by the LVM and since patch 190 is also added by the SAP Host Agent, end reexecute your process."
    The issue should be resolved in a future patch for SAP LVM, but if you encounter this error before it has been released, then try to downgrade to 189 like i did.
    Thanks to SAP Support v/ Jochen for resolving this issue.
    Best Regards,
    Stian Eiken

  • File does not exist: /www/public_html/null, referer:  - error in log file from IE8 and IE9

    I just updated an existing slide show that was created several months back. Since loading my new set of files to our web server, we keep getting the following error in our server log files when someone loads our page in IE8 and IE9:
    File does not exist: /www/public_html/null, referer:
    One of our programmers tracked the error to this line of code, but we're not sure what's causing it:
    Muse.Utils.addSelectorFn('#slideshowu70', function(elem) { new WebPro.Widget.ContentSlideShow(elem, {autoPlay:true,displayInterval:6000,slideLinkStopsSlideShow:false,transitionStyle:'horizo ntal',lightboxEnabled_runtime:false,shuffle:false}); });/* #slideshowu70 */
    Any suggestions?
    Thanks!

    The errors aren't showing up on the client side, only in the server access logs. Every time the page is loaded with IE8 or IE9 and the Muse.Utils.addSelectorFn with the #slideshowu70 line is hit, it generates four errors in the server access logs.
    Our admin guy pulled the errors from where it was clicked on from the Adobe forum. He said it also appeared that you tested it with www added to the url (if this was when you were testing it)...
    [Mon Feb 04 13:58:57 2013] [error] [client 121.242.198.2] File does not exist: /www/public_html/null, referer:http://stingrayboats.com/
    [Mon Feb 04 13:59:00 2013] [error] [client 121.242.198.2] File does not exist: /www/public_html/null, referer:http://stingrayboats.com/
    [Mon Feb 04 13:59:08 2013] [error] [client 121.242.198.2] File does not exist: /www/public_html/null, referer:http://stingrayboats.com/
    [Mon Feb 04 13:59:33 2013] [error] [client 121.242.198.2] File does not exist: /www/public_html/null, referer:http://www.stingrayboats.com/
    [Mon Feb 04 13:59:43 2013] [error] [client 121.242.198.2] File does not exist: /www/public_html/null, referer:http://www.stingrayboats.com/
    The reason we find it strange is because it did not result in errors with the previous version of the slideshow. The only thing that changed this time around is that I removed some slides and added additional ones, but I noticed that there is a lot of difference in the code and even the scripts that are used. And I did copy over all of the scripts, css files, etc., so it's using all of the new source files.
    I was hoping it was something that you guys had noticed if you review the access logs when you're testing. While it works perfectly on the client side, the IT guys do go through our access logs, so it would be nice to eliminate the errors.
    Thanks for looking at it!

  • Statspack: High log file sync timeouts and waits

    Hi all,
    Please see an extract from our statpack report:
    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~ % Total
    Event Waits Time (s) Ela Time
    log file sync 349,713 215,674 74.13
    db file sequential read 16,955,622 31,342 10.77
    CPU time 21,787 7.49
    direct path read (lob) 92,762 8,910 3.06
    db file scattered read 4,335,034 4,439 1.53
    Avg
    Total Wait wait Waits
    Event Waits Timeouts Time (s) (ms) /txn
    log file sync 349,713 150,785 215,674 617 1.8
    db file sequential read 16,955,622 0 31,342 2 85.9
    I hope the above is readable. I'm concerned with the very high number of Waits and Timeouts, particulary around the log file sync event. From reading around I suspect that the disk our redo log sits on isn't fast enough.
    1) Is this conclusion correct, are these timeouts excessively high (70% seems high...)?
    2) I see high waits on almost every other event (but not timeouts), is this pointing towards an incorrect database database setup (give our very high loads of 160 executes second?
    Any help would be much appreciated.
    Jonathan

    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~ % Total
    Event Waits Time (s) Ela Time
    log file sync 349,713 215,674 74.13
    db file sequential read 16,955,622 31,342 10.77
    CPU time 21,787 7.49
    direct path read (lob) 92,762 8,910 3.06
    db file scattered read 4,335,034 4,439 1.53
    Avg
    Total Wait wait Waits
    Event Waits Timeouts Time (s) (ms) /txn
    log file sync 349,713 150,785 215,674 617 1.8
    db file sequential read 16,955,622 0 31,342 2 85.9What's the time frame of this report on?
    It looks like your disk storage can't keep up with the volume of I/O requests from your database.
    The first few thing need to look at, what're IO intensive SQLs in your database. Are these SQLs doing unnecessary full table scan?
    Find out the hot blocks and the objects they belong.
    Check v$session_wait view.
    Is there any other suspicious activity going on in your Server ? Like other program other than Oracle doing high IO activities? Are there any core dump going on?

  • I am geting errors in my database alert log file ora-12012 and ora 44003

    Hi all,
    I am using oracle 10g ...10.2.0.1.0 and OS is solaris 10.......
    i am geting these errors in my database alert log file last two days..i am not geting why it is coming?
    ORA-12012: error on auto execute of job 8887
    ORA-44003: invalid SQL name
    and trace file is i got i am posting here please help any one is aware about this....
    System name: SunOS
    Node name: obrmzone1
    Release: 5.10
    Version: Generic_137111-06
    Machine: sun4v
    Instance name: ------
    Redo thread mounted by this instance: 1
    Oracle process number: 144
    Unix process pid: 26874, image:--------- (J000)
    *** ACTION NAME:(AUTO_SPACE_ADVISOR_JOB) 2011-06-10 22:00:49.372
    *** MODULE NAME:(DBMS_SCHEDULER) 2011-06-10 22:00:49.372
    *** SERVICE NAME:(SYS$USERS) 2011-06-10 22:00:49.372
    *** SESSION ID:(110.7114) 2011-06-10 22:00:49.372
    *** 2011-06-10 22:00:49.372
    ORA-12012: error on auto execute of job 8887
    ORA-44003: invalid SQL name
    thank you in advance,
    Dav

    I just google with "ORA-44003: invalid SQL name" and found below link:
    http://www.oracle-base.com/articles/10g/dbms_assert_10gR2.php
    SIMPLE_SQL_NAME
    The SIMPLE_SQL_NAME function checks the input string conforms to the basic characteristics of a simple SQL name:
    The first character of the name is alphabetic.
    The name only contains alphanumeric characters or the "_", "$", "#"
    Quoted names must be enclosed by double quotes and may contain any characters, including quotes provided they are represented by two quotes in a row ("").
    The function ignores leading and trailing white spaces are ignored
    The length of the input string is not validated.

  • Log file sequential read  and RFS ping/write - among Top 5 event

    I have situation here to discuss. In a 3-node RAC setup which is Logical standby DB; one node is showing high CPU utilization around 40~50%. The CPU utilization was less than 20% 10 days back but from 9th oldest day it jumped and consistently shows the double figure. I ran AWR reports on all three nodes and found one node with high CPU utilization and shows below tops events-
    EVENT WAITS TIME(S) AVG WAIT(MS) %TOTAL CALL TIME WAIT CLASS
    CPU time 5,802 34.9
    RFS ping 15 5,118 33,671 30.8 Other
    Log file sequential read 234,831 5,036 21 30.3 System I/O
    Sql*Net more data from
    client 24,171 1,087 45 6.5 Network
    Db file sequential read 130,939 453 3 2.7 User I/O
    Findings:-
    On AWR report(file attached) for node= sipd207; we can see that "RFS PING" wait event takes 30% of the waits and "log file sequential read" wait event takes 30% of the waits that occurs in database.
    Environment :- (Oracle- 10.2.0.4.0, O/S - AIX .3)
    1)other node awr shows "log file sync" - is it due to oversized log buffer?
    2)Network wait events can be reduced by tweaking SDU & TDU values based on MDU.
    3) Why ARCH processes taking much to archives filled redo logs; is it issue with slow disk I/O?
    Regards
    WORKLOAD REPOSITORY report for<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<DB Name DB Id Instance Inst Num Release RAC Host
    XXXPDB 4123595889 XXX2p2 2 10.2.0.4.0 YES sipd207
    Snap Id Snap Time Sessions Curs/Sess
    Begin Snap: 1053 04-Apr-11 18:00:02 59 7.4
    End Snap: 1055 04-Apr-11 20:00:35 56 7.5
    Elapsed: 120.55 (mins)
    DB Time: 233.08 (mins)
    Cache Sizes
    ~~~~~~~~~~~ Begin End
    Buffer Cache: 3,728M 3,728M Std Block Size: 8K
    Shared Pool Size: 4,080M 4,080M Log Buffer: 14,332K
    Load Profile
    ~~~~~~~~~~~~ Per Second Per Transaction
    Redo size: 245,392.33 10,042.66
    Logical reads: 9,080.80 371.63
    Block changes: 1,518.12 62.13
    Physical reads: 7.50 0.31
    Physical writes: 44.00 1.80
    User calls: 36.44 1.49
    Parses: 25.84 1.06
    Hard parses: 0.59 0.02
    Sorts: 12.06 0.49
    Logons: 0.05 0.00
    Executes: 295.91 12.11
    Transactions: 24.43
    % Blocks changed per Read: 16.72 Recursive Call %: 94.18
    Rollback per transaction %: 4.15 Rows per Sort: 53.31
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer Nowait %: 99.99 Redo NoWait %: 100.00
    Buffer Hit %: 99.92 In-memory Sort %: 100.00
    Library Hit %: 99.83 Soft Parse %: 97.71
    Execute to Parse %: 91.27 Latch Hit %: 99.79
    Parse CPU to Parse Elapsd %: 15.69 % Non-Parse CPU: 99.95
    Shared Pool Statistics Begin End
    Memory Usage %: 83.60 84.67
    % SQL with executions>1: 97.49 97.19
    % Memory for SQL w/exec>1: 97.10 96.67
    Top 5 Timed Events Avg %Total
    ~~~~~~~~~~~~~~~~~~ wait Call
    Event Waits Time (s) (ms) Time Wait Class
    CPU time 4,503 32.2
    RFS ping 168 4,275 25449 30.6 Other
    log file sequential read 183,537 4,173 23 29.8 System I/O
    SQL*Net more data from client 21,371 1,009 47 7.2 Network
    RFS write 25,438 343 13 2.5 System I/O
    RAC Statistics DB/Inst: UDAS2PDB/udas2p2 Snaps: 1053-1055
    Begin End
    Number of Instances: 3 3
    Global Cache Load Profile
    ~~~~~~~~~~~~~~~~~~~~~~~~~ Per Second Per Transaction
    Global Cache blocks received: 0.78 0.03
    Global Cache blocks served: 1.18 0.05
    GCS/GES messages received: 131.69 5.39
    GCS/GES messages sent: 139.26 5.70
    DBWR Fusion writes: 0.06 0.00
    Estd Interconnect traffic (KB) 68.60
    Global Cache Efficiency Percentages (Target local+remote 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer access - local cache %: 99.91
    Buffer access - remote cache %: 0.01
    Buffer access - disk %: 0.08
    Global Cache and Enqueue Services - Workload Characteristics
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Avg global enqueue get time (ms): 0.5
    Avg global cache cr block receive time (ms): 0.9
    Avg global cache current block receive time (ms): 1.0
    Avg global cache cr block build time (ms): 0.0
    Avg global cache cr block send time (ms): 0.1
    Global cache log flushes for cr blocks served %: 2.9
    Avg global cache cr block flush time (ms): 4.6
    Avg global cache current block pin time (ms): 0.0
    Avg global cache current block send time (ms): 0.1
    Global cache log flushes for current blocks served %: 0.1
    Avg global cache current block flush time (ms): 5.0
    Global Cache and Enqueue Services - Messaging Statistics
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Avg message sent queue time (ms): 0.1
    Avg message sent queue time on ksxp (ms): 0.6
    Avg message received queue time (ms): 0.0
    Avg GCS message process time (ms): 0.0
    Avg GES message process time (ms): 0.1
    % of direct sent messages: 31.57
    % of indirect sent messages: 5.17
    % of flow controlled messages: 63.26
    Time Model Statistics DB/Inst: UDAS2PDB/udas2p2 Snaps: 1053-1055
    -> Total time in database user-calls (DB Time): 13984.6s
    -> Statistics including the word "background" measure background process
    time, and so do not contribute to the DB time statistic
    -> Ordered by % or DB time desc, Statistic name
    Statistic Name Time (s) % of DB Time
    sql execute elapsed time 7,270.6 52.0
    DB CPU 4,503.1 32.2
    parse time elapsed 506.7 3.6
    hard parse elapsed time 497.8 3.6
    sequence load elapsed time 152.4 1.1
    failed parse elapsed time 19.5 .1
    repeated bind elapsed time 3.4 .0
    PL/SQL execution elapsed time 0.7 .0
    hard parse (sharing criteria) elapsed time 0.3 .0
    connection management call elapsed time 0.3 .0
    hard parse (bind mismatch) elapsed time 0.0 .0
    DB time 13,984.6 N/A
    background elapsed time 869.1 N/A
    background cpu time 276.6 N/A
    Wait Class DB/Inst: UDAS2PDB/udas2p2 Snaps: 1053-1055
    -> s - second
    -> cs - centisecond - 100th of a second
    -> ms - millisecond - 1000th of a second
    -> us - microsecond - 1000000th of a second
    -> ordered by wait time desc, waits desc
    Avg
    %Time Total Wait wait Waits
    Wait Class Waits -outs Time (s) (ms) /txn
    System I/O 529,934 .0 4,980 9 3.0
    Other 582,349 37.4 4,611 8 3.3
    Network 279,858 .0 1,009 4 1.6
    User I/O 54,899 .0 317 6 0.3
    Concurrency 136,907 .1 58 0 0.8
    Cluster 60,300 .0 41 1 0.3
    Commit 80 .0 10 130 0.0
    Application 6,707 .0 3 0 0.0
    Configuration 17,528 98.5 1 0 0.1
    Wait Events DB/Inst: UDAS2PDB/udas2p2 Snaps: 1053-1055
    -> s - second
    -> cs - centisecond - 100th of a second
    -> ms - millisecond - 1000th of a second
    -> us - microsecond - 1000000th of a second
    -> ordered by wait time desc, waits desc (idle events last)
    Avg
    %Time Total Wait wait Waits
    Event Waits -outs Time (s) (ms) /txn
    RFS ping 168 .0 4,275 25449 0.0
    log file sequential read 183,537 .0 4,173 23 1.0
    SQL*Net more data from clien 21,371 .0 1,009 47 0.1
    RFS write 25,438 .0 343 13 0.1
    db file sequential read 54,680 .0 316 6 0.3
    DFS lock handle 97,149 .0 214 2 0.5
    log file parallel write 104,808 .0 157 2 0.6
    db file parallel write 143,905 .0 149 1 0.8
    RFS random i/o 25,438 .0 86 3 0.1
    RFS dispatch 25,610 .0 56 2 0.1
    control file sequential read 39,309 .0 55 1 0.2
    row cache lock 130,665 .0 47 0 0.7
    gc current grant 2-way 35,498 .0 23 1 0.2
    wait for scn ack 50,872 .0 20 0 0.3
    enq: WL - contention 6,156 .0 14 2 0.0
    gc cr grant 2-way 16,917 .0 11 1 0.1
    log file sync 80 .0 10 130 0.0
    Log archive I/O 3,986 .0 9 2 0.0
    control file parallel write 3,493 .0 8 2 0.0
    latch free 2,356 .0 6 2 0.0
    ksxr poll remote instances 278,473 49.4 6 0 1.6
    enq: XR - database force log 2,890 .0 4 1 0.0
    enq: TX - index contention 325 .0 3 11 0.0
    buffer busy waits 4,371 .0 3 1 0.0
    gc current block 2-way 3,002 .0 3 1 0.0
    LGWR wait for redo copy 9,601 .2 2 0 0.1
    SQL*Net break/reset to clien 6,438 .0 2 0 0.0
    latch: ges resource hash lis 23,223 .0 2 0 0.1
    enq: WF - contention 32 6.3 2 62 0.0
    enq: FB - contention 660 .0 2 2 0.0
    enq: PS - contention 1,088 .0 2 1 0.0
    library cache lock 869 .0 1 2 0.0
    enq: CF - contention 671 .1 1 2 0.0
    gc current grant busy 1,488 .0 1 1 0.0
    gc current multi block reque 1,072 .0 1 1 0.0
    reliable message 618 .0 1 2 0.0
    CGS wait for IPC msg 62,402 100.0 1 0 0.4
    gc current block 3-way 998 .0 1 1 0.0
    name-service call wait 18 .0 1 57 0.0
    cursor: pin S wait on X 78 100.0 1 11 0.0
    os thread startup 16 .0 1 53 0.0
    enq: RO - fast object reuse 193 .0 1 3 0.0
    IPC send completion sync 652 99.2 1 1 0.0
    local write wait 194 .0 1 3 0.0
    gc cr block 2-way 534 .0 0 1 0.0
    log file switch completion 17 .0 0 20 0.0
    SQL*Net message to client 258,483 .0 0 0 1.5
    undo segment extension 17,282 99.9 0 0 0.1
    gc cr block 3-way 286 .7 0 1 0.0
    enq: TM - contention 76 .0 0 4 0.0
    PX Deq: reap credit 15,246 95.6 0 0 0.1
    kksfbc child completion 5 100.0 0 49 0.0
    enq: TT - contention 141 .0 0 2 0.0
    enq: HW - contention 203 .0 0 1 0.0
    RFS create 2 .0 0 115 0.0
    rdbms ipc reply 339 .0 0 1 0.0
    PX Deq Credit: send blkd 452 20.1 0 0 0.0
    gcs log flush sync 128 32.8 0 2 0.0
    latch: cache buffers chains 128 .0 0 1 0.0
    library cache pin 441 .0 0 0 0.0
    Wait Events DB/Inst: UDAS2PDB/udas2p2 Snaps: 1053-1055
    -> s - second
    -> cs - centisecond - 100th of a second
    -> ms - millisecond - 1000th of a second
    -> us - microsecond - 1000000th of a second
    -> ordered by wait time desc, waits desc (idle events last)

    We only apply on one node in a cluster so I would expect that the node running SQL Apply would have much higher usage and waits. Is this what you are asking?
    Larry

  • Getting empty log files with log4j and WebLogic 10.0

    Hi!
    I get empty log files with log4j 1.2.13 and WebLogic 10.0. If I don't run the application in the application server, then the logging works fine.
    The properties file is located in a jar in the LIB folder of the deployed project. If I change the name of the log file name in the properties file, it just creates a new empty file.
    What could be wrong?
    Thanks!

    I assume that when you change the name of the expected log file in the properties file, the new empty file is that name, correct?
    That means you're at least getting that properties file loaded by log4j, which is a good sign.
    As the file ends up empty, it appears that no logger statements are being executed at a debug level high enough for the current debug level. Can you throw in a logger.error() call at a point you're certain is executed?

  • 903/Enterprise; log files grow boundlessly and weekly bounces are needed

    This feed back is the result of an eval of 903 IAS
    enterprise on Solaris, 2 box cluster.
    This feed back is FYI.
    After installing IAS Enterprise 903 many times, testing
    stop/start/reboot use cases I noticed that many internal
    log files grow boundlessly.
    I've seen the notes on turning down log levels. This
    is still far from adequate. Log files in a 24x365
    architecture must be managable and have finite maximum disk
    space consumption sizes setable.
    Framework error events also need to be hooked into an Enterprise
    error and alert protocol. SNMP is not my favorite, but it's
    a start. JMX is my favorite. Also check out the free
    Big Brother monitor; www.bb4.com and it's wire protocol.
    Thanks and best of luck,
    Curt

    Just curious whether other folks have any opinions on
    OEM's suitability for a high rel. environment?
    On log file management?
    curt

  • GC.log File Size Limitation and Rotation

    Hi,
    Is it possible to limit the size of a GC.log file and to rotate it without causing havoc?
    Best regards,
    Katzensee

    James Roller1 wrote:
    You say, if you use 1024 x 768 images... that's just it: I'm just using the images without resizing. Some may be 4MB files when dragged in; others may be 300kb. I am not doing anything to them. Should I?  Does that help? Or does IBA "handle all that" when I drag and drop?
    It's not well documented. I get the impression that, if you drop an image that has resolution higher than 2048 x 1536, it is down-sampled to that size. You should at least crop the image outside IBA before dropping it in. If you want to save space, down-sample the image first to 1024 x 768 (after cropping).
    James Roller1 wrote:
    Also for the video: we're using iMovie and we just "Share" to iTunes and choose Large. Should I be choosing a lower setting or, again, does IBA not care and will take the Large and resize it?
    Instead of sharing to iTunes, export the movie to iPad format (File -> Export) and select 640 x 480. That'll give you reasonable quality at around 12 MB per minute.
    James Roller1 wrote:
    I have to say I chuckled at the wisdom and simplicity of the 2 Volume suggestion. I'd like to keep it as one, and I'd like to sell it on the store, but your idea is ingenious in its simplicity: just make 2!
    Thanks! Two's better'n one
    Michi.

  • System.log file are large, and cause a kernel panic when deleting

    Hello everyone,
    I am trying to solve a dillema.  I discovered that over 60GB of my 120GB SSD drive are taken up by system.log files.  All 3 are around 20GB.  I left my laptop running so none of them are the current one.  They are system.log.0, system.log.3 and system.log.4.  I have tried several ways to delete them:
    Using Disksweeper
    At a terminal prompt using sudo rm
    At a terminal prompt using sudo su - (to get root access) and then rm -f ....
    Rebooting the system in single user mode, mounting the drive and then attempting to delete them
    Each time I do, the system reboots with a kernel panic.  I have not looked at the panic log yet, but can post that if you need to.
    I did look at the end of the current log file to discover there were some issues with programs, and I have fixed it so that these size log files "shouldn't" be generated anymore.
    Looking for advice as to how to delete these files!  This is driving me nuts.  It shouldn't be this hard.
    Thanks for any help or insight you can give.

    OK, I actually was able to solve this myself.  My problem was that ther kernel panic was caused by the journaling.  So here is what I had to do:
    I temporarily disabled journaling, see here: http://forums.macrumors.com/showthread.php?t=373067
    Then I was able to delete the files by starting a terminal session and using sudo rm to delete the files.
    Re-enabled journaling (I don't know if it is needed, but it was on before).
    Hope this can help someone else in the future.

  • How can stop the bdump and udump trace files from accumulating?

    The files are growing so until it is filling up the user's c: drive. It's hundreds of files over 20 megs.

    Angela,
    The management of those files is a reposnsibilty of the DBA. ie The files dumped there are from background processes, session traces and your alert log.
    The alert log in bdump should be checked daily and errors investigated. Some errors will result in trace files being dumped. The alert log will always grow during normal operation and you can simply delete it or better yet zip it up and store it somewhere for a period of time.
    You can clean up the udump user traces files by simply deleting them periodically. For other traces you should have a look to see what they are for. Any ORA 600 traces should be investigated.
    If you are not a DBA and this seems to much trouble then you can just delete the files to reclaim the space.

  • Trying to install Photoshop Elements 12 from a DVD. I keep getting an error message that says "Error opening installation log file.  Verify that the specified log file location exists and is writable."  Have tried the solutions suggested on line but get t

    Have tried turning explorer of and on again and other recommended solutions but nothing seems to work.  Any suggestions?

    David,
    Can you please try Right-Click Setup.exe  > Run As Administrator
    and see if it helps.
    Thanks,
    Anwesha

  • Question regarding alert log file and trace files

    What should be the alert log file size ? And when should it be deleted? And for how many days user trace files should be kept?
    Also will anyone please tell me the importance of these files.
    Thanks

    This may help: http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14231/manproc.htm#sthref729
    There are a few discussions on it here:
    Re: Alert Log File
    alert log file contents viewing
    Re: how to read alert log file? is there any tool available?

Maybe you are looking for

  • How can I edit budgeted cost/unit in excel and then import into P6?

    I would like to edit resource assignements (budgeted cost/unit) month by month in an excel file, and then import it into P6. Is this possible? It seems that the resource assignments month by month cannot be exported to an excel file. Is there anyone

  • IPOD not recognized by itunes when its plugged in and recharged

    can't get itunes to recognize my shuffle when its plugged in. ipod works with what already loaded on it but I can't change anything cuz itunes doesn't see it! I installed latest itunes updater and ipod updater to no avail and everything else the webs

  • Opening a System-Form with selected Data

    Hi, i´m looking for a way to open a System-Form, e.g. Orders with selected Data. In the moment i do the following:      application.ActivateMenuItem("2050");  // orders      SAPbouiCOM.Form belegForm = application.Forms.ActiveForm;      belegForm.Mod

  • Can't show content of Invoke-WebRequest in Unicode

    I am trying to get Japanese web pages through  Invoke-WebRequest but the contents get garbled in Console. PS> $JP_MSSITE = Invoke-WebRequest http://www.microsoft.co.jp PS> $JP_MSSITE.ParsedHtml.title a?\a?¬a??a??a? ̄a?-a??a??a??a??Microsoft Japan: a??

  • Client on Server installation best practice

    Hi all, I wonder on this subject, searched and found nothing relevant, so I ask here : Is there any best practice/state of the art when you have a client application installed on the same machine as the database ? I know the client app use the server