Oracle 11g Performance Issue

We have Oracle 11g (11.1.0.6) on HP-UX environment with CC&B application. It was working fine and after we upgrade the database to 11.1.0.7 last week, since then the database performance is really slow. After we upgrade the database we notice that xdb component became invalid.
We are not sure how to investigate this issue, any help would be appreciated.

WORKLOAD REPOSITORY report for
DB Name         DB Id    Instance     Inst Num Startup Time    Release     RAC
CCBPROD       3218377102 ccbprod             1 10-Jan-12 20:26 11.1.0.7.0  NO
Host Name        Platform                         CPUs Cores Sockets Memory(GB)
huccbhp5         HP-UX IA (64-bit)                   4     4       2      23.97
              Snap Id      Snap Time      Sessions Curs/Sess
Begin Snap:     43912 11-Jan-12 10:00:31       157      96.1
  End Snap:     43913 11-Jan-12 11:00:35       186     101.3
   Elapsed:               60.08 (mins)
   DB Time:              653.40 (mins)
Cache Sizes                       Begin        End
~~~~~~~~~~~                  ---------- ----------
               Buffer Cache:    12,160M    12,096M  Std Block Size:         8K
           Shared Pool Size:       704M       704M      Log Buffer:    58,764K
Load Profile              Per Second    Per Transaction   Per Exec   Per Call
~~~~~~~~~~~~         ---------------    --------------- ---------- ----------
      DB Time(s):               10.9                0.5       0.02       0.02
       DB CPU(s):                3.6                0.2       0.01       0.01
       Redo size:          529,539.6           24,544.1
   Logical reads:          155,545.5            7,209.5
   Block changes:            2,047.0               94.9
  Physical reads:              204.6                9.5
Physical writes:               96.0                4.5
      User calls:              698.7               32.4
          Parses:               77.8                3.6
     Hard parses:                0.1                0.0
W/A MB processed:          406,873.3           18,858.5
          Logons:                0.4                0.0
        Executes:              456.1               21.1
       Rollbacks:               19.6                0.9
    Transactions:               21.6
Instance Efficiency Percentages (Target 100%)
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
            Buffer Nowait %:  100.00       Redo NoWait %:   99.97
            Buffer  Hit   %:   99.87    In-memory Sort %:  100.00
            Library Hit   %:   99.93        Soft Parse %:   99.89
         Execute to Parse %:   82.95         Latch Hit %:   99.99
Parse CPU to Parse Elapsd %:    0.01     % Non-Parse CPU:   99.93
Shared Pool Statistics        Begin    End
             Memory Usage %:   89.78   90.02
    % SQL with executions>1:   86.04   84.87
  % Memory for SQL w/exec>1:   86.22   86.03
Top 5 Timed Foreground Events
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
                                                           Avg
                                                          wait   % DB
Event                                 Waits     Time(s)   (ms)   time Wait Class
virtual circuit wait              2,170,696      15,761      7   40.2 Network
DB CPU                                           13,115          33.5
db file sequential read             362,802       6,304     17   16.1 User I/O
enq: TX - row lock contention           115       1,118   9721    2.9 Applicatio
log file sync                         8,183         818    100    2.1 Commit
Host CPU (CPUs:    4 Cores:    4 Sockets:    2)
~~~~~~~~         Load Average
               Begin       End     %User   %System      %WIO     %Idle
                3.28      3.93      93.1       2.3       2.8       4.6
Instance CPU
~~~~~~~~~~~~
              % of total CPU for Instance:  91.1
              % of busy  CPU for Instance:  95.4
  %DB time waiting for CPU - Resource Mgr:   0.0
Memory Statistics
~~~~~~~~~~~~~~~~~                       Begin          End
                  Host Mem (MB):     24,545.7     24,545.7
                   SGA use (MB):     13,312.0     13,312.0
                   PGA use (MB):        935.9        949.3
    % Host Mem used for SGA+PGA:        58.05        58.05

Similar Messages

  • Oracle 11g performance issue ( BITMAP CONVERSION TO ROWIDS)

    I have two instance of oracle 11g.
    in both instance i fired same query.
    one instance returns the result in 1sec but other instance returns the result in 10 sec
    following is explain plan for bot instance
    instance 1
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 1 | 143 | 59 (2)| 00:00:01 |
    | 1 | HASH GROUP BY | | 1 | 143 | 59 (2)| 00:00:01 |
    | 2 | VIEW | VM_NWVW_2 | 1 | 143 | 59 (2)| 00:00:01 |
    | 3 | HASH UNIQUE | | 1 | 239 | 59 (2)| 00:00:01 |
    | 4 | NESTED LOOPS | | | | | |
    | 5 | NESTED LOOPS | | 1 | 239 | 58 (0)| 00:00:01 |
    PLAN_TABLE_OUTPUT
    | 6 | NESTED LOOPS | | 1 | 221 | 57 (0)| 00:00:01 |
    | 7 | NESTED LOOPS | | 1 | 210 | 55 (0)| 00:00:01 |
    | 8 | NESTED LOOPS | | 1 | 184 | 54 (0)| 00:00:01 |
    | 9 | NESTED LOOPS | | 1 | 158 | 53 (0)| 00:00:01 |
    | 10 | NESTED LOOPS | | 1 | 139 | 52 (0)| 00:00:01 |
    | 11 | NESTED LOOPS | | 1 | 105 | 50 (0)| 00:00:01 |
    |* 12 | INDEX RANGE SCAN | year_field | 1 | 29 | 2 (0)| 00:00:01 |
    | 13 | SORT AGGREGATE | | 1 | 8 | | |
    | 14 | INDEX FULL SCAN (MIN/MAX)| idx_bf_creation_date | 1 | 8 | 2 (0)| 00:00:01 |
    |* 15 | TABLE ACCESS BY INDEX ROWID| OHRT_bugs_fact | 1 | 76 | 48 (0)| 00:00:01 |
    |* 16 | INDEX RANGE SCAN | idx_bf_creation_date | 76 | | 1 (0)| 00:00:01 |
    PLAN_TABLE_OUTPUT
    |* 17 | TABLE ACCESS BY INDEX ROWID | OHRT_all_time_dimension | 1 | 34 | 2 (0)| 00:00:01 |
    |* 18 | INDEX UNIQUE SCAN | unique_alltime_bug_instance_id | 1 | | 1 (0)| 00:00:01 |
    | 19 | TABLE ACCESS BY INDEX ROWID | OHRT_all_time_dimension | 1 | 19 | 1 (0)| 00:00:01 |
    |* 20 | INDEX UNIQUE SCAN | unique_alltime_bug_instance_id | 1 | | 1 (0)| 00:00:01 |
    |* 21 | INDEX RANGE SCAN | bugseverity_instance_id_ref_id | 1 | 26 | 1 (0)| 00:00:01 |
    |* 22 | INDEX UNIQUE SCAN | unique_alltime_bug_instance_id | 1 | 26 | 1 (0)| 00:00:01 |
    | 23 | INLIST ITERATOR | | | | | |
    |* 24 | TABLE ACCESS BY INDEX ROWID | OHMT_ANL_BUCKET | 1 | 11 | 2 (0)| 00:00:01 |
    |* 25 | INDEX UNIQUE SCAN | SYS_C0053213 | 5 | | 1 (0)| 00:00:01 |
    |* 26 | INDEX RANGE SCAN | FK_BUCKET_TYPE | 6 | | 0 (0)| 00:00:01 |
    |* 27 | TABLE ACCESS BY INDEX ROWID | OHMT_ANL_BUCKET | 1 | 18 | 1 (0)| 00:00:01 |
    instance 2
    Plan
    SELECT STATEMENT ALL_ROWS Cost: 22 Bytes: 142 Cardinality: 1
    32 HASH GROUP BY Cost: 22 Bytes: 142 Cardinality: 1
    31 VIEW VIEW SYS.VM_NWVW_2 Cost: 22 Bytes: 142 Cardinality: 1
    30 HASH UNIQUE Cost: 22 Bytes: 237 Cardinality: 1
    29 NESTED LOOPS
    27 NESTED LOOPS Cost: 21 Bytes: 237 Cardinality: 1
    25 NESTED LOOPS Cost: 20 Bytes: 219 Cardinality: 1
    21 NESTED LOOPS Cost: 18 Bytes: 208 Cardinality: 1
    19 NESTED LOOPS Cost: 17 Bytes: 183 Cardinality: 1
    17 NESTED LOOPS Cost: 16 Bytes: 157 Cardinality: 1
    14 NESTED LOOPS Cost: 15 Bytes: 138 Cardinality: 1
    11 NESTED LOOPS Cost: 13 Bytes: 104 Cardinality: 1
    3 INDEX RANGE SCAN INDEX REPORTSDB.year_field Cost: 2 Bytes: 29 Cardinality: 1
    2 SORT AGGREGATE Bytes: 8 Cardinality: 1
    1 INDEX FULL SCAN (MIN/MAX) INDEX REPORTSDB.idx_bf_creation_date Cost: 3 Bytes: 8 Cardinality: 1
    10 TABLE ACCESS BY INDEX ROWID TABLE REPORTSDB.OHRT_bugs_fact Cost: 13 Bytes: 75 Cardinality: 1
    9 BITMAP CONVERSION TO ROWIDS
    8 BITMAP AND
    5 BITMAP CONVERSION FROM ROWIDS
    4 INDEX RANGE SCAN INDEX REPORTSDB.idx_OHRT_bugs_fact_2product Cost: 2 Cardinality: 85
    7 BITMAP CONVERSION FROM ROWIDS
    6 INDEX RANGE SCAN INDEX REPORTSDB.idx_bf_creation_date Cost: 2 Cardinality: 85
    13 TABLE ACCESS BY INDEX ROWID TABLE REPORTSDB.OHRT_all_time_dimension Cost: 2 Bytes: 34 Cardinality: 1
    12 INDEX UNIQUE SCAN INDEX (UNIQUE) REPORTSDB.unique_alltime_bug_instance_id Cost: 1 Cardinality: 1
    16 TABLE ACCESS BY INDEX ROWID TABLE REPORTSDB.OHRT_all_time_dimension Cost: 1 Bytes: 19 Cardinality: 1
    15 INDEX UNIQUE SCAN INDEX (UNIQUE) REPORTSDB.unique_alltime_bug_instance_id Cost: 1 Cardinality: 1
    18 INDEX UNIQUE SCAN INDEX (UNIQUE) REPORTSDB.unique_alltime_bug_instance_id Cost: 1 Bytes: 26 Cardinality: 1
    20 INDEX RANGE SCAN INDEX REPORTSDB.bugseverity_instance_id_ref_id Cost: 1 Bytes: 25 Cardinality: 1
    24 INLIST ITERATOR
    23 TABLE ACCESS BY INDEX ROWID TABLE OPSHUB.OHMT_ANL_BUCKET Cost: 2 Bytes: 11 Cardinality: 1
    22 INDEX UNIQUE SCAN INDEX (UNIQUE) OPSHUB.SYS_C0040939 Cost: 1 Cardinality: 5
    26 INDEX RANGE SCAN INDEX OPSHUB.FK_BUCKET_TYPE Cost: 0 Cardinality: 6
    28 TABLE ACCESS BY INDEX ROWID TABLE OPSHUB.OHMT_ANL_BUCKET Cost: 1 Bytes: 18 Cardinality: 1
    in both explain plan only difference is
    9 BITMAP CONVERSION TO ROWIDS
    8 BITMAP AND
    5 BITMAP CONVERSION FROM ROWIDS
    but is bitmap degrading performance lot?
    or suggest me what other parameter i can see so 2nd instance gives me better performace.

    I see more differences.
    In plan 1:
    * 16      INDEX RANGE SCAN      idx_bf_creation_date      76           1 (0)     00:00:01
    in Plan 2:
    1 INDEX FULL SCAN (MIN/MAX) INDEX REPORTSDB.idx_bf_creation_date Cost: 3 Bytes: 8 Cardinality: 1
    So this is not about "bitmap" good/bad, it about the access strategy which changed due to differences in data statistics etc. To analyze more, I'd help a LOT if those plans would be formated in a good and same way, use around it to do so.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Newbie question about Oracle 11G performance

    I have a situation where a newer better faster server running Oracle 11gR1 is running considerably slower than an older slower server running Oracle 10gR2.
    Both of these servers have the same web applications and same dumpfiles loaded into them.
    I personally installed the Oracle 11g on the new server, however, I didn't install the 10gR2 on the older server, some one else did...
    I have checked what I think are the basics: the tables are not missing constraints, the automatic memory management is enabled.
    I know this is a really broad question.
    Newer server:
    Quad-Core AMD Opteron Processor 2352 2.10 GHz (2 processors)
    16 GB memory
    Windows Server 2008, 64 bit
    Raid 5
    Older Server:
    Intel Xeon 1.8 GHz (2 processors)
    4 GB memory
    Windows Server 2003, 32 bit
    Raid 5

    Hi Derik,
    I know this is a really broad question. No, it happens all the time! Here is a similar issue:
    http://blog.tuningknife.com/2008/09/26/oracle-11g-performance-issue/
    +"In the end, nothing I tried could dissuade 11g from emitting the “PARSING IN CURSOR #d+” message for each insert statement. I filed a service request with Oracle Support on this issue and they ultimately referred the issue to development as a bug. Note that Support couldn’t reproduce the slowness I was seeing, but their trace files did reflect the parsing messages I observed."+
    I would:
    1 - Start by examining historical SQL execution plans (stats$sql_plan or dba_hist_sql_plan). Try to isolate the exact nature of the decreased performance.
    Are different indexes being used? Are you geting more full-table scans?
    2 - Migrate-in your old 10g CBO statistics
    3 - Confirm that all init.ora parms are identical
    4 - Drill into each SQL with a different execution plan . . .
    Raid 5 Don't believe that crap that all RAID-5 is evil. . . .
    http://www.dba-oracle.com/t_raid5_acceptable_oracle.htm
    But at the same time, consider using the Oracle standard, RAID-10 . . .
    Hope this helps . . .
    Donald K. Burleson
    Oracle Press author
    Author of "Oracle Tuning: The Definitive Reference"
    http://www.rampant-books.com/t_oracle_tuning_book.htm
    "Time flies like an arrow; Fruit flies like a banana".

  • Oracle 11g - Date Issue?

    Oracle 11g - Date Issue?
    =====================
    I hope this is the right forum for this issue, if not let me, where to go.
    We are using Oracle 11g (11.2.0.1.0) on (Platform : solaris[tm] oe (64-bit)), Sql Developer 3.0.04.
    Our NLS_DATE_FORMAT = 'DD-MON-RR'
    Using 'RR' in the format forces two-digit years less than or equal to 49 to be interpreted as years in the 21st century (2000–2049), and years 50 and over, as years in the 20th century (1950–1999). Setting the RR format as the default for all two-digit year entries allows you to become year-2000 compliant. For example:
    We have a date '01-JUN-31' in our source system. It treat this date as 01-JUN-2031' instead of '01-JUN-1931'. One century forward.
    Do we able to resolve using NLS_DATE_FORMAT change?
    How do we resolve this issue?
    Thanks in helping.

    qwe16235 wrote:
    Our source is Oracle data base, where S_date is defined as DATE. Why did you say STRING , when it defined as DATE data type?I doubt you source is an Oracle database. You may have it stored in an Oracle database, but it cam form somewhere else, and was very likely inserted into the table as a string, wherever it came from. Given a string that resembles a date, Oracle will try to convert it to a date using the nls_date_format parameter for the session (which can either be set in the session or inherited form the database). Perhaps this will help explain:
    SQL> create table t (conv_type varchar2(10), dt date);
    Table created.
    SQL> alter session set nls_date_format = 'dd-mon-rr';
    Session altered.
    SQL> insert into t values ('Implicit', '01-jun-31');
    1 row created.
    SQL> insert into t values ('Explicit', to_date('01-jun-1931', 'dd-mon-yyyy'));
    1 row created.
    SQL> commit;
    Commit complete.
    SQL> select conv_type, to_char(dt, 'dd-mon-yyyy') dt
      2  from t;
    CONV_TYPE  DT
    Implicit   01-jun-2031
    Explicit   01-jun-1931So, unless you are really inserting dates, not strings that look like dates, you are going to have problems.
    John

  • Oracle BPEL 11G performance issue

    Hi
    We are facing performance issues in executing our composite process in Oracle SOA 11g .
    We have installed an admin server and 2 managed servers in cluster in same box. The machine utilization reached almost 95% when i started the admin server and 2 managed server (min n max size of heap given as 1GB each in start up). So i shut down one managed server and increased the JVM size of other to 2 GB and found that the heap size reaches 1.5 GB on start up (observed the heap size using Jconsole)
    The machine capacity is windows server with 4 GB RAM.
    Our process requries multiple records to be processed which are retrieved using Database query.
    We have created 2 composites
    the first composite has 2 BPEL process. the First BPEL 1 executes the DB query and retrieves the result and based on result retrieved we invoke the second BPEL2
    which does around 4 DB calls and passed the result to the next composite. The final BPEL process 3 has multiple select and update query involving DB intensive process.
    When we retrieve 500 records from the BPEL 1 and process , half way through we face out of memory exception. So we are using throttling but even then while executing the process of BPEL3 we are facing out of memory excetion.
    Can you let me know how to find the memory space taken from heap by each BPEL process during it execution. Where in console can i get the memory used details so that i can find which BPEL sis consuming more memory and we can work on optimising.
    Actually we are expecting around 1Lakh and above messages per day and need to check on how this process can handle and also how to increase or determine the capacity of the windows box.
    any immediate help is highly appreciated
    thanks

    Always raise a case with Oracle Support for such issues.
    Regards,
    Anuj

  • "oracle database 11g performance issues"

    Hai everybody,
    In oracle 11g 11.2.0.1.0 we are developing business application using java, Our developers said Database performance is very poor it takes more time to retrieve values from database, they check froantend and middleware that has no problem, if the query returns less values or rows it takes little more time. How can i solve this problem? please help me
    Regards Benk

    hai,
    sorry for the delay my @$ORACLE_HOME/rdbms/admin/ashrpt.sql output shown below
    ASH Report For ORCL/orcl
    DB Name DB Id Instance Inst Num Release RAC Host
    ORCL 1295420332 orcl 1 11.2.0.1.0 NO node6.node6-
    CPUs SGA Size Buffer Cache Shared Pool ASH Buffer Size
    2 1,561M (100%) 480M (30.7%) 432M (27.7%) 4.0M (0.3%)
    Analysis Begin Time: 04-Oct-12 16:44:16
    Analysis End Time: 04-Oct-12 16:59:33
    Elapsed Time: 15.3 (mins)
    Begin Data Source: V$ACTIVE_SESSION_HISTORY
    End Data Source: V$ACTIVE_SESSION_HISTORY
    Sample Count: 3
    Average Active Sessions: 0.00
    Avg. Active Session per CPU: 0.00
    Report Target: None specified
    Top User Events DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Avg Active
    Event Event Class % Event Sessions
    null event Other 33.33 0.00
    Top Background Events DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Avg Active
    Event Event Class % Activity Sessions
    CPU + Wait for CPU CPU 33.33 0.00
    os thread startup Concurrency 33.33 0.00
    Top Event P1/P2/P3 Values DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Service/Module DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Service Module % Activity Action % Action
    SYS$BACKGROUND UNNAMED 66.67 UNNAMED 66.67
    SYS$USERS UNNAMED 33.33 UNNAMED 33.33
    Top Client IDs DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top SQL Command Types DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Phases of Execution DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Avg Active
    Phase of Execution % Activity Sessions
    SQL Execution 33.33 0.00
    Top SQL with Top Events DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top SQL with Top Row Sources DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top SQL using literals DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Parsing Module/Action DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top PL/SQL Procedures DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Java Workload DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Call Types DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Sessions DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    -> '# Samples Active' shows the number of ASH samples in which the session
    was found waiting for that particular event. The percentage shown
    in this column is calculated with respect to wall clock time
    and not total database activity.
    -> 'XIDs' shows the number of distinct transaction IDs sampled in ASH
    when the session was waiting for that particular event
    -> For sessions running Parallel Queries, this section will NOT aggregate
    the PQ slave activity into the session issuing the PQ. Refer to
    the 'Top Sessions running PQs' section for such statistics.
    Sid, Serial# % Activity Event % Event
    User Program # Samples Active XIDs
    10, 1 33.33 CPU + Wait for CPU 33.33
    SYS [email protected] (ARC2) 1/917 [  0%] 0
    19, 148 33.33 null event 33.33
    SYS [email protected] (J000) 1/917 [  0%] 0
    139, 3 33.33 os thread startup 33.33
    SYS [email protected] (CJQ0) 1/917 [  0%] 0
    Top Blocking Sessions DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Sessions running PQs DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top DB Objects DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top DB Files DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Latches DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Activity Over Time DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    -> Analysis period is divided into smaller time slots
    -> Top 3 events are reported in each of those slots
    -> 'Slot Count' shows the number of ASH samples in that slot
    -> 'Event Count' shows the number of ASH samples waiting for
    that event in that slot
    -> '% Event' is 'Event Count' over all ASH samples in the analysis period
    Slot Event
    Slot Time (Duration) Count Event Count % Event
    16:48:00 (2.0 min) 1 CPU + Wait for CPU 1 33.33
    16:50:00 (2.0 min) 2 null event 1 33.33
    os thread startup 1 33.33
    End of Report
    Regards Benk

  • Oracle SOA 11g Performance Issue

    Hi,
    We have set up Oracle SOA Suite on AIX environment. Java which we are using is IBM Jdk 1.6. Recently we are hit with performance issue. Frequently we are getting out of memory exception and we need to restart the server and sometimes physically reboot the machine, because out of 16 GB of RAM 4GB we have given as heap space to Admin Server, 7 GB to SOA Server but it is taking more than 7 GB as heap space. On stopping or killing both the services memory is not getting released
    SOA Suite Version : 11.1.1.3
    Instance Node: Single Node
    I collected the logs and tried to analyze in Thread Dump Analyzer and i could see objects(Reserved) is taking 100% of the CPU Utilization.
    We are getting the following error highlighed in the analyzer. There are about 200+ threads got stuck.
    "HTTPThreadGroup-42" prio=10 tid=0x6382ba28 nid=0x20bf4 waiting on condition [0x6904f000..0x6904fb94]
    at sun.misc.Unsafe.park(Native Method)
    at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:146)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireNanos(AbstractQueuedSynchronizer.java:772)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireNanos(AbstractQueuedSynchronizer.java:1087)
    at java.util.concurrent.SynchronousQueue$Node.waitForPut(SynchronousQueue.java:291)
    at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:443)
    at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:475)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:674)
    at java.lang.Thread.run(Thread.java:595)
    "HTTPThreadGroup-41" prio=10 tid=0x6ae3cce0 nid=0x20bf0 waiting on condition [0x68d8f000..0x68d8fc14]
    at sun.misc.Unsafe.park(Native Method)
    at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:146)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireNanos(AbstractQueuedSynchronizer.java:772)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireNanos(AbstractQueuedSynchronizer.java:1087)
    at java.util.concurrent.SynchronousQueue$Node.waitForPut(SynchronousQueue.java:291)
    at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:443)
    at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:475)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:674)
    at java.lang.Thread.run(Thread.java:595)
    Has anyone faced same issue? We are badly hit with this performance issue in UAT
    Please concider this on high priority and someone please help us
    Regards,
    Sundar

    Always raise a case with Oracle Support for such issues.
    Regards,
    Anuj

  • Crystal Reports 11g Performance Issue

    We just upgraded our database to Oracle 11g from Oracle 10g and we are having significant performance issues with Crystal Reports.
    Our DEV and TEST environments are on 11g and are very slow to connect to the database and attach to specific tables.  It is unusable.  Our PROD environment is still 10g and works fine.  
    We  have tested with both the lastest version -  Crystal Reports 2008 V1  and  Crystal Reports XI R2 SP6.  We have also tested on several different machines.
    We are using Oracle 10g ODBC drivers.
    Does anyone have any recommendations?

    You could also try our Data direct drivers we have available on our WEB site using this [link|https://smpdl.sap-ag.de/~sapidp/012002523100008666562008E/cr_datadirect53_win32.zip].
    Those drivers are the most recent and do support 11g. It also has a wired driver that doesn't require the client to be installed.
    Also, highly recommended that when you do update the Oracle client to uninstall 10 first. There have been issues with CR and Oracle mixing the Oracle dependencies and causing problems.
    Thank you
    Don

  • Oracle 11g upgrade issues

    Hello :
    We recently migrated our database from 9.2.07 (RBO) to 11.2.0.1 (CBO) ( Solaris 10 Sparc 64 bit ). Ever since the migration happened we are experiencing the following issues:
    1. Execution plans are very costly. ( Please note that we moved from RBO to CBO )
    2. High Gets / Reads
    3. High CPU Utilization
    4. High waits on 'DB FILE SEQUENTIAL READ'
    5. This situation is vitually killing all our applications performance.
    I know that oracle 11g explicitly does not support RULE based optimizer , but is there any parameters that can force it go the RBO way.
    I would really appreciate your help.
    Regards,
    BMP

    Officially it is no more documented http://download.oracle.com/docs/cd/E11882_01/server.112/e10820/initparams165.htm#i1131532
    but it is still accepted:
    SQL> select * from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    PL/SQL Release 11.2.0.1.0 - Production
    CORE    11.2.0.1.0      Production
    TNS for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production
    SQL> alter system set optimizer_mode=rule;
    System altered.
    SQL> show parameter optimizer
    NAME                                 TYPE        VALUE
    optimizer_capture_sql_plan_baselines boolean     FALSE
    optimizer_dynamic_sampling           integer     2
    optimizer_features_enable            string      11.2.0.1
    optimizer_index_caching              integer     0
    optimizer_index_cost_adj             integer     100
    optimizer_mode                       string      RULE
    optimizer_secure_view_merging        boolean     TRUE
    optimizer_use_invisible_indexes      boolean     FALSE
    optimizer_use_pending_statistics     boolean     FALSE
    optimizer_use_sql_plan_baselines     boolean     TRUEEdited by: P. Forstmann on 1 sept. 2010 20:21

  • Oracle  11g Performance tuning approach ?

    Hello Experts,
    Is it the right forum to follow oracle performance tuning discussions ? If not, let me know what will be the forum to pick up some thread on this subject.
    I am looking for performance tuning approach for oracle 11g. I learned there are some new items in 11g in this regard. For persons, who did tuning in earlier versions of Oracle,
    what will be the best way adopt to 11 g?
    I reviewed the 11g performance tuning guide, but I am looking for some white papers/blogs with case studies and practical approaches. I hope that you have used them.
    What are the other sources to pick up some discussions?
    Do you mind, share your thoughts?
    Thanks in advance.
    RI

    The best sources of information on performance tuning are:
    1. Jonathan Lewis: http://jonathanlewis.wordpress.com/all-postings/
    2. Christian Antognini: http://www.antognini.ch/
    3. Tanel Poder: http://blog.tanelpoder.com/
    4. Richard Foote: http://richardfoote.wordpress.com/
    5. Cary Millsap: http://carymillsap.blogspot.com/
    and a few dozen others whose blogs you will find cross-referenced in those above.

  • Oracle 11g performace issue

    hi all,
    am using oracle 11g enterprise edition., i didn't ran any query but in ADDR report it was running some sytem based query(SYS,DBSNMP,SYSMAN etc)
    but in previous 10g it was working fine.. can anyone tell me how to stop those query..
    this is the query was running most frequently :
    select end_time, wait_class#, (time_waited_fg)/(intsize_csec/100), (time_waited)/(intsize_csec/100), 0
    from v$waitclassmetric union all
    select fg.end_time, -1, fg.value, bg.value, dbtime.value
    from v$sysmetric fg, v$sysmetric bg, v$sysmetric dbtime
    where bg.metric_name = 'Background CPU Usage Per Sec' and bg.group_id = 2 and fg.metric_name = 'CPU Usage Per Sec' and fg.group_id = 2 and dbtime.metric_name = 'Average Active Sessions' and dbtime.group_id = 2 and bg.end_time = fg.end_time and fg.end_time = dbtime.end_time order by end_time,wait_class#
    UPDATE MGMT_JOB_EXECUTION SET STEP_STATUS=:B2 WHERE STEP_ID=:B1

    Velsjeya wrote:
    hi all,
    am using oracle 11g enterprise edition., i didn't ran any query but in ADDR report it was running some sytem based query(SYS,DBSNMP,SYSMAN etc)
    but in previous 10g it was working fine.. can anyone tell me how to stop those query..
    this is the query was running most frequently :
    select end_time, wait_class#, (time_waited_fg)/(intsize_csec/100), (time_waited)/(intsize_csec/100), 0
    from v$waitclassmetric union all
    select fg.end_time, -1, fg.value, bg.value, dbtime.value
    from v$sysmetric fg, v$sysmetric bg, v$sysmetric dbtime
    where bg.metric_name = 'Background CPU Usage Per Sec' and bg.group_id = 2 and fg.metric_name = 'CPU Usage Per Sec' and fg.group_id = 2 and dbtime.metric_name = 'Average Active Sessions' and dbtime.group_id = 2 and bg.end_time = fg.end_time and fg.end_time = dbtime.end_time order by end_time,wait_class#
    UPDATE MGMT_JOB_EXECUTION SET STEP_STATUS=:B2 WHERE STEP_ID=:B1Why are you so concerned about background queries? Do you really think these queries uses high resource and have high elapsed time then you custom queries? The god way to diagnose a performance problem is to identify most resource intensive/time consuming program/module/code rather then running after background queries. Yes sometimes background queries too run slowly in rear cases and then you need to run dbms_stats.gather_dictionary_stats or dbms_stats.gather_fixed_objects_stats (https://blogs.oracle.com/optimizer/entry/fixed_objects_statistics_and_why)

  • Oracle 11g : Sequence Issues after impdp

    Hi All,
    We are migrating from Oracle 10g to Oracle 11g. As part of this, we take a expdp from 10g and successfully did an import using impdp on to Oracle 11g database.
    But the problem is, there were few Primary key violations which occurred and all of them relate to sequences.
    The Maximum data in the tables and the last number in the sequences differed which caused the issue.
    Upon investigation and browsing the Web, http://www.nerdliness.com/article/2009/03/18/my-oracle-sequencedatapump-shenanigans ; I understood it could be because of the export taken while the application is online and writing to the database.
    I reset all the failing sequences manually and it is fine now.
    My questions now are
    *1) Can we ascertain that the export taken when the source database is offline would eliminate the sequences issue.*
    2) As this is being done in Production, I would like to make few checks to ensure that the sequences are properly imported. ---
    Again reading few websites and oracle Forums, i found the below sql's..
    select table_name, column_name, utl_raw.cast_to_number(high_value) as highval
    from dba_Tab_columns
    where owner = 'PRODUCTION_OWNER'
    AND DATA_TYPE= 'NUMBER'
    AND (OWNER, TABLE_NAME, COLUMN_NAME) IN
    (SELECT CC.OWNER, CC.TABLE_NAME, CC.COLUMN_NAME
    FROM DBA_CONS_COLUMNS CC
    JOIN DBA_CONSTRAINTS C
    ON CC.OWNER=C.OWNER
    AND CC.CONSTRAINT_NAME=C.CONSTRAINT_NAME
    WHERE C.CONSTRAINT_TYPE ='P'
    ORDER BY 3;
    SELECT SEQUENCE_NAME, MIN_VALUE, MAX_VALUE, LAST_NUMBER
    FROM DBA_SEQUENCES
    WHERE SEQUENCE_OWNER = 'PRODUCTION_OWNER'
    ORDER BY LAST_NUMBER
    If I relate the last_number with the highval and if they are same, does that mean the sequences are imported properly.
    Note: We have sequence caching done and we are on RAC.
    Edited by: ramakrishnavydyula on Jan 30, 2013 9:36 AM

    We are migrating from Oracle 10g to Oracle 11g. As part of this, we take a expdp from 10g and successfully did an import using impdp on to Oracle 11g database.
    But the problem is, there were few Primary key violations which occurred and all of them relate to sequences.
    The Maximum data in the tables and the last number in the sequences differed which caused the issue.
    it could be because of the export taken while the application is online and writing to the database.I don't know what's your database's size , and what's acceptable downtime for you, but I could never propose such a migration tecnique to any of my customers, due to excessive downtime when databases are quite large.
    It's quite obvious that you get such errors when the export is done while people are working, unless you use FLASHBACK_SCN or FLASHBACK_TIME parameter while exporting. But in this case, assuming you don't get errors (e.g. ORA-01555), you'll probably lose a lot ot transactions.
    Assuming your 10g database is running in archivelog mode (which should be the default in production DBs) did you think of using RMAN to do the migration ?

  • Oracle XQuery performance issue in XMLType column

    Dear All,
    As for oracle I'm using oracle 11g to measure the performance.
    I'm using data from XMark project which is a >100MB data of XML for bencmarking purposes.
    I make a table that contains an XMLType column and upload the data into that column, after doing that I try to do a query like this:
    select xmlquery(
    'for $i in /site/people/person
    where $i/id = "person0"
    return $i/name'
    passing BookXMLContent Returning Content)
    from Book;
    The purpose of this query is to retrieve the name of a person that have id = 'person0'
    My questions are:
    1. Did I do something wrong with my query?
    2. Is there any setting on the database that I should done prior to doing the query to done significantly better result?
    3. Is there any other approach that are much better than I currently used?
    Regards,
    Anthony Steven
    Edited by: mdrake on Nov 4, 2009 6:01 AM

    Anthony
    First, please read the licencing terms for Oracle ( And I suspect DB2, MSFT) . You are not allowed to publish externally (in any form, including forum posts :) ) the results of any benchmarking activities. I have edited your post accordingly. I hope this research is not part of a thesis or similar work that would intend making public as you and your institution would be in violation of your licence agreeement were you to do so.
    Now back to your question, how can you improve performance for XMark
    #1. Can you show us the create table statement you used, so we can see how you created your XMLType column BOOKXMLCONTENT.
    #2. Did you create any indexes
    #3. Did you look at the explain plan output.
    -Mark
    Edited by: mdrake on Nov 4, 2009 6:06 AM

  • Oracle 10g performance issues

    Hi,
    We were using Oracle 9i in Solaris 5.8 and it was working fine with some minor performance issues. We formatted the Solaris server with new Solaris 5.10 and installed Oracle 10g.
    Now we are experiencing some performance issues in Oracle 10g. This issue is arising when using through Websphere 5.1.
    We have analyzed the schema, index is rebuild, SGA is 4.5 GB, PGA is 2.0 GB, Solaris RAM is 16 GB. Also we are having some Mat Views (possibly this may cause performance issues - not sure) due to refresh.
    Also I have changed some parameters in init.ora file like query_rewrite = STALE_TOLERATED, open_cursors = 1500 etc.
    Is is something due to driver from which the data is accessed. I guess it is not utilizing the indexes on the table.
    Can anyone please suggest, what could be the issue ?

    <p>There are a lot of changes to the optimizer in the upgrade from 9i to 10g, and you need to be aware of them. There are also a number of changes to the default stats collection mechanism, so after your upgrade your statistics (hence execution paths) could change dramatically.
    </p>
    <p>
    Greg Rahn has a useful entry on his blog about stats collection, and the blog al,so points to an Oracle white paper which will give you a lot of ideas about where the optimizer changes - which may help you spot your critical issues.
    </p>
    <p>Otherwise, follow triggb's advice about using Statspack to find the SQL that is the most expensive - it's reasonably likely to be this SQL that has changed execution plans in the upgrade.
    </p>
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk

  • Oracle 11g decode issue with null

    Hi,
    we want to migrate from Oracle 10g to Oracle 11g and have an issue with decode.
    The database has the following character set settings:
    NLS_CHARACTERSET = AL32UTF8 in Oracle 11g and UTF8 in Oracle 10g
    NLS_NCHAR_CHARACTERSET = AL16UTF16
    If I try a select with decode which has null as first result argument I will get a wrong value.
    select decode(id, null, null, name) from tab1;
    ("name" is a NVARCHAR2 field. Table tab1 has only one entry and "id" is not null.)
    This select returns a value with characters which are splitted by 0 bytes.
    In Oracle 10g the value without 0 bytes is delivered.
    If I suround the decode with dump I get following results:
    select dump(decode(id, null, null, name), 1016) from tab1;
    Oracle 10g: Typ=1 Len=6 CharacterSet=AL32UTF8: 4d,61,72,74,69,6e
    Oracle 11g: Typ=1 Len=12 CharacterSet=US7ASCII: 0,4d,0,61,0,72,0,74,0,69,0,6e
    NLS_LANG has no effect on the character set of 'null' in Oracle 11g.
    Non null literals work:
    select dump(decode(id, null, 'T', name), 1016) from tab1;
    Oracle 10g: Typ=1 Len=6 CharacterSet=UTF8: 4d,61,72,74,69,6e
    Oracle 11g: Typ=1 Len=6 CharacterSet=AL32UTF8: 4d,61,72,74,69,6e
    select dump(decode(id, null, N'T', name), 1016) from tab1;
    Oracle 10g: Typ=1 Len=12 CharacterSet=AL16UTF16: 0,4d,0,61,0,72,0,74,0,69,0,6e
    Oracle 11g: Typ=1 Len=12 CharacterSet=AL16UTF16: 0,4d,0,61,0,72,0,74,0,69,0,6e
    Here the scripts for creating the table and the entry:
    create table tab1 (
    id NUMBER(3),
    name NVARCHAR2(10)
    insert into tab1 (id, name) values (1, N'Martin');
    commit;
    Is it possible to change the character set?
    Could you please help me?
    Regards
    Martin

    This doesn't have the problem.looks this doesn't solve the problem (of returning a value with characters which are splitted by 0 bytes):
    SQL> select * from v$version where rownum = 1
    BANNER                                                                         
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production         
    1 row selected.
    SQL> select dump(decode(id, null, null, name), 1016) from tab1
    union all
    select dump(case id when null then null else name end, 1016) cs from tab1
    DUMP(DECODE(ID,NULL,NULL,NAME),1016)                                           
    Typ=1 Len=12 CharacterSet=US7ASCII: 0,4d,0,61,0,72,0,74,0,69,0,6e              
    Typ=1 Len=12 CharacterSet=AL16UTF16: 0,4d,0,61,0,72,0,74,0,69,0,6e             
    2 rows selected.You need to explicitly convert the third parameter to char:
    SQL> select dump(decode(id, null, to_char(null), name), 1016) from tab1
    DUMP(DECODE(ID,NULL,TO_CHAR(NULL),NAME),1016)                                  
    Typ=1 Len=6 CharacterSet=WE8MSWIN1252: 4d,61,72,74,69,6e                       
    1 row selected.

Maybe you are looking for

  • Get error message 2753 to download to Lightroom 3.6

    Error messagr 2753 when downloading Lightroom 3.6

  • Ise 1.2.1 User session time limit?

    Is it possible to set a dot1x user to have a time limited session? Scenario is a wireless PEAP client connects to a wlc controlled access point Management wants that session to last no longer than 10 minutes and then is disconnected and not automatic

  • WPA Enterprise Question - Convince My School to Accept Macs on Wi-Fi!

    It's been a whole year since I've been able to use my MBP on my school's wifi. Currently, they are using WPA2 Personal (which I think gives you a good definition of the idiots I'm dealing with). I've just heard about WPA Enterprise and what is allows

  • Should we create the same IDOC type and message type?

    Hi all, if we want to send an IDOC from one R/3 system to another R/3 system type, should we have to create the IDOC type in one system, then repeate the steps in another system? If we don't have to, what shoud we do?

  • Zen Touch Upgr

    Hi, I have had a Zen Touch 20MB for a year which I think is fantastic and has run 00% with no problems whatsoever. I have now run our of space and want to upgrade. I'd appreciate your views on whether the Zen Touch 40MB or the Creative NOMAD Jukebox