ST03 statistics details :- Response time distribution

Hi Experts,
We have requirement to display the Percentage (cumulated) of response time distribution view of ST03 transaction code.
There is one standard infocube 0CCMSATTI.
Can anybody pls explain about the significance of fields like CNT01 to CNT09 of 0CCMSATTI infocube?
Thanks
Kind Regards
Anukul

hi,
the definition of the infoobjects are stored in strucutre SAPWLTSK.
CNT001 -- SAP Workload: Number of Steps with Response Time < 0.1s
CNT002 -- SAP Workload: No. of Steps with Response Time 0.1s< RT< 0.2s
CNT003 -- SAP Workload: No. of Steps with Response Time 0.2s< RT< 0.3s
CNT004 -- SAP Workload: No. of Steps with Response Time 0.3s< RT< 0.4s
CNT005 -- SAP Workload: No. of Steps with Response Time 0.4s< RT< 0.5s
CNT006 -- SAP Workload: No. of Steps with Response Time 0.5s< RT< 1.0s
CNT007 -- SAP Workload: No. of Steps with Response Time 1s< RT< 2s
CNT008 -- SAP Workload: No. of Steps with Response Time 2s< RT< 3s
CNT009 -- SAP Workload: No. of Steps with Response Time 3s< RT< 10s
regards,
Arvind.

Similar Messages

  • Any way to get more statistics and response-time info from cache processes?

    I'm currently involved in an evaluation of Coherence for use as a global in-core cache/DB for data in the GB range. We'll try to run multiple cache process per cache in order to stay below 512MB heap-space per cache process.
    In order to tune cache access times and understand the dynamic behaviour of the app we'd like to see some statistics info for cache access, e. g.
    - requests processed per cache (process)
    - requests answer time (per request, min, max, avg, median etc.)
    - request answer size (number of cache entries returned, size of entries etc.)
    - whatever else that helps ...
    We are in position to get this kind of statistics info on the application side (when we call Coherence-APIs) but we'd need to see the statistics on a technical level per cache process.
    We know about JMX/MBeans but that does not allow us to come up with log-files for statistical analysis and there is no way to correlate requests from the application to any statistics information (what request did cause what effect).
    Any way to tell the cache processes to provide this kind of (logging) info (mesurements per request) and some summary.

    David,
    In addition to statistics that you can collect at the client tier (measuring the latency of individual cache calls), you could use a number of attributes exposed by the ServiceMBean.
    If you look at a ServiceMBean for your partitioned cache service, the RequestAverageDuration and RequestMaxDuration attributes describe the latency of the underlying communication tier as seen from the client and does not include the serialization/deseriation cost.
    The TaskAverageDuration, TaskMaxBacklog and ThreadAverageActiveCount expose the latency and expenses as seen by the server side, excluding the network tier expenses.
    Also, there are number of statistics exposed by the CacheMBean related to the underlying backing map efficiency such as AverageGetMillis, AverageHitMillis, AverageMissMillis and AveragePutMillis.
    You can also create your own subclasses for backing map implementation and implement custom MBeans that would expose any addition information you wish to collect.
    Regards,
    Gene

  • Query Tuning - Response time Statistics collection

    Our Application is Load tested for a period of 1 hour with peak load.
    For this specific period of time, say thousands of queries gets executed in the database,
    What we need is say for one particular query " select XYZ from ABC" within this span of 1 hour, we need statistics like
    Number of times Executed
    Average Response time
    Maximum response time
    minimum response time
    90th percentile response time ( sorted in ascending order, 90th percentile guy)
    All these statistics are possible if i can get all the response times for that particular query for that period of 1 hour....
    I tried using sql trace and TKPROF but unable to get all these statistics...
    Application uses connection pooling, so connections are taken as and when needed...
    Any thoughts on this?
    Appreciate your help.

    I don't think v$sqlarea can help me out with the exact stats i needed, but certainly it has lot of other stats to take. B/w there is no dictionary view called v$sqlstats.
    There are other applications which share the same database where i am trying to capture for my application, so flushing cache which currently has 30K rows is not feasible solution.
    Any more thoughts on this?

  • Average Response Time Different in SLR/EWA (in solman) and ST03 (in satelli

    Hi All,
    We find that the average dialog response time shown in SLR and EWA reports of Solman is much lower (250 ms) whereas the dialog response time shown in ST03 transaction of the system (XI instance) is high 2000 ms.
    Which one is correct and how can I correlate these two values if both are right in their own way.
    regards

    Hello,
    The data source is ST03, therefore unless there is an issue with the collectors, you should assume the ST03 data is correct.
    Pleae run RTCCTOOL on the managed system and ensure ST-PI and ST-A/PI are current.. Also make sure your SAPOSCOL is current.
    There is a possibility that there was an error during preparing the transport to send the data back to Solution Manager, so scan the SDCCN task logs for any errors/warnings on the managed system. You can also check the transport to see what vaues are being returned to Solman's service session. If the have the same value as ST03 at this point, then there is an issue witht eh EWA/SLR provcessing. If the data is not the same here as ST03 then there was an issue with the SDCCN tasks.
    If the data collection is not at issue and the transfering of that data to Solution Manager service session had no errors, then Please follow Jared's recommendation, as this would then be an issue processing the service sessions, and this could be due to outstanding Service Content updates. This can be checked with transaction code AGS_UPDATE.
    Regards,
    Paul

  • Table name for dialog response time without GUI in ST03?

    In SWNC_COLLECTOR_GET_AGGREGATES, cnt00x (x =1,2,3... 9) gives me the proportion of transaction steps with a response time between 0s and the upper  limit of the individual response time categories (includes GUI time), broken down by task type. But where is the data without GUI time store? any ideas?

    Hello Surya
    here screenshots from my test:
    ST03 for my instance before I have executed something on it, no dialog task type:
    here I have connected to the instance through SM51, still no dialog task type:
    and here I have connected directly through SAP Logon to this instance, now my actions on the instance are displayed as dialog task type:
    If your users connect through RFC to the instance, then all their actions are RFC task type actions. As soon as a user logons directly to the instance, dialog task type will appear in ST03.
    regards,
    Alwina

  • Any suggession on Tools for Performance testing of application designed for Blackberry phones.. which can get some details on CPU usage, memory, response time?? any help on the same is much appriciated.

    Hi Team,
    Any suggession on Tools for Performance testing of application designed for Blackberry phones.. which can get some details on CPU usage, memory, response time?? any help on the same is much appriciated.
    Thank You,
    Best Regards,
    neeraj

    Hi Team,
    Any suggession on Tools for Performance testing of application designed for Blackberry phones.. which can get some details on CPU usage, memory, response time?? any help on the same is much appriciated.
    Thank You,
    Best Regards,
    neeraj

  • Report to calculate avg response time for a transaction using ST03.

    Hi Abap Gurus ,
    I want to develop a report which calculates the average response time(ST 03) for a transaction on hourly basis.
    I have read many threads like in which users are posting which tables/FM to use to extract data such as  dialog step and total response time .
    I am sure many would have created a report like this , would appeciate if you can share pseudo code for same.Any help regarding the same is highly appreciated...
    Cheers,
    Karan

    http://jakarta.apache.org/jmeter/

  • How to get query response time from ST03 via a script ?

    Hello People,
    I am trying to get average query response time for BW queries with a script (for monitoring/historisation).
    I know that this data can be found manually in ST03n in the "BI workload'.
    However, I don't know how to get this stat from a script.
    My idea is to run a SQL query to get this information, here is the state of my query :
    select count(*) from sapbw.rsddstat_olap
    where calday = 20140401
    and (eventid = 3100 or eventid = 3010)
    and steptp = 'BEX3'
    The problem is that this query is not returning the same number of navigations as the number shown in ST03n.
    Can you help me to set the correct filters to get the same number of navigation as in ST03n ?
    Regards.

    Hi Experts,
    Do you have ideas for this SQL query ?
    Regards.

  • How to Tune the Transactions/ Z - reports /Progr..of High response time

    Dear friends,
    in <b>ST03</b> work load anlysis menu.... there are some z-reports, transactions, and some programmes are noticed contineously that they are taking the <b>max. response time</b> (and mostly >90%of time is  DB Time ).
    how to tune the above situation ??
    Thank u.

    Siva,
    You can start with some thing like:
    ST04  -> Detail Analysis -> SQL Request (look at top disk reads and buffer get SQL statements)
    For the top SQL statements identified you'd want to look at the explain plan to determine if the SQL statements is:
    1) inefficient
    2) are your DB stats up to date on the tables (note up to date stats does not always means they are the best)
    3) if there are better indexes available, if not would a more suitable index help?
    4) if there are many slow disk reads, is there an I/O issue?
    etc...
    While you're in ST04 make sure your buffers are sized adequately.
    Also make sure your Oracle parameters are set according to this OSS note.
    Note 830576 - Parameter recommendations for Oracle 10g

  • High RFC response time in SAP BW system

    Hello all,
    How to analyze and fine tune high RFC response time in SAP BW system ?
    Regards,
    Archana

    Hi
    Kindly Check follows
    1. Check the RFC connections are correctly Configured or not? You can execute the program “RSRFCTRC” and get a full log of the RFC connections details
    2 you can check the BW queries are right optimized? Is this any network issues?
    3.  In which time are you facing the high rfc response? (During the 24 hours which time)?
    4. Kindly refer the SCN & SAP Notes for overall system performance
    Short Notes on PERFORMANCE MONITORING - ABAP Development - SCN Wiki
    1063061 - Information about response time in STAD/ST03
    948066 - Performance Analysis: Transactions to use
    Regards
    Sriram

  • Response time of query utterly upside down because of small where clause change

    Hello,
    I'm wondering why a small change on a where clause in a query has a dramatic impact on its response time.
    Here is the query, with its plan and a few details:
    select * from (
    SELECT xyz_id, time_oper, ...
         FROM (SELECT 
                        d.xyz_id xyz_id,
                        TO_CHAR (di.time_operation, 'DD/MM/YYYY') time_oper,
                        di.time_operation time_operation,
                        UPPER (d.delivery_name || ' ' || d.delivery_firstname) custname,
                        d.ticket_language ticket_language, d.payed,
                        dsum.delivery_mode delivery_mode,
                        d.station_delivery station_delivery,
                        d.total_price total_price, d.crm_cust_id custid,
                        d.bene_cust_id person_id, d.xyz_num, dpe.ers_pnr ers_pnr,
                        d.delivery_name,
                        TO_CHAR (dsum.first_travel_date, 'DD/MM/YYYY') first_traveldate,
                        d.crm_company custtype, UPPER (d.client_name) partyname,
                        getremark(d.xyz_num) remark,
                        d.client_app, di.work_unit, di.account_unit,
                        di.distrib_code,
                        UPPER (d.crm_name || ' ' || d.crm_firstname) crm_custname,
                       getspecialproduct(di.xyz_id) specialproduct
                   FROM xyz d, xyz_info di, xyz_pnr_ers dpe, xyz_summary dsum
                  WHERE d.cancel_state = 'N'
                 -- AND d.payed = 'N'
                    AND dsum.delivery_mode NOT IN ('DD')
                    AND dsum.payment_method NOT IN ('AC', 'AG')
                    AND d.xyz_blocked IS NULL
                    AND di.xyz_id = d.xyz_id
                    AND di.operation = 'CREATE'
                    AND dpe.xyz_id(+) = d.xyz_id
                    AND EXISTS (SELECT 1
                                  FROM xyz_ticket dt
                                 WHERE dt.xyz_id = d.xyz_id)
                    AND dsum.xyz_id = di.xyz_id
               ORDER BY di.time_operation DESC)
        WHERE ROWNUM < 1002
    ) view
    WHERE view.DISTRIB_CODE in ('NS') AND view.TIME_OPERATION > TO_DATE('20/5/2013', 'dd/MM/yyyy')
    plan with "d.payed = 'N'" (no rows, *extremely* slow):
    | Id  | Operation                          | Name             | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                   |                  |  1001 |  4166K| 39354   (1)| 00:02:59 |
    |*  1 |  VIEW                              |                  |  1001 |  4166K| 39354   (1)| 00:02:59 |
    |*  2 |   COUNT STOPKEY                    |                  |       |       |            |          |
    |   3 |    VIEW                            |                  |  1001 |  4166K| 39354   (1)| 00:02:59 |
    |   4 |     NESTED LOOPS OUTER             |                  |  1001 |   130K| 39354   (1)| 00:02:59 |
    |   5 |      NESTED LOOPS SEMI             |                  |   970 |   111K| 36747   (1)| 00:02:47 |
    |   6 |       NESTED LOOPS                 |                  |   970 |   104K| 34803   (1)| 00:02:39 |
    |   7 |        NESTED LOOPS                |                  |   970 | 54320 | 32857   (1)| 00:02:30 |
    |*  8 |         TABLE ACCESS BY INDEX ROWID| XYZ_INFO         |    19M|   704M| 28886   (1)| 00:02:12 |
    |   9 |          INDEX FULL SCAN DESCENDING| DNIN_IDX_NI5     | 36967 |       |   296   (2)| 00:00:02 |
    |* 10 |         TABLE ACCESS BY INDEX ROWID| XYZ_SUMMARY      |     1 |    19 |     2   (0)| 00:00:01 |
    |* 11 |          INDEX UNIQUE SCAN         | SB11_DSMM_XYZ_UK |     1 |       |     1   (0)| 00:00:01 |
    |* 12 |        TABLE ACCESS BY INDEX ROWID | XYZ              |     1 |    54 |     2   (0)| 00:00:01 |
    |* 13 |         INDEX UNIQUE SCAN          | XYZ_PK           |     1 |       |     1   (0)| 00:00:01 |
    |* 14 |       INDEX RANGE SCAN             | DNTI_NI1         |    32M|   249M|     2   (0)| 00:00:01 |
    |  15 |      TABLE ACCESS BY INDEX ROWID   | XYZ_PNR_ERS      |     1 |    15 |     4   (0)| 00:00:01 |
    |* 16 |       INDEX RANGE SCAN             | DNPE_XYZ         |     1 |       |     2   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
      1 - filter("DISTRIB_CODE"='NS' AND "TIME_OPERATION">TO_DATE(' 2013-05-20', 'syyyy-mm-dd'))
      2 - filter(ROWNUM<1002)
      8 - filter("DI"."OPERATION"='CREATE')
    10 - filter("DSUM"."DELIVERY_MODE"<>'DD' AND "DSUM"."PAYMENT_METHOD"<>'AC' AND "DSUM"."PAYMENT_METHOD"<>'AG')
    11 - access("DSUM"."XYZ_ID"="DI"."XYZ_ID")
    12 - filter("D"."PAYED"='N' AND "D"."XYZ_BLOCKED" IS NULL AND "D"."CANCEL_STATE"='N')
                  ^^^^^^^^^^^^^^
    13 - access("DI"."XYZ_ID"="D"."XYZ_ID")
    14 - access("DT"."XYZ_ID"="D"."XYZ_ID")
    16 - access("DPE"."XYZ_ID"(+)="D"."XYZ_ID")
    plan with "d.payed = 'N'" (+/- 450 rows, less than two minutes):
    | Id  | Operation                          | Name             | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                   |                  |  1001 |  4166K| 58604   (1)| 00:04:27 |
    |*  1 |  VIEW                              |                  |  1001 |  4166K| 58604   (1)| 00:04:27 |
    |*  2 |   COUNT STOPKEY                    |                  |       |       |            |          |
    |   3 |    VIEW                            |                  |  1002 |  4170K| 58604   (1)| 00:04:27 |
    |   4 |     NESTED LOOPS OUTER             |                  |  1002 |   130K| 58604   (1)| 00:04:27 |
    |   5 |      NESTED LOOPS SEMI             |                  |  1002 |   115K| 55911   (1)| 00:04:14 |
    |   6 |       NESTED LOOPS                 |                  |  1476 |   158K| 52952   (1)| 00:04:01 |
    |   7 |        NESTED LOOPS                |                  |  1476 | 82656 | 49992   (1)| 00:03:48 |
    |*  8 |         TABLE ACCESS BY INDEX ROWID| XYZ_INFO         |    19M|   704M| 43948   (1)| 00:03:20 |
    |   9 |          INDEX FULL SCAN DESCENDING| DNIN_IDX_NI5     | 56244 |       |   449   (1)| 00:00:03 |
    |* 10 |         TABLE ACCESS BY INDEX ROWID| XYZ_SUMMARY      |     1 |    19 |     2   (0)| 00:00:01 |
    |* 11 |          INDEX UNIQUE SCAN         | AAAA_DSMM_XYZ_UK |     1 |       |     1   (0)| 00:00:01 |
    |* 12 |        TABLE ACCESS BY INDEX ROWID | XYZ              |     1 |    54 |     2   (0)| 00:00:01 |
    |* 13 |         INDEX UNIQUE SCAN          | XYZ_PK           |     1 |       |     1   (0)| 00:00:01 |
    |* 14 |       INDEX RANGE SCAN             | DNTI_NI1         |    22M|   168M|     2   (0)| 00:00:01 |
    |  15 |      TABLE ACCESS BY INDEX ROWID   | XYZ_PNR_ERS      |     1 |    15 |     4   (0)| 00:00:01 |
    |* 16 |       INDEX RANGE SCAN             | DNPE_XYZ         |     1 |       |     2   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       1 - filter("DISTRIB_CODE"='NS' AND "TIME_OPERATION">TO_DATE(' 2013-05-20', 'syyyy-mm-dd'))
       2 - filter(ROWNUM<1002)
       8 - filter("DI"."OPERATION"='CREATE')
      10 - filter("DSUM"."DELIVERY_MODE"<>'DD' AND "DSUM"."PAYMENT_METHOD"<>'AC' AND "DSUM"."PAYMENT_METHOD"<>'AG')
      11 - access("DSUM"."XYZ_ID"="DI"."XYZ_ID")
      12 - filter("D"."XYZ_BLOCKED" IS NULL AND "D"."CANCEL_STATE"='N')
      13 - access("DI"."XYZ_ID"="D"."XYZ_ID")
      14 - access("DT"."XYZ_ID"="D"."XYZ_ID")
      16 - access("DPE"."XYZ_ID"(+)="D"."XYZ_ID")
    XYZ.PAYED values breakdown:
    P   COUNT(1)
    Y   12202716
    N    9430207
    tables nb of records:
    TABLE_NAME           NUM_ROWS
    XYZ                  21606776
    XYZ_INFO            186301951
    XYZ_PNR_ERS           9716471
    XYZ_SUMMARY          21616607
    Everything that comes inside the "select * from(...) view" parentheses is defined in a view. We've noticed that the line "AND d.payed = 'N'" (commented above) is the guilty clause: the query takes one or two seconds to return between 400 and 500 rows if this line is removed, when included in the query, the response time then switches to *hours* -sic !- but then the result set is empty (no rows returned). The plan is exactly the same whether this "d.payed = 'N'" is added or removed, I mean the nb of steps, access paths, join order etc., only the rows/bytes/cost columns values change, as you can see.
    We've found no other way of solving this perf issue but by taking out this "d.payed = 'N'" condition and setting it outside the view along with view.DISTRIB_CODE and view.TIME_OPERATION.
    But we would like to understand why such a small change on the XYZ.PAYED column turns everything upside down that much, and we'd like to be able to tell the optimizer to perform this check on payed = 'N' by itself in the end, just like we did, through the use of a hint if possible...
    Anybody ever encountered such a behaviour before ? Do you have any advice regarding the use of a hint to reach the same response time as that we've got by setting the payed = N condition outside of the view definition ??
    Thanks a lot in advance.
    Regards,
    Seb

    I am really sorry I couldn't get back earlier to this forum...
    Thanks to you all for your answers.
    First I'd just like to correct a small mistake I made, when writing
    "the query takes one or two seconds": I meant one or 2 *minutes*. Sorry.
    > What table/columns are indexed by "DNTI_NI1"?
    aaaa.dnti_ni1 is an index ON aaaa.xyz_ticket(xyz_id, ticket_status)
    > And what are the indexes on xyz table?
    Too many:
    XYZ_ARCHIV_STATE_IND           ARCHIVE_STATE
    XYZ_BENE_CUST_ID_IND           BENE_CUST_ID
    XYZ_BENE_TTL_IND               BENE_TTL
    XYZ_CANCEL_STATE_IND           CANCEL_STATE
    XYZ_CLIENT_APP_NI              CLIENT_APP
    XYZ_CRM_CUST_ID_IND            CRM_CUST_ID
    XYZ_DELIVE_MODE_IND            DELIVERY_MODE
    XYZ_DELIV_BLOCK_IND            DELIVERY_BLOCKED
    XYZ_DELIV_STATE_IND            DELIVERY_STATE
    XYZ_XYZ_BLOCKED                XYZ_BLOCKED
    XYZ_FIRST_TRAVELDATE_IND       FIRST_TRAVELDATE
    XYZ_MASTER_XYZ_IND             MASTER_XYZ_ID
    XYZ_ORG_ID_NI                  ORG_ID
    XYZ_PAYMT_STATE_IND            PAYMENT_STATE
    XYZ_PK                         XYZ_ID
    XYZ_TO_PO_IDX                  TO_PO
    XYZ_UK                         XYZ_NUM
    For ex. XYZ_CANCEL_STATE_IND on CANCEL_STATE seems superfluous to me, as the column may only contain Y or N (or be null)...
    > Have you traced both cases to compare statistics? What differences did it reveal?
    Yes but it only shows more of *everything* (more tables blocks accessed, the same
    for indexes blocks, for almost all objects involved) for the slowest query !
    Greping WAIT on the two trc files made for every statement and counting the
    object IDs access show that the quicker query requires much less I/Os; the
    slowest one overall needs much more blocks to be read (except for the indexes
    DNSG_NI1 or DNPE_XYZ for example). Below I replaced obj# with the table/index
    name, the first column is the figure showing how many times the object was
    accessed in the 10053 file (I ctrl-C'ed my second execution ofr course, the
    figures should be much higher !!):
    [login.hostname] ? grep WAIT OM-quick.trc|...|sort|uniq -c
        335 XYZ_SUMMARY
      20816 AAAA_DSMM_XYZ_UK (index on xyz_summary.xyz_id)
        192 XYZ
       4804 XYZ_INFO
        246 XYZ_SEGMENT
          6 XYZ_REMARKS
         63 XYZ_PNR_ERS
        719 XYZ_PK           (index on xyz.xyz_id)
       2182 DNIN_IDX_NI5     (index on xyz.xyz_id)
        877 DNSG_NI1         (index on xyz_segment.xyz_id, segment_status)
        980 DNTI_NI1         (index on xyz_ticket.xyz_id, ticket_status)
        850 DNPE_XYZ         (index on xyz_pnr_ers.xyz_id)
    [login.hostname] ? grep WAIT OM-slow.trc|...|sort|uniq -c
       1733 XYZ_SUMMARY
      38225 AAAA_DSMM_XYZ_UK  (index on xyz_summary.xyz_id)
       4359 XYZ
      12536 XYZ_INFO
         65 XYZ_SEGMENT
         17 XYZ_REMARKS
         20 XYZ_PNR_ERS
       8598 XYZ_PK
       7406 DNIN_IDX_NI5
         29 DNSG_NI1
       2475 DNTI_NI1
         27 DNPE_XYZ
    The overwhelmingly dominant wait event is by far 'db file sequential read':
    [login.hostname] ? grep WAIT OM-*elect.txt|cut -d"'" -f2|sort |uniq -c
         36 SQL*Net message from client
         38 SQL*Net message to client
    107647 db file sequential read
          1 latch free
          1 latch: object queue header operation
          3 latch: session allocation
    > It will be worth knowing the estimations...
    It show the same plan with a higher cost when PAYED = N is added:
    SQL> select * from sb11.dnr d
      2* where d.dnr_blocked IS NULL and d.cancel_state = 'N'
    SQL> /
    | Id  | Operation                   | Name                 | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT            |                      |  1002 |   166K|    40   (3)| 00:00:01 |
    |*  1 |  TABLE ACCESS BY INDEX ROWID| XYZ                  |  1002 |   166K|    40   (3)| 00:00:01 |
    |*  2 |   INDEX RANGE SCAN          | XYZ_CANCEL_STATE_IND |       |       |     8   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       1 - filter("D"."XYZ_BLOCKED" IS NULL)
       2 - access("D"."CANCEL_STATE"='N')
    SQL> select * from sb11.dnr d
      2  where d.dnr_blocked IS NULL and d.cancel_state = 'N'
      3* and d.payed = 'N'
    SQL> /
    Execution Plan
    Plan hash value: 1292668880
    | Id  | Operation                   | Name                 | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT            |                      |  1001 |   166K|    89   (3)| 00:00:01 |
    |*  1 |  TABLE ACCESS BY INDEX ROWID| XYZ                  |  1001 |   166K|    89   (3)| 00:00:01 |
    |*  2 |   INDEX RANGE SCAN          | XYZ_CANCEL_STATE_IND |       |       |    15   (0)| 00:00:01 |

  • PRVF-5636 : The DNS response time for an unreachable node exceeded "15000"

    Friends,
    OS: OEL 6.3 64 bit
    DB: 11.2.0.3
    Environment: VirtualBox 4.1.20
    i have 3 nodes. one is for dns and other two for rac1 and rac2
    i have done all the preliminary steps. then i started
    [oracle@rac1 grid]$ ./runcluvfy.sh stage -pre crsinst -n rac1,rac2 -verboseand all are passed except the above subjected error.
    Checking DNS response time for an unreachable node
      Node Name                             Status                 
      rac2                                  failed                 
      rac1                                  failed                 
    PRVF-5636 : The DNS response time for an unreachable node exceeded "15000" ms on following nodes: rac2
    File "/etc/resolv.conf" is not consistent across nodesok....now to solve this error...i tried many options...but not able to succeed.
    Try 1:
    [root@mydns ~]# time nslookup rac1
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac1.sathy.net
    Address: 192.168.1.3
    real    0m0.051s
    user    0m0.003s
    sys     0m0.011s
    [root@mydns ~]# [oracle@rac1 grid]$ time nslookup mydns
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   mydns.sathy.net
    Address: 192.168.1.2
    real    0m0.019s
    user    0m0.005s
    sys     0m0.008s
    [oracle@rac1 grid]$
    [root@rac2 ~]# time nslookup rac1
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac1.sathy.net
    Address: 192.168.1.3
    real    0m0.019s
    user    0m0.008s
    sys     0m0.005s
    [root@rac2 ~]# Try 2:
    [root@mydns ~]# nslookup sathy-scan
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.8
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.9
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.7
    [root@mydns ~]# nslookup rac1
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac1.sathy.net
    Address: 192.168.1.3
    [root@mydns ~]# nslookup rac2
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac2.sathy.net
    Address: 192.168.1.5
    [root@mydns ~]#
    [oracle@rac1 grid]$ nslookup sathy-scan
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.9
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.7
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.8
    [oracle@rac1 grid]$ nslookup rac2
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac2.sathy.net
    Address: 192.168.1.5
    [oracle@rac1 grid]$ nslookup rac1
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac1.sathy.net
    Address: 192.168.1.3
    [oracle@rac1 grid]$
    [root@rac2 ~]# nslookup sathy-scan
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.7
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.8
    Name:   sathy-scan.sathy.net
    Address: 192.168.1.9
    [root@rac2 ~]# nslookup rac1
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac1.sathy.net
    Address: 192.168.1.3
    [root@rac2 ~]# nslookup rac2
    Server:         192.168.1.2
    Address:        192.168.1.2#53
    Name:   rac2.sathy.net
    Address: 192.168.1.5
    [root@rac2 ~]# Try 3:
    [root@mydns ~]# cat /etc/named.conf
    // Default named.conf generated by install of bind-9.2.4-30.el4_7.2
    options {
            listen-on port 53 { any; };
            directory "/var/named";
            dump-file "/var/named/data/cache_dump.db";
            statistics-file "/var/named/data/named_stats.txt";
            forwarders { 192.168.1.1; };
    include "/etc/rndc.key";
    zone "sathy.net" IN {
    type master;
    file "sathy.net.zone";
    allow-update { none; };
    zone "1.168.192.in-addr.arpa." IN {
    type master;
    file "1.168.192.in-addr.arpa";
    allow-update { none; };
    zone "." in {
    type hint;
    file "/dev/null";
    [root@mydns ~]# Try 4:
    [root@mydns ~]# cat /etc/selinux/config
    # This file controls the state of SELinux on the system.
    # SELINUX= can take one of these three values:
    #     enforcing - SELinux security policy is enforced.
    #     permissive - SELinux prints warnings instead of enforcing.
    #     disabled - No SELinux policy is loaded.
    SELINUX=disabled
    # SELINUXTYPE= can take one of these two values:
    #     targeted - Targeted processes are protected,
    #     mls - Multi Level Security protection.
    #SELINUXTYPE=targeted
    [root@mydns ~]#
    [oracle@rac1 grid]$ cat /etc/selinux/config
    # This file controls the state of SELinux on the system.
    # SELINUX= can take one of these three values:
    #     enforcing - SELinux security policy is enforced.
    #     permissive - SELinux prints warnings instead of enforcing.
    #     disabled - No SELinux policy is loaded.
    SELINUX=disabled
    # SELINUXTYPE= can take one of these two values:
    #     targeted - Targeted processes are protected,
    #     mls - Multi Level Security protection.
    #SELINUXTYPE=targeted
    [oracle@rac1 grid]$
    [root@rac2 ~]# cat /etc/selinux/config
    # This file controls the state of SELinux on the system.
    # SELINUX= can take one of these three values:
    #     enforcing - SELinux security policy is enforced.
    #     permissive - SELinux prints warnings instead of enforcing.
    #     disabled - No SELinux policy is loaded.
    SELINUX=disabled
    # SELINUXTYPE= can take one of these two values:
    #     targeted - Targeted processes are protected,
    #     mls - Multi Level Security protection.
    #SELINUXTYPE=targeted
    [root@rac2 ~]# what else i have to check to solve this error?
    for the past 2 days..i am trying to solve this error. may be somewhere a small mistake is happened from my side, but i am not able to find out.
    hope somebody will find out and help me.
    thanks

    ok...here it is....
    [root@mydns ~]# cat /etc/resolv.conf
    # Generated by NetworkManager
    search sathy.net
    nameserver 192.168.1.2
    [root@mydns ~]#
    [root@rac1 ~]# cat /etc/resolv.conf
    # Generated by NetworkManager
    search sathy.net
    nameserver 192.168.1.2
    [root@rac1 ~]#
    [root@rac2 ~]# cat /etc/resolv.conf
    # Generated by NetworkManager
    search sathy.net
    nameserver 192.168.1.2
    [root@rac2 ~]#
    [root@mydns ~]# cat /etc/hosts
    #127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
    #::1         localhost localhost.localdomain localhost6 localhost6.localdomain6
    127.0.0.1               localhost.localdomain localhost
    ::1            localhost6.localdomain6 localhost6
    # Public Network – (eth0)
    192.168.1.3     rac1.sathy.net    rac1
    192.168.1.5     rac2.sathy.net    rac2
    # Public Virtual IP (VIP) addresses – (eth0:1)
    192.168.1.4     rac1-vip.sathy.net rac1-vip
    192.168.1.6     rac2-vip.sathy.net rac2-vip
    # Private Interconnect – (eth1)
    192.168.2.3      rac1-priv.sathy.net       rac1-priv
    192.168.2.5      rac2-priv.sathy.net       rac2-priv
    #Compute Node Client Access SCAN Interface details
    #192.168.1.7   sathy-scan.sathy.net  sathy-scan
    #192.168.1.8   sathy-scan.sathy.net  sathy-scan
    #192.168.1.9   sathy-scan.sathy.net  sathy-scan
    [root@mydns ~]#
    [root@rac1 ~]# cat /etc/hosts
    #127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
    #::1         localhost localhost.localdomain localhost6 localhost6.localdomain6
    127.0.0.1               localhost.localdomain localhost
    ::1            localhost6.localdomain6 localhost6
    # Public Network – (eth0)
    192.168.1.3     rac1.sathy.net    rac1
    192.168.1.5     rac2.sathy.net    rac2
    # Public Virtual IP (VIP) addresses – (eth0:1)
    192.168.1.4     rac1-vip.sathy.net rac1-vip
    192.168.1.6     rac2-vip.sathy.net rac2-vip
    # Private Interconnect – (eth1)
    192.168.2.3      rac1-priv.sathy.net       rac1-priv
    192.168.2.5      rac2-priv.sathy.net       rac2-priv
    #Compute Node Client Access SCAN Interface details
    #192.168.1.7   sathy-scan.sathy.net  sathy-scan
    #192.168.1.8   sathy-scan.sathy.net  sathy-scan
    #192.168.1.9   sathy-scan.sathy.net  sathy-scan
    [root@rac1 ~]#
    [root@rac2 ~]# cat /etc/hosts
    #127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
    #::1         localhost localhost.localdomain localhost6 localhost6.localdomain6
    127.0.0.1               localhost.localdomain localhost
    ::1            localhost6.localdomain6 localhost6
    # Public Network – (eth0)
    192.168.1.3     rac1.sathy.net    rac1
    192.168.1.5     rac2.sathy.net    rac2
    # Public Virtual IP (VIP) addresses – (eth0:1)
    192.168.1.4     rac1-vip.sathy.net rac1-vip
    192.168.1.6     rac2-vip.sathy.net rac2-vip
    # Private Interconnect – (eth1)
    192.168.2.3      rac1-priv.sathy.net       rac1-priv
    192.168.2.5      rac2-priv.sathy.net       rac2-priv
    #Compute Node Client Access SCAN Interface details
    #192.168.1.7   sathy-scan.sathy.net  sathy-scan
    #192.168.1.8   sathy-scan.sathy.net  sathy-scan
    #192.168.1.9   sathy-scan.sathy.net  sathy-scan
    [root@rac2 ~]#
    [oracle@rac1 ~]$ ssh rac2 date
    Sun Sep  9 22:04:45 IST 2012
    [oracle@rac1 ~]$ su - grid
    Password:
    [grid@rac1 ~]$ ssh rac2 date
    Sun Sep  9 22:04:54 IST 2012
    [grid@rac1 ~]$
    [oracle@rac2 ~]$ ssh rac1 date
    Sun Sep  9 22:04:18 IST 2012
    [oracle@rac2 ~]$ su - grid
    Password:
    [grid@rac2 ~]$ ssh rac1 date
    Sun Sep  9 22:04:31 IST 2012
    [grid@rac2 ~]$ Now the error is ...
    PRVF-5637 : DNS response time could not be checked on following nodes: rac2,rac1what else?
    i wonder how it was successful in my previous installation which in oel 5.7....
    Edited by: OraDB on Sep 9, 2012 9:44 AM

  • Weekly response time of workproces

    hi
    1.how to check weekly  response time of a particular work process?
    2. does abap program which is usually developed by our abapers is client specific or cross client?

    Hi Ramesh,
    ABAP programs are cross client. They are repository objects!!!.
    Dont know if it is possible to check repsonse time of a process as such. Yes you can do an performance analysis of dialog/background,update WPs etc. as a bunch but individual process analysis in terms of weekly data.. I have no idea. Anyways it makes sense to do this analysis for type of workprocess and not a workprocess in itself. It is not as if a workprocess is performing well you are going to replace it with something new :-). For system performance analysis ST02,ST03,ST03N,ST04 and ST06 are best transactions though check ST03 and ST03N in your case of process response time analysis. May be you will find historical details of an individual work process using ST03N. Wish u luck.
    Please award points for useful info.
    Regards.
    Ruchit.

  • ST03N Tcode not showing last 3 days average response time

    Hi,
    In my Production system not showing last 3 days average response time.In our production 1 (CI+DB) and 3 dialog instances are there. In all servers sapcollector working fine but in all servers not showing last 3 days values.Please advise solution.
    Thanks,
    Srinivas.

    The following are the prerequisites needed to get the workload monitor(st03) working.
    1)      The system time zone of the SAP and the Operating system has to be same. This can be done by customizing time zones of the SAP using the Tcode STZAC.
    2)     All the standard background jobs has to be scheduled to run. This can be done by scheduling the standard background jobs using the Tcode SM36.
    3)     The following parameters has to be added to the instance profile.
    Parameter                                                  Value(default)                                                  Description
    i.     stat/as_collect                                                          = 2                                          Application statistic: detailed records.
    ii.     stat/as_level                                                              = 1                                          Records for application statistic are active.
    iii.     stat/max_files                                                           = 48                                        Max. no. of stat files to be kept.
    iv.     stat/as_max_files                                                    = 48                                        Max. no. of appl. stat files to be kept.
    This can be done by editing the profiles using the Tcode RZ10.

  • How to get Query response Time?

    II am on BI 7.0. I ran some queries using RSRT command. I want to find how much time the queries took.
    I went to
    st03 -> expert mode -> BI system load-> select today / week/month according to the query runtime day
    I do not see any Info Providers. Query was on a cube so why no Info Providers.
    Does something have to turned on InfoPorvider to show.
    When I look  in RSDDSTAT_OLAP table, I do see many rows but cannot make any sense. Is there some documentation on how to get total  query time from this table?
    Is there any other way to get query response time?
    Thanks a lot.

    HI,
    why not use RSRT ? You can add database statistics option in "Execut & Debug" and you get all the runtime metrics of your query
    In transaction RSRT, enter the query name and press u2018Execute +Debugu2019.
    Selecting u2018Display Statistics Datau2019 .
    After executing the query will return a list of the measured metrics.
    The event id / text describes the steps  (duration in seconds):
    "OLAP: Read data" gives the SQL statements repsonse time (ok - because the SAP
    application server acts as an Oracle client a little network traffic from the db server is included,
    but as far as you not transferring zillions of rows it can be ignored)
    But it gives you much more (i.e. if the OLAP cache gets used or not )...
    In the "Aggreagate statistcs" you get all the infoproviders involved in that query.
    bye
    yk

Maybe you are looking for

  • Hmm, to the person who asked why the store carries cover versions...

    Not sure what caused that thread to disappear... Apple don't monitor these forums for feedback nor answer questions directly. You may send your thoughts to iTunes Feedback if you wish. There is a long tradition of different artists recording the same

  • Having Problems opening After Effects CC

    Having Problems opening After Effects CC - wont open and I get an error message 0cx00000005

  • Photosmart 6520: disconnect​s from EPRINT when it sleeps

    I just bought this printer today.  I got it set up just fine, and I can print wirelessly just fine, and I can also print using the email address just fine.  The printer seems to work perfectly fine.  HOWEVER, every time it goes to sleep, it disconnec

  • How to make a Java Applet work beyond proxy?

    Hi All, I have problems with my applet working through proxy.I am using a client applet which makes a socket connection to a Java application running on the same pc as the web server. Everything works fine when I am directly connected.However it does

  • About standard text

    1.How do u transport STANDARD TEXT from development client to quality client or to development?