More response time - STMS_QA

HI,
The transaction code STMS_QA is taking more time to execute ( nealry 2 mins). while analysing in the STAD it is showing it due to more processing time. Can some one help how to solve this issue.
Thansks,
Prashanth

Hi Deepak,
Thanks for you reply. Below are the answers for your questions.
1) Yes trans directory is mounted and currently in Development system.
2)  All the TR's.
3)  We have enough space in trans directory.
Could you please suggest the better solution to solve this issue ASAP?
Thanks,
Prashant Raja.

Similar Messages

  • More response time

    when program execute at time my program takes more response time so what is reason.

    hi krupa,
    pls check with transaction SE30[Run time analysis] and identify the area where it is taking time.
    pls Re-check u r code with below points.
    1) should not write any select statement in side the loop and endloop.
    2) Use Select into corresponding fields insted of select *
    3) refresh your internal tables after data is processed.
    4) use move corresponding insted of move single
    5)use type insted of like while declaring the parameters.
    6) it would be better if u use workarea
    this is from ABAP side.
    from basis side,
    pls check whether the server is working fine or not??
    if u want more help, pls post u r code.

  • SESSION_MANAGER  using more response time

    Dear Experts ,
    My SAP Production Server performance these days show very slow.I check in Early Watch Report that Top 5 Transistion.
    The "Top 5 transactions" diagram  show the average response time in dialog tasks for the top 5 transactions
    In this upper one is SESSION_MANAGER .
    Rest all is for Z development.
    Transaction     Type     Dialog Steps     Total Resp. Time in %     Avg. Resp. Time in ms     Avg. CPU Time in ms     Avg. DB Time in ms     Avg. GUI Time in ms
    Total          752009     100,0     1.213,7     177,1     758,1     229,5
    SESSION_MANAGER     DIA     34155     5,7     1.524,6     164,7     595,5     627,5
    Now How to analyses and solve  for  SESSION_MANAGER .
    thanks in advanced
    pooja

    in most companies, SESSION_MANAGER always show as one of the slowest transaction. This is usually not an issue.
    check other reports and analyze them but do not lose your time on SESSION_MANAGER.

  • Query taking more response time

    Hi Masters,
    Today i am tuning One Query
    The Following points are the Info regarding my query.
    Table1- has 100 million records.
    Table2 - has 7000 records.
    I am Using table2 as Driving table.
    I used the Ordered, index hint, use_nl hint and all
    My Query:
    select
    max(t1.date)
    from table2 t2*, table1 t1*
    where t2.col1=12354
    and
    t1.col2=t2.col2 -- Is the  Join b/w the two tables.
    and t1.col3=0
    Regarding Index Details:
    There is a composite index on table1 (col2,col3).
    Also there is an index on table1(col1).
    The Query using the Right Index Only.
    I found out from the explain Plan.
    Then what is the Problem with the above Query.
    Do i need to add any more Hints to work better?
    One more Info man,
    When i tried to retrieve only the t1.date column not as max(t.date), it will respond in subsequent mill seconds.
    select
    t1.date
    from table2 t2, table1 t1
    where t2.col1=12354
    and
    t1.col2=t2.col2 -- Is the Join b/w the two tables.
    and t1.col3=0
    And the result contains, 257000 records.
    For finding the max value only it will take several hundred times more!!
    Try to give some useful feedback!!
    Edited by: user10376641 on Aug 21, 2009 1:32 AM

    If i am harsh....sorry for that.
    even after Sybrand's post, you asked for "anyother feedback". so i thought you had not gone through the link completely.
    but good to hear that you are analyzing the things. please post your findings, so that we can help you out.
    once again, sorry for my words.

  • High Response time

    Hi All ,
    In my SANDBOX , in time of transaction code execution it's OK. But after that in the time of any customization, table viewing it is using much more response time & even sometimes session is going terminated . How to fix & solve this problem. Can anybody help me ?
    Regards
    Asad

    Hi,
    I would look into following areas in st03n which would give you idea which transactions are taking more time and which area its struggling such as db time, cpu time, number of db reads, gui time and this would give an idea.
    Also check if your system has sufficient resources in terms of memory, cpu and paging files(swap size). Please check the paging out/paging rates as this would indicate problem areas.
    You can do a trace as mentioned earlier and check for the st22 dumps and system logs sm21 on what the errors messages indicate. This should really make it clear on what the core issues are and if you have any queries, pls post it here so that we can check it for you.
    Cheers Sam

  • Higher response time after adding more CPU

    Hello all,
    have you ever experienced that after incresing SAPs (adding 2 more CPU + 4 GB of memory) on the VMware ESX 3.5 the DB response time went from 150 to 400-500 ms ? NO DB parameter were changed. In app server, only extended memory and program buffer were enlarged.
    We are running ECC6 system on MAXDB 7.6.6.3 (SLES9). MAXCPU is set to 1
    Any idea ?
    Thank you.
    Pavol
    Edited by: Pavol Simko on Jul 28, 2009 10:35 AM

    > We are running ECC6 system on MAXDB 7.6.6.3 (SLES9). MAXCPU is set to 1
    I'd always set MAXCPU to the number of physical CPUs/cores you have. It's VERY unlikely that the database will use more CPU than the application.
    The database can't scale, if one CPU is used other requests have to wait.
    Markus

  • Response time much more in migrated AIX database than in Windows database

    We migrated a 9.2.0.6 database from windows to AIX (on same oracle version). Initially we were getting really slow response on direct sqls and batch sqls fom application server. Then we tuned aix database a bit and got a better response time for direct sqls on aix database.
    Now queries run faster on aix database than the db on windows.
    However, the queries from applciation (ACCURATE application) are still twice slower on aix database compared to windows database.
    Can anyone suggest a possible cause or tuning area?
    We have checked on network level and found no bottlenecks.
    Regards,
    Garima

    user650817 wrote:
    We migrated a 9.2.0.6 database from windows to AIX (on same oracle version). Initially we were getting really slow response on direct sqls and batch sqls fom application server. Then we tuned aix database a bit and got a better response time for direct sqls on aix database.
    Now queries run faster on aix database than the db on windows.
    However, the queries from applciation (ACCURATE application) are still twice slower on aix database compared to windows database.
    Can anyone suggest a possible cause or tuning area?Garima,
    if you have particular queries that you know of that are behaving differently, you could start right away by comparing the execution plans and tracing the execution using the SQL trace facility and the "tkprof" trace file analyzer.
    If you want to get an overview of the database performance, you should consider installing STATSPACK on both systems if you haven't done so yet and create snapshots to compare statspack reports from both systems for a representative time period. This way you can gather information about the top wait events and top SQLs of the periods analyzed.
    By the way, how did you "migrate" the database? Did you use export/import to migrate? What about the statistics for the cost based optimizer, could these be different between the two systems or are you still using the rule based optimizer?
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • HP Pavilion hard drive response time is very slow but tests pass (evenutally)

    I have owned a HP Pavilion M1199a originally running Windows XP, but has been running Windows Vista 32-bit for the last 3 years or more.
    The PC started to lock up for 30 seconds or more (2 weeks ago), with the disk light constantly on then go again for no apparent reason. These lock ups have become more frequent and for longer durations. I removed anti-virus software, and a few other applications/services but generally this computer is fairly clean as I use it as a media center PC. 
    I have tried checking the disk (SATA) for errors, and it passes although the tests take a lot longer than they should. The resource monitor shows no excess CPU usage and there is plenty of memory available. The disk monitor shows response times of 5000-20000 ms.
    What is the best way to proceed from here? Is there a SATA controller or motherboard (ASUS PTGD1-LA) test? Should I buy another SATA drive and try that? I suspect that it is either the drive, drive controller and/or the motherboard that is failing but I don't know how to isolate the problem. The computer hardware configuration has remained the same for years.
    The OS has the automatic updates enabled and I uninstalled the recent ones in case they were somehow causing an issue. 

    tr3v wrote:
    Thanks for replying.
    1/ No, but I am running Vista, so assume this will be the same? Or should I just look at the tmp and temp environment variables to see what folders are being used?
     In Vista type temp in the Search programs and files box and double click on the temp files icon that appears above. Delete all the files in the folder that you can. It is safe as they are exactly what they are called and that is temporary files. If you haven't done this ever or in quite a while there should be a noticeable increase in the operating system's response time.
    2/ Yes. It completes after a very long time. 
    5/ No - but will try this out too. 
    ****Please click on Accept As Solution if a suggestion solves your problem. It helps others facing the same problem to find a solution easily****
    2015 Microsoft MVP - Windows Experience Consumer

  • SAP GoLive : File System Response Times and Online Redologs design

    Hello,
    A SAP Going Live Verification session has just been performed on our SAP Production environnement.
    SAP ECC6
    Oracle 10.2.0.2
    Solaris 10
    As usual, we received database configuration instructions, but I'm a little bit skeptical about two of them :
    1/
    We have been told that our file system read response times "do not meet the standard requirements"
    The following datafile has ben considered having a too high average read time per block.
    File name -Blocks read  -  Avg. read time (ms)  -Total read time per datafile (ms)
    /oracle/PMA/sapdata5/sr3700_10/sr3700.data10          67534                         23                               1553282
    I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    2/
    We have been asked  to increase the size of the online redo logs which are already quite large (54Mb).
    Actually we have BW loading that generates "Chekpoint not comlete" message every night.
    I've read in sap note 79341 that :
    "The disadvantage of big redo log files is the lower checkpoint frequency and the longer time Oracle needs for an instance recovery."
    Frankly, I have problems undertanding this sentence.
    Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right ?
    But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    Thank you.
    Any useful help would be appreciated.

    Hello
    >> I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    The recommended ("standard") values are published at the end of sapnote #322896.
    23 ms seems really a little bit high to me - for example we have round about 4 to 6 ms on our productive system (with SAN storage).
    >> Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right?
    Correct.
    >> But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    A checkpoint is occured on every logswitch (of the online redologfiles). On a checkpoint event the following 3 things are happening in an oracle database:
    Every dirty block in the buffer cache is written down to the datafiles
    The latest SCN is written (updated) into the datafile header
    The latest SCN is also written to the controlfiles
    If your redologfiles are larger ... checkpoints are not happening so often and in this case the dirty buffers are not written down to the datafiles (in the case of no free space in the buffer cache is needed). So if your instance crashes you need to apply more redologs to the datafiles to be in a consistent state (roll forward). If you have smaller redologfiles more log switches are occured and so the SCNs in the data file headers (and the corresponding data) are closer to the newest SCN -> ergo the recovery is faster.
    But this concept does not really fit the reality because of oracle implements some algorithm to reduce the workload for the DBWR in the case of a checkpoint.
    There are also several parameters (depends on the oracle version) which control that a required recovery time is kept. (for example FAST_START_MTTR_TARGET)
    Regards
    Stefan

  • Response time for Error Messages - Please Help

    Hi
    I have a PRO C application talking to an Oracle database.
    The Response time for successful query is within desirable limits.
    But when there is a error condition (eg SQLError -3113,or connection refused) it takes more than 9 minutes for the database to respond with the error code.
    This condition is observed with only one database while the others are working fine.
    What is the reason for this? Can’t it be reduced?
    Regards
    David

    ever been faced with the same problem ?
    Why deleting ? It that only one way to fix this problem ?
    What are the others doing in such cases. Or am I the only one person
    where has this special problem on the world. Becides I dont believe
    in solving the problem through removing mentioned directory and
    reinstalling. Nevertheless I will try it. I let you know about the result.
    bye
    sas

  • How to find the Response time for a particular Transaction

    Hello Experts,
            Am implementing a BAdI to achieve some customer enhancement for XD01 Transaction . I need to confirm to customer that after the implementation and before implementation what is the response time of the system
    Response time BEFORE BAdI Implementation
    Response time AFTER BAdI Implementation
    Where can i get this.
    Help me in this regard
    Best Regards
    SRiNi

    Hello,
    Within STAD, enter the time range that the user was executing the transaction within as well as the user name. The time field indicates the time when the transaction would have ended. STAD adds some extra time on using your time interval. Depending on how long the transaction ran, you can set the length you want it to display. This means that if it is set to 10, STAD will display statistical records from transactions that ended within that 10 minute period.
    The selection screen also gives you a few options for display mode.
    - Show all statistic records, sorted by star
    This shows you all of the transaction steps, but they are not grouped in any way.
    -Show all records, grouped by business transaction
    This shows the transaction steps grouped by transaction ID (shown in the record as Trans. ID). The times are not cumulative. They are the times for each individual step.
    -Show Business Transaction Tots
    This shows the transaction steps grouped by transaction ID. However, instead of just listing them you can drill from the top level down. The top level will show you the overall response time, and as you drill down, you can get to the overall response time.
    Note that you also need to add the user into the selection criteria. Everything else you can leave alone in this case.
    Once you have the records displayed, you can double click them to get a detailed record. This will show you the following:
    - Breakdown of response time (wait for work process, processing time, load time, generating time, roll time, DB time, enqueue time). This makes STAD a great place to start for performance analysis as you will then know whether you will need to look at SQL, processing, or any other component of response time first.
    - Stats on the data selected within the execution
    - Memory utilization of the transaction
    - RFCs executed (including the calling time and remote execution time - very useful with performance analysis of interfaces)
    - Much more.
    As this chain of comments has previously indicated, you are best off using STAD if you want an accurate indication of response time. The ST12 (combines SE30 ABAP trace and ST05 SQL trace) trace times are less accurate that the values you get from ST12. I am not discounting the value of ST12 by any means. This is a very powerful tool to help you tune your transactions.
    I hope this information is helpful!
    Kind regards,
    Geoff Irwin
    Senior Support Consultant
    SAP Active Global Support

  • Response time too slow when there is competition in the same connector PCo

    Dear all, I'm having a problem with using PCo 2.1.
    I created all the conections in PCo with OSIsoft PI and when I create a TagQuery
    on MII to get the values of a tag is not occur problem in response time.
    But when I have another TagQuery or Transaction, which is collecting data at the same
    time, sharing the same connector, the time response is very slow.
    Using the UDS this problem does not happen.
    So problem is that when there is competition in the same connector, the response time is extremely slow.
    Has anyone experienced this?
    Regards!

    Hi Udayan,
    Thanks again for attention,
    I found this problem on the two MII versions, 12.0 and 12.1.
    The UDS and PCo are on different servers and using different ports too.
    I didn't find errors on netweaver about this, but I'll verify one more time.
    On true, on version 12.0 of MII there is two connections with PI, one PCo and another UDS, that are using the same port. But this does not happen on the MII version 12.1, there I'm using different ports.
    Thank you very much!

  • How to Tune the Transactions/ Z - reports /Progr..of High response time

    Dear friends,
    in <b>ST03</b> work load anlysis menu.... there are some z-reports, transactions, and some programmes are noticed contineously that they are taking the <b>max. response time</b> (and mostly >90%of time is  DB Time ).
    how to tune the above situation ??
    Thank u.

    Siva,
    You can start with some thing like:
    ST04  -> Detail Analysis -> SQL Request (look at top disk reads and buffer get SQL statements)
    For the top SQL statements identified you'd want to look at the explain plan to determine if the SQL statements is:
    1) inefficient
    2) are your DB stats up to date on the tables (note up to date stats does not always means they are the best)
    3) if there are better indexes available, if not would a more suitable index help?
    4) if there are many slow disk reads, is there an I/O issue?
    etc...
    While you're in ST04 make sure your buffers are sized adequately.
    Also make sure your Oracle parameters are set according to this OSS note.
    Note 830576 - Parameter recommendations for Oracle 10g

  • How to obtain the Query Response Time of a query?

    Given the Average Length of Row of tables and the number of rows in each table,
    is there a way we get the query response time of a query involving
    those tables. Query includes joins as well.
    For example, suppose there 3 tables t1, t2, t3. I wish to obtain the
    time it takes for the following query:
    Query
    SELECT t1.col1, t2.col2
    FROM t1, t2, t3
    WHERE t1.col1 = t2.col2
    AND t1.col2 IN ('a', 'c', 'd')
    AND t2.col1 = t3.col2
    AND t2.col1 = t1.col1 (+)
    ORDER BY t1.col1
    Given are:
    Average Row Length of t1 = 200 bytes
    Average Row Length of t2 = 100 bytes
    Average Row Length of t3 = 500 bytes
    No of rows in t1 = 100
    No of rows in t2 = 1000
    No of rows in t3 = 500
    What is required is the 'query response time' for the said query.

    I do not know how to do it myself. But if you are running Oracle 10g, I believe that there is a new tool called: SQL Tuning Advisor which might be able to help.
    Here are some links I found doing a google search, and it looks like it might meet your needs and even give you more information on how to improve your code.
    http://www.databasejournal.com/features/oracle/article.php/3492521
    http://www.databasejournal.com/features/oracle/article.php/3387011
    http://www.oracle.com/technology/obe/obe10gdb/manage/perflab/perflab.htm
    http://www.oracle.com/technology/pub/articles/10gdba/week18_10gdba.html
    http://www.oracle-base.com/articles/10g/AutomaticSQLTuning10g.php
    Have fun reading:
    You can get help from teachers, but you are going to have to learn a lot by yourself, sitting alone in a room ....Dr. Seuss
    Regards
    Tim

  • Query Tuning - Response time Statistics collection

    Our Application is Load tested for a period of 1 hour with peak load.
    For this specific period of time, say thousands of queries gets executed in the database,
    What we need is say for one particular query " select XYZ from ABC" within this span of 1 hour, we need statistics like
    Number of times Executed
    Average Response time
    Maximum response time
    minimum response time
    90th percentile response time ( sorted in ascending order, 90th percentile guy)
    All these statistics are possible if i can get all the response times for that particular query for that period of 1 hour....
    I tried using sql trace and TKPROF but unable to get all these statistics...
    Application uses connection pooling, so connections are taken as and when needed...
    Any thoughts on this?
    Appreciate your help.

    I don't think v$sqlarea can help me out with the exact stats i needed, but certainly it has lot of other stats to take. B/w there is no dictionary view called v$sqlstats.
    There are other applications which share the same database where i am trying to capture for my application, so flushing cache which currently has 30K rows is not feasible solution.
    Any more thoughts on this?

Maybe you are looking for