Performance analysis on interposed functions

Hi,
This is a repost of http://forum.java.sun.com/thread.jspa?threadID=5156578 mistakenly posted in Sun Studio C++ forum.
I'm using the example code given in the following article (and the Linker and Library Guide) to define my own time function as a shared library:
http://developers.sun.com/solaris/articles/time_stamp.html
This works properly. Now I wish to analyze performance of a program using this library using collect and analyzer. Two problems arise:
1. The #pragma init does not seem to work. I have to put a static flag variable and call the init_func myself inside my time() function if the flag is not set.
2. Even after doing this, the dlsym() call returns an error indicating the original time() function could not be found.
Is performance analysis of such interposed functions possible ? The test code only has a bunch of time() calls.
If this can work, I plan to use it to analyze some code (including some third party libraries) that needs the current time to be set to a fixed value in order to ensure identical results.
I'm using Sun Studio 10 on Solaris 8.
Regards,
Paul

Sorry for a delayed response.
I'm not sure if this constitutes a bug in the dynamic linker. Please try setting LD_PRELOAD environment variable instead of LD_FLAGS. It appears that setting both (collect sets LD_PRELOAD) confuses the dynamic linker and only one setting takes effect.
-- Oleg

Similar Messages

  • BOM Performance Analysis using SAT or SE30 t-code?

    Hi,
        Currently I'm doing performance analysis of BOM to find out the bottlenecks. I'm doing the analysis using SAT or SE30 t-code.
    From my understanding, BOM performance depends upon
    1. no. of BOM items
    2. BOM Level
    3. BOM Evolution over time
    Can anyone provide me a hint to find out the bottlenecks in BOM performance. Thanks!!
    Regards,
    Saravana

    I have done the BOM Performance Analysis. Based on the most execution time and top performance consumers, please find my observations…
    Functionally, Performance of BOM execution depends upon
    Number of BOM items available in the BOM
    Number of Assemblies present in the BOM
    BOM Levels
    BOM Evolution over time
    Technically, BOM execution time is purely depends upon
    fetching the data from the SAP tables
    hitting the same table again and again when goes down the level
    loop at the tables / Nested functions used in SAP Program
    Hence, I'm closing this thread.....

  • Performance Analysis Notes

    Hi,
    I want some "Performance Analysis Notes or material"..
    Plz give me some links or send me on my id: [email protected]
    Thanks...

    Hi, this may help you.
    1) Dont use nested select statements
    2) If possible use for all entries in addition
    3) In the where addition make sure you give all the primary key
    4) Use Index for the selection criteria.
    5) You can also use inner joins
    6) You can try to put the data from the first select statement into an Itab and then in order to select the data from the second table use for all entries in.
    7) Use the runtime analysis SE30 and SQL Trace (ST05) to identify the performance and also to identify where the load is heavy, so that you can change the code accordingly
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/5d0db4c9-0e01-0010-b68f-9b1408d5f234
    ABAP performance depends upon various factors and in devicded in three parts:
    1. Database
    2. ABAP
    3. System
    Run Any program using SE30 (performance analys) to improve performance refer to tips and trics section of SE30, Always remember that ABAP perfirmance is improved when there is least load on Database.
    u can get an interactive grap in SE30 regarding this with a file.
    also if u find runtime of parts of codes then use :
    Switch on RTA Dynamically within ABAP Code
    *To turn runtim analysis on within ABAP code insert the following code
    SET RUN TIME ANALYZER ON.
    *To turn runtim analysis off within ABAP code insert the following code
    SET RUN TIME ANALYZER OFF.
    Always check the driver internal tables is not empty, while using FOR ALL ENTRIES
    Avoid for all entries in JOINS
    Try to avoid joins and use FOR ALL ENTRIES.
    Try to restrict the joins to 1 level only ie only for tables
    Avoid using Select *.
    Avoid having multiple Selects from the same table in the same object.
    Try to minimize the number of variables to save memory.
    The sequence of fields in 'where clause' must be as per primary/secondary index ( if any)
    Avoid creation of index as far as possible
    Avoid operators like <>, > , < & like % in where clause conditions
    Avoid select/select single statements in loops.
    Try to use 'binary search' in READ internal table. Ensure table is sorted before using BINARY SEARCH.
    Avoid using aggregate functions (SUM, MAX etc) in selects ( GROUP BY , HAVING,)
    Avoid using ORDER BY in selects
    Avoid Nested Selects
    Avoid Nested Loops of Internal Tables
    Try to use FIELD SYMBOLS.
    Try to avoid into Corresponding Fields of
    Avoid using Select Distinct, Use DELETE ADJACENT
    Check the following Links
    Re: performance tuning
    Re: Performance tuning of program
    http://www.sapgenie.com/abap/performance.htm
    http://www.thespot4sap.com/Articles/SAPABAPPerformanceTuning_PerformanceAnalysisTools.asp
    check the below link
    http://www.sap-img.com/abap/performance-tuning-for-data-selection-statement.htm
    See the following link if it's any help:
    http://www.thespot4sap.com/Articles/SAPABAPPerformanceTuning_PerformanceAnalysisTools.asp
    Check also http://service.sap.com/performance
    and
    books like
    http://www.sap-press.com/product.cfm?account=&product=H951
    http://www.sap-press.com/product.cfm?account=&product=H973
    http://www.sap-img.com/abap/more-than-100-abap-interview-faqs.htm
    http://www.thespot4sap.com/Articles/SAPABAPPerformanceTuning_PerformanceAnalysisTools.asp
    Performance tuning for Data Selection Statement
    http://www.sap-img.com/abap/performance-tuning-for-data-selection-statement.htm
    Debugger
    http://help.sap.com/saphelp_47x200/helpdata/en/c6/617ca9e68c11d2b2ab080009b43351/content.htm
    http://www.cba.nau.edu/haney-j/CIS497/Assignments/Debugging.doc
    http://help.sap.com/saphelp_erp2005/helpdata/en/b3/d322540c3beb4ba53795784eebb680/frameset.htm
    Run Time Analyser
    http://help.sap.com/saphelp_47x200/helpdata/en/c6/617cafe68c11d2b2ab080009b43351/content.htm
    SQL trace
    http://help.sap.com/saphelp_47x200/helpdata/en/d1/801f7c454211d189710000e8322d00/content.htm
    CATT - Computer Aided Testing Too
    http://help.sap.com/saphelp_47x200/helpdata/en/b3/410b37233f7c6fe10000009b38f936/frameset.htm
    Test Workbench
    http://help.sap.com/saphelp_47x200/helpdata/en/a8/157235d0fa8742e10000009b38f889/frameset.htm
    Coverage Analyser
    http://help.sap.com/saphelp_47x200/helpdata/en/c7/af9a79061a11d4b3d4080009b43351/content.htm
    Runtime Monitor
    http://help.sap.com/saphelp_47x200/helpdata/en/b5/fa121cc15911d5993d00508b6b8b11/content.htm
    Memory Inspector
    http://help.sap.com/saphelp_47x200/helpdata/en/a2/e5fc84cc87964cb2c29f584152d74e/content.htm
    ECATT - Extended Computer Aided testing tool.
    http://help.sap.com/saphelp_47x200/helpdata/en/20/e81c3b84e65e7be10000000a11402f/frameset.htm
    Just refer to these links...
    performance
    Performance
    Performance Guide
    performance issues...
    Performance Tuning
    Performance issues
    performance tuning
    performance tuning
    You can go to the transaction SE30 to have the runtime analysis of your program.Also try the transaction SCI , which is SAP Code Inspector.
    1 Always check the driver internal tables is not empty, while using FOR ALL ENTRIES
    2 Avoid for all entries in JOINS
    3 Try to avoid joins and use FOR ALL ENTRIES.
    4 Try to restrict the joins to 1 level only ie only for 2 tables
    5 Avoid using Select *.
    6 Avoid having multiple Selects from the same table in the same object.
    7 Try to minimize the number of variables to save memory.
    8 The sequence of fields in 'where clause' must be as per primary/secondary index ( if any)
    9 Avoid creation of index as far as possible
    10 Avoid operators like <>, > , < & like % in where clause conditions
    11 Avoid select/select single statements in loops.
    12 Try to use 'binary search' in READ internal table. Ensure table is sorted before using BINARY SEARCH.
    13 Avoid using aggregate functions (SUM, MAX etc) in selects ( GROUP BY , HAVING,)
    14 Avoid using ORDER BY in selects
    15 Avoid Nested Selects
    16 Avoid Nested Loops of Internal Tables
    17 Try to use FIELD SYMBOLS.
    18 Try to avoid into Corresponding Fields of
    19 Avoid using Select Distinct, Use DELETE ADJACENT.
    Regards
    Madhu

  • How to do Performance analysis of any ABAP report

    Hello Friends,
                         I have to do the performance analysis of ABAP report, it takes time to execute. I am  not aware about, how to do performance analysis? I know only we have check index, do ST04 and ST05 analysis. but how to do these things  that i don't know.
                        Can anyone please guide me in this regards.
    Thanks

    HI,
    Please Check this,
    System Trace: Transaction ST01 lets you do various levels of system trace such as authorization checks, SQL traces, table/buffer trace etc. It is a general Basis tool but can be leveraged for BW.
    Workload Analysis: You use transaction code ST03
    Database Performance Analysis: Transaction ST04 gives you all that you need to know about whatu2019s happening at the database level.
    Performance Analysis: Transaction ST05 enables you to do performance traces in different are as namely SQL trace, Enqueue trace, RFC trace and buffer trace.
    ABAP Runtime Analysis Tool: Use transaction SE30 to do a runtime analysis of a transaction, program or function module. It is a very helpful tool if you know the program or routine that you suspect is causing a performance bottleneck.
    Rgds
    Sabu

  • Mailbox Migration Performance Analysis

    I'm attempting to follow the steps per this website
    http://ppe.blogs.technet.com/b/exchange/archive/2014/03/24/mailbox-migration-performance-analysis.aspx in Exchange PowerShell and Step 4. is returning the following.
    ProcessStats : The term 'ProcessStats' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again.
    At line:1 char:1
    + ProcessStats -stats $stats -name ProcessedStats1
    + ~~~~~~~~~~~~
        + CategoryInfo          : ObjectNotFound: (ProcessStats:String) [], CommandNotFoundException
        + FullyQualifiedErrorId : CommandNotFoundException
    Any suggestions?

    No luck with that format for step 1.
    PS D:\temp> ..\AnalyzeMoveRequestStats.ps1
    ..\AnalyzeMoveRequestStats.ps1 : The term '..\AnalyzeMoveRequestStats.ps1' is not recognized as the name of a cmdlet,
    function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the
    path is correct and try again.
    At line:1 char:1
    + ..\AnalyzeMoveRequestStats.ps1
    + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        + CategoryInfo          : ObjectNotFound: (..\AnalyzeMoveRequestStats.ps1:String) [], CommandNotFoundException
        + FullyQualifiedErrorId : CommandNotFoundException
    PS D:\temp>

  • SoD Analysis with Portal functions

    HI all:
    Can anyone point me to a place where i can find information on proces of creating SoD rules for Portal functions?  We want to perform SoD analysis on Portal functions which do not necessarily have transaction codes.
    We have iviews that we have restricted by role placement, very basic.
    However, we would like to be able to analyse SoD based on Webdynpro execution.  Is this possible and if so, how can we do this?
    Thanks,
    Margaret

    Hi Margaret,
    the Portal SoD Analysis is either based on UME actions or iView names, that's it.
    Here you will find a guide:
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/502a14db-6261-2c10-22b5-95117ab0e5ed
    Best,
    Frank

  • What is performance analysis

    hi gurus
    can any one suggest me
    what is performance analysis
    for which purpose it is used
    how to use that one ...
    thank you
    regards
    kals.

    hai
      kalyan
           In SE30, u can see perform analysis. it consists of two parts.
      1. Reording performance data
      2. Analyzing the performance data
    1.  Reording performance data :
             The system measures the transaction, program, or  procedure, and writes the result to a performance data file.
    The following steps to be followed.
    Procedure:
    From the initial screen of the runtime analysis transaction:
    1. Enter a transaction code, a program name, or the name of a function module.
    2. In the Comment line, enter a short description of the performance data file.
    3. Restrict the analysis to the options that you need. Refer also to Selecting the Measurement Options.
    4. To start the measurement, choose Measure runtime.
    5. Run the transaction, program, or function module as normal.
    6. Return to the initial screen of the runtime analysis transaction.
    To do so, either leave the transaction, program, or function module as normal, or start the runtime analysis again.
      2. Analyzing the performance data
    The analysis overview contains general information about the performance of your program.
    It displays the overall runtime of the program in microseconds, both in figures and as a bar chart, as well as telling you how much memory each particular modularization unit required.
    The bar chart displays the total runtime required for
    The ABAP interpreter
    The database
    System statements.
    The time taken by above three types statements in ur program / object  is shown in that bargraph.
    if useful, reward points.
    Thank you,
    G.V.K.Prasad

  • Vendor performance analysis report

    Hi all,
    requirement to create an ALV report for vendor performance analysis...  wat r d tables n feilds which are to be used fr this... any help is appreciated..

    Refer the links -
    vendor performance report !!!
    vendor performance report
    vendor performance report
    I need standard vendor performance report

  • Performance issue for this function-module(HR_TIM_REPORT_ABSENCE_DATA)

    Hi Friends
    I am having performance issue for this function-module(HR_TIM_REPORT_ABSENCE_DATA) and one my client got over 8 thousend employees . This function-module taking forever to read the data. is there any other function-module to read the absences data IT2001 .
    I did use like this .if i take out this F.M 'HR_TIM_REPORT_ABSENCE_DATA_INI' its not working other Function-module.please Suggest me .
    call function 'HR_TIM_REPORT_ABSENCE_DATA_INI'
    exporting "Publishing to global memory
    option_string = option_s "string of sel org fields
    trig_string = trig_s "string of req data
    alemp_flag = sw_alemp "all employee req
    infot_flag = space "split per IT neccessary
    sel_modus = sw_apa
    importing
    org_num = fdpos_lines "number of sel org fields
    tables
    fieldtab = fdtab "all org fields
    field_sel = fieldnametab_m. "sel org fields
    To Read all infotypes from Absences type.
    RP_READ_ALL_TIME_ITY PN-BEGDA PN-ENDDA.
    central function unit to provide internal tables: abse orgs empl
    call function 'HR_TIM_REPORT_ABSENCE_DATA'
    exporting
    pernr = pernr-pernr
    begda = pn-begda
    endda = pn-endda
    IMPORTING
    SUBRC = SUBRC_RTA
    tables
    absences = absences_01
    org_fields = orgs
    emp_fields = empl
    REFTAB =
    APLTAB =
    awart_sel_p = awart_s[]
    awart_sel_a = awart_s[]
    abstp_sel = abstp_s[]
    i0000 = p0000
    i0001 = p0001
    i0002 = p0002
    i0007 = p0007
    i2001 = p2001
    i2002 = p2002
    i2003 = p2003.
    Thanks & Regards
    Reddy

    guessing will not help you much, check with SE30 to get a better insight
    SE30
    The ABAP Runtime Trace (SE30) -  Quick and Easy
    what is the total time, what are the Top 10 in the hitlist.
    Siegfried

  • NAM Report Analyze -- Wan Optimization -- Application Performance Analysis Report

    Hello,
    Transaction Time (Client Exprerience) report does not show optimized traffic. For this graph what we expect is to show side by side the transaction time of optimized and non optimized traffic side by side to give information about the baseline and optimized application performance.
    May it be that the optimization should be disabled for some time for the Wan Sites, and after some time manually reenabled in order to compare these values?  This should be difficult for large deployments since manually disabling and reenabling for performance analysis time taking task.
    Best Regards,

    Mohammed,
    It is common, in many of the Cisco Express 8.5 environments we have looked at, for the Total Incoming Calls given on a Traffic Analysis report to be a higher number than an Application Report.
    The Traffic Analysis Report counts every unique sessionID (unique call) that is inbound (contact type of 1).  The Application Reports do a similar thing but qualify (filter) only the records that have an application assigned.
    There are simply times where inbound calls have been directed to an "agent" without having an applicaiton assigned.
    The best thing the reporting user can do is to run a query on his or her database such as: 
         select * from ContactCallDetail where contactType=1and startDateTime > '2012-11-16 10:00:00' and startDateTime < '2012-11-16 11:00:00';
    Usually when an application is not assigned to the record in the ContactCallDetail table it is because the destination type is equal to 1, which is an 'Agent' instead of a 'route point'.
    So if you modify your select statement to filter by destinationType, you can quickly find the records that don't have the application assigned. 
    Example:  select * from ContactCallDetail where contactType = 1 and destinationType = 1 and startDateTime > '2012-11-16 10:00:00 and startDateTime < 2012-11-16 11:00:00';
    When you look at these records, you will see the agent that took the call from the destinationID field.  The number in that field should match up with the field called 'resourceID' in a table called 'resouce';
         Example:  select * from resource where resouce=6011; where 6011 was the number you found in the destinationID field.
    If there is still confusion about the source of the call - then talk to that agent and find out what is was.
    Good Luck and let me know if you need further help.
    Ron Reif
    [email protected]

  • Import Performance Analysis report

    Hi,
    I ran the import analysis report in FDM and it tells me the time taken by each step of import process. My question is why does it take time for "Memo Reassign Time". I am not doing anything with Memo. Is there a way to avoid the time it takes to reassign memo.
    Thanks, AJ

    Import Performance Analysis
    File Archive Time 1
    Memo Reassign Time 53
    Total Import Process 254 Seconds
    Total Map Process 65 Seconds
    Almost 1/5th of the total time is taken by Memo Reassign Step for nothing. Is there a way to bring it to 0 seconds.

  • Performance analysis and monitoring of a Forteapplication

    Hello,
    It would be good if one could do some performance analysis and monitoring of
    a Forte application at production time.
    By performance analysis I mean measure the time some selected methods take
    to return. In a CS application such a method would be some selected method
    of a key remote service representative of the application's activity.
    One would like to mesure the min, max, average time, and also give a
    threshhold which when reached will automatically generate an alarm or some
    pre-defined processing.
    The most powerful way would be to use a SNMP application through a Forte -
    SNMP gateway. (see the G. Puterbaugh paper "Building a Forte-SNMP gateway"
    in the 96' Forte Forum proceedings).
    But before going that far some simple means accessible through EConsole
    would already be great.
    A collegue of mine when to that Forum and reported that Puterbaugh said that
    such an agent is currently missing, but that its implementation is not
    difficult.
    I looked at all the agents and their instruments. I came to the following
    conclusions :
    1) instrumented data are available at the granularity level of a partition,
    not at a smaller granularity. For example, the DistObjectMgr agent gives you
    very useful information : the number of events (sent/received) and the
    number of (remote) methods (called/invoked), but this for the entire
    partition. Thus it prevents making tuned observations (unless you partition
    in a special way your application to put in a dedicated partition the thing
    you want to observe and only this thing).
    2) there is no instrumented data related to processing time.
    This leads me to the point that no information observed by the standard
    agents help me figuring out my performances. Thus I have to add at
    development time some lines of code to the methods I potentially want to
    observe later at production time to generate the appropriate information a
    custom agent will then display (process) with the appropriate instruments.
    Does someone share this position ?
    Has someone implemented such an agent and assotiated means ?
    PS: I will probably implement my own one if no other way around.
    best regards,
    Pierre Gelli
    ADP GSI
    Payroll and Human Resources Management
    72-78, Grande Rue, F-92310 SEVRES
    phone : +33 1 41 14 86 42 (direct) +33 1 41 14 85 00 (reception desk)
    fax : +33 1 41 14 85 99

    Hello,
    It would be good if one could do some performance analysis and monitoring of
    a Forte application at production time.
    By performance analysis I mean measure the time some selected methods take
    to return. In a CS application such a method would be some selected method
    of a key remote service representative of the application's activity.
    One would like to mesure the min, max, average time, and also give a
    threshhold which when reached will automatically generate an alarm or some
    pre-defined processing.
    The most powerful way would be to use a SNMP application through a Forte -
    SNMP gateway. (see the G. Puterbaugh paper "Building a Forte-SNMP gateway"
    in the 96' Forte Forum proceedings).
    But before going that far some simple means accessible through EConsole
    would already be great.
    A collegue of mine when to that Forum and reported that Puterbaugh said that
    such an agent is currently missing, but that its implementation is not
    difficult.
    I looked at all the agents and their instruments. I came to the following
    conclusions :
    1) instrumented data are available at the granularity level of a partition,
    not at a smaller granularity. For example, the DistObjectMgr agent gives you
    very useful information : the number of events (sent/received) and the
    number of (remote) methods (called/invoked), but this for the entire
    partition. Thus it prevents making tuned observations (unless you partition
    in a special way your application to put in a dedicated partition the thing
    you want to observe and only this thing).
    2) there is no instrumented data related to processing time.
    This leads me to the point that no information observed by the standard
    agents help me figuring out my performances. Thus I have to add at
    development time some lines of code to the methods I potentially want to
    observe later at production time to generate the appropriate information a
    custom agent will then display (process) with the appropriate instruments.
    Does someone share this position ?
    Has someone implemented such an agent and assotiated means ?
    PS: I will probably implement my own one if no other way around.
    best regards,
    Pierre Gelli
    ADP GSI
    Payroll and Human Resources Management
    72-78, Grande Rue, F-92310 SEVRES
    phone : +33 1 41 14 86 42 (direct) +33 1 41 14 85 00 (reception desk)
    fax : +33 1 41 14 85 99

  • [request] Intel VTune Performance Analyser

    Hello, I am wondering if someone have made a pkgbuild for the Intel VTune Performance Analyser: http://www.intel.com/cd/software/produc … 239143.htm. At the moment I do not have the skills or time to make a PKGBUILD, so I am hoping that someone already have done it. I am planning too learn the PKGBUILD system, so if do not have any success with request, I may prowide a PKGBUILD, in a distant future!

    OK.
    ThinkPad Edge E530 (3259-AD9)
    i5-3210M(2.5Ghz), 6GB RAM, 750GB 5400rpm HD, 15.6in 1366x768 LCD, Intel HD Graphics, 802.11bgn wireless, 1Gb Ethernet, UltraNav, Fingerprint reader, Camera, 6c Li-Ion, Win7 Home Premium 64
    First, in case you don't have it, here's your complete driver matrix: http://support.lenovo.com/en_US/research/hints-or-tips/detail.page?&DocID=HT073274
    From there, please confirm that you have installed the most recent release of the Intel HD Graphics drivers:
    http://support.lenovo.com/en_US/research/hints-or-tips/detail.page?&DocID=HT073274 from November 2013.
    Regards.
    English Community   Deutsche Community   Comunidad en Español   Русскоязычное Сообщество
    Community Resources: Participation Rules • Images in posts • Search (Advanced) • Private Messaging
    PM requests for individual support are not answered. If a post solves your issue, please mark it so.
    X1C3 Helix X220 X301 X200T T61p T60p Y3P • T520 T420 T510 T400 R400 T61 Y2P Y13
    I am not a Lenovo employee.

  • Performance Analysis in Oracle Enterprise Manger console..

    Hi,
    Following is a findings of performance analysis section of Oracle Enterprise Manager.
    SQL statements consuming significant database time were found.
    ======
    While clicking on above link system opens another page and it should following details.
    Action Run SQL Tuning Advisor on the SQL statement with SQL_ID "5z50bh2wj1svt".
    SQL Text SELECT /*+FIRST_ROWS INDEX(a ttdpur041730$idx4)*/ a.t$amta,a.t$bqua,a.t$btsp,a.t...
    SQL ID 5z50bh2wj1svt
    Rationale SQL statement with SQL_ID "5z50bh2wj1svt" was executed 732 times and had an average elapsed time of 2.9 seconds.
    The impact of this finding is 77% ...
    THis is an erp apps where we don't have any source code ...
    Any suggestions to reduce impact form 77%
    SSM

    I don't think that the system tablespace datafile is 92% being filled is a problem,but his what you can do:
    set oracle_sid=<your SID>
    I take it the path includes the correct <oracle_home>\bin
    sqlplus / as sysdba
    alter database datafile 'E:\path_to\system.dbf' AUTOEXTEND ON NEXT 1M MAXSIZE 2000M; replace 'E:\path_to by the correct disk and path.
    Changing a datafile size (resize) or changing a datafile so it is able to autoexextend can also be done by the EM itself.
    The controlfile can grow over time but I haven't seen any problems with it's size.
    Sorry, I can't explain the backup issue.
    Eric

  • Tuxedo Performance Analysis - Query - tried txrpt not receiving report

    Hi All,
    We have currently migrated from Solaris-8 - Tuxedo 8 environment to Solaris 10 - Tuxedo 11gr1 Staging environment.
    All the application binaries were running fine.
    However we conducted a performance testing for 500 users we could see the performance coming down, when compared to our old performance report of
    Solaris 8 - tux8 environment.
    However we doubt it may be some issue with the network or some other server via which the requests are hitting or responses leaving our system.
    Can you suggest any utility or its usage by which we can check that our current Tuxedo version is working fine than the older one?
    Like how much time tuxedo is taking now to process,the throughputs.
    can you please suggest some way to monitor or generate some sort of report to compare tuxedo 8 performance and Tuxedo 11g R1 performance for same application.
    tried the txrpt
    I5_RQMG_0002_00
    SRVGRP=AQUREQGRP
    SRVID=2000
    CLOPT="-r -s PSY_MG_0002_00:I5_RQMG_0002_00 -o logs/slog/I5_RQMG_0002_00.out -e logs/slog/I5_RQMG_0002_00.err"
    MIN=1 MAX=1
    RQADDR="I5_RQMG_0012_00Q2"
    REPLYQ=Y
    for this above server :
    -rwxrwxrwx 1 tuxedo tuxedo 71022518 May 29 14:56 I5_RQMG_0002_00.err
    I get the above file generated.
    Executed the command:
    tuxconsolas1:/apps/tuxapps/intserv/logs/slog>txrpt -nPSY_MG_0002_00 -d05/29 -s02:00 -e14:00 I5_RQMG_0002_00.err > txrpt.out
    I am just receiving :
    tuxconsolas1:/apps/tuxapps/intserv/logs/slog>less txrpt.out
    START AFTER: Tue May 29 02:00:00 2012
    END BEFORE: Tue May 29 14:00:00 2012
    in the report and the process hangs for me.
    Am i doing any error?
    Edited by: 835542 on May 29, 2012 10:27 AM

    Hi,
    If you want detailed performance analysis of your Tuxedo 11gR1 application, you can use TSAM (Tuxedo System and Application Monitor) to get very detailed performance information including time spent on network links, time spent waiting in IPC queues, etc. Unfortunately you won't be able to use the same tool for your Tuxedo 8.0 application as it is too old for TSAM to support.
    Regards,
    Todd Little
    Oracle Tuxedo Chief Architect

Maybe you are looking for

  • JNI / Garbage Collection question

    I have a C++ library I am calling into from our application. The question I have is I want the library to return a jintarray or a jstringarray (not sure which yet). If they create that array in the function and then return it to me as the return type

  • CS3 randomly indexing mpeg2

    Hello, Time being of the essence, we mostly work with native mpeg2 files. We are using 2 workstations: - 1 running PPro2 mainly for editing - 1 running CS3 mainly for rendering This system has been working like a charm until we changed the PC running

  • Mail embeds my images

    I have posted about this before, but thought i try again now that im on leopard and im STILL having this truly annoying problem. I am sending an email. I attach an image (jpg), make sure to click the windows friendly button since that is who i am sen

  • How to enable apple ID from it being disabled?

    Hello, my Apple ID is disabled from entering my password incorrect to many times. I have reset my password 5 times and my account is still disabled. How can I enable it?

  • I have a macbook pro core 2 duo (late 2006) that i need to install a OS on (It doesn't have one at all).

    I spoke with an Apple worker and they said snow leopard would work but he didnt know if the upgrade disk would work or not. I need help with this. I will need 5 copies so i hope i can spend 20 or 30 instead of 600. Thanks. Email is ****** <Edited by