System performance issue due to multiple submission of a job

For month end recoincilation users run few critical reports which are quite resource consuming.
In order to control this, want to restrict the usage of such  reports. For example if one session is active (Foreground or background) user couldn't submit another job or gets a pop up with an error mesage.
Searched SDN for this and couldn't find much.
anya

Anya since this is the ABAP forum and I happen to be an ABAP programmer I can give you an ABAP solution. This would involve changes to the code in all these reports so you would need the help of an ABAP programmer if you are not one.
a) Create a Z table containing 3 fields
1) Client of type MANDT  (Primary key field)
2)Program Name  of type PROGRAM_ID (Primary key field)
3) User name of type XUBNAME (Regular field).
b) Create table maintenance for this internal table.
c) Create one record for each of the programs you want to regulate. You only need to enter the program name initially and leave the user name blank.
d) In all the ABAP programs make the following change.
1) Under the START-OF-SELECTION event of the report lock the use function module ENQUEUE_E_TABLEE to lock the record in the Ztable for the program being executed. Look at the sample code below.
TABLES: <ztable>.
data: w_message(100) TYPE c,
        w_locked(1) TYPE c.
CALL FUNCTION 'ENQUEUE_E_TABLEE'
  EXPORTING
    tabname              = <ztable name>
    varkey               = <concatenation of mandt and sy-cprog>
  EXCEPTIONS
   foreign_lock         = 1
   system_failure       = 2
   OTHERS               = 3.
IF sy-subrc EQ 1.
  SELECT SINGLE *
    FROM <ztable>
    INTO <ztable>
    WHERE program FIELD name EQ sy-cprog.
  IF sy-subrc EQ 0.
    CONCATENATE 'Program'
                sy-cprog
                'is currently being used by'
                <ztable>-<user name>
      INTO w_message SEPARATED BY space.
    MESSAGE e208(00) WITH w_message.
  ENDIF.
ELSEIF sy-subrc EQ 0.
  w_locked = 'X'.
  SELECT SINGLE *
    FROM <ztable>
    INTO <ztable>
    WHERE program FIELD name EQ sy-cprog.
  IF sy-subrc EQ 0.
    <ztable>-user name> = sy-uname.
    MODIFY <ztable> FROM <ztable>.
  ENDIF.
ENDIF.
e) At the event END-OF-SELECTION (at the end of the program) use function module DEQUEUE_E_TABLEE to unlock the record. Look at sample code below.
CASE w_locked.
  WHEN 'X'.
    CALL FUNCTION 'DEQUEUE_E_TABLEE'
      EXPORTING
        tabname          = <ztable name>
        varkey             = <concatenation of mandt and sy-cprog>.
ENDCASE.
This code is designed to allow just one user or job to run the program at a time. The second person will be issued an error message informing him/her that the program is being used by <user name>.

Similar Messages

  • Possible solution to wireless/bluetooth/and system performance issues.

    I had a TON of issues with Leopard after my Tiger Install, network, bluetooth, video playback, iTunes playback, system performance and freezing just to name a few. I tried everything (reseting PRAM, etc.) to fix the problem and have done a variety if installations half a dozen times. Yesterday I took out my 3rd party memory I had installed (Corsair if anyone cares) and did an erase and install. I've had no problems so far and everything seems to be running great. I'll keep everyone posted on any progress. You may want to try this if you think you may have the same issue.

    Hi
    I may have the solution to slow performance when using multi core processing. I was testing out realflow 2013 and noticed simulations were talking a long time to finish when using maximum threads 32. I set the threads to 8 and the simumlation was twice as fast which did not make any sense to me. The simulation was taking 1min 24 secs to finish using 32 threads and 45sec using 8 threads. I realized their has to be a problem here. So I went into my BIOS and disabled Hyper Threading. So now instead of having 32 threads I only have 16 threads. When i went back to try the same simulation again this time the 16 threads was only taking 24secs to finish. and the 8 threads was taking 35secs to finish. This made a lot more sence to me.
    Anyway so once I realized that turning off Hyper Threading fixed my simulation problem maybe it might be the cause of my slow performance in AE using multi processing. I tried AE and now multi processing is working great very fast now and stable. MP renders start right away now instead of getting stuck on frame 2 or saying initializing background processes and never ever doing anything. Before I had to keep disabling MP and re-enabling it in AE to get it to start working again and still it wasen't very stable
    If anybody is having issues with slow performance from their Multi processor systems try disabling Hyper Threading in the BIOS it might just fix your problem.
    Hope this helps 

  • Cisco ASR 1002- performance issue due to access list

    Hi,
    We are planning to implement inbound access-list to block subnets from particular country. Since the subnets are not contiguous, we have about 16000 lines of acl entries.
    I want to know, would there be any performance or latency issues after applying 16k lines of acl?
    Is there a good document where I can read more about ACL limitations and performance issues on ASR.
    This is for ASR1002, running IOS-XE 15.3(1)S1.
    Thanks

    Disclaimer
    The Author of this posting offers the information contained within this posting without consideration and with the reader's understanding that there's no implied or expressed suitability or fitness for any purpose. Information provided is for informational purposes only and should not be construed as rendering professional advice of any kind. Usage of this posting's information is solely at reader's own risk.
    Liability Disclaimer
    In no event shall Author be liable for any damages whatsoever (including, without limitation, damages for loss of use, data or profit) arising out of the use or inability to use the posting's information even if Author has been advised of the possibility of such damage.
    Posting
    Sorry, I don't know the answer to your questions, but I'm writing to mention a 7200 feature, that if supported on the ASR, might help in your situation.  See http://www.cisco.com/c/en/us/support/docs/security/ios-firewall/23602-confaccesslists.html#turbo

  • Avoiding performance issue due to loop within loop on internal tables

    Hi Experts,
                    I have a requirement where in i want to check whether each of the programs stored in one internal table are called from any of the programs stored in another internal table. In this case i am looping on two internal tables (Loop within a loop) which is causing a major performance issue. Program is running very very slow.
    Can any one advise how to resolve this performance issue so that program runs faster.
    Thanks in advance.
    Regards,
    Chetan.

    Forget the parallel cursur stuff, it is much to complicated for general usage and helps nearly nothing. I will publish a blog in the next days where this is shown in detail.
    Loop on loop is no problem if the inner table is a hashed or sorted table.
    If it must be a standard table, then you must make a bit more effort and faciliate a binary search (read binary search / loop from index exit)
    see here the exact coding Measurements on internal tables: Reads and Loops:
    /people/siegfried.boes/blog/2007/09/12/runtimes-of-reads-and-loops-on-internal-tables
    And don't forget, the other table must not be sorted, the loop reaches anyway every line. The parallel cursor requires both tables to be sorted. The additional sort
    consumes nearly the whole advantage of the parallel cursor compared to the simple but good loop in loop solutions.
    Siegfried

  • Performance issue due to RFC calls. (R/3 to R/3 system)

    Hi,
    My application face serious performance problem because of RFC calls (R/3 system to R/3 system).
    1)is there any transaction code for doing performance analysis on RFC calls in R/3 system?
    2)How far large volume of Data Transfer (mainly due to internal tables) in a RFC call affects the performance? is there any limit for data transfer size at a time in a RFC call? if so how to calculate for best performance?
    Thanks and regards,
    Prakash.

    Hi Andreas,
    Suppose an RFC enabled Fm having an internal table as importing parameter.During execution of this Fm with 4000 enteries for the internal table results in performance degrading.is there any way to improve the performance?.
    Thanks and regards,
    Prakash.

  • UoM issue due to multiple backend systems

    Hi All,
    We have multiple backend systems hooked up with the SRM system. One issue that we have is that many of the UoMs are different in the backend systems. For example, for THOUSANDS, we have TH and THO in different systems. As part of optimization we intend to include only 1 entry per UoM in SRM box.
    The problem is if we put TH as the unit, then if it is chosen by a user linked to the backend system that does not have this Commercial unit, then the Shopping Cart fails when trying to create a follow-on document.
    Any advice how we can get this resolved will be greatly appreciated.
    Best Regards
    Gaurav

    Hi Gaurav,
    Common problem when using multiple back-ends, which can be resolved rather easily.  The best way to solve this is by a business change. That isu2026.  synchronize the use of the UoMu2019s throughout the organization.  This would be the most elegant solution. All back-ends will then have the same valueu2019s in their respective CUNI tables.
    Off course the feasibility of this will depend on your organizational structure and the hitting the u2018right stringsu2019 in order to get every one to understand the importance of this.
    As a back-up an other solution would be to create a Z-table which will translate the UoM used in the SRM system to the corresponding UoM in the respective back-endsu2026.. The table will have three columns.  u2018SRM UoMu2019 u2013 u2018Back-end Systemu2019 u2013 Back-end UoM.  Depending on the implementation scenario U use, you can have the interface to the back-end carry out the translation. For this you can use the available Badiu2019s: BBP_CREATE_BE_PO_NEW, BBP_CREATE_BE_RQ_NEW, BBP_CREATE_BE_RS_NEW
    Downside of this solution will be the extra maintenance effort for the Z-table. Best choice in my perspective would be the business change.
    Kind Regards,
    Skander

  • SAP MaxDB 7.7.07.16 - performance issue due to IOWait(R) task

    We are running MaxDB 7.7.07.16 unicode in a non SAP environment (is basis for a web application).
    Database Server has 4 Cpu's, 8 Gb RAM, operating system Debian-Linux 6.0.
    We have some performance problems in not regular periods of time.
    The performance problem occurs for a time frame between 3 and 15 minutes and blocks more or less the complete system.
    After a long phase of checking and evaluating possible issues, we think that we are very near to the issue.
    We identified the database tasks
    - TblPrfC  - Prefetch Table Coordinator:
    - PrefPag - Prefetch Pages
    having the current state "IO Wait (R)".
    As long these tasks are in state "IO Wait (R)" the users (at the web application) gets awful slow respond times on their queries.
    On 9th of April 2011 we changed the DB parameters:
    - ReadAheadTableThreshold         old 0 => new 128
    But today we got the same performance problem again.
    If you have any hints or tips, you're welcome!
    Best regards
       Hannes

    Hello Hannes,
    ok brief but hopefully usefull answer to your post:
    - if the database I/O generated by SQL statements alone brings the system to a grinding halt, then your storage system is undersized. How should it be able to handle additional I/O, e.g. for backups if it isn't capable to handle the standard I/O?
    - the prefetching tasks are used to speed up large scan operations. So, the next thing to do here is to find statements that perform table scans and check whether or not these can be tuned to KEY or Index accesses
    - As a workaround, I would propose to increase the value of ReadAheadTableThreshold to, say, 500, so that the prefetching is done only for really large scans.
    regards,
    Lars

  • Performance issue due to column formula and filters

    Hi,
    I am facing strange issue with performance for my OBIEE reports. I have two sets of reports Static and Dynamic. Both runs against same tables. The only difference between these reports is that the Static reports would run against all the data for given aggregation level e.g. Year, Month, Date and so on. Where as for Dynamic one I have range prompts to filter data. Other difference is that I have a column formula for one of the column in the Dynamic report, which is nothing but Go URL to show another page with certain parameters.
    The static report takes around 14-15 Seconds where as the Dynamic one takes around 3.5 min. The amount of data and range is same here. From the logs I could see that for the Static reports, i.e. reports without filters it applys group by at SQL level where as it is not doing so for the dynamic one. Is this expected ?
    Second issue is, even if I say remove the filters and just have report with column formula in one and no formula in other there is significant time difference in the processing at Presentation service layer. Again this is taken from the log. it takes 8 second to get data from DB but shows almost 218 Seconds as response time at Presentation layer.
    Below are conceptual details about table and reports -
    Table 1 (It is date dimension) : Date_Dim
    DateCode Date
    Day Number
    MonthCode Varchar2
    YearCode Varchar2
    Table 2 (It is aggregate table at year level) : Year_Aggr
    DateCode Date (FK to Table1 above)
    Measure1
    Measure2
    Measure3
    Measure4
    Measure5
    Report 1
    Date_Dim.YearCode | Year_Aggr.Measure1 | Year_Aggr.Measure2 | Year_Aggr.Measure3 | Year_Aggr.Measure4
    Report 2
    Dashboard Filter : Dimension1 | Dimension2 | Year Start | Year End |
    Date_Dim.YearCode | Year_Aggr.Measure1 | Year_Aggr.Measure2 | Year_Aggr.Measure3 | Year_Aggr.Measure4
    Column formula for Date_Dim.YearCode is something like :
    '<a href="saw.dll?Dashboard&PortalPath=somepath and parameters  target=_self>'  || Date Dim"."YearCode" || '</a">'
    Filters :
    Dimension1 is prompted...
    Dimension2 is prompted...
    cast("Date Dim"."YearCode" as Int) is greater than or equal to @{Start_Year}
    cast("Date Dim"."YearCode" as Int) is greater than or equal to @{End_Year}
    Note : I need to apply cast to int as column is varchar2, legacy problem.+
    How can I fix this? Am I missing something? In the result of report2 the DB SQL doesn't show the year in where thought it is displayed in the logical sql.
    Let me know if anybody had faced this and have fixed. Or suggetion to make changes to fix this.
    Thannks,
    Ritesh</a>

    Hi Ritesh,
    I think you right about the root cause of your problem. The first request does the group by in the database which returns fewer records to the BI Server for processing. The second request does not do the group by and sends significantly more records back to the BI server forcing it to do the group by. Compound that with the fact that pivot table views are relatively expensive computationally and that explains the difference between the execution times.
    Assuming that the execution time of the first report is satisfactory, I would recommend you try to experiment with a few settings to see if you can get the second report to do the group by in the database.
    Are the two filters identical except for the following conditions?
    cast("Date Dim"."YearCode" as Int) is greater than or equal to @{Start_Year}
    cast("Date Dim"."YearCode" as Int) is greater than or equal to @{End_Year}
    Best regards,
    -Joe

  • Performance issue due to localization code in SSRS 2008

    The reports I am working on consist lot of data and all the customers use it frequently.
    Report title and columns are localized by expressions .
    This takes long time for report rendering and exporting to csv. When I test without the localization code it doesn't take that long .
    Can someone help me to optimize the report. SSRS 2008 R2
    Archana

    Hi Archana,
    In Reporting Services, the total time to generate a reporting server report (RDL) can be divided into 3 elements:
    Total time = (TimeDataRetrieval) + (TimeProcessing) + (TimeRendering)
    TimeRendering means the number of milliseconds spent after the Rendering Object Model is exposed to the rendering extension. It includes the Time spent in on-demand expression evaluations (e.g. TextBox.Value, Style.*). So it is make sense that the report
    with localization code takes longer time.
    Besides, we can improve the report performance form other aspects. For example, we can add filter, sorting and aggregation in dataset query, because filter, sort, and aggregation is more efficient on the data source than during report processing. For more
    details about report performance, please refer to the following article:
    http://technet.microsoft.com/en-us/library/bb522806(v=sql.105).aspx
    If there are any other questions, please feel free to ask.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Performance Issues due to Loading of ADF/JClient View Objects

    Hi,
    I developed a two-tier ADF/JClient application.
    When the ADF/JClient application opens for the first time, there's a big delay
    However opening it subsequently, takes less than half of the time to open the screen.
    I think it might the xml files of different view objects which the framework loads for the first time, and thus takes a lot of time. And after that, the time to open the application reduces.
    We have around 200 view objects in our application. Can this be the reason?
    If yes, then can we load the view objects before ahead, or can we just load them at the very time they are requested ?
    Thanks,
    Makrand Pare

    Hi Makrand,
    Check this out:
    "Limiting Fetching of Business Components Attributes in ADF Swing
    http://www.oracle.com/webapps/online-help/jdeveloper/10.1.3/state/content/navId.4/navSetId._/vtTopicFile.developing_jclient_applications%7Cjui_plimitingfetching%7Ehtml/
    Note: In local mode deployment (the client and Business Components reside in the same VM), the fetching of attribute properties is not supported.
    You can optimize startup time for a Business Components client application and the remotely deployed Business Components by specifying the list of view object attributes that your client uses. If you create a project without the metadata, by coding to the API, you will want to add fetchAttributeProperties() to the bootstrap code of the client forms with a list of only the attributes used by the form. Without this method call, your client form would fetch all control hint properties (including the attributes format and label for example) for all the attributes of the named view objects in the application module, in a single network roundtrip.
    For example, when you do not intend to use all the attributes of the ADF Swing form's bound view object, with the fetchAttributeProperties() method, your ADF Swing form fetches only the information required to layout your forms, while ignoring the attributes you do not require.
    Note: In local mode deployment (the client and Business Components reside in the same VM), the fetching of attribute pro
    Calling fetchAttributeProperties() will prevent property methods such as getFormat() or getLabel() from being called on the Business Components attribute definition whenever the form is created."
    Vlad

  • Report Performance Issue - Activity

    Hi gurus,
    I'm developing an Activity report using Transactional database (Online real time object).
    the purpose of the report is to list down all contacts related activities and activities NOT related to Contact by activity owner (user id).
    In order to fullfill that requirment I've created 2 report
    1) All Activities related to Contact -- Report A
    pull in Acitivity ID , Activity Type, Status, Contact ID
    2) All Activities not related to Contact UNION All Activities related to Contact (Base report) -- Report B
    to get the list of activities not related to contact i'm using Advanced filter based on result of another request which is I think is the part that slow down the query.
    <Activity ID not equal to any Activity ID in Report B>
    Anyone encountered performance issue due to the advanced filter in analytic before?
    any input is really appriciated
    Thanks in advanced,
    Fina

    Fina,
    Union is always the last option. If you can get all record in one report, do not use union.
    since all records, which you are targeting, are in the activity subject area, it is not nessecery to combine reports. add a column with the following logic
    if contact id is null (or = 'Unspecified') then owner name else contact name
    Hopefully, this is helping.

  • Performance issues -- related to printing

    Hi All,
    I am haviing production system performance issues related to printing. endusers are telling the printing is slow for almost printers. We are having more that 40 to 50 printers in landscape.
    As per my primary investigation I didnt find any issues in TSP01 & TSP02 tables. But I can see the table TST01 and TST03 table having many number of entries (more that lakh). I dont have idead about this table. Is ther eany thing related to this table where the print causes slowness or any other factors also makes this printing issue .. Please advice ..
    thanks in advance

    Hai,
    Check the below link...
    http://help.sap.com/saphelp_nw70/helpdata/en/c1/1cca3bdcd73743e10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/fc/04ca3bb6f8c21de10000000a114084/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/86/1ccb3b560f194ce10000000a114084/content.htm
    TemSe cannot administer objects that require more than two gigabytes of storage space, regardless of whether the objects are stored in the database or in the file system. Spool requests of a size greater than two gigabytes must therefore be split into several smaller requests.
    It is enough if you perform the regular background jobs and Temse consistency checks for the tables.
    This will help in controlling the capacity problems.
    If you change the profile parameter rspo/store_location parameter value to 'G' this will make the performance better. The disadvantages are TemSe data must be backed up and restored separately from the database using operating system tools, In the event of problems, it can be difficult to restore consistency between the data held in files and the TemSeu2019s object management in the database. Also you have take care of the Hard disk requirements because in some cases, spool data may occupy several hundred megabytes of disk storage. If you use the G option, you must ensure that enough free disk space is available for spool data.
    Regards,
    Yoganand.V

  • Convert smartform to PDF Convert_OTF - Performance Issue

    Hi gurus,
    We have a custom development for Compensation statement developed smartform that could have easily been done in Adobe PDF. The downside of this is ithat whenever we want to download to PDF, the program first coverts this smart form to PDF usinf funciton module "Convert_OTF" and then downloads it.
    Did any one face this performance issue. Is the convert to OTF a major time consuming process
    We are having system performance issue when downloading the comp statement and want to analysize how big this CONVERT_OTF is a factor.
    please suggest

    Ok
    Iìve never had particular performance problem with that function, so:
    have you tried to measure the time for the printing (smartform), for the convertion (convert_otf) and for the downloading separatly?
    Max

  • High volume of batches with Split valuation - impact on system performance

    Hi!
    I have a client that is intending to load a new material type from their legacy system which will be automatically batch managed with split valuation.  So, Valuation category will be 'x' and the valuation type will also be the batch number as automatically created on GR.
    The concern of the client is the impact on system performance.  Having up to 80,000 batches per material master record (so, 80,000 valuation types will be mainatined with a unique price in the Accounting 1 tab of the MMR) and overall around 1 million batches a year.  I'm not aware of any system performance issues around this myself but there seems to be anecdotal evidence that SAP has advised against using this functionality with high volumes of batches. 
    Could you please let me know of any potential problems I might encounter, having 1 million batches with split valuation may cause?  Logically, this would increase to tens of millions of batches over time until archived off via SARA.
    Many thanks!
    Anthony

    I currently have about 1.5 million batches with split valuation in my system (but it is not the X split), and we archive yearly.
    having many batches for one material ( lets say 1000) causes dramatic performace issues during automatic batch determination.
    it took about 5 minutes until a batch was returned into a delivery. if the user then wants a different batch and has to carry out batch determination again, then he just works for 10 to 15 minutes on one delivery.
    This is mainly caused by the storage location segment of the batches. if one batch gets movedd within a plant thru 3 different locations, then the batch has 3 records in table MCHB. But SAP has a report to reorganize the MCHB table that have zero stock.
    The X split has more effect, it is not only the batch table that makes issues in this case. With the x-split SAP adds an MBEW record (material master valuation view) for each new batch.
    However, if the design is made to get a certain functionality (here valution at batch level),   then you have to get a proper hardware in place that can give you the performance that is needed.

  • Classification data organization and system performance

    We are planning on setting up classification for the material master. Reviewing discussions on the topic indicate the potential for system performance issues with search response and file sizes. Performance is dependent on several factors such as the number of material numbers, classes and characteristics, use of interval values, depth of hierarchies ect. 
    We have a significant number of material numbers, looking to have at least 15 classes with about 200 total characteristics.  The extent of superior class linkage is undetermined at this time.
    I am looking for a guideline or tool that would help us understand if our design is efficient or will lead to long term performance issues. Is there a resource or documentation that gives some examples of situatinos where classification got our of hand and why?
    Thank you
    Brian

    Hello Brian,
    There is no document or guidlines as such however you should keep the tables upto date . Also we recommend to use CLST for updating statistics regularly. Are you looking to assign 15 classes for material ? It should not affect the performance as much but you should keep characteristics to minimal in that case. Also are you using VC as there are lot of recommendations for VC but in general for classification keeping the stats upto date is most important.
    Thanks
    Amber

Maybe you are looking for

  • Closing an iPad app in 7.0.4

    I updated my iPad to ios 7.0.4 yesterday. Previously, one could double-tap the button at the bottom, and then hold a finger on the resulting icons, and get a little X to close the application.  I had been tolD (at the PeachMac store where I bought it

  • How to export  the value  t.code me2s  to an excel sheet ?

    HI ,        THIS IS  sandipan.  I am very  new in this field .  My requirement is when i am  going me2s . then i am giving   purchasing group ,  plant ,  Service  Number as input.  then it showing purchase doc . ,  item , porg , matl... ,  plant , cr

  • Problem with tabbed panel

    I have a problem with tabbed panel...this panel has 3 columns, I copied to each of them a diffrent text (diffrent height)...I expected to obtain diffrent working area of column, for example  1st column-height 7500 px, 2nd-1500 px, 3rd-2500 px, but I

  • Also Question about create Random number

    What's the best way to create random number in main method, which will allow the number (random) to be stored into content of queue. I was thinking somewhere along the road, is this the appropriate way to handle this? import java.util.*; public stati

  • Why won't my CS5 Design Premium Update?

    I am unable to update any CS5 Design Premium program apart from Acrobat 9 Pro.  Support recommendation was to uninstall all Adobe programs, run the Cleaner Tool and then reinstall.  After doing this three times, deleting every add-on, finally only re