Need the memory consumption details of a interface

Hi,
We have scenario that there are 5000 messages were trigered at a time ie.., during peak time.
Is there any way in XI 3.0 to analyse the memory and CPU consumption of a single interface.
I am aware of the option to pull the entries from Performance monitoring in RWB. Is there any file where we can get the memory dumps for every 10 min.
The scenario consits of BPM where 5000 messages are triggered and get processed.
Please help.
Regards.
MM

You can check this in tcode ST06N,ST06 for time basis but for interface wise you can check user wise if uses different users for RFC connection in tcode STAD.
Thanks!

Similar Messages

  • Need the Mass Run Details

    Hi Team ,
    I have a Requirement , I need to  show the Mass Run details in the form of report  . Can any one suggest in this regard .
    Mass Run with Respect to Ea26 , EA38  and the  details must include portion and Division
    I have done few R and D in this regard .
    1)  Try to use the Application log but not able to fetch the document numbers tthat it has created in that RUN
    2) Whether FQEVENTS can be used in this regard .
    Please guide in this regard .
    Regards,
    K.Vinay Kumar

    Hi All ,
    I have done little RD on your suggestion , I have found that we can use the function module APPL_LOG_READ_DB , For reading
    all the messages that has occurred in the mass run . This function module is giving the details with   division and document number created  and failed and categorized in different way  like  WARNING , SUCCESS , STATISTICS etc .. . But some how I have tried on EMMA As well  but not able to use it as per my requirement .
    Thanks for every one on their inputs and suggestion .
    Regards,
    K.Vinay Kumar .
    Edited by: William Eastman on Feb 24, 2012 3:28 PM

  • Understand the memory consumption in Solaris

    Hi All,
    There is some things that I don't understand on Solaris, regarding the memory utilization.
    In my research on google, I have found this very usefull command : echo "::memstat" | mdb -k
    the output is :
    root@localhost > echo "::memstat" | mdb -k
    Page Summary Pages MB %Tot
    Kernel 290375 1134 7%
    ZFS File Data 55829 218 1%
    Anon 82841 323 2%
    Exec and libs 9844 38 0%
    Page cache 10370 40 0%
    Free (cachelist) 15677 61 0%
    Free (freelist) 3726068 *14554* 89%
    Total 4191004 16371
    Physical 4101523 16021
    I already know the vmstat command :
    root@PA-OFC-SRV-UAT-1 > vmstat 2
    kthr memory page disk faults cpu
    r b w swap free re mf pi po fr de sr lf s0 s1 -- in sy cs us sy id
    0 0 0 16286044 14528236 24 106 0 0 0 0 1 0 -0 2 0 704 950 410 0 0 100
    0 0 0 16390660 14629360 6 24 0 0 0 0 0 0 0 0 0 680 346 390 0 0 100
    0 0 0 16390572 14629308 1 2 0 0 0 0 0 0 0 0 0 685 219 362 0 0 100
    0 0 0 16390572 *14629308* 1 1 0 0 0 0 0 0 0 0 0 680 274 342 0 0 100
    ^C
    The thing that I don't understand is the difference between the 2 output for the free memory:
    FREE for mdb : 14554 MB -> 14.2 Go
    FREE for vmstat : 14629308 kb -> 14286.4MO -> 13.9Go
    I can tell that nothing is runing on my system (except the OS)
    The difference is not very big 0.3 as on this server I have :
    Memory size: 16380 Megabytes. (prtconf)
    A strange thing is that I have successfully started a java process with 15go (so 15360 Mo), even if I have 14554 MB:
    root@localhost > java -d64 -Xms15G -version
    java version "1.6.0_23"
    Java(TM) SE Runtime Environment (build 1.6.0_23-b05)
    Java HotSpot(TM) 64-Bit Server VM (build 19.0-b09, mixed mode)
    root@localhost > /home/ullink/COMMON/JAVA/latest/bin/java -d64 -Xms16G -version
    Error occurred during initialization of VM
    Could not reserve enough space for object heap
    BUT, on a servers which I have 73720 Megabytes (prtconf), the difference is bigger:
    root@host > echo "::memstat" | mdb -k
    Page Summary Pages MB %Tot
    Kernel 662611 2588 3%
    ZFS File Data 998688 3901 5%
    Anon 2139403 8357 11%
    Exec and libs 13959 54 0%
    Page cache 35561 138 0%
    Free (cachelist) 97387 380 1%
    Free (freelist) 14922431 *58290* 79%
    Total 18870040 73711
    Physical 18362749 71729
    vmstat:
    root@host > vmstat 2
    kthr memory page disk faults cpu
    r b w swap free re mf pi po fr de sr lf s1 s2 s3 in sy cs us sy id
    0 0 0 50601596 60622968 64 179 0 0 0 0 0 0 1 0 0 1112 11666 959 0 0 100
    0 0 0 48801836 58047100 59 78 0 0 0 0 0 0 0 0 0 1163 1200 860 0 0 100
    3 0 0 48801756 58047480 246 929 0 0 0 0 0 0 0 0 0 1160 194171 968 1 0 99
    0 0 0 48801556 *58047224* 149 656 0 0 0 0 0 0 0 0 0 1186 183253 980 1 0 99
    ^C
    so here I have:
    FREE for mdb -> 58290mb --> 56.9Go
    FREE for vmstat -> 58047224 b --> 55.35Go
    so here I have a difference of 1.5Go....
    I have the feeling that when more we have memory, more the difference is big.
    I investigate on this because on this (last) server, the production team of my compagny told me that they don't know "where" are some GB of RAM on this server....so I try to understand....
    I don't say that all my interpretation are good, but there is something I don't understand...
    so the question I have are :
    - why was I able to start a java process with 15G?
    - what is the difference between mdb and vmstat?
    Thank you a lot in advance

    How big is you SGA ? This is the memory Oracle is using.
    (if you don't know: see init.ora
    -> shared_pool
    -> large_pool
    -> java_pool
    -> buffer_cache * db_block_size
    -> log_buffer )

  • Need the location where ODI scenario and interfaces are physically stored

    Hi,
    I am new to ODI and would like to know in which folder ODI stores the interfaces, packages and scenarios.
    Any document reference or details will be very helpful.
    Regards,
    User1672911

    Hi User1672911,
    ODI will store all information in metadata table in your underlying WORK repository.
    Thanks,
    Guru

  • Need FM to get the last consumption date and value

    Hello Friends,
       i am developing a report for non moving material , i have got the relevant fields from MSEG,MKPF, MCHB, MCH1,,, but i need the last consumption date for that non moving material batch and value (total amount ) of that batch...
    so i need a Function Modules, to get this values .
    Please help me out
    Regards
    Mukesh

    You have to join MSEG and MKPF ( where MKPF has the posting date ) , ( day,month,year ) . You can calculate the month.moreover, also see S031,S032,S033
    else.
    check it at Mseg-lgort, and movement types for the given matnr.
    from there u can find the consumtion according to movement types .
    for particular period carry mblnr from mkpf and check the movement types which r equal to this mblnr and matnr in mseg.
    so u get the all movement types in that period for particular matnr.
    consumtion = issues - receipt.
    hope this helps u

  • How to measure memory consumption during unit tests?

    Hello,
    I'm looking for simple tools to automate measurement of overall memory consumption during some memory-sensitive unit tests.
    I would like to apply this when running a batch of some test suite targetting tests that exercise memory-sensitive operations.
    The intent is, to verify that a modification of code in this area does not introduce regression (raise) of memory consumption.
    I would include it in the nightly build, and monitor evolution of summary figure (a-ah, the "userAccount" test suite consumed 615Mb last night, compared to 500Mb the night before... What did we check-in yesterday?)
    Running on Win32, the system-level info of memory consumed is known not to be accurate.
    Using perfmon is more accurate but it seems an overkill - plus it's difficult to automate, you have to attach it to an existing process...
    I've looked in the hprof included in Sun's JDK, but it seems to be targetted at investigating problems rather than discovering them. In particular there isn't a "summary line" of the total memory consumed...
    What tools do you use/suggest?

    However this requires manual code in my unit test
    classes themselves, e.g. in my setUp/tearDown
    methods.
    I was expecting something more orthogonal to the
    tests, that I could activate or not depending on the
    purpose of the test.Some IDEs display mmeory usage and execution time for each test/group of tests.
    If I don't have another option, OK I'll wire my own
    pre/post memory counting, maybe using AOP, and will
    activate memory measurement only when needed.If you need to check the memory used, I would do this.
    You can do the same thing with AOP. Unless you are using an AOP library, I doubt it is worth additional effort.
    Have you actually used your suggestion to automate
    memory consumption measurement as part of daily builds?Yes, but I have less than a dozen tests which fail if the memory consumption is significantly different.
    I have more test which fail if the execution time is siginificantly different.
    Rather than use the setUp()/tearDown() approach, I use the testMethod() as a wrapper for the real test and add the check inside it. This is useful as different test will use different amounts of memory.
    Plus, I did not understand your suggestion, can you elaborate?
    - I first assumed you meant freeMemory(), which, as
    you suggest, is not accurate, since it returns "an
    approximation of [available memory]"freeMemory gives the free memory from the total. The total can change so you need to take the total - free as the memory used.
    - I re-read it and now assume you do mean
    totalMemory(), which unfortunately will grow only
    when more memory than the initial heap setting is
    needed.more memory is needed when more memory is used. Unless your test uses a significant amount of memory there is no way to measure it reliably. i.e. if a GC is perform during a test, you can have the test appear to use less memory than it consumes.
    - Eventually, I may need to inlcude calls to
    System.gc() but I seem to remember it is best-effort
    only (endless discussion) and may not help accuracy.if you do a System.gc(); followed by a Thread.yield() at the start it can improve things marginally.

  • How to overcome the Memory leakage issue in crystal report 2008 SP2 setup.

    I have developed the small windows based application tool with help of  Visual studio 2008 for identify the memory consumption of crystal report object. It helps to load the crystal report objects in the memory and then released the object from the memory. The tool simply does the u201CLoading and Unloadingu201D the objects in the memory.
    The tool will be started once u201CTest_MemoryConsumption.Exeu201D executed. The u201CTest_MemoryConsumption.Exeu201D consumes u201C9768 KBu201D memory before load the crystal report object in memory. It means, 9768 KB is normal memory consumption for run the tool.
    Crystal report object initiated by the tool and object help to load the report in memory once the tool initiated the crystal report object. Now u201CTest_MemoryConsumption.Exeu201D consumes u201C34980 KBu201D memory during the crystal report object creation and report load process. The actual memory consumption of crystal report object is 34980u20139768=u201C25212u201DKBu201D. 
    The memory consumption u201C34980 KBu201D will be continued till the end of the process. The memory consumption will be reduced to u201C34652 KBu201D from u201C34980 KBu201D once report load process completed. It means, u201C328 KBu201D memory only released from the memory consumption. Tool enables the Release command for the crystal report object. But crystal report object does not respond to the command and will not release his memory consumption.
    The memory consumption u201C34652 KBu201D will be stayed in the memory once job ends.  If i again initiate the crystal report object then it crystal report object start to consume the memory from 34652 KB.
    Database objects and crystal report objects are properly used in the tool. The object release commands properly  communicated to crystal report setup. But the u201CCrystal report service pack 2u201D setup unable to respond the commands which has enabled from .Net Tool.  Crystal report objects are properly initiated and disposed in the tool. But the crystal report unable to release from the server.
    The memory consumption will be reduced once the server restarted or kill the application.
    Crystal report 2008 and crystal report 2008 SP2 setup available in the server.
    Microsoft .Net Framework 2.0 SP2, Microsoft .Net Framework 3.0 SP2 and Microsoft .Net Framework 3.5 SP1 are available in the server,
    Could you please suggesst how to avoid the memory consumption keep increasing and  how to release the memory consumption  once the crystal object disposed???

    Hi Don..
    My case is different one. I hope, the problem with Run time Installation setup file (Crystal report 2008 Serivce Pack2 installer) which we installed in the server.
    Let me explain with Live scenario which our client faced in crystal report 2008 Service pack2 Installer.
    Our client is using a application to help to print their reports. The application is developed with Windows service.
    Windows service keep on running in the server. Windows service executes the client 's crystal reports( Labels Report, Stock  report) which designed for clients need and the reports will be printed from printer. 
    10 Same type report (Label Report) will be printed in 1 minute. Reports are not printing during non business hours. But the windows service keep on running.  Memory cosumption of application will be 160 MB in business hours.
    For Example, On Monday the application memory consumption starts with 160 MB. The Memory consumption will be reached 165 MB  in peak business hours. Then the memory will be ended in163 MB in the End of Monday. It means, The memory consumption will be in 163 MB during the non business hours. Reports will not be printed in non business hours.
    On Tuesday, the application memory consumption starts with 163 MB and it will be reached 168 MB during the peak hours. The Memory consumption will be ended in 165 MB in the end of Tuesday.  The same process contiues till friday. End of friday, the memory consumption of the application will be ended with 170 MB.
    Application Memory Consumption slowly increasing in the server. In 5 days, Memory consumption reached Threshold value (170 MB) of the server. Application gets hanged up once the memory consumption reached 170 MB. We got the error messages as "Attempted to read write protected memory " / "Not Enough memory for process".  If we restart server / If we restart the service then memory consumption of application get reduced to 160 MB.
    From the above scenario, We came know that the either the problem with Application object or the problem crystal report object. In the application, We have checked dispose methods of application objects completly. I am sure that  application objects are properly disposed in the application. I hope the problem not with application objects. The problem with Crystal report objects.
    Application properly communicates the dispose methods to crystal report objects. Crystal report objects are not released from
    the memory.
    Crystal report 2008 Serive Pack 2 setup installed in the server. 
    As you said, If Crystal report runtime is not released from memory then memory consumption keep increase???  In service oriented architecture application, how to unload the crystal report runtime??
    Do you any fix for this kind of issue??
    Willl Crystal report 2008 service pack 3 help on this issue??

  • Query on memory consumption during SQL

    Hi SAP Gurus,
    Could I kindly request for your inputs concerning the following scenario?
    To put it quite simply, we have a program where we're required to retrieve all the fields from a lengthy custom table, i.e. the select statement uses an asterisk.  Unfortunately, there isn't really a way to avoid this short of a total overhaul of the code, so we had to settle with this (for now).
    The program retrieves from the database table using a where clause filtering only to a single value company code.  Kindly note that company code is not the only key in the table.  In order to help with the memory consumption, the original developer had employed retrieval by packages (also note that the total length of each record is 1803...).
    The problem encountered is as follows:
    - Using company code A, retrieving for 700k entries in packages of 277, the program ran without any issues.
    - However, using company code B, retrieving for 1.8m in packages of 277, the program encountered a TSV_TNEW_PAGE_ALLOC_FAILED short dump.  This error is encountered at the very first time the program goes through the select statement, ergo it has not even been able to pass through any additional internal table processing yet.
    About the only biggest difference between the two company codes is the number of corresponding records they have in the table.  I've checked if company code B had more values in its columns than company code A.  However, they're just the same.
    What I do not quite understand is why memory consumption changed just by changing the company code in the selection.  I thought that the memory consumed by both company codes should be the same... at least, in the beginning, considering that we're retrieving by packages, so we're not trying to get all of the records all at once.  However, the fact that it failed at the very beginning has shown me that I'm gravely mistaken.
    Could someone please enlighten me on how memory is consumed during database retrieval?
    Thanks!

    Hi,
    with FAE (FOR ALL ENTRIES) the whole query even for a single record in the itab is executed and all results for
    the company code are transfered from the database to the DBI since the duplicates will be removed by the DBI
    not by the database.
    If you use package size the resultset is buffered in a system table in the DBI (which allocates memory from your user quota). And from there on the package sizes are built and handed over to your application (into table lt_temp).
    see recent ABAP documentation:
    Since duplicate rows are only removed on the application server, all rows specified using the WHERE condition are sometimes transferred to an internal system table and aggregated here. This system table has the same maximum size as the normal internal tables. The system table is always required if addition PACKAGE SIZE or UP TO n ROWS is used at the same time. These do not affect the amount of rows transferred from the database server to the application server; instead, they are used to transfer the rows from the system table to the actual target area.
    What you should do:
    calculate the size needed for your big company code B. How many row  multiplied with line length.
    That is the minimum amount you need for your user memory quota. (quotas can be checked with
    ABAP report RSMEMORY) If the amount of memory is sufficient then try without packagesize.
    SELECT * FROM <custom table>
    INTO TABLE lt_temp
    FOR ALL ENTRIES IN lt_bukrs
    WHERE bukrs = lt_bukrs-bukrs
    ORDER BY primary key.
    This might actually use less memory than the package size option for the FOR ALL ENTRIES.
    Since with FAE it is buffered anyway in the DBI (and subtracted from your quota) you can
    do it right away and avoid double saving portions (the DBI buffer and a portion of that in the
    packe in lt_temp).
    If the amount of memory is still too big, you have to either increase the quotas or select
    less data (additional where conditions) or avoid using FAE in this case in order to not read all
    the data in one go.
    Hope this helps,
    Hermann

  • Check Process memory consumption and Kill it

    Hello
    I have just installed Orchestrator and have a problem that I think is perfekt for Orchestrator to handle.
    I have a process that sometimes hangs and the only way to spot it is that the memory consumption has stoped.
    The process is started every 15 minutes and scans a folder, if it finds a file it reads the file to a system. You can see that it is working by the increasing Memory consumption. If the read fails then the memory consumption stops. The process is still working
    and is responding but is hung.
    I'm thinking about doing a runbook that checks the memory-consumption every 5 minutes and compares it with the previous value. if the last three values are the same then I will kill the process and start it again.
    My problem is that I have not found a way to check the memory consumption of a process.
    I have set up a small test, just verify that I get the correct process, with the activity Monitor process -> Get Process Status -> Append Line (process name).
    But How do I get the process memory consumption?
    /Anders

    Now that I think about it a bit more I don't think there will be an easy way to set up a monitor for your situation in SCOM. Not that it couldn't be done, just not easily. Getting back to SCORCH. What you are trying to do isn't an every day kind of
    scenario. I don't think there is a built in activity for this.
    The hardest thing to overcome whether you use SCORCH or SCOM is likely going to be determining the error condition of three consecutive samples of the same memory usage. you'll need a way to track the samples. I can't think of a good way to do
    this without utilizing scripting.

  • Query on Memory consumption of an object

    Hi,
    I am able to get information on the number of instances loaded, the memory occupied by those instances using heap histogram.
    Class      Instance Count      Total Size
    class [C      10965      557404
    class [B      2690      379634
    class [S      3780      220838
    class java.lang.String      10807      172912 Is there way to get detailed info like, String object of which class consume much memory.
    In other words,
    The memory consumption of String is 172912. can I have a split up like
    String Objects of Class A - 10%
    String Objects of Class B - 90%
    Thanks

    I don't know what profiler you are using but many memory profilers can tell you where the strings are allocated.

  • Query memory consumption

    Hi,
    Need some expert in SQL here. May i know how much memory (RAM) consumption for a simple query like 'SELECT SUM(Balance) FROM OCRD' cost.
    What about query like
    select (select sum(doctotal) from ordr) + (select sum(doctotal) from odln) + (select sum(doctotal) from oinv)
    How much memory would it normally takes? The reason is that i have a query that is quite similar to this and it would be run quite often. So i wonder if it is feasible to use this type of queries withought making the server to a crawl.
    Please note that the real query would include JOINS and such. Thanks
    Any information is appreciated

    Hi Melvin,
    Not sure I'd call myself an expert but I'll have a go at an answer
    I think you are going to need to set up a test environment and then stress test your solution to see what happens. There are so many different variables that affect the memory consumption that no-one is likely to be able to say just what the impact will be on your server. SQL Server, by default will allocate 1024Kb to each query but, of course, quite a number of factors will affect whether SQL needs more memory than this to execute a particular query (e.g. the number of joins, the locks created, whether the data is grouped or sorted, the size of the data etc etc). Also, SQL will release memory as soon as it can (based on its own algorithms) so a query that is run periodically has much less impact on the server than a query that will be run concurrently by multiple users. For these reasons, the impact can only really be assessed if you test it in a real-world scenario.
    If you've ever seen SQL Server memory usage when XL Reporter is running a very large report then you'll know that this is a very memory hungry operation. XL Reporter bombards SQL with a huge number of separate little queries and SQL Server starts grabbing significant amounts of memory to fulfill these queries. As the queries are coming so fast, SQL hasn't yet got around to releasing the memory used by previous queries so SQL instead grabs available memory from the server.
    You'll get better performance and scaleability by using stored procedures but SDK certification does not allow the use of SPs in the SBO databases.
    Hope this helps,
    Owen

  • Memory consumption is more with oracle database compared to Sybase

    Hi,
    We are executing the same java source code with backend sybase and the oracle database. But with oracle database it is consuming more memory than Sybase.
    Currently using 11g R2 with ojdbc6.jar driver
    can you please provide the information to optimize the memory consumption when using the oracle.
    Thanks,
    Nagaraj

    user12569889 wrote:
    We are executing the same java source code with backend sybase and the oracle database. But with oracle database it is consuming more memory than Sybase. That is not saying anything at all.
    What memory in Oracle? Shared pool? Buffer cache? PGA? UGA? Library cache? Something else?
    What memory in Sybase did you compare it to?
    What type of client-server connection to Oracle was made? Dedicated or shared?
    What commands/method were used to determine the memory consumption in Oracle and then Sybase?
    What serves as the baseline for comparison?
    Comparing product A with product B is a COMPLEX thing to do. And IMO, beyond the abilities of the majority of developers - as they lack the technical expertise to extract usable metrics and correctly compare these between products that can work VERY differently.
    And unless you can provide technical details to backup your claim that you are observing that "+Memory consumption is more with oracle database compared to Sybase+", I would say that you have no idea what you are actually observing and in no position to deduce that Oracle consumes more memory.

  • Excessive memory consumption when loading Customers through Component Inter

    Hi All,
    I'm facing a big problem with the high memory consumption when loading Customers, Companies and Sites using the Component Interfaces delivered by the product (RD_CONSUMER_CI_API, RD_COMPANY_CI_API, RD_SITE_CI_API) within Application Engine programs. I'm loading about 7 million customers, an amount that is not so big in my opinion, but the memory consume is too high.
    We have 3 Batch Servers, each one running under Red Hat OS with 32 GB RAM memory plus 32 GB Swap memory in each server. We are running 2 Process by Server and with a day and half the servers crash with 100% of memory consumed (RAM and SWAP).
    There's a good pratice to use Component Interface in a heavy load process?
    There are parameters in the process scheduler configuration file that could help to reduce the memory consumption?
    There's a way to free the memory through PeopleCode or by running another process?
    Thanking you in advance.

    You may want to try and cut down on the input data to ascertain that the load might be a problem.
    You may try and use the GarbageCollector, but might not help in your case.
    To get an idea of the size allocated in buffer for the Rowset being used, you may want to check out the memory overhead ...
    Also, could check, which is the process that is consuming a lot of memory

  • Measuring Swing GUI memory consumption

    Hi community,
    First , I beg you for excuse for posting this here and not in the Swing forum. I think the question is not clear GUI question, but a more generic one.
    I have a GUI that seems to consume a lot of memory(i.e. opening several screens causes OutOfMemoryError and data received on network measures to a very small amount (<5%) compared to the total memory shown by Task Manager).
    I would like to take these screens apart to discover which of their parts are heavy so that I can concentrate my efforts on optimizing them. I cannot use ObjectOutputStream serialization since it seems there are lots of classes that do not implement Serializable, so I need some other way of determining component memory sizes. I believe that XMLEncoder serialzation will leed to nowhere because it contains a lot of text describibg properties that cannot be accurately measured.
    What can I use to measure the memory consumption for a given component tree?
    Thank you for your time
    Mike

    Swing related questions should be...
    Umm have you considered the use of a Profiler? I kind of doubt it is the GUI per se but a model or something that is leaking memory. I mean there are lot's of fairly complex GUI's out there that don't have these issues so I suspect there is a very deep or complex structure somewhere that is eating all of the memory.
    But in short I think running in some sort of profiler would be good becuase I also suspect that this may be one of this issues where looking at each component may not help because it's as the whole thing is used at runtime where some complexity comes into play and it dies.
    Try http://www.manageability.org/blog/stuff/open-source-profilers-for-java for a start.

  • Continuously refreshing a tab after an interval leads to high memory consumption (400MB to 800MB in 30 seconds for 3 refreshes at 10 secs interval), why?

    Environment:
    MAC OSX 10.9.5
    Firefox 32.0.3
    Firefox keeps consuming lot of memory when you keep refreshing a tab after an interval.
    I opened a single tab in my firefox and logged into my gmail account on that. At this stage the memory consumption was about 400MB. I refreshed the page after 10 seconds and it went to 580MB. Again i refreshed after 10 seconds and this time it was 690MB. Finally, when i refreshed 3rd time after 10 seconds, it was showing as 800MB.
    Nothing was changed on the page (no new email, chat conversation, etc. nothing). Some how i feel that the firefox is not doing a good job at garbage collection. I tested this use case with lot of other applications and websites and got the similar result. Other browsers like Google chrome, safari, etc. they just work fine.
    For one on of my application with three tabs open, the firefox literally crashed after the high memory consumption (around 2GB).
    Can someone tell me if this is a known issue in firefox? and is firefox planning to fix it? Right now, is there any workaround or fix for this?

    Hi FredMcD,
    Thanks for the reply. Unfortunately, i don't see any crash reports in about:crashes. I am trying to reproduce the issue which will make browser to crash but somehow its not happening anymore but the browser gets stuck at a point. Here is what i am doing:
    - 3 tabs are open with same page of my application. The page has several panels which has charts and the javascript libraries used for this page are backbone.js, underscore.js, require.js, highcharts.js
    - The page automatically reloads after every 30 seconds
    - After the first loading of there three tabs, the memory consumption is 600MB. But after 5 minutes, the memory consumption goes to 1.6GB and stays at this rate.
    - After sometime, the page wont load completely for any of the tabs. At this stage the browser becomes very slow and i have to either hard refresh the tabs or restart the browser.

Maybe you are looking for