JVM memory problem

I have a memory problem which it exceeds a maximum limit and my application will suspend to restart my service of this application, after that everything is okay.
so i need ask about,
How can I know if my Garbage Collection is working or not??
How can I force Garbage Collection to work???
How can I force Garbage Collection to stop???
if there are addetional way to trace Memory performance beside Runtime class??

Cross post:
http://forum.java.sun.com/thread.jspa?threadID=693125

Similar Messages

  • JVM memory breakup

    Hi,
    My java application (with jdk1.5.0_03) shoots upto 169 MB of Process size (seen in Task manager)
    and the heap size when printed (Runtime.totalMemory()-Runtime.freeMemory())
    shows as 68 MB. I assume that the process size is the total JVM memory.
    I would like to know exact breakup of JVM memory; is it possible to find out the individual parameters that contribute to the total process memory.
    Thanks
    Srila

    Hi,
    I have a similar kind of problem.
    We have a web-application running on jdk1.5 and tomcat 5. I want that my application should not take more than 48 MB memory on to my machine. I used the -Xmx48m option, but it just controls the heap memory. But what about the non-heap memory? Here are few things I have noticed
    -> when I use JConsole to monitor the application, I found that my applications non-heap memory Max is set to 96M.
    -> Another thing I have noticed that when I use "top" command to monitor memory footprint of my application, the RSS (which shows the PHYSICAL MEMORY USED ) shows an alarming 125MB of usage ( Sometimes when i use -server option it used to go upto 250M ) , which may harm my other applications which are memory prone. Even when the GC runs java do not free any memory for other application.
    -> I have tried many options like
         -XX:+UseConcMarkSweepGC
         -XX:NewSize=8m -XX:MaxNewSize=8m -XX:SurvivorRatio=2 -Xms48m -Xmx48m -Xss64k
         -XX:MaxHeapFreeRatio=20 -XX:MinHeapFreeRatio=10 -XX:NewSize=32m -XX:SurvivorRatio=32 -Xss256k -Xms48m -Xmx48m
         -XX:+UseParallelGC -XX:GCTimeRatio=20 -Xms30m -Xmx30m -Xss2048k
         but failed to restrict it to 48M
    Before using Sun JDK5 we are using IBM jdk1.3 with tomcat 3 with green thread. So we were able to configure it with just one option -mx48m. I am really stuck. I cannot believe that Sun have not given us any control on the maximum memory usage. Please help me .
    Regards
    Purav Gandhi

  • Stringbuffer memory problem

    Hi guys
    I am developing an applet which reads a 25mb text file into a stringbuffer and outputs the stringbuffer content to a text area. the file is an auto generated report.
    File f = new File ("c:\\my_file.txt");
    FileReader fr = new FileReader(f);
    BufferedReader br = new BufferedReader(fr);
    StringBuffer sb = new StringBuffer();          
    String oLine;     
    while ((oLine = br.readLine()) != null) {
         sb.append(oLine+"\n");
    br.close();               
    TextArea data = new TextArea(sb.toString(),50,110);This works fine when I increase my JVM size to 128mb. But runs out of memory when using the default JVM size.
    Is there a more efficient way I can do this so that it works in default JVM memory size? A datatype that takes up less memory than stringbuffer perhaps?
    Thanks in advance.

    The question is: what are 25 MB of text doing in an applet? Who's supposed to look at all that info? Trim down the amount of information you display. You couldn't even navigate through that pile of text.
    A datatype that takes up less memory than stringbuffer perhaps?How would that work? If you have 25 MB of data, you'll have 25 MB of data.
    I guess your problem comes up when you create a String from the StringBuffer - I don't know if or how StringBuffer and String are sharing their backing arrays, but I'd assume for simplicity's sake they don't. So you end up with a 25 MB String and a 25 MB StringBuffer.

  • What would make a JVM memory size 1.9GB ?

    This is on Solaris 2.6, JDK 1.2.2_05a, JDK1.2.2_07, possibly JDK 1.3 (it crashed).
    WLS 5.1, SP8. With inline_instrs_jit=0.
    This is the [heap] segment shown by /usr/proc/bin/pmap. This is native heap, it is
    NOT the Java Heap.
    I know it could be the Type 2 db drivers - but I doubt it.
    With JDK1.3 we set -XX:MaxPermSize=128m - it crashes after an hour.
    One part of the application creates its own threads for some parallel processing.
    Is this a known
    problem with the JVMs?
    Any help is much appreciated.
    Mike Reiche

    What is the 'Break difference' that the debug JVM shows? Is that the Java Heap? Or
    is memory in the C Heap that is wasted by fragmentation?
    ============== C Heap Report ========================
    mmap space 300515328
    Committed space 95690752
    Allocated space 34049316
    Break difference 66060288
    Breakdown by type:
    ClassSpace 15252128
    JITCompiledStackMaps 5277188
    StackMap 1892672
    Dependency 1856760
    utf8_space 1854838
    untagged 1783520
    basicBlocksState 1664992
    utf8_table 1055903
    Class 811520
    JITPrestub 610480
    NearClass 531360
    CardObjectTable 386048
    JavaStack 257024
    CheckingDone 200367
    SummaryCardTable 193024
    name_loader_cache_entry 65976
    ScratchRecord 65548
    basic_block_t 50896
    IntfMethodTable 48960
    stringTableOverflowRecord 46480
    BitVectorElem 32768
    ObjectMap 30724
    ClassName 28407
    Clinit 18752
    Table_mem 18628
    LoaderConstraintsTableEntry 11668
    JITInlineByteCode 1928
    LoaderConstraintsTableLoaderEntry 520
    stateVecBuf 237
    =====================================================
    Total GC time: 2987981 ms
    - Mike
    "[email protected]" Mike wrote:
    >
    >
    Aslo - GC[1] are taking way long (Debug JVM) - I'm used to seeing them taking
    500
    ms with the regular JVM. With -ms32m -m256m - I wanted GC[1] to run often
    because
    that has helped with other memory problems in the past - it doesn't work
    so hot here.
    GC[1] in 5514 ms: (38Mb, 0% free) -> (38Mb, 69% free)
    GC[1] in 6300 ms: (38Mb, 0% free) -> (38Mb, 65% free)
    GC[1] in 7130 ms: (38Mb, 0% free) -> (38Mb, 64% free)
    GC[1] in 6158 ms: (38Mb, 2% free) -> (38Mb, 62% free)
    GC[1] in 6171 ms: (38Mb, 0% free) -> (38Mb, 65% free)
    GC[1] in 6885 ms: (38Mb, 0% free) -> (38Mb, 64% free)
    GC[1] in 7403 ms: (38Mb, 0% free) -> (38Mb, 61% free)
    GC[1] in 689101:25:49 PM PST {218} - Evaluation is: true
    GC[1] in 6545 ms: (38Mb, 1% free) -> (38Mb, 61% free)
    GC[1] in 7459 ms: (38Mb, 0% free) -> (38Mb, 62% free)
    GC[1] in 6861 ms: (38Mb, 2% free) -> (38Mb, 58% free)
    GC[1] in 6090 ms: (38Mb, 0% free) -> (38Mb, 56% free)
    GC[1] in 6413 ms: (38Mb, 0% free) -> (38Mb, 60% free)
    GC[1] in 8230 ms: (38Mb, 1% free) -> (38Mb, 58% free)
    GC[1] in 7649 ms: (38Mb, 0% free) -> (38Mb, 54% free)
    GC[1] in 7912 ms: (38Mb, 0% free) -> (38Mb, 55% free)
    GC[1] in 7696 ms: (38Mb, 1% free) -> (38Mb, 57% free)
    GC[1] in 6543 ms: (38Mb, 1% free) -> (38Mb, 60% free)
    GC[1] in 7036 ms: (38Mb, 1% free) -> (38Mb, 59% free)
    GC[1] in 7483 ms: (38Mb, 0% free) -> (38Mb, 56% free)
    GC[1] in 5132 ms: (38Mb, 1% free) -> (38Mb, 59% free)
    GC[1] in 8186 ms: (38Mb, 0% free) -> (38Mb, 53% free)
    GC[1] in 5691 ms: (38Mb, 0% free) -> (38Mb, 58% free)
    GC[1]01:35:49 PM PST {575} - Evaluation is: true
    GC[1] in 8176 ms: (38Mb, 0% free) -> (38Mb, 46% free)
    GC[1] in 8463 ms: (38Mb, 0% free) -> (38Mb, 47% free)
    GC[1] in 7342 ms: (38Mb, 0% free) -> (42Mb, 41% free)
    GC[1] in 6920 ms: (42Mb, 0% free) -> (49Mb, 41% free)
    GC[1] in 8610 ms: (49Mb, 1% free) -> (60Mb, 40% free)
    GC[1] in 9581 ms: (60Mb, 1% free) -> (60Mb, 50% free)
    GC[1] in 10514 ms: (60Mb, 0% free) -> (60Mb, 59% free)
    GC[1] in 9865 ms: (60Mb, 0% free) -> (60Mb, 60% free)
    GC[1] in 11323 ms: (60Mb, 0% free) -> (60Mb, 57% free)
    GC[1] in 18768 ms: (60Mb, 0% free) -> (60Mb, 60% free)
    GC[1] in 11147 ms: (60Mb, 0% free) -> (60Mb, 55% free)
    GC[1] in 7293 ms: (60Mb, 0% free) -> (60Mb, 42% free)
    GC[1] in 9372 ms: (60Mb, 0% free) -> (60Mb, 53% free)
    GC[1] in 8814 ms: (60Mb, 0% free) -> (60Mb, 55% free)
    GC[1] in 9695 ms: (60Mb, 1% free) -> (60Mb, 54% free)
    GC[1] in 7309 ms: (60Mb, 0% free) -> (60Mb, 53% free)
    GC[1] in 8713 ms: (60Mb, 0% free) -> (60Mb, 54% free)
    GC[1] in 11955 ms: (60Mb, 0% free) -> (60Mb, 55% free)
    GC[1] in 12685 ms: (60Mb, 1% free) -> (60Mb, 55% free)
    GC[1] in 18906 ms: (60Mb, 0% free) -> (60Mb, 58% free)
    GC[1] in 13333 ms: (60Mb, 0% free) -> (60Mb, 51% free)
    "[email protected]" <[email protected]> wrote:
    Rob -
    Thanks for your attention -
    We create lots of threads - up to five threads for each DB hit (i didn't
    design it).
    The multi-threading will be disabled/removed and tested later on today.
    Both 'top'
    and pstack show around 85 threads.
    I've read articles on the bea newsgroups about WL hanging on to threadreferences
    so that the threads could not be GC'ed, but it sounded more like a customer's
    impression
    of what might be going on than something from BEA engineering.
    Right now we are running with the 1.2.2_05a 'Debug' JVM with -verbose:gc
    -verbose:gc
    -verbose:gc
    JVMARGS=verbose_c_heap,inline_instrs_jit=0
    Unfortunately, some of the code being tested is now broken and not being
    excercised
    today. I will keep you posted.
    - Mike
    Rob Woollen <[email protected]> wrote:
    How many threads are you creating?
    Each thread uses some memory (stack etc.), but 1.9GB would be a lot of
    thread creation.
    -- Rob
    Mike Reiche wrote:
    This is on Solaris 2.6, JDK 1.2.2_05a, JDK1.2.2_07, possibly JDK 1.3
    (it
    crashed).
    WLS 5.1, SP8. With inline_instrs_jit=0.
    This is the [heap] segment shown by /usr/proc/bin/pmap. This is nativeheap, it is
    NOT the Java Heap.
    I know it could be the Type 2 db drivers - but I doubt it.
    With JDK1.3 we set -XX:MaxPermSize=128m - it crashes after an hour.
    One part of the application creates its own threads for some parallelprocessing.
    Is this a known
    problem with the JVMs?
    Any help is much appreciated.
    Mike Reiche--
    Coming Soon: Building J2EE Applications & BEA WebLogic Server
    by Michael Girdley, Rob Woollen, and Sandra Emerson
    http://learnweblogic.com

  • CFC method memory problem

    Hi all,
    I use CFCs a lot, so when I was creating a tool to migrate
    some database tables, I built the actual data transfer function as
    a method of a CFC. The basic code just reads a block of 1000
    records from the old db, does some minor manipulation of the data,
    and then inserts the records into the new db a record at a time
    inside a cfloop. Lather, rinse, repeat, until all of the 500K+
    records have been transferred.
    I was surprised to find that this leads to Java out of memory
    errors around the 200Kth record (given my current JVM memory
    settings). Having run across a similar problem when sending
    broadcast emails from a CFC method, I simply moved the code out of
    the CFC method and into the calling cfm page, and it ran through
    all of the records without any problems.
    The only explanation I can think of for this behavior is that
    when the code is inside a CFC method, the query object used to read
    the blocks of 1000 records keeps getting reallocated rather than
    reused, so when it gets to the point of trying to store 200 blocks
    of 1000 records in memory, it runs out of room. If that's the case,
    it would seem to be a significant bug that needs to be fixed.
    Does anyone know if that's what is really going on?
    Thanks,
    Russ

    > Actually, there is a good reason to put the code in a
    cfc: there are almost 100
    > tables to migrate, so the code gets rather long and
    cumbersome to wade through
    > if it isn't broken up.
    Fair enough. Youd didn't really tell us that part before,
    though.
    So you have a calling template which calls an equivalent to
    this function
    for 100-odd tables? Yikes. I'll reiterate my question whether
    this can't
    possibly be done DB to DB or via a bulk insert.
    > thing, but ultimately my question really has nothing to
    do with this particular
    > tool. I would really just like to understand why the out
    of memory issue exists
    > when the code is in a CFC method versus when it is in a
    CFM. This issue has
    Well it kinda is about this particular tool, because it's
    this code that's
    causing the problems (or something about the calling code, or
    something
    like that). Obviously there's some idiosyncasy of it which
    gives you
    problems when it's in a CFC method as opposed to mainline
    code, but I would
    not immediately say it's a generic problem with CFCs vs CFMs.
    Just yet.
    Is this the first table that's being migrated? Or is it
    buried somewhere
    within the other 100-odd? Is it always this one, or can it
    happen on any
    of them?
    Can you inspect the memory usage between each table's
    function, and see if
    there's anything unexpected going on (other than it running
    out of memory,
    which is - in itself - unexpected!).
    Do you have 100-odd methods (one for each table) in the CFC,
    or lots of
    individual CFCs?
    I guess you could call a GC in between each table's
    processing, and see if
    you can keep check on memory usage that way. Or maybe check
    memory levels
    within that loop you've got, and if it starts flaring up, doa
    GC then.
    It's not advised to call GCs "by hand", but I have found it's
    helped some
    times.
    Have you sued CF's server monitor or something like
    FusionReactor to check
    if anything unexpected is consuming RAM?
    I think you should try adding <cfqueryparam> tags to
    your queries: CF might
    be cachinng some info about them, and that could be leaking
    memory or
    something. Maybe try <cfobjectcache action = "clear">
    between each method
    callor something.
    I wonder if there's something about the fact you're replacing
    oldInfo with
    a new query within a loop that has a condition based on
    oldInfo's record
    count. Possibly this is forcing CF to maintain the previous
    oldInfo
    queries in the background, so you're not actually overwriting
    the previous
    one with the next one when you requery, you're actually just
    getting a new
    pointer assigned. Do me a favour, and do this:
    <cfset iRecCount = OldInfo.RecordCount>
    <cfloop condition="iRecCount GT 0">
    <!--- Load the old info --->
    <cfquery name="OldInfo" datasource="dsOld"
    maxrows="#GetRecsCount#">
    </cfquery>
    <cfset iRecCount = OldInfo.RecordCount>
    </cfloop>
    This is a slightly far-fetched notion, but it doesn't hurt to
    try: it's a
    pretty easy change.
    > come up enough times now that I'd like to determine if
    there is a workaround.
    > Due to the deadline I'm on, I haven't tried simply
    assigning the query object
    > to an empty string at the end of each iteration yet, but
    I will try to do so
    > soon so I can report on the result.
    It might pay to apply a test rig to this notion first:
    0) on a dev machine without any other activity
    1) grab the heap usage
    2) load a big query in
    3) grab the heap usage (should losely reflect the fact you've
    just stuck a
    chunk of data in RAM)
    4) set the query variable to be an empty string
    5) grab the heap usage. Check to see if the memory is
    reclaimed straight
    away.
    You might need to do a GC between 4+5 for it to take effect.
    I speculate
    that the RAM won't get freed up straight away, and you might
    be consuing it
    faster than the JVM can clear it.
    Adam

  • Display jvm memory use of the server

    hello ,
    I'm trying to display a graph of jvm memory use of the server.
    my problem is how to communicate data between the server and the client ,
    should I use applet to display the graph ? , the use of memory change with the time , how manage that ?
    thanks

    A method to retrieve the call graph would be great!
    I just discovered that NI Idea Exchange only has a forum for LabVIEW. Maybe it's time to add TestStand the the Idea Exchange. Where do I suggest this idea?!
    Message Edited by Phillip Brooks on 08-19-2009 01:56 PM
    Now is the right time to use %^<%Y-%m-%dT%H:%M:%S%3uZ>T
    If you don't hate time zones, you're not a real programmer.
    "You are what you don't automate"
    Inplaceness is synonymous with insidiousness

  • Newbie : memory problem with file uploading

    Hello
    I'm exdending a web app based on jsf and , being new to it and java in general, i'm having a problem with file uploading.
    The customer needs to upload large file (more than 30 MB). The application accomplish this using a fileupload component, to let the user select the file, and passing the array of bytes to a web service method.
    I'm using netbeans 6, tomcat 5.5.
    When i try to upload a file too large i got an outofmemory on the local tomcat. So i've rised jvm memory and going on with tests i've found that growing the size i got an http 500 error from the service (the developer of the service has found an out of memory too ).
    I think the service is not modifiable, so i were asking if threre were a way to optimize the file uploading process (a component, a library, a pattern ...) or my only chioice is to rise available memory on both sides.
    I've notice the wsdl declares as base64binary type the data parameter , while netbeans 6 generates a client proxy with a byte[] parameter. Is correct or is a netbeans problem?
    searching with google i've found few thing i haven't fully understood due to my lack of experience:
    the mtom , how can evaluate if it's suitable to help in my problem (i suppose it involves more the web service rather than the client)
    something about the size of a soap body.
    thank you in advance
    Stefano

    Which component exactly are you using? Declaring the stream as a raw byte[] is certainly not efficient. Write code yourself, don't let it autogenerate.
    I can highly recommend you the Tomahawk t:inputFileUpload component. It is easy to integrate in Mojarra. You can find here an useful article: [http://balusc.blogspot.com/2008/02/uploading-files-with-jsf.html].

  • CF 8 JVM memory is not being garbage collected.

    I am baffled by something I am seeing on my QA server. I have
    an app that we load tested but when the test completed the JVM
    memory used was not released. I used CF Server Monitor to watch the
    memory usage and sometimes it spiked to the max and either the app
    failed or I got timeout exceptions.
    This is the only app running on this server and the testing
    completed over an hour ago but the memory has not been released
    yet.
    CF Admin settings:
    Maximum JVM Heap Size (MB) 512
    The CF Server JVM Setting arguments include: -server
    -Dsun.io.useCanonCaches=false -XX:MaxPermSize=192m
    -XX:+UseParallelGC
    I found a script that uses java.lang.Runtime and
    java.lang.management.managementFactory that dumps a JVM memory
    usage profile The latest dump follows:
    JVM Monitor - ColdFusion Server - Enterprise v8,0,1,195765
    JVM Memory Monitor - struct
    Heap Memory Usage - Committed 481 MB
    Heap Memory Usage - Initial 0.00 MB
    Heap Memory Usage - Max 493 MB
    Heap Memory Usage - Used 437 MB
    JVM - Free Memory 44.0 MB
    JVM - Max Memory 493 MB
    JVM - Total Memory 481 MB
    JVM - Used Memory 449 MB
    Memory Pool - Code Cache - Used 8.80 MB
    Memory Pool - PS Eden Space - Used 6.37 MB
    Memory Pool - PS Old Gen - Used 428 MB
    Memory Pool - PS Perm Gen - Used 52.4 MB
    Memory Pool - PS Survivor Space - Used 3.50 MB
    Non-Heap Memory Usage - Committed 62.8 MB
    Non-Heap Memory Usage - Initial 18.3 MB
    Non-Heap Memory Usage - Max 240 MB
    Non-Heap Memory Usage - Used 61.2 MB
    According to the CF Server Monitor JVM memory usage builds up
    to 477 MB then the app fails or timesout.
    Session Scope memory usage: 0.27 KB
    Application Scope memory usage: 1.370 KB
    Server Scope memory usage: 3.12 KB
    Since the test ended JVM Memory Usage has dropped back to 438
    MB?
    Besides CFAdmin nothing else is running on this CF Server.
    I've read several other memory related topics but none of them have
    helped.
    Can someone tell me why the memory isn't being release? How
    can I further troubleshoot the problem?
    Thx
    pwp

    > Adam Cameron wrote:
    >
    The maximum stable heap size I've managed to get is around
    > 1.0-1.2GB, on a win32 system. On Solaris (running a
    32-bit JVM),
    > about 1.4GB. It *seems* like GC doesn't actually clear
    out RAM
    > properly if more than that much RAM is being
    addressed.
    >
    > Yes, there is a well-known
    >
    http://kb.adobe.com/selfservice/viewContent.do?externalId=tn_19359&sliceId=1
    Not really what I was talking about. One might be able to get
    the CF
    instance to *start* with 1.8GB allocated to the heap, but it
    won't actually
    work. I've managed to get a server to idle for a reasonable
    length of time
    on 1.5GB, but as soon as the thing started to ramp up, it
    face-planted,
    once it started actually trying to *use* the higher end of
    the RAM
    allocated to it. At 1.2GB, it'll seem to run OK for a
    reasonable length of
    time, but eventually it starts leaking memory; at around 1GB,
    it was pretty
    stable.
    Hence my comment about it being *stable* at that allocation.
    Not that "it
    simply won't start if more than 1.8GB is allocated to it".
    My point was that your rule of thumb:
    maximum heap size(Xmx) = RAM(in MB) / (2 * number of servers
    using the
    JVM)
    Is not a very good one. Plug 4GB RAM (so a small server) and
    one CF
    instance into that equation. Your rule suggests I should be
    allocating 2GB
    to the heap. Which - as you yourself pointed out - won't
    work.
    Adam

  • Memory Problem with SEt and GET parameter

    hi,
    I m doing exits. I have one exit for importing and another one for changing parameter.
    SET PARAMETER exit code is ....
    *data:v_nba like eban-bsart,
           v_nbc like eban-bsart,
           v_nbo like eban-bsart.
           v_nbc = 'CAPX'.
           v_nbo = 'OPEX'.
           v_nba = 'OVH'.
    if im_data_new-werks is initial.
      if im_data_new-knttp is initial.
        if im_data_new-bsart = 'NBC' or im_data_new-bsart = 'SERC' or im_data_new-bsart = 'SERI'
           or im_data_new-bsart = 'SER' or im_data_new-bsart = 'SERM' or im_data_new-bsart = 'NBI'.
          set parameter id 'ZC1' field v_nbc.
        elseif im_data_new-bsart = 'NBO' or im_data_new-bsart = 'NBM' or im_data_new-bsart = 'SERO'.
          set parameter id 'ZC2' field v_nbo.
        elseif im_data_new-bsart = 'NBA' or im_data_new-bsart = 'SERA'.
          set parameter id 'ZC3' field  v_nba.
        endif.
      endif.
    endif. *
    and GET PARAMETER CODE IS....
      get parameter id 'ZC1' field c_fmderive-fund.
      get parameter id 'ZC2' field c_fmderive-fund.
      get parameter id 'ZC3' field c_fmderive-fund.
    FREE MEMORY ID 'ZC1'.
      FREE MEMORY ID 'ZC2'.
       FREE MEMORY ID 'ZC3'.
    In this code i m facing memory problem.
    It is not refreshing the memory every time.
    So plz give me proper solution.
    Its urgent.
    Thanks
    Ranveer

    Hi,
       I suppose you are trying to store some particular value in memory in one program and then retieve it in another.
    If so try using EXPORT data TO MEMORY ID 'ZC1'. and IMPORT data FROM MEMORY ID 'ZC1'.
    To use SET PARAMETER/GET PARAMETER the specified parameter name should be in table TPARA. Which I don't think is there in your case.
    Sample Code :
    Data declarations for the function codes to be transferred
    DATA : v_first  TYPE syucomm,
           v_second TYPE syucomm.
    CONSTANTS : c_memid TYPE char10 VALUE 'ZCCBPR1'.
    Move the function codes to the program varaibles
      v_first  = gv_bdt_fcode.
      v_second = sy-ucomm.
    Export the function codes to Memory ID
    EXPORT v_first
           v_second TO MEMORY ID c_memid.        "ZCCBPR1  --- Here you are sending the values to memory
    Then retrieve it.
    Retrieve the function codes from the Memory ID
      IMPORT v_first  TO v_fcode_1
             v_second TO v_fcode_2
      FROM MEMORY ID c_memid.                                   "ZCCBPR1
      FREE MEMORY ID c_memid.                                   "ZCCBPR1
    After reading the values from memory ID free them your problem should be solved.
    Thanks
    Barada
    Edited by: Baradakanta Swain on May 27, 2008 10:20 AM

  • Memory Problems with Adobe PDF iFilter for 64-bit

    In preparation to rebuild my Windows Search Index, I installed the Adobe PDF iFilter for 64-bit on my system (Vista Business 64).  When I finally rebuilt the index, I wasn't too surprised by what I saw happen, namely, the SearchFilter.exe process would kick in whenever I wasn't using the system and just eat RAM.  One time I turned it on and it had allocated over 4,000 MB (and my system only has 4,030 MB available) so of course it was forcing all the other processes to hard fault (ie. everything was moving like molasses--for example, it took 20 minutes to put the thing to sleep).  But I just let it do it's work, figuring that perhaps this was to be expected relative to the small library of PDF's that I've accumulated on my computer, ranging from LaTeX generated text files, to containers for hi-res scans.  So, after a day and a half of basically not using my laptop, everything finally calmed down and I enjoyed the benefits of searching the content of my library from the Windows Start menu--for a short while.
    However, to my dismay I've encountered the problem that this freezing of my computer would now occur after everytime I download a new PDF (in this particular case they were Google Books scans) and then left the computer to idle.  Again, the SearchFilter.exe would allocate all of my RAM for itself and just push everything else onto the Virtual RAM, which means the SLOWEST possibly fetching you can get.  I had to uninstall as this was making my computer unusable for 15-30 minutes after each idle. Everything is back in working order without the iFilter, but I would like to know if anyone has reported such problems on x64 systems.  Obviously, I will also report the problem to Microsoft, since the search engine should certainly have the precaution to handle such memory problems.   However, it is a problem that is created by the Adobe PDF iFilter interacting with the Windows Search engine.

    Hello,
    We believe we have figured this out.  It looks like it has to do with the length of the default folder location for the Adobe iFilter.
    I was able to reproduce the issue and the following resolved it for me.  See if this resolves it for you all as well.
    Here is how to get Adobe Version 11 PDF filter to work.
     1 . If you haven’t already, run the following in SQL Server:
    Sp_fulltext_service ‘Load_os_resources’, 1
    Go
    --you might also need to run: 
    sp_fulltext_service ‘Verify_signature’,0  --This is used to validate trusted iFilters. 0 disables it. So use with caution.
    --go
    2. Stop SQL Server.  (Make sure FDHost.exe stops)
    3.  
    Uninstall the Adobe ifilter (because it defaulted to having spaces or the folder name is too long).
    4.  
    Reinstall the Adobe iFilter and when it prompts for where to install it, change it to: C:\Program Files\Adobe\PDFiFilter
    5.  Once the installation finishes, go the computer’s Environment variables. Add the following to the PATH.
    C:\Program Files\Adobe\PDFiFilter\BIN
    NOTE: it must include the BIN folder
    NOTE: If you had the OLD location that included spaces, remove it from the path environment variable.
    6. Start SQL Server
    7.  IF you had an existing Full-text index on PDFs, drop the full-text index and recreate it.
    8. You should now get results when you run sys.dm_fts_index_keywords('db','tblname')  --Note: Change db to be the actual database name and tblname to be the actual table name.
     Give this a try and see if this fixes yours. 
    Sincerely,
    Rob Beene, MSFT

  • Memory problems with PreparedStatements

    Driver: 9.0.1 JDBC Thin
    I am having memory problems using "PreparedStatement" via jdbc.
    After profiling our application, we found that a large number oracle.jdbc.ttc7.TTCItem objects were being created, but not released, even though we were "closing" the ResultSets of a prepared statements.
    Tracing through the application, it appears that most of these TTCItem objects are created when the statement is executed (not when prepared), therefore I would have assumed that they would be released when the ResultSet is close, but this does not seem to be the case.
    We tend to have a large number of PreparedStatement objects in use (over 100, most with closed ResultSets) and find that our application is using huge amounts of memory when compared to using the same code, but closing the PreparedStatement at the same time as closing the ResultSet.
    Has anyone else found similar problems? If so, does anyone have a work-around or know if this is something that Oracle is looking at fixing?
    Thanks
    Bruce Crosgrove

    From your mail, it is not very clear:
    a) whether your session is an HTTPSession or an application defined
    session.
    b) What is meant by saying: JSP/Servlet is growing.
    However, some pointers:
    a) Are there any timeouts associated with session.
    b) Try to profile your code to see what is causing the memory leak.
    c) Are there references to stale data in your application code.
    Marilla Bax wrote:
    hi,
    we have some memory - problems with the WebLogic Application Server
    4.5.1 on Sun Solaris
    In our Customer Projects we are working with EJB's. for each customer
    transaction we create a session to the weblogic application server.
    now there are some urgent problems with the java process on the server.
    for each session there were allocated 200 - 500 kb memory, within a day
    the JSP process on our server is growing for each session and don't
    reallocate the reserved memory for the old session. as a work around we
    now restart the server every night.
    How can we solve this problem ?? Is it a problem with the operating
    system or the application server or the EJB's ?? Do you have problems
    like this before ?
    greetings from germany,

  • Dual Channel memory problems on Neo 875P

    I am having considerable trouble getting the NEO 875P to work with dual channel DDR400 DIMMs. I have two Kingston KVR400X64C25 DDR400 256MB memory modules, each of which works correctly by itself. However, once I move one
    of the modules to channel B, I cannot boot my operating system (currently Windows XP.) The problem also occurs in Windows 2000, which I also have on the system. I receive blue screen errors that indicate paging/memory problems. I have upgraded my BIOS to the latest version (1.3)* in the hope that it would solve my problems, but it has not. The POST screen detects that there is 512MB of memory working in Dual Channel mode, but even when it does get through the O/S booting process (rarely), it crashes within minutes. Putting just one of the chips in makes the system perfectly stable. If anyone has any information, please let me know!
    Thanks,
    Jim Keller
    http://www.centerfuse.net
    * My board was dead after trying to flash the 1.3 BIOS, but after about 10 attempts of using CTRL+HOME method to recover the BIOS, it finally worked with version 1.0. I then re-flashed with 1.3, and it rebooted fine. Are there any updates about all of the problems with v1.3 ?

    >I can only answer based on what I've read on this forum
    I figured as much, I was just wondering if you knew whether MSI themselves frequents the forums or answers emails sufficiently.
    I've had an interesting development though. I stumbled upon Tom's Hardware Guide's article about 875P motherboards at http://www17.tomshardware.com/motherboard/20030519/i875p-01.html
    They mentioned that they needed to increase the memory voltage from 2.5 to 2.6 in the BIOS to get some memory modules to work. I did this, and for the first time, I am running in Dual channel DDR mode. However, the article also mentions that "the system does not run completely stably", so I guess I'll have to wait and see if it crashes again. However, I'm far enough beyond the XP splash screen to post this message, so I guess we're getting somewhere!
    -Jim Keller
    http://www.centerfuse.net

  • MS-6380 Memory Problem

    I have a MS-6380E with the latest BIOS and chipset drivers. Athlon 1.5 Ghz processor, Geo4 Ti series card with 64MB ram and latest drivers and an SBLive Platinum 5.1 card, with the latest drivers. Windows XP Pro is the operating system. The system has 2 PC2100 256MB sticks of DDR ram on the board. This past weekend I tried to install a 3rd chip, PC2100 512MB chip to bring the system up to a Gig of ram. The system would reboot immediatly upon logging into Windows. I took the memory back and exchanged it for another. Installed it with similar results. I tried all three slots with the new memory and the two existing. No, I didn't try the new memory by itself. I finally go it to appear stable for a little bit but I also play Dark Age of Camelot and the latest symptoms were that the system would kick me back to the desktop as soon as I got to the initial "connecting" screen in Dark Age. I also had other odd behaviour with the additional ram installed like my modem taking longer to validate my connection. When I went back to the pre-existing configuration 512MB total, everything would stabilize. The DDR brand was Centron, I know it's not the same manufacturor as my other two sticks but I don't recall the brand on those.  Also, just in case you needed to know Dark Age of Camelot is an MMORPG. I checked their tech section and nothing on this problem and their game. I've since returned the memory and have resigned to maxing out at 512MB tho the system specs say I can go to 3GB. I really like this board but am frustrated at the memory problem.

    http://www.msi.com.tw/program/service/forum/search.php?boardid=
    6380 covers a hell of a lot of boards
    getting same make and model of ram is generally advisable to

  • Nokia C5-03 Low memory problems ( Other Files )

    Please help. I have a nokia C5-03 the handset is showing phone memory is full. I have checked and uninstalled all unecessay applications. But problem is still there. when i check the phone memory it indicates that i have other files of 42mb installed and using phone memory. however i am unable to check what other files are installed or classfied as other files. Can you please assist me so i can know what these other files consist of or where they are located.

    Hi Thetao,
    I typed a more extensive reaction before, but it got lost when I pressed "post". Therefore I just respond to the main points that you mentioned (and some I found out myself).
    Strange: I can't find the 40 MB Maximum User Storage on the Nokia website anymore (nor the 75 MB). But it sounds very familiar to me. It looks to me as if they removed this from the phone specs, also of other Smartphones by the way.
    Yesterday, I deleted some small apps that I don't use (anymore) such as InternetRadio and I also removed Nokia email. Although the apps were below 2 MB together, this freed up over 7 MB of Phone memory (24 MB free now)! I think there were still some old emails stored in C: which I couldn't delete any other way. This helped me a great deal already but I tried your suggestions as well.
    1. No map data or CITIES folder on C: 2. Switched messages memory to phone (and phone to offline mode) and I did indeed find a forgotten email account with 30 email messages. Not much but I had 24,7 MB free after that. Of course, I put messages memory back to the memorycard. 3. Used the free edition of Y-Browser to manually delete the cache folder. Not much data in that, but 25,1 MB free after that. Nice tool, with which you can reach folders that normally stay hidden! Used YBrowser to search all C: for files over 300 kB. Only 2 files: boot_space.txt in C:\ (500 kB, contains only the X repeatedly as far as I see, but is probably essential for the operating system) and C:\resource\python25\pyton25.zip (1 MB). It looks like an installation package, but I'm not sure if I can delete it. By the way: YBrowser hasn't made a shortcut in one of the menus. Only way that I found to start it was to look for it using the Phone's search function. Is there a way to make this shortcut myself?
    4. Yes I did. No Images folder on C: anymore, nor other big files (see point 3)
    5. I use Bluetooth for file transfer sometimes, mainly for installation files (such as YBrowser.sis, but I did this one via USB-cable). However, no big files are left on C: so I don't think I have this problem.
    6. I tried to delete Nokia Chat yesterday as well (with the other apps), but it won't be uninstalled the normal way as it says "Uninstall cancelled" (not sure about the exact translation since my phone 'speaks' Dutch) Do you know if there's another way to get rid of this 3 MB app that I don't use at all?
    I think I may have found an explanation and a solution for the memory problem while navigating. You mentioned the "memory in use" in the map settings. Above that option there's a slide bar for the % of memory that the navigation can use. Standard is 70%. I always thought this was about storage memory on (in my case) the memorycard. Another topic mentioned that this the working memory (so the RAM) that the navigation may use. Setting it to 70% means there's only 30% left for other apps that run in the background. The other topic states that this is nog enough so the slider should be set to for instance 30% for navigation leaving 70% free for "the phone". From behind my computer, navigation seems much more stable. I'll try this setting in my car soon and let you know how it works.
    Thanks a lot for thinking along with me so far! There's already 25,1 MB of space, which is great since it was only 7 MB last Sunday. And navigation looks more stable. I'd appreciate if you have some more answers to my latest questions, but if not I think my phone will work a lot better already!
    Regards, Paul

  • Memory problem with jdk/jre 1.1.8

    My name is BERGMANN Yannick.
    I'm working for IRM in Li?ge and we developped an application (user interface for an industrial measurement system) in Java (JDK/JRE version : 1.1.8).
    We have a big memory problem with this application :
    - This user interface is running on a WINDOWS NT PC with 128MB.
    - This is the command to lanch our application :
    C:\Program Files\JavaSoft\JRE\1.1\bin\jrew.exe" -ms32m -mx32m -cp "\Program Files\HMI\HMI.zip;\Velocis\Add_On\Jdbc\raima.jar;\Program Files\Swing-1.1.1\swingall.jar" be.irm.hmi.kernel.HMI -t15 -d"Velocis rdstcp" -newdb -mf1m -mr20
    - When our application is running, everything seems to be OK in memory for it. The garbage collector seems to work properly and our application has always at least 5MB free memory (We use the java instruction "Runtime.getRuntime().freeMemory()" to know this).
    - But when we look in the "Windows NT task manager" for the "jrew" application, the memory increases ALWAYS.
    - After 5 days our application is completely frozen and blocked ...???
    - here is a memory map of our Windows NT PC :
         "jre.exe"     "commit total"     "commit limit"     "commit peak"     "physical total"     "physical available"     "physical file cache"     
    Monday     92264     109256     194944     109424     130484     19492     6216     
    Thuesday     106196     123072     194944     123348     130484     6072     5840     
    Wednesday     110836     132288     194944     132416     130484     4408     5140     
    Thursday     108200     144980     194944     145140     130484     4888     5148     
    Friday     109440     158319     194944     161334     130484     4911     4992     
    Monday     111600     209060     228548     209148     130484     5184     3484     
    Have you any idea of what is happening with "jrew" in memory ?
    We have had this problem for six month and we are totaly out of idea.
    If you can give us any idea, we'll appreciate a lot.
    Thanks in advance,
    BERGMANN Yannick
    IRM SA - Software Engineer
    Tel. (32)4/239.90.10
    Tel. (32)4/239.90.74 (direct)
    Fax (32)4/263.40.97
    E-mail [email protected]

    We had a memory problem with a swing applet in our company. The major reason for this was that we added new components in a JTree and removed them later again, and the components we removed were never garbage collected. This was because with these components we added different listeners, and we didn't remove the listeners after we didn't need the components anymore. After we corrected this, the components where garbage collected.
    Perhaps it's a similar problem you have, or I have no idea. Check that you remove actionlisteners, mouselisteners etc from components you want to be garbage collected.
    You could also test your application with OptimizeIt to see what objects you create and how many you get of them over time: http://www.vmgear.com

Maybe you are looking for