Sudden increase in native heap memory

Hi all,
we are using JRockit 26.4 and currently configured with 1800MB heap on windows 2003 with /3GB switch.
We noticed that after the weblogic server is started the total memory (java heap + native memory) lingers around 2550 MB... (We monitored this using perfmon -> virtual bytes of java process)
after couple of days run, the native memory suddenly increased by 300 MB, taking the total virtual bytes to 2850 MB.
Do we know what could be the reason for such a sudden increase?
Thanks,
- Pritam.

I don't believe this is documented anywhere unfortunately, and it might vary depending on the exact JRockit release. But the key thing to watch out for is growth, regardless of which component it is. If you see that, post a question here and we'll find someone to tell you what the issue is...
That said, here's an example printout and some guesses from my side.
[JRockit] memtrace is collecting data...
[JRockit] *** 0th memory utilization report
(all numbers are in kbytes)
Total mapped                         ;;;;;;; 860976
; Total in-use                        ;;;;;;  89816
;;  executable                         ;;;;;   4400
;;;   java code                         ;;;;    384;   8.7%
;;;;    used                             ;;;    236;   61.6%
;;  shared modules (exec+ro+rw)        ;;;;;   8496
;;  guards                             ;;;;;    180
;;  readonly                           ;;;;;   3052
;;  rw-memory                          ;;;;;  78132
;;;   Java-heap                         ;;;;  65536;   83.9%
;;;   Stacks                            ;;;;   1976;   2.5%
;;;   Native-memory                     ;;;;  10619;   13.6%
;;;;    java-heap-overhead               ;;;   2056
;;;;    codegen memory                   ;;;    576
;;;;    classes                          ;;;   2048;   19.3%
;;;;;     method bytecode                 ;;    219
;;;;;     method structs                  ;;    277    (#5929)
;;;;;     constantpool                    ;;    788
;;;;;     classblock                      ;;     69
;;;;;     class                           ;;    143    (#443)
;;;;;     other classdata                 ;;    338
;;;;;     overhead                        ;;     91
;;;;    threads                          ;;;      6;   0.1%
;;;;    malloc:ed memory                 ;;;   1445;   13.6%
;;;;;     codeinfo                        ;;     60
;;;;;     codeinfotrees                   ;;     24
;;;;;     exceptiontables                 ;;      5
;;;;;     metainfo/livemaptable           ;;    243
;;;;;     codeblock structs               ;;      0
;;;;;     constants                       ;;      0
;;;;;     livemap global tables           ;;    135
;;;;;     callprof cache                  ;;      0
;;;;;     paraminfo                       ;;     30    (#506)
;;;;;     strings                         ;;    426    (#8402)
;;;;;     strings(jstring)                ;;      0
;;;;;     typegraph                       ;;     27
;;;;;     interface implementor list      ;;      5
;;;;;     thread contexts                 ;;      8
;;;;;     jar/zip memory                  ;;    497
;;;;;     native handle memory            ;;      5
;;;;    unaccounted for memory           ;;;   4493;   42.3%;3.11
---------------------!!!Total mapped means mapped virtual memory, eg the largest size the JVM can be expected to grow to. This is typically on the same order of magnitude as the Java heap size (Xmx).
Total in-use is the actual memory in use, this is the current "memory footprint" of your process.
executable refers to JIT compiled code, I believe.
Java-heap is the heap (duh).
Stacks should be the thread stacks (where local variables are stored).
classes is native memory used to store bytecode etc. If this grows, it may be because you are generating classes dynamically and they are not getting cleaned up because there are still live instances. Or maybe you're running with -Xnoclassgc.
threads can grow if you start a lot of threads and don't let them die.
Stuff under malloc:ed memory is mainly misc JVM metadata.
unaccounted for memory is...everything else. The most common leak involving unaccounted-for is native byte buffers (DirectByteBuffer).
-- Henrik

Similar Messages

  • Changing heap memory paramter  and Additonal Java node configuration

    Hi All,
    Kindly provide me the procedure to change the heap memory parameter and location (in Visual admin). How to install/configure 2nd Java server node in PI system.
    Thanks in advance.

    Hi Ramesh,
    You would have posted this question in Netweaver Administrator forum...
    Well here are the answers.
    Login to Configtool.
    Expand Cluster-Data->instance<xx>->dispatcher. On the right hand side you can change the heap size for the dispatcher
    Follow the same procedure to increase the server heap memory.
    To add the server node click on instance<xx>. Then click Server->Add server.
    Cheers....
    Raghu

  • Can I increase heap memory without specify any class or jar file??

    Hi,
    I tried to increase my heap memory in this way :
    java -Xms256m -Xmx256m
    but I got an error ... it's seem that I must specify a class java or a .jar file ...
    This is the error :
    Usage: java [-options] class [args...]
    (to execute a class)
    or java [-options] -jar jarfile [args...]
    (to execute a jar file)
    where options include:
    -client to select the "client" VM
    -server to select the "server" VM
    -hotspot is a synonym for the "client" VM [deprecated]
    The default VM is client.
    -cp <class search path of directories and zip/jar files>
    -classpath <class search path of directories and zip/jar files>
    A ; separated list of directories, JAR archives,
    and ZIP archives to search for class files.
    -D<name>=<value>
    set a system property
    -verbose[:class|gc|jni]
    enable verbose output
    -version print product version and exit
    -version:<value>
    require the specified version to run
    -showversion print product version and continue
    -jre-restrict-search | -jre-no-restrict-search
    include/exclude user private JREs in the version search
    -? -help print this help message
    -X print help on non-standard options
    -ea[:<packagename>...|:<classname>]
    -enableassertions[:<packagename>...|:<classname>]
    enable assertions
    -da[:<packagename>...|:<classname>]
    -disableassertions[:<packagename>...|:<classname>]
    disable assertions
    -esa | -enablesystemassertions
    enable system assertions
    -dsa | -disablesystemassertions
    disable system assertions
    -agentlib:<libname>[=<options>]
    load native agent library <libname>, e.g. -agentlib:hprof
    see also, -agentlib:jdwp=help and -agentlib:hprof=help
    -agentpath:<pathname>[=<options>]
    load native agent library by full pathname
    -javaagent:<jarpath>[=<options>]
    load Java programming language agent, see java.lang.instrument
    -splash:<imagepath>
    show splash screen with specified image
    can I increase heap memory without specify any class or jar file??
    thx

    chiara wrote:
    Hi,
    I tried to increase my heap memory in this way :
    java -Xms256m -Xmx256m
    but I got an error ... it's seem that I must specify a class java or a .jar file ...
    This is the error :
    Usage: java [-options] class [args...]
    (to execute a class)
    or java [-options] -jar jarfile [args...]
    (to execute a jar file)
    can I increase heap memory without specify any class or jar file??The job of java.exe is to execute java bytecode.
    What is it supposed to do with your request to use 256m of memory for heap
    when you are not giving it a class or a jar to run?

  • Sudden increase in memory space

    We have developed an application using the Hyperion 9.3.1.
    We are trying to load data and the data load itself is taking lot of time.
    Again, the memory space is getting increased drastically (almost to 100%).
    We see lot of page files being generated.
    What can be the reason for this sudden increase in memory space utilization.

    you should consider the folllowing factors for sudden increase in memory space
    1.Minimize the no of hierarchies otherwise it leads to increase in the size of aggregated data and slows down view selection
    2. remove fragmentation
    3.Compact the outline to remove the records of members that we have been deleted
    4.create transparent partition if neccessary which leads to enable write-back partition.By doing this we can reducethe size of the Database and reduces calculation time.
    5.use compression if values are grouped together in consecutive members on dimensions or hierarchies rather than spread throughout the outline with lots of
    #missing data between values.

  • How to increase jvm heap memory

    hello all,
    we r facing "outofmemory" exception. could any one kindly help us in increasing the jvm's memory in runtime.
    thanks alot
    raghu

    use java with -Xms and -Xmx options.
    see http://java.sun.com/j2se/1.4/docs/tooldocs/windows/java.html
    nk

  • How to find out what is using the native heap of a process running a JVM?

    Hello,
    I am not sure where to post this question so I am starting here.
    I am troubleshooting a Java application using some native calls (32 bits Java running on Solaris 10). The size of the process (as reported by prstat) is slowly increasing day after day.
    The size of the 'Java heap' is fixed at the start (-Xms and -Xmx are set to the same value on the command line when launching the Java app) and the GC is workling fine. No memory complaints from the Java side of the application.
    It is the size of the 'native' heap (as reported by 'pmap') that is increasing:
    root@mas01 # pmap 5382
    5382:/apps/java/bin/java -server -Xms207M -Xmx207M -XX:MaxNewSize=24M -XX:N
    00010000 64K r-x-- /apps/jdk1.5.0_19/bin/java
    0002E000 16K rwx-- /apps/jdk1.5.0_19/bin/java
    00032000 3896K rwx-- [ heap ]
    00400000 389120K rwx-- [ heap ]
    18000000 2784K rwx-- [ heap ]
    DCAF4000 48K rw--R [ stack tid=169 ]
    DCBF6000 40K rw--R [ stack tid=161 ]
    DCCF8000 32K rw--R [ stack tid=160 ]
    My first reaction was to search for a memory leak. Found a minor leak in the JVM with the ::findleak function (called within the mdb debugger). Upgraded to a later release of Java 5 (Java 1.5.0_19) where the leak is fixed but the heap is still increasing.
    Many parts of the process allocate memory in the native heap. The JVM itself, the native calls made to a C++ library part of our Java application and maybe also some 3rd party software.
    I would like to know what is the best way to find out what is consuming more and more memory in the native heap. I started looking a DTraces but I am new to this. I also thought maybe the Solaris Perftools might be of use but I never used them. Before plunging into a tool more or less blindly, I am asking for advices on how to tackle this issue. Can someone recommend a tool/method to see what is allocated in the heap?
    Regards,
    Stéphan
    Edited by: StephanDupont on Sep 22, 2009 8:47 AM

    After googling a lot I managed to run my application with libumem, generated a core file and succeeded to find some leak with mdb even if ::findleak reported nothing.
    Does anyone knows if the ::findleak (you need libumem and mdb) is supposed to find leak in the native part of the memory and a Java application using the JNI interface?
    Regards,
    Stéphan

  • ODSEE 11g - DPS Directory proxy server suddenly increase load average

    Hi all
    Recently upgraded from directory server 5.2 to ODSEE 11g, one directory proxy configure to one master directory server and one consumer directory server.
    all the three instances are in the same sparc t3 machine.
    Directory proxy server alerts server load average on the machine is above 6.00 normally it is 0.66. I'm not sure what is causing the sudden burst in the load ? the traffic is normal there is no abnormal requests coming to the server. proxy performance degrades over the span of 24 hours ....and Once i restart the proxy services (dpsadm restart) all load averages comes to normal and directory proxy runs normal for the next two - three weeks. The same cycle continues ...I'm not sure what was causing the sudden load increase.
    I increased the JVM heap size from 1GB to 2 GB still continue to have the problem ..did anyone else experience similar problem. How did you fix it....
    Any input or advise in the right direction is much appreciated.
    Thank you.

    server load I'm referring to "prstat command" - server load average suddenly shoot up from 0.66 to 6.00 ie) the CPU usage. Alert is from our server monitoring tool not related to directory proxy.
    Clients report connections time out (etime goes from etime=0 ..2..4.....) over 24 hours i can see the etime increases and eventually the proxy server get hung and non responsive. Once I restart all the performance back to normal at-least for another two weeks.
    I suspect there might be a memory leak or JVM Garbage collection issue -- any expert input how to figure this out will help.
    Here is the JVM args in the proxy server "Xms2g -Xmx2g -Xmn1g -XX:SurvivorRatio=4 -XX:+UseParNewGC -XX:+UseConcMarkSweepGC"
    Here is a jstat during the problem
    ./jstat -gcutil -t 25365 2s 30
    Timestamp S0 S1 E O P YGC YGCT FGC FGCT GCT
    982106.4 0.00 26.17 4.26 92.25 59.52 523 60.979 689 1002.587 1063.566
    982108.4 0.00 26.17 4.40 92.25 59.52 523 60.979 689 1002.587 1063.566
    982110.4 0.00 26.17 4.80 92.25 59.52 523 60.979 689 1002.587 1063.566
    982112.4 0.00 26.17 5.10 92.25 59.52 523 60.979 690 1002.719 1063.698
    982114.4 0.00 26.17 5.15 92.25 59.52 523 60.979 690 1002.719 1063.698
    982116.4 0.00 26.17 5.32 92.25 59.52 523 60.979 691 1003.009 1063.988
    982118.4 0.00 26.17 5.72 92.25 59.52 523 60.979 691 1003.009 1063.988
    982120.4 0.00 26.17 5.80 92.25 59.52 523 60.979 691 1003.009 1063.988
    982122.4 0.00 26.17 5.93 92.25 59.52 523 60.979 692 1003.168 1064.146
    982124.4 0.00 26.17 6.03 92.25 59.52 523 60.979 692 1003.168 1064.146
    982126.4 0.00 26.17 6.15 92.25 59.52 523 60.979 693 1003.481 1064.460
    982128.5 0.00 26.17 6.18 92.25 59.52 523 60.979 693 1003.481 1064.460
    982130.5 0.00 26.17 6.25 92.25 59.52 523 60.979 693 1003.481 1064.460
    982132.5 0.00 26.17 6.29 92.25 59.52 523 60.979 694 1003.656 1064.635
    982134.5 0.00 26.17 6.31 92.25 59.52 523 60.979 694 1003.656 1064.635
    982136.5 0.00 26.17 6.36 92.25 59.52 523 60.979 695 1003.988 1064.967
    982138.5 0.00 26.17 6.89 92.25 59.52 523 60.979 695 1003.988 1064.967
    982140.5 0.00 26.17 6.99 92.25 59.52 523 60.979 695 1003.988 1064.967
    982142.5 0.00 26.17 7.08 92.25 59.52 523 60.979 696 1004.187 1065.165
    982144.5 0.00 26.17 7.31 92.25 59.52 523 60.979 696 1004.187 1065.165
    982146.5 0.00 26.17 7.82 92.25 59.52 523 60.979 697 1004.553 1065.531
    982148.5 0.00 26.17 7.92 92.25 59.52 523 60.979 697 1004.553 1065.531
    982150.5 0.00 26.17 8.01 92.25 59.52 523 60.979 697 1004.553 1065.531
    982152.5 0.00 26.17 8.17 92.25 59.52 523 60.979 698 1004.786 1065.764
    982154.5 0.00 26.17 8.26 92.25 59.52 523 60.979 698 1004.786 1065.764
    982156.5 0.00 26.17 8.38 92.25 59.52 523 60.979 699 1005.174 1066.153
    982158.5 0.00 26.17 8.74 92.25 59.52 523 60.979 699 1005.174 1066.153
    982160.5 0.00 26.17 8.88 92.25 59.52 523 60.979 699 1005.174 1066.153
    982162.5 0.00 26.17 8.96 92.25 59.52 523 60.979 700 1005.433 1066.412
    982164.5 0.00 26.17 9.09 92.25 59.52 523 60.979 700 1005.433 1066.412
    jstat after the restart
    ./jstat -gcutil -t 10084 2s 30
    Timestamp S0 S1 E O P YGC YGCT FGC FGCT GCT
    40312.6 0.00 25.13 88.49 1.98 63.68 21 2.366 0 0.000 2.366
    40314.6 0.00 25.13 88.58 1.98 63.68 21 2.366 0 0.000 2.366
    40316.6 0.00 25.13 88.71 1.98 63.68 21 2.366 0 0.000 2.366
    40318.6 0.00 25.13 88.99 1.98 63.68 21 2.366 0 0.000 2.366
    40320.6 0.00 25.13 89.31 1.98 63.68 21 2.366 0 0.000 2.366
    40322.6 0.00 25.13 89.36 1.98 63.68 21 2.366 0 0.000 2.366
    40324.6 0.00 25.13 89.42 1.98 63.68 21 2.366 0 0.000 2.366
    40326.6 0.00 25.13 89.53 1.98 63.68 21 2.366 0 0.000 2.366
    40328.6 0.00 25.13 89.60 1.98 63.68 21 2.366 0 0.000 2.366
    40330.6 0.00 25.13 89.72 1.98 63.68 21 2.366 0 0.000 2.366
    40332.6 0.00 25.13 90.11 1.98 63.68 21 2.366 0 0.000 2.366
    40334.6 0.00 25.13 90.56 1.98 63.68 21 2.366 0 0.000 2.366
    40336.6 0.00 25.13 90.67 1.98 63.68 21 2.366 0 0.000 2.366
    40338.6 0.00 25.13 90.75 1.98 63.68 21 2.366 0 0.000 2.366
    40340.6 0.00 25.13 91.09 1.98 63.68 21 2.366 0 0.000 2.366
    40342.6 0.00 25.13 91.36 1.98 63.68 21 2.366 0 0.000 2.366
    40344.6 0.00 25.13 91.47 1.98 63.68 21 2.366 0 0.000 2.366
    40346.6 0.00 25.13 91.53 1.98 63.68 21 2.366 0 0.000 2.366
    40348.7 0.00 25.13 91.64 1.98 63.68 21 2.366 0 0.000 2.366
    40350.7 0.00 25.13 91.77 1.98 63.68 21 2.366 0 0.000 2.366
    40352.7 0.00 25.13 91.87 1.98 63.68 21 2.366 0 0.000 2.366
    40354.7 0.00 25.13 91.95 1.98 63.68 21 2.366 0 0.000 2.366
    40356.7 0.00 25.13 92.11 1.98 63.68 21 2.366 0 0.000 2.366
    40358.7 0.00 25.13 92.19 1.98 63.68 21 2.366 0 0.000 2.366
    40360.7 0.00 25.13 92.24 1.98 63.68 21 2.366 0 0.000 2.366
    40362.7 0.00 25.13 92.85 1.98 63.68 21 2.366 0 0.000 2.366
    40364.7 0.00 25.13 93.19 1.98 63.68 21 2.366 0 0.000 2.366
    40366.7 0.00 25.13 93.40 1.98 63.68 21 2.366 0 0.000 2.366
    40368.7 0.00 25.13 93.44 1.98 63.68 21 2.366 0 0.000 2.366
    40370.7 0.00 25.13 93.47 1.98 63.68 21 2.366 0 0.000 2.366
    Any one else had similar behavior. Any input to the right direction is highly appreciated.
    Thanks.

  • Any way to increase the default Heap size for all Java VMs in Solaris 8

    Hello,
    I have a java product that deals with large databases under Solaris 8. It is a jar file, started by a cron job every night. Some nights it will fail because it runs out of Heap memory depending on the amount of records it has to deal with. I know that I could increase the java VM heap size with "java -jar -mx YY JARFILE" command but I have other java products that are showing the same behavior, and I would like to correct them all in one shot if possible.
    What I would like to do is find a system or configuration parameter that forces all Java VMs to use a larger MAX Heap size than the default 16M specified in the Man page for Java. Is there a way to accomplish that?
    TIA
    Maizo

    You could always download the source and modify it.

  • Can off JVM heap Memory used in the Near-Cache front-tier

    I had tried to config a near-Cache used nio-manager(off JVM heap) in the Front-tier.
    <near-scheme>
          <scheme-name>CohApp-near</scheme-name>
          <front-scheme>
            <external-scheme>
            </external-scheme>
          </front-scheme>
          <back-scheme>
            <distributed-scheme>
              <scheme-ref>CohApp-distributed</scheme-ref>
            </distributed-scheme>
          </back-scheme>
          <invalidation-strategy>auto</invalidation-strategy>
          <autostart>true</autostart>
        </near-scheme>
    when start 'com.tangosol.net.DefaultCacheServer' for this config, error as:
    Oracle Coherence Version 3.7.1.0 Build 27797
    Enterprise Edition: Development mode
    Copyright (c) 2000, 2011, Oracle and/or its affiliates. All rights reserved.
    2014-03-30 16:34:17.518/1.201 Oracle Coherence EE 3.7.1.0 <Error> (thread=main,
    member=n/a): Error org.xml.sax.SAXParseException: cvc-complex-type.2.4.a: Invali
    d content was found starting with element 'external-scheme'. One of '{"http://xm
    lns.oracle.com/coherence/coherence-cache-config":local-scheme, "http://xmlns.ora
    cle.com/coherence/coherence-cache-config":class-scheme}' is expected. - line 92
    Exception in thread "main" (Wrapped: Failed to load the factory) (Wrapped: Missi
    ng or inaccessible constructor "com.tangosol.net.DefaultConfigurableCacheFactory
    (String)"
    <configurable-cache-factory-config>
      <class-name>com.tangosol.net.DefaultConfigurableCacheFactory</class-name>
      <init-params>
        <init-param>
          <param-type>java.lang.String</param-type>
          <param-value>coherence-cache-config.xml</param-value>
        </init-param>
      </init-params>
    </configurable-cache-factory-config>) java.lang.reflect.InvocationTargetExceptio
    n
            at com.tangosol.util.Base.ensureRuntimeException(Base.java:288)
            at com.tangosol.net.ScopedCacheFactoryBuilder.getDefaultFactory(ScopedCa
    cheFactoryBuilder.java:311)
            at com.tangosol.net.DefaultCacheFactoryBuilder.getSingletonFactory(Defau
    ltCacheFactoryBuilder.java:48)
            at com.tangosol.net.DefaultCacheFactoryBuilder.getFactory(DefaultCacheFa
    ctoryBuilder.java:121)
            at com.tangosol.net.ScopedCacheFactoryBuilder.getConfigurableCacheFactor
    y(ScopedCacheFactoryBuilder.java:112)
            at com.tangosol.net.CacheFactory.getConfigurableCacheFactory(CacheFactor
    y.java:126)
            at com.tangosol.net.DefaultCacheServer.getDefaultConfigurableCacheFactor
    y(DefaultCacheServer.java:364)
            at com.tangosol.net.DefaultCacheServer.main(DefaultCacheServer.java:197)
    Caused by: (Wrapped: Missing or inaccessible constructor "com.tangosol.net.Defau
    ltConfigurableCacheFactory(String)"
    <configurable-cache-factory-config>
      <class-name>com.tangosol.net.DefaultConfigurableCacheFactory</class-name>
      <init-params>
        <init-param>
          <param-type>java.lang.String</param-type>
          <param-value>coherence-cache-config.xml</param-value>
        </init-param>
      </init-params>
    </configurable-cache-factory-config>) java.lang.reflect.InvocationTargetExceptio
    n
            at com.tangosol.util.Base.ensureRuntimeException(Base.java:288)
            at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2652)
            at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2536)
            at com.tangosol.net.ScopedCacheFactoryBuilder.getDefaultFactory(ScopedCa
    cheFactoryBuilder.java:273)
            ... 6 more
    Caused by: java.lang.reflect.InvocationTargetException
            at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
            at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstruct
    orAccessorImpl.java:39)
            at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingC
    onstructorAccessorImpl.java:27)
            at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
            at com.tangosol.util.ClassHelper.newInstance(ClassHelper.java:694)
            at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2611)
            ... 8 more
    Caused by: (Wrapped: Failed to load cache configuration: coherence-cache-config.
    xml) (Wrapped) java.io.IOException: Exception occurred during schema validation:
    cvc-complex-type.2.4.a: Invalid content was found starting with element 'externa
    l-scheme'. One of '{"http://xmlns.oracle.com/coherence/coherence-cache-config":l
    ocal-scheme, "http://xmlns.oracle.com/coherence/coherence-cache-config":class-sc
    heme}' is expected.
            at com.tangosol.util.Base.ensureRuntimeException(Base.java:288)
            at com.tangosol.run.xml.XmlHelper.loadResourceInternal(XmlHelper.java:34
    1)
            at com.tangosol.run.xml.XmlHelper.loadFileOrResource(XmlHelper.java:283)
            at com.tangosol.net.DefaultConfigurableCacheFactory.loadConfig(DefaultCo
    nfigurableCacheFactory.java:439)
            at com.tangosol.net.DefaultConfigurableCacheFactory.loadConfig(DefaultCo
    nfigurableCacheFactory.java:425)
            at com.tangosol.net.DefaultConfigurableCacheFactory.<init>(DefaultConfig
    urableCacheFactory.java:155)
            ... 14 more
    Caused by: (Wrapped) java.io.IOException: Exception occurred during schema valid
    ation:
    cvc-complex-type.2.4.a: Invalid content was found starting with element 'externa
    l-scheme'. One of '{"http://xmlns.oracle.com/coherence/coherence-cache-config":l
    ocal-scheme, "http://xmlns.oracle.com/coherence/coherence-cache-config":class-sc
    heme}' is expected.
            at com.tangosol.run.xml.XmlHelper.loadXml(XmlHelper.java:122)
            at com.tangosol.run.xml.XmlHelper.loadXml(XmlHelper.java:157)
            at com.tangosol.run.xml.XmlHelper.loadResourceInternal(XmlHelper.java:32
    2)
            ... 18 more
    Caused by: java.io.IOException: Exception occurred during schema validation:
    cvc-complex-type.2.4.a: Invalid content was found starting with element 'externa
    l-scheme'. One of '{"http://xmlns.oracle.com/coherence/coherence-cache-config":l
    ocal-scheme, "http://xmlns.oracle.com/coherence/coherence-cache-config":class-sc
    heme}' is expected.
            at com.tangosol.run.xml.SimpleParser.parseXml(SimpleParser.java:212)
            at com.tangosol.run.xml.SimpleParser.parseXml(SimpleParser.java:93)
            at com.tangosol.run.xml.SimpleParser.parseXml(SimpleParser.java:162)
            at com.tangosol.run.xml.SimpleParser.parseXml(SimpleParser.java:115)
            at com.tangosol.run.xml.XmlHelper.loadXml(XmlHelper.java:118)
            ... 20 more
    Caused by: org.xml.sax.SAXParseException: cvc-complex-type.2.4.a: Invalid conten
    t was found starting with element 'external-scheme'. One of '{"http://xmlns.orac
    le.com/coherence/coherence-cache-config":local-scheme, "http://xmlns.oracle.com/
    coherence/coherence-cache-config":class-scheme}' is expected.
            at com.sun.org.apache.xerces.internal.util.ErrorHandlerWrapper.createSAX
    ParseException(ErrorHandlerWrapper.java:195)
            at com.sun.org.apache.xerces.internal.util.ErrorHandlerWrapper.error(Err
    orHandlerWrapper.java:131)
            at com.sun.org.apache.xerces.internal.impl.XMLErrorReporter.reportError(
    XMLErrorReporter.java:384)
            at com.sun.org.apache.xerces.internal.impl.XMLErrorReporter.reportError(
    XMLErrorReporter.java:318)
            at com.sun.org.apache.xerces.internal.impl.xs.XMLSchemaValidator$XSIErro
    rReporter.reportError(XMLSchemaValidator.java:417)
            at com.sun.org.apache.xerces.internal.impl.xs.XMLSchemaValidator.reportS
    chemaError(XMLSchemaValidator.java:3182)
            at com.sun.org.apache.xerces.internal.impl.xs.XMLSchemaValidator.handleS
    tartElement(XMLSchemaValidator.java:1806)
            at com.sun.org.apache.xerces.internal.impl.xs.XMLSchemaValidator.startEl
    ement(XMLSchemaValidator.java:705)
            at com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.scan
    StartElement(XMLNSDocumentScannerImpl.java:400)
            at com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImp
    l$FragmentContentDriver.next(XMLDocumentFragmentScannerImpl.java:2756)
            at com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl.next(X
    MLDocumentScannerImpl.java:648)
            at com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.next
    (XMLNSDocumentScannerImpl.java:140)
            at com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImp
    l.scanDocument(XMLDocumentFragmentScannerImpl.java:511)
            at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(X
    ML11Configuration.java:808)
            at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(X
    ML11Configuration.java:737)
            at com.sun.org.apache.xerces.internal.jaxp.validation.StreamValidatorHel
    per.validate(StreamValidatorHelper.java:144)
            at com.sun.org.apache.xerces.internal.jaxp.validation.ValidatorImpl.vali
    date(ValidatorImpl.java:111)
            at javax.xml.validation.Validator.validate(Validator.java:127)
            at com.tangosol.run.xml.SaxParser.validateXsd(SaxParser.java:236)
            at com.tangosol.run.xml.SimpleParser.parseXml(SimpleParser.java:206)
    So I think that if can off JVM heap Memory used in the Near-Cache front-tier?
    Or can help how to config the  off JVM heap Memory used in the Near-Cache front-tier.
    Thanks.

    Only local-scheme and class-scheme can be used in the front-scheme of a near cache. 

  • OAS Heap memory issue: An error "java.lang.OutOfMemoryError: GC overhead

    OAS - 10.1.3.4.0
    We are running out of Heap memory and seeing lots of full GC and out of memory events
    Verbose GC is on.
    Users don't know what they are doing to cause this
    We have 30-40 users per server and 1.5 GB heap memory allocated
    There are no other applications on the machine. Only the PRD instance with 1.5 GB allocated to the JVM. We do not have any issue with memory on the server and we could increase the heap but we dont want to go over the 1.5 GB since that is what I understood to be the high end of what is recommended. we only have 30-40 users on each machine. There are 8 servers and a typical heavy usage day we may have 1 or two machines that have the out of memory or continuous full GC in the logs. When this occurs the phones light up with the people on that machine experiencing slowness.
    below is an example of what we see in a file created in the OPMN log folder on the JAS server then this occurs. I think this is the log created when Verbose GC is turned on. I can send you the full log or anything else you need. Thanks
    1194751K->1187561K(1365376K), 4.6044738 secs]
    java.lang.OutOfMemoryError: GC overhead limit exceeded
    Dumping heap to java_pid10644.hprof ...
    [Full GC 1194751K->1188321K(1365376K), 4.7488200 secs]
    Heap dump file created [1326230812 bytes in 47.602 secs]
    [Full GC 1194751K->1177641K(1365376K), 5.6128944 secs]
    [Full GC 1194751K->986239K(1365376K), 4.6376179 secs]
    [Full GC 1156991K->991906K(1365376K), 4.5989155 secs]
    [Full GC 1162658K->1008331K(1365376K), 4.1139016 secs]
    [Full GC 1179083K->970476K(1365376K), 4.9670050 secs]
    [GC 1141228K->990237K(1365376K), 0.0561096 secs]
    [GC 1160989K->1012405K(1365376K), 0.0920553 secs]
    [Full GC 1012405K->1012274K(1365376K), 4.1170216 secs]
    [Full GC 1183026K->1032000K(1365376K), 4.4166454 secs]
    [Full GC 1194739K->1061736K(1365376K), 4.4009954 secs]
    [Full GC 1194739K->1056175K(1365376K), 5.1124431 secs]
    [Full GC 1194752K->1079807K(1365376K), 4.5160851 secs]
    in addition to the 'overhead limit exceded' we also will see :
    [Full GC 1194751K->1194751K(1365376K), 4.6785776 secs]
    [Full GC 1194751K->1188062K(1365376K), 5.4413659 secs]
    [Full GC 1194751K->1194751K(1365376K), 4.5800033 secs]
    [Full GC 1194751K->1194751K(1365376K), 4.4951213 secs]
    [Full GC 1194751K->1194751K(1365376K), 4.5227857 secs]
    [Full GC 1194751K->1171773K(1365376K), 5.5696274 secs]
    11/07/25 11:07:04 java.lang.OutOfMemoryError: Java heap space
    [Full GC 1194751K->1183306K(1365376K), 4.5841678 secs]
    [Full GC 1194751K->1184329K(1365376K), 4.5469164 secs]
    [Full GC 1194751K->1184831K(1365376K), 4.6415273 secs]
    [Full GC 1194751K->1174738K(1365376K), 5.3647290 secs]
    [Full GC 1194751K->1183878K(1365376K), 4.5660217 secs]
    [Full GC 1194751K->1184651K(1365376K), 4.5619460 secs]
    [Full GC 1194751K->1185795K(1365376K), 4.4341158 secs]

    There's an Oracle support note with a very similar MO :
    WebLogic Server: Getting "java.lang.OutOfMemoryError: GC overhead limit exceeded" exception with Sun JDK 1.6 [ID 1242994.1]
    If I search for "java.lang.OutOfMemoryError: GC overhead" on Oracle Support it returns at least 12 documents
    Might be bug 6065704. Search Oracle support for this bug number.
    Best Regards
    mseberg

  • Increase Java Max Heap Size

    Hello ,
    How can I increase java max heap size to more than 2G .I tried a bunch of different sizes and anything under 2G works just fine so this appears to be a hard limit.
    system info:
    Linux server with process data model 32-bit ,Java 1.5 , total physical memory 9GB and available virtual and physical memory close to 4GB
    I would appreciate any guidance on how to resolve this.
    Regards,
    Omid

    warnerja wrote:
    Any application wanting to allocate that much of a heap size should be considered a rogue application and a system hog of gigantic proportions, not desirable to have running on anyones machine. Focus on designing it correctly so that it doesn't require so much memory in the first place.By very virtue of the fact that the author has a test box with 9 gigs of RAM (I'm inferring), it seems unlikely that this app will be distributed to just "anyone."
    However, if this is meant to be some server app that scales to an enormous memory capacity, you're probably better off not using the java launcher directly and instead using some enterprise container.
    I'm a complete newb when it comes to anything meant to be scalable though, so warnerja would probably have better advice than I would.

  • Allocated heap memory goes up even when there is enough free memory

    Hi,
    Our Java application's memory usage keep growing. Further analysis of the heap memory using JProbe shows that the allocated heap memory goes up even when there is enough free memory available in the heap.
    When the process started, allocated heap memory was around 50MB and the memory used was around 8MB. After few hours, the inuse memory remains at 8MB (slight increase in KBs), but the allocated memory went upto 70MB.
    We are using JVM 1.5_10. What could be the reason for heap allocation going up even when there is enough free memory available in heap?
    -Rajesh.

    Hi Eric,
    Please check if there is any error or warning in the Event Viewer based on the data time?
    If there is any error, please post the event ID to help us to troubleshoot.
    Best Regards,
    Anna

  • MemoryOutOfError in jvm native heap

    hi,
    i am getting MemoryOutOfError at JVM native heap.
    anybody tell me why this is coming ?
    thanks
    Rams

    You need be more specific about your phenomenon.
    For example:
    OS Info
    JVM Info
    JVM Options
    Full Error Message
    GC Dump
    There are many possible reasons for OOEM in native heap.
    Thread stack size, JNI library, virtual memory space(somtimes by too big java heap size), file/socket open count, ...
    Even highly-specialized expert might not be able to tell which one is the main reason without more detailed information.
    Anyway, JVM forum might be more appropriate place for your question.

  • Native heap exaustion under Linux 2.6.x

    All,
    We have an application that extensively uses Runtime.exec() in a multi-threaded
    environment. Under both JRockit 1.4.2_8 and 1.4.2_10 x86 we observe steady growth
    of resident and virtual memory occupied by JVM, as reported by top, that eventually
    leads to the death of JVM.
    Java heap behaves normally. The application that is repeatedly executed by Runtime.exec()
    doesn't leak. That was confirmed by running it in a bash cycle.
    I am wondering if this is a known problem and if there is a fix or a workaround
    for this issue. If not, what are the ways to track down the problem?
    Thanks.
    Slava

    Calling deleteOnExit() on a File adds that file to a list of files that the JVM should delete once it exits. Files are never
    removed from this list. Thus, if you keep adding files to the list (by calling deleteOnExit()) you will run out of memory
    eventually.That may be what the JVM code does. Here is what the documentation says:
    ================================================================
    deleteOnExit
    public void deleteOnExit()
    Requests that the file or directory denoted by this abstract pathname be deleted when the
    virtual machine terminates. Deletion will be attempted only for normal termination of the virtual
    machine, as defined by the Java Language Specification.
    Once deletion has been requested, it is not possible to cancel the request. This method
    should therefore be used with care.
    Note: this method should not be used for file-locking, as the resulting protocol cannot
    be made to work reliably. The FileLock facility should be used instead.
    Throws:
    SecurityException - If a security manager exists and its SecurityManager.checkDelete(java.lang.String)
    method denies delete access to the file
    Since:
    1.2
    See Also:
    delete()"
    ================================================================
    As you can see, there is nothing in here that would alert that either there is
    a list that may cause *native heap* overflow nor that actual deletions of the
    file don't remove a file from the list. For all practical reasons it is a bug that
    was lurking there since 2001:
    http://bugs.sun.com/bugdatabase/view_bug.do;:YfiG?bug_id=4513817
    Other JVM vendors such as IBM have taken care about this problem (see #97403):
    http://www-1.ibm.com/support/docview.wss?uid=swg21221304
    It would be nice if you guys looked at this issue.
    Apart from that I am very pleased with JRockIt speed and improved stability - good
    job!
    Regards,
    Slava Imeshev

  • Heap dumps on very large heap memory

    We are experiencing memory leak issues with one of our application deployed on JBOSS (SUN JVM 1.5, Win 32 OS). The application is already memory intensive and consumes the maximum heap (1.5 GB) allowed on a 32-bit JVM on Win32.
    This leaves very few memory for heap dump and the JVM crashes whenever we try adding the heap dump flag (-agentlib:), with a "malloc error"..
    Has anyone faced a scenario like this?
    Alternatively for investigation purpose, we are trying to deploy it on a Windows X64 - but the vendor advises only to run on 32 bit JVM. Here is my question:
    1) Can we run 32bit JVM on Windows 64? Even if we run, can i allocate more than 2 GB for heap memory?
    2) I dont see the rational why we cannot run on 64bit JVM - because, JAVA programs are supposed to be 'platform-independent' and the application in form of byte code should be running no matter if it is a 32bit or 64-bit JVM?
    3) Do we have any other better tools (except HPROF heapdumps) to analyze memory leaks - we tried using profiling tools but they too fail becos of very few memory available.
    Any help is really appreciated! :-)
    Anush

    anush_tv wrote:
    1) Can we run 32bit JVM on Windows 64? Even if we run, can i allocate more than 2 GB for heap memory?Yes, but you're limited to 2GB like any other 32-bit process.
    2) I dont see the rational why we cannot run on 64bit JVM - because, JAVA programs are supposed to be 'platform-independent' and the application in form of byte code should be running no matter if it is a 32bit or 64-bit JVM?It's probably related to JBoss itself, which is likely using native code. I don't have experience with JBoss though.
    3) Do we have any other better tools (except HPROF heapdumps) to analyze memory leaks - we tried using profiling tools but they too fail becos of very few memory available.You could try "jmap", which can dump the heap.

Maybe you are looking for

  • SAP report for movement type and material type

    Hi! I would like to use a standard SAP report to give me a list of all the movement types (much like MB51) but instead of material I would like to select on Material type e.g. ROH. The result should for example be a list of the movement types 501 on

  • Play Button with Transition

    I am trying to create an initial slide that will look similar to a video played on Youtube. I would like the user to click on a Play Icon which triggers a transition to fade out and then start the presentation. Any ideas how I would do this?

  • Satellite A660: HW setup. Error1316 Network error occured

    I have an annoying problem on my Satellite A660 that I cant fix. I have lost the use of my function keys as well as the flash card function. It seems to be related to the HW setup utility which wont open. When I click on the icon this message comes u

  • How to capture LabVIEW value through pda

    工程師先生您好﹔ 現在我要寫一個程式是利用pda來抓取我LABVIEW上的值,server端的數值有改變,但無法將資料傳入pda,請問我這個程式是哪裡有寫錯,麻煩各位LABVIEW幫我解決這個問題,謝謝. 若今日我想要把LabVIEW上有多個顯示元件,我該如何撰寫才能將這些顯示元件的值讀入pda note:Data Server - Windows Mobile (電腦) Data Client - Windows Mobile         (pda) Attachments: Data S

  • How to stop Accounting Entries for certain Movement Types

    Hello, We are using Movement Type 511 to record free of charge receipts from vendors. Still it is generating Accounting Entries with Standard Material Price. Is there any way to stop Accounting Entries for 511 and other Movement Types, where external