RT-Target running out of memory while writing to Network Stream

Hey there,
I have a program, that transfers acquired data from the FPGA to the Host-PC. The RT-VI reads the data from the DMA-FIFO and writes it onto a Network Stream (BlockDiagram.png).
Now I am experiencing a phenomenon, that the RT-Target loads its RAM until it's full, and crashes.
I have no idea, why this happens, the buffer of the Network Stream is empty, all elements are read by the Host, and there is no array built by indexing or so.
Has anybody an idea, how I can handle this?
Best regards,
Solved!
Go to Solution.
Attachments:
BlockDiagram.png ‏43 KB
DSM.png ‏78 KB

Hey there,
I got the problem solved,
the problem was, the buffer of the sender endpoint was too big. Unlike this problem: http://digital.ni.com/public.nsf/allkb/784CB8093AE30551862579AB0050C429, it wasn't memory growth because of dynamic memory allocation,
it's just the normal speed of the cRIO while allocating the buffer memory. Setting the sender buffer much smaller, memory growth stops at a specific level (DSM2.png).
It's only strange, that memory usage grows that slowly, despite creating the endpoint with preallocated buffer, while usage sinks rapidly when the VI-execution stops...
Best regards...
Attachments:
DSM2.png ‏63 KB

Similar Messages

  • Problem : Application run out of memory while processing image I/O.

    Hi,
    I have coded an application (utility) for creating scaled (JPEG) images (thumb nails) and writing (saving) them to files. I tried this with both imageio package of jdk1.4 and com.sun.image.codec.jpeg.* classes of earlier version. Application works with both versions but when I execute it for more (say more than 10 images) images, system becomes very slow and application starts throwing a memory related (OutOfMemoryError) error. Is it that objects of java image related classes consume too much memory ? How to solve this problem ? (My system run P4 processor with 256 MB RAM).
    Awaiting solution,
    Thanx and regards,
    IB

    So I am not alone in this...
    I wrote the jpeg thumbnail generating code over a year ago and now when I am ready to use it in the finished app. I noticed that each process (as monitored using Forte 3's execution window) that instantiates a new ImageIcon object is never killed by the JVM. As more and more processes are started for distinct tasks they build up endlessly. Here's the method which is almost identical to the test code provided in sun's tutorial for generating thumbnails from jpeg and gif image files (http://developer.java.sun.com/developer/TechTips/1999/tt1021.html) :
    public boolean createThumbImage(int maxDim) throws EntegraEntityException {
         * Reads an image in a file and creates
         * a thumbnail in another file.
         * @param this.getImagePath() The name of image file.
         * @param thumb The name of thumbnail file. 
         * Will be created if necessary.
         * @param maxDim The width and height of
         * the thumbnail must
         * be maxDim pixels or less.
           String thumbsource =  this.path + this.filename;
           String thumbdest = this.path + "thumbs_" + maxDim + File.separator/*"\\"*/ + Utilities.replaceString(Utilities.replaceString(this.filename," ","_","ALL").trim(),"[.][a-zA-Z]*",".jpg","ALL");
       //  System.out.println("thumbsource in createthumbimage(): " + thumbsource + "\n");
       //  System.out.println("thumbdest in createthumbimage(): " + thumbdest + "\n");
            try {
                // Get the image from a file.
                java.awt.Image inImage = new ImageIcon (thumbsource).getImage();
                // Determine the scale.
             double scale = (double)maxDim/(double)inImage.getHeight(null);
                if (inImage.getWidth(null) > inImage.getHeight(null)) {
                    scale = (double)maxDim/(double)inImage.getWidth(null);
                // Determine size of new image.
                //One of them
                // should equal maxDim.
                int scaledW = (int)(scale*inImage.getWidth(null));
                int scaledH = (int)(scale*inImage.getHeight(null));
                // Create an image buffer in
                //which to paint on.
                BufferedImage outImage =
                  new BufferedImage(scaledW, scaledH,
                    BufferedImage.TYPE_INT_RGB);
                // Set the scale.
                AffineTransform tx =
                  new AffineTransform();
                // If the image is smaller than
                //the desired image size,
                // don't bother scaling.
                if (scale < 1.0d) {
                    tx.scale(scale, scale);
                // Paint image.
                Graphics2D g2d =
                 outImage.createGraphics();
                g2d.drawImage(inImage, tx, null);
                g2d.dispose();
                // JPEG-encode the image
                //and write to file.
                Utilities.fileMakeDirs(this.path + "thumbs_" + maxDim + File.separator/*"\\"*/);
                Utilities.fileMakeFile(thumbdest);
                OutputStream os =
                 new FileOutputStream(thumbdest);
                JPEGImageEncoder encoder = JPEGCodec.createJPEGEncoder(os);
                encoder.encode(outImage);
                os.close();
                return true;
            } catch (IOException e) {
                throw new EntegraEntityException("IO exception creating thumbnail for Image object ID: " + this.getId() + " ::",e);
            } catch (EntegraFileIOException fi) {
                throw new EntegraEntityException("File IO exception creating thumbnail for Image object ID: " + this.getId() + " ::",fi);
        //    return false; //Never reached. Since it doesn't live in the "try" block but must be here to ensure method body gets a return value.
    }:During my troubleshooting of the method execution yesterday I discovered that the stuck process only occurs as a consequence of the line shown in bold below :
    java.awt.Image inImage = new ImageIcon (thumbsource).getImage();
    :this indicates to me that there is something wrong with how either the Image object or the ImageIcon file resources are allocated. At first I tried using the "flush()" method of the java.awt.Image class to release the resources to no avail, later I was able to confirm that the error does not occur for the default constructor of ImageIcon only for the one that specifies the source path (as shown in the line above) this indicates that a fileIO stream is being created and probably not being released inside ImageIcon or Image, but since that stream is probably private to those core classes we can't access it to close() it properly. I sure hope I am wrong on this or there is an alternative to using these classes..
    Please let me know if any of you have other ideas on how to quash this bug or can otherwise find flaws in my logic for it's occurance.
    Regards,
    Sent2null

  • Running out of memory while using cursored stream with large data

    We are following the suggestions/recommendations for the cursored stream:
    CursoredStream cursor = null;
              try
                   Session session = getTransaction();
                   int batchSize = 50;
                   ReadAllQuery raq = getQuery();
                   raq.useCursoredStream(batchSize, batchSize);
                   int num = 0;
                   ArrayList<Request> limitRequests = null;
                   int totalLimitRequest = 0;
                   cursor = (CursoredStream) session.executeQuery(raq);
                   while( !cursor.atEnd() )
                        Request request = (Request) cursor.read() ;
                        if( num == 0 )
                             limitRequests = new ArrayList<Request>(batchSize);
                        limitRequests.add(request);
                        totalLimitRequest++;
                        num++;
                        if( num >= batchSize )
                             log.warn("Migrating batch of " + batchSize + " Requests.");
                             updateLimitRequestFillPriceForBatch(limitRequests);
                             num = 0;
                             cursor.releasePrevious();
                   if( num > 0 )
                        updateLimitRequestFillPriceForBatch(limitRequests);
                   cursor.close();
    We are committing every 50 records in the unit of work, if we set DontMaintianCache on the ReadAllQuery we are getting PrimaryKeyExceptions intermittently, and we do not see much difference in the IdentityMap size.
    Any suggestions/ideas for dealing with large data sets? Thanks

    Hi,
    If I use read-only classes with CursoredStream and execute the query within UOW, should I be saving any memory?
    I had to use UOW because when I use Session to execute the query I get
    6115: ISOLATED_QUERY_EXECUTED_ON_SERVER_SESSION
    Cause: An isolated query was executed on a server session: queries on isolated classes, or queries set to use exclusive connections, must not be executed on a ServerSession or in CMP outside of a transaction.
    I assume marking the descriptor as read-only will avoid registering in UOW, but I want to make sure that this is the case while using CursoredStream.
    We are running in OC4J(OAS10.1.3.4) with BeanManagedTransaction.
    Please suggest.
    Thanks
    -Raam
    Edited by: Raam on Apr 2, 2009 1:45 PM

  • Target has run out of memory on LM3s8962

    I'm using the LM3s8962 evaluation kit to record data from the ADC's.  I have the system set up so that I use the elemental nodes of the four adc's in a while loop, and replace the values in four different arrays.  The arrays are initialize (1x1000 elements) before entering the loop.  This works fine.
    THE PROBLEM:  When I try to make the arrays larger (i.e. initial arrays larger than 1000 points, 4 individual arrays), I get the following error:
    Error: Memory allocation failed. The target has run out of memory. [C:\Program Files (x86)\National Instruments\LabVIEW 2011\CCodeGen\libsrc\blockdiagram\CCGArrSupport2.c at line 253: 2 3
    OR
    Error: Memory allocation failed. The target has run out of memory. [C:\Program Files (x86)\National Instruments\LabVIEW 2011\CCodeGen\libsrc\blockdiagram\CCGArrSupport2.c at line 173: 2 3
    Any suggestions?

    Th0r wrote:
    It looks like you're filling up the flash memory on the LM3S8962 with all of these array initializations.  According to page 263 of the LM3S8962 datasheet, that microcontroller has 256 KB of flash memory which you can use to fill up with your code.  In addition to your array initializations, some of this space is taken up by the LabVIEW Embedded Module-specific code as well.  What datatype are you using in these arrays?  Does this error occur upon building or running your code?  Thanks for any additional information you can provide!  
    That's probably it.  The error occurs when building the code, before it's actually able to run.  If reduce the array size, I'm able to run the code no problem.  At the moment,  I'm using a long 32 bit integer, which I know realize I can reduce significantly, as my ADC only reads at 10 bits.  Do you know if there's a way that I can preallocate the array to a place other than flash?
    I've found a fix around it since I last posted, in which I set up a buffer (smaller) and then save the buffer values on the SD card.  This works well and I can sample for long periods of time, but it does slow down my overall sampling rate, so I'd like to fix the above problem nonetheless. 

  • System hanging when it runs out of memory

    Hello,
    my system has a finite amount of RAM and swap (it doesn't matter, to my purpose, if it's 16GB or 128MB, I'm not interested in increasing it anyway).
    Sometimes my apps completely use all the available memory. I would expect that in these cases the kernel kills some apps to keep working correctly. The OOM Killer exists just for this, doesn't it?
    What happens, instead, is that the system hangs. Even the mouse stops working. Sometimes it manages to get back to life in a few seconds/minutes, other times hours pass by and nothing changes.
    I don't want to add more memory, I just want that the kernel kills some application when it's running out of memory.
    Why isn't this happening?
    Right now I'm writing a bash script that will kill the most memory-hungry process when available memory gets below 10%, because I'm sick of freezing my machine.
    But why do I have to do this? Why do I need an user space tool polling memory usage and sentencing applications according to a cheap policy? What the hell is wrong with my OOM Killer, why isn't it doing is job?!

    Alright, you won, now quit pointing out my ignorance
    Your awkish oom killer is a lot cooler than mine, switching to it, thanks!
    I did some testing (initially just to test the oom killing script) and found out that if a program tries to allocate all the memory it can, it gets eventually killed by linux's oom killer. If instead it stops allocating new memory when there are less than 4MB of free memory (or similar values) the OOM won't do anything, and the system will get stuck, like if a forkbomb was running.
    That's it, this program with MINSIZE=1 will be killed, while with MINSIZE=4MB will force me to hard reboot:
    #include <string.h>
    #include <stdlib.h>
    #define MINSIZE (1024*1024*4) // 4MB
    int main( )
    int block = 1024*1024*1024; // 1GB
    void *p;
    while( 1 ) {
    p = malloc( block );
    if( p ) {
    memset( p, 85, block );
    else if( block > MINSIZE ) {
    block /= 2;
    return 0;
    Guess I'd need to go deeper to understand why Linux' oom killer works like that, but I won't (assuming the oom killing script will behave).

  • Oracle 9i running out of memory

    Folks !
    I have a simple 3 table schema with a few thousand entries each. After dedicating gigabytes of hard disk space and 50% of my 1+ GB memory, I do a few simple Oracle Text "contains" searches (see below) on these tables and oracle seems to grow some 25 MB after each query (which typically return less than a dozen rows each) till it eventually runs out of memory and I have to reboot the system (Sun Solaris).
    This is on Solaris 9/Sparc with Oracle 9.2 . My query is simple right outer join. I think the memory growth is related to Oracle Text index/caching since memory utilization seems pretty stable with simple like '%xx%' queries.
    "top" shows a dozen or so processes each with about 400MB RSS/SIZE. It has been a while since I did Oracle DBA work but I am nothing special here. Databse has all the default settings that you get when you create an Oracle database.
    I have played with SGA sizes and no matter how large or small the size of SGA/PGA, Oracle runs out of memory and crashes the system. Pretty stupid to an Enterprise databas to die like that.
    Any clue on how to arrest the fatal growth of memory for Oracle 9i r2?
    thanks a lot.
    -Sanjay
    PS: The query is:
    SELECT substr(sdn_name,1,32) as name, substr(alt_name,1,32) as alt_name, sdn.ent_num, alt_num, score(1), score(2)
    FROM sdn, alt
    where sdn.ent_num = alt.ent_num(+)
    and (contains(sdn_name,'$BIN, $LADEN',1) > 0 or
    contains(alt_name,'$BIN, $LADEN',2) > 0)
    order by ent_num, score(1), score(2) desc;
    There are following two indexes on the two tables:
    create index sdn_name on sdn(sdn_name) indextype is ctxsys.context;
    create index alt_name on alt(alt_name) indextype is ctxsys.context;

    I am already using MTS.
    Atached is the init.ora file below.
    may be I should repost this article with subject "memory leak in Oracle" to catch developer attention. I posted this a few weeks back in Oracle Text groiup and no response there either.
    Thanks for you help.
    -Sanjay
    # Copyright (c) 1991, 2001, 2002 by Oracle Corporation
    # Cache and I/O
    db_block_size=8192
    db_cache_size=33554432
    db_file_multiblock_read_count=16
    # Cursors and Library Cache
    open_cursors=300
    # Database Identification
    db_domain=""
    db_name=ofac
    # Diagnostics and Statistics
    background_dump_dest=/space/oracle/admin/ofac/bdump
    core_dump_dest=/space/oracle/admin/ofac/cdump
    timed_statistics=TRUE
    user_dump_dest=/space/oracle/admin/ofac/udump
    # File Configuration
    control_files=("/space/oracle/oradata/ofac/control01.ctl", "/space/oracle/oradata/ofac/control02.ctl", "/space/oracle/oradata/ofac/control03.ctl")
    # Instance Identification
    instance_name=ofac
    # Job Queues
    job_queue_processes=10
    # MTS
    dispatchers="(PROTOCOL=TCP) (SERVICE=ofacXDB)"
    # Miscellaneous
    aq_tm_processes=1
    compatible=9.2.0.0.0
    # Optimizer
    hash_join_enabled=TRUE
    query_rewrite_enabled=FALSE
    star_transformation_enabled=FALSE
    # Pools
    java_pool_size=117440512
    large_pool_size=16777216
    shared_pool_size=117440512
    # Processes and Sessions
    processes=150
    # Redo Log and Recovery
    fast_start_mttr_target=300
    # Security and Auditing
    remote_login_passwordfile=EXCLUSIVE
    # Sort, Hash Joins, Bitmap Indexes
    pga_aggregate_target=25165824
    sort_area_size=524288
    # System Managed Undo and Rollback Segments
    undo_management=AUTO
    undo_retention=10800
    undo_tablespace=UNDOTBS1

  • Running out of memory despite having set je.maxMemory to a moderate value

    I have set je.maxMemory to 20MB (je.maxMemory=20000000) and allowed a max heap size of 512MB (-Xms256M -Xmx512M).
    After two hours of running my web service, I'm running out of memory. After having profiled my service (using Yourkit Java Profiler 1.10.6), I can see the following:
    Name                                               Objects ShallowSize  RetainedSize
    byte[]                                               16711   124124880     124124880
    com.sleepycat.je.tree.BIN                              181       24616     116254200
    com.sleepycat.je.tree.Node[]                           187       98736     115743184
    com.sleepycat.je.tree.LN                              7092      226944     115253600
    java.util.concurrent.ConcurrentHashMap$HashEntry       554       17728      78328944
    java.util.concurrent.ConcurrentHashMap$HashEntry[]    1053       34728      77489632
    java.util.concurrent.ConcurrentHashMap                 117        5616      71812072
    java.util.concurrent.ConcurrentHashMap$Segment[]       118       10304      71807912
    java.util.concurrent.ConcurrentHashMap$Segment        1052       42080      71798808
    com.sleepycat.je.tree.IN                                 6         672      45592352
    java.lang.String                                    135888     4348416      14152664The memory profiler claims further, that com.sleepycat.je.tree.BIN is responsible for 71% of all heap memory.
    In any case, com.sleepycat.je.tree.BIN claims ~ 116MB of heap memory, which is by any goodwill, exceeded the limit of 20MB.
    How can this be?
    How is JE ensuring that the limit is not exceeded? Is there a timer (thread) running which once a while checks the memory used and then cleans up ; or is memory usage checked creating a com.sleepycat.je.tree.BIN object?
    My environment:
    BDB JE 4.0.92 - used as cache loader within Jboss Cache (3.2.7.GA), running on a JBOSS Application Server, Java 1.6 (IBM) on Linux. Further details are listed in the system properties below (except some deleted security items).
    System properties:
    (java.lang.String, int, java.lang.StringBuffer, int)=contains
    DestroyJavaVM helper thread=(java.lang.String, java.security.KeyStore$Entry, java.security.KeyStore$ProtectionParameter)
    base.collection.name=CD2JAVA
    bind.address=10.12.25.130
    catalina.base=/work/ocrgws_test/server0
    catalina.ext.dirs=/work/ocrgws_test/server0/lib
    catalina.home=/work/ocrgws_test/server0
    catalina.useNaming=false
    com.arjuna.ats.arjuna.objectstore.objectStoreDir=/work/ocrgws_test/server0/data/tx-object-store
    com.arjuna.ats.jta.lastResourceOptimisationInterface=org.jboss.tm.LastResource
    com.arjuna.ats.tsmx.agentimpl=com.arjuna.ats.internal.jbossatx.agent.LocalJBossAgentImpl
    com.arjuna.common.util.logger=log4j_releveler
    com.arjuna.common.util.logging.DebugLevel=0x00000000
    com.arjuna.common.util.logging.FacilityLevel=0xffffffff
    com.arjuna.common.util.logging.VisibilityLevel=0xffffffff
    com.ibm.cpu.endian=little
    com.ibm.jcl.checkClassPath=
    com.ibm.oti.configuration=scar
    com.ibm.oti.jcl.build=20100326_1904
    com.ibm.oti.shared.enabled=false
    com.ibm.oti.vm.bootstrap.library.path=/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs:/opt/ibm/java-x86_64-60/jre/lib/amd64
    com.ibm.oti.vm.library.version=24
    com.ibm.util.extralibs.properties=
    com.ibm.vm.bitmode=64
    common.loader=${catalina.home}/lib,${catalina.home}/lib/*.jar
    epo.jboss.deploymentscanner.extradirs=/work/ocrgws_test/app/
    external.cert.ldap.* = ***************
    file.encoding=UTF-8
    file.separator=/
    flipflop.activation.time=16:30
    hibernate.bytecode.provider=javassist
    ibm.signalhandling.rs=false
    ibm.signalhandling.sigchain=true
    ibm.signalhandling.sigint=true
    ibm.system.encoding=UTF-8
    jacorb.config.log.verbosity=0
    java.assistive=ON
    java.awt.fonts=
    java.awt.graphicsenv=sun.awt.X11GraphicsEnvironment
    java.awt.printerjob=sun.print.PSPrinterJob
    java.class.path=/work/ocrgws_test/config:/usr/local/jboss-eap-4.3-cp07/bin/run.jar:/opt/ibm/java-x86_64-60/lib/tools.jar
    java.class.version=50.0
    java.compiler=j9jit24
    java.endorsed.dirs=/usr/local/jboss-eap-4.3-cp07/lib/endorsed
    java.ext.dirs=/opt/ibm/java-x86_64-60/jre/lib/ext
    java.fullversion=JRE 1.6.0 IBM J9 2.4 Linux amd64-64 jvmxa6460sr8-20100401_55940 (JIT enabled, AOT enabled)
    J9VM - 20100401_055940
    JIT - r9_20100401_15339
    GC - 20100308_AA_CMPRSS
    java.home=/opt/ibm/java-x86_64-60/jre
    java.io.tmpdir=/tmp
    java.jcl.version=20100408_01
    java.library.path=/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs:/opt/ibm/java-x86_64-60/jre/lib/amd64:/usr/lib64/mpi/gcc/openmpi/lib64:/usr/lib
    java.naming.factory.initial=org.jnp.interfaces.NamingContextFactory
    java.naming.factory.url.pkgs=org.jboss.naming:org.jnp.interfaces
    java.net.preferIPv4Stack=true
    java.protocol.handler.pkgs=org.jboss.net.protocol
    java.rmi.server.codebase=http://10.12.25.130:8083/
    java.rmi.server.hostname=10.12.25.130
    java.rmi.server.randomIDs=true
    java.runtime.name=Java(TM) SE Runtime Environment
    java.runtime.version=pxa6460sr8-20100409_01 (SR8)
    java.security.krb5.conf=/usr/local/jboss/etc/krb5.conf
    java.specification.name=Java Platform API Specification
    java.specification.vendor=Sun Microsystems Inc.
    java.specification.version=1.6
    java.util.prefs.PreferencesFactory=java.util.prefs.FileSystemPreferencesFactory
    java.vendor.url=http://www.ibm.com/
    java.vendor=IBM Corporation
    java.version=1.6.0
    java.vm.info=JRE 1.6.0 IBM J9 2.4 Linux amd64-64 jvmxa6460sr8-20100401_55940 (JIT enabled, AOT enabled)
    J9VM - 20100401_055940
    JIT - r9_20100401_15339
    GC - 20100308_AA_CMPRSS
    java.vm.name=IBM J9 VM
    java.vm.specification.name=Java Virtual Machine Specification
    java.vm.specification.vendor=Sun Microsystems Inc.
    java.vm.specification.version=1.0
    java.vm.vendor=IBM Corporation
    java.vm.version=2.4
    javax.management.builder.initial=org.jboss.mx.server.MBeanServerBuilderImpl
    javax.net.ssl.trustStore=/usr/local/jboss/etc/ldap.truststore
    javax.net.ssl.trustStorePassword=password
    jboss.bind.address=10.12.25.130
    jboss.home.dir=/usr/local/jboss-eap-4.3-cp07
    jboss.home.url=file:/usr/local/jboss-eap-4.3-cp07/
    jboss.identity=30df88bc0a52e350x6e2ff59cx136c17794d5x-8000757
    jboss.lib.url=file:/usr/local/jboss-eap-4.3-cp07/lib/
    jboss.messaging.controlchanneludpaddress=239.1.200.4
    jboss.messaging.datachanneludpaddress=239.1.200.4
    jboss.partition.name=ocrgws_test_Partition
    jboss.partition.udpGroup=239.1.200.4
    jboss.remoting.domain=JBOSS
    jboss.remoting.instanceid=30df88bc0a52e350x6e2ff59cx136c17794d5x-8000757
    jboss.remoting.jmxid=luu002t.internal.epo.org_1334685694459
    jboss.remoting.version=22
    jboss.security.disable.secdomain.option=true
    jboss.server.config.url=file:/work/ocrgws_test/server0/conf/
    jboss.server.data.dir=/work/ocrgws_test/server0/data
    jboss.server.home.dir=/work/ocrgws_test/server0
    jboss.server.home.url=file:/work/ocrgws_test/server0/
    jboss.server.lib.url=file:/work/ocrgws_test/server0/lib/
    jboss.server.log.dir=/work/ocrgws_test/server0/log
    jboss.server.name=luu002t_ocrgws_test_server0
    jboss.server.temp.dir=/work/ocrgws_test/server0/tmp
    jboss.tomcat.udpGroup=239.1.200.4
    jbossmx.loader.repository.class=org.jboss.mx.loading.UnifiedLoaderRepository3
    je.maxMemory=20000000
    jgroups.bind_addr=10.12.25.130
    jmx.console.bindcredential=3bpwdmpc
    jmx.console.binddn=cn=jbossauth-ro,ou=accounts,ou=auth,dc=epo,dc=org
    jmx.console.rolesctxdn=ou=roles-test,ou=jboss,ou=applications,ou=internal,dc=epo,dc=org
    jndi.datasource.name=java:MainframeDS
    jnp.disableDiscovery=true
    jxe.current.romimage.version=15
    jxe.lowest.romimage.version=15
    line.separator=
    mainframelogin.password=720652a1e842fc7f
    mainframelogin.username=test_t
    org.apache.commons.logging.Log=org.apache.commons.logging.impl.Log4JLogger
    org.apache.tomcat.util.http.ServerCookie.VERSION_SWITCH=true
    org.epo.jboss.application.home=/work/ocrgws_test
    org.hyperic.sigar.path=/work/ocrgws_test/server0/./deploy/hyperic-hq.war/native-lib
    org.jboss.ORBSingletonDelegate=org.jacorb.orb.ORBSingleton
    org.omg.CORBA.ORBClass=org.jacorb.orb.ORB
    org.omg.CORBA.ORBSingletonClass=org.jboss.system.ORBSingleton
    org.w3c.dom.DOMImplementationSourceList=org.apache.xerces.dom.DOMXSImplementationSourceImpl
    os.arch=amd64
    os.name=Linux
    os.version=2.6.32.46-0.3-xen
    package.access=sun.,org.apache.catalina.,org.apache.coyote.,org.apache.tomcat.,org.apache.jasper.,sun.beans.
    package.definition=sun.,java.,org.apache.catalina.,org.apache.coyote.,org.apache.tomcat.,org.apache.jasper.
    path.separator=:
    poll.interval.milliseconds=300000
    program.name=run.sh
    server.loader=
    shared.loader=
    spnego.config=/usr/local/jboss/etc/spnego.properties
    sun.arch.data.model=64
    sun.boot.class.path=/usr/local/jboss-eap-4.3-cp07/lib/endorsed/xercesImpl.jar:/usr/local/jboss-eap-4.3-cp07/lib/endorsed/xalan.jar:/usr/local/jboss-eap-4.3-cp07/lib/endorsed/serializer.jar:/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs/jclSC160/vm.jar:/opt/ibm/java-x86_64-60/jre/lib/annotation.jar:/opt/ibm/java-x86_64-60/jre/lib/beans.jar:/opt/ibm/java-x86_64-60/jre/lib/java.util.jar:/opt/ibm/java-x86_64-60/jre/lib/jndi.jar:/opt/ibm/java-x86_64-60/jre/lib/logging.jar:/opt/ibm/java-x86_64-60/jre/lib/security.jar:/opt/ibm/java-x86_64-60/jre/lib/sql.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmorb.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmorbapi.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmcfw.jar:/opt/ibm/java-x86_64-60/jre/lib/rt.jar:/opt/ibm/java-x86_64-60/jre/lib/charsets.jar:/opt/ibm/java-x86_64-60/jre/lib/resources.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmpkcs.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmcertpathfw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjgssfw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjssefw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmsaslfw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjcefw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjgssprovider.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjsseprovider2.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmcertpathprovider.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmxmlcrypto.jar:/opt/ibm/java-x86_64-60/jre/lib/management-agent.jar:/opt/ibm/java-x86_64-60/jre/lib/xml.jar:/opt/ibm/java-x86_64-60/jre/lib/jlm.jar:/opt/ibm/java-x86_64-60/jre/lib/javascript.jar:/tmp/yjp201202191932.jar
    sun.boot.library.path=/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs:/opt/ibm/java-x86_64-60/jre/lib/amd64
    sun.io.unicode.encoding=UnicodeLittle
    sun.java.command=org.jboss.Main -b 10.12.25.130 -Djboss.server.home.dir=/work/ocrgws_test/server0 -Djboss.server.home.url=file:/work/ocrgws_test/server0 -Djboss.server.name=luu002t_ocrgws_test_server0 -Djboss.partition.name=ocrgws_test_Partition -Depo.jboss.deploymentscanner.extradirs=/work/ocrgws_test/app/ -Dorg.epo.jboss.application.home=/work/ocrgws_test
    sun.java.launcher.pid=17781
    sun.java.launcher=SUN_STANDARD
    sun.java2d.fontpath=
    sun.jnu.encoding=UTF-8
    sun.rmi.dgc.client.gcInterval=3685000
    sun.rmi.dgc.server.gcInterval=3685000
    system=java.io.ObjectStreamField
    tomcat.util.buf.StringCache.byte.enabled=true
    user.country=US
    user.dir=/work/ocrgws_test
    user.home=*****************
    user.language=en
    user.name=***********
    user.timezone=Europe/Berlin
    user.variant=

    The memory profiler claims further, that com.sleepycat.je.tree.BIN is responsible for 71% of all heap memory. In any case, com.sleepycat.je.tree.BIN claims ~ 116MB of heap memory, which is by any goodwill, exceeded the limit of 20MB. >
    I'm not sure whether the profiler is reporting live objects only (referenced) or all objects (including those not yet reclaimed). If the latter, it isn't telling you how much memory is actually referenced by the JE cache.
    Please look at the JE stats to see what the cache usage is, from JE's point of view.
    If you believe there is a bug in JE cache management, you'll need to write a small standalone test to demonstrate it and submit it to us, since we don't know of any such bug. Also note that we'll have difficulty supporting JE 4.0 (without a support contract anyway). Please use JE 5.0, or at least 4.1.
    Eviction occurs as objects are allocated, as well as in background threads. Eviction in background threads and concurrent eviction were greatly improved in JE 4.1.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Aperture runs out of memory generating previews

    Aperture 3.2.3
    65K photos in library
    OSX 10.7.3
    Model Identifier:          MacPro1,1
    Processor Name:          Dual-Core Intel Xeon
    Processor Speed:          2.66 GHz
    Total Number of Cores:          4
    L2 Cache (per Processor):          4 MB
    Memory:          10 GB
    After moving my photoes to an external drive as referenced images, and a library rebuild, Aperture is attempting to recreatea all of my previews.  After generateing a couple of hunderd previews, it runs out of memory and I can see the following errors in the system console:
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: Aperture(37843,0xb071b000) malloc: *** mmap(size=1464881152) failed (error code=12)
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** error: can't allocate region
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** set a breakpoint in malloc_error_break to debug
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: Aperture(37843,0xb071b000) malloc: *** mmap(size=1464881152) failed (error code=12)
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** error: can't allocate region
    5/11/12 6:03:09.040 PM [0x0-0xfe0fe].com.apple.Aperture: *** set a breakpoint in malloc_error_break to debug
    5/11/12 6:03:09.041 PM [0x0-0xfe0fe].com.apple.Aperture: Aperture(37843,0xb071b000) malloc: *** mmap(size=1464881152) failed (error code=12)
    5/11/12 6:03:09.041 PM [0x0-0xfe0fe].com.apple.Aperture: *** error: can't allocate region
    5/11/12 6:03:09.041 PM [0x0-0xfe0fe].com.apple.Aperture: *** set a breakpoint in malloc_error_break to debug
    Inactive memory quickly uses up all available memory and is not released.  Quitting Aperture only releases half of the inactive memory.
    Any ideas on how to get previews generated without running out of memory?  I've tried running aperture in 32-bit mode, with no difference in behavior.

    Are Previews being generated as you import, or later?  I am asking because I can identify with memory issues when editing numerous images from my D800 in succession (which are huge), but i am not seeing it while importing.
    Ernie

  • Problems with PNGs... Overall compression + Running out of memory!

    We're having a number of issues with PNGs while working on our first iPhone project, and any assistance would be greatly appreciated!
    Our game is using a large number of PNG assets, some of which are full-frame (though the full-frame files tend to mostly be transparent/use alphas, though apparently that doesn't help memory issues much).
    We're running into two huge problems -
    1) We're running out of memory on device when calling to these full frame sequences, which tend to be anywhere from 10-40 frames each at anywhere from 50 to 250kb in size.
    2) Our overall package size is huge, sitting at around 60mb. I've already compressed the pngs through photoshop to the best of my ability, and I'm not having much luck with downloadable compressors like pngcrush. IS there a way to compress all the pngs through xcode/c++/objective C? The programmers are informing me right now that the only compression possible is whatever I apply directly to my pngs on my end - nothing through code.
    I'm stumped as to how I'm seeing seemingly-complex apps with plenty of content at 1-5mbs, and running smoothly with full-frame animations. I'm imagining the problem is that we're not using a proprietary engine to properly manage things, but I'm wondering if there is a simple solution.
    Thanks in advance, guys!
    P.S. - Already done a bunch of research on my own, and not having much luck. Just wondering if there is something obvious that both the programmers and I are missing!

    * Make image frame smaller? A fully transparent border around an image is pure overhead -- there is still data in the actual pixels (even though you cannot see them), plus the transparent border itself. All it needs is an x/y coordinate and a tiny code adjustment.
    * Use less transparency bits? Perhaps you could do with 1-bit transparency on some images. It'll use 1/8 times the memory (well, globally).
    * Use less colors? 24 bit color might compress down to 16 bits without visual artifacts (esp. when you don't have lots of gradients). Perhaps even 8 bit palettized.
    * Make images smaller? You might be able to get away with enlarging some images.
    Just a few things you could check right away, without major rewrites.

  • Why does the iPad 2 run out of memory

    Why does the iPad 2 run out of memory?

    First ...read deggie's response. Makes sense - even a computer running a large amount of RAM can run short on memory depending upon the operation you are performing so sometimes you have to do what is available to "clear" the memory.
    You can ...close apps and restart the iPad on occasion. You close apps by double tapping the home button to bring up your recent apps in the multitasking tray. Tap and hold down on any of the icons until they all begin to wiggle. Then tap the minus signs in the corner of the apps to close them. Close all of your apps in that area. Also take note of how many apps are down there. That may have something to do with why the iPad is running out of memory. Tap the home button twice to return to the home screen. Then restart your iPad by holding down on the sleep button until the red slider appears and slide to shut off. To power up, hold the sleep button until the Apple logo appears and let go of the button.
    If you do this once in a while, it may make a difference in how your iPad performs. In any event, it will not hurt anything.

  • ImageFetcher running out of memory?

    I'm using images in my program and after awhile I run out of memory. I've never really used images before but is there a way to maybe clear out whatever buffer is holding on to images that are no longer being used?
    Exception in thread "Image Fetcher 0" java.lang.OutOfMemoryError: Java heap space
         at java.awt.image.DataBufferInt.<init>(Unknown Source)
         at java.awt.image.Raster.createPackedRaster(Unknown Source)
         at java.awt.image.DirectColorModel.createCompatibleWritableRaster(Unknown Source)
         at sun.awt.image.ImageRepresentation.createBufferedImage(Unknown Source)
         at sun.awt.image.ImageRepresentation.setPixels(Unknown Source)
         at sun.awt.image.ImageDecoder.setPixels(Unknown Source)
         at sun.awt.image.JPEGImageDecoder.sendPixels(Unknown Source)
         at sun.awt.image.JPEGImageDecoder.readImage(Native Method)
         at sun.awt.image.JPEGImageDecoder.produceImage(Unknown Source)
         at sun.awt.image.InputStreamImageSource.doFetch(Unknown Source)
         at sun.awt.image.ImageFetcher.fetchloop(Unknown Source)
         at sun.awt.image.ImageFetcher.run(Unknown Source)

    Image img = ((ImageShape) (i.getShape())).getImage();
    if (i.getShape() != null)
            myMedia.addImage(img, 1);
    myMedia.waitForAll();
    myMedia.removeImage(img);That would be correct. Here's a slightly more elaborate example
    private static MediaTracker tracker;
    private static int NEXT_ID;
    /**Loads the specified image if it is not already loaded. Note that
         * Images created with the default toolkit need to be loaded.  A loaded
         * image is one that is ready for drawing. If it is already loaded then
         * this method does nothing. This method blocks until the image is loaded.
         * @throws NullPointerException If image is null.
         * @return true if the image was loaded or false if the image for whatever
         *         reason could not be loaded.*/
        public static boolean loadImage(Image image)
            if(image == null) throw new NullPointerException("null image!");
            if(image.getWidth(null) == -1 || image.getHeight(null) == -1)
                if (tracker == null)
                    tracker = new MediaTracker(new java.awt.Container());
                int id;
                synchronized (tracker) {
                    id = ++NEXT_ID;
                tracker.addImage(image, id);
                try {
                    tracker.waitForID(id, 0);
                } catch (InterruptedException e) {
                    System.err.println("INTERRUPTED while loading Image");
                    tracker.removeImage(image, id);
                    return false;
                int loadStatus = tracker.statusID(id, false);
                tracker.removeImage(image, id);
                if ((loadStatus & MediaTracker.COMPLETE) == MediaTracker.COMPLETE)
                    return true;
                else
                    return false;
           return true;
        }And as DB suggested, look into the javax.imageio.ImageIO class and its read methods.

  • Finale running out of memory

    Mac Pro 3.5GHz 6-core Intel Xeon E5
    Memory:  16GB 1867 MHz DDR3
    IT of Flash
    I am running OSX 10.9.5
    The problem occurs when I use Finale 2014 3.4736  (latest version) music notation application with Garritan for Finale
    Shortly after starting work on a small vocal score with piano accompaniment I am faced with the warning:
    "Finale has run out of memory. Please save your work and quit".
    No other applications are running.
     How can this be?
    Help is needed
    Thanks Sion
    Back

    Hi ..
    If you can, don't run other memory intensive apps while using the Finale app.
    Open the Activity Monitor located in HD > Applications > Utilities then select the Memory tab.
    From there you can see which apps are using the most RAM. (memory)

  • I am running out of memory on my hard drive and need to delete files. How can I see all the files/applications on my hard drive so I can see what is taking up a lot of room?

    I am running out of memory on my hard drive and need to delete files. How can I see all the files/applications on my hard drive so I can see what is taking up a lot of room?
    Thanks!
    David

    Either of these should help.
    http://grandperspectiv.sourceforge.net/
    http://www.whatsizemac.com/
    Or search 'disk size' in the App Store.
    Be carefull with what you delete & have a backup BEFORE you start, you may also want to reboot to try to free any memory that may have been written to disk.

  • How can I avoid running out of memory when creating components dynamically

    Hello everyone,
    Recently, I am planning to design a web application. It will be used by all middle school teachers in a region to make examination papers and it must contain the following main functions.
    1)Generate test questions dynamically. For instance, a teacher who logs on the web application will only see a select one menu and a Next Quiz button. The former is used for determining the number of options for the current multiple/single choice question. The later is dedicated to creating appropriate input text elements according to the selected option number. That is to say, if the teacher selects 4 in the menu and presses the Next Quiz button, 5 input text form elements will appear. The first one is for the question to be asked such as "1.What is the biggest planet in the solar system?", the others are optional answers like a)Uranus. b) Saturn. c)Jupiter. d)Earch. Each answer stands for an input text elements. When the teacher fills in the fourth answer, another select one menu and Next Quiz button will emerge on the fly just under this answer, allowing the teacher to make the second question. The same thing repeats for the following questions.
    2)Undo and Redo. Whenever a teacher wants to roll back or redo what he has done, just press the Undo or[i] Redo button. In the previous example, if the teacher selects the third answer and presses the Delete button to drop this answer, it will delete both the literal string content[i] and the input text element, changing the answer d to c automatically. After that, he decides to get back the original answer c, Jupiter, he can just click the Undo button as if he hadn�ft made the deleting operation.
    3)Save the unfinished working in the client side. If a teacher has done half of his work, he can choose to press the Save button to store what he has done in his computer. The reason for doing so is simply to alleviate the burden of the server. Although all finished test papers must be saved in a database on the server, sometimes the unfinished papers could be dropped forever or could form the ultimate testing papers after several months. So if these papers keep in the server, it will waste the server computer�fs room. Next time the teacher can press the Restore button on the page to get the previously stored part of the test paper from his own computer and continue to finish the whole paper.
    4)Allow at least 1,000 teachers to make test papers at the same time. The maximum question number per examination paper is 60.
    Here are my two rough solutions,
    A.Using JSF.
    B.Using JavaScript and plain JSP[b] without JSF.
    The comparison of the two solutions:
    1)Both schemas can implement the first and the second requirements. In JSF page I could add a standard panelGird tag and use its binding attribute. In the backing bean, the method specified by the binding attribute is responsible for generating HtmlInput objects and adding them to the HtmlPanelGird object on the fly. Every HtmlInput object is corresponding to a question subject or an optional answer. The method is called by an actionListener, which is registered in the Next Quiz commandButton, triggering by the clicking on this button in the client side. Using JSF can also be prone to managing the HtmlInput objects, e.g. panelGird.getChildren().add(HtmlInput) and panelGird.getChildren().remove(HtmlInput) respond to the undoing operation of deleting an optional answer and the redoing operation of the deleting action respectively. I know JavaScript can also achieve these goals. It could be more complex since I don�ft know well about JavaScript.
    2)I can not find a way to meet the third demand right now. I am eager to know your suggestion.
    3)Using JSF, I think, can�ft allow 1,000 teachers to do their own papers at the same time. Because in this scenario, suppose each questionnaire having 60 questions and 4 answers per question, there will be approximately 300,000 HtmlInput objects (1,000X60X(4+1)) creating on the server side. The server must run out of memory undoubtedly. To make things better, we can use a custom component which can be rendered as a whole question including its all optional answers. That is to say, a new custom component on the server side stands for a whole question on the client side. Even so, about 60,000(1,000X60) this type of custom components will be created progressively and dynamically, plus other UISelectOne and UICommand objects, it also can�ft afford for most servers. Do I have to use JavaScript to avoid occupying the server's memory in this way? If so, I have to go back and use JavaScript and plain JSP without JSF.
    Thank you in advance!
    Best Regards, Ailsa
    2007/5/4

    Thank you for your quick response, BalusC. I really appreciate your answer.
    Yes, you are right. If I manually code the same amount of those components in the JSF pages instead of generating them dynamically, the server will still run out of memory. That is to say, JSF pages might not accommodate a great deal of concurrent visiting. If I upgrade the server to just allow 1,000 teachers making their own test papers at the same time, but when over 2,000 students take the same questionnaire simultaneously, the server will need another upgrading. So I have to do what you have told me, using JS+DOM instead of upgrading the server endlessly.
    Best Regards, Ailsa

  • Lightroom 5 permanently runs out of memory

    Lightroom 5 on Windows 7 32 Bit and 8 Gigabytes of memory (more than the 32 Bit system can use) permanently runs out of memory when doing some more complex edits on a RAW file, especially when exporting to 16 Bit TIFF. The RAW files were created by cameras with 10 up to 16 megapixel sensors with bit depths between 12 and 14.
    After exporting one or two images to 16 Bit uncompressed TIFF an error message "Not enough memory" will be displayed and only a Lightroom restart solves that - for the next one to two exports. If an image has much brush stroke edits, every additional stroke takes more and more time to see the result until the image disappears followed by the same "Not enough memory" error message.
    A tab character in the XMP sidecar file is *not* the reason (ensured that), as mentioned in a post. It seems that Lightroom in general does not allocate enough memory and frees too less/late allocated.
    Please fix that bug, it's not productive permanently quit and restart Lightroom when editing/exporting a few RAW files. Versions prior to Lightroom 4 did not have that bug.
    P.S. Posting here, because it was not possible to post it at http://feedback.photoshop.com/photoshop_family/topics/new It's very bad design, to let a user take much time to write and then say: "Log in", but a log in with the Adobe ID and password does not work (creating accounts on Facebook etc. is not an acceptable option, Adobe ID should be enough). Also a bugtracker such as Bugzilla would be a much better tool for improving a software and finding relevant issues to avoid duplicate postings.

    First of all: I personally agree with your comments regarding the feedback webpage. But that is out of our hands since this is a user-to-user forum, and there is nothing we users can do about it.
    Regarding your RAM: You are running Win7 32-bit, so 4 GB of your 8 GB of RAM sit idle since the system cannot use it. And, frankly, 4 GB is very scant for running Lr, considering that the system uses 1 GB of that. So there's only 3 GB for Lr - and that only if you are not running any other programs at the same time.
    Since you have a 8 GB system already, why don't you go for Win7 64-bit. Then you can also install Lr 64-bit and that - together with 8 GB of RAM - will bring a great boost in Lr performance.
    Adobe recommends to run Lr in the 64-bit version. For more on their suggestion on improving Lr performance see here:
    http://helpx.adobe.com/lightroom/kb/performance-hints.html?sdid=KBQWU
    for more: http://forums.adobe.com/thread/1110408?tstart=0

Maybe you are looking for