Running out of handles in LabVIEW 7.1

Hi,
I'm having a problem with a LabVIEW 7.1 application running on Linux (2.6.12). About once a week the application crashes locking up the entire system.
While reproducing the problem at the office, it looked like the application ran out of handles. There was 1 network connection open, the listen task couldn't spawn a new one, untill the open one was closed. It didn't even matter if the connection was made from another PC. There also were problems with opening files (log files weren't updated anymore). The application has a couple of tasks, some of them weren't running anymore, some were.
Are there any known bugs for LabVIEW 7.1 for Linux?
Can anyone give me some pointers in solving this problem?
thanx,
Dennis

Dennisvr wrote:
Hi,   I'm
having a problem with a LabVIEW 7.1 application running on Linux
(2.6.12). About once a week the application crashes locking up the
entire system.   While reproducing the
problem at the office, it looked like the application ran out of
handles. There was 1 network connection open, the listen task couldn't
spawn a new one, untill the open one was closed. It didn't even matter
if the connection was made from another PC. There also were problems
with opening files (log files weren't updated anymore). The application
has a couple of tasks, some of them weren't running anymore, some were.   Are there any known bugs for LabVIEW 7.1 for Linux?   Can anyone give me some pointers in solving this problem?
Are
you sure you are closing all LabVIEW refnums properly you ever open.
TCP Listener refnums, TCP connection refnums, file refnums, each of
them will bind in internally to a system file handle under Unix and as
long as that LabVIEW refnum is not closed, will use up one of those
handles. LabVIEW only cleans up unclosed refnums after the top level VI
in whose hierarchy the refnum was created goes idle.
One thing to check out may be that the according Close function does
really close a refnum even though its "error in" contains an error
indication. That should be how it works but I've found in the past
areas where this didn't always be the case.
Rolf Kalbermatter
Message Edited by rolfk on 02-15-2006 11:02 PM
Rolf Kalbermatter
CIT Engineering Netherlands
a division of Test & Measurement Solutions

Similar Messages

  • I have an iMac and I am running out of disc space. I am considering an external drive to house my iTunes library. If I buy a second drive to operate as a back up will time machine back up the disc on the IMac and the external drive with the iTunes library

    I have an iMac and I am running out of disc space. I am considering an external drive to house my iTunes library. If I buy a second drive to operate as a back up will time machine back up the disc on the IMac and the external drive with the iTunes library?

    If you have a considerable photograph and/or movie collection you may want to consider off loading those to the EHD too. Here are Apple's article showing how handle iTunes & iPhoto libraries.
    iTunes: library on EHD
    iPhoto: How to move the Library to EHD
    Roger

  • Running out of memory despite having set je.maxMemory to a moderate value

    I have set je.maxMemory to 20MB (je.maxMemory=20000000) and allowed a max heap size of 512MB (-Xms256M -Xmx512M).
    After two hours of running my web service, I'm running out of memory. After having profiled my service (using Yourkit Java Profiler 1.10.6), I can see the following:
    Name                                               Objects ShallowSize  RetainedSize
    byte[]                                               16711   124124880     124124880
    com.sleepycat.je.tree.BIN                              181       24616     116254200
    com.sleepycat.je.tree.Node[]                           187       98736     115743184
    com.sleepycat.je.tree.LN                              7092      226944     115253600
    java.util.concurrent.ConcurrentHashMap$HashEntry       554       17728      78328944
    java.util.concurrent.ConcurrentHashMap$HashEntry[]    1053       34728      77489632
    java.util.concurrent.ConcurrentHashMap                 117        5616      71812072
    java.util.concurrent.ConcurrentHashMap$Segment[]       118       10304      71807912
    java.util.concurrent.ConcurrentHashMap$Segment        1052       42080      71798808
    com.sleepycat.je.tree.IN                                 6         672      45592352
    java.lang.String                                    135888     4348416      14152664The memory profiler claims further, that com.sleepycat.je.tree.BIN is responsible for 71% of all heap memory.
    In any case, com.sleepycat.je.tree.BIN claims ~ 116MB of heap memory, which is by any goodwill, exceeded the limit of 20MB.
    How can this be?
    How is JE ensuring that the limit is not exceeded? Is there a timer (thread) running which once a while checks the memory used and then cleans up ; or is memory usage checked creating a com.sleepycat.je.tree.BIN object?
    My environment:
    BDB JE 4.0.92 - used as cache loader within Jboss Cache (3.2.7.GA), running on a JBOSS Application Server, Java 1.6 (IBM) on Linux. Further details are listed in the system properties below (except some deleted security items).
    System properties:
    (java.lang.String, int, java.lang.StringBuffer, int)=contains
    DestroyJavaVM helper thread=(java.lang.String, java.security.KeyStore$Entry, java.security.KeyStore$ProtectionParameter)
    base.collection.name=CD2JAVA
    bind.address=10.12.25.130
    catalina.base=/work/ocrgws_test/server0
    catalina.ext.dirs=/work/ocrgws_test/server0/lib
    catalina.home=/work/ocrgws_test/server0
    catalina.useNaming=false
    com.arjuna.ats.arjuna.objectstore.objectStoreDir=/work/ocrgws_test/server0/data/tx-object-store
    com.arjuna.ats.jta.lastResourceOptimisationInterface=org.jboss.tm.LastResource
    com.arjuna.ats.tsmx.agentimpl=com.arjuna.ats.internal.jbossatx.agent.LocalJBossAgentImpl
    com.arjuna.common.util.logger=log4j_releveler
    com.arjuna.common.util.logging.DebugLevel=0x00000000
    com.arjuna.common.util.logging.FacilityLevel=0xffffffff
    com.arjuna.common.util.logging.VisibilityLevel=0xffffffff
    com.ibm.cpu.endian=little
    com.ibm.jcl.checkClassPath=
    com.ibm.oti.configuration=scar
    com.ibm.oti.jcl.build=20100326_1904
    com.ibm.oti.shared.enabled=false
    com.ibm.oti.vm.bootstrap.library.path=/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs:/opt/ibm/java-x86_64-60/jre/lib/amd64
    com.ibm.oti.vm.library.version=24
    com.ibm.util.extralibs.properties=
    com.ibm.vm.bitmode=64
    common.loader=${catalina.home}/lib,${catalina.home}/lib/*.jar
    epo.jboss.deploymentscanner.extradirs=/work/ocrgws_test/app/
    external.cert.ldap.* = ***************
    file.encoding=UTF-8
    file.separator=/
    flipflop.activation.time=16:30
    hibernate.bytecode.provider=javassist
    ibm.signalhandling.rs=false
    ibm.signalhandling.sigchain=true
    ibm.signalhandling.sigint=true
    ibm.system.encoding=UTF-8
    jacorb.config.log.verbosity=0
    java.assistive=ON
    java.awt.fonts=
    java.awt.graphicsenv=sun.awt.X11GraphicsEnvironment
    java.awt.printerjob=sun.print.PSPrinterJob
    java.class.path=/work/ocrgws_test/config:/usr/local/jboss-eap-4.3-cp07/bin/run.jar:/opt/ibm/java-x86_64-60/lib/tools.jar
    java.class.version=50.0
    java.compiler=j9jit24
    java.endorsed.dirs=/usr/local/jboss-eap-4.3-cp07/lib/endorsed
    java.ext.dirs=/opt/ibm/java-x86_64-60/jre/lib/ext
    java.fullversion=JRE 1.6.0 IBM J9 2.4 Linux amd64-64 jvmxa6460sr8-20100401_55940 (JIT enabled, AOT enabled)
    J9VM - 20100401_055940
    JIT - r9_20100401_15339
    GC - 20100308_AA_CMPRSS
    java.home=/opt/ibm/java-x86_64-60/jre
    java.io.tmpdir=/tmp
    java.jcl.version=20100408_01
    java.library.path=/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs:/opt/ibm/java-x86_64-60/jre/lib/amd64:/usr/lib64/mpi/gcc/openmpi/lib64:/usr/lib
    java.naming.factory.initial=org.jnp.interfaces.NamingContextFactory
    java.naming.factory.url.pkgs=org.jboss.naming:org.jnp.interfaces
    java.net.preferIPv4Stack=true
    java.protocol.handler.pkgs=org.jboss.net.protocol
    java.rmi.server.codebase=http://10.12.25.130:8083/
    java.rmi.server.hostname=10.12.25.130
    java.rmi.server.randomIDs=true
    java.runtime.name=Java(TM) SE Runtime Environment
    java.runtime.version=pxa6460sr8-20100409_01 (SR8)
    java.security.krb5.conf=/usr/local/jboss/etc/krb5.conf
    java.specification.name=Java Platform API Specification
    java.specification.vendor=Sun Microsystems Inc.
    java.specification.version=1.6
    java.util.prefs.PreferencesFactory=java.util.prefs.FileSystemPreferencesFactory
    java.vendor.url=http://www.ibm.com/
    java.vendor=IBM Corporation
    java.version=1.6.0
    java.vm.info=JRE 1.6.0 IBM J9 2.4 Linux amd64-64 jvmxa6460sr8-20100401_55940 (JIT enabled, AOT enabled)
    J9VM - 20100401_055940
    JIT - r9_20100401_15339
    GC - 20100308_AA_CMPRSS
    java.vm.name=IBM J9 VM
    java.vm.specification.name=Java Virtual Machine Specification
    java.vm.specification.vendor=Sun Microsystems Inc.
    java.vm.specification.version=1.0
    java.vm.vendor=IBM Corporation
    java.vm.version=2.4
    javax.management.builder.initial=org.jboss.mx.server.MBeanServerBuilderImpl
    javax.net.ssl.trustStore=/usr/local/jboss/etc/ldap.truststore
    javax.net.ssl.trustStorePassword=password
    jboss.bind.address=10.12.25.130
    jboss.home.dir=/usr/local/jboss-eap-4.3-cp07
    jboss.home.url=file:/usr/local/jboss-eap-4.3-cp07/
    jboss.identity=30df88bc0a52e350x6e2ff59cx136c17794d5x-8000757
    jboss.lib.url=file:/usr/local/jboss-eap-4.3-cp07/lib/
    jboss.messaging.controlchanneludpaddress=239.1.200.4
    jboss.messaging.datachanneludpaddress=239.1.200.4
    jboss.partition.name=ocrgws_test_Partition
    jboss.partition.udpGroup=239.1.200.4
    jboss.remoting.domain=JBOSS
    jboss.remoting.instanceid=30df88bc0a52e350x6e2ff59cx136c17794d5x-8000757
    jboss.remoting.jmxid=luu002t.internal.epo.org_1334685694459
    jboss.remoting.version=22
    jboss.security.disable.secdomain.option=true
    jboss.server.config.url=file:/work/ocrgws_test/server0/conf/
    jboss.server.data.dir=/work/ocrgws_test/server0/data
    jboss.server.home.dir=/work/ocrgws_test/server0
    jboss.server.home.url=file:/work/ocrgws_test/server0/
    jboss.server.lib.url=file:/work/ocrgws_test/server0/lib/
    jboss.server.log.dir=/work/ocrgws_test/server0/log
    jboss.server.name=luu002t_ocrgws_test_server0
    jboss.server.temp.dir=/work/ocrgws_test/server0/tmp
    jboss.tomcat.udpGroup=239.1.200.4
    jbossmx.loader.repository.class=org.jboss.mx.loading.UnifiedLoaderRepository3
    je.maxMemory=20000000
    jgroups.bind_addr=10.12.25.130
    jmx.console.bindcredential=3bpwdmpc
    jmx.console.binddn=cn=jbossauth-ro,ou=accounts,ou=auth,dc=epo,dc=org
    jmx.console.rolesctxdn=ou=roles-test,ou=jboss,ou=applications,ou=internal,dc=epo,dc=org
    jndi.datasource.name=java:MainframeDS
    jnp.disableDiscovery=true
    jxe.current.romimage.version=15
    jxe.lowest.romimage.version=15
    line.separator=
    mainframelogin.password=720652a1e842fc7f
    mainframelogin.username=test_t
    org.apache.commons.logging.Log=org.apache.commons.logging.impl.Log4JLogger
    org.apache.tomcat.util.http.ServerCookie.VERSION_SWITCH=true
    org.epo.jboss.application.home=/work/ocrgws_test
    org.hyperic.sigar.path=/work/ocrgws_test/server0/./deploy/hyperic-hq.war/native-lib
    org.jboss.ORBSingletonDelegate=org.jacorb.orb.ORBSingleton
    org.omg.CORBA.ORBClass=org.jacorb.orb.ORB
    org.omg.CORBA.ORBSingletonClass=org.jboss.system.ORBSingleton
    org.w3c.dom.DOMImplementationSourceList=org.apache.xerces.dom.DOMXSImplementationSourceImpl
    os.arch=amd64
    os.name=Linux
    os.version=2.6.32.46-0.3-xen
    package.access=sun.,org.apache.catalina.,org.apache.coyote.,org.apache.tomcat.,org.apache.jasper.,sun.beans.
    package.definition=sun.,java.,org.apache.catalina.,org.apache.coyote.,org.apache.tomcat.,org.apache.jasper.
    path.separator=:
    poll.interval.milliseconds=300000
    program.name=run.sh
    server.loader=
    shared.loader=
    spnego.config=/usr/local/jboss/etc/spnego.properties
    sun.arch.data.model=64
    sun.boot.class.path=/usr/local/jboss-eap-4.3-cp07/lib/endorsed/xercesImpl.jar:/usr/local/jboss-eap-4.3-cp07/lib/endorsed/xalan.jar:/usr/local/jboss-eap-4.3-cp07/lib/endorsed/serializer.jar:/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs/jclSC160/vm.jar:/opt/ibm/java-x86_64-60/jre/lib/annotation.jar:/opt/ibm/java-x86_64-60/jre/lib/beans.jar:/opt/ibm/java-x86_64-60/jre/lib/java.util.jar:/opt/ibm/java-x86_64-60/jre/lib/jndi.jar:/opt/ibm/java-x86_64-60/jre/lib/logging.jar:/opt/ibm/java-x86_64-60/jre/lib/security.jar:/opt/ibm/java-x86_64-60/jre/lib/sql.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmorb.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmorbapi.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmcfw.jar:/opt/ibm/java-x86_64-60/jre/lib/rt.jar:/opt/ibm/java-x86_64-60/jre/lib/charsets.jar:/opt/ibm/java-x86_64-60/jre/lib/resources.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmpkcs.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmcertpathfw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjgssfw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjssefw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmsaslfw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjcefw.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjgssprovider.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmjsseprovider2.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmcertpathprovider.jar:/opt/ibm/java-x86_64-60/jre/lib/ibmxmlcrypto.jar:/opt/ibm/java-x86_64-60/jre/lib/management-agent.jar:/opt/ibm/java-x86_64-60/jre/lib/xml.jar:/opt/ibm/java-x86_64-60/jre/lib/jlm.jar:/opt/ibm/java-x86_64-60/jre/lib/javascript.jar:/tmp/yjp201202191932.jar
    sun.boot.library.path=/opt/ibm/java-x86_64-60/jre/lib/amd64/compressedrefs:/opt/ibm/java-x86_64-60/jre/lib/amd64
    sun.io.unicode.encoding=UnicodeLittle
    sun.java.command=org.jboss.Main -b 10.12.25.130 -Djboss.server.home.dir=/work/ocrgws_test/server0 -Djboss.server.home.url=file:/work/ocrgws_test/server0 -Djboss.server.name=luu002t_ocrgws_test_server0 -Djboss.partition.name=ocrgws_test_Partition -Depo.jboss.deploymentscanner.extradirs=/work/ocrgws_test/app/ -Dorg.epo.jboss.application.home=/work/ocrgws_test
    sun.java.launcher.pid=17781
    sun.java.launcher=SUN_STANDARD
    sun.java2d.fontpath=
    sun.jnu.encoding=UTF-8
    sun.rmi.dgc.client.gcInterval=3685000
    sun.rmi.dgc.server.gcInterval=3685000
    system=java.io.ObjectStreamField
    tomcat.util.buf.StringCache.byte.enabled=true
    user.country=US
    user.dir=/work/ocrgws_test
    user.home=*****************
    user.language=en
    user.name=***********
    user.timezone=Europe/Berlin
    user.variant=

    The memory profiler claims further, that com.sleepycat.je.tree.BIN is responsible for 71% of all heap memory. In any case, com.sleepycat.je.tree.BIN claims ~ 116MB of heap memory, which is by any goodwill, exceeded the limit of 20MB. >
    I'm not sure whether the profiler is reporting live objects only (referenced) or all objects (including those not yet reclaimed). If the latter, it isn't telling you how much memory is actually referenced by the JE cache.
    Please look at the JE stats to see what the cache usage is, from JE's point of view.
    If you believe there is a bug in JE cache management, you'll need to write a small standalone test to demonstrate it and submit it to us, since we don't know of any such bug. Also note that we'll have difficulty supporting JE 4.0 (without a support contract anyway). Please use JE 5.0, or at least 4.1.
    Eviction occurs as objects are allocated, as well as in background threads. Eviction in background threads and concurrent eviction were greatly improved in JE 4.1.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Running out of OS level space.

    Hi,
    Can you guys help me out in this problem. I am using UNiX as OS and running 9i database Running on SUN Machines. For some reason my oracle partition at OS level is running out of space. It's not because my datafiles are full or any file is reaching on the edge but for some reason any file is eating the space of my Oracle Partition at OS level ...Now I am really having hard time to find out what is the cause of this problem.
    If anybody can help me out and if possible give me any Unix command or any sort of resource from which i can find out which thing is eating my Oracle partition space at OS level
    REGARDS,

    Hello,
    You can have a small shell script to monitor the filesystem freespace and notify via email/pager. Ofcourse you have to handle couple things on database side.
    You might have to change the script a bit as per your environment.
    << Sample Script>>
    #!/bin/ksh
    SPACE_THRESHOLD=$1
    ALERTCNT=0
    NOTIFY=<your email id>
    PAGER=<your pager id>
    if [ `ls -l a.log | wc -l` -eq 1 ] ;
    then
    `rm -f a.log
    fi
    if [ `ls -l a.pag | wc -l` -eq 1 ] ;
    then
    `rm -f a.pag
    fi
    # Logic to check the FS
    for i in `df -k | grep -v /cdrom/ | grep '%' | awk '{ print $6 }'`
    do
    SPACE=`df -k $i | grep '%' | awk '{ print $5 }' | sed s/"%"//g`
    MNTPT=`df -k $i | awk '{print $1}'`
    if [ ${SPACE} -gt ${SPACE_THRESHOLD} ];
    then
    echo " $i - $SPACE% Full ">> a.log
    echo " $i - $SPACE% Full ">> a.pag
    ALERTCNT=`expr ${ALERTCNT} + 1`
    fi
    done
    # Notify in case of issues
    if [ ${ALERTCNT} -gt 0 ];
    then
    mailx -s "FS Report `date`" ${NOTIFY}<a.log
    mailx -s "FS Report `date`" ${PAGER}<a.pag
    else
    exit;
    fi
    All the best.
    -Sri

  • Workfow Iterate to subprocess runs out of memory

    I have a workflow that returns all suspended tasks and then calls a subprocess for each task. The subprocess decides whether the task needs to be deleted and if so it processes the task in various ways before deleting the taskinstance.
    I have no issues when there are not too many tasks returned by the query but when the workflow returns 2000+ items, I run out of memory.
    What is the best way to workflow to call the subprocess without running out of memory?
    Do I need to cleanup something at the end of subprocess?
    Do I need to add something in the workflow to beakup the list of tasks into smaller chunks?
    <Activity id='3' name='ProcessTasks'>
    <Action id='0' name='processTasks' process='processTheTask'>
    <Iterate for='taskInstanceName' in='mytasks'/>
    <Argument name='taskInstanceName' value='$(taskInstanceName)'/>
    </Action>
    Edited by: user1937458 on Mar 14, 2012 3:12 PM

    I didnt think that this would put that much stress on the system.
    1) Use IDM best practice to generate low memory tasks, use exposedVariables and extendedVariables in manual actions to generate low memory tasks, that will save lots of memory.
    No manual action. This is a scheduled task.
    2) Run this workflow on dedicated server which is responsible to run this task only.
    I have run this when no one else was using the system but that did not help either.
    3) You can put some more conditions to get the limited return data which your server can handle in one go.
    we normally have 8000 tasks in the system. About 5000 are completed so I can ignore those in the workflow. The rest need to be looked at to determine if we need to update the request. Let's say that I can use a rule to determine that in the workflow before the subprocess is called and I end up with a list of 500 taskinstance names, I think that the process will still run out of memory unless there is some other solution.
    2000 task names in a list should not take up that much space. I am pretty sure that the subprocess which determines if the task needs to be deleted is chewing up resources. This is going to be a scheduled task with no manual actions.
    My thinking was that workflow calls the subprocess and the subprocess does a lot of work as far as canceling a request, disabling accounts in some cases, auditing and notifying users that their request was cancelled. Upon return to the workflow to get the next taskinstance name, there is probably some variable that keeps getting larger with each iteration.
    I have run smaller lists and the flow diagram that returns at the end shows the flowchart for every item that was deleted so that is probably 1 place where the variable keeps getting larger.
    Is there a way to clean everything so that each subprocess acts as if it was the 1st and only time it was getting called?
    I tried the following at the end of the subprocess but that did not help:
    <Action id='0' name='CleanUp'>
    <expression>
    <set name='WF_CASE_RESULT'/>
    </expression>
    </Action>
    I will try to debug and see what variables are getting larger and larger but any other suggestions are appreciated.

  • Run a cmd command in LabVIEW with parameters

    I have an executable program, for example myprogram.exe, located in "C:\Documents and Settings\myname\myprogram.exe" (I use windows xp 32 bit)
    When I run the program (by running cmd and type "myprogram.exe > to_text.txt" on the command line), the program will print out the result to that text file.
    But when I try to run the whole thing on LabVIEW, I can't manage to do so. LabVIEW only supports using the "Run Command" of Windows, not the cmd. (using the SystemExec vi)
    So how can I run "myprogram.txt > to_text.txt" on LabVIEW as I run it in cmd? Moreover, how can I send the break event (by pressing Ctrl+C in the cmd) by LabVIEW to stop the program?
    With my appreciations!

    I made this a while ago and haven't tested it too much but it should do what you want.  Provide the application path, then the switches in the array.  It does work with built in commands like dir and copy.
    Unofficial Forum Rules and Guidelines - Hooovahh - LabVIEW Overlord
    If 10 out of 10 experts in any field say something is bad, you should probably take their opinion seriously.
    Attachments:
    Execute Command Line With Switches.vi ‏24 KB

  • Target has run out of memory on LM3s8962

    I'm using the LM3s8962 evaluation kit to record data from the ADC's.  I have the system set up so that I use the elemental nodes of the four adc's in a while loop, and replace the values in four different arrays.  The arrays are initialize (1x1000 elements) before entering the loop.  This works fine.
    THE PROBLEM:  When I try to make the arrays larger (i.e. initial arrays larger than 1000 points, 4 individual arrays), I get the following error:
    Error: Memory allocation failed. The target has run out of memory. [C:\Program Files (x86)\National Instruments\LabVIEW 2011\CCodeGen\libsrc\blockdiagram\CCGArrSupport2.c at line 253: 2 3
    OR
    Error: Memory allocation failed. The target has run out of memory. [C:\Program Files (x86)\National Instruments\LabVIEW 2011\CCodeGen\libsrc\blockdiagram\CCGArrSupport2.c at line 173: 2 3
    Any suggestions?

    Th0r wrote:
    It looks like you're filling up the flash memory on the LM3S8962 with all of these array initializations.  According to page 263 of the LM3S8962 datasheet, that microcontroller has 256 KB of flash memory which you can use to fill up with your code.  In addition to your array initializations, some of this space is taken up by the LabVIEW Embedded Module-specific code as well.  What datatype are you using in these arrays?  Does this error occur upon building or running your code?  Thanks for any additional information you can provide!  
    That's probably it.  The error occurs when building the code, before it's actually able to run.  If reduce the array size, I'm able to run the code no problem.  At the moment,  I'm using a long 32 bit integer, which I know realize I can reduce significantly, as my ADC only reads at 10 bits.  Do you know if there's a way that I can preallocate the array to a place other than flash?
    I've found a fix around it since I last posted, in which I set up a buffer (smaller) and then save the buffer values on the SD card.  This works well and I can sample for long periods of time, but it does slow down my overall sampling rate, so I'd like to fix the above problem nonetheless. 

  • TM: When external hard drive space runs out, can I add another drive?

    I am using an external hard drive 1TB to back-up my new iMac with TM. When the external drive space runs out, do I have to buy another drive with much more space (e.g. 2-3TB) and start over, or can I just add a second external hard drive 1TB (will TM recognize the second one as an addition)?
    Thanks for any help.

    First off you should know that when Time Machine runs out of space on an external drive, it deletes the oldest backup you have. So if you are running out of space you probably don't need to worry as Time Machine will handle it. Just make sure the Time Machine backup drive has a higher capacity than the amount of data you have on your internal drive and you'll be fine. However, if you'd like to keep a lot of backups over a period of time you may want to consider getting a whole new drive. The only way you can add more drives to back up to is by using RAID 0 which I wouldn't recommend unless you have to due to the fact that you'd be more likely to lose data. The easiest option would be to buy a 2TB drive and use it with Time Machine.

  • Running out of connections...

     

    Hello,
    I think I'm having the same problem ..
    Did you solve it ? How ?
    Thanks a lot.
    Franco
    Francesco Costa
    BEA Systems Italia
    [email protected]
    Phone +39 02 3327 3000
    Mobile + 39 335 1246931
    Fax +39 02 3327 3001
    "Gerhard Henning" <[email protected]> ha scritto nel messaggio
    news:[email protected]...
    >
    Hi,
    we seem to have a similar problem. Can you please tell me,
    which service pack solves the problem.
    We are using SP8 but the problem seems to be still there!
    Thanks
    Gerhard
    Joseph Weinstein <[email protected]> wrote:
    Gene Chuang wrote:
    Hi Joe,
    Good news; bug ISOLATED and FIXED! First off, I was able to determinethis bug only manifests if
    we are running clustered apps. under a single app server, this bugdoes not appear, and under
    clustered app this bug always appears.
    Furthermore, I appologize if I may have lead you on a wild goosechasewith all the subsequent
    patches you sent me. As it turns out, when I received each patch Ihanded it off to our tester, who
    applied the patch and probably only bounced a node in our server
    cluster
    before testing, hence
    getting negative results. Today I applied your latest patch,
    clean-bounced
    all the nodes, and the
    bug disappeared! Thanks Joe!
    Now the 2 questions of the day:
    1) Is your latest patch production worthy? Can I apply it on ourlive site?
    Yes.
    2) If not, when's the next service pack coming out with this fix?It will be in future service packs.
    Joe
    Gene
    "Joseph Weinstein" <[email protected]> wrote in message
    news:[email protected]...
    >>>>
    >>>>
    Gene Chuang wrote:
    Hi Joe,
    We tried it and found the same exception raised, except for this
    time it took a lot longer to
    bug
    out; the server hung for almost 2 minutes before dumping the
    following.
    However I was just
    informed by another developer that we may have a recursive bugon our side, and that we are
    doing a
    non-terminating db search. Perhaps this is causing this
    out-of-resource
    allocation bug, i.e.
    our
    bug exposed your bug! I'm sure once we fix our recursion bug we
    won't see this again, but maybe
    something should also be done on your side that'll handle thismore gracefully...
    I'll let you know what happens once we fix our bug; thanks forthe help!
    DO let me know. In fact, please use the attached jar (in place ofall the others:-)
    because the line ResourceAllocator.java:561 is not executable code.There may be a
    JVM problem... I'm fairly sure there's no way our code can causean NPE in this
    area...
    Joe
    javax.ejb.FinderException: Couldn't build CallableStatement from
    JDBCCommandImpl:
    java.sql.SQLException: Exception raised by connection pool
    :java.lang.NullPointerException
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:56
    1)
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:55
    5)
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:50
    2)
    atweblogic.jdbc.common.internal.ConnectionPool.reserve(ConnectionPool.java:172
    atweblogic.jdbc.common.internal.ConnectionPool.reserveWaitSecs(ConnectionPool.
    java:145)
    atweblogic.jdbcbase.jts.Connection.openConnectionIfNecessary(Connection.java:5
    87)
    atweblogic.jdbcbase.jts.Connection.prepareCall(Connection.java:159)
    atweblogic.jdbc20.rmi.internal.ConnectionImpl.prepareCall(ConnectionImpl.java:
    89)
    atweblogic.jdbc20.rmi.SerialConnection.prepareCall(SerialConnection.java:71)
    atcom.kiko.db.JDBCConnection.getCallableStatement(JDBCConnection.java:158)
    >>>>>
    Gene Chuang
    Join Kiko.com!
    "Joseph Weinstein" <[email protected]> wrote in message
    news:[email protected]...
    >>>>>>
    >>>>>>
    Gene Chuang wrote:
    Hi Joe,
    We were finally able to consistently replicate the
    ResourceAllocator.reserve()
    stacktrace,
    although
    from a slightly different route. I tried your jar, and now
    have a slightly different
    exception:
    Hi Gene. Thank you for your patience. Here's one last jar file.
    Ditch the previous one. I appreciate your energy and willingness
    to help. Let me know,
    Joe
    javax.ejb.EJBException: Failed to select row 0: Exception
    raised
    by connection pool
    :java.lang.NullPointerException
    at
    weblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:56
    1)
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:55
    5)
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:50
    2)
    atweblogic.jdbc.common.internal.ConnectionPool.reserve(ConnectionPool.java:172
    atweblogic.jdbc.common.internal.ConnectionPool.reserveWaitSecs(ConnectionPool.
    java:145)
    atweblogic.jdbcbase.jts.Connection.openConnectionIfNecessary(Connection.java:5
    87)
    atweblogic.jdbcbase.jts.Connection.prepareStatement(Connection.java:135)
    atweblogic.jdbc20.rmi.internal.ConnectionImpl.prepareStatement(ConnectionImpl.
    java:80)
    atweblogic.jdbc20.rmi.SerialConnection.prepareStatement(SerialConnection.java:
    55)
    atcom.kiko.db.JDBCConnection.getPreparedStatement(JDBCConnection.java:141)
    >>>>>>>
    Gene Chuang
    Join Kiko.com!
    "Joseph Weinstein" <[email protected]> wrote in message
    news:[email protected]...
    Hi Gene.
    Would you put the attached jar to the head of your
    weblogic.classpath,
    and let me know if this fixes the problem? thanks,
    Joe
    Gene Chuang wrote:
    WL 5.1 sp 6, running on Solaris 2.7 over Oracle 8i
    Gene Chuang
    Join Kiko.com!
    "Joseph Weinstein" <[email protected]> wrote in message
    news:[email protected]...
    Gene Chuang wrote:
    Hi Joe,
    Now I'm getting this slightly different stack trace...
    do u think it's the same
    "out of
    connection"
    error, or something else?It's something else. What version of the product is this?
    Joe
    Couldn't build PreparedStatement from JDBCCommandImpl:
    java.sql.SQLException:
    Exception
    raised
    by
    connection pool :java.lang.NullPointerException
    at
    weblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:56
    1)
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:55
    5)
    atweblogic.common.internal.ResourceAllocator.reserve(ResourceAllocator.java:50
    2)
    atweblogic.jdbc.common.internal.ConnectionPool.reserve(ConnectionPool.java:172
    atweblogic.jdbc.common.internal.ConnectionPool.reserveWaitSecs(ConnectionPool.
    java:145)
    atweblogic.jdbcbase.jts.Connection.openConnectionIfNecessary(Connection.java:5
    87)
    atweblogic.jdbcbase.jts.Connection.prepareStatement(Connection.java:135)
    atweblogic.jdbc20.rmi.internal.ConnectionImpl.prepareStatement(ConnectionImpl.
    java:80)
    atweblogic.jdbc20.rmi.SerialConnection.prepareStatement(SerialConnection.java:
    55)
    >>>>>>>>>>>
    Gene Chuang
    Join Kiko.com!
    "Joseph Weinstein" <[email protected]> wrote in message
    news:[email protected]...
    Hi.
    The server won't let an EJB wait indefinitely for
    a connection to become
    available in a pool. All the while it waits, it iscommandeering one
    of the fixed number of execute-threads, doing nothingwhile the thread
    could possibly run another task which already hasit's pool connection,
    and could release it if it ran. The default waitallowed is 5 seconds
    the server decides to devote it's resources to thework it can do, and
    not let overflow requests cause an unintentionalor malicious denial-
    of-service attack. The fundamental solution is tosupply enough DBMS
    connections to serve the load you have.
    Joe
    Gene Chuang wrote:
    Hi,
    Here's our pool init info:
    weblogic.jdbc.connectionPool.oraclePool=\
    driver=oracle.jdbc.driver.OracleDriver,\
    url=jdbc:oracle:oci8:@xxx,\
    loginDelaySecs=0,\
    initialCapacity=4,\
    maxCapacity=10,\
    capacityIncrement=2,\
    allowShrinking=true,\
    shrinkPeriodMins=15,\
    refreshMinutes=10,\
    testTable=dual,\
    props=user=xxx;password=xxxx;server=xxx
    I run a db-intensive data migration program on
    my ejbs (hence JDBC connection
    requests
    are
    made
    on
    the serverside). After a while, I run out of
    connections,
    and the exception
    pasted
    below is
    thrown.
    Why is this exception thrown? If the pool is at
    maxCapacity and all connections
    are
    taken,
    shouldn't weblogic place the connection-requestor
    on wait until a connection is
    available?
    And
    I'm
    pretty sure I return the connections because I
    wrap connection.close() on a
    finally
    block...
    JDBCPoolConnection.connect: Problem connecting
    Exception:
    java.sql.SQLException: Pool connect failed: Noneavailable
    java.sql.SQLException: Pool connect failed: Noneavailable
    at
    weblogic.jdbcbase.pool.Driver.connect(Driver.java:184)
    atweblogic.jdbcbase.jts.Driver.connect(Driver.java:233)
    at
    weblogic.jdbc20.common.internal.RmiDataSource.getConnection(RmiDataSourc
    e.java:55)
    at
    weblogic.jdbc20.common.internal.RmiDataSource_ServiceStub.getConnection(
    RmiDataSource_ServiceStub.java:179)
    at
    com.kiko.db.JDBCPoolConnection.connect(JDBCPoolConnection.java:24)ABMEB.
    ejbActivate: Getting bean out of pooled state.
    Gene Chuang
    Join Kiko.com!--
    PS: Folks: BEA WebLogic is in S.F. with both entryand advanced positions for
    people who want to work with Java and E-Commerceinfrastructure products. Send
    resumes to [email protected]
    The Weblogic Application Serverfrom BEA
    JavaWorld Editor's Choice Award: Best WebApplication Server
    Java Developer's Journal Editor's Choice Award:Best Web Application Server
    Crossroads A-List Award: Rapid Application
    Development
    Tools for Java
    Intelligent Enterprise RealWare: Best ApplicationUsing a Component Architecture
    >
    http://www.bea.com/press/awards_weblogic.html
    >>>>>>>>>>
    PS: Folks: BEA WebLogic is in S.F. with both entry andadvanced positions for
    people who want to work with Java and E-Commerce
    infrastructure
    products. Send
    resumes to [email protected]
    The Weblogic Application Server fromBEA
    JavaWorld Editor's Choice Award: Best Web
    Application
    Server
    Java Developer's Journal Editor's Choice Award: BestWeb Application Server
    Crossroads A-List Award: Rapid Application
    Development
    Tools for Java
    Intelligent Enterprise RealWare: Best Application Usinga Component Architecture
    >
    http://www.bea.com/press/awards_weblogic.html
    >>>>>>>>
    PS: Folks: BEA WebLogic is in S.F. with both entry andadvanced
    positions for
    people who want to work with Java and E-Commerce
    infrastructure
    products. Send
    resumes to [email protected]
    The Weblogic Application Server fromBEA
    JavaWorld Editor's Choice Award: Best Web
    Application
    Server
    Java Developer's Journal Editor's Choice Award: Best WebApplication Server
    Crossroads A-List Award: Rapid Application DevelopmentTools for Java
    Intelligent Enterprise RealWare: Best Application Using aComponent Architecture
    >
    http://www.bea.com/press/awards_weblogic.html
    >>>>>>
    PS: Folks: BEA WebLogic is in S.F. with both entry and advancedpositions for
    people who want to work with Java, XML, SOAP and E-Commerce
    infrastructure
    products. Send resumes to [email protected]
    The Weblogic Application Server from BEA
    JavaWorld Editor's Choice Award: Best Web ApplicationServer
    Java Developer's Journal Editor's Choice Award: Best Web
    Application
    Server
    Crossroads A-List Award: Rapid Application Development Toolsfor Java
    Intelligent Enterprise RealWare: Best Application Using a
    Component
    Architecture
    http://www.bea.com/press/awards_weblogic.html--
    PS: Folks: BEA WebLogic is in S.F. with both entry and advanced
    positions
    for
    people who want to work with Java, XML, SOAP and E-Commerce
    infrastructure
    products. Send resumes to [email protected]
    The Weblogic Application Server from BEA
    JavaWorld Editor's Choice Award: Best Web Application Server
    Java Developer's Journal Editor's Choice Award: Best WebApplication
    Server
    Crossroads A-List Award: Rapid Application Development Toolsfor Java
    Intelligent Enterprise RealWare: Best Application Using a ComponentArchitecture
    http://www.bea.com/press/awards_weblogic.html--
    PS: Folks: BEA WebLogic is in S.F. with both entry and advanced positions
    for
    people who want to work with Java, XML, SOAP and E-Commerceinfrastructure
    products. Send resumes to [email protected]
    The Weblogic Application Server from BEA
    JavaWorld Editor's Choice Award: Best Web Application Server
    Java Developer's Journal Editor's Choice Award: Best Web Application
    Server
    Crossroads A-List Award: Rapid Application Development Tools for
    Java
    Intelligent Enterprise RealWare: Best Application Using a ComponentArchitecture
    http://www.bea.com/press/awards_weblogic.html

  • JCo Connections in WD Model running out

    Hello,
    we have WD application that uses RFC model.
    During a load test we have observed that the connection
    pools are running out of JCo connections. The result is a an error message in the trace file
    and Http 500 response code to the end user.
    Number of concurrent users: 20
    Max Pool Size: 20
    Max Connections: 20
    Exception:
    [code]aused by: com.sap.mw.jco.JCO$Exception: (106) JCO_ERROR_RESOURCE: Connection pool ME_LP_MODELDATA_TSTWEB1_PT_useSSO is exhausted. The current pool size limit is 20 connections.
         at com.sap.mw.jco.JCO$Pool.getClient(JCO.java:5150)
         at com.sap.mw.jco.JCO$PoolManager.getClient(JCO.java:5849)
         at com.sap.mw.jco.JCO$PoolManager.getClient(JCO.java:5799)
         at com.sap.mw.jco.JCO.getClient(JCO.java:8076)
         at com.sap.tc.webdynpro.serverimpl.core.sl.AbstractJCOClientConnection.getClient(AbstractJCOClientConnection.java:393)
         at com.sap.tc.webdynpro.modelimpl.dynamicrfc.AiiModel.getCurrentlyUsedJcoClient(AiiModel.java:191)[/code]
    Questions:
    Is it necessary to have equal number of Max Connections as
    number of potential concurrent users?
    Or is there a way to manually release not used connections?
    Best Regards, Maik

    Hi,
    It is not necessary to create as many connections as there are concurrent users.
    We have to handle the scope of the connection properly and also close the connections once we complete the task.
    Re: How to close a model object connection for Adaptive RFC?
    Regards
    Bharathwaj

  • I'M RUNNING OUT OF EVERYTHING

    I'm near the end of my rope. I've been more than patient dealing with jdev problems because every major problem has been solved with elbow grease, education, blood, sweat, and tears and, at least, a workaround. But I'm not too sure with this one.
    10.1.3 EA 1 ADF warnings and I'm not using ADF.  How-to get rid of them?
    For some time I've been developing an extensive JSF JSP app with JDev and everything was ok until a couple of days ago when these warnings started (after making navigation and managed bean changes to faces-config.xml with the graphics tool). This "new" problem causes deployment failures to standalone oc4j which is a major problem for us.
    But now I've run out of things to try, I'm at my wits end, and I'm almost out of patience with JDev. I'm going to give it one more good try tomorrow and if the end of the rope is at hand I'm downloading eclipse or netbeans or whatever.
    Could be time to clean house and sweep out Oracle AS Enterprise Edition and JDev.
    How appropriate --- I tried to post this to the forum and got this response ---
    Internal Server Error
    The server encountered an internal error or misconfiguration and
    was unable to complete your request.
    Please contact the server administrator, [email protected]
    and inform them of the time the error occurred, and anything you
    might have done that may have caused the error.
    More information about this error may be available in the server error log.
    Additionally, a 500 Internal Server Error error was encountered while
    trying to use an ErrorDocument to handle the request.
    Regards,
    Al Malin
    (269) 923-6301

    Sorry, Al, that this has gotten so frustrating. It sounds like you are up against it, which is a place we've all been and all hate to be.
    I think, though, that you will be more likely to get help from others in this forum if you tone down the rant a few notches. My guess is that a lot of otherwise helpful folks, including some from Oracle, won't even bother to select your posts because of the flaming all caps titles and, although we've all gotten to the place a time or two of being ready to dump the tools we're using, the threat of that is more effective at showing how you're feeling than at getting you any sympathy.
    Breath, brother. Bump your posts every couple of days. Plead every now and then. Open a ticket with MetaLink. Ask your co-workers for help thinking about the problem. Answer others questions where you can (just to build up a little forum karma). Tell you boss you need her/him to back off ...
    (What's the saying about "free advice is worth every penny you spent on it" ?)
    Hope things work out.
    Johnny Lee

  • Comment out some sections of Labview programs during debuging

    In other programming such as VC or VB, we can always comment out some sections of our program so that the commented section are not executed during debuging. Is there any way to comment out some sections of Labview program during debugging?

    YanWong wrote:
    You can use a case structure to surround the section of the program you do not want to execute and then wire a false constant into the input so the case will never execute.
    Also remember that doing this will only work if the code you "commenting" out must be executable. A broken run arrow will still be there after you put it in the Case Structure. If you are trying to comment out some broken code, you're out of luck, at least now.
    Ed
    Ed Dickens - Certified LabVIEW Architect - DISTek Integration, Inc. - NI Certified Alliance Partner
    Using the Abort button to stop your VI is like using a tree to stop your car. It works, but there may be consequences.

  • We updated our phones and now we are running out of memory.  Never happened before the update and we have not added anything new.  We have hardly anything on our phones.

    We updated our phones and now we are running out of memory.  Never happened before the update and we have not added anything new.  We have hardly anything on our phones.  Why am I having to remove stuff that has always been on my phone. 

    Thanks for the reply TJBUSMC1973. 
    I guess that means back to the shop.  Ridiculous to sell a phone that can't handle the new iOS more efficiently (or vice versa).  I can't imagine anyone going near a 5C now (I have a feeling mine had already been used and taken back once and sold as new at O2).
    Charlie

  • My computer is crashed after 3 hours, when am running an Executable file from LabView 6.0.2

    My VI is using DAQ6025e and GPIB cards, it is countiniusly writing data to files and talking with 2 GPIB instruments and one Serial instrument. Sudenly after 3 hours of perfect working it is crash my computer.
    Does any body know why ??? Help? S.O.S?

    Some questions to help us troubleshoot:
    1)How crashed is your computer when it crashes?
    Blue Screen Of Death crashed? If you don't know about BSOD then that
    is when your computer presents a blue screen (white in XP) and tells
    you that the system is halted, gives you a register dump and the
    memory address where it stopped (all of which is useless unless you
    are an assembly language wizard) and then you have to reboot the
    machine.
    Application error that kills LabVIEW.exe? Usually a pop-up MSGBOX
    style dialog box will appear with a such and such error occurred in
    LabVIEW.exe or another application and an OK button.
    VI Slows to a crawl w/hourglass in the app window? Program doesn't
    stop executing, it just stops responding to user inputs.
    LabVIEW error cluster error that causes the VI to quit?
    2) What O/S are you running? Win 3.1, Win 95, Win 98, Win NT 3.1, Win
    NT 3.5, Win NT 4.0, Win2000, WinME, WinXPHome, Win XPPro?, Linux?, Mac
    O/S?
    In the Microsoft family, WinNT and Win2000 are by far the most stable
    and reliable of the various versions of Windows with Win95 and Win98
    having a lot of problems running for long periods of time. Make sure
    that you get WinNT 4.0 updated to service pack 6 if you are still
    using 4.0 but haven't updated the service pack level yet. There were
    quite a number of problems with it when it first came out that have
    since been cured.
    3) Have you run the LabVIEW VI profiler yet? This will point out
    memory leaks that are due to dynamically built arrays using shift
    registers etc. or appending to strings, or improper IMAQ management,
    etc.
    4) Have you brought up the task manager and looked at the memory
    allocated for LabVIEW.exe and for the whole system as well? Is it
    growing? Sometimes there can be memory leaks that don't show in the
    profiler that will show in the task manager when you look. These can
    be due to bugs in LabVIEW itself or some driver for some hardware you
    are using, etc.
    If you are not running NT or a later derivative, then you won't have
    task manager available but you can get diagnostic tools that are very
    similar to task manager from off the web.
    5) How much memory do you have in your PC? How fast is your PC? What
    does the CPU % usage look like under task manager when you first start
    your app? If it is running >40% CPU usage for any significant amount
    of time at the start then perhaps your PC is overloaded with work and
    you need a faster PC or your code needs redesigning. I like data
    acquisition systems to operate <25% generally speaking. This number
    seems to improve the reliability of the code that is running on them.
    6) How much H/D space is available? You say you are writing data to
    files. Your h/d space isn't running out is it?
    7) Is it possible to run one instrument at a time for several hours to
    see if it is a particular instrument that is causing the crash? Make
    sure that your GPIB cards are properly configured so that they don't
    share resources with any other devices(IRQ, DMA, memory addresses,
    etc.) It may be possible that they behave fine most of the time even
    with shared resources but if they ever try to send a command
    simultaneously and they share resources, they could "collide" and
    crash your PC. You can check all this out in the control panel and/or
    MAX.
    8) Is it possible that your serial port is getting hung up after
    running a long period of time? Serial port UARTS are notoriously
    finicky especially over longer line lengths and higher data rates. It
    could be that your serial instrument is locking up your app or your PC
    serial port is locking up your app and causing the crash. Can you put
    a serial port analyzer in between your serial instrument and your PC
    to see what's going on when it crashes?
    9) Are you trying to read or write to the serial port in more than one
    thread or while loop or dynamically launched vi? If you are you MUST
    protect the serial port access with a semaphore or a VISA resource
    lock. You CANNOT use a boolean global to protect access. Regular
    globals aren't atomic and you can have race conditions that result in
    unintended serial port collisions between differen threads. I have
    seen this lock up multi-threaded serial programs before to the point
    of requiring rebooting.
    This problem can be especially frustrating if you haven't done
    multi-threaded before. The port won't necessarily lock up right away
    or at the same time every time. You MUST make sure that serial port
    access is limited to ONLY ONE execution thread at a time.
    10) Is there another application or service running in the background
    with your application that may be trying to access the serial port or
    GPIB cards when you aren't expecting it? This could cause a collision
    that could crash your system as well.
    Douglas De Clue
    LabVIEW developer
    [email protected]
    Yan wrote in message news:<[email protected]>...
    > My VI is using DAQ6025e and GPIB cards, it is countiniusly writing
    > data to files and talking with 2 GPIB instruments and one Serial
    > instrument. Sudenly after 3 hours of perfect working it is crash my
    > computer.
    > Does any body know why ??? Help? S.O.S?

  • Running out of cursors, memory and performance

    1. using xsql servlet .9.9.1 command-line with a query that involves cursor expressions, grouping and a join, i run into the following error:
    <ERROR>oracle.xml.sql.OracleXMLSQLException: ORA-00604: error occurred at recursive SQL level 1
    ORA-01000: maximum open cursors exceeded
    </ERROR>
    using a quick program with OracleXMLQuery from XSU 1.2, i get the same error with the same query.
    the database i'm testing is supposed to return 1 million+ rows; only 300 or so can be returned if i specify maxrows to be 300. anything beyond that gives me the cursor error.
    SQL*Plus handles the query just fine.
    2. if i take away the cursor expression in the query, and remove the maxrow spec, i get a java out of memory error.
    3. i can't enhance the performance via using OracleConnection object and wanting to use cursor expressions with XSU at the same time. i get a "protocol error" instead.
    So the questions are:
    1. how can i resolve the cursor issue for the result set size i'm trying to retrieve?
    2. how can i resolve the memory issue for the result set size i'm trying to retrieve? since XSU tries to load the entire result set into memory to build the DOM tree, is it possible to stream it out so that it doesn't have to be capped by memory limitations? if so, do you have sample code i can take a look at?
    3. is it possible to tune the oracle jdbc in conjunction of the use of XSU?
    null

    We are encountering the same problem. We are just inserting via http and no matter how we do things we end up running out of cursors. We are using the production version of the servlet. Have you been able to resolve your problem and do you know if there is a fix. Thanks for any information.
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by alfred h:
    1. using xsql servlet .9.9.1 command-line with a query that involves cursor expressions, grouping and a join, i run into the following error:
    <ERROR>oracle.xml.sql.OracleXMLSQLException: ORA-00604: error occurred at recursive SQL level 1
    ORA-01000: maximum open cursors exceeded
    </ERROR>
    using a quick program with OracleXMLQuery from XSU 1.2, i get the same error with the same query.
    the database i'm testing is supposed to return 1 million+ rows; only 300 or so can be returned if i specify maxrows to be 300. anything beyond that gives me the cursor error.
    SQL*Plus handles the query just fine.
    2. if i take away the cursor expression in the query, and remove the maxrow spec, i get a java out of memory error.
    3. i can't enhance the performance via using OracleConnection object and wanting to use cursor expressions with XSU at the same time. i get a "protocol error" instead.
    So the questions are:
    1. how can i resolve the cursor issue for the result set size i'm trying to retrieve?
    2. how can i resolve the memory issue for the result set size i'm trying to retrieve? since XSU tries to load the entire result set into memory to build the DOM tree, is it possible to stream it out so that it doesn't have to be capped by memory limitations? if so, do you have sample code i can take a look at?
    3. is it possible to tune the oracle jdbc in conjunction of the use of XSU?
    <HR></BLOCKQUOTE>
    null

Maybe you are looking for