Is this really an out of memory error?

I get the following output from running the pl/sql script below. I can't believe this is an actual out of memory error because the query involved returns only 51 rows. If I run the cursor query from the command line it takes about 5sec and returns 51 rows. The server this is running on is doing nothing else. We successfully run pl/sql scripts that use huge amounts of memory on this server and have never seen this error.
Any idea what could be going on here?
SQL> exec test_obis_admin.find_invalid_dates('royals_census_1984')
find_invalid_dates
year collected=1953
year collected=1953
year collected=1953
year collected=1953
year collected=1953
year collected=1953
year collected=1983
year collected=1983
year collected=1996
year collected=1996
year collected=1996
year collected=1996
year collected=1996
year collected=1996
year collected=1996
year collected=1972
year collected=1972
year collected=1996
year collected=1996
year collected=1996
year collected=1996
year collected=1972
year collected=1972
year collected=1976
year collected=1976
year collected=1976
year collected=1976
year collected=1976
year collected=1976
BEGIN test_obis_admin.find_invalid_dates('royals_census_1984'); END;
ERROR at line 1:
ORA-04030: out of process memory when trying to allocate 16396 bytes (koh-kghusessi,pl/sql vc2)
-- find and identify rows containing invalid date component
PROCEDURE find_invalid_dates ( res_name varchar2 ) IS
     date_temp DATE ;
     CURSOR date_components IS SELECT
     a.id,
     a.yearcollected,
     a.monthcollected,
     a.daycollected,
     a.startyearcollected,
     a.startmonthcollected,
     a.startdaycollected,
     a.endyearcollected,
     a.endmonthcollected,
     a.enddaycollected
     FROM expandeddigircache a WHERE a.res_name=res_name ;
     date_fields date_components%ROWTYPE ;
BEGIN
     print('find_invalid_dates') ;
    OPEN date_components ;
    LOOP
        fetch date_components into date_fields ;
        exit when date_components%NOTFOUND ;
        print('year collected=' || date_fields.yearcollected ) ;
   END loop;
    CLOSE date_components ;
END find_invalid_dates ;\

As it turns out this probably was a genuine out of memory error. While I was trying to figure out what was going on, I accidently ran the exact query from the pl/sql procedure.
     CURSOR date_components IS SELECT
     a.id,
     a.res_name ,
     a.yearcollected,
     a.monthcollected,
     a.daycollected,
     a.startyearcollected,
     a.startmonthcollected,
     a.startdaycollected,
     a.endyearcollected,
     a.endmonthcollected,
     a.enddaycollected
     FROM expandeddigircache a WHERE a.res_name=res_name           I was expecting 51 records, I got considerable more before the out of memory error occurred.
     I believe the query as written returns everything in the table. In this case 20mil+ records.
     The fundamental problem was that the parameter to the procedure was called res_name and the where clause tested a column called res_name. pl/sql did not consider them two different names. So I think the net result was a where clause of the form WHERE 'a'='a'
     When I changed the parameter name to p_res_name and the where clause to WHERE res_name=p_res_name the procedure worked as expected.

Similar Messages

  • Out of memory errors

    I'm having problems starting 8i. When started from /etc/rc.d/
    init.d/dbora during bootup or manually by root, I get an out of
    memory error.
    If I start it from the dbadmin user (oracle), I either get the
    same out of memory error, or I end up with several hundred shell
    logins and the database still doesn't respond.
    This is RedHat Linux 6.0, kernel 2.2.5-22smp. Here is a sample of
    what happens when I get the out of memory error:
    Mem: 160448K av, 53488K used, 106960K free, 26264K shrd,
    6532K buff
    Swap: 656360K av, 0K used, 656360K free
    36328K cached
    Oracle Server Manager Release 3.1.5.0.0 - Production
    (c) Copyright 1997, Oracle Corporation. All Rights Reserved.
    Oracle8i Release 8.1.5.0.0 - Production
    With the Java option
    PL/SQL Release 8.1.5.0.0 - Production
    SVRMGR> Connected.
    SVRMGR> ORA-27102: out of memory
    Linux Error: 22: Invalid argument
    SVRMGR>
    Server Manager complete.
    Database "ORCL" warm started.
    null

    It turns out that the problem I was having with a large number of
    shell processes being created was due to the use of oraenv in my
    .bashrc file (so much for following the instructions!) It was
    calling itself recursively until the process limit was reached.
    However, even with this fixed, the out of memory error still
    exists.
    max (guest) wrote:
    : dan....
    : check your init.ora......
    Aside from comments, it has these lines, which were created by
    dbassist:
    db_name = test
    instance_name = ORCL
    service_names = test
    control_files = ("/u02/oradata/test/control01.ctl", "/u02/oradata/
    test/control02.ctl")
    db_block_buffers = 8192
    shared_pool_size = 4194304
    log_checkpoint_interval = 10000
    log_checkpoint_timeout = 1800
    # I reduced processes to see if it would help
    processes = 10
    log_buffer = 163840
    background_dump_dest = /u01/admin/test/bdump
    core_dump_dest = /u01/admin/test/cdump
    user_dump_dest = /u01/admin/test/udump
    db_block_size = 2048
    remote_login_passwordfile = exclusive
    os_authent_prefix = ""
    compatible = "8.1.0"
    : also check ulimit
    Here's ulimit -a:
    core file size (blocks) 1000000
    data seg size (kbytes) unlimited
    file size (blocks) unlimited
    max memory size (kbytes) unlimited
    stack size (kbytes) 8192
    cpu time (seconds) unlimited
    max user processes 256
    pipe size (512 bytes) 8
    open files 1024
    virtual memory (kbytes) 2105343
    Everything looks pretty large to me.
    : Dan Wilga (guest) wrote:
    : : I'm having problems starting 8i. When started from /etc/rc.d/
    : : init.d/dbora during bootup or manually by root, I get an out
    : of
    : : memory error.
    : : If I start it from the dbadmin user (oracle), I either get
    the
    : : same out of memory error, or I end up with several hundred
    : shell
    : : logins and the database still doesn't respond.
    : : This is RedHat Linux 6.0, kernel 2.2.5-22smp. Here is a
    sample
    : of
    : : what happens when I get the out of memory error:
    : : Mem: 160448K av, 53488K used, 106960K free, 26264K shrd,
    : : 6532K buff
    : : Swap: 656360K av, 0K used, 656360K free
    : : 36328K cached
    : : Oracle Server Manager Release 3.1.5.0.0 - Production
    : : (c) Copyright 1997, Oracle Corporation. All Rights Reserved.
    : : Oracle8i Release 8.1.5.0.0 - Production
    : : With the Java option
    : : PL/SQL Release 8.1.5.0.0 - Production
    : : SVRMGR> Connected.
    : : SVRMGR> ORA-27102: out of memory
    : : Linux Error: 22: Invalid argument
    : : SVRMGR>
    : : Server Manager complete.
    : : Database "ORCL" warm started.
    null

  • Out of Memory error while builng HTML String from a Large HashMap.

    Hi,
    I am building an HTML string from a large map oject that consits of about 32000 objects using the Transformer class in java. As this HTML string needs to be displayed in the JSP page, the reponse time was too high and also some times it is throwing out of memory error.
    Please let me know how i can implement the concept of building the library tree(folder structure) HTML string for the first set of say 1000 entries and then display in the web page and then detect an onScroll event and handle it in java Script functions and come back and build the tree for the next set of entries in the map and append this string to the previous one and accordingly display it.
    please let me know whether
    1. the suggested solution was the advisable one.
    2. how to build tree(HTML String) for a set of entries in the map while iterating over the map.
    3. How to detect a onScroll event and handle it.
    Note : Handling the events in the JavaScript functions and displaying the tree is now being done using AJAX.
    Thanks for help in Advance,
    Kartheek

    Hi
    Sorry,
    I haven't seen any error in the browser as this may be Out of memory error which was not handled. I got the the following error from the web logic console
    org.apache.struts.actions.DispatchAction">Dispatch[serviceCenterHome] to method 'getUserLibraryTree' returned an exceptionjava.lang.reflect.InvocationTargetException
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:324)
         at org.apache.struts.actions.DispatchAction.dispatchMethod(DispatchAction.java:276)
         at org.apache.struts.actions.DispatchAction.execute(DispatchAction.java:196)
         at org.apache.struts.action.RequestProcessor.processActionPerform(RequestProcessor.java:421)
         at org.apache.struts.action.RequestProcessor.process(RequestProcessor.java:226)
         at org.apache.struts.action.ActionServlet.process(ActionServlet.java:1164)
         at org.apache.struts.action.ActionServlet.doPost(ActionServlet.java:415)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at weblogic.servlet.internal.ServletStubImpl$ServletInvocationAction.run(ServletStubImpl.java:996)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:419)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:315)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:6452)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:118)
         at weblogic.servlet.internal.WebAppServletContext.invokeServlet(WebAppServletContext.java:3661)
         at weblogic.servlet.internal.ServletRequestImpl.execute(ServletRequestImpl.java:2630)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:219)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:178)
    Caused by: java.lang.OutOfMemoryError
    </L_MSG>
    <L_MSG MN="ILHD-1109" PID="adminserver" TID="ExecuteThread: '14' for queue: 'weblogic.kernel.Default'" DT="2012/04/18 7:56:17:146" PT="WARN" AP="" DN="" SN="" SR="org.apache.struts.action.RequestProcessor">Unhandled Exception thrown: class javax.servlet.ServletException</L_MSG>
    <Apr 18, 2012 7:56:17 AM CDT> <Error> <HTTP> <BEA-101017> <[ServletContext(id=26367546,name=fcsi,context-path=/fcsi)] Root cause of ServletException.
    *java.lang.OutOfMemoryError*
    Please Advise.
    Thanks for your help in advance,
    Kartheek

  • Out of Memory error caused by huge result set

    Hi,
    I'm using version 3.3.4 of KODO with the enterpise license. Our
    application has some places where we're using straight SQL to create a
    query using the following statement:
    Query query = pm.newQuery ("javax.jdo.query.SQL", queryString);
    There is one query that performs an inner join and returns over 2 billion
    results. Of course this throws an out of memory error when we try to add
    the results to a list. I printed out the exact SQL statement that is
    being executed and it should return 0 results. The application is hitting
    an Oracle 9i database, and running the query through the Oracle tool
    SQLPlus returns the correct result. I'm stumped as to why this massive
    result set is being returned when we use KODO. Here is the SQL string for
    your examination:
    SELECT T_ORDERITEM.* FROM T_ORDERITEM INNER JOIN T_REVENUE ON
    T_ORDERITEM.ORDER_HOID = T_REVENUE.HOID WHERE T_ORDERITEM.COMMODITY_HOID
    = '1871816913731466' AND T_ORDERITEM.DATESHIPPED BETWEEN
    TO_DATE('12/25/2005', 'MM/DD/YYYY') AND TO_DATE('01/24/2006',
    'MM/DD/YYYY') AND T_REVENUE.CUSTOMER_HOID IN (1844203535241210)
    Thanks for your help,
    Bennett

    I'm also using JDK 1.4.2 so I'm not sure why you wouldn't be running into
    the issue using the ArrayList constructor.
    -Bennett
    Paul Mogren wrote:
    What result list implementation are you using? What's your JDK
    version? I'm sure I've written new ArrayList(results) numerous times
    without trouble, though I have not used Kodo 3.3.4. I had a look at
    the JDK 1.4.2 source code for ArrayList(Collection) and
    ArrayList.addAll(Collection), and it looks to me like the latter would
    consume more memory than the former. However, it might be worth noting
    that the former invokes size() on the argument.
    -Paul Mogren, CommerceHub
    On Thu, 19 Jan 2006 22:10:50 +0000 (UTC), [email protected]
    (Bennett Hunter) wrote:
    Nevermind, issue resolved. This wasn't due to the query, but rather how
    we were creating a list with the results returned. This issue is probably
    already posted, but I'll give the info. again:
    You can't create an arraylist by passing in the kodo collection of results
    to the constructor of ArrayList. For example: List l = new
    ArrayList(results) .... that will throw the out of memory error because of
    the way that the constructor builds the kodo collection. Maybe the bug is
    with the kodo collection. However, the way around this is to create the
    list with an empty constructor and then do an addAll() operation.
    Bennett Hunter wrote:
    Hi,
    I'm using version 3.3.4 of KODO with the enterpise license. Our
    application has some places where we're using straight SQL to create a
    query using the following statement:
    Query query = pm.newQuery ("javax.jdo.query.SQL", queryString);
    There is one query that performs an inner join and returns over 2 billion
    results. Of course this throws an out of memory error when we try to add
    the results to a list. I printed out the exact SQL statement that is
    being executed and it should return 0 results. The application is hitting
    an Oracle 9i database, and running the query through the Oracle tool
    SQLPlus returns the correct result. I'm stumped as to why this massive
    result set is being returned when we use KODO. Here is the SQL string for
    your examination:
    SELECT T_ORDERITEM.* FROM T_ORDERITEM INNER JOIN T_REVENUE ON
    T_ORDERITEM.ORDER_HOID = T_REVENUE.HOID WHERE T_ORDERITEM.COMMODITY_HOID
    = '1871816913731466' AND T_ORDERITEM.DATESHIPPED BETWEEN
    TO_DATE('12/25/2005', 'MM/DD/YYYY') AND TO_DATE('01/24/2006',
    'MM/DD/YYYY') AND T_REVENUE.CUSTOMER_HOID IN (1844203535241210)
    Thanks for your help,
    Bennett

  • Adobe X Pro 10.1.10 Out of Memory error message

    Hello One, Hello All
    Since updating our Adobe X Pro machines to version update 10.1.10, we occasionally receive an "Out of Memory" error message, which depending on what we are doing, may force us to shutdown all Adobe windows and re-open Adobe, or simply click OK and continue working with the PDF document like nothing is wrong. It seems to have no rhyme or reason to when or why it occurs. I do not see any warnings or errors in the Windows Event Log. It happens with files we have created, and also happens with files received from internal and external sources through email. It is affecting our high end machines: Lenovo X Series and W Series laptops with Windows 8.1 3rd/4th gen i5 CPUs, 8+GB RAM, 128+GB SSD's. We have all system and Windows updates applied. We have Trend Micro and Malwarebytes real-time protection and system scans do not find any malware.
    I have found a few other recent threads on Adobe forum related to this error message but the responses are weak at best with no definitive fix. The system and user temp folders size are less than 100MB each so this cannot be the issue. When the error occurs I check task manager and system utilization, including memory, is well below 100%.
    We did not have this issue prior to version update 10.1.10.
    Really hoping Adobe can step in to help here and hoping boilerplate responses are not used.

    Hi wayne103,
    We released a new security update yesterday that is v10.1.11
    Please update to this version and check the performance.
    I have seen this error message occur while opening a 3rd party created pdf that has corrupted file structure.
    Please let me know if the issue occurs for some specific files or random
    Also let me know the PDF producer for these pdfs.
    Regards,
    Rave

  • JavaScript Out of Memory Error on Portal timeout.

    Hello All,
    I am using jsf and Inline navigation in all our portlets and when user leave the browser idle for portal timeout we have 2 problems. 1: Login portlet shows in that specific portlet. 2: we get a javascript alert saying out of memory at line 40. and the porltet shows error message as "Gateway was not able to access requested content. If the error persists, contact your portal Administrator."
    We are using Plumtree 5.0.4 Java version.
    any help is highly appreciated.
    Thanks
    A.J.

    Both are valid behaviors unfortunately.
    1) login portlet is showing up in specific portlet b/c inline navigation allows for you to create and load pages without affecting the overall portal.
    This happens when you use iframes (which behave in a similar fashion).
    - your only workaround is really to write some javascript function to "listen" to the portal login page getting loaded and then throwing the session into the parent browser (which is Portal). At least this is the only solution that I ever came up with when using Iframes.
    2) Don't know about out of memory error actually, but getting the "gatewy was not able to access requested content" is valid b/c the session died.
    - javascript errors require javascript solutions. Sorry I couldn't be more helpful than that.
    Maybe someone else will have better suggestions.
    The other suggestion is to use your app server to listen to the logout event and redirect appropriately to somewhere else, or have it do what you want it to do in situations as this.

  • E71 keeps giving out of memory error; I can't find...

    My Nokia E71 (it's black so it may be an E71x, I guess) keeps giving me out of memory errors. I use my phone for nothing but voice calls, text messages, and emails. I have cleared all call logs, deleted all emails except 5 (they have no attachments) and cleared deleted emails, and deleted all text messages. I went into file manager and made sure there are no images, videos, etc. on the phone. I set up options to use memory card for emails. I've turned the phone off, waited 15 or 20 seconds, and turned it back on.
    I did all this a couple weeks ago. Ever since then. every couple days I get these out of memory errors, and the phone memory shows me 119 MB used and something like 300K free. I'm at the point where every few emails that come in, I have to delete all emails, erase all text messages, etc. or the phone is just too low on memory to operate.
    What else can I try? I don't have any additional applications installed. I've even tried to delete built-in apps such as RealPlayer but it didn't seem like that was possible.
    This is getting really annoying. Please help!

    @EricHarmon
    Although all user data would be deleted, have you tried resetting device to "Out of box" state by keying in *#7370# followed by 12345 (default Nokia lock code unless altered by user)?
    Happy to have helped forum with a Support Ratio = 42.5

  • N97 out of memory errors

    The product is N97, you don’t have in you drop down list!!!
    I have had my n97 for about 10 days now and everything was going brilliant. It is a stunning piece of kit.
    However on day 11. I repowered the n97, it kept on hanging the eventually got to boot after 20 attempts, but it was useless could not look up contacts. So I had taskman installed and killed loads of processes and uninstalled all the apps. It became usable as a very basic phone, but sluggish. Could use browser, maps, photos.
    The problem is I keep on experiencing out of memory errors on the phone, when I try to do basic things. I don’t know how to resolve this. I have rebooted, reset the phone to factory settings.
    Deleted data, uninstalled more apps and still no joy!! I cannot connect to ovi/pc suite to up grade the phone or roll it back.
    Symptoms the phone is sluggish it is still virtually unusable. It just barely make phones and is slow to respond to touch-screen presses and key presses.
    The taskman app says the is 47 mb free on C: drive. So I cant see what the problem is!!!
    I have always bought nokia for 10 years, never experience this sort of thing before.  I am really disappointed with this.
    Anybody else experienced this? Any ideas?
    Many thanks
    Gd
    Solved!
    Go to Solution.

    Copyless thanks for the tip.  Kudos added :-)
    I had the same problem a friend was trying to bluetooth some videos to me and I was getting Memory problem, How embarasing after a few min of my mates being impressed with the phone, and me praising it and talking about so much space and blah blah blah.
    I could not receive videos via bluetooth because they were 5MB.  WHY on EARTH can they not default this to the E: drive.
    Whoever made that decision in NOKIA should be fired what an idiot.
    Also  when I finally got to get the pictures on the phone via a computer I could not see them via the "Photos"  I had to go via the File Manager and see one by one.  AARRRRRRRRGHHH  I am getting to the point that I am going to return this phone.
    N97 Black unlocked using the 3 Network (UK). Product Code 0586308. Firmware V20.0.019. Made In Finland.Custom version V20.019.235.10.

  • Why do I get a Track out of memory error while running open loop frequency response?

    MatrixX Build 61mx1411: I get a "Track out of memory" error when I run the Open Loop Frequency Response from the MatrixX pull down tools. What can I do to prevent this? We are running on an HP B1000 with 768 MB of RAM under HP-UX 10.2.

    In the old days of Mx say Version 5 and prior the user actually selected the amount of memory that would be allocated. Depending on the size of the model etc. you would have to allocate memory. In version 6.0 and going forward there is no need for the user to manually allocate the memory.
    Build {rstack=50000,istack=200000,sstack=50000,cstack=50​0 000}
    If this is a command in a script file that you are running and the error is resulting from that then I would try commenting out everything after the letter d in the word build and then starting it back up.
    i.e. only use Build
    I don't believe that there is a way to manually allocate the initial SystemBuild Stack size.
    I believe initially the stack size is set to 10010.
    However, one way
    you can manually set the initial SystemBuild stack size,is to create a large StateSpace as soon as you start up SystemBuild. This will prevent piece-meal reallocs while using SystemBuild.
    You can created a new SuperBlock in SystemBuild and then drop down a StateSpace Block with 199 inputs and 199 Outputs and 1 State and entered ones(200,200)as the StateSpace Matrix without any problems. This would resize this internal stack to at least 40000.
    You really should not have to do this but if that helps then you might think about doing this in your startup.ms file you could use SBA or load the file then you could delete the superblock and begin working.
    "Bob" gave me this little tid bit.
    Please let me know if any of this is of use.
    Garrett
    Garrett Thurston
    [email protected]
    Phone: 781.993.5540

  • FCP - Out of Memory Error in 10.6.6

    Running a MacBook Pro 17" (A1261) and while I edit, I got a "Out of Memory" Error and an "Invalid Operation". Can't reopen the sequence on my MacBook but it opens with no problem on a Mac Pro.
    Gonna re-install the OS and see if that helps.

    I just had something very similar happen. A sequence that I've been editing for over a month now no longer opens (since installing the 10.6.6 update) I'm getting the same "out of memory" and I/O error codes.
    I'm at a loss what to do. Other editing projects open fine. But so far I haven't been able to open
    any of the project files I've vaulted from the last six weeks.
    Does anyone have any idea what changes 10.6.6 might have made to cause this and how
    it could be rectified. I really need to get back into my project file.

  • Out of memory error - large project

    I'm consulting on a feature doc edit, and the primary editor (Avid guy) is having serious problems accessing anything from the original project.
    It's an hour and 15 minute show, with probably close to 100 hours of footage.
    The box is a D2.3 G5 with 1.5 g of RAM, and the media is on two G-Tech drives: a G-RAID and a G-Drive. Plenty of headroom on both (now) and the system drive is brand new, having been replaced after the original died, and there's nothing loaded on it but FC Studio. The FCP version is 5.1.4. The project file is well over 100 MB.
    We started getting Out of Memory errors with this large project, and I checked all of the usual suspects: CMYK graphics, hard drive space, sufficient RAM... all checked out okay, except possibly the less-than-ideal amount of RAM.
    I copied the important sequences and a couple of select bins to a new project, and everything seems workable for now. The project is still 90 MB, and I've suggested breaking it up into separate projects and work on it as reels, but we can edit and trims work efficiently at the moment. However, the other editor has gotten to a point now where he can't even open bins in the old, big project. He keeps getting the OOM error whenever he tries to do anything.
    I have no similar problems opening the same project on my G5, which is essentially identical except I have 2.5 G RAM (1 G extra). Can this difference in RAM really make this big a deal? Is there something else I'm missing? Why can't this editor access even bins from the other project?
    G4   Mac OS X (10.2.x)  

    Shane's spot on.
    What I often do with large projects is pare down, just what you have done. But 90 out of 100 is not a big paredown by any stretch. In the new copy throw away EVERYTHING that's outdated: old sequences are the big culprit. Also toss any render files and re-render.
    Remember that, to be effective fcp keeps EVERYTHING in ram, so that it can instantly access anything in your project. The more there is to keep track of the slower you get.

  • "out of memory error" will more memory help?

    I seem to be reading mixed reviews. I have an employee doing a really simple editing task with FCP 4.5 on an iMac we have that is 1.8 GHZ with 512 MB RAM. she keeps getting "Out of Memory" errors in FCP 4.5. So we're buying 2 GB RAM. Now I'm reading in other forums that the error may still continue. Any thoughts?

    Aside from the lack of ram - 512 MB is really the minimum for the OS - out of memory messages can also come about when using CMYK color space graphics in FCP. All images should be in the RGB color space.
    This usually doesn't appear as an issue unless you are using images that were intended for offset printing or someone inadvertently changed the color space from RGB.
    Good luck.
    x

  • Out of Memory Error, Buffer Sizes, and Loop Rates on RT and FPGA

    I'm attempting to use an FPGA to collect data and then transfer that data to a RT system via a FIFO DMA, and then stream it from my RT systeam over to my host PC. I have the basics working, but I'm having two problems...
    The first is more of a nuisance. I keep receiving an Out of Memory error. This is a more recent development, but unfortunately I haven't been able to track down what is causing it and it seems to be independent of my FIFO sizes. While not my main concern, if someone was able to figure out why I would really appreciate it.
    Second, I'm struggling with overflows. My FPGA is running on a 100 MHz clock and my RT system simply cannot seem to keep up. I'm really only looking at recording 4 seconds of data, but it seems that no matter what I do I can't escape the problem without making my FIFO size huge and running out of memory (this was before I always got the Out of Memory error). Is there some kind of tip or trick I'm missing? I know I can set my FPGA to a slower clock but the clock speed is an important aspect of my application.
    I've attached a simplified version of my code that contains both problems. Thanks in advance for taking a look and helping me out, I appreciate any feedback!

    David-A wrote:
    The 7965 can't stream to a host controller faster than 800MB/s. If you need 1.6GB/s of streaming you'll need to get a 797x FPGA module.  
    I believe the application calls for 1.6 GB over 4s, so 400 MB/s, which should be within the capabilities of the 7965.
    I was going to say something similar about streaming over ethernet. I agree that it's going to be necessary to find a way to buffer most if not all of the data between the RT Target and the FPGA Target. There may be some benefit to starting to send data over the network, but the buffer on the host is still going to need to be quite large. Making use of the BRAMS and RAM on the 7965 is an interesting possibility.
    As a more out there idea, what about replacing the disk in your 8133 with an SSD? I'm not entirely sure what kind of SATA connection is in the 8133, and you'd have to experiment to be sure, but I think 400 MB/s or close to that should be quite possible. You could write all the data to disk and then send it over the network from disk once the aquisiton is complete. http://digital.ni.com/public.nsf/allkb/F3090563C2FC9DC686256CCD007451A8 has some information on using SSDs with PXI Controllers.
    Sebastian

  • CS6 Running out of memory Error while rendering

    Hey everyone,
    I recently upgraded to After Effects CS6 thinking it would be a great investment. So far, I have been horribly wrong!
    Here's my computer's specs: (this is a huge upgrade from the computer I was running 5.5 on without any problems other than it being slow)
    Intel i7-3615QM 2.3GHz Ivy Bridge processor
    8GB GDDR5 RAM
    NVIDIA GeForcs 650M w/ 2GB GDDR5 dedicated RAM
    Here's what's happening:
    I've been doing recording for my job at 1080p in front of a green screen. I record 16 minute sections. I use Keylight to replace the background with a still picture, and then export the video at half resolution, with audio as H.264 (.mp4) files. Normally I would set up the 16 minute sections to render in queue, but it kept having the "out of memory" error shortly after starting the one second in line. So then I started just doing one at a time... and now it won't even make it through one!
    I have noticed that even after rendering is done, AE is still using 6 gigs of RAM! Usually I can purge the ram, but now even doing that doesn't change anything - I have to close the program to get my RAM back. I have tried enlarging my page file, I've tried rendering in different resolutions and file types - I feel like I've tried everthing. I've looked all over the internet and I haven't found any real solutions (just talk from 2008 about "secret" menus that I can't get to work).
    Please help!!
    Thank you!!
    Dominic

    I wouldn't mind getting more RAM except that the computer I used to do this very same work with (except with 5.5) only had 6GHz of RAM and a pathetic processor and GPU in comparison.
    I've tried doing the same editing in PrP but it just doesn't have enough tools . Changing my output format hasn't helped either unfortunately. I think the program has a memory leak.
    I really think I might just go back to 5.5. I called support today and it took forever just to get the poor girl helping me to understand my name has an "m" in it. She wasn't really able to understand the problem. As far as I can tell, CS6 has been a 100% downgrade.
    Thank you both for the help!
    Dominic

  • Will 64-bit office fix out of memory error?

    I've been troubleshooting an out of memory error in Excel 2010 for some time. I've read quite a few articles on forums and on MS sites (including here.) I find many hits but none seems to offer a solution that works. One idea that seems to show
    up often is that 64-bit Office versions will have a LOT MORE memory to work with than 32-bit versions. I'm running a 64-bit version of Windows 7, so I'm considering giving Office 2010 64-bit a try. However, I also find a lot of caveats in those articles that
    concern me. On the other hand most of those articles are 2 to 3 years old, so I'm wondering if most of those issues have been dealt with. For example, VBA issues; third-party add-ins that do not (did not) support 64-bit; ActiveX and COM issues.
    Sorry to be overly verbose. My questions pretty much come down to this:
    1. Is 64-bit likely to solve my out of memory problem?
    2. What issues are still unresolved in 64-bit Excel with Windows 7?
    TIA,
    Phil

    If you are an Excel power user working with huge amounts of data, then you would benefit from 64-bit Office being able to utilize more memory.
    MS is recommending 32-bit Office as the default installation on both 32-bit and 64-bit Windows mainly due to compatibility with existing 32-bit controls, add-ins, and VBA.
    This is really not an issue that can be resolved on Office side, it depends on whether you are using any 32-bits controls, add-ins and VBA. The question is those existing 32-bit controls, add-ins and vba need an update to adapter to the 64bit of Office.
    If all your controls, add-ins and VBA is 64 bits, or designed to work with 64 bits of Office, then you are good.
    Bhasker Timilsina (ManTechs Inc)

Maybe you are looking for

  • ECC6.0 Upgrade and BW Extractors

    Hi Experts, We are doing a Source system upgrade from 4.6c to ECC6.0 just a Technical upgrade. Can you please let me know if we need to delete all the delta queues and re-initialize delta. Following are the 4.6c Pre- Export Activities  in relation BW

  • Windows 7, Magic mouse, two finger swipe action

    I'm using the 27" iMac. Does the newer bootcamp supports two finger swipe action on the magic mouse on Windows 7? Also, the scrolling action works but I can't seem to control the speed of the scrolling. Are these normal at this point?

  • Obtain only top two nodes in JTree

    I have a JTree (say, tree1) with 6 levels. I want another JTree (say, tree2) which has only the top two rows of tree1 i.e, I want the root and it's children only (no grand-children). How do I get it? Thanks. Pavan

  • Stop IMAP junk GMail?

    My question is simple: Is there ANY way to configure Apple Mail or GMail such that I don't have to see all the junk mail that GMail puts in the Spam mailbox? Any solution is welcome--either if GMail could automatically and immediately delete all inco

  • JSP, APPLET and JNLP

    OK, here is my question, we have a simple (sic) application, which consists of jsps, applets ( Runtime.exec ), I have been suggested to use jnlp and jws, A typical JSP page looks like this. <% %> :Bunch of JSP /HTML stuff <APPLET> </APPLET> Bunch of