Out of Memory error sure looks familiar!

I have a 1:42:00 video in AVI form (captred in PP2 and then output as a movie). I opened it up in Encore 2, made a timeline, it played find so then I started chugging away at transcoding the 19 gig file into a 2 hr mode DVD. About halfway through the transcoding, I got the old familiar erorr from Encore saying it was out of memory.
The whole idea of transcoding the AVI was so Encore could see up front what it had to work with and automatically pick the compression that would allow it to fit on a 4.7 gig disk.
It seems to say that as it was telling me in the build box that it would be writing to 4.7 gigs or I'm assuming it was saying it would fill all the space it had available and would compress accordingly.
Well, we had these same errors often in the old software and I'm sorry to seem them again. Does anyone know a workaround for this?
My system is an "approved" HP W-8200 with Dual P4 3.6 Xeons and 4 gigs of ram writing to a 1 terabyte sata raid 5. I don't think my machine ran out of memory.

Try  eliminating the <include logicalentitytype =  "OrganizationalUnit".../> from the extract class. If you haven't  changed any information about the OU, you shouldn't need it; the  <reference .../> takes care of the relationship between  the person and his/her OUs.
Some of the default extract classes  seem to err on the side of being too inclusive. This looks like one of  them. See, for another example, the classes given for groups, which has  the same type of issue: compare c

Similar Messages

  • Out of Memory error end erratic behavior

    Brief description of the issue: while importing images at their current location, some of them would show in grid view with an exclamation mark, and LR would not allow me to adjust them. Removing and importing the same set of images again, the subset of troubled images changed. At some stage, trying to develop one of the images, I got an "Out of Memory" written in red in the center pane. PS CS opens the same images without any problem, and Windows Task Manager indicates there is plenty of available physical and virtual memory.
    Here what I did in detail.
    While I was importing images at their current location into the library from DVDs, I started to see en exclamation mark in the upper right corner of several images in the grid view. When I selected any of those images and hit "D" to develop them, I got "An unknown error occurred" in the Develop module, and LR would not allow me to do any adjustment.
    I have tried removing the images from the Library and importing them again, and then some of the troubled images were not the same as before, and LR would allow me to develop some of the previously troubled images.
    PS CS File Browser shows them all correctly, and also PS can open any of the images without any issue.
    I have copied some of the troubled images to my hard disk, and then tried to import them into LR from there. The same image sometimes worked, sometimes didn't. Finally, when trying to develop one of those, I got an "Out of Memory" error message printed in red and upside down in the middle of the picture.
    Windows Task Manager was reporting 833824 KB of Available Physical Memory out of 2095536 KB, and was reporting as Commit Charge: Total 1281084 KB, Limit 5080272 KB, Peak 1707720 KB.
    Any idea what is going on? It looks to me like instability after I imported a few hundred pictures in a row.

    You are ABSOLUTELY right that I had explorer windows open when I got the export failed messages!!!! I do all my printing at costco, so I have Lr export my final pics to a 'jpeg finals' directory, and then individual sub-directories for each folder they came from, and then manually copy from each of these sub directories to the root directory on my CF card for transport to Costco.
    Will see if it happens in the future without that window open. Still wouldnt help the upside down OOM messages, tho thanks for the update that the team is working on it.
    These two issues and the UI lag are really all that I can ***** about too much, which is pretty amazing given my beta frustrations. I still wish I wasnt locked to the DB. I still cannot in good conscience recommend Lr to most new DSLR users. But for my own use I'm pretty happy. I wish I could get better highlight and shadow recovery like I got in RSP but I'm sure that is user error and n00bishness more than a failure of the app. I'm also not happy that I can't toggle the crop tool to the way I want to work permanently but I guess that ship has sailed too.

  • Out of Memory Error, Buffer Sizes, and Loop Rates on RT and FPGA

    I'm attempting to use an FPGA to collect data and then transfer that data to a RT system via a FIFO DMA, and then stream it from my RT systeam over to my host PC. I have the basics working, but I'm having two problems...
    The first is more of a nuisance. I keep receiving an Out of Memory error. This is a more recent development, but unfortunately I haven't been able to track down what is causing it and it seems to be independent of my FIFO sizes. While not my main concern, if someone was able to figure out why I would really appreciate it.
    Second, I'm struggling with overflows. My FPGA is running on a 100 MHz clock and my RT system simply cannot seem to keep up. I'm really only looking at recording 4 seconds of data, but it seems that no matter what I do I can't escape the problem without making my FIFO size huge and running out of memory (this was before I always got the Out of Memory error). Is there some kind of tip or trick I'm missing? I know I can set my FPGA to a slower clock but the clock speed is an important aspect of my application.
    I've attached a simplified version of my code that contains both problems. Thanks in advance for taking a look and helping me out, I appreciate any feedback!

    David-A wrote:
    The 7965 can't stream to a host controller faster than 800MB/s. If you need 1.6GB/s of streaming you'll need to get a 797x FPGA module.  
    I believe the application calls for 1.6 GB over 4s, so 400 MB/s, which should be within the capabilities of the 7965.
    I was going to say something similar about streaming over ethernet. I agree that it's going to be necessary to find a way to buffer most if not all of the data between the RT Target and the FPGA Target. There may be some benefit to starting to send data over the network, but the buffer on the host is still going to need to be quite large. Making use of the BRAMS and RAM on the 7965 is an interesting possibility.
    As a more out there idea, what about replacing the disk in your 8133 with an SSD? I'm not entirely sure what kind of SATA connection is in the 8133, and you'd have to experiment to be sure, but I think 400 MB/s or close to that should be quite possible. You could write all the data to disk and then send it over the network from disk once the aquisiton is complete. http://digital.ni.com/public.nsf/allkb/F3090563C2FC9DC686256CCD007451A8 has some information on using SSDs with PXI Controllers.
    Sebastian

  • Out of memory error - WLW Schema Project build

    hi
    Wasn't sure which group (workshop vs xmlbeans) but here goes.
    I'm doing a persomal eval of the WLS 8.1.1. platform.
    My current focus is on XML Schema support, XML Beans, transformations etc
    and in particular how it all copes with large schema.
    I'm using Justice XML Data Dictionary (3.0.0.0) as an example of a large
    schema
    (http://it.ojp.gov/jxdd/)
    I am getting an out of memory error when I build this within WLW.
    The build appears to be in the javac phase (and occurs when the memory goes
    over 256M) so I think it's the ANT task.
    WLW doesn't seem to be running out of memory and the build machine is not
    running out of resources.
    Where do I configure this? The <xmlbean> tag ? Is there any doco?
    regards
    Jim Nicolson
    =============== Build trace below ====================
    Build project JXDDSchema started.
    BUILD STARTED
    build:
    check-uptodate:
    build-sub:
    Updating property file:
    C:\data\projects\bea\integration\wlitest\.workshop\.ide\JXDDSchema\build.pro
    perties
    Deleting directory
    C:\DOCUME~1\JIMNIC~1\LOCALS~1\Temp\.wlw-temp\wlw_compile38940\JXDDSchema
    Created dir:
    C:\DOCUME~1\JIMNIC~1\LOCALS~1\Temp\.wlw-temp\wlw_compile38940\JXDDSchema
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\jxdds_3.0.0.0_full-doc.x
    sd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\iso_639-2t_1.0.0.0_full-
    doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ut_offender-tracking-mis
    c_1.0.0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-gun_1.0.0.0_fu
    ll-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-other-transact
    ions_1.0.0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-uniform-offens
    e_1.0.0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\mn_offense_1.0.0.0_full-
    doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\dod_jcs-pub2.0-misc_1.0.
    0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\cap_1.0.0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\nibrs_misc_1.0.0.0_full-
    doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\fips_6-4_1.0.0.0_full-do
    c.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\iso_3166_1.0.0.0_full-do
    c.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\dod_misc_1.0.0.0_full-do
    c.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-misc_1.0.0.0_f
    ull-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-boat_1.0.0.0_f
    ull-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-securities_1.0
    .0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-article_1.0.0.
    0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\unece_rec20-misc_1.0.0.0
    _full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-state-country_
    1.0.0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\iso_4217_1.0.0.0_full-do
    c.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\fips_10-4_1.0.0.0_full-d
    oc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ansi_d20_1.0.0.0_full-do
    c.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\iso_639-2b_1.0.0.0_full-
    doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\fips_5-2_1.0.0.0_full-do
    c.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-vehicle_1.0.0.
    0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\dod_exec-12958_1.0.0.0_f
    ull-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\ncic_2000-personal-descr
    iptors_1.0.0.0_full-doc.xsd
    Loading schema file
    C:\data\projects\bea\integration\wlitest\JXDDSchema\usps_states_1.0.0.0_full
    -doc.xsd
    Time to build schema type system: 6.75 seconds
    The system is out of resources.
    Consult the following stack trace for details.
    java.lang.OutOfMemoryError
    BUILD FAILED
    BUILD FAILED

    hi Kevin
    Thanks I'll take a look there.
    I think it's a good move to make XMLBeans open source.
    (I've been using Castor for the same purpose for the last two years).
    I was mainly focussed on XMLBeans as they are used in WLW context and my
    testing suggested that I would probably want to know more about the ANT
    build/tasks etc. (Just to be able to deal with unexpected issues)
    It's not urgent - I'm just doing a personal eval to come up to speed with
    WLS 8.1 platform.
    thanks
    Jim Nicolson
    "Kevin Krouse" <[email protected]> wrote in message
    news:[email protected]..
    Sorry, the first URL should have been:
    http://cvs.apache.org/viewcvs.cgi/*checkout*/xml-xmlbeans/v1/xkit/anttask.html?rev=1.2
    >
    --k
    On Tue, 30 Sep 2003 22:35:58 -0700, Kevin Krouse wrote:
    Hi Jim,
    Well, the XMLBean docs are in a process of moving into the Apache
    XMLBeans
    project (http://xml.apache.org/xmlbeans), so I can point you to where
    they
    are in the cvs repository. You can also look at the source if you'dlike!
    >>
    Here's an html doc for the xmlbean task:
    http://cvs.apache.org/viewcvs.cgi/*checkout*/xml-xmlbeans/v1/xkit/anttask.html?rev=1.2http://cvs.apache.org/viewcvs.cgi/*checkout*/xml-xmlbeans/v1/xkit/anttask.html?rev=1.2
    >>
    Here's the code for the xmlbean task:
    http://cvs.apache.org/viewcvs.cgi/xml-xmlbeans/v1/src/xmlcomp/org/apache/xmlbeans/impl/tool/XMLBean.java?rev=HEAD&content-type=text/vnd.viewcvs-markup
    >>
    >>
    As to your second question about WLW becoming unresponsive, after the
    Schemas.jar is built, the source for the entire project is re-scanned
    which could take several minutes. The performance impact of rescanning
    the project is improved in SP2 and will be addressed even more in future
    releases. Stay tuned.
    --k

  • Out of Memory error, though nothing has changed?

    I have two large projects, 65mb each, which I have had open in FCP 7 with no problems for months.
    Then I left them closed for a month and worked on other projects.
    Now I can't open either of those two earlier projects without receiving the Out of Memory error (I'm trying to open each one individually). There are many things to look for when this error appears, as found all over this forum and others. So here is my checklist of facts:
    Neither project contains any PSD files or any sequences with text files in them.
    One of those projects contains no sequences whatsoever, only clips.
    All media for these projects is Apple AVC-Intra 100M 1280x720.
    The RAID drive containing the media for these projects has not been used since the last time the projects were open and working fine.
    The edit system has not changed in any way:
    - FCP 7.0.3
    - Mac Pro
    - OS 10.6.6
    - 2x2.66 Quad Core
    - 6 GB RAM
    - LaCie 8TB RAID, with 1.7TB free
    All RAM is functioning fine.
    I've trashed FCP preferences and received the same results anyway.
    Again, both projects were working fine on this same computer, same version of FCP, even with BOTH projects open at once.
    Also, I've been working with other projects on this same system with no problems, some nearly as large as these two.
    So... none of the obvious culprits are at work here to cause this error, unless I'm missing something.
    Anyone have any ideas? This is a bit scary, as these are documentary projects which are only going to get bigger, and I've done projects bigger than this several times without problems. Hopefully I'm missing something simple...
    Thanks.

    I think I found the culprit, and it's an obvious one:
    My original big project, 65mb, had a bin that was created by an assistant before I began work on the project and I never saw it. That bin has sequence strings for every tape/card loaded into the project, so it had about 65 rather long sequences. Once I found it, I moved that bin to its own project and deleted it from the original project.
    Now the original project opens, and I can open multiple projects at once, no Out of Memory error.
    So it's the bin full of sequences that did it. If I had seen that bin before I wouldn't have had to run this up the flagpole here. But at least it reinforces that word of advice that has appeared in many previous posts about this error: Don't overload a project with long sequences. Break up the sequences into smaller projects and archive them.
    And my own addition to that advice: If you're an assistant setting up a project for an editor, make sure your editor knows what sequences you've tucked away in bins hidden inside bins inside bins, etc.
    All good to go! (Until the next error... stand by...)

  • Out of Memory error caused by huge result set

    Hi,
    I'm using version 3.3.4 of KODO with the enterpise license. Our
    application has some places where we're using straight SQL to create a
    query using the following statement:
    Query query = pm.newQuery ("javax.jdo.query.SQL", queryString);
    There is one query that performs an inner join and returns over 2 billion
    results. Of course this throws an out of memory error when we try to add
    the results to a list. I printed out the exact SQL statement that is
    being executed and it should return 0 results. The application is hitting
    an Oracle 9i database, and running the query through the Oracle tool
    SQLPlus returns the correct result. I'm stumped as to why this massive
    result set is being returned when we use KODO. Here is the SQL string for
    your examination:
    SELECT T_ORDERITEM.* FROM T_ORDERITEM INNER JOIN T_REVENUE ON
    T_ORDERITEM.ORDER_HOID = T_REVENUE.HOID WHERE T_ORDERITEM.COMMODITY_HOID
    = '1871816913731466' AND T_ORDERITEM.DATESHIPPED BETWEEN
    TO_DATE('12/25/2005', 'MM/DD/YYYY') AND TO_DATE('01/24/2006',
    'MM/DD/YYYY') AND T_REVENUE.CUSTOMER_HOID IN (1844203535241210)
    Thanks for your help,
    Bennett

    I'm also using JDK 1.4.2 so I'm not sure why you wouldn't be running into
    the issue using the ArrayList constructor.
    -Bennett
    Paul Mogren wrote:
    What result list implementation are you using? What's your JDK
    version? I'm sure I've written new ArrayList(results) numerous times
    without trouble, though I have not used Kodo 3.3.4. I had a look at
    the JDK 1.4.2 source code for ArrayList(Collection) and
    ArrayList.addAll(Collection), and it looks to me like the latter would
    consume more memory than the former. However, it might be worth noting
    that the former invokes size() on the argument.
    -Paul Mogren, CommerceHub
    On Thu, 19 Jan 2006 22:10:50 +0000 (UTC), [email protected]
    (Bennett Hunter) wrote:
    Nevermind, issue resolved. This wasn't due to the query, but rather how
    we were creating a list with the results returned. This issue is probably
    already posted, but I'll give the info. again:
    You can't create an arraylist by passing in the kodo collection of results
    to the constructor of ArrayList. For example: List l = new
    ArrayList(results) .... that will throw the out of memory error because of
    the way that the constructor builds the kodo collection. Maybe the bug is
    with the kodo collection. However, the way around this is to create the
    list with an empty constructor and then do an addAll() operation.
    Bennett Hunter wrote:
    Hi,
    I'm using version 3.3.4 of KODO with the enterpise license. Our
    application has some places where we're using straight SQL to create a
    query using the following statement:
    Query query = pm.newQuery ("javax.jdo.query.SQL", queryString);
    There is one query that performs an inner join and returns over 2 billion
    results. Of course this throws an out of memory error when we try to add
    the results to a list. I printed out the exact SQL statement that is
    being executed and it should return 0 results. The application is hitting
    an Oracle 9i database, and running the query through the Oracle tool
    SQLPlus returns the correct result. I'm stumped as to why this massive
    result set is being returned when we use KODO. Here is the SQL string for
    your examination:
    SELECT T_ORDERITEM.* FROM T_ORDERITEM INNER JOIN T_REVENUE ON
    T_ORDERITEM.ORDER_HOID = T_REVENUE.HOID WHERE T_ORDERITEM.COMMODITY_HOID
    = '1871816913731466' AND T_ORDERITEM.DATESHIPPED BETWEEN
    TO_DATE('12/25/2005', 'MM/DD/YYYY') AND TO_DATE('01/24/2006',
    'MM/DD/YYYY') AND T_REVENUE.CUSTOMER_HOID IN (1844203535241210)
    Thanks for your help,
    Bennett

  • Out of Memory Error While deploying as EAR file

    Hai,
    I was trying to deploy an EAR file of size 63 MB which inturn containing about 60 EJB.jars. No WARs. application.xml has all the entries for the JARs. While I am deploying it is giving Out of Memory Error. Is there any way to tweak this problem. I am using my own hand written java application which uses the SunONE deployment APIs for deployment. Can u please tell how to tackle this problem. I am running my application through a batch file which uses jdk1.4.
    Please help me regarding this issue.

    You can set the initial heap size and maximum heap size for the JVM, either in the app-server admin console, or maybe in one of your scripts. You look-up the syntax!...
    I had this error yesterday. I too had run out of memory (150Mb). You simply need to allocate more to the app-server.

  • Photoshop CS6 has out of memory errors under Mac OS X 10.7.4

    Has anyone who has a 2008 MacBoo Pro with the NVIDIA GeForce 8600M GT 512 MB graphics card tried running Photoshop CS6 under Mac OS X 10.7.4?
    On my MacBook Pro, PS CS6 launches; however, any action results in an out of memory error.  The CS6 versions of After Effects, Premiere Pro and Illustrator all launch and run as expected under 10.7.4.
    If I switch to my Mac OS X 10.6.8 startup volume, Photoshop CS6 runs as expected.
    I have support cases open with Adobe, NVidia and Apple.  Adobe thinks I need to update the display drivers.  NVidia says display drivers must be provided by Apple.  Apple says that the display drivers are current.  The Apple support agent said that he can escalate the case if the issue can be reproduced on more than one computer.
    I'd love to keep my MacBook Pro running another year or so without having to reboot in 10.6.8 every time I need to use Photoshop.  I'd stay with 10.6.8, but of course iCloud requires 10.7.4.
    I have not tried 10.8 yet, but it's on my troubleshooting list.
    Also, PS CS6 runs fine under Mac OS X 10.7.4 on my 2010 iMac.
    Thanks in advance for any feedback.
    - Warren

    It looks like it was indeed the display driver for the NVIDIA GeForce 8600 GT 512 MB graphics card inside my 2008 MacBook Pro.
    It seems that this driver only gets installed if you upgrade from Mac OS X 10.6.7 to 10.7.4 while the startup drive is connected to the MacBook Pro itself.  I had updated while the startup drive was connected to my 2010 iMac. Accordingly, I could have reinstalled Photoshop 100 times without it ever launching as expected.  Having used the external startup drive with three other Apple computers (all 2010 or newer machines) with Photoshop CS6 launching as expected, I had just assumed (mistakenly) that it would work as expected with my 2008 computer.
    So, this was simple enough to resolve by upgrading the OS on the external startup drive while having it connected to the MacBook Pro.   In hindsight, this makes perfect sense.
    -Warren

  • E71 keeps giving out of memory error; I can't find...

    My Nokia E71 (it's black so it may be an E71x, I guess) keeps giving me out of memory errors. I use my phone for nothing but voice calls, text messages, and emails. I have cleared all call logs, deleted all emails except 5 (they have no attachments) and cleared deleted emails, and deleted all text messages. I went into file manager and made sure there are no images, videos, etc. on the phone. I set up options to use memory card for emails. I've turned the phone off, waited 15 or 20 seconds, and turned it back on.
    I did all this a couple weeks ago. Ever since then. every couple days I get these out of memory errors, and the phone memory shows me 119 MB used and something like 300K free. I'm at the point where every few emails that come in, I have to delete all emails, erase all text messages, etc. or the phone is just too low on memory to operate.
    What else can I try? I don't have any additional applications installed. I've even tried to delete built-in apps such as RealPlayer but it didn't seem like that was possible.
    This is getting really annoying. Please help!

    @EricHarmon
    Although all user data would be deleted, have you tried resetting device to "Out of box" state by keying in *#7370# followed by 12345 (default Nokia lock code unless altered by user)?
    Happy to have helped forum with a Support Ratio = 42.5

  • N97 out of memory errors

    The product is N97, you don’t have in you drop down list!!!
    I have had my n97 for about 10 days now and everything was going brilliant. It is a stunning piece of kit.
    However on day 11. I repowered the n97, it kept on hanging the eventually got to boot after 20 attempts, but it was useless could not look up contacts. So I had taskman installed and killed loads of processes and uninstalled all the apps. It became usable as a very basic phone, but sluggish. Could use browser, maps, photos.
    The problem is I keep on experiencing out of memory errors on the phone, when I try to do basic things. I don’t know how to resolve this. I have rebooted, reset the phone to factory settings.
    Deleted data, uninstalled more apps and still no joy!! I cannot connect to ovi/pc suite to up grade the phone or roll it back.
    Symptoms the phone is sluggish it is still virtually unusable. It just barely make phones and is slow to respond to touch-screen presses and key presses.
    The taskman app says the is 47 mb free on C: drive. So I cant see what the problem is!!!
    I have always bought nokia for 10 years, never experience this sort of thing before.  I am really disappointed with this.
    Anybody else experienced this? Any ideas?
    Many thanks
    Gd
    Solved!
    Go to Solution.

    Copyless thanks for the tip.  Kudos added :-)
    I had the same problem a friend was trying to bluetooth some videos to me and I was getting Memory problem, How embarasing after a few min of my mates being impressed with the phone, and me praising it and talking about so much space and blah blah blah.
    I could not receive videos via bluetooth because they were 5MB.  WHY on EARTH can they not default this to the E: drive.
    Whoever made that decision in NOKIA should be fired what an idiot.
    Also  when I finally got to get the pictures on the phone via a computer I could not see them via the "Photos"  I had to go via the File Manager and see one by one.  AARRRRRRRRGHHH  I am getting to the point that I am going to return this phone.
    N97 Black unlocked using the 3 Network (UK). Product Code 0586308. Firmware V20.0.019. Made In Finland.Custom version V20.019.235.10.

  • Out of memory error - JS Runtime: How many users can one connect?

    Not talking video here.  Talking interactive apps, like chat.  Ours crashes at about 500 connected users.  When I report this I'm told "make sure you're not creating too many objects serverside" or "increase the JSRuntimeSize setting in your application.xml file to the max".
    Have now done both of those things but still get this out of memory error.  Let's say I optomized my app and got 100% more connection capacity.  That would be 1,000 connected users - still nowhere near enough.
    Are my dreams of 6,000 or 10,000 connected users enjoying all of the fruits of the FMS interactivity pipe dreams?  Is it not meant for sessions of that size?  Where does one find documentation or advice or application assistance on this issue?
    How do large social media applications connect so many people concurrently.
    Thoughts appreciated.
    Thanks

    Yes.  I'm using the max.
    <RuntimeSize>51200</RuntimeSize>
    See:
    http://help.adobe.com/en_US/FlashMediaServer/3.5_AdminGuide/WS5b3ccc516d4fbf351e63e3d119f2 926bcf-7ff0.html#WS5b3ccc516d4fbf351e63e3d119f2926bcf-7ed2
    Don't think 100MB or 200MB would be valid settings.

  • Out of Memory Errors when Exporting Slideshows

    Hi -
    I've been using LR for sometime now. Today, for the first time ever, I attempted to export a slideshow as a .pdf file, 90% quality setting, 48 slides total. LR created a 41MB file. However, when I played it back, there were many slides which only showed an empty frame with the message "out of memory error" upside down in the lower left corner of the frame. The same slides displayed the message each time I played it, so I'm sure the problem stemmed from creating the file, not showing it.
    I have a core 2 duo machine with 4GB Ram and no other applications running at the time I exported the slideshow.
    Any thoughts on this? BTW, I'm running version 1.2
    Thanks!
    E-

    Try restarting and reexporting. OOM errors have little to do with how much memory you have or how much is available.

  • Out of memory error - from parsing a "fixed width file"

    This may be fairly simple for someone out there but I am trying to write a simple program that can go through a "fixed width" flat txt file and parse it to be comma dilmeted.
    I use a xml file with data dictionary specifications to do the work. I do this because there are over 430 fields that need to be parsed from a fixed width with close to 250,000 lines I can read the xml file fine to get the width dimensions but when I try to apply the parsing instructions, I get an out of memory error.
    I am hoping it is an error with code and not the large files. If it is the latter, does anyone out there know some techniques for getting at this data?
    Here is the code
       import java.io.*;
       import org.w3c.dom.Document;
       import org.w3c.dom.*;
       import javax.xml.parsers.DocumentBuilderFactory;
       import javax.xml.parsers.DocumentBuilder;
       import org.xml.sax.SAXException;
       import org.xml.sax.SAXParseException;
        public class FixedWidthConverter{
          String[] fieldNameArray;
          String[] fieldTypeArray;
          String[] fieldSizeArray;      
           public static void main(String args []){
             FixedWidthConverter fwc = new FixedWidthConverter();
             fwc.go();
             fwc.loadFixedWidthFile();
            //System.exit (0);
          }//end of main
           public void go(){
             try {
                DocumentBuilderFactory docBuilderFactory = DocumentBuilderFactory.newInstance();
                DocumentBuilder docBuilder = docBuilderFactory.newDocumentBuilder();
                Document doc = docBuilder.parse (new File("files/dic.xml"));
                // normalize text representation            doc.getDocumentElement ().normalize ();
                System.out.println ("Root element of the doc is " +
                     doc.getDocumentElement().getNodeName());
                NodeList listOfFields = doc.getElementsByTagName("FIELD");
                int totalFields = listOfFields.getLength();
                System.out.println("Total no of fields : " + totalFields);
                String[] fldNameArray = new String[totalFields];
                String[] fldTypeArray = new String[totalFields];
                String[] fldSizeArray = new String[totalFields];
                for(int s=0; s<listOfFields.getLength() ; s++){
                   Node firstFieldNode = listOfFields.item(s);
                   if(firstFieldNode.getNodeType() == Node.ELEMENT_NODE){
                      Element firstFieldElement = (Element)firstFieldNode;
                      NodeList firstFieldNMList = firstFieldElement.getElementsByTagName("FIELD_NM");
                      Element firstFieldNMElement = (Element)firstFieldNMList.item(0);
                      NodeList textFNList = firstFieldNMElement.getChildNodes();
                      //System.out.println("Field Name : " +
                               //((Node)textFNList.item(0)).getNodeValue().trim());
                      //loads values into an array
                      //fldNameArray[s] = ((Node)textFNList.item(0)).getNodeValue().trim();
                      NodeList typeList = firstFieldElement.getElementsByTagName("TYPE");
                      Element typeElement = (Element)typeList.item(0);
                      NodeList textTypList = typeElement.getChildNodes();
                      //System.out.println("Field Type : " +
                               //((Node)textTypList.item(0)).getNodeValue().trim());
                      //loads values into an array
                      //fldTypeArray[s] = ((Node)textTypList.item(0)).getNodeValue().trim(); 
                      NodeList sizeList = firstFieldElement.getElementsByTagName("SIZE");
                      Element sizeElement = (Element)sizeList.item(0);
                      NodeList textSizeList = sizeElement.getChildNodes();
                      //System.out.println("Field Size : " +
                               //((Node)textSizeList.item(0)).getNodeValue().trim());
                      //loads values into an array
                      fldSizeArray[s] = ((Node)textSizeList.item(0)).getNodeValue().trim();   
                   }//end of if clause
                }//end of for loop with s var
                //setFldNameArray(fldNameArray);
                //setFldTypeArray(fldTypeArray);
                setFldSizeArray(fldSizeArray);
                 catch (SAXParseException err) {
                   System.out.println ("** Parsing error" + ", line "
                      + err.getLineNumber () + ", uri " + err.getSystemId ());
                   System.out.println(" " + err.getMessage ());
                 catch (SAXException e) {
                   Exception x = e.getException ();
                   ((x == null) ? e : x).printStackTrace ();
                 catch (Throwable t) {
                   t.printStackTrace ();
          }//end go();
           public void setFldNameArray(String[] s){
             fieldNameArray = s;
          }//end setFldNameArray
           public void setFldTypeArray(String[] s){
             fieldTypeArray = s;
          }//end setFldTypeArray
           public void setFldSizeArray(String[] s){
             fieldSizeArray = s;
          }//end setFldSizeArray
           public String[] getFldNameArray(){
             return fieldNameArray;
          }//end setFldNameArray
           public String[] getFldTypeArray(){
             return fieldTypeArray;
          }//end setFldTypeArray
           public String[] getFldSizeArray(){
             return fieldSizeArray;
          }//end setFldSizeArray 
           public int getNumLines(){
             int countLines = 0;
             try {
                    //File must be in same director and be the name of the string below
                BufferedReader in = new BufferedReader(new FileReader("files/FLAT.txt"));
                String str;
                while ((str = in.readLine()) != null) {
                   countLines++;
                in.close();
                 catch (IOException e) {}    
             return countLines;
          }//end of getNumLines
           public void loadFixedWidthFile(){
             int c = getNumLines();
             int i = 0;
             String[] lineProcessed = new String[c];
             String chars;
             try {
                    //File must be in same director and be the name of the string below
                BufferedReader in = new BufferedReader(new FileReader("files/FLAT.txt"));
                String str;
                while ((str = in.readLine()) != null) {
                   //System.out.println(str.length());
                   lineProcessed[i] = parseThatLine(str);
                   i++;
                in.close();
                 catch (IOException e) {}     
                //write out the lineProcess[] array to another file
             writeThatFile(lineProcessed);
          }//end loadFixedWidthFile()
           public void writeThatFile(String[] s){
             try {
                BufferedWriter out = new BufferedWriter(new FileWriter("files/outfilename.txt"));
                for(int i = 0; i < s.length -1; i++){
                   out.write(s);
    }//end for loop
    out.close();
    catch (IOException e) {}
    }//end writeThatFile
    public String parseThatLine(String s){
    int start = 0;
    int end = 0;
    String parsedLine = "";
    int numChars = getFldSizeArray().length;
    //Print number of lines for testing
    //System.out.println(numChars);
    String[] oArray = getFldSizeArray();
    //String chars = oArray[0];
    //System.out.println(chars.length());
    //oArray
    for(int i = 0; i < numChars -1; i++ ){
    if(i == 0){
    start = 0;
    end = end + Integer.parseInt(oArray[i])-1;
    else
    start = end;
    end = end + Integer.parseInt(oArray[i]);
    parsedLine = parsedLine + s.substring(start, end) + "~";
    }//end for loop
    return parsedLine;
    }//End of parseThatLine
    I have tried to illeminate as many arrays as I can thinking that was chewing up the memory but to no avail.
    Any thoughts or ideas?
    Message was edited by:
    SaipanMan2005

    You should not keep a String array of all the lines of the file read.
    Instead for each line read, parse it, then write the parsed line in the other file:      public void loadFixedWidthFile() {
             BufferedReader in = null;
             BufferedWriter out = null;
             try {
                //File must be in same director and be the name of the string below
                in = new BufferedReader(new FileReader("files/FLAT.txt"));
                out = new BufferedWriter(new FileWriter("files/outfilename.txt"));
                String str;
                while ((str = in.readLine()) != null) {
                   //System.out.println(str.length());
                   str = parseThatLine(str);
                   //write out the parsed str to another file
                   out.write(str);
             catch (IOException e) {
                e.printStackTrace(); // At least print the exception - never swallow an exception
             finally { // Use a finally block to be sure of closing the files even when exception occurs
                try { in.close(); }
                catch (Exception e) {}
                try { out.close(); }
                catch (Exception e) {}
          }//end loadFixedWidthFile()Regards

  • Out of Memory error in  JTextPane

    Hi all,
    A application contains four JTextPanes.The features supported are foreground,background colouring,editing etc...
    When i try to load a file
    containing 10,000 lines i get "out of memory error".
    What is the maximum number of characters or lines that can be
    represented in a JTextPane.
    I find a lot of objects related to JTextPane getting created.
    Is there any work around to avoid this?.
    Has any one faced this problem already?.
    The forum already contains the question but no replies to it.
    Thanks.

    Has any one faced this problem already?.assuming you're having the same problem (sure sounds similar)...
    see my thread here:
    http://forum.java.sun.com/thread.jsp?forum=57&thread=340872
    there's a bug in jtextarea, jtextpane and jeditorpane. a link to the initial bug report (jtextarea) is in there. i duplicated the bug in all three components by creating a blank jframe with the component in question on it. memory use skyrocketed with all three
    i already cast two votes for for the bug to be fixed. i don't know how sun can expect people to adopt java when there are debilitating problems such as these...
    anyway, try using an awt text area instead...that is, if you don't need to make use of the extra features of jtextpane. using the awt text area solved my problems.

  • Out of Memory Error in ODI

    We are working on Siebel to EBiz OIC PIP. This is direct integration using ODI.
    When we are trying to move 4 lakh records from Siebel to EBiz OIC for PSR testing, Then the ODI process continue to run for ever and in the back ground (where the agent is started) we saw an out-of-memory error message.
    Did any one of you come across such cases

    If the issue come from ODI buffer.
    You can maybe try to increase the MAX_HEAP_SIZE parameter on the ODIPARAMS.bat (odi_install --> oracledi --> bin).
    Then restart your agent.
    EDIT:
    After taking a look to the documentation, if your agent is a windows Service and after updated the ODIPARams.bat the service
    should be un-installed and then re-installed.
    Edited by: BriceM on 6 nov. 2009 06:06

Maybe you are looking for