Understanding a FCP memory issue

I am compositing an animated film in FCP.
I am using output from Maya and 3DS Max
It is about half imported and cut (the total film will be 9mins).
I recently started running into all kinds of issues. First of all it take about 30mins to load to project, and now when the project finally loads I get a "Operation Not Allowed" "Out of Memory" message.
I started by importing the raw TIFF files into FCP (there were thousands!) Mike Grenadier gave me a helpful pointer NOT to import the TIFFs directly, but to first composite the TIFFS into an uncompressed .mov using QT.
I did that and the project is alot cleaner, but I'm still getting the memory issue.
What's interesting, is that I'm looking at the Activity Monitor as the project loads and it never seems to approach the 4gig limit.
Here's an example of the readings: Real memory Size: 442MB, Virtual Memory Size: 1.21GB, Private Memory Size 362MB, Virtual Private Memory: 597MB.
Can someone shed some light for me so I can get back to work?!?
Many thanks!
Steven Galvano

Think I found the culprit -- not a memory issue that all, but an element in the timeline that FCP didn't like.
If you are getting erratic behavior from FCP, or are getting memory errors, crashes, quits, etc. -- examine what's in your sequence. In my case, it was a group of nested TIFFs.
thanks
Steve

Similar Messages

  • FCP Memory issue on a new MBP

    I just bought a new 17" macbook pro with 4gb of ram. It's only a few days old. I'm having an issue with final cut pro 6 while trying to render or compress a file. Over the course of a few minutes, It transfers all of my memory from free to inactive memory and then either crashes or gives me an out of memory error. Meanwhile only a small portion is actually active. The same thing happens when I export using compressor.
    I tried compresing the same exact file on my old powerbook G4 that has 1.5ghz of ram and it processes fine with plenty of free memory left over.
    I have absolutely no clue what's causing this. I've adjusted the preferences, deleted them, deleted render files, ran FCP rescue, and nothings having an effect.
    I don't think anything is wrong with the ram, because every other program works great.
    Has anyone heard of this problem before? Any advice to what can be causing the problem? Please help!!

    If you don't feel comfortable with this process, Brian's advice seems the best way to go.
    That's exactly why I suggested it. Swapping RAM sticks in a MBP is a little bit tricky. I'm experienced at all that stuff, but the first time I did it on a MBP, I was a bit nervous.
    And if you don't practice safe static control, you could do harm to the good sticks as well. (And you might never know you did it, as it doesn't always show up right away.)

  • Memory issues

    I have a 2008 Macpro...remarkable up-to -date for its age...I plan to use it for some time yet. Lately I have been a having what seems to be a memory issue. There are 12 gig of Ram installed...6x2 gig each in pairs. Sometimes on boot the screen only shows a flashing folder. I shut down and reboot, and it seems fine...but only 8gig of ram shows installed. I opened it up, unseated the boards that hod the ram, and unseated and reset all he ram, and replaced the boards. Great, all he ram shows, 12 gig again. This morning, a few days later, it did it again, showing 8. So, when I run Crucial's memory analyzer, it shows first four slots full, last four empty. A few questions...are the first four slots the top board, the last four the bottom board? This, BTW, is what it was also showing before I reseated the ram. If so, the ram in question s on a separate board. Opinions, is the ram bad, or the board, or neither? Should I reseat again and flush with contact/tuner cleaner/degreaser? Could this hurt? (it is made to clean electronics) the machine does seem ever so slightly unstable. FCP crashed once in a 5 hour edit, which it has done before but rarely. The machine itself did not crash, just the program. Can using the machine like this do more damage, or will it just not boot? I feel it is a seating or connection issue, as I have never seen ram come and go, once it is fried it's fried. The ram is still pretty pricy, but I was going to buy 4 more gigs just to fill it out. Are the two board that hold the ram identical, top and bottom? If so, is a good diagnostic tool to swap them and see what happens? Move all the ram and swap board positions, or do I risk blowing ram ? Remember, the ram did show up when I reseated, so it seems like it is I damaged. My goal would be to see if top ram stops appearing. Does ram need to be filled in order? I know in pairs, but can I just swap the boards and leave the ram, which should look like filling slots 1,2,5,6,7,8...as opposed to 1,2,3,4,5,6.  Can the boards be swapped, or are they set somehow to be 1-4 and 5-8, or is that determined automatically by their upper or lower insertion? Any other advice is appreciated. Funds are low, so I need to keep this machine running. Other info, 3.06ghz (2 quad core) Xeon CPUs, 1tb system drive, 4 TB raid5' OSX 10.6.8 dual graphics display with 512MB (can't remember card model right now). Also, can I update graphic card in this machine? I looked at it a few years ago, but the I for I found was confusing and I never got to it. I figure I might be able to get a better card pretty cheap by now. Thanks all.

    You have typed a huge list of ideas all as one paragraph with many questions in it without a single bit of white space. My eyes are not good enough to read a question, attempt to answer it, then dig through the sea of letters to find the next question.
    Your writing needs white space, which gives it a peculiar overall shape, and allows a reader with poor eyesight to return to a particular spot in the text. So I will attempt to answer some parts of your query, but will be forced to do so from memory of what you wrote.
    The Mac Pro uses Error Correcting Code (ECC) RAM. Eight extra bits are stored with each word, and checked when read back. Single-bit errors are detected, and the check bits are used to make a correction on the fly. More complicated errors are detected, and can cause your Mac to halt on a kernel panic to prevent bad data from propagating.
    ECC RAM means your Mac can detect bad memory easily at Startup, and can declare pairs of DIMMs "absent" to avoid using them.
    Contact corrosion, which was an issue with much older DIMMs, is rarely an issue with these FBDIMMs. Once re-Inserted ONCE, corrosion will not recur for a year or more.
    The upper and lower Riser-cards are identical.
    The FBDIMMs in these machines are subject to overheating and death. If Mac OS X or any program reports DIMMs as missing, at least one of the DIMMs in that pair has died (but since they are installed in pairs, both are marked absent).

  • Heap Memory Issue in weblogic 9.2 for a JSF 1.1 web application

    Hi,
    We are running a JSF application (Myfaces, facelets, tomahawk, richfaces & iBATIS) in weblogic 9.2 server on Solaris 10. This application is deployed in production and works fine under normal circumstances. But when there is a heavy user load we are facing a memory issue. The memory usage is gradually increasing and when it reaches to max, Full GC kicks in again & again which choks up all requests. We don't save anything in session scope. All our backing beans are saved in request scope hence they should be garbage collected after each request done, but this is not happening.
    We took the heap dump from production after this issue and analyzed it. After my analysis, I found all objects which are set in request object not being garbage collected and the root referers of all these objects is weblogic.servlet.internal.MuxableSocketHTTP.
    I reproduced the similar behaviour in one of our development environment using JMeter. I ran 100 concurrent users in JMeter for almost 1 hour and saw the similar behaviour. Below is the result of all weblogic objects which are still hanging in heap after test was over (I also ran manual Garbage Collector from admin server).
    1) weblogic.servlet.internal.MuxableSocketHTTP - 1774 objects - retained heap (1 GB)
    2) weblogic.servlet.internal.ServletRequestImpl - 1774 objects - retained heap (1 GB)
    My understanding is that every request made to weblogic server goes through the MuxableSocketHTTP object which creates the ServletRequestImpl to serve it. Once the request is served these objects are suppose to be removed. As a result of that whatever is saved in your request will still be hanging.
    I am not able to understand why these objects are hanging after request is done. Could anybody answer to my question. I appreciate your help in advance.
    The GC setting for weblogic server while startup is:
    -XX:MaxTenuringThreshold=15 -XX:+PrintTenuringDistribution -XX:+AggressiveHeap -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:PermSize=128m -XX:MaxPermSize=128m -Xms3g -Xmx3g -XX:NewSize=512m -XX:MaxNewSize=1024m
    Thanks MaKK

    What happened with this issue? We are seeing something similar on WebLogic 9.2 MP1 in Solaris (Jdk 1.5. patch 10. 32 bit). Out of Memory's with thousands if instances of weblogic.socket.MuxableSocket hanging around.
    Our thinking was initally the Java heap, then we thought that maybe the sockets weren't being closed properly, possible in WebLogic or in LiveCycle.
    Any info would be greatly appreciated.
    Snippet of our stack trace:
    <16-Feb-2010 04:30:13 o'clock GMT> <Error> <Kernel> <BEA-000802> <ExecuteRequest failed
    java.lang.OutOfMemoryError: Java heap space.
    java.lang.OutOfMemoryError: Java heap space
    >
    javax.ejb.EJBException: EJB encountered System Exception: : java.lang.OutOfMemoryError: Java heap space
         at weblogic.ejb.container.internal.EJBRuntimeUtils.throwEJBException(EJBRuntimeUtils.java:145)
         at weblogic.ejb.container.internal.BaseLocalObject.postInvokeCleanup(BaseLocalObject.java:550)
         at weblogic.ejb.container.internal.BaseLocalObject.postInvokeCleanup(BaseLocalObject.java:496)
         at com.adobe.idp.um.businesslogic.directoryservices.DirectorySynchronizationManagerBean_f5g74_ELOImpl.synchronizeProviders(DirectorySynchronizationManagerBean_f5g74_ELOImpl.java:267)
    Joel

  • Low memory issue

    I have an HP Pavilion 2 Quad Proc 2.5 Ghz. Vista 32 bit home prem, 4 GB DDR2, 512 MB Nvidia  8500GT, Realtek Audio (updated driver), Blu ray read/write drive, Two 500 GB interinal drives 7200rpm SATA 3GB, Virtual mem set to windows control. Program PE7
    I shoot video on a Canon SX1 IS. in 1920 x 1080 30f/s h264 MOV files. If I edit less than 4 Min of Video in PE7 all is well. 5 min of video causes low mem warning from PE7. No message from Windows. Adobe tech support said it is a windows virtual memory issue talk to Microsoft, but nothing else. I have HP support but they said it is outside the scope of support. For $ 60 I can get better support one time but no guarantee. I tried Reg cleaners, manual setting virtual mem at 18 GB. That helped  a little. Rebooting helps., but in the end I get "program not responding" or other failure messages. I have read other posts but they were talking 40 min of video or more. I have had no problems with video from my old Sony I have edited over 1 1/2 hr. but 5 min of DH video brings it to a halt.Is anybody editing from this model Canon with sucess? I can play video from the camera with HDMI cable and it is beautiful. Any ideas?
    Richard

    Steve
    I can understand how it would help process time but helping memory is a strange concept. Since less compression means bigger files. Memory management is way beyond my understanding. I have been working with computers since the PET and TRS80 with 2 kb upgrade and I want to use a hammer some days. I feel Adobe has a way to go with Memory management so that we can just open the program and edit video. I can understand taking a long time to process data in virtual memory but why the "program stopped" message? It is as if virtual memory is not used. Also data not used at the moment could be sent to the disk. It would be a lot slower but no memory crash. I can remember running large databases on 8 bit machines with 64 kb of memory. Of course a large database was smaller than one or two sec of video.
    Thanks for the reply
    Richard

  • LR2 Memory Issues

    LR2 seems to have memory issues. My most frequent experiences with these are:
    - Exporting about 60-100 images at a time. Sometimes LR2 will export all of them, sometimes it will fail and tell me that it is out of memory. Exiting and restarting LR2 allows the export to complete. (I *do* like that it remembers which images were selected prior to exiting, it saves having to figure out what I had just tried to export. Kudos!)
    - Sometimes when I export I get corrupted jpegs. Usually this is just before (or during) one of the above mentioned out of memory errors. I have to be sure to check everything that was exported to make sure it exported cleanly (major pain in the kazoo).
    - Sometimes LR2 hangs and I have to use Task Manager to shut it down.
    - Sometimes LR2 simply "gives up the ghost" and goes away on its own.
    While I don't have proof of the last two being memory issues, I highly suspect that they are.
    All of the above are on a Windows XP Home machine with 4G of memory.

    Michael Shaffer -- the 1.5 to 2x RAM recommendation by Adobe engineers was specifically the Windows VM settings and has nothing to do with the scratch disk space (temporary files). Photoshop uses Windows' virtual memory management and page file, but it supplements that with its own management of the memory space allocated by the OS, rather than continually giving back memory when it's not immediately in use for image data.
    Kenneth Daves -- I'm not sure what the URLs you cited are supposed to tell me. The first one confirms that the page file is used for disk storage of virtual memory and that memory is continually swapped or paged between RAM and disk; it also confirms that the maximum per-process memory space for Win32 programs is 2GB (unless they are compiled to use large addresses and the /3GB switch is used). The remaining ones discuss how memory usage differs in 64 bit OSs.
    I agree that virtual address space is the better choice of words to describe the overall memory space used in Windows. Virtual Memory is typically used to refer more specifically to the page file. And now that I understand that your reference to 2 GB of User Space and Kernel Space was intended to describe the the default
    maximum of each of those spaces (absent the /3GB switch and large address space compilation), I agree completely.
    Yes, when an application has used all of its permissible virtual address space, needs a memory allocation, and cannot release any of the existing memory, it is "out of memory." However, Adobe complicates this somewhat, because when there is insufficient space for its scratch files, it displays an error message stating that it is "out of memory (RAM)" which is misleading and incorrect in numerous ways. I don't specifically know how Lightroom employs memory and/or temporary file space -- other than that it often uses as much virtual address space as it can get.

  • Continual Memory Issues

    Continual Error - "Can't Finish Previewing.  there Isn't enough Memory ID=-108"
    Illustrator CS5; windows 7; 2x quad core xeon; 24GB RAM.
    Only a small amount of system resources appear to be in use.  Is this a bug or do i need to modify a setting?

    My limited understanding is that memory is distributed evenly between cores.
    I could go into some detail about thread handling and memory allocation, but let's just say that's not how things work. That aside, AI's memory handling is just not efficient - whatever that could possibly mean. That 1.8 GB is a clear sign that you are approaching the overall limit of 2.37 GB or so (content and the app itself) and the app simply struggles to find any free memory at all. As Jesse suggested, seriously consider upgrading to CS6, if all those issues seem memory related. There are other and some new annoyances in CS6, but at least being able to work at all goes a long way to improve ones mood...
    Mylenium

  • Application is not working due to memory issue

    Hi Friesnds,
    Kindly help me regarding settings in java. Scenario is like that. I have one server having two jboss versions (jboss-4.2.3 GA and jboss- 3.2.6). Each having java memory (JVM settings) 1 GB. The total memory of that server is 3 GB.
    Problem is that every two days my application is not working due to memory issue. Once i freed the memory (through run the commands - (1). sync
    (2). echo 3 > /proc/sys/vm/drop_caches
    Application works fine.
    I heared that the above command (echo 3 > /proc/sys/vm/drop_caches) can't run frequently because of server crash.
    Kindly help me regarding this issue and provide the resolution.
    Let me know if you need more information from my side.
    Thanks
    Ashish Shukla

    All of the above.  I tested on 4 different networks and had no luck, I also had friends test my network and the other networks with their iPhones and all had no issues using FaceTime.
    Also after I wiped my phone (erased all content and settings) I was able to successfully initiate a FaceTime call with no changes to my network and was also able to initiate another call when connected to another previously tested network that didn't work before.  Once I restored from backup though, FaceTime stopped working again.
    I did find this discussion: https://discussions.apple.com/thread/5163024?start=0&tstart=0
    and tried the suggestions found there too.  That discussion describes my issue I am having as well.

  • How to correct internal memory issue

    This is related to the standard SAP program SAPRCKM_MR11 in which there is one include RCKM_MR11F01 which contains a select statement :
    SELECT *
         INTO CORRESPONDING FIELDS OF TABLE t_bhistory
         FROM v_ckmlgrir
         WHERE bukrs EQ p_bukrs
         AND   ebeln IN r_ebeln
         AND   ebelp IN r_ebelp
         AND   lifnr IN r_lifnr
         AND   ekorg IN r_ekorg
         AND   ekgrp IN r_ekgrp
         AND   werks IN r_werks
         AND   bedat IN r_bedat
         AND   bsart IN r_bsart
         AND   bstyp IN ht_bstyp
         AND   vgabe IN ('1', '2', '3')  "Wareneingang/Rechnung/Nachbel.
         AND   loekz_k NE 'L'            "Loeschkz. Bestellkopf
         AND   loekz_p NOT IN ('L', 'S') "Loeschkz. Bestellposition
         AND ( frgrl   EQ space          "Freigabe unvollständig
         OR    frgrl   IS NULL )
         AND ( memory  EQ space          "Bestellung noch nicht komplett
         OR    memory  IS NULL )         "falls Feld nicht initialisiert
         AND ( xwoff   EQ space          "Wertbildung offen
         OR    xwoff   IS NULL )         "falls Feld nicht initialisiert
    bei Dienstleistungsbestellungen werden mehrfach kontierte
    Bestell-Pos. gegen das WE/RE-Konto gebucht, deshalb dürfen
    mehrfach kontierte Pos. nicht generell überlesen werden.
          AND   vrtkz   EQ space          "keine Mehrfachkontierung
         AND   wepos   EQ 'X'            "Wareneingang wird erwartet
         AND   repos   EQ 'X'            "Rechnung wird erwartet
         AND   weunb   EQ space          "Wareneingang unbewertet
         AND   stapo   EQ space.         "Item is not statistical
    There is one job FMT612P which contains this standard program and the job is going to abend
    because of the select statement given above which is because of some memory issue with the internal table t_bhistory, please give us a suggestion how we can avoid the same.

    Hi,
    I am also facing the same issue. We had copied this report into a custom report and are trying to improve performance by breaking the view into separate queries on EKKO, EKPO and EKBE.
    Can someone advise the best way to deal with this view v_ckmlgrir or a best approach to break down the view into separate queries.
    Thanks
    Edited by: Shreyas Shrikant Kulkarni on Jun 29, 2009 4:56 PM

  • Memory issue with Podcast !!!

    Hi,
    I have several issues with the Podcast App.
    - i cannot sync podcast that i have downloaded on my laptop using itunes to the iphone,
    - when i download some podcast directly on the iphone, i cannot play them, i still have to stream them,
    - i have uge memory issues and i think that the downloaded version of the podcast might be somewhere on the iphone, but cannot access / delete them (over 16 GO, i have more than 4.5 GO used for 'other', that are not related to anything...)
    i have tried to delete / reinstall the app, but this does not change anything...
    Any help ?
    Rgds

    Quote
    I have a MSI 975X Platinum rev 2 mb with a e6300 processor overclocked to 2.24ghz.
    If your CPU operates @2.24 GHz, it means that you increased FSB frequency to about 348 MHz.
    To make sure there is no misunderstanding here, let me point out the following:
    Memory frequency is ALWAYS linked to FSB frequency by the FSB/DRAM Ratio you can select in BIOS.  If you select 1:1 for example, the BIOS will show that memory frequency is 533 MHz. This value however, only applies to the default FSB clock speed of 266 MHz:
    266 MHz x 1 = 266 MHz (or 533 MHz effective DDR2-frequency).
    If your system is overclocked, what counts is only FSB/DRAM ratio, not the memory speed displayed in BIOS.  That means, if you set FSB/DRAM ration to 1:1 and FSB frequency to 348 MHz, your RAM will operate @:
    348 MHz x 1 = 348 MHz (or 696 MHz effective DDR2-frequency).
    The main question is:
    What are you talking about when you say, you can't make your RAM operate @800 MHz?
    The BIOS does not offer a proper divider to get your RAM to 800 MHz @348 FSB clock speed to begin with.
    You have the following choices if your overclock your system:
    FSB=300 + FSB/DRAM ratio = 1.33
    300 MHz x 1.33 ~ 400 MHz (or 800 MHz effective DDR2-frequency)
    FSB=320 MHz + FSB/DRAM ratio = 1:1.25
    320 MHz x 1.25 = 400 MHz (or 800 MHz effective DDR2-frequency)
    FSB=400 MHz + FSB/DRAM ratio = 1:1
    400 MHz x 1 = 400 MHz (or 800 MHz effective DDR2-frequency)
    Use CPU-Z to monitor the DRAM frequency that is actually set if you are overclocking.

  • Question for those with memory issues on k8n neo plat

    I have been having issues with my memory on the k8n neo platinum board since i built the system last month and i just recently discovered a fact I had overlooked.
    I Have discovered that  if i test the system while it is cool (having sat turned off overnight for example), then memtest will pass the ram just fine and the system will be stable for a short period of time, generally longer if no intensive applications are used, ie. doom3 vs. websurfing).
    This system, once it heats up ( seems mostly to be the nvidia 3 nb that needs to be "hot"- that is, a probe placed in contact with the module reaches about 58+ degrees c) then the system becomes unstable and memtest, if used, fails.
    Prime95, if used on it when its cool, will usually take about an hour or so to fail, but if its used on the system when its been running a bit ( temps from nb at 58c+ and memory module hot to the touch) then it fails after about 20-30 seconds.
    I have placed an active hs on the nb in an attempt to correct this, but temps still remain high with failures occuring.
    I have spent the better part of three days testing the sytem this way and the above facts have held true in all these tests.
    I have  tested at stock ratings ( not using "auto" in bios, but using "approved ratings" of 10 x 200 for fsb and timings of 2-3-3-7 for my ram at 1:1 (which is recommended as well) since i am testing for base stability at this point;Voltages have been left at auto settings, however.
    I was wondering if anyone else with memory issues could try this out and post if this seems to be the case for you as well.
    Perhaps the memory issues on these boards are directly related to a problem with the nvidia 3 chipset running too hot?
    Does anyone have a memory issue with these boards who is using water cooling on the NB ?
     Lack of issues from people using 'problem" memory, but having no failures with water cooled nb's would support my assumption.

    Yes i did mean chipset.  
    I am using a stock hs/fan from amd, which seems to run a tad hot but, considering that I wont overclock until I can get board stable, its not an issue. (runs 35-40 idle and up to 72 intensive). I need the warranty more then the extra mhz at present.
    I removed the stock heatsink from the nvidia chipset and bought a generic chipset cooler (green, very little in the way of fins and seems to be aluminum) with a small fan.
     I used a generic silver paste from Compusa to adhere both the cpu hs and the chipset.
    I didnt lap either hs, just installed each as per instructions on artic silvers website (figured it had to go on the same way even generic).
    I also bought a kingwin thermal center which i ran probes from to the cpu, case, and nb chipset.
     I am not using it to adjust fans so those arent connected - its just a reference temp I can see at a glance, esp considering the funky bios reading this boards been known to give.
    I assume from these responces that even a nvidia 3 250 chipset shouldnt get as hot as mine?
    I asked about that from the retailer i bought this from but he seemed to think it was "ok" at 50ish celcius, but then again he didnt look too confident when he said it, lol.
    Idle right now reads on Corecenter at 28 celcius for cpu and 33 celcius for the chipset, although, I have readings of 37 c and 43 c from the kingwin thermal center, which i believe to be more accurate.
    My voltage reads at 11.67 for the 12v line at idle right now.
    The psu i got is stick from the xblade case i bought in store. (http://www.xoxide.com/xblade2.html)
    I have a dvd burner, cd drive, and two maxtor ide hd's attached to the system;The only card is the agp.
    Is there a way to check and see if the cpu ispower starved? software monitor or even an external probe perhaps?
    I havent put a probe on it, but the ram came in its own heatspreader, which gets very hot to the touch as well.
    What temps should I be looking for on these items?

  • Firefox 3.6.8 and memory issues

    After the last update to 3.6.6, I disabled plugin-container per the instructions given on the forum, and my RAM problems went away. However, after the automatic update to 3.6.8, the memory problems came back, with memory usage going up to 92%. I downgraded to 3.6.4, and turned plugin-container off again; while memory usage is currently at 84%, it is not as low as it was before the upgrade to 3.6.8. (79-80%) (There is very little difference in the websites being opened on the various tabs.)
    == This happened ==
    Every time Firefox opened
    == I upgraded to 3.6.8

    ffs guys. It's a free browser. It's cool. We love it. But can we do something about memory issues? Almost every release has huge memory leaks.
    How about halving the release points and focus on memory issues for the other half of the time. I am a long time user since way back in the day and it seems to me that instead of trying to release so many versions, we should be focused more fixing the current release.
    I say put a moratorium on new features until bloat can be fixed. These experiences are starting to give me a hardon for IE again.
    Screw new features. I can deal with plain old tabs like they are. I don't need more bells and whistles. This browser has been kicking ass and taking names for a loonnnggggg time. We don't need more bloat. Keep the security fixes coming but everything else can wait. Take a breath. Stop and smell the roses for a few weeks.

  • Regarding Memory issues((error) while scheduling the job

    Hi Friends
    I am  facing memory issue while rerunning BODS jobs in production.As we have rapid mart when jobs got failed i rerun the
    job suddenly job got failed and  in logs i found Memory issue
    please help me what i have to do in steps
    Thanks

    I think there is no one solution to buffer pool issue. Buffer pool issue happens due to many reasons like how you design the data flow, requirement of push down or even insufficient memory in the running environment.
    You can check by checking 'collect statistics for optimization' during running the job. Also in the data flow, try changing from pageable cache to memory cache by right clicking and selecting properties.  The below link give some details on caching, might be helpful.
    https://wiki.sdn.sap.com/wiki/display/BOBJ/CachinginDI
    Arun

  • Memory issues while fetching content

    HI all,
    I am using SAP KM Apis. however i am facing some memory issues while using them.
    I am using getContent() api on a resource. and then getInputStream on the Icontent object. I am using a ByteArrayOutputStream for writing the data. here is the code snippet:
    byte[] buf= new byte[4096];
    InputStream inputStream = content.getInputStream();
    ByteArrayOutputStream os = new ByteArrayOutputStream(1024)
    while ((count = inputStream.read(buf)) != -1)
        os.write(buf, 0, count);
    byte[] arr=os.toByteArray();
    THis code consuimes a lot of memory . we need to optimize this. we also tried removing BYteArrayOutputSteam and directly read the bytes in chunk :
    int n; 
    do 
         n = inputStream.read(buf, 0,(int)content.getContentLength()); 
    while (n != -1); 
    however this approach has not helped us.
    please suggest an approach where memory consumption is less and the entire content is fed.
    thanks

    Hi
    This code worked for me for reading from a KM res...
    BufferedReader reader = new BufferedReader(new InputStreamReader(is));
    StringBuilder sb = new StringBuilder();
    String line = null;
            try {
                while ((line = reader.readLine()) != null) {
                    sb.append(line + "\n");
            } catch (IOException e) {
                e.printStackTrace();
            } finally {
                try {
                    is.close();
                } catch (IOException e) {
                    e.printStackTrace();
    return sb.toString();
    Regards
    BP

  • Memory issues - Bring demand paging to X6.

    does anyone at nokia remember a few years ago a firmware update that added demand paging to N95 and enabled the phone to do multi-tasking properly without having always memory issues? Please bring it to the X6. It's frustating to run only one app (in this case, opera mobile) and having memory warnings always poping up... After a fresh start of the phone.
    In the wise words of a famous Top Gear presenter: how hard can it be?

    The numbers below were constantly moving as I copied them so they're not exact but here's what task manager reported. Funny thing is that it always seems to show around 3 GB available.
    Physical:
    Total - 12581952 Kb
    Available - 3377056 Kb
    System Cache - 3237124 Kb
    Kernel Memory:
    Total - 204288 Kb
    Paged - 143704 Kb
    Non-Paged - 60056 Kb
    I just saw that OEM was advising "Increase the size of the SGA by setting the parameter "sga_target" to 10752 M"

Maybe you are looking for

  • RAM issues with K7N420

    I bought two extra sticks of 256 MB DDR 2100 RAM and I've installed them into slots 2 and 3 (along with the original 256 I had in slot 1) and the system doesn't recognise that they are there. I read the manual and it shows in the appendix the configu

  • Align By Page in 11.5

    I have noticed in 11.5 that the option to align a group of objects by the page margins while keeping the objects relative to each other is gone. Please tell me this option is just hidden somewhere and is still available! Blanchard

  • SAP Business Workflow Vs UWL

    As a part of our first release - we had HR and SRM and had major Portal usage. UWL was leveraged for the HR & SRM tasks. Currently for our next release - we dont have any HR / SRM - But core business processes around SD, MM, FI modules and not any fu

  • Issues with connecting to BI presentation services

    Hello, I have downloaded the latest 11g version with the bi extensions. I am trying to connect to BI presentation services. However I am getting the following: Any idea???? java.util.MissingResourceException: Can't find resource for bundle oracle.jde

  • Default ssl context init failed: Cannot resolve key

    Hi, I get this SSL Exception when I try to run my server using ssl socket: "default ssl context init failed: Cannot resolve key" it is thrown at this line: "sslServerFactory.createServerSocket(port)" I created a kestore and trustore files using 'keyt