Business Connector - Out of Memory error

Hello,
We have several services running from business connector and some which are called from SAP as remote function calls.  One error which keeps resurfacing is:
java.lang.reflect.InvocationTargetException:OutOfMemoryError
Can anyone tell me how to fix this?  At the moment we restart BC and move some log files and it seems to do the trick but we need to start being pro-active to minimise disruption to our clients.
Any suggestions welcome.
Thanks,
Gill

Hi,
you should take care that you sweep regularly the SAP-transaction-log. There is a transaction 'sweepTransaction' which can be scheduled on a regurlar basis to get rid of old transactions.
Another possible reason is that your service is consuming vast amounts of main memory, this could happen with very large objects, i.e. extremely big xml-files.
Hope you find a solution.
Regards, Astrid

Similar Messages

  • "returning error page for JRun too busy or out of memory"

    Hello,
    I've got some trouble to update a ColdFusion linux based.
    We have already tried to update from ASPB11-04 and ASPB11-14 but we always have the folloing erreor message :
    "returning error page for JRun too busy or out of memory"
    We really need a solution ASAP. Thanks!
    Jeff

    Can you share the exception and application logs having the error entries.
    Regards,
    Anit Kumar

  • Acrobat XI Pro "Out of Memory" Error.

    We just received a new Dell T7600 workstation (Win7, 64-bit, 64GB RAM, 8TB disk storage, and more processors than you can shake a stick at). We installed the Adobe CS6 Master Collection which had Acrobat Pro X. Each time we open a PDF of size greater than roughly 4MB, the program returns an "out of memory" error. After running updates, uninstalling and reinstalling (several times), I bought a copy of Acrobat XI Pro hoping this would solve the problem. Same problem still exists upon opening the larger PDFs. Our business depends on opening very large PDF files and we've paid for the Master Collection, so I'd rather not use an freeware PDF reader. Any help, thoughts, and/or suggestions are greatly appreciated.
    Regards,
    Chris

    As mentioned, the TEMP folder is typically the problem. MS limits the size of this folder and you have 2 choices: 1. empty it or 2. increase the size limit. I am not positive this is the issue, but it does crop up at times. It does not matter how big your harddrive is, it is a matter of the amount of space that MS has allocated for virtual memory. I am surprised that there is an issue with 64GB of RAM, but MS is real good at letting you know you can't have it all for use because you might want to open up something else. That is why a lot of big packages turn off some of the limits of Windows or use Linux.

  • Report with more than 600 kb image - BO Server getting Out Of memory Error

    Hi,
       We have a report which displays images and size is above 600 KB.We are getting an "Out Of memory" Error while previewing this report in Business Object Server.
    In another situation we tried by giving a dynamic path to the OLE object, but the report is not showing error. We are not able to view the image whose size is more than 600KB
    Notes: 1. Image is stored as a BLOB  in Oracle Database
               2. Connection used inside the report is OLEDB
               3. UNIX Environment
    Regards,
    Sathish

    Please re-post if this is still an issue to the Business Objects Forum or if you have a valid support contract create a case on line.

  • Out of Memory Error in iplanet 6.1

    While starting iplanet 6.1 sp2 in HP-UX 11.00 , out of memory error occurs. Value for JVM heapsize is set to min 128 MB and Max of 512 MB. Tried changing both the values to 512 MB, still the error occurs.
    Please revert with any solution.

    Please find the values of JVM HeapSIze and HP-UX process and address space limits.
    JVM Heap size has been changed to
    Min 128MB
    Max 2GB
    max_thread_proc 2048
    maxdsiz 1073741824
    maxssiz 401604608
    maxtsiz 1073741824
    After the modification of the above changes , out of memory errors occurs.Find the logfile below.
    [21/Feb/2005:09:37:31] failure ( 8528): for host 163.38.174.17 trying to POST /wect/servlets/com.citicorp.treasury.westerneurope.maintenance.PageLinksServlet, service-j2ee reports: StandardWrapperValve[PageLinksServlet]: WEB2792: Servlet.service() for servlet PageLinksServlet threw exception
    javax.servlet.ServletException: WEB2664: Servlet execution threw an exception
    at org.apache.catalina.core.StandardWrapperValve.invokeServletService(StandardWrapperValve.java:793)
    at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:322)
    at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
    at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:212)
    at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
    at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:209)
    at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
    at com.iplanet.ias.web.connector.nsapi.NSAPIProcessor.process(NSAPIProcessor.java:161)
    at com.iplanet.ias.web.WebContainer.service(WebContainer.java:578)
    ----- Root Cause -----
    java.lang.OutOfMemoryError
    [21/Feb/2005:09:45:04] warning ( 8528): HTTP3039: terminating with 6 sessions st
    ill in progress
    [21/Feb/2005:09:45:04] failure ( 8528): CORE3189: Restart functions not called since 6 sessions still active
    Please suggest.

  • Oracle Service Bus For loop getting out of memory error

    I have a business service that is based on a JCA adapter to fetch an undertimed amout of records from a database.  I then need to upload those to another system using a webservice designed by an external source.  This web service will only accept upto to x amount of records.
    The process:
    for each object in the Jca Response
          Insert object into Service callout Request body
          if object index = number of objects in jca response or object index = next batch index
               Invoke service callout
               Append service callout Response to a total response object (xquery transform)
               increase next batch index by Batch size
               reset service callout to empty body
           endif
    end for
    replace body  with total response object.
    If I use the data set that only has 5 records  and use a batch size of 2 the process works fine.
    If I use  a data set with 89 records  and a batch size of 2 I get the below out of memory error  after about 10 service callouts
    the quantity of data in the objects is pretty small, less than 1kB for each JCA Object
    Server Name:
    AdminServer
    Log Name:
    ServerLog
    Message:
    Failed to process response message for service ProxyService Sa/Proxy Services/DataSync:
    java.lang.OutOfMemoryError: allocLargeObjectOrArray:
    [C, size 67108880 java.lang.OutOfMemoryError: allocLargeObjectOrArray:
    [C, size 67108880 at org.apache.xmlbeans.impl.store.Saver$TextSaver.resize(Saver.java:1700)
    at org.apache.xmlbeans.impl.store.Saver$TextSaver.preEmit(Saver.java:1303) at
    org.apache.xmlbeans.impl.store.Saver$TextSaver.emit(Saver.java:1234)
    at org.apache.xmlbeans.impl.store.Saver$TextSaver.emitXmlns(Saver.java:1003)
    at org.apache.xmlbeans.impl.store.Saver$TextSaver.emitNamespacesHelper(Saver.java:1021)
    at org.apache.xmlbeans.impl.store.Saver$TextSaver.emitElement(Saver.java:972)
    at org.apache.xmlbeans.impl.store.Saver.processElement(Saver.java:476)
    at org.apache.xmlbeans.impl.store.Saver.process(Saver.java:307)
    at org.apache.xmlbeans.impl.store.Saver$TextSaver.saveToString(Saver.java:1864)
    at org.apache.xmlbeans.impl.store.Cursor._xmlText(Cursor.java:546)
    at org.apache.xmlbeans.impl.store.Cursor.xmlText(Cursor.java:2436)
    at org.apache.xmlbeans.impl.values.XmlObjectBase.xmlText(XmlObjectBase.java:1500)
    at com.bea.wli.sb.test.service.ServiceTracer.getXmlData(ServiceTracer.java:968)
    at com.bea.wli.sb.test.service.ServiceTracer.addDataType(ServiceTracer.java:944)
    at com.bea.wli.sb.test.service.ServiceTracer.addDataType(ServiceTracer.java:924)
    at com.bea.wli.sb.test.service.ServiceTracer.addContextChanges(ServiceTracer.java:814)
    at com.bea.wli.sb.test.service.ServiceTracer.traceExit(ServiceTracer.java:398)
    at com.bea.wli.sb.pipeline.debug.DebuggerTracingStep.traceExit(DebuggerTracingStep.java:156)
    at com.bea.wli.sb.pipeline.PipelineContextImpl.exitComponent(PipelineContextImpl.java:1292)
    at com.bea.wli.sb.pipeline.MessageProcessor.finishProcessing(MessageProcessor.java:371)
    at com.bea.wli.sb.pipeline.RouterCallback.onReceiveResponse(RouterCallback.java:108)
    at com.bea.wli.sb.pipeline.RouterCallback.run(RouterCallback.java:183)
    at weblogic.work.ContextWrap.run(ContextWrap.java:41)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:545)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256) at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)
    Subsystem:
    OSB Kernel
    Message ID:
    BEA-382005
    It appears to be the service callout that is the problem (it calls another OSB service that logins and performs the data upload to the External service)  because If I change the batch size up to 100  the loop will load all the 89 records into the callout request and execute it fine.  If I have a small batch size then I run out of memory.
    Is there some settings I need to change?  Is there a better way in OSB (less memory intensive than service callout in a for loop)?
    Thanks.

    hi,
    Could you please let me know if you get rid off this issue as we are also facing the same issue.
    Thanks,
    SV

  • InDesign - out of memory error

    Hi All,
    One of a clients is experiencing a issue whilst using InDesign CS4 version 6.0.5, they are constantly interrupted by an “Out of Memory” error while using InDesign.
    They have been experiencing this issue for some time now and we have performed memory upgrades and also upgraded to version 6.0.5 (which was meant to solve the issue) several months ago and yet they are still getting this issue.
    The issue that they are experiencing is the following error that was said to have been resolved in the 6.0.5 update.
    Document containing hundreds of text frames and custom baseline grids takes a long time to open and
    causes an “out of memory” error, followed by an unexpected quit. [2253219]
    Has there been another hot fix or a known resolution for this issue? Anyone else experiencingthis?
    Cheers,
    Allan

    Hi Scott,
    Their system has the following specs;
    OS:     Windows Vista Business 32-Bit
    RAM:  3GB DDR2
    CPU:   E8400 @ 3.00Ghz (Dual Core)
    Garphics: Onboard intel G33/G31
    Regards,
    Allan

  • Acrobat XI Pro "Out of Memory" error after Office 2010 install

    Good Afternoon,
    We recently pushed Office 2010 to our users and are now getting reports of previous installs of Adobe Acrobat XI Pro no longer working but throwing "Out of Memory" errors.
    We are in a Windows XP environment. All machines are HP 8440p/6930p/6910 with the same Service pack level (3) and all up to date on security patches.
    All machines are running Office 2010 SP1.
    All machines have 2GB or 4GB of RAM (Only 3.25GB recognized as we are a 32bit OS environment).
    All machines have adequate free space (ranging from 50gb to 200gb of free space).
    All machines are set to 4096mb initial page file size with 8192mb maximum page file size.
    All machines with Acrobat XI Pro *DO NOT* have Reader XI installed alongside. If Reader is installed, it is Reader 10.1 or higher.
    The following troubleshooting steps have been taken:
    Verify page file size (4096mb - 8192mb).
    Deleted local user and Windows temp files (%temp% and c:\WINDOWS\Temp both emptied).
    Repair on Adobe Acrobat XI Pro install. No change.
    Uninstall Acrobat Pro XI, reboot, re-install. No change.
    Uninstall Acrobat Pro XI Pro along with *ALL* other Adobe applications presently installed (Flash Player, Air), delete all Adobe folders and files found in a full search of the C drive, delete all orphaned Registry entries for all Adobe products, re-empty all temp folders, reboot.
    Re-install Adobe Acrobat XI Pro. No change.
    Disable enhanced security in Acrobat XI Pro. No change.
    Renamed Acrobat XI's plug_ins folder to plug_ins.old.
    You *can* get Acrobat to open once this is done but when you attempt to edit a file or enter data into a form, you get the message, "The "Updater" plug-in has been removed. Please re-install Acrobat to continue viewing the current file."
    A repair on the Office 2010 install and re-installing Office 2010 also had no effect.
    At this point, short of re-imaging the machines (which is *not* an option), we are stumped.
    We have not yet tried rolling back a user to Office 2007 as the upgrade initiative is enterprise-wide and rolling back would not be considered a solution.
    Anyone have any ideas beyond what has been tried so far?

    As mentioned, the TEMP folder is typically the problem. MS limits the size of this folder and you have 2 choices: 1. empty it or 2. increase the size limit. I am not positive this is the issue, but it does crop up at times. It does not matter how big your harddrive is, it is a matter of the amount of space that MS has allocated for virtual memory. I am surprised that there is an issue with 64GB of RAM, but MS is real good at letting you know you can't have it all for use because you might want to open up something else. That is why a lot of big packages turn off some of the limits of Windows or use Linux.

  • RoboHelp 9 gives an out of memory error and crashes when I try to import or link a Frame 10 file or

    I have Tech Suite 3. If I start a new RoboHelp project and try to import or link Frame files, RoboHelp tries for a while, then I get an Out of Memory error and the program crashes.
    I opened one of the sample projects and was able to link to one of my frame files without any problem, so it seems to be an issue with creating something new.
    Any suggestions?

    It happens when I create a new project and then try to import or link frame docs to make up the content. It starts scanning, then crashes. I did get it to the conversion setting page once, but no further.
    It does not happen if I open one of the supplied example projects and link a file. But then it doesn't let me choose, during import, any style mapping. And I can't delete the sample project fold
    Twice now it has told me when I tried to import (not link, but import) that my .fm file could not be opened, and told me to verify that Frame is installed (it is) and that the file is a valid frame file (it is).
    The docs and project are in separate folders on my C: drive.

  • Out of Memory Error While deploying as EAR file

    Hai,
    I was trying to deploy an EAR file of size 63 MB which inturn containing about 60 EJB.jars. No WARs. application.xml has all the entries for the JARs. While I am deploying it is giving Out of Memory Error. Is there any way to tweak this problem. I am using my own hand written java application which uses the SunONE deployment APIs for deployment. Can u please tell how to tackle this problem. I am running my application through a batch file which uses jdk1.4.
    Please help me regarding this issue.

    You can set the initial heap size and maximum heap size for the JVM, either in the app-server admin console, or maybe in one of your scripts. You look-up the syntax!...
    I had this error yesterday. I too had run out of memory (150Mb). You simply need to allocate more to the app-server.

  • Uploading large files from applet to servlet throws out of memory error

    I have a java applet that needs to upload files from a client machine
    to a web server using a servlet. the problem i am having is that in
    the current scheme, files larger than 17-20MB throw an out of memory
    error. is there any way we can get around this problem? i will post
    the client and server side code for reference.
    Client Side Code:
    import java.io.*;
    import java.net.*;
    // this class is a client that enables transfer of files from client
    // to server. This client connects to a servlet running on the server
    // and transmits the file.
    public class fileTransferClient
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    // this method transfers the prescribed file to the server.
    // if the destination directory is "", it transfers the file to
    "d:\\".
    //11-21-02 Changes : This method now has a new parameter that
    references the item
    //that is being transferred in the import list.
    public static String transferFile(String srcFileName, String
    destFileName,
    String destDir, int itemID)
    if (destDir.equals(""))
    destDir = "E:\\FTP\\incoming\\";
    // get the fully qualified filename and the mere filename.
    String fqfn = srcFileName;
    String fname =
    fqfn.substring(fqfn.lastIndexOf(File.separator)+1);
    try
    //importTable importer = jbInit.getImportTable();
    // create the file to be uploaded and a connection to
    servlet.
    File fileToUpload = new File(fqfn);
    long fileSize = fileToUpload.length();
    // get last mod of this file.
    // The last mod is sent to the servlet as a header.
    long lastMod = fileToUpload.lastModified();
    String strLastMod = String.valueOf(lastMod);
    URL serverURL = new URL(webadminApplet.strServletURL);
    URLConnection serverCon = serverURL.openConnection();
    // a bunch of connection setup related things.
    serverCon.setDoInput(true);
    serverCon.setDoOutput(true);
    // Don't use a cached version of URL connection.
    serverCon.setUseCaches (false);
    serverCon.setDefaultUseCaches (false);
    // set headers and their values.
    serverCon.setRequestProperty("Content-Type",
    "application/octet-stream");
    serverCon.setRequestProperty("Content-Length",
    Long.toString(fileToUpload.length()));
    serverCon.setRequestProperty(FILENAME_HEADER, destDir +
    destFileName);
    serverCon.setRequestProperty(FILELASTMOD_HEADER, strLastMod);
    if (webadminApplet.DEBUG) System.out.println("Connection with
    FTP server established");
    // create file stream and write stream to write file data.
    FileInputStream fis = new FileInputStream(fileToUpload);
    OutputStream os = serverCon.getOutputStream();
    try
    // transfer the file in 4K chunks.
    byte[] buffer = new byte[4096];
    long byteCnt = 0;
    //long percent = 0;
    int newPercent = 0;
    int oldPercent = 0;
    while (true)
    int bytes = fis.read(buffer);
    byteCnt += bytes;
    //11-21-02 :
    //If itemID is greater than -1 this is an import file
    transfer
    //otherwise this is a header graphic file transfer.
    if (itemID > -1)
    newPercent = (int) ((double) byteCnt/ (double)
    fileSize * 100.0);
    int diff = newPercent - oldPercent;
    if (newPercent == 0 || diff >= 20)
    oldPercent = newPercent;
    jbInit.getImportTable().displayFileTransferStatus
    (itemID,
    newPercent);
    if (bytes < 0) break;
    os.write(buffer, 0, bytes);
    os.flush();
    if (webadminApplet.DEBUG) System.out.println("No of bytes
    sent: " + byteCnt);
    finally
    // close related streams.
    os.close();
    fis.close();
    if (webadminApplet.DEBUG) System.out.println("File
    Transmission complete");
    // find out what the servlet has got to say in response.
    BufferedReader reader = new BufferedReader(
    new
    InputStreamReader(serverCon.getInputStream()));
    try
    String line;
    while ((line = reader.readLine()) != null)
    if (webadminApplet.DEBUG) System.out.println(line);
    finally
    // close the reader stream from servlet.
    reader.close();
    } // end of the big try block.
    catch (Exception e)
    System.out.println("Exception during file transfer:\n" + e);
    e.printStackTrace();
    return("FTP failed. See Java Console for Errors.");
    } // end of catch block.
    return("File: " + fname + " successfully transferred.");
    } // end of method transferFile().
    } // end of class fileTransferClient
    Server side code:
    import java.io.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    import java.util.*;
    import java.net.*;
    // This servlet class acts as an FTP server to enable transfer of
    files
    // from client side.
    public class FtpServerServlet extends HttpServlet
    String ftpDir = "D:\\pub\\FTP\\";
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    public void doGet(HttpServletRequest req, HttpServletResponse resp)
    throws ServletException,
    IOException
    doPost(req, resp);
    public void doPost(HttpServletRequest req, HttpServletResponse
    resp)
    throws ServletException,
    IOException
    // ### for now enable overwrite by default.
    boolean overwrite = true;
    // get the fileName for this transmission.
    String fileName = req.getHeader(FILENAME_HEADER);
    // also get the last mod of this file.
    String strLastMod = req.getHeader(FILELASTMOD_HEADER);
    String message = "Filename: " + fileName + " saved
    successfully.";
    int status = HttpServletResponse.SC_OK;
    System.out.println("fileName from client: " + fileName);
    // if filename is not specified, complain.
    if (fileName == null)
    message = "Filename not specified";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // open the file stream for the file about to be transferred.
    File uploadedFile = new File(fileName);
    // check if file already exists - and overwrite if necessary.
    if (uploadedFile.exists())
    if (overwrite)
    // delete the file.
    uploadedFile.delete();
    // ensure the directory is writable - and a new file may be
    created.
    if (!uploadedFile.createNewFile())
    message = "Unable to create file on server. FTP failed.";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // get the necessary streams for file creation.
    FileOutputStream fos = new FileOutputStream(uploadedFile);
    InputStream is = req.getInputStream();
    try
    // create a buffer. 4K!
    byte[] buffer = new byte[4096];
    // read from input stream and write to file stream.
    int byteCnt = 0;
    while (true)
    int bytes = is.read(buffer);
    if (bytes < 0) break;
    byteCnt += bytes;
    // System.out.println(buffer);
    fos.write(buffer, 0, bytes);
    // flush the stream.
    fos.flush();
    } // end of try block.
    finally
    is.close();
    fos.close();
    // set last mod date for this file.
    uploadedFile.setLastModified((new
    Long(strLastMod)).longValue());
    } // end of finally block.
    } // end - the new file may be created on server.
    } // end - we have a valid filename.
    // set response headers.
    resp.setContentType("text/plain");
    resp.setStatus(status);
    if (status != HttpServletResponse.SC_OK)
    getServletContext().log("ERROR: " + message);
    // get output stream.
    PrintWriter out = resp.getWriter();
    out.println(message);
    } // end of doPost().
    } // end of class FtpServerServlet

    OK - the problem you describe is definitely what's giving you grief.
    The workaround is to use a socket connection and send your own request headers, with the content length filled in. You may have to multi-part mime encode the stream on its way out as well (I'm not about that...).
    You can use the following:
    http://porsche.cis.udel.edu:8080/cis479/lectures/slides-04/slide-02.html
    on your server to get a feel for the format that the request headers need to take.
    - Kevin
    I get the out of Memory Error on the client side. I
    was told that this might be a bug in the URLConnection
    class implementation that basically it wont know the
    content length until all the data has been written to
    the output stream, so it uses an in memory buffer to
    store the data which basically causes memory issues..
    do you think there might be a workaround of any kind..
    or maybe a way that the buffer might be flushed after
    a certain size of file has been uploaded.. ?? do you
    have any ideas?

  • Photoshop CS6 has out of memory errors under Mac OS X 10.7.4

    Has anyone who has a 2008 MacBoo Pro with the NVIDIA GeForce 8600M GT 512 MB graphics card tried running Photoshop CS6 under Mac OS X 10.7.4?
    On my MacBook Pro, PS CS6 launches; however, any action results in an out of memory error.  The CS6 versions of After Effects, Premiere Pro and Illustrator all launch and run as expected under 10.7.4.
    If I switch to my Mac OS X 10.6.8 startup volume, Photoshop CS6 runs as expected.
    I have support cases open with Adobe, NVidia and Apple.  Adobe thinks I need to update the display drivers.  NVidia says display drivers must be provided by Apple.  Apple says that the display drivers are current.  The Apple support agent said that he can escalate the case if the issue can be reproduced on more than one computer.
    I'd love to keep my MacBook Pro running another year or so without having to reboot in 10.6.8 every time I need to use Photoshop.  I'd stay with 10.6.8, but of course iCloud requires 10.7.4.
    I have not tried 10.8 yet, but it's on my troubleshooting list.
    Also, PS CS6 runs fine under Mac OS X 10.7.4 on my 2010 iMac.
    Thanks in advance for any feedback.
    - Warren

    It looks like it was indeed the display driver for the NVIDIA GeForce 8600 GT 512 MB graphics card inside my 2008 MacBook Pro.
    It seems that this driver only gets installed if you upgrade from Mac OS X 10.6.7 to 10.7.4 while the startup drive is connected to the MacBook Pro itself.  I had updated while the startup drive was connected to my 2010 iMac. Accordingly, I could have reinstalled Photoshop 100 times without it ever launching as expected.  Having used the external startup drive with three other Apple computers (all 2010 or newer machines) with Photoshop CS6 launching as expected, I had just assumed (mistakenly) that it would work as expected with my 2008 computer.
    So, this was simple enough to resolve by upgrading the OS on the external startup drive while having it connected to the MacBook Pro.   In hindsight, this makes perfect sense.
    -Warren

  • Final Cut Pro 6 - 'Out of Memory/Error Code'

    Hey All,
    I'm in the final stages of editing a documentary and I keep running into an 'Out of Memory' error code in Final Cut Pro 6.0 when trying to export a reference file. The finished film is 81 minutes long and contains 19 sequences. I have all 19 sequences nested in a main sequence of a seperate FCP project file, so I'm not wasting any memory on clips or additional backups of sequences. This Nested sequence is the only sequence that exists in this 'other' project file. All of the photos used are under 4k resolution and are RGB. Even the lower thirds are animation QT's with alphas. Any thoughs would be greatly appreciated.
    Nick
    System Specs:
    Quad-Core 2GHZ MacBook Pro
    4GB DDR3 Memory
    MAC 10.7.2 (Lion)
    FCP v6.0.6

    First troubleshooting step
    https://discussions.apple.com/docs/DOC-2491
    If that doesn't solve the problem, you may have a corrupt sequence
    Try creating a new sequence with the same settings and copy the contents from the problematic sequence
    If that doesn't work.  Mark an in and an out for the first half of the sequence and do an export.  If that works, mark an in and an out for the second half and export.  Keep narrowing down the sections til you find the problematic clip. 

  • Adobe X Pro 10.1.10 Out of Memory error message

    Hello One, Hello All
    Since updating our Adobe X Pro machines to version update 10.1.10, we occasionally receive an "Out of Memory" error message, which depending on what we are doing, may force us to shutdown all Adobe windows and re-open Adobe, or simply click OK and continue working with the PDF document like nothing is wrong. It seems to have no rhyme or reason to when or why it occurs. I do not see any warnings or errors in the Windows Event Log. It happens with files we have created, and also happens with files received from internal and external sources through email. It is affecting our high end machines: Lenovo X Series and W Series laptops with Windows 8.1 3rd/4th gen i5 CPUs, 8+GB RAM, 128+GB SSD's. We have all system and Windows updates applied. We have Trend Micro and Malwarebytes real-time protection and system scans do not find any malware.
    I have found a few other recent threads on Adobe forum related to this error message but the responses are weak at best with no definitive fix. The system and user temp folders size are less than 100MB each so this cannot be the issue. When the error occurs I check task manager and system utilization, including memory, is well below 100%.
    We did not have this issue prior to version update 10.1.10.
    Really hoping Adobe can step in to help here and hoping boilerplate responses are not used.

    Hi wayne103,
    We released a new security update yesterday that is v10.1.11
    Please update to this version and check the performance.
    I have seen this error message occur while opening a 3rd party created pdf that has corrupted file structure.
    Please let me know if the issue occurs for some specific files or random
    Also let me know the PDF producer for these pdfs.
    Regards,
    Rave

  • Getting out of memory errors in Indesign 5.5. What can I do to fix it?

    Getting out of memory errors in Indesign 5.5. What can I do to fix it?

    Tell your dumb friend to pay you for a new phone as he damaged it. You cannot get help here for a phone that has been taken apart, as it is not user servicible. Your dumb friend also voided your warranty and, even if the warranty were expired Apple will never touch that phone.
    Time to get smarter friends.

Maybe you are looking for

  • Error while running Customer Interest Report

    Hello All, I am configuering Interest on Customer arrears/balance & I am getting an error "NO TIME-DEPENDENT CONDITIONS FOR INDICATOR EF" while running FINT. I have checked all the threads but there is no such query exept one.  Please guide me to sol

  • Reset password for RPD in OBIEE 11G

    Hi. Is there a way to reset or bypass the password to open a RPD in offline mode ? I'm using OBIEE 11.1.1.5 linux, and using the Administrator standalone client in windows. Thanksx !!

  • User Exit to Search for Duplicate BP Records

    Hai Guys, I want to use a "User Exit" that gives me a warning message when I am creating a BP with the same name and zipcode which is already presnet in the system. Right now we are not getting any warning messgae and hence we are having duplicate BP

  • Alarm clock for iOs 3.2.2

    I need an app to wake up me. But i dont want to upgrade for some reasons So it has to work on iOS 3.2.2 and i want my ipad not to be on but locked with an alarm which still works !!!! I tried my alarm clock it does not work

  • Delete Handling Units

    Hi Experts, We met a problem with delivery processing. Due to the mistake in packing instruction, we have 14000 handling unit in a single delivery. And when we try to delete all these handling units, we keep receiving runtime errors. Reading from run