Unknown File, Out of Memory

When opening my project sequence across a gigabit Ethernet connection to an Xsan, I first get: File Error: Unknown File, followed by: Error: Out of Memory.
The Sequence is full of Jpegs, Tiffs, and some 10-Bit Uncompressed. How can I get to view this project..?
Si

As mentioned above, picture files, such as jpegs require RAM in order to playback. You mentioned you have a lot of TIFFs and JPEGs in this project. I'm still willing to bet that is exactly what is causing the Out of Memory error.
Sure, you can copy files over the same network with great speeds, but try taking all of those jpegs and try transferring them simultaneously. I doubt you'll get great speeds doing that.
You could have your editor use the Media Manager to copy the project to an external drive.

Similar Messages

  • Can't open .ppt file: "out of memory" error on Mavericks

    Hello,
    I have a couple of .ppt files that can't open/import in keynote v5.3. It says "import error: out of memory". The files are pretty small, 500kb to 8mb, no videos or music, only images and plain text. I have dozens of free gb.
    This error started popping out after I updated to Mavericks a few days ago.
    Keynote works fine with other .ppt.
    The .ppt can be opened in Power Point (in a Windows PC, I don't have it installed) and with Previews.
    I read that if I update to the new iWork the old version is kept in a separate folder. Do you confirm that I can still use both the old and the new version?
    What can I do about that?
    Thanks for your help.

    This error is usually caused by a corrupt ppt file or an incompatible font.
    A quick work around is to recreate the presentation in Keynote from scratch using the images copied from the original.

  • No Movie in file-Out of memory errors!

    I can't open my file! or related autosave files!
    I get :
    "no movie in file" and "error: out of memory"
    please help!

    I have had a similar kind of message when I've had a file passed between soundtrack or live type back to FCP that has become corrupt. If you are getting a name of a particular file that has no movie, you might want to try deleting it - that may give you an offline message, and it should then allow the project file to open.

  • Rendering print resolution files -- out of memory, out of luck?

    Hey guys!
    Trying to render a 7500x5000 pixel image for a print piece. And not surprisingly, I'm getting out of memory buffer errors. Am I just out of luck trying to render something so massive out of AE (until it goes 64bit)? Trying to render just a small ROI isn't working either.

    I hear your point. Technically it makes since, but I'd have to disagree. After all, the monitors we use can't display 32bpc color space anyway. Almost all the motion graphics we make ends up in 8bpc color on TV or Web. So we're always viewing some sort of drastically reduced color space.
    As a test, I had AE render out a 1/10 res TIFF in 32bpc, reduced the file to 8bpc in Photoshop and did a CYMK soft proof and it looked remarkably faithful to what After Effects was displaying. I took the low-res sample file to the printer and they did a test printing and it looked exactly like the on-screen soft proof. So it's just the effects and transfer modes I want AE to process in HDR space.
    So what I ended up doing was rendering out each layer individually and then used Photoshop to assemble all the layers in 32bpc. While I lost the interactivity of working with live plug-ins in After Effects, I was able to get the look I wanted.
    Looking forward to the day when After Effects goes 64-bit

  • I am running out of memory on my hard drive and need to delete files. How can I see all the files/applications on my hard drive so I can see what is taking up a lot of room?

    I am running out of memory on my hard drive and need to delete files. How can I see all the files/applications on my hard drive so I can see what is taking up a lot of room?
    Thanks!
    David

    Either of these should help.
    http://grandperspectiv.sourceforge.net/
    http://www.whatsizemac.com/
    Or search 'disk size' in the App Store.
    Be carefull with what you delete & have a backup BEFORE you start, you may also want to reboot to try to free any memory that may have been written to disk.

  • RoboHelp 9 gives an out of memory error and crashes when I try to import or link a Frame 10 file or

    I have Tech Suite 3. If I start a new RoboHelp project and try to import or link Frame files, RoboHelp tries for a while, then I get an Out of Memory error and the program crashes.
    I opened one of the sample projects and was able to link to one of my frame files without any problem, so it seems to be an issue with creating something new.
    Any suggestions?

    It happens when I create a new project and then try to import or link frame docs to make up the content. It starts scanning, then crashes. I did get it to the conversion setting page once, but no further.
    It does not happen if I open one of the supplied example projects and link a file. But then it doesn't let me choose, during import, any style mapping. And I can't delete the sample project fold
    Twice now it has told me when I tried to import (not link, but import) that my .fm file could not be opened, and told me to verify that Frame is installed (it is) and that the file is a valid frame file (it is).
    The docs and project are in separate folders on my C: drive.

  • Can't open vcf file. "The file may not exist ... Out of memory or system resources"

    Trying to open a vcf file (vCard) from a user.  After double-clicking file, get message:
    Cannot open file: C:\folder\filename.vcf. The file may not exist, you may not have permission to open it, or it may be open in another program. Right-click the folder that contains
    the file, and then click Properties to check your permissions for the folder. Out of memory or system resources. Close some windows or programs and try again.
    I have Outlook 2010 with SP2.
    Similar error, different problem.  Has workaround:
    http://community.office365.com/en-us/f/172/t/198029.aspx
    I turn OFF cached exchange mode.  I double-click the vcf file, it opens fine.
    I turn ON cached exchange mode.  I double-click the vcf file, I get the erorr message again.
    We have hosted Exchange, so we can't turn off cached exchange mode in Outlook.
    Is this a bug?  Anybody know of a fix?
    By the way, if I save one of my contact as a vcf file
    and double-click it, it opens fine.

    Hi,
    I'm marking the reply as answer as there has been no update for a couple of days.
    If you come back to find it doesn't work for you, please reply to us and unmark the answer.
    Best Regards,
    Steve Fan
    Forum Support
    Come back and mark the replies as answers if they help and unmark them if they provide no help.
    If you have any feedback on our support, please click
    here

  • I have a file where I am running out of memory can anyone take a look at this file and see?

    I am trying to make this file 4'x8'.
    Please let me know if anyone can help me get this file to that size.
    I have a quad core processor with 6 gig of ram and have gotten the file to 50"x20", but I run out of memory shortly thereafter.  Any help would be appreciated.
    Thanks,

    Where to begin? You should look into using a swatch pattern instead of those repeating circles. Also, I see that each circle in your pattern is actually a stack of four circles, but I see no reason why. Perhaps Illustrator is choking on the huge number of objects required to make that patterns as you haave constructed it.
    Here is a four foot by eight foot Illustrator file using a swatch pattern. Note that, despite the larger size, the file is less than one 16th the size.

  • Out of Memory Error While deploying as EAR file

    Hai,
    I was trying to deploy an EAR file of size 63 MB which inturn containing about 60 EJB.jars. No WARs. application.xml has all the entries for the JARs. While I am deploying it is giving Out of Memory Error. Is there any way to tweak this problem. I am using my own hand written java application which uses the SunONE deployment APIs for deployment. Can u please tell how to tackle this problem. I am running my application through a batch file which uses jdk1.4.
    Please help me regarding this issue.

    You can set the initial heap size and maximum heap size for the JVM, either in the app-server admin console, or maybe in one of your scripts. You look-up the syntax!...
    I had this error yesterday. I too had run out of memory (150Mb). You simply need to allocate more to the app-server.

  • Server goes out of memory when annotating TIFF File. Help with Tiled Images

    I am new to JAI and have a problem with the system going out of memory
    Objective:
    1)Load up a TIFF file (each approx 5- 8 MB when compressed with CCITT.6 compression)
    2)Annotate image (consider it as a simple drawString with the Graphics2D object of the RenderedImage)
    3)Send it to the servlet outputStream
    Problem:
    Server goes out of memory when 5 threads try to access it concurrently
    Runtime conditions:
    VM param set to -Xmx1024m
    Observation
    Writing the files takes a lot of time when compared to reading the files
    Some more information
    1)I need to do the annotating at a pre-defined specific positions on the images(ex: in the first quadrant, or may be in the second quadrant).
    2)I know that using the TiledImage class its possible to load up a portion of the image and process it.
    Things I need help with:
    I do not know how to send the whole file back to servlet output stream after annotating a tile of the image.
    If write the tiled image back to a file, or to the outputstream, it gives me only the portion of the tile I read in and watermarked, not the whole image file
    I have attached the code I use when I load up the whole image
    Could somebody please help with the TiledImage solution?
    Thx
    public void annotateFile(File file, String wText, OutputStream out, AnnotationParameter param) throws Throwable {
    ImageReader imgReader = null;
    ImageWriter imgWriter = null;
    TiledImage in_image = null, out_image = null;
    IIOMetadata metadata = null;
    ImageOutputStream ios = null;
    try {
    Iterator readIter = ImageIO.getImageReadersBySuffix("tif");
    imgReader = (ImageReader) readIter.next();
    imgReader.setInput(ImageIO.createImageInputStream(file));
    metadata = imgReader.getImageMetadata(0);
    in_image = new TiledImage(JAI.create("fileload", file.getPath()), true);
    System.out.println("Image Read!");
    Annotater annotater = new Annotater(in_image);
    out_image = annotater.annotate(wText, param);
    Iterator writeIter = ImageIO.getImageWritersBySuffix("tif");
    if (writeIter.hasNext()) {
    imgWriter = (ImageWriter) writeIter.next();
    ios = ImageIO.createImageOutputStream(out);
    imgWriter.setOutput(ios);
    ImageWriteParam iwparam = imgWriter.getDefaultWriteParam();
    if (iwparam instanceof TIFFImageWriteParam) {
    iwparam.setCompressionMode(ImageWriteParam.MODE_EXPLICIT);
    TIFFDirectory dir = (TIFFDirectory) out_image.getProperty("tiff_directory");
    double compressionParam = dir.getFieldAsDouble(BaselineTIFFTagSet.TAG_COMPRESSION);
    setTIFFCompression(iwparam, (int) compressionParam);
    else {
    iwparam.setCompressionMode(ImageWriteParam.MODE_COPY_FROM_METADATA);
    System.out.println("Trying to write Image ....");
    imgWriter.write(null, new IIOImage(out_image, null, metadata), iwparam);
    System.out.println("Image written....");
    finally {
    if (imgWriter != null)
    imgWriter.dispose();
    if (imgReader != null)
    imgReader.dispose();
    if (ios != null) {
    ios.flush();
    ios.close();
    }

    user8684061 wrote:
    U are right, SGA is too large for my server.
    I guess oracle set SGA automaticlly while i choose default installion , but ,why SGA would be so big? Is oracle not smart enough ?Default database configuration is going to reserve 40% of physical memory for SGA for an instance, which you as a user can always change. I don't see anything wrong with that to say Oracle is not smart.
    If i don't disincrease SGA, but increase max-shm-memory, would it work?This needs support from the CPU architecture (32 bit or 64 bit) and the kernel as well. Read more about the huge pages.

  • Uploading large files from applet to servlet throws out of memory error

    I have a java applet that needs to upload files from a client machine
    to a web server using a servlet. the problem i am having is that in
    the current scheme, files larger than 17-20MB throw an out of memory
    error. is there any way we can get around this problem? i will post
    the client and server side code for reference.
    Client Side Code:
    import java.io.*;
    import java.net.*;
    // this class is a client that enables transfer of files from client
    // to server. This client connects to a servlet running on the server
    // and transmits the file.
    public class fileTransferClient
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    // this method transfers the prescribed file to the server.
    // if the destination directory is "", it transfers the file to
    "d:\\".
    //11-21-02 Changes : This method now has a new parameter that
    references the item
    //that is being transferred in the import list.
    public static String transferFile(String srcFileName, String
    destFileName,
    String destDir, int itemID)
    if (destDir.equals(""))
    destDir = "E:\\FTP\\incoming\\";
    // get the fully qualified filename and the mere filename.
    String fqfn = srcFileName;
    String fname =
    fqfn.substring(fqfn.lastIndexOf(File.separator)+1);
    try
    //importTable importer = jbInit.getImportTable();
    // create the file to be uploaded and a connection to
    servlet.
    File fileToUpload = new File(fqfn);
    long fileSize = fileToUpload.length();
    // get last mod of this file.
    // The last mod is sent to the servlet as a header.
    long lastMod = fileToUpload.lastModified();
    String strLastMod = String.valueOf(lastMod);
    URL serverURL = new URL(webadminApplet.strServletURL);
    URLConnection serverCon = serverURL.openConnection();
    // a bunch of connection setup related things.
    serverCon.setDoInput(true);
    serverCon.setDoOutput(true);
    // Don't use a cached version of URL connection.
    serverCon.setUseCaches (false);
    serverCon.setDefaultUseCaches (false);
    // set headers and their values.
    serverCon.setRequestProperty("Content-Type",
    "application/octet-stream");
    serverCon.setRequestProperty("Content-Length",
    Long.toString(fileToUpload.length()));
    serverCon.setRequestProperty(FILENAME_HEADER, destDir +
    destFileName);
    serverCon.setRequestProperty(FILELASTMOD_HEADER, strLastMod);
    if (webadminApplet.DEBUG) System.out.println("Connection with
    FTP server established");
    // create file stream and write stream to write file data.
    FileInputStream fis = new FileInputStream(fileToUpload);
    OutputStream os = serverCon.getOutputStream();
    try
    // transfer the file in 4K chunks.
    byte[] buffer = new byte[4096];
    long byteCnt = 0;
    //long percent = 0;
    int newPercent = 0;
    int oldPercent = 0;
    while (true)
    int bytes = fis.read(buffer);
    byteCnt += bytes;
    //11-21-02 :
    //If itemID is greater than -1 this is an import file
    transfer
    //otherwise this is a header graphic file transfer.
    if (itemID > -1)
    newPercent = (int) ((double) byteCnt/ (double)
    fileSize * 100.0);
    int diff = newPercent - oldPercent;
    if (newPercent == 0 || diff >= 20)
    oldPercent = newPercent;
    jbInit.getImportTable().displayFileTransferStatus
    (itemID,
    newPercent);
    if (bytes < 0) break;
    os.write(buffer, 0, bytes);
    os.flush();
    if (webadminApplet.DEBUG) System.out.println("No of bytes
    sent: " + byteCnt);
    finally
    // close related streams.
    os.close();
    fis.close();
    if (webadminApplet.DEBUG) System.out.println("File
    Transmission complete");
    // find out what the servlet has got to say in response.
    BufferedReader reader = new BufferedReader(
    new
    InputStreamReader(serverCon.getInputStream()));
    try
    String line;
    while ((line = reader.readLine()) != null)
    if (webadminApplet.DEBUG) System.out.println(line);
    finally
    // close the reader stream from servlet.
    reader.close();
    } // end of the big try block.
    catch (Exception e)
    System.out.println("Exception during file transfer:\n" + e);
    e.printStackTrace();
    return("FTP failed. See Java Console for Errors.");
    } // end of catch block.
    return("File: " + fname + " successfully transferred.");
    } // end of method transferFile().
    } // end of class fileTransferClient
    Server side code:
    import java.io.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    import java.util.*;
    import java.net.*;
    // This servlet class acts as an FTP server to enable transfer of
    files
    // from client side.
    public class FtpServerServlet extends HttpServlet
    String ftpDir = "D:\\pub\\FTP\\";
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    public void doGet(HttpServletRequest req, HttpServletResponse resp)
    throws ServletException,
    IOException
    doPost(req, resp);
    public void doPost(HttpServletRequest req, HttpServletResponse
    resp)
    throws ServletException,
    IOException
    // ### for now enable overwrite by default.
    boolean overwrite = true;
    // get the fileName for this transmission.
    String fileName = req.getHeader(FILENAME_HEADER);
    // also get the last mod of this file.
    String strLastMod = req.getHeader(FILELASTMOD_HEADER);
    String message = "Filename: " + fileName + " saved
    successfully.";
    int status = HttpServletResponse.SC_OK;
    System.out.println("fileName from client: " + fileName);
    // if filename is not specified, complain.
    if (fileName == null)
    message = "Filename not specified";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // open the file stream for the file about to be transferred.
    File uploadedFile = new File(fileName);
    // check if file already exists - and overwrite if necessary.
    if (uploadedFile.exists())
    if (overwrite)
    // delete the file.
    uploadedFile.delete();
    // ensure the directory is writable - and a new file may be
    created.
    if (!uploadedFile.createNewFile())
    message = "Unable to create file on server. FTP failed.";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // get the necessary streams for file creation.
    FileOutputStream fos = new FileOutputStream(uploadedFile);
    InputStream is = req.getInputStream();
    try
    // create a buffer. 4K!
    byte[] buffer = new byte[4096];
    // read from input stream and write to file stream.
    int byteCnt = 0;
    while (true)
    int bytes = is.read(buffer);
    if (bytes < 0) break;
    byteCnt += bytes;
    // System.out.println(buffer);
    fos.write(buffer, 0, bytes);
    // flush the stream.
    fos.flush();
    } // end of try block.
    finally
    is.close();
    fos.close();
    // set last mod date for this file.
    uploadedFile.setLastModified((new
    Long(strLastMod)).longValue());
    } // end of finally block.
    } // end - the new file may be created on server.
    } // end - we have a valid filename.
    // set response headers.
    resp.setContentType("text/plain");
    resp.setStatus(status);
    if (status != HttpServletResponse.SC_OK)
    getServletContext().log("ERROR: " + message);
    // get output stream.
    PrintWriter out = resp.getWriter();
    out.println(message);
    } // end of doPost().
    } // end of class FtpServerServlet

    OK - the problem you describe is definitely what's giving you grief.
    The workaround is to use a socket connection and send your own request headers, with the content length filled in. You may have to multi-part mime encode the stream on its way out as well (I'm not about that...).
    You can use the following:
    http://porsche.cis.udel.edu:8080/cis479/lectures/slides-04/slide-02.html
    on your server to get a feel for the format that the request headers need to take.
    - Kevin
    I get the out of Memory Error on the client side. I
    was told that this might be a bug in the URLConnection
    class implementation that basically it wont know the
    content length until all the data has been written to
    the output stream, so it uses an in memory buffer to
    store the data which basically causes memory issues..
    do you think there might be a workaround of any kind..
    or maybe a way that the buffer might be flushed after
    a certain size of file has been uploaded.. ?? do you
    have any ideas?

  • Out of memory error when opening an .xlsx file in numbers v1.6.2 on ipad 2

    The file sizes are too small (38KB) for this to be a true "out of memory" issue. Any suggestions? I didn't have any problems before updating numbers to the current version.
    I haven't downloaded the new iOS 6 operating system. Is that the problem?

    Ended up being a problem with the Excel file we were using. Recreated the file and all was ok.

  • Advice needed: The way to solve out of memory problem (or the way to work with big csv files)

    Hello:)
    I'm in trouble: I have a big csv file (over 5gb of web-analytics data) and my 64 bit excel (and 6gb ram)
    I cant load file to data model because of it's size. There is an error "out of memory" in power query. 
    This is the first time when I encountered such a problem.
    What options do I have to work with such a file? To increase memory in my computer? Would it solve the problem? How much do I need to work with 6gb csv? 
    Or may be I can upload my data somewhere to azure and work with it there? 
    So the problem - is there any way to deal with big files using power query? Or I need to become a developer and learn sql or other languages? 
    Thanks in advance.
    Max

    Hi Miguel!
    Thanks for your answer. 
    I've tried to load this file on virtual pc from azure cloud with this config:
    I have increased memory limit in power query settings:
    And still, the proble is the same:
    What I do wrong? 

  • Out of memory when big files are being sent to N80...

    i downloaded a few big mp3 files (15-20mb) onto my 128 card i got with the phone, as soon as my new 2gb phone came, i didnt have a reader for my pc and my wife said she wanted these songs on her S.E.810i. i sent them over to her using bluetooth (with no probs whatsoever) in the hope i could send them back to my phone with my new 2gb card in. when i try to send them back from the 810i with bluetooth, i get out of memory message but i have next to nothing on phone mem - no vid/images/sound etc. can anyone tell me if i would be right in saying that when i download from my home network the files are temp stored on mem card until you confirm where u want to save whereas bluetooth tries to store on phone mem temporarily first? hope u got all that! if this is the case, is there anyway i can change this? if anyone is still with me, and knows what i am talking about, please enlighten me!! its driving me mad!
    thanks

    1. All messages will be stored/moved to the memory card.
    2. I doubt it. The phone will still save files received via bluetooth in a messaging folder. If this folder is stored on the phone memory there will never be enough space to receive really big files. dmz

  • Out of memory error to view AVI files

    Hi everyone,
    I've imported 6 AVI files from the same source into the browser. All work fine except for 2 of them. When I click on those files in the browser to preview them in the viewer, I receive 2 messages: the first one is 'general error', the second 'out of memory'. What strikes me is that in the browser, the duration marks 0 although the files are respectively 4MB and 1,6MB. I've compared them to the 4 others and they are similar. Any idea would be welcome. Many thanks in advance.

    Welcome to the forum!
    AVI files don't work with FCP. FCP is based in Quicktime. You will need to convert all you AVI files to an quicktime editing codec like ProRes or DV-NTSC You can find a list of editing codecs in FCP's Easy Setup. You can use Compressor to convert the files.

Maybe you are looking for