Working with  large Avi (DivX) File - PE 2

Hi there
I'm trying to use PE2 to edit & generate scene markers in a 1.24GB avi file.
The problem I'm having is that frames are very slow to update on the timeline making it very difficult to work with, probably due to the size of the file file & my low system specs.
Any thoughts on how to best tackle this? Would it be best to split the avi file into more manageable pieces? If so, how would I do that?
Any suggestions would be much appreciated.
Thanks & regards
Stephen

Export the DivX clip as a DV-AVI file using File>Export>Movie. Then replace the DivX clip on the timeline with the DV-AVI clip.

Similar Messages

  • ADT Not Working With Large Amount Of Files

    Hi,
    I posted a similar issue in the Flash Builder forum; it seems these tools are not meant for anything too big. Here is my problem; I have an application that runs on the iPad just fine, however I cannot seem to be able to package any of the resource files it uses, such as swf, xml, jpg, and flv. There are a lot of these files, maybe around a 1000. This app won't be on the AppStore and just used internally at my company. I tried to package it using Flash Builder, but it couldn't handle it, and so now I have tried the adt command line tool. The problem now is that it is running out of memory:
    Exception in thread "main" java.lang.OutOfMemoryError
            at java.util.zip.Deflater.init(Native Method)
            at java.util.zip.Deflater.<init>(Unknown Source)
            at com.adobe.ucf.UCFOutputStream.addFile(UCFOutputStream.java:428)
            at com.adobe.air.ipa.IPAOutputStream.addFile(IPAOutputStream.java:338)
            at com.adobe.ucf.UCFOutputStream.addFile(UCFOutputStream.java:273)
            at com.adobe.ucf.UCFOutputStream.addFile(UCFOutputStream.java:247)
            at com.adobe.air.ADTOutputStream.addFile(ADTOutputStream.java:367)
            at com.adobe.air.ipa.IPAOutputStream.addFile(IPAOutputStream.java:161)
            at com.adobe.air.ApplicationPackager.createPackage(ApplicationPackager.java:67)
            at com.adobe.air.ipa.IPAPackager.createPackage(IPAPackager.java:163)
            at com.adobe.air.ADT.parseArgsAndGo(ADT.java:504)
            at com.adobe.air.ADT.run(ADT.java:361)
            at com.adobe.air.ADT.main(ADT.java:411)
    I tried increasing the java memory heap size to 1400m, but that did not help. What is disturbing though is why this day and age, the adt tool is written solely for 32 bit java runtimes, when you can no longer buy a 32bit computer. Almost everything is 64bit now. Being able to utilize a 64bit java runtime would eliminate this.
    Has anyone else had to deal with packaging a large number of files within an iPad app?
    thanks

    Hi,
    Yes the bug number is 2896433, logged against Abobe Air. I also logged one against Flash Builder for the same issue basically, FB-31239. In the tests I have done here, both issues are the result of including a large number of files in the package. By reducing the number of files included, these bugs do not appear.
    I would liketo try out the new SDK, but it won't be available until a few more weeks, correct?
    The main reason I want to include all these files is because the iPads will not always have a wireless connection.
    thx

  • Encore CS5 constantly locks up/freezes working with large H.264 file

    This is driving me mad, I'm trying to author my first Blu Ray with this program, a single layer output with H.264/AVC video. I've encoded the video elsewhere and I have a 21GB .264 file. It's a BR legal file but as soon as I import it into Encore CS5 the fun and games begin. Firstly it takes a good five minutes to import and Encore appears to have crashed. Finally though it does import and looking at 'Blu Ray transcode' it says 'Don't transcode', so it's a legal file. However if I then try to do anything with the file in Encore, like add it to a timeline the whole program locks up again. 10 minutes later it's added as a timeline but trying to navigate that timeline to add chapters is impossible, the program locks up (Not responding) if I try and move the slider or pretty much do anything. It means making a BR with this file is pretty much impossible.
    My PC spec is as follows:
    Core i5 2500k 3.3Ghz
    8GB DDR3 1600mhz RAM
    120GB Crucial SSD boot drive with Encore installed on
    500GB Hard Disk & 2TB Hard Disk
    Windows 7 64-Bit Home Premium
    Is there anything I can do to actually make Encore usable with this large file and finally finish authoring a disc? I've tried loading the file from all three drives and it's the same. The media cache is currently on my 2TB drive because I want to save space on my SSD and minimize writes to it.

    Taking 5 minutes to import is normal.  Encore has to index the entire file so it knows where the keyframes are.  Don't try to do anything while that's happening.
    Once in, however, Encore should respond normally.  At this point, I'm drawing a blank as to why it isn't.

  • Working with large Artboards/Files in Illustrator

    Hello all!
    I'm currently designing a full size film poster for a client. The dimensions of the poster are 27" x 40" (industry standard film poster).
    I am a little uncertain in working with large files in Illustrator, so I have several problems that have come up in several design projects using similar large formats.
    The file size is MASSIVE. This poster uses several large, high-res images that I've embedded. I didn't want them to pixelate, so I made sure they were high quality. After embedding all these images, along with the vector graphics, the entire .ai file is 500MB. How can I reduce this file size? Can I do something with the images to make the .ai file smaller?
    I made my artboard 27" x 40" - the final size of the poster. Is this standard practice? Or when designing for a large print format, are you supposed to make a smaller, more manageable artboard size, and then scale up after to avoid these massive file sizes?
    I need to upload my support files for the project, including .ai and .eps - so it won't work if they're 500MB. This would be good info to understand for all projects I think.
    Any help with this would be appreciated. I can't seem to find any coherent information pertaining to this problem that seems to address my particular issues. Thank you very much!
    Asher

    Hi Asher,
    It's probably those high-res images you've embedded. Firstly, be sure your images are only as large as you need them Secondly, a solution would be to use linked images while you're working instead of embedding them into the file.
    Here is a link to a forum with a lot of great discussion about this issue, to get you started: http://www.cartotalk.com/lofiversion/index.php?t126.html
    And another: http://www.graphicdesignforum.com/forum/archive/index.php/t-1907.html
    Here is a great list of tips that someone in the above forum gave:
    -Properly scale files.  Do not take a 6x6' file then use the scaling tool to make it a 2x2'  Instead scale it in photoshop to 2x2 and reimport it.  Make a rule like over 20%, bring it back in to photoshop for rescaling.
    -Check resolutions. 600dpi is not going to be necessary for such and such printer.
    -Delete unused art.  Sloppy artists may leave old unused images under another image.  The old one is not being used but it still takes up space, therefore unecessarily inflating your file.
    -Choose to link instead of embedd.  This is your choice.  Either way you still have to send a large file but many times linking is less total MB then embedding.  Also embedding works well with duplicated images. That way multiple uses link to one original, whereas embedding would make copies.
    -When you are done, using compression software like ZIP or SIT (stuffit)
    http://www.maczipit.com/
    Compression can reduce file sizes alot, depending on the files.
    This business deals with alot of large files.  Generally people use FTP's to send large files, or plain old CD.  Another option is using segmented compression.  Something like winRAR/macRAR or dropsegment (peice of stuffit deluxe) compresses files, then breaks it up into smaller manageble pieces.   This way you can break up a 50mb file into say 10x 5mb pieces and send them 5mb at a time. 
    http://www.rarlab.com/download.htm
    *make sure your client knows how to uncompress those files.  You may want to link them to the site to download the software."
    Good luck!

  • Speed up Illustrator CC when working with large vector files

    Raster (mainly) files up to 350 Mb. run fast in Illustrator CC, while vector files of 10 Mb. are a pain in the *blieb* (eg. zooming & panning). When reading the file it seems to freeze around 95 % for a few minutes. Memory usage goes up to 6 Gb. Processor usage 30 - 50 %.
    Are there ways to speed things up while working with large vector files in Illustrator CC?
    System:
    64 bit Windows 7 enterprise
    Memory: 16 Gb
    Processor: Intel Xeon 3,7 GHz (8 threads)
    Graphics: nVidia Geforce K4000

    Files with large amounts vector points will put a strain on the fastest of computers. But any type of speed increase we can get you can save you lots of time.
    Delete any unwanted stray points using  Select >> Object >> stray points
    Optimize performance | Windows
    Did you draw this yourself, is the file as clean as can be? Are there any repeated paths underneath your art which do not need to be there from live tracing or stock art sites?
    Check the control panel >> programs and features and sort by installed recently and uninstall anything suspicious.
    Sorry there will be no short or single answer to this, as per the previous poster using layers effectively, and working in outline mode when possible might the best you can do.

  • Photoshop CS6 keeps freezing when I work with large files

    I've had problems with Photoshop CS6 freezing on me and giving me RAM and Scratch Disk alerts/warnings ever since I upgraded to Windows 8.  This usually only happens when I work with large files, however once I work with a large file, I can't seem to work with any file at all that day.  Today however I have received my first error in which Photoshop says that it has stopped working.  I thought that if I post this event info about the error, it might be of some help to someone to try to help me.  The log info is as follows:
    General info
    Faulting application name: Photoshop.exe, version: 13.1.2.0, time stamp: 0x50e86403
    Faulting module name: KERNELBASE.dll, version: 6.2.9200.16451, time stamp: 0x50988950
    Exception code: 0xe06d7363
    Fault offset: 0x00014b32
    Faulting process id: 0x1834
    Faulting application start time: 0x01ce6664ee6acc59
    Faulting application path: C:\Program Files (x86)\Adobe\Adobe Photoshop CS6\Photoshop.exe
    Faulting module path: C:\Windows\SYSTEM32\KERNELBASE.dll
    Report Id: 2e5de768-d259-11e2-be86-742f68828cd0
    Faulting package full name:
    Faulting package-relative application ID:
    I really hope to hear from someone soon, my job requires me to work with Photoshop every day and I run into errors and bugs almost constantly and all of the help I've received so far from people in my office doesn't seem to make much difference at all.  I'll be checking in regularly, so if you need any further details or need me to elaborate on anything, I should be able to get back to you fairly quickly.
    Thank you.

    Here you go Conroy.  These are probably a mess after various attempts at getting help.

  • Using  Adobe Premiere CS3 (ver. 3.2) with Microsoft AVI DV files (4:3, 48Khz) on timeline now ask render (red line) and before didnt.Proyect is DV NTSC standard 48Khz. What it could be?

    Hi,
    Normally during months the edition was done with Microsoft AVI DV Files (4:3 NTSC 48 Khz). No effects (just cuts). Proyect settings as DV-NTSC Standard 48 KHz
    Since today when the work is done in a new or old proyect (DV-NTSC Standard 48 KHz) and  if the same file (Microsoft AVI DV) is loaded to timeline now is showing a red line over the video on timeline (so requires render). There no effects. This never happens before. it is very weird.
    The Premiere is CS3 ver 3.2.0.
    This is  happening to 2 computers with the same version.So far I know there is no changes on the systems
    I was thinking to reisntall 3.0 and then test and the update to 3.2...
    Any idea?

    Hi,
    It worked once in one of the computers. (AVi DV didnt show render red over line). BUt afterwards the problem got back. The only solution so far was to reinstall premiere, update to 3.2.0 in both pc.
    It worked for a while then problems was again. Reinstalled for second time, so far 2 days still ok.. It is very weird. I unchecked updates from cs3 and windowsfor just in case has any relation.
    Thanks for your tip anyway.

  • Is anyone working with large datasets ( 200M) in LabVIEW?

    I am working with external Bioinformatics databasesa and find the datasets to be quite large (2 files easily come out at 50M or more). Is anyone working with large datasets like these? What is your experience with performance?

    Colby, it all depends on how much memory you have in your system. You could be okay doing all that with 1GB of memory, but you still have to take care to not make copies of your data in your program. That said, I would not be surprised if your code could be written so that it would work on a machine with much less ram by using efficient algorithms. I am not a statistician, but I know that the averages & standard deviations can be calculated using a few bytes (even on arbitrary length data sets). Can't the ANOVA be performed using the standard deviations and means (and other information like the degrees of freedom, etc.)? Potentially, you could calculate all the various bits that are necessary and do the F-test with that information, and not need to ever have the entire data set in memory at one time. The tricky part for your application may be getting the desired data at the necessary times from all those different sources. I am usually working with files on disk where I grab x samples at a time, perform the statistics, dump the samples and get the next set, repeat as necessary. I can calculate the average of an arbitrary length data set easily by only loading one sample at a time from disk (it's still more efficient to work in small batches because the disk I/O overhead builds up).
    Let me use the calculation of the mean as an example (hopefully the notation makes sense): see the jpg. What this means in plain english is that the mean can be calculated solely as a function of the current data point, the previous mean, and the sample number. For instance, given the data set [1 2 3 4 5], sum it, and divide by 5, you get 3. Or take it a point at a time: the average of [1]=1, [2+1*1]/2=1.5, [3+1.5*2]/3=2, [4+2*3]/4=2.5, [5+2.5*4]/5=3. This second method required far more multiplications and divisions, but it only ever required remembering the previous mean and the sample number, in addition to the new data point. Using this technique, I can find the average of gigs of data without ever needing more than three doubles and an int32 in memory. A similar derivation can be done for the variance, but it's easier to look it up (I can provide it if you have trouble finding it). Also, I think this funtionality is built into the LabVIEW pt by pt statistics functions.
    I think you can probably get the data you need from those db's through some carefully crafted queries, but it's hard to say more without knowing a lot more about your application.
    Hope this helps!
    Chris
    Attachments:
    Mean Derivation.JPG ‏20 KB

  • Working w/ large scale AI files

    What is the best way to work with large scale (wall mural) files that contain many gradients, shadows. etc. Any manipulations take a considerable time to redraw. Files are customer created, and we are then manipulating from there in-house. We have some fairly robust machines ( mac pro towers- 14GB RAM, RAID scratch disk).. I would guess there is a way to work in a mode that does not render effects, allowing for faster manipulation? Any help would be great- and considerably help reduce our time.
    First post- sorry if did something wrong- question / title wise/
    THX!

    In a perfect world, the customers would be creating their Illustrator artwork with the size & scale of the final image in mind. It's very difficult to get customers (who often have no formal graphic design training and are self taught at learning Illustrator & Photoshop) to think about the basic rule of designing for the output device -a graphics 101 sort of thing.
    Something like a large wall mural, especially one that is reproduced outdoors, can get by just fine with Illustrator artwork using raster effects settings of 72ppi or even less than that. Lots of users will have a 300ppi setting applied. 300ppi is fine for a printed page being viewed up close. 300ppi is sheer overkill for large format use.
    Mind you, Adobe Illustrator has a 227" X 227" maximum artboard size, so anything bigger than that will have to be designed at a reduced scale with appropriate raster effects settings applied. For example, I'll design a 14' X 48' billboard at a 1" = 1' scale. A 300ppi raster effects setting is OK in that regard. When blown up in a large format printing RIP the raster based effects have an effective 25ppi resolution, but that's good enough for a huge panel being viewed by speeding vehicles on city streets or busy highways.
    Outside of that, the render speed of vector-based artwork will depend on the complexity of artwork. One "gotcha" I tend to watch are objects with large numbers of anchor points (like anything above 4000 points). At a certain point the RIP might process only part of the object or completely disregard it.

  • How does the sync functionality work with large libraries on small devices?

    How does the sync functionality work with large libraries?
    Say I sync 100gb of photos with the new Photos app and turn on sync on a 16gb iphone. Will it fill the device up to 16gb? Can I tell it to limit to xgb so I leave room for music and apps? How does this work? Will it slow down my phone if its trying to sync 100gb across smaller devices?

    "Will the Apple TV now read directly from the Time Capsule?" ATV does not 'read' from the TC. It connects to the Mac and itunes library associated with the Mac. Of course the itunes program can 'point' to a library on the TC. You'll still need to have itunes open and the Mac powered on and not sleeping with the TC mounted to the Desktop to use ATV properly. So bottom line the use of TC just adds one more step to view files on the ATV.

  • Lag with large brushes and files

    I typically work with large files around 10000 pixels, 300 dpi.  But I'm getting a lot of lag when I use brushes larger than 1000 px as well as a long wait to open and save the files.
    Is this a RAM issue?
    Here are my specs:
    OS Name             Microsoft Windows 7 Home Premium
    Version 6.1.7601 Service Pack 1 Build 7601
    System Manufacturer      Alienware
    System Model    M18xR1
    System Type       x64-based PC
    Processor            Intel(R) Core(TM) i7-2630QM CPU @ 2.00GHz, 2001 Mhz, 4 Core(s), 8 Logical Processor(s)
    Installed Physical Memory (RAM)               8.00 GB
    Total Physical Memory    7.98 GB
    Available Physical Memory            4.43 GB
    Total Virtual Memory      16.0 GB
    Available Virtual Memory              11.2 GB
    Page File Space  7.98 GB
    And two GPUs: NIVIDIA Geforce GTX 460m
    I'm upgrading my RAM and GPU soon anyway but wanted to check and make sure it wasnt some other underlying problem?

    I'm having some trouble deciphering what is going on here. I'm useing Paul Tuersley's response as a starting point. Here is my modification not that it matters since it isn't working at the moment. I don't actually know what some of the script is doing. Right now "theFiles" is returning all of the files that are not real files, folders and .DS_Store when I want it to be an array of all of the files in the folder.
    Could you help me break it down?
    // Log the selected footage
    var selectedFiles = app.project.selection;
    // Choose the folder to scan
    var theFolder = Folder.selectDialog("Please select folder.");
    if(theFolder != null){
        var resultArray = new Array();
        folderFiles = checkForFiles(theFolder, "");
        $.writeln(resultArray.toString().replace(new RegExp(",", "g"), ""));
        // Compare selected files to files in the folder
        for (var i = 0; i < selectedFiles.length; i++){
            compareFiles (selectedFiles[i], folderFiles);
    function compareFiles(file, folderFiles) {
              for (var i = 0; i < folderFiles.length; i++){
                             if (folderFiles[i].name==file.name) {
                                            alert(folderFiles[i].name + " and " + file.name + " match!");
    function checkForFiles(folderItem, tabString) {
        var theFiles = folderItem.getFiles();
        for(var c = 0; c < theFiles.length; c++){
            resultArray.push(tabString + theFiles[c].name + "\r");
            if (theFiles[c] instanceof Folder) {
                checkForFiles(theFiles[c], tabString + "\t");
        return(theFiles);

  • Cant work with 32 bit image files?

    Hi there, I am unable to work with 32 bit image files on our cs6 ? I am fallowing a tutorial and a person creates a new file with 32 bit channel selected from dropdown menu, and starts painting on it he is using CS5. When I create a 32 bit file and select the brush tool, it comes on with circle and cross line and error comes up saying that the brush tool is not supported with 32 bit image files, in fact, I cant use any tools on the 32 bit images, I can't even create a new layer on the file. There is one thing I can do, which is a great feature, I can save it! But it's not much of a use now is it? If the image is plane white sheet. Is this a special feature that you need to purchase seperetly? Could you please advise. Regards

    Hello again,
    so resetting the preference file makes no difference to the problem.
    I checked 6 of our macs. 3 of the macs can create the 32 bit images and allow paint and other tools with no problem. Jet the other 3 macs have the same problem as I have. I checked my 2008 macbook which has CS5, and it can also edit 32 bit files. Checked my PC yesterdey which has CS6 installed, can't work with 32 bit image files.
    I find this very odd, some work and some don't. Any suggestions would be great. Please also find a screenshot of the mac with the issue.
    Kind regards,

  • Work with ADF Library jar file

    Hi
    My English isn't very good
    I use jdeveloper11.1.1.3.0
    I want to work with ADF Library jar file.
    I created two generic application. In each application I have a View Obj(one for employees table and one for department table ) and added them to App Module,
    then I made a ADF Library jar file for each application: adflib_emp, adflib_dep
    Then I created two separated application and in each application I created a BTF .
    In emp_TF1 I added adflib_emp jar file to application and in dep_TF2 I added adflib_dep jar file to application.
    These two applications work well. Then I created ADF Library jar file for each of these applications named: adflib_TF_1 and adflib_TF_2
    Finally I designed a main page to use these two jar file(adflib_TF_1 , adflib_TF_2). I added thses two jar file to application and then I dragged and dropped each TF to main page as a region:
    <af:panelSplitter id="ps1" splitterPosition="500">
    <f:facet name="first">
    <af:region value="#{bindings.emp_TF1.regionModel}" id="r1"/>
    </f:facet>
    <f:facet name="second">
    <af:region value="#{bindings.dep_TF2.regionModel}" id="r2"/>
    </f:facet>
    </af:panelSplitter>
    When I run main application both TF show employees table, but one TF should shows department table and the other one should shows employees table.
    I read oracle document and created my application on base of it. But I don't know why this happen and how to solve it?
    Habib
    Edited by: Habib Eslami on Apr 7, 2013 8:46 PM

    Hi,
    most likely the names of the application modules are not different but the same, which means you have a naming problem. Application modules (if they should be used as is) must have a unique name
    Frank

  • Working with Large Numbers

    Hi there,
    I am currently doing a school assignment and not looking for answers but just a little guidance.
    I am working with large numbers and the modulo operator.
    I might have some numbers such as :
    int n = 221;
    int e = 5;
    int d = 77;
    int message = 84;
    int en = (int) (Math.pow(message, e) % n);
    int dn = (int) (Math.pow(en, d) % n);Would there be a better way to do this kind of calculation. The dn value should come out the same as message. But I always get something different and I think I might be losing something in the fact that an int can only hold smaller values.

    EJP wrote:
    It might make sense in some contexts to have a positive and negative infinity.
    Yes, perhaps that's a better name. Guess I was harking back to old COBOL days :-).(*)
    But the reason these things exist in FP is because the hardware can actually deliver them. That rationale doesn't apply to BIgInteger.Actually, it does. All I'm talking about is a value that compares higher or lower than any other. That could be done either by a special internal sign value (my slight preference) or by simply adding code to compareTo(), equals() and hashCode() methods that takes the two constants into account (as they already do with ZERO and ONE).
    Don't worry, I'm not holding my breath; but I have come across a few situations in which values like that would have been useful.
    Winston
    Edited by: YoungWinston on Mar 22, 2011 9:07 AM
    (*) Actually, '&plusmn;infinity' tends to suggest a valid arithmetic value, and I wasn't thinking of changing existing BigInteger/BigDecimal maths (except perhaps to throw an exception if either value is involved).

  • Working with Large List in sharepoint 2010

    Hi All
    I have a list with almost 10k records in my sharepoint list and based on some business requirement i am binding (almost 6k records) the data to asp.net grid view and this will visible on the home page of the portal where most of the users will access. Can
    someone please guide the best method to reduce the performance inorder the program to hit the SP list everytime the page loads...
    Thanks & Regards
    Rakesh Kumar

    Hi,
    If you are Working with large data retrieval from the content database (SharePoint list), the points below for your reference:
    1. Limit the number of returned items.
    SPQuery query = new SPQuery();
    query.RowLimit =6000; // we want to retrieve2000 items
    query.ListItemCollectionPosition = prevItems.ListItemCollectionPosition; // starting at a previous position
    SPListItemCollection items = SPContext.Current.List.GetItems(query);
    2. Limit the number of returned columns.
    SPQuery query = new SPQuery();
    query.ViewFields = "";
    3. Query specific items using CAML (Collaborative Markup Language).
    SPQuery query = new SPQuery();
    query.Query = "15";
    4.Use ContentIterator class
    https://spcounselor-public.sharepoint.com/Blog/Post/2/Querying-a--big-list--with-ContentIterator-and-multiple-filters
    5. Create a Stored Procedure in Database to get the special data, create a web service to get the data, when create a web part to show the data in home page.
    Best Regards
    Dennis Guo
    TechNet Community Support

Maybe you are looking for

  • Why is the dot in find my iPhone purple?

    Why is the dot in find my IPhone purple?

  • Ios8 on my iphone5 have problem!!!!

    when are appointment reminders on the bottom of the screen. It will remain that way Can not push to make it go away. To make the access to be lost.

  • Text for Professional Printing

    I am preparing a newsletter that will be sent out for professional printing.  Will Adobe's "Open Face" fonts work or do I need to download post script fonts from somewhere. I appreciate your help.

  • Downloading Patches using Application Manager

    Hi, I am currently preparing for oracle apps dba exam and doing self studdy fot his. I have installed oracle apps on my personal computer on linux which is directly connected to internet with a home internet connection.(without any proxy). I submitte

  • Workflow Launcher Configuration Exclude List

    Hi, I have a situation where I need to have the main DAM Update Asset workflow ignore two specific directories inside of /content/dam and their contents.  I've tried entring their paths in various formats in the Exclude List field, but nothing seems