File Data Rate?

Dear PrP CC Users,
I work with several file formats ranging from ProRes (422HQ, 4444), to H264 to R3D. Within my project browser it never displays the metadata File Data Rate, it is always blank. This for me is a useful piece of information when dealing with footage and knowing what drive speeds I require; having migrated to PrP CC from FCP7, this is something I sorely miss.
Is there some way to actively see the File Data Rate? How come PrP CC fails to calculate the File Data Rate?
Thank you.

Is there some way to actively see the File Data Rate?
Not yet.
http://forums.adobe.com/thread/1369233?tstart=60

Similar Messages

  • How do you fix error message "data rate for this file is too high for DVD.  You must replace this file with one of a lower data rate".

    When trying to burn a DVD it will go through the encoding step and at 98% we see the message 'data rate for this file is too high for DVD.  You must replace this file with one of a lower data rate".  We need help to correct this so we can complete burning to DVD. 

    What did you export from Premiere?
    Did you use the MPEG2-DVD preset... and did you make any changes to the preset?
    CS5-thru-CC PPro/Encore tutorial list http://forums.adobe.com/thread/1448923 may help

  • PGC has an error--data rate of this file is too high for DVD

    Getting one of those seemingly elusive PGC errors, though mine seems to be different from many of the ones listed here. Mine is telling me that the data rate of my file is too high for DVD. Only problem is, the file it's telling me has a datarate that is too high, is a slideshow which Encore has built using imported jpg files. I got the message, tried going into the slideshow and deleting the photo at the particular spot in the timeline where it said it had the problem, now getting the same message again with a different timecode spot in the same slideshow. The pictures are fairly big, but I assumed that Encore would automatically resize them to fit an NTSC DVD timeline. Do I need to open all the pictures in Photoshop and scale them down to 720x480 before I begin with the slideshows?

    With those efforts, regarding the RAM, it would *seem* that physical memory was not the problem.
    I'd look to how Windows is managing both the RAM addresses and also its Virtual Memory. To the former, I've seen programs/Processes that lock certain memory addresses upon launch (may be in startup), and do not report this to Windows accurately. Along those lines, you might want to use Task Manager to see what Processes are running from startup on your machine. I'll bet that you've got some that are not necessary, even if IT is doing a good job with the system setup. One can use MSCONFIG to do a trial of the system, without some of these.
    I also use a little program, EndItAll2 for eliminating all non-necessary programs and Processes, when doing editing. It's freeware, has a tiny footprint and usually does a perfect job of surveying your running programs and Processes, to shut them down. You can also modify its list, incase it wants to shut down something that IS necessary. I always Exit from my AV, spyware, popup-blocker, etc., as these progams will lie to EndItAll2 and say that they ARE necessary, as part of their job. Just close 'em out in the Tasktray, then run EndItAll2. Obviously, you'll need to do this with the approval of IT, but NLE machines need all available resources.
    Now, to the Virtual Memory. It is possible that Windows is not doing a good job of managing a dynamic Page File. Usually, it does, but many find there is greater stability with a fixed size at about 1.5 to 2.5x the physical RAM. I use the upper end with great results. A static Page File also makes defragmenting the HDD a bit easier too. I also have my Page File split over two physical HDD's. Some find locating to, say D:\ works best. For whatever reason, my XP-Pro SP3 demanded that I have it on C:\, or split between C:\ and D:\. Same OS on my 3 HDD laptop was cool having it on D:\ only. Go figure.
    These are just some thoughts.
    Glad that you got part of it solved and good luck with the next part. Since this seems to affect both PrPro and En, sounds system related.
    Hunt
    PS some IT techs love to add all sorts of monitors to the computers, especially if networkded. These are not usually bad, but are usually out of the mainstream, in that most users will never have most of these. You might want to ask about any monitors. Also, are you the only person with an NLE computer under the IT department? In major business offices, this often happens. Most IT folk do not have much, if any, experience with graphics, or NLE workstations. They spend their days servicing database, word processing and spreadsheet boxes.

  • Reducing data rate written to a LVM file

    I have a LabVIEW 7.1 question regarding data logging. I have designed a system that reads the pressure of 6 4-20Ma pressure transducers. This data is then filtered to remove any noise, and displayed to the user on the computer's screen.
    I am currently reading the data at 1000HZ using the DAQ Assistant, which seems to give a clear picture of the pressure changes. My supervisor has asked me to allow the system to log the results for what may be 3 or 4 days. As you ca imagine, 48 hours at 1000 samples a second is a lot of data! I need some method to reduce the data rate written to the file. Ideally, I would like to take one second of data, find the highest reading, and then file a single entry and a timestamp interval of 1 second. So, for each hour, there would be 3600 rows with 7 columns (1 time stamp and 6 sensors). It would be nice if the 1 second average could be adjustable, so that if long term logging is to be done (a week, for example), the interval can be changed to 5 seconds or whatever.
    I have tried reducing the data read rate as much as possible, but it seems that 100hz is the smallest that is possible. That is still too high... as the hard drive would fill up in no time.
    Any help in this area would be appreciated. I have tried a few things, and all to no success. I have included a copy of the code for anyone to review.
    Hardware:
    - 1 X P3 Laptop running LabVIEW 7.1 data acquisition software
    - 1 X NI DAQCard-6062E interface PCard
    - 1 X NI SC-2345 data acquisition hardware with
    - 3 X SCC-CI20 current input module
    - 6 X Omega PX 0-600PSIA 4-20mA pressure transducer
    thanks so much!
    Andrew
    Attachments:
    Testing.vi ‏812 KB

    You would have to talk with your supervisor first to determine what he intends to do with the log data and what degree of resolution he actually needs first. You probably want the 1000Hz Sampling rate so you can do decent filtering on the signal (hopefully your pressures aren't actually changing at a faster rate than that). I'm assuming you are returning a single result for those 1000 readings for each sensor. Specify some file logging duration of n-seconds (or minutes or whatever). Between file writes, pop each filtered measurement into an array (either 1 array of 6 dimensions or 6 1-dimension arrays). After n-seconds have passed, determine the min, avg, and max values from each sensor (array) and log those with your timestamp. So if you set your log timer for 1 minute, you would log a single min, max, and average reading of 60 readings for 6 sensors (this would only require 1 row with say a timestamp and 3x6 (18) columns for each sensor's min, max, average data). After a 24hour period you will have logged 1440 rows of data. In 3 days that would only be 4320 rows of data. All that is as easy as using a timer and a case structure around your logging function which would be triggered every n-seconds. Everything else you're doing would be the same. None of this really has much to do with labview as it is more of a logical explanation of how and when to acquire and log your data. What method are you using for storing your data? CSV, BIN, etc.. If you also want to display the data in a chart, I would recommend charting the same data you're logging, otherwise your chart will probably crash your system at 1000samps/second for 60 hours.. Once again, it depends on how your supervisor is analyzing the logged data. Make your log duration programmable and change it until he is happy. If he's(she's?) using Excel, your maximum log timer would be 9 seconds (Equates to ~6.67 Logs per Minute, ~400 Logs per Hour, ~9600 Logs Per Day, for a total of 28,800 Logs(rows) for 3 days -- Excel is limited to 32000 rows).

  • The data rate of this file is too high encore

    Hi,
    Im new to all this, and I need a real quick solution as this needs to be done for the next few days, please help.
    I get a "the data rate of this file is too high encore"the data rate of this file is too high for DVD" message when I try to brun to disk.`
    I assume this means the bitrate I used when exporting from premiere? I only have one file to go on the DVD, what bitrate should I use?
    Thanks

    What did you export from Premiere?
    Did you use the MPEG2-DVD preset... and did you make any changes to the preset?
    CS5-thru-CC PPro/Encore tutorial list http://forums.adobe.com/thread/1448923 may help

  • Gettin info about Flash file's data rate (and more) on Mac

    In Quicktime one can get all sorts of info like frame rate,
    data rate, codecs and more, but flah players don't do this. Right?
    I see there is special software for this for a PC, but what about
    Mac? Is there any way to get this info for files I didn't make
    myself.

    I am referring to video files that I am making out of .mov
    files. I am trying to find best cmpression options, but to do that
    it would be nice to study others' files.

  • Files failing to adhere to set data rate on export from Quicktime

    I am hoping somebody can shed some light on this.
    I open an uncompressed .mov file (output from FinalCut) with Quicktime with the intention of compressing it for the web. After File: export, in the options dialog box, in the field where you can restrict the data rate, I put a number, say 488 kbs/sec. But every time I do this, the finished compressed file has a data rate wildly higher, like 2000 kbs/sec, and so is too large for my purposes.
    I feel like I used to do this all the time- I have plenty of .mov files with data rates around 500 kbs/sec that I posted on my website, so this had to have worked at some time in the past.
    Anybody? Thanks

    Just a bump-

  • Data rate of AAC file converted with iTunes 7.5.0.20

    I have just noticed data rate of files converted to AAC with iTunes revision 7.5.0.20 deviate slightly from preset converting data rate in random manner. For example browser shows files with preset 192kBit/s data rate as 193 or 195kBit/s data rate, ones with 320kBit/s preset as 321, 322, 325kBit/s etc.
    Is this a encoder bug or a music content dependent data rate adjustment?
    Anyone a suggestion?
    Thanks and best regards

    Refer to one of the many threads already posted about this:
    http://discussions.apple.com/thread.jspa?threadID=1238169

  • How to achieve the maximum file transfer rate from PXI to local host?

    I will have to copy a huge amount of data (as big as 50 GB) from a PXI-8106 unit on site to a laptop. Currently it would take several hours. It is critical to max out the data transfer rate.
    The standard method we've been using is just drag and drop using Windows Explorer via FTP. I tried to use FTP VIs with Filezilla Server as well but the transfer rate is only like 1.5 MB/s. Is this normal? Once I saw it was 3.0 MB/s but for some reason, that isn't happening now.
    For another option, I tried to transfer data to a USB flash drive but it's even slower. As far as I know, USB 2.0 transfer rate is supposed to be around 60 MB/s but why would it be slower than 1.5 MB/s? What could be the speed limiting factor in my file transfer setup?
    Any tips on improving file transfer rate would be appreciated!

    Sustained 40 MB/s? That is my dream speed! The maximum speed I've ever seen here was about 25 MB/s, which is still very good, and that happened only after copying back and forth, which I explained above. Let me go through what I did with screencaps first.
    These are the test tdms files to transfer in the PXI: an original file witten in the PXI and its copy file. I copied the original file from PXI to laptop and renamed it (added '_copy') and then copied it back to PXI.
    FTP module is transferring the original file to the Filezilla server on laptop. Note the low speed.
    FTP module is transferring the copy file to the Filezilla server on laptop. Note the higher speed.
    This is the FTP sub VI I'm using. It transfers files in sub-folders (one level lower), too. The data connection is set to 'passive' on the FTP Put Multiple Files VI but it doesn't seem to make any difference. I attach the sub VI.
    You tried the file transfer with a 500 MB text file. Would there be any difference if the file type was tdms? Would the RAM size matter? It is 2GB here.
    And the LED is orange, which means Gigabit.
    Another question: Is there any way to programmatically stop file transfer in progress? Currently, I just have to restart the PXI.
    Thank you!
    Attachments:
    FTP.vi ‏40 KB

  • I have a question about Data Rates.

    Hello All.
    This is a bit of a noob question I'm sure. I don't think I really understand Data Rates and how it applies to Motion... therefore I'm not even sure what kind of questions to ask. I've been reading up online and thought I would ask some questions here. Thanks to all in advance.
    I've never really worried about Data Rates until now. I am creating an Apple Motion piece with about 15 different video clips in it. And 1/2 of them have alpha channels.
    What exactly is Data Rate? Is it the rate in which video clip data is read (in bits/second) from the Disc and placed into my screen? In Motion- is the Data Rate for video only? What if the clip has audio? If a HDD is simply a plastic disc with a dye read by "1" laser... how come my computer can pull "2" files off the disc at the same time? Is that what data transfer is all about? Is that were RAM comes into play?
    I have crunched my clips as much as I can. They are short clips (10-15seconds each). I've compressed them with the Animation codec to preserve the Alpha channel and sized them proportionally smaller (320x240). This dropped their data rate significantly. I've also taken out any audio that was associated with them.
    Is data rate what is slowing my system down?
    The data rates are all under 2MBs. Some are as low as 230Kbs. They were MUCH higher. However, my animation still plays VERY slowly.
    I'm running a 3GigRam Powerbook Pro 2.33GHz.
    I store all my media on a 1TB GRaid Firewire 800 drive. However for portability I'm using a USB 2 smartdisk external drive. I think the speed is 5200rpm.
    I'm guessing this all plays into the speed at which motion can function.
    If I total my data rate transfer I get somewhere in the vicinity of 11MBs/second. Is that what motion needs for it to play smoothly a 11MBs/second data connection? USB 2.0 is like what 480Mbs/second. So there is no way it's going to play quickly. What if I played it from my hard drive? What is the data rate of my internal HDD?
    I guess my overall question is.
    #1. Is my thinking correct on all of these topics? Do my bits, bytes and megs make sense. Is my thought process correct?
    #2. Barring getting a new machine or buying new hardware. What can I do to speed up this workflow? Working with 15 different video clips is bogging Motion down and becoming frustrating to work with. Even if only 3-4 of the clips are up at a time it bogs things down. Especially if I throw on a glow effect or something.
    Any help is greatly appreciated.
    -Fraky

    Data rate DOES make a difference, but I'd say your real problem has more to do with the fact that you're working on a Powerbook. Motion's real time capabilities derive from the capability of the video card. Not the processor. Some cards do better than others, but laptops are not even recommended for running Motion.
    To improve your workflow on a laptop will be limited, but there are a few things that you can try.
    Make sure that thumbnails and previews are turned off.
    Make sure that you are operating in Draft Mode.
    Lower the display resolution to half, or quarter.
    Don't expect to be getting real time playback. Treat it more like After Effects.
    Compressing your clips into smaller Animations does help because it lowers the data rate, but you're still dealing with the animation codec which is a high data rate codec. Unfortunately, it sounds necessary in your case because you're dealing with alpha channels.
    The data rate comes into play with your setup trying to play through your USB drive. USB drives are never recommended for editing or Motion work. Their throughput is not consistent enough for video work. a small FW drive would be better, though your real problem as I said is the Powerbook.
    If you must work on the powerbook, then don't expect real-time playback. Instead, build your animation, step through it, and do RAM previews to view sections in real time.
    I hope this helps.
    Andy

  • H.264 All-Intra Data Rates Significantly Higher

    Does the built-in H.264 codec encode I-frame only files differently? I am trying to determine the optimal GOP length for high bitrate exports. Image quality seems to degrade, even in the I-frames, when using key frame distances greater than one.
    After performing a series of tests to characterize the Adobe H.264 encoder, I discovered that exported files are significantly larger when key frame distance equals one frame (N=1). The average video data rate for a test file rendered with the Adobe H.264 encoder is as follows:
    N=1 : 2.17 bpp : 24I
    N=2 : 0.66 bpp : 12I + 12P
    N=3 : 0.59 bpp : 8I + 8B + 8P
    Note how the data rate drops 70% (from 2.17 to 0.66 bpp) even though 50% of the I-frames still exist when N=2. By comparison, here is the video data rate when exporting with QuickTime H.264:
    N=1 : 0.89 bpp : 24I
    N=2 : 0.70 bpp : 12I + 12P
    N=3 : 0.64 bpp : 8I + 16P
    The following chart shows data rates at key frame distances from 1-48 frames for Adobe H.264, QuickTime H.264 (via Adobe), QuickTime Pro, and Expression Encoder 4 Pro. Data rates are consistent among all encoders at GOP lengths greater than one. There is an anomoly with the Adobe H.264 codec compressing all-intra files.
    The observed behavior occurs in all profiles, which were tested at Levels 4.1, 4.2, 5.0, and 5.1:
    Image quality is better in the Adobe H.264 all-intraframe file, especially with respect to detail retention. The pictures below show sections of two consecutive frames magnified 400%. The file with N=2 is less accurate and contains noticeable blocking. Even the I-frames don't look as good in the files where N>1.
    The test file was a seventeen second Premiere Pro sequence consisting of H.264, MPEG-2, and AE files with effects applied. Exports were rendered from the Premiere Pro timeline and from a V210 uncompressed 4:2:2 intermediate file of the sequence. The following settings were used:
    Format: H.264
    Width: 1280
    Height: 720
    Frame Rate: 24 fps
    Field Order: Progressive
    Aspect: Square Pixels (1.0)
    TV Standard: NTSC
    Profile: Baseline, Main, and High
    Levels: 4.1, 4.2, 5.0, 5.1
    Render at Maximum Bit Depth: Enabled
    Bitrate Encoding: VBR, 2-Pass
    Target Bitrate: Maximized for each Profile/Level
    Maximum Bitrate: Maximized for each Profile/Level
    Key Frame Distance: 1, 2, 3, 4, 5, 6, 7, 8, 12, 24, 48
    Use Maximum Render Quality: Enabled
    Multiplexer: MP4
    Stream Capability: Standard
    Software:
    Adobe Media Encoder CS6 Creative Cloud v6.0.3.1 (64-bit)
    Premiere Pro CS6 Creative Cloud v6.0.3
    Windows 7 SP1
    QuickTime Pro for Windows v7.6.9 (1680.9)
    MediaInfo 0.7.62 (for GOP and data rate information)

    I think the short answer is yes,
    a h.264 encoder does I-frame only differently. I frames are complete expressions of a picture with no temporal compression information.
    P frames use _P_redictive information. IE information from prior frames.
    B frames use _B_i-directional predictive frame information.
    h.264 gets the majority of it's bit saving from use of B and P frames. When you do I-frame only you only get the block compression and none of the advantages of P and B frames. Thus the GOP N=1 doesn't get very good bit's per pixel.
    Having said all that I do find your comment...
    Profitic wrote:
    Note how the data rate drops 70% (from 2.17 to 0.66 bpp) even though 50% of the I-frames still exist when N=2. By comparison, here is the video data rate when exporting with QuickTime H.264:
    ... very interesting. Indeed, why is the datarate 70% less when it should at best be 50% for GOP N=2. 50% less should be the same I-frame information plus 0 bytes for the B frame between them. (GOP = I,B,I). Any more than that and it is throwing away bits from the I-frame. So, this seems to be to be a ratecontrol bug.
    http://en.wikipedia.org/wiki/Group_of_pictures
    "The GOP structure is often referred by two numbers, for example, M=3, N=12. The first number tells the distance between two anchor frames (I or P): it is the GOP size. The second one tells the distance between two full images (I-frames): it is the GOP length. For the example M=3, N=12, the GOP structure is IBBPBBPBBPBBI. Instead of the M parameter the maximal count of B-frames between two consecutive anchor frames can be used."

  • What is the Data Rate of the Cache in Adobe Premiere CS6

    Hi,
    Trying to build a raid array to edit with and want to know how much speed i need for caching read/write drive for premiere and After Effects. I'm of the impression that the cache doesn't need more than a single 7200rpm drive (~120MB/s) but really i have no clue what the data rate that adobe caches to is or what the data rate to read is
    so it's 2 questions
    how fast does premiere write the cache? (i know this is system dependent but a ballpark figure would be good for say a sandy bridge i5 2500k system with a 2 raid 7200rpm drive)
    and
    what's the data rate to read the cache in a 1080 24p file?
    also, when doing the final export does premiere/after effects read from the Cache or the original files or both?
    thx,
    Jayson
    youtube.com/AWDEfilms

    ok i've figured out the physics of this question.
    The Cache Read data rate is always larger than the Cache Write data rate, because the computer would have to be rendering to Cache faster than realtime for the Write rate to be higher, which would make it unnecessary to render to cache in the 1st place. So I'm really only worried about the Cache read data rate. Does adobe have a paper that tells us what the data rate is for different sequences.
    my 3 common workflows are
    canon h.264 1080 24p
    AVChd 1080 24p from my GH2 with a 44mb
    and
    r3d 5k epic footage 24p - (this is painful to edit )
    anyone know where this info is?
    thx,
    Jayson
    youtube.com/AWDEfilms

  • In Aperture 3.4 Export Slideshow to a video, what are the actual Data Rates used for "Best", "High", ... "Least" quality for a given resolution?

    My Photo website host (SmugMug) converts uploaded video files at a specific Data Rate in Mbps before installing them. I would like to compress my slideshow video file to the same rate before I upload it to reduce file size and upload time. When I choose the "Custom" setting for an Export, I can choose 1 of the 5 Quality choices and see the estimated file size. But, I cannot know what the actual Data Rate is until after I wait a long time to export the slideshow (hours for a long slideshow) and then examine the resulting file in QuickTime Player's Inspector.

    I ran a few tests using a short slideshow (16 images, 1:23 mins/secs) at 1,728 x 1,080 resolution to find out the bit rates for various quality level choices.
    Export setting              Quality          Resulting bit rate          File Size
    HD 1080p                    default                20.68 Mbps              214.8 MBytes
    Custom                       Best                   20.49                       212.8
    Custom                       High                     6.25                         65.0
    Custom                       Medium                3.97                         41.3
    However, I don't know if those bit rates will be the same for different length slideshows or for different output resolutions. My SmugMug host site uses an 8.0 Mbps rate for a 1728 x 1080 video file. If I choose Custom/Best, my file will be almost 3 times bigger and much higher quality than necessary, but if I choose Custom/High, my file will be smaller and lower quality than SmugMug's converted version.
    I have installed MPEG StreamClip that will let me convert an exported Aperture slideshow video file, and StreamClip allows me to choose a specific bit rate in Mbps. But, I would prefer not having to do a 2 step process (Export from Aperture, then convert in StreamClip).

  • Problem making slideshow. "Data rate to high"

    Using Encore CS3 on Win XP. Trying make dvd on a short film and some slideshows. Total c 2,5 GB. Building stops telling me: PGC "Lenas" has an error at 00:01:31:08. The data rate is too high for DVD. You must replace the file with one of a lower rate.....ref, EPGC...." All pictures in slideshow "Lenas" are the same size, 2-2,5 MB, nothing special with the picture at 00:01:31:08. What does the info say?
    Preview with no problems.
    Regards Gunnar Löwenhielm

    What are the pixel x pixel dimensions of those still images?
    What is your Encore Project setting?
    How many images do you have per SlideShow?
    Good luck,
    Hunt

  • Video (or Video + Audio) Data Rate

    Wondered if anyone here might know of a software utility that has the ability to scan QuickTime compatible files and either display the video (or video + audio) data rate excursion or, as an alternative, graph the instantaneous variation in the data rate over time.

    Apple +I (show info) in QT player will show you the framerate and bitrate while playing.
    Thanks for your suggestion, Rick. However, I must point out that the data rate so displayed is simply the total average for the entire clip and not the instantaneous (i.e., constantly changing) variance with time. Preliminary observations tend to indicate the single pass H.264 algorithm begings roughly 10% under the requested data rate and quickly settles very near the target. The multipass algorithm, on the other hand, appears to begin in the vicinity of 300% above the targeed data rate and monotonically decreases throughout the remainder of the movie clip. So much for a quick qualitative analysis based on observations during the actualy encoding process.
    What I now wish to do is actually perform a bit of quantitative analysis and, based on those results, correlate actual data rates with the ability of a given H.264 multipass file to sync to an iPod. Basically, I am trying to determine whether or not a user data rate input is used as a comparator during the initial phase of multipass H.264 coding and, if so, how it is implemented. (I.e., how are the delta values handled -- linearly, exponentially, etc.) In addition, I wish to compare QT v7.0.3 with v7.0.4 clips to determine why the latter are less iPod compatible. Too, there remains a question as to whether or not any data rate information is now embedded in the clips themselves since various work arounds tried (i.e., clipping file lengths) have proven unsuccessful.

Maybe you are looking for