What is the recommendations on rendering ?

I've just really gotten going on using AE, and i got frustrated pretty quickly. Mostly with render times, and overall performance.
I had a GeForce GTX 560 Ti and while it did ok, if I rendered anything that i used a heavy plugin, like Particular, it would take a tremendous amount of time to render.
After researching a little, and squeezing the wallet, I just couldn't afford one of those billion dollar cards you guys use, so i bought a GTX 770 SC.
I read that the program says it's not supported yet, but will be in October. i changed the file in the system to recognize my card, and that worked. So iit is definitely using the card.
But one issue I'm struggling with as a newbie is understanding rendering. I have been, up until tonight, dropping my projects into Adobe Media Encoder. Tonight, when i installed this new card, I put a GPU meter on my desktop so i could monitor it's use. I noticed when i put the project in media encoder it was barely touching the graphics card. I guess it's trying to render everything from the CPU. Which i don't undetrstand because the 770has the actual architecture for h264 encoding right?
So i put the same file in the render queue of AE and it immediately pegged the card, so i'm confident it's using it. I did the tweaks i see recommended and set it to use th max cpu etc.
What i don't quite get i guess is why i should render in the program. I mean, i assumed the media queue would be the best. Why would there be a dedicted app for it? The other reason, is using the regular render machine I don't see options for the h264 so i never really knew how to configure it. whereas the encoder puts out a perfect result for 1080p upload to YouTube. The same file it renders and gives me at 50m will be 3g if I render directly to an avi.
So I guess my point is, what is the best way to render? is there a reason I can't get the encoder program itself to utilize the GPU? What i will end up doing is rendering this as a goliath file, then bringing it back into the encorder to crunch it up and reduce the file size to leave me a fair quality that's reaasonable to upload.
I'm eger for any suggestions. i don't mind using the AE queue at all, I just don't know which is the best option.
Thanks everyone.
p.s. i did change the prefes for simultaneous frames, etc, and raised the allocation to the max. So i think i'm good there.
I'm running a test on a very heavy encoding project now. This is what my 'stuff' is doing. This looks about right to me, no?
By the way, the picture is slighly decieving. The GPU pegs to 100% then cycles down then back up.  the SS shows it in the middle.

> I had a GeForce GTX 560 Ti and while it did ok, if I rendered anything that i used a heavy plugin, like Particular, it would take a tremendous amount of time to render.
After researching a little, and squeezing the wallet, I just couldn't afford one of those billion dollar cards you guys use, so i bought a GTX 770 SC.
is there a reason I can't get the encoder program itself to utilize the GPU?
The GPU has very little to do with the speed of processing in After Effects, and it has nothing to do with the speed of processing in Adobe Media Encoder. The CPU, RAM, and disk setup are all much more important.
See this page for details of the few things that the GPU is used for:
http://adobe.ly/AE_CUDA_OpenGL_GPU
See this page for resources about making After Effects work faster: http://adobe.ly/eV2zE7
See this page for information about hardware for After Effects: http://adobe.ly/pRYOuk

Similar Messages

  • Is there a risk of setting a console connection timeout and what is the recommended setting?

    Is there a risk of setting a console connection timeout and what is the recommended setting? Please suggest if there is any best prctice documentation that can be referred.

    Hi Henrik
    depend on what you need or what your security policy says for my lab gear i use 60 minutes. because i know how can access this. if you have gear outside in insecure space set it to a minimum or disable the console. everybody how can access your gear can break in. simple restart and boot w/o config. and you are in.
    it realy depends how secure is your space and how much security you need.
    and than the settings for policy have to match, what sec do you have if your console login and logout is secure. but when you restart you can simple break in by starting w/o config and than load it.
    HTH
    Patrick

  • What is the recommended way to do multiple channel, single point sampling for control with an NI PCI-6255 in RLP?

    Hello,
    I am writing a driver for the M-series NI PCI-6255 for QNX. I have downloaded the MHDDK and have all the examples working. I have also enhanced the examples to do interrupt handling (e.g. on AI_FIFO interrupt or DMA Ring Buffer interrupt). My ultimate goal is to write a driver that I can use for closed-loop control at 500 Hz using all 80 channels of the NI PCI-6255. I may also need to synchronize each scan with a NI PCIe-7841R card for which I've already written a driver. I want an interrupt-driven solution (be it programmed I/O on an interrupt or DMA that generates an interrupt) so that the CPU is available to other threads while the 80 analog inputs are being read (since it takes quite a while). I also want to minimize the number of interrupts. Basically, I will need to collect one sample from all 80 channels every 2 milliseconds.
    There are many different options available to do so, but what is the recommended technique for the NI PCI-6255 card? I tried using the AI FIFO interrupt without DMA, but it seems to interrupt as soon as any data is in the AI FIFO (i.e. not empty condition), rather than when all 80 channels are in the FIFO, so more interrupts are generated than necessary. I tried using DMA in Ring Buffer mode to collect a single sample of 80 channels and interrupting on the DMA Ring Buffer interrupt, which appears to work better except that this technique runs into problems if I cannot copy all the data out of the DMA buffer before the next AI scan begins (because the DMA will start overwriting the buffer as it is in ring buffer mode). If the DMA is not in ring buffer mode or I make the ring buffer larger than one 80-channel sample then I don't have a way to generate an interrupt when one sample has been acquired (which I need, because I'm doing control).
    I saw something in the documentation about a DMA Continue mode in which it looks like you can switch between two different buffers (by programming the Base Count/Address with a different address than the current address) automatically and thereby double-buffer the DMA but there is no real documentation or examples on this capability. However, I think it would work better than the Ring Buffer because I could interrupt on the DMA CONT flag presumably and be copying data out of one buffer while it is filling the other buffer.
    Another option would be DMA chaining, but again, I cannot find any information on these features specific to the NI DAQs.
    I tried interrupting on AI STOP figuring that I could get a single interrupt for each scan, but that doesn't appear to work as expected.
    I know that DAQmx on Windows has the ability to do such single sample, multiple channel tasks at a fixed rate so the hardware must support it.
    Any suggestions would be appreciated.
    Thanks.
    Daniel Madill

    Hello,
    The interrupt that will happen nearest the times that you need is the AI_Start_Interrupt in the Interrupt_A group. This interrupt will occur with each sample clock. By the second time this interrupt fires, the AI FIFO should have the samples from the first conversion. If it is easier to use programmed IO, you can read the samples out of the FIFO until you get all 80.
    Additionally, you can set the DMA to send samples as soon as the FIFO is no longer empty...instead of waiting for half full or full. This change will reduce latency for your control loop. You can set AI_FIFO_Mode in AI_Mode_3_Register to 0. By the second time this interrupt fires, you should be able to check how much data is in the DMA ring buffer and read the 80 samples when they are available. You can make the ring buffer larger than 80 samples if you see data getting overwritten.
    There is no interrupt associated with 80 samples being available in the FIFO or 80 samples being available/transferred by DMA to the host. X Series has much more flexibility with these interrupts.
    I hope this helps!
    Steven T.

  • What is the recommended length of the cable in between the GPIB instrument and GPIB-USB card.

    hi,
    we are planning to use GPIB-USB Controller in our application by replacing the existing GPIB-PCI controller. Our GPIB instrument doesn't have the standard IEEE GPIB Connector. So, we will have to connect one more cable in between GPIB-USB controller and the GPIB Instrument. I just want to know what is the recommended lenth of the cable that can be connected in between the GPIB instrument and the GPIB-USB card. Could anybody please help me?
    Thanks,
    Rajaram

    Page 730 in the 2001 catalog gives detailed cabling requirements for the GPIB bus. It's usually best to keep the length 2meters or less.
    Randy Solomonson
    Application Engineer
    National instruments

  • What is the recommended ratio of publish nodes to dispatcher web servers for CQ5.5?

    What is the recommended ratio of publish nodes to dispatcher web servers for CQ5.5?
    We are getting conflicting opinions on this form the field. I realize this is a loaded question. I'd like to see how Adobe presents this opinion. In the past scaling at the webserver level was only loosely tied to the publishing layer. Now it seems as dynamic functionality increases the ratio is getting closer to 1:1.
    Also,I'd like to know in what use cases this load on publisher increases dramatically.

    Hello, this is an evergreen. Just call setPickOnBounds(false) on the CustomHexagon.
    An issue tracking this problem is open here: https://javafx-jira.kenai.com/browse/RT-17024

  • What is the recommended number of clients per Mac server? Also what are some recommended specs when purchasing an Apple machine that will have Mac OS X server installed?

    What is the recommended number of clients per Mac server? Also what are some recommended specs when purchasing an Apple machine that will have Mac OS X server installed? We have around 300 clients that need to be enroled on the Mac server. I want to know what is the recommended amount of clients a Mac server should contain. Also what are some recommended specs to make sure the server will flawlessly?

    Hello cpreasbeck,
    Thank you for contacting Apple Support Communities.
    I was able to find the following transition guide for Xserve that provides some workload guidance to determine performance when planning a server deployment.
    Transition Guide Xserve
    http://images.apple.com/xserve/pdf/L422277A_Xserve_Guide.pdf
    On page 9, Performance there is a chart that provides maximum numbers of connected users for various activities such as file sharing, mail, web, calendar, directory services and Time Machine and the CPU used as a server (Xserve, Mac Pro, Mac Mini). This information is a bit dated as the referenced software is Snow Leopard Server (OS X 10.6), and the hardware is older also, but it should give you a general idea of what you might need to look for.
    Regards,
    Jeff D.

  • What is the recommended way to obtain tracking data from carriers post XSI

    We currently run an old version of SAP Business Connector. We are in the process of migrating all interfaces off BC onto PI. The one remaining interface we have problems is the XSI (Express Delivery Interface) interface we have with ECC06 and UPS via the BC server. The interface works but is not stable and we would like to decommission it if we are able.
    I'm not 100% clear but it appears that XSI is no longer the recommended solution for obtaining tracking data from carriers. What is the recommend method today? We'd be happy to use a PI or ABAP solution but would prefer a standard solution that was supported by SAP and UPS.

    Using Time Machine is the simplest way to back up.
    debasencio wrote:
    They don't fit on a standard data DVD and when I try to back it up to my 500GB external hard drive it says its not formatted to files that size.
    How is the drive formatted?
    If it's DOS/FAT32, you can only size file sizes up to 4GB.
    If you are using it on a Mac only, format it Mac OS X HFS+.

  • What is the recommended way to upgrade system ruby and python binaries and their libraries?

    I'd like to upgrade my system ruby to 1.9.2 instead of old 1.8.7. Also was wondering what's the recommended way of upgrading any scripting languages installed on system?

    Hi Sogaard,
    see below:
    How does SAP recommend to integrate Webi-report in the SAP Portal? Is it through an URL iview, the iview templates (thumbnail, folder and alert) or through the Master Iview?
    The sample iViews are just samples for a particular case. The iView template allows you to use any application on the BusinessObjects system - including the viewing of content. What you are looking for is the creation process for WebI. So you have 2 options: you can use the KM integration and you can build a java application that offers the workflow you looking for and integrate it with the iView template
    Depending on what method should be used, I'd like to know which settings to be focused on. For instance, if the Master Iview is to be used, should opendocument be used instead of reports? And what other customizations will have to be done in order to integrate a Webi-report instead of a Crystal-Report?
    OpenDocument is for viewing reports. The Installation Guide outlines the integration of the iView template and the KM part.
    What is the intended use for the BOBJ repositoy manager? Would that be the way to integrate the Info-view?
    This is the integration into the KM repository which provides a richer functionality then just an iView.
    Ingo

  • What is the recommended way for persisting JMS messages?

    What is the recommended way for persisting JMS messages?. As per the IMQ admin documentation , using the default built-in persistence type which is through unix flat files is much efficient and faster, compared to the database persistence .
    Tried setting up the jdbc stuff for database persistence on iAS 6.5 . I am getting the following
    error .
    [24/Apr/2002:16:09:20 PDT] [B1060]: Loading persistent data...
    [24/Apr/2002:16:09:21 PDT] Using plugged in persistent store: database connection
    url=jdbc:oracle:thin:@dbatool.mygazoo.com:1521:qa1 brokerid=ias01
    [24/Apr/2002:16:09:23 PDT] [B1039]: Broker "jmqbroker" ready.
    [24/Apr/2002:16:11:56 PDT] ERROR [B4012]: Failed to persist interest
    SystemManager%3ASystemManagerEngine%2BiMQ+Destination%0AgetName%28%29%3A%09%09SM_Response%0AClass%3A%09%09%09com.sun.messaging.Topic%0AgetVERSION%28%29%3A%09%092.0%0AisReadonly%28%29%3A%09%09false%0AgetProperties%28%29%3A%09%7BJMQDestinationName%3DSM_Response%2C+JMQDestinationDescription%3DA+Description+for+the+Destination+Object%7D:
    java.sql.SQLException: ORA-01401: inserted value too large for column
    [24/Apr/2002:16:11:56 PDT] WARNING [B2009]: Creation of consumer SM_Response to destination 1
    failed:com.sun.messaging.jmq.jmsserver.util.BrokerException: Failed to persist interest
    SystemManager%3ASystemManagerEngine%2BiMQ+Destination%0AgetName%28%29%3A%09%09SM_Response%0AClass%3A%09%09%09com.sun.messaging.Topic%0AgetVERSION%28%29%3A%09%092.0%0AisReadonly%28%29%3A%09%09false%0AgetProperties%28%29%3A%09%7BJMQDestinationName%3DSM_Response%2C+JMQDestinationDescription%3DA+Description+for+the+Destination+Object%7D:
    java.sql.SQLException: ORA-01401: inserted value too large for column
    Any thoughts?

    From the output, you are using imq 2.0. In that release
    the key used to persist a durable subscriber in the database
    table has a limit of 100 characters. The output shows that
    your value is:
    SystemManager%3ASystemManagerEngine%2BiMQ+Destination%0AgetName%28%29%3A%09%09SM_Res
    ponse%0AClass%3A%09%09%09com.sun.messaging.Topic%0AgetVERSION%28%29%3A%09%092.0%0Ais
    Readonly%28%29%3A%09%09false%0AgetProperties%28%29%3A%09%7BJMQDestinationName%3DSM_R
    esponse%2C+JMQDestinationDescription%3DA+Description+for+the+Destination+Object%7D:
    which is much longer than 100 characters.
    You might want to shorten the string you use for the
    durable name.
    And yes, the default file-based persistence store is
    more efficient when compared to the plugged-in persistence
    through a database.

  • What is the recommended way to truncate tables in ODI?

    I want to create a separate step to truncate the result tables, before the start of the actual job. What is the recommended way of doing this?
    I am currently putting the truncate statements in the ODI procedure, but that has a lot of typing. Is there a odi command in the toolbox that I can use?
    Thanks.

    Ok,
    If the table will be loaded by interfaces, you have the "Truncate" option at the IKM's, just change it to "Yes".
    If you need to Truncate but they won't be loaded by Interfaces a possible way is:
    - requirements: it will be necessary to have some common "string" at tables names.
    1) create a procedure
    2) create an step
    3) at source tab put:
    Select table_name from user_tables where table_name in '%THE_STRING%'
    4) at target tab put:
    Truncate table #table_name
    If you don't have a common "string" you, instead, can create a table with all table names that you need to truncate and change the select command at 3).
    Does it help you?
    Message was edited by:
    Cezar_Santos

  • What is the recommended way of connecting to repository out of WebDAV, RMI, JNDI and JCA connector ?

    What is the recommended way of connecting to repository out of WebDAV, RMI, JNDI, and JCA connector possibilities provided by CQ 5.5?

    Hi dp_adusumalli,
    I recognized your list of ~8 questions you posted at around the same time, as I received that same list in our customer implementation from Arif A., from the India team, visiting San Jose. :-)
    I provided him feedback for most of the questions, so please check back with Arif for that info.
    For this particular question, can you provide specifics for the types of interactions you are interested in?
    Understanding the kinds of things you need to achieve will help determine which of the CQ/CRX interfaces is best suited for the task(s).
    I've collated a few points on this subject on this page:
    Manipulating the Adobe WEM/CQ JCR
    Regards,
    Paul

  • What's the recommended guideline on backing up a recovery catalog?

    while I understand "Benefits of Using the Recovery Catalog as the RMAN Repository" as stated in
    http://download-uk.oracle.com/docs/cd/B10501_01/server.920/a96566/rcmquick.htm#442214
    as well as things to take note if using control file as the sole rman repository,
    as stated in "Monitoring the Overwriting of Control File Records" in http://download-uk.oracle.com/docs/cd/B10501_01/server.920/a96566/rcmrepos.htm#446924
    I do not understand why in Backup and Recovery Advanced User's Guide
    http://download.oracle.com/docs/cd/B14117_01/server.101/b10734/rcmrepos.htm
    "Back Up the Recovery Catalog Often"
    Even if you have to restore the recovery catalog database using a control file autobackup, you can then use the full record of backups in your restored recovery catalog database to restore the target database without using a control file autobackup for the target database.
    =>probably implies that we can use another recovery catalog let's call it backupbackup to restore the recovery catalog
    "Choosing the Appropriate Method for Physical Backups"
    Follow these guidelines when developing an RMAN backup strategy for the recovery catalog database:
    * Run the recovery catalog database in ARCHIVELOG mode so that you can do point-in-time recovery if needed.
    * Set the retention policy to a REDUNDANCY value greater than 1.
    * Back up the database onto two separate media (for example, disk and tape).
    * Run BACKUP DATABASE PLUS ARCHIVELOG at regular intervals, to a media manager if available, or just to disk.
    * Do not use another recovery catalog as the repository for the backups.=> this contradicts the previous statement
    * Configure the control file autobackup feature to ON.
    so may I know what's the recommended guideline on backing up a recovery catalog? should we use another recovery catalog to keep the rman repositories of the the main primary catalog or just use a control file autobackup feature?
    thanks

    I understand the question and the confusion.
    Please note, however, that this forum is for discussing problems and errors with the documentation in general. Your question is related to a specific product and is, as far as I can tell, not a note about a documentation error or a correction.
    The appropriate place to discuss this is the Database (General) forum or - perhaps surprisingly - the RMAN forum. These can both be found by
    1) Clicking on the 'Forum Home' link at the top of this page;
    2) Scrolling down to the category 'Database' (a category is identified by a special row format that does not list the number of topics)
    3) Expanding the category by clicking on the 'More ...' link.
    4) Scrolling down to the "Database - General" or "Recovery Manager" forum and clicking on it's link.
    Note that forums have a search function that may be used to find answers if someone has already asked a similar question.

  • What is the recommended virus software for my MacBook?

    what is the recommended virus software for my MacBook?

    None, actually, is needed as there are no known viruse affecting OS X.
    Helpful Links Regarding Virus Protection
    An excellent link to read is Tom Reed's Mac Malware Guide.
    Also, visit The XLab FAQs and read Detecting and avoiding malware and spyware.
    If you require anti-virus protection I recommend using ClamXav.

  • What's the recommended setting for "Process memory usage" ("process virtual" in UI) for a 64-bit host on a 64-bit OS?

    Hi gurus
    In resource based throttling, what's the recommended setting for "Process memory usage" ("process virtual" in the resource based throttling tab of the UI) for a 64-bit host
    on a 64-bit Windows OS?
    According to MS (http://msdn.microsoft.com/en-us/library/ee308808(v=bts.10).aspx):
    "By default, the
    Process memory usage throttling threshold is set to 25. If this value is exceeded and the BizTalk process memory usage is more than 300 MB, a throttling condition may occur. On a 32-bit
    server, you can increase the Process memory usage value to 50. On a 64-bit server, you can increase this value to 100. This allows for more memory consumption by the BizTalk process before throttling
    occurs."
    Does this mean that 100 is the recommended setting for a 64-bit host on a 64-bit Windows?
    Thanks
    Michael Brandt Lassen

    Hi Michael,
    Recommended setting is the default setting which is 25 .dot.
    If your situation is abnormal and you see message delivery throttling state to “4” when the following performance counters are high or if you expect any of you integration
    process could have impact on following counters, then you can consider the suggestion by Microsoft. Don’t change the default setting.
    High process memory
    Process memory usage (MB)
    Process memory usage threshold (MB)
    You can see these counters under “BizTalk:MessageAgent”
    You can gauge these performance counter and its maximum values if have done any regression/performance testing in your test servers. If you have seen these counters having
    high values and causing throttling, then you can update the Process memory usage.
    Or unexpectedly you’re process high throughput messages in production which is causing these counters to go high and cause throttling, then up can update the Process memory
    usage.
    The above two cases where I know my expected process usage (by doing performance testing) or suddenly my production server processing has gone high due to unexpected business
    hike (or any reasons) which caused throttling, then do changes to default throttling setting.
    Just changing the default setting without actual reason could have adverse effect where you end up allocating 
    more processing capacities but the actual message processing message usage ever is low means you end up investing in underutilised resources.
    Regards,
    M.R.Ashwin Prabhu
    If this answers your question please mark it accordingly. If this post is helpful, please vote as helpful by clicking the upward arrow mark next to my reply.

  • With Premiere elements 11, when Publish/Share (saving) a finished video, selecting H.264  .mov file type, what are the recommended Codec and Aspect settings for a crisp video and a small file size?

    With Premiere Elements 11, when Publish/Share (saving) a finished video, selecting H.264  .mov file type, what are the recommended Aspect and Codec settings , for a crisp video with a small file size?

    Gary Krouse
    We will customize a very detailed reply, but
    (a) What are the properties of your source media?
    (b) What did you or the program set as the project preset to match the properties of the source media?
    (c) What specific requirements do you have for this H.264.mov export file?
    File size can be a compromise between bitrate, file size, and quality.
    Thanks.
    ATR

Maybe you are looking for

  • How to Attach files in background .

    HI all , How can I attach files in background in CRM . Thanks , Naval Bhatt .

  • Check in Billing Transaction

    Dear All, I want to place a Check in Billing Transaction that the Billing Date must be equal to the ACTUAL GI DATE of the Delivery Document. There are chances that sometimes User changes the Billing date while generating the Billing Document. So plea

  • Fusion Middleware Console

    Hi all, I have WLS 11g running. The Welcome page is working (http://<hostname>:7001), the WL Admin console is working (http://<hostname>:7001/console) and even my B2B console is working (http://<hostname>:3311/b2bconsole). Problem is the Middleware c

  • Can't move photos to smart album

    I have created a smart album based on face recognition, But I am falling to move new photos to the album by dragging them or by copying them. Also, I can't see the album name when choosing "albums" from the "add to" button. Any help will be appreciat

  • On opening a new window tab shows - "browser.startup.homepage=about:home"

    When I open a new window all I see is this "browser.startup.homepage=about:home". And address bar says chrome://branding/locale/browserconfig.properties What is happening and how do I fix this?