Best converters for logic and DP

Im going to start tracking and editing in DP 4.5 and mixing and programming with logic pro... Im trying to figure out wich converters would do the best job for both of these apps. i basically need the lowest latency possible when tracking while still maintaining awesome sound quality... also the card must have adat ins. any ideas?

I use trhe fireface
metric halo is good I hear ... Lynx
also the Saffire ive seen and is very nice ... but not adats

Similar Messages

  • The best DAW for Logic and big sample libraries... no more system overload

    I'm planning on going big and getting the Mac Pro Two 2.66GHz Quad-Core Intel Xeon, 16 GB of RAM and add two additional 1 TB drives, one for samples and one for projects. It's been suggested that I include a RAID 5 and that it will make streaming of the samples faster and also safer and more reliable for backing up data.
    With my current setup I constantly have to freeze tracks of sampled instruments or I get system overload.
    Do I need RAID or am I better off without it?

    Gabe Garza wrote:
    With my current setup I constantly have to freeze tracks of sampled instruments
    I'm not sure why you consider this a bad thing.
    Having to constantly freeze/unfreeze takes time and disrupts workflow. It's a good option to have if you need it, but much nicer if you don't.
    Johnny,
    I would strongly advise against diving straight in with 16GB of RAM. There are a lot threads around that discuss how much RAM Logic can even recognize, you should check them out. I don't remember the details very well, but I feel like 16GB was more that it can handle (maybe that was just for ESX...). In any case, don't buy a Mac pre-installed with all that memory; Apple charges way too much for their RAM, and you can find perfectly functional, guaranteed RAM from third-party sellers. I bought my Mac Pro with the base amount of RAM and have added as needed, and I recommend you do the same.
    When you play your projects, look at your CPU window.
    Look at the Audio and Disk I/O meters.
    If Audio is peaking, you need more CPU's
    If Disk I/O is peaking, a faster I/O on your hard drive is needed.
    I would also recommend doing this. Are you getting "core audio overload" messages, or another type? If you get core audio overload messages while your System Performance meters aren't peaking (or getting close to it), getting a faster machine isn't necessarily the answer. Core audio overloads are kind of tricky--some people with slower machines get them rarely, others with very fast machines get them often. If you haven't read up on this issue, I recommend you do so.
    I'm not too savvy on RAID arrays, but I have a few thoughts. How likely are you to need your PCI slots for other cards? The specs for the Mac Pro lists 3 PCI slots, which I believe means 3 empty ones, i.e. doesn't include the one that holds your graphics card. Do you expect to add a second graphics card? PCI based audio interface? Anything else like that? A RAID setup will occupy one of those PCI slots, just keep that in mind. Second, while a RAID 5 array is a pretty secure way to store your data, you ought to have an external backup drive anyway, so maybe it wouldn't be worth it ($700 for a RAID card, yikes!). Lastly, it's likely that any of the most popular 1TB drives (search barefeats.com for some articles on these) will be sufficient to do what you need them too. I don't know how involved your Logic sessions are, but I've never maxed out my disk I/O while using my Western Digital Caviar 640.

  • Can I check whole DB for logical and physical consistency?

    Assume I get somehow the impression that a part of a DB is somehow corrupted.
    E.g. due to hard disc sector errors or logical reference errors/changed foreign key values.
    How can I tell Oracle DB to check itself resp. to verify all the content inside for logical and physical integrity?
    Peter

    Use DBVerify to check corruption in the datafiles.
    Asif Momen
    http://momendba.blogspot.com

  • Best Practice for Planning and BI

    What's the best practice for Planning and BI infrastructure - set up combined on one box or separate? What are the factors to consider?
    Thanks in advance..

    There is no way that question could be answered with the information that has been provided.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • How can i get best practice for SD and MM

    Please, can any body tell me how can i get best practices for SD and MM for functional approach?
    Thanks
    Utpal

    Hello Utpal,
    I am really surprised, in just 10 minutes you searched that site and found it not useful. <b>Check out my previous reply "you will not find screen shot in this but you can add it in this"</b>
    You will not find readymade document, you need to add this as per your requirement.
    btw, the following link gives you some more link for new SAP guys, this will be helpful. <b>Check out HOW to BASIC transaction</b>
    New to Materials Management / Warehouse Management?
    Hope this helps.
    Regards
    Arif Mansuri

  • Best practice for Plan and actual data

    Hello, what is the best practice for Plan and actual data?  should they both be in the same app or different?
    Thanks.

    Hi Zack,
    It will be easier for you to maintain the data in a single application. Every application needs to have the category dimension, mandatorily. So, you can use this dimension to maintain the actual and plan data.
    Hope this helps.

  • DNS best practices for hub and spoke AD Architecture?

    I have an Active Directory Forest with a forest root such as joe.co and the root domain of the same name, and root DNS servers (Domain Controllers) dns1.joe.co and dns2.joe.co
    I have child domains with names in the form region1.joe.com, region2.joe.co and so on, with dns servers dns1.region1.joe.co and so on.
    Each region has distribute offices that may have a DC in them, servers named in the form dns1branch1.region1.joe.co
    Over all my DNS tests out okay, but I want to get the general guidelines for setting up new DCs correct.
    Configuration:
    Root DC/DNS server dns1.joe.co adapter settings points DNS to itself, then two other root domain DNS/DCs dns2.joe.co and dns3.joe.co.
    The other root domain DNS/DCs adapter settings point to root server dns1.joe.co and then to itself dns2.joe.co, and then 127.0.0.1
    The regional domains have a root dns server dns1.region1.joe.co with adapter that that points to root server dns1.joe.co then to itself.
    The additional region domain DNS/DCs adapter settings point to dns1.region1.joe.co then to itself then to dn1.joe.co
    What would you do to correct this topology (and settings) or improve it?
    Thanks in advance
    just david

    Hi,
    According to your description, my understanding is that you need suggestion about your DNS topology.
    In theory, there is no obvious problem. Except for the namespace and server plaining for DNS, zone is also needed to consideration. If you place DNS server on each domain and subdomain, confirm that if the traffic browsed by DNS will affect the network performance.
    Besides, fault tolerance and security are also necessary.
    We usually recommend that:
    DC with DNS should point to another DNS server as primary and itself as secondary or tertiary. It should not point to self as primary due to various DNS islanding and performance issues that can occur. And when referencing a DNS server on itself, a DNS client
    should always use a loopback address and not a real IP address. detailed information you may reference:
    What is Microsoft's best practice for where and how many DNS servers exist? What about for configuring DNS client settings on DC’s and members?
    http://blogs.technet.com/b/askds/archive/2010/07/17/friday-mail-sack-saturday-edition.aspx#dnsbest
    How To Split and Migrate Child Domain DNS Records To a Dedicated DNS Zone
    http://blogs.technet.com/b/askpfeplat/archive/2013/12/02/how-to-split-and-migrate-child-domain-dns-records-to-a-dedicated-dns-zone.aspx
    Best Regards,
    Eve Wang
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • What's the best app for editing and viewing office doc's

    What is the best App for editing and viewing office doc's, that will be compatable with IPhone 5S and Air pad?

    Search the app store, find one that you like and that works for you.

  • I will be buying the new IMac 27 inch in December. Would like input on the best configuration for photographic and possibly video processing-dream machine!

    I will be buying an IMac 27 inch in December. What is the best configuration for photo and possibly video processing.Does one need a Mac Pro? Dream machine!

    Just to explain what I just wrote to Laundry Bleach:
    Why is there no iDVD on my new Mac?
    https://discussions.apple.com/docs/DOC-3673
    UPDATE & ADDENDUM:
    But even though you can still buy iLife 11 that includes iDVD 7 from Amazon, Apple now make it difficult to install:
    Poster jhb21939 posted this in another thread:
    “when I attempted to load iDVD into a new iMac. A notice came up on the screen stating that the 'Authorisation Licence' had expired on 25 March this year (2012).
    I contacted the Apple support team and eventually, I was told that the Licence had been withdrawn and could no longer be used.”
    In other words Apple are now so adamant that we don’t use iDVD that they have tried to make it impossible to install.
    In response, Old Toad posted this solution:
    “You can still use it one all of your Macs.  If you get an invalid certificate message just set your Mac's clock to sometime before early 2011 and run the installer.  After you're done reset the time back to the correct time.” He added this comment:
    “It began after iDVD and iWeb were discontued and they were dropped from the Apple Store. All I can think of is the certificate was set to expire after a certain time period after the intitial iLife disc was released.
    I've been able to use the installer even without setting back the date.  I just clicked on the Continue button and it would work as expected.  For some it would not continue unless the date was set back.”
    The latest anorexic iMacs just announced do not even include a CD drive! Proof positive that Apple virtually prohibit the use of DVDs - although the newly announced Mac Minis do include a Superdrive.
    Yet, they still include iMovie! Heaven alone knows or understands what you are supposed to do with your newly edited masterpiece - except make a low quality version for YouTube?

  • Best program for monitoring and blocking internet use on my kids mobile devices, iPads, iPhones from my MAC

    Whats the best program for monitoring and blocking internet sites and usage for my kids on their mobile devices ipads, iphones and ipods, all from my desktop mac?

    Can't be done from your Mac remotely.
    But you can enable parental controls directly on iOS devices.
    iOS: Understanding Restrictions (parental controls)

  • Kernel: PANIC! -- best practice for backup and recovery when modifying system?

    I installed NVidia drivers on my OL6.6 system at home and something went bad with one of the libraries.  On reboot, the kernel would panic and I couldn't get back into the system to fix anything.  I ended up re-installing the OS to recovery my system. 
    What would be some best practices for backing up the system when making a change and then recovering if this happens again?
    Would LVM snapshots be a good option?  Can I recovery a snapshot from a rescue boot?
    EX: File system snapshots with LVM | Ars Technica -- scroll down to the section discussing LVM.
    Any pointers to documentation would be welcome as well.  I'm just not sure what to do to revert the kernel or the system when installing something goes bad like this.
    Thanks for your attention.

    There is often a common misconception: A snapshot is not a backup. A snapshot and the original it was taken from initially share the same data blocks. LVM snapshot is a general purpose solution which can be used, for example, to quickly create a snapshot prior to a system upgrade, then if you are satisfied with the result, you would delete the snapshot.
    The advantage of a snapshot is that it can be used for a live filesystem or volume while changes are written to the snapshot volume. Hence it's called "copy on write (COW), or copy on change if you want. This is necessary for system integrity to have a consistent data status of all data at a certain point in time and to allow changes happening, for example to perform a filesystem backup. A snapshot is no substitute for a disaster recovery in case you loose your storage media. A snapshot only takes seconds, and initially does not copy or backup any data, unless data changes. It is therefore important to delete the snapshot if no longer required, in order to prevent duplication of data and restore file system performance.
    LVM was never a great thing under Linux and can cause serious I/O performance bottlenecks. If snapshot or COW technology suits your purpose, I suggest you look into Btrfs, which is a modern filesystem built into the latest Oracle UEK kernel. Btrfs employs the idea of subvolumes and is much more efficient that LVM because it can operate on files or directories while LVM is doing the whole logical volume.
    Keep in mind however, you cannot use LVM or Btrfs with the boot partition, because the Grub boot loader, which loads the Linux kernel, cannot deal with LVM or BTRFS before loading the Linux kernel (catch22).
    I think the following is an interesting and fun to read introduction explaining basic concepts:
    http://events.linuxfoundation.org/sites/events/files/slides/Btrfs_1.pdf

  • Pls help putting together the BEST SOLUTION for PP and AE CS4

    Hi guys
    I was tasked to put together the best machine (machines) for video editing / rendering .. money is not an issue.
    My client is a TV broadcasting network and they do lots of videos everyday .. They want to render their work as FAST as possible because sometimes they have tight deadlines and rendering and re rendering (if there are mistakes) has cost them .. so now they're serious about getting the best technology to achieve best rendering time possible, but in a bit of practical sense of course.. i mean there's no sense spending 1000USD or more for a piece of technology that offers under 10 second faster rendering only ..
    What they do and want
    - edit/put together/ sequence lots of videos constantly, under premiere cs3 (will upgrade to cs4 on the new machine)
    - use affter effects for some effects and 3D stuffs.. but not that much.. but increasingly as time goes by and their video editors get more affter effects training..
    - needs really really fast rendering!! Right now it takes them typically 1-2 hours to render stuffs, and what to cut this time by half or more..
    Existing Setups on work stations:
    Intel QuadCore (Q6600, i7, Q9550, Xeon x3220, and the likes)
    4GB RAM DDR2 1066mhz
    Highend motherboards
    GeForce 9500GT , 9800GT, and GTX 260 216sp
    GSKILL Falcon 64GB SSD Drive as primary
    1TB Western Digital BLACK edition drives for storage
    Software:
    Windows XP 32 bit
    Adobe Premiere Pro CS3 (80% of all their work done here)
    AE CS3
    etc..
    Running NOD32 Antivirus
    I am an somewhat of an expert in Hardware and stuff, i build custom PC's machines most of my life but i have limiited knowledge on optimizing hardware for VIDEO EDITING stuffs.. so i turn to you all.
    I do understand that w/ video editing CPU is the most important, followed by RAM, and ofcourse by HDD for disc caches and stuff.. and where videocard may come to an advantage as well. and normally i'd suggest the fastest components available . i7 , extreme processors, skulltrail dual quad setups perhaps? RAID 0 SSD Drive? etc.. but I do not know w/c gives the most advantage, and w/c is a waste of money..
    I also read about quadro cards offering really big performance improvements using CUDA tech etc..
    again im not the expert at this type of requirment.. so i need your advises..
    with unlimitted PRACTICAL budget given to me to come up w/ a solution I initially thought a rendering farm would be the solution, but upon reading around it seems there's no rendering farm solution for PP CS4.. just the adobe after effects part.. so i had to go back to the drawing board and suggest really powerful work stations instead..and perhaps have them migrate to 64bit OS for that "extra" performance.. i myself am on win7 64bit and it's looking mighty good.
    1) is this the right move?
    2) is it still worth it to put together rendering farm for AE renders only? or might as well concentrate funds on the machines/workstations?
    Also along the PRACTICALITY LINE, would it be practical to maintain all the existing machines, and instead build a manned RENDERING MACHINE w/ PP and AE (basically desktop or Server if they are faster at rendering w/ the best specs for the purpose etc) where they can just throw the workload to some guy there for rendering? Atleast the editors can continue using the workstations while the rendering machine guy is at work? and even perhaps setup a gigabit lan network where the RENDERING MACHINES can dynamically pull the work load via the network , not needing to transfer the files to the rendering machine first ? etc
    as you can see im not only looking for the best machine, but the best STRATEGY for my client.
    Good suggestions would be much appreciated.
    regards to all

    in a nut shell no. if you do not have to "Share" then Avid is pointless Adobe is a far better answer.
    Avid solutions still need a computer. The Mojo.Nitris are a joke and Avid no longer provides computers even the Symphany comes with your choice of Apple Mac Pro or HP 800. (Dual Xeon)
    there are better "hardware" answers that what Avid sells for Adobe. Decklink, Aja to name a few.
    if you have to share then you have no choice. this is the one area Abode does not lead.
    renders have to do primarily with 2 things CPU power (Xeon will render faster than single i7)
    and Drives. the faster your drive set up the faster the render. Memory does play a part and so does the video card depending on what codec and to what
    it also depends on WHAT you are rendering and from what for the HDDs.
    without knowing your work flow format and codec its impossible to suggest something.
    in terms of rendering
    Core i7
    Dual Xeon (faster unless still doing DV/HDV thru firewire)
    Sonnet raid array
    Scott
    ADK

  • What are best practice for packaging and deploying j2EE apps to iAS?

    We've been running a set of J2EE applications on a pair of iAS SP1b for about a year and it has been quite stable.
    Recently however we have had a number of LDAP issues, particularly when registering and unregistering applications (registering ear files sometimes fails 1st time but may work 2nd time). Also We've noticed very occasionally that old versions of classes sometimes find their way onto our machines.
    What is considered to be best practice in terms of packaging and deployment, specifically:
    1) Packaging - using the deployTool that comes with iAS6 SP1b to package is a big manual task, especially when you have 200+ jsp files. Are people out there using this or are they scripting it with a build tool such as Ant?
    2) Deploying an existing application to multiple iAS's. Are you guys unregistering old application then reregistering new application? Are you shutting down iAS whilst doing the deployment?
    3) Deploying ear files can take 5 to 10 mins, is this normal?
    4) In a clustered scenario where HTTPSession is shared what are the consequences of doing deployments to data stored in session?
    thanks in asvance for your replies
    Owen

    You may want to consider upgrading your application server environment to a newer service pack. There are numerous enhancements involving the deployment tool and run time layout of your application that make clear where you're application is loading its files from.
    If you've at a long running application server environment, with lots of deployments under your belt, you might start to notice slow downs in deployment and kjs start time. Generally this is due to garbage collecting in your iAS registry.
    You can do several things to resolve this. The most complete solution is to reinstall the application server. This will guarantee a clean ldap registry. Of course you've got to restablish your configurations and redeploy your applications. When done, backup your application server install space with the application server and directory server off. You can use this backup to return to a known configuation at some future time.
    For the second method: <B>BE CAREFUL - BACKUP FIRST</B>
    There is a more exhaustive solution that involves examining your deployed components to determine the active GUIDS. You then search the NameTrans section of the registry searching for Applogic Servlet *, and Bean * entries that represent your previously deployed components but are represented in the set of deployed GUIDs. Record these older GUIDs, remove them from ClassImp and ClassDef. Finally remove the older entries from NameTrans.
    Best practices for deployment depend on your particular environmental needs. Many people utilize ANT as a build tool. In later versions of the application server, complete ANT scripts are included that address compiling, assembly and deployment. Ant 1.4 includes iAS specific targets and general J2EE targets. There are iAS specific targets that can be utilized with the 1.3 version. Specialized build targets are not required however to deploy to iAS.
    Newer versions of the deployment tool allow you to specify that JSPs are not to be registered automatically. This can be significant if deployment times lag. Registered JSP's however benefit more fully from the services that iAS offers.
    2) In general it is better to undeploy then redeploy. However, if you know that you're not changing GUIDs, recreating an existing application with new GUIDs, or removing registered components, you may avoid the undeploy phase.
    If you shut down the KJS processes during deployment you can eliminate some addition workload on the LDAP server which really gets pounded during deployment. This is because the KJS processes detect changes and do registry loads to repopulate their caches. This can happen many times during a deployment and does not provide any benefit.
    3) Deploying can be a lengthy process. There have been improvements in that performance from service pack to service pack but unfortunately you wont see dramatic drops in deployment times.
    One thing you can do to reduce deployment times is to understand the type of deployment. If you have not manipulated your deployment descriptors in any way, then there is no need to deploy. Simply drop your newer bits in to the run time space of the application server. In later service packs this means exploding the package (ear,war, or jar) in to the appropriate subdirectory of the APPS directory.
    4) If you've changed the classes of objects that have been placed in HTTPSession, you may find that you can no longer utilize those objects. For that reason, it is suggested that objects placed in session be kept as simple as possible in order to minimize this effect. In general however, is not a good idea to change a web application during the life span of a session.

  • Best practice for PK and indexes?

    Dear All,
    What is the best practice for making Primary Key and indexes? Should we keep them in the same tablespace as table or should we create a seperate tableapce for all indexes and Primary Key? Please note I am talking about a table that has 21milion rows at the moment and increasing 10k to 20k rows daily. This table is also heavily involved in daily reports and causing slow performance. Currently the complete table with all associated objects such as indexes and PK is stored in one seperate tablespace. If my way is right then please advise me how can I improve the performance of retrival or DML operation on this table?
    Thanks in advance..
    Zia Shareef

    Well, thanks for valueable advices... I am using Oracle 8i and let me tell you exact problem...
    My billing database has two major tables having almost 21 millions rows each... one has collection data and other one for invoices... many reports are showing the data with the joining of Customer + Collection + Invoices tables.
    There are 5 common fields in between invoices(reading) and collection tables
    YEAR, MONTH, AREA_CODE, CONS_CODE, BILL_TYPE(adtl)
    My one of batch process has following update and it is VERY VERY SLOW:
    UPDATE reading r
    SET bamount (SELECT sum(camount)
    FROM collection cl
    WHERE r.ryear = cl.byear
    AND r.rmonth = cl.bmonth
    AND r.area_code = cl.area_code
    AND r.cons_code = cl.cons_code
    AND r.adtl = cl.adtl)
    WHERE area_code = 1
    tentatively area_code(1) is having 20,000 consumers
    each consuemr may have 72 invoices and against these invoices it may have 200 rows in collection tables (system have provision to record partial payment against one invoice)
    NOTE: Please note presently my process is based on cursors so the above query runs for one consumer at one time but just for giving an idea I have made it for whole area.
    Mr. Yingkuan, can you please tell me how can I check that the table' statistics is not current and how can I make it current. Is it really effect performance?

  • Best Practice for SUP and WSUS Installation on Same Server

    Hi Folks,
    I have a question, I am in process of deploying SCCM 2012 R2... I was in process of deploying Software Update Point on SCCM with one of the existing WSUS server installed on a separate server from SCCM.
    A debate has started with of the colleague who says that the using remote WSUS server is recommended by Microsoft because of the scalability security  that WSUS will be downloading the updates from Microsoft and SCCM should be working as downstream
    server to fetch updates from WSUS server.
    but according to my consideration it is recommended to install WSUS server on the same server where SCCM is installed... actually it is recommended to install WSUS on a site system and you can used the same SCCM server to deploy WSUS.
    please advice me the best practices for deploying SCCM and WSUS ... what Microsoft says about WSUS to be installed on same SCCM server OR WSUS should be on a separate server then the SCCM server ???
    awaiting your advices ASAP :)
    Regards, Owais

    Hi Don,
    thanks for the information, another quick one...
    the above mentioned configuration I did is correct in terms of planning and best practices?
    I agree with Jorgen, it's ok to have WSUS/SUP on the same server as your site server, or you can have WSUS/SUP on a dedicated server if you wish.
    The "best practice" is whatever suits your environment, and is a supported-by-MS way of doing it.
    One thing to note, is that if WSUS ever becomes "corrupt" it can be difficult to repair and sometimes it's simplest to rebuild the WSUS Windows OS. If this is on your site server, that's a big deal.
    Sometimes, WSUS goes wrong (not because of ConfigMgr)..
    Note that if you have a very large estate, or multiple primary site servers, you might have a CAS, and you would need a SUP on the CAS. (this is not a recommendation for a CAS, just to be aware)
    Don
    (Please take a moment to "Vote as Helpful" and/or "Mark as Answer", where applicable.
    This helps the community, keeps the forums tidy, and recognises useful contributions. Thanks!)

Maybe you are looking for

  • Problem with conversion  AVCHD

    I have the Sony SR-7 and I convert the AVCHD clips with Log & Transfer (Final Cut Express 4.0.1) but the converted clips are in fast motion (3X or 4X about): A clip about 10 sec is converted in a 3 second clip approximately. And the same thing with t

  • Critical patches for oracle enterprise linux

    Hello, I'm very new to this so here I go. Where do I go to find out all the critical patches there are for oracle enterprise linux. Please provide the link, so I can read about all the critical patches. Thanks a million for your time.

  • How do you EXTRACT those Firmwares?

    I am running a linksys wireless rounter WRT54GS v2 with three computers, two desktop and one laptop.  Not all computers are on at the same time.  Normally only one computer is on but sometime two. two computers are running Windows XP and one is runni

  • Script for quitting an application when switching users?

    Is there any way to create an applescript that will quit iTunes for me when I switch users? I have my computer set up with two users, one for work and one for "play". I use iTunes in both users and they both use the same library. Because of this, iTu

  • Needs suggestion and help regarding my preparation

    I am looking for new job from my existing role. Here all the interviews are I am getting are scenario based and I am not able to reply most of them and as a result getting rejecting in every instance. . In my day to day work I am a part of DBA Monito