Best FS for 500X5GB hardware RAID5?

I would like to hear your thoughts on this.  I have a 500x5GB RAID5 array currently formatted with NTFS (from Windows), and I'm currently moving all of the data over to another RAID0 partition in order to allow me to delete the partition and reformat it in a different file system.  The reason for this is that NTFS  doesn't allow me to prevent accidental file deletion using an immutable flag using chattr:
sudo chattr -R +i /media/Perc\ RAID5/Music
Here is some info on it:
Storing documents, music, movies, and television shows.  Most of the storage it will be larger mkv files (so >5GB).
It will be for archiving only, where the data is mirrored on my RAID0 /home partition, which is where I point Amarok and browse to read documents, etc. 
I will occasionally watch Movies and TV shows from it, but other than that it's strictly for archiving.  That being said I need
I would like to choose a FS that I can still read from Windows 7 in order to play video files from Windows 7 x64.  So far I'm reading up between ext4, XFS, JFS, and Reiser.  I don't really need write support in Windows 7.
I know none of these FSes are natively support in Windows (which is a joke, but I digress....), however I've used programs that implement ext in the past.  I tried to get ext2fs to work with my /home ext4 partition, but it couldn't read it.  It was able to detect it and assign a drive letter, but it said it needed to be formatted before. 
If anyone has any tips/thoughts on any of this I would appreciate it.  Right now I'm backing everything up and reading on the different file systems, then afterward I'll go ahead and format.  I may also recreate the array if anyone has suggestions for a different stripe size to go with the FS....currently I have 64 kB stripe size IIRC.
Thanks!

I'm glad you've got things working out for you!
I've heard JFS is very stable and good for dealing with LOTS of file entries.  I use XFS.  Good with big files.  Both JFS and XFS have the down-side of potential file corruption with power loss as they only journal metadata.  There are good wikipedia articles on JFS and XFS, as well as the Arch wiki articles here.  XFS has a community website: Xfs.org  There is a lot you can do to improve XFS performance when creating the XFS partiton.
You should also keep your eyes on ZFS and BTRFS.  ZFS is currently being ported to linux.  BTRFS is kinda like the linux version of ZFS, but is still in alpha-- I think.  Both are filesystems of the future!
CoLinux is very cool, btw.  I'll think of it the next time I try to get a friend to use Linux
Last edited by eliphant0723 (2010-10-20 08:58:33)

Similar Messages

  • Best practice for migration to new hardware

    Hi,
    We are commissioning new hardware for our Web Server. Our current webserver is version 6.1SP4, and for the new server we've decided to stay with 6.1 but install SP7.
    Is there a best practice for migrating content from one physical server to another?
    What configuration files should I watch out for?
    Hopefully the jump from SP4 to SP7 won't cause too many problems.
    Thanks,
    John

    unfortunately, there is no quick solution for migrating from 1 server to other. you will need to carefully reconstruct
    - acl rules
    - server hostname configurations
    - any certificates that have been created on the old machine

  • Pls help putting together the BEST SOLUTION for PP and AE CS4

    Hi guys
    I was tasked to put together the best machine (machines) for video editing / rendering .. money is not an issue.
    My client is a TV broadcasting network and they do lots of videos everyday .. They want to render their work as FAST as possible because sometimes they have tight deadlines and rendering and re rendering (if there are mistakes) has cost them .. so now they're serious about getting the best technology to achieve best rendering time possible, but in a bit of practical sense of course.. i mean there's no sense spending 1000USD or more for a piece of technology that offers under 10 second faster rendering only ..
    What they do and want
    - edit/put together/ sequence lots of videos constantly, under premiere cs3 (will upgrade to cs4 on the new machine)
    - use affter effects for some effects and 3D stuffs.. but not that much.. but increasingly as time goes by and their video editors get more affter effects training..
    - needs really really fast rendering!! Right now it takes them typically 1-2 hours to render stuffs, and what to cut this time by half or more..
    Existing Setups on work stations:
    Intel QuadCore (Q6600, i7, Q9550, Xeon x3220, and the likes)
    4GB RAM DDR2 1066mhz
    Highend motherboards
    GeForce 9500GT , 9800GT, and GTX 260 216sp
    GSKILL Falcon 64GB SSD Drive as primary
    1TB Western Digital BLACK edition drives for storage
    Software:
    Windows XP 32 bit
    Adobe Premiere Pro CS3 (80% of all their work done here)
    AE CS3
    etc..
    Running NOD32 Antivirus
    I am an somewhat of an expert in Hardware and stuff, i build custom PC's machines most of my life but i have limiited knowledge on optimizing hardware for VIDEO EDITING stuffs.. so i turn to you all.
    I do understand that w/ video editing CPU is the most important, followed by RAM, and ofcourse by HDD for disc caches and stuff.. and where videocard may come to an advantage as well. and normally i'd suggest the fastest components available . i7 , extreme processors, skulltrail dual quad setups perhaps? RAID 0 SSD Drive? etc.. but I do not know w/c gives the most advantage, and w/c is a waste of money..
    I also read about quadro cards offering really big performance improvements using CUDA tech etc..
    again im not the expert at this type of requirment.. so i need your advises..
    with unlimitted PRACTICAL budget given to me to come up w/ a solution I initially thought a rendering farm would be the solution, but upon reading around it seems there's no rendering farm solution for PP CS4.. just the adobe after effects part.. so i had to go back to the drawing board and suggest really powerful work stations instead..and perhaps have them migrate to 64bit OS for that "extra" performance.. i myself am on win7 64bit and it's looking mighty good.
    1) is this the right move?
    2) is it still worth it to put together rendering farm for AE renders only? or might as well concentrate funds on the machines/workstations?
    Also along the PRACTICALITY LINE, would it be practical to maintain all the existing machines, and instead build a manned RENDERING MACHINE w/ PP and AE (basically desktop or Server if they are faster at rendering w/ the best specs for the purpose etc) where they can just throw the workload to some guy there for rendering? Atleast the editors can continue using the workstations while the rendering machine guy is at work? and even perhaps setup a gigabit lan network where the RENDERING MACHINES can dynamically pull the work load via the network , not needing to transfer the files to the rendering machine first ? etc
    as you can see im not only looking for the best machine, but the best STRATEGY for my client.
    Good suggestions would be much appreciated.
    regards to all

    in a nut shell no. if you do not have to "Share" then Avid is pointless Adobe is a far better answer.
    Avid solutions still need a computer. The Mojo.Nitris are a joke and Avid no longer provides computers even the Symphany comes with your choice of Apple Mac Pro or HP 800. (Dual Xeon)
    there are better "hardware" answers that what Avid sells for Adobe. Decklink, Aja to name a few.
    if you have to share then you have no choice. this is the one area Abode does not lead.
    renders have to do primarily with 2 things CPU power (Xeon will render faster than single i7)
    and Drives. the faster your drive set up the faster the render. Memory does play a part and so does the video card depending on what codec and to what
    it also depends on WHAT you are rendering and from what for the HDDs.
    without knowing your work flow format and codec its impossible to suggest something.
    in terms of rendering
    Core i7
    Dual Xeon (faster unless still doing DV/HDV thru firewire)
    Sonnet raid array
    Scott
    ADK

  • What is teh best solution for my Problem??

    hello everyone
    i'am planing to develop a Patient Management System (PMS)
    since i 'm not that guru in the J technologies i would like you to get me some help with this decision:
    what is the best solution for developeing an intranet "PMS" which should handles data (connection to postgresql) and runs some algorithm (for planing puropose) im background ??
    1 developeing a web based solution (servlet and jsp and and ..)
    2 developeing an application wich will be manging everything ??? (a kind of programm with its own GUI)
    3 you may have an other idea solution and ill be gald for any advices
    ps sorry for my english its not that great ;)

    Web applications are all the rage these days and IMO rightly so as they greatly reduce deployment cost and headaches.
    They do have some drawbacks though which for you might be important.
    I can imagine that sooner or later (probably sooner in my experience) a requirement to interface with hardware installed on the client machines will emerge.
    Whether this is a barcode scanner or a chipcard reader to get client information from some form of card issued to the client, or maybe a label printer to print out labels for medication bottles doesn't matter, you need to access the client machine directly.
    Web applications have great difficulty with that (in fact, it's all but impossible).
    You could go for an applet solution but that's contrived at best and still leaves the same problems.
    So in your situation an old-fashioned client/server application (maybe using web services to communicate with the serverside application) is likely the best longterm solution.
    You'd have the database remote so everyone has access to the same data, being accessed through processes run on the server and called by the clients. These can also do most of the business logic, leaving the clients relatively dumb responsible mainly for data validation and entry and presentation of the results.

  • Best Practices for Using Photoshop (and Computing in General)

    I've been seeing some threads that lead me to realize that not everyone knows the best practices for doing Photoshop on a computer, and in doing conscientious computing in general.  I thought it might be a good idea for those of us with some exprience to contribute and discuss best practices for making the Photoshop and computing experience more reliable and enjoyable.
    It'd be great if everyone would contribute their ideas, and especially their personal experience.
    Here are some of my thoughts on data integrity (this shouldn't be the only subject of this thread):
    Consider paying more for good hardware. Computers have almost become commodities, and price shopping abounds, but there are some areas where spending a few dollars more can be beneficial.  For example, the difference in price between a top-of-the-line high performance enterprise class hard drive and the cheapest model around with, say, a 1 TB capacity is less than a hundred bucks!  Disk drives do fail!  They're not all created equal.  What would it cost you in aggravation and time to lose your data?  Imagine it happening at the worst possible time, because that's exactly when failures occur.
    Use an Uninterruptable Power Supply (UPS).  Unexpected power outages are TERRIBLE for both computer software and hardware.  Lost files and burned out hardware are a possibility.  A UPS that will power the computer and monitor can be found at the local high tech store and doesn't cost much.  The modern ones will even communicate with the computer via USB to perform an orderly shutdown if the power failure goes on too long for the batteries to keep going.  Again, how much is it worth to you to have a computer outage and loss of data?
    Work locally, copy files elsewhere.  Photoshop likes to be run on files on the local hard drive(s).  If you are working in an environment where you have networking, rather than opening a file right off the network, then saving it back there, consider copying the file to your local hard drive then working on it there.  This way an unexpected network outage or error won't cause you to lose work.
    Never save over your original files.  You may have a library of original images you have captured with your camera or created.  Sometimes these are in formats that can be re-saved.  If you're going to work on one of those files (e.g., to prepare it for some use, such as printing), and it's a file type that can be overwritten (e.g., JPEG), as soon as you open the file save the document in another location, e.g., in Photoshop .psd format.
    Save your master files in several places.  While you are working in Photoshop, especially if you've done a lot of work on one document, remember to save your work regularly, and you may want to save it in several different places (or copy the file after you have saved it to a backup folder, or save it in a version management system).  Things can go wrong and it's nice to be able to go back to a prior saved version without losing too much work.
    Make Backups.  Back up your computer files, including your Photoshop work, ideally to external media.  Windows now ships with a quite good backup system, and external USB drives with surprisingly high capacity (e.g., Western Digital MyBook) are very inexpensive.  The external drives aren't that fast, but a backup you've set up to run late at night can finish by morning, and if/when you have a failure or loss of data.  And if you're really concerned with backup integrity, you can unplug an external drive and take it to another location.
    This stuff is kind of "motherhood and apple pie" but it's worth getting the word out I think.
    Your ideas?
    -Noel

    APC Back-UPS XS 1300.  $169.99 at Best Buy.
    Our power outages here are usually only a few seconds; this should give my server about 20 or 25 minutes run-time.
    I'm setting up the PowerChute software now to shut down the computer when 5 minutes of power is left.  The load with the monitor sleeping is 171 watts.
    This has surge protection and other nice features as well.
    -Noel

  • Best Practices for new iMac

    I posted a few days ago re failing HDD on mid-2007 iMac. Long story short, took it into Apple store, Genius worked on it for 45 mins before decreeing it in need of new HDD. After considering the expenses of adding memory, new drive, hardware and installation costs, I got a brand new iMac entry level (21.5" screen,
    2.7 GHz Intel Core i5, 8 GB 1600 MHz DDR3 memory, 1TB HDD running Mavericks). Also got a Superdrive. I am not needing to migrate anything from the old iMac.
    I was surprised that a physical disc for the OS was not included. So I am looking for any Best Practices for setting up this iMac, specifically in the area of backup and recovery. Do I need to make a boot DVD? Would that be in addition to making a Time Machine full backup (using external G-drive)? I have searched this community and the Help topics on Apple Support and have not found any "checklist" of recommended actions. I realize the value of everyone's time, so any feedback is very appreciated.

    OS X has not been officially issued on physical media since OS X 10.6 (arguably 10.7 was issued on some USB drives, but this was a non-standard approach for purchasing and installing it).
    To reinstall the OS, your system comes with a recovery partition that can be booted to by holding the Command-R keys immediately after hearing the boot chimes sound. This partition boots to the OS X tools window, where you can select options to restore from backup or reinstall the OS. If you choose the option to reinstall, then the OS installation files will be downloaded from Apple's servers.
    If for some reason your entire hard drive is damaged and even the recovery partition is not accessible, then your system supports the ability to use Internet Recovery, which is the same thing except instead of accessing the recovery boot drive from your hard drive, the system will download it as a disk image (again from Apple's servers) and then boot from that image.
    Both of these options will require you have broadband internet access, as you will ultimately need to download several gigabytes of installation data to proceed with the reinstallation.
    There are some options available for creating your own boot and installation DVD or external hard drive, but for most intents and purposes this is not necessary.
    The only "checklist" option I would recommend for anyone with a new Mac system, is to get a 1TB external drive (or a drive that is at least as big as your internal boot drive) and set it up as a Time Machine backup. This will ensure you have a fully restorable backup of your entire system, which you can access via the recovery partition for restoring if needed, or for migrating data to a fresh OS installation.

  • Networking "best practice" for setting up a farm

    Hi all.
    We would like to set an OracleVM farm, and I have a question about "best practice" for
    configuring the network. Some background:
    - The hardware I have is comprised of machines with 4 gig-eth NICs each.
    - The storage will be coming primarily from a backend NAS appliance (Netapp, FWIW).
    - We have already allocated a separate VLAN for management.
    - We would like to have HA capable VMs using OCFS2 (on top of NFS.)
    I'm trying to decide between 2 possible configurations. The first would keep physical separation
    between the mgt/storage networks and the DomU networks. The second would just trunk
    everything together across all 4 NICs, something like:
    Config 1:
    - eth0 - management/cluster-interconnect
    - eth1 - storage
    - eth2/eth3 => bond0 - 8021q trunked, bonded interfaces for DomUs
    Config 2:
    - eth0/1/2/3 => bond0
    Do people have experience or recommendation about the best configuration?
    I'm attracted to the first option (perhaps naively) because CI/storage would benefit
    from dedicated bandwidth and this configuration might also be more secure.
    Regards,
    Robert.

    user1070509 wrote:
    Option #4 (802.3ad) looks promising, but I don't know if this can be made to work across
    separate switches.It can, if your switches support cross-switch trunking. Essentially, 802.3ad (also known as LACP or EtherChannel on Cisco devices) requires your switch to be properly configured to allow trunking across the interfaces used for the bond. I know that the high-end Cisco and Juniper switches do support LACP across multiple switches. In the Cisco world, this is called MEC (Multichassis EtherChannel).
    If you're using low-end commodity-grade gear, you'll probably need to use active/passive bonds if you want to span switches. Alternatively, you could use one of the balance algorithms for some bandwitch increase. You'd have to run your own testing to determine which algorithm is best suited for your workload.
    The Linux Foundation's Net:Bonding article has some great information on bonding in general, particularly on the various bonding methods for high availability:
    http://www.linuxfoundation.org/en/Net:Bonding

  • Best setup for a non-Mac file server?

    I have a dual xeon server, with a SATA RAID5 I want to use as a file server in a cross-platform environment.
    *What I've tried and the issues...*
    At first I considered using Windows 2003 (Win2k3) but Services for Macintosh (SFM) is an older version of AFP and thus only supports 31 character filenames. With all our Macs supporting SMB/Samba/CIFS and Apple touting that "Macs and PCs can co-exist harmoniously on the same network" I figured I would give that a try.
    SMB doesn't work.
    Sure I can create a connection, but transferring files is a completely different story. I'm trying to backup application and system data, but companies such as Adobe and Apple have named some of their files with special characters that can't be transferred over SMB. I know NTFS doesn't support these characters, but I though a Linux box using SMB would work fine. It doesn't. It's the protocol which keeps me from transferring the data. I end up with the lovely error message of "You cannot copy some of these items to the destination because their names are too long or contain invalid characters for the destination..." (what's sad is, if you google for " because their names are too long or contain invalid characters for the destination" you only get 6 results.)
    So I thought I would give NFS a try. Apple says "Viewed from Mac OS X, [connecting via NFS] is just like connecting to an Apple or Windows server." No. It's not. NFS shares don't even show up in the Finder's Network listing. There are also a pile of other hurdles which are only tackled by savvy, command-line using users.
    So that leaves me with AFP. Win2k3 doesn't support filenames longer than 31 characters, and Win2k8 is dropping SFM altogether. Off to choose a *nix flavor, but that requires Netatalk. It hasn't been updated in years, it has many bad performance reviews... and most distros have removed it. I can download and install it. Oh, but that requires I get the kernel source files. Then I have to create an RPM an that's not working... now I'm several levels deep in trying to figure out how to get Netatalk working and I'm not even sure it will work.
    *What's the best setup for a non-Mac file server?*
    FreeNAS seems promising, but it's in alpha/beta and they have all sorts of warnings regarding potential data loss. Sure there's ExtremeZ-IP, but I really don't want to spend $675 do something Apple claims OS X can already do. I can put just about any non-Mac OS on this thing... what's the best way to set it up so it works?
    Thanks much.

    Rick may be right because although i didnt think of it before i tend to have notoriuosly long classnames for my php classes and i have used samba on occasion (when rsync is out of the question for one reason or another) and never had a problem. I use kubuntu (feisty at the moment )with an ext3 filesystem. if i have a chance this evening ill give it a try and see what happens.
    You could also possibly use FUSE to use an ssh filesystem for the shares... i don tknow how that would figure in your back up though.
    Also if worse comes to worse you could tar or dmg the the necessary files... just some thoughts.
    Ill be interested to know what you end up implementing....
    OH one last thought... Compile Darwin from source and use that as your server

  • Best software for doing full backups to ext. hard disk to to be used to restore system

    As a new PC owner I wanted to find out what the best software for doing full backups to external hard disk,  to to be used to restore the system if primary primary disk crashes. Im aware that Windows, Acronis and Paragon all have software that doing varying degress of back and resore functions.

    lasvideo,
    My favorites are ShadowProtect to do an "image backup" of boot drive and Beyond Compare for backing up files for all other "data" arrays. ShadowProtect does not seem to have much market share compared with Ghost and Acronis, but I really love it.
    What I like about ShadowProtect:
    - Extremely fast (complete image backup for 61GB of files on my boot SSD RAID 0 array to a RAID data array takes less than 3 minutes)
    - Can do image backups across a LAN; restores from boot CDROM and USB device
    - Allows for backups from one controller and restore of OS on a different controller and/or RAID configuration (has HAL tools; Hardware ? Layer I think)
    - Can "mount" an image backup as a drive letter and see (and copy) all directories, files, etc. that are contained in the archive
    What I like about Beyond Compare:
    - Can choose what to backup and what to ignore very easily (for example you can set it to ignore certain files or directory names to ignore and then save how you set things up as a "session")
    - Only needs to copy over files that are new or have changed since your last backup
    - Makes it very easy to selectively delete files from your backup
    Regards,
    Jim

  • Best Practice for a Print Server

    What is the best practice for having a print server serving over 25 printers 10 of which are colour lasers and the rest black and white lasers.
    Hardware
    At the moment we have one server 2Ghz Dual G5 with 4GB Ram and xserve RAID. The server is also our main Open directory server, with about 400+ clients.
    I want to order a new server and want to know the best type of setup for the optimal print server.
    Thanks

    Since print servers need RAM and spool space, but not a lot of processing power, I'd go with a Mac Mini packed with ram and the biggest HD you can get into it. Then load a copy of Xserver Tiger on it and configure your print server there.
    Another option, if you don't mind used equipment, is to pick up an old G4 or G5 Xserve, load it up with RAM and disk space, and put tiger on that.
    Good luck!
    -Gregg

  • Best Practice for SRST deployment at a remote site

    What is the best practice for a SRST deployment at a remote site? Should a separate router such as a 3800 series be deployed for telephony in addition to another router to be deployed for Data? Is there a need for 2 different devices?

    Hi Brian,
    This is typically done all on one ISR Router at the remote site :)There are two flavors of SRST. Here is the feature comparison;
    SRST Fallback
    This feature enables routers to provide call-handling support for Cisco Unified IP phones if they lose connection to remote primary, secondary, or tertiary Cisco Unified Communications Manager installations or if the WAN connection is down. When Cisco Unified SRST functionality is provided by Cisco Unified CME, provisioning of phones is automatic and most Cisco Unified CME features are available to the phones during periods of fallback, including hunt-groups, call park and access to Cisco Unity voice messaging services using SCCP protocol. The benefit is that Cisco Unified Communications Manager users will gain access to more features during fallback ****without any additional licensing costs.
    Comparison of Cisco Unified SRST and
    Cisco Unified CME in SRST Fallback Mode
    Cisco Unified CME in SRST Fallback Mode
    • First supported with Cisco Unified CME 4.0: Cisco IOS Software 12.4(9)T
    • IP phones re-home to Cisco Unified CME if Cisco Unified Communications Manager fails. CME in SRST allows IP phones to access some advanced Cisco Unified CME telephony features not supported in traditional SRST
    • Support for up to 240 phones
    • No support for Cisco VG248 48-Port Analog Phone Gateway registration during fallback
    • Lack of support for alias command
    • Support for Cisco Unity® unified messaging at remote sites (Distributed Exchange or Domino)
    • Support for features such as Pickup Groups, Hunt Groups, Basic Automatic Call Distributor (BACD), Call Park, softkey templates, and paging
    • Support for Cisco IP Communicator 2.0 with Cisco Unified Video Advantage 2.0 on same computer
    • No support for secure voice in SRST mode
    • More complex configuration required
    • Support for digital signal processor (DSP)-based hardware conferencing
    • E-911 support with per-phone emergency response location (ERL) assignment for IP phones (Cisco Unified CME 4.1 only)
    Cisco Unified SRST
    • Supported since Cisco Unified SRST 2.0 with Cisco IOS Software 12.2(8)T5
    • IP phones re-home to SRST router if Cisco Unified Communications Manager fails. SRST allows IP phones to have basic telephony features
    • Support for up to 720 phones
    • Support for Cisco VG248 registration during fallback
    • Support for alias command
    • Lack of support for features such as Pickup Groups, Hunt Groups, Call Park, and BACD
    • No support for Cisco IP Communicator 2.0 with Cisco Unified Video Advantage 2.0
    • Support for secure voice during SRST fallback
    • Simple, one-time configuration for SRST fallback service
    • No per-phone emergency response location (ERL) assignment for SCCP Phones (E911 is a new feature supported in SRST 4.1)
    http://www.cisco.com/en/US/prod/collateral/voicesw/ps6788/vcallcon/ps2169/prod_qas0900aecd8028d113.html
    These SRST hardware based restrictions are very similar to the number of supported phones with CME. Here is the actual breakdown;
    Cisco 880 SRST Series Integrated Services Router
    Up to 4 phones
    Cisco 1861 Integrated Services Router
    Up to 8 phones
    Cisco 2801 Integrated Services Router
    Up to 25 phones
    Cisco 2811 Integrated Services Router
    Up to 35 phones
    Cisco 2821 Integrated Services Router
    Up to 50 phones
    Cisco 2851 Integrated Services Router
    Up to 100 phones
    Cisco 3825 Integrated Services Router
    Up to 350 phones
    Cisco Catalyst® 6500 Series Communications Media Module (CMM)
    Up to 480 phones
    Cisco 3845 Integrated Services Router
    Up to 730 phones
    *The number of phones supported by SRST have been changed to multiples of 5 starting with Cisco IOS Software Release 12.4(15)T3.
    From this excellent doc;
    http://www.cisco.com/en/US/prod/collateral/voicesw/ps6788/vcallcon/ps2169/data_sheet_c78-485221.html
    Hope this helps!
    Rob

  • Best practice for RDGW placement in RDS 2012 R2 deployment

    Hi,
    I have been setting up a RDS 2012 R2 farm deployment and the time has come for setting up the RDGW servers. I have a farm with 4 SH servers, 2 WA servers, 2 CB servers and 1 LS.
    Farm works great for LAN and VPN users.
    Now i want to add two domain joined RDGW servers.
    The question is; I've read a lot on technet and different sites about how to set the thing up, but no one mentions any best practices for where to place them.
    Should i:
    - set up WAP in my DMZ with ADFS in LAN, then place the RDGW in the LAN and reverse proxy in
    - place RDGW in the DMZ, opening all those required ports into the LAN
    - place the RDGW in the LAN, then port forward port 443 into it from internet
    Any help is greatly appreciated.
    This posting is provided "AS IS" with no warranties or guarantees and confers no rights

    Hi,
    The deployment is totally depends on your & company requirements as many things to taken care such as Hardware, Network, Security and other related stuff. Personally to setup RD Gateway server I would not prefer you to select 1st option. But as per my research,
    for best result you can use option 2 (To place RDG server in DMZ and then allowed the required ports). Because by doing so outside network can’t directly connect to your internal server and it’s difficult to break the network by any attackers. A perimeter
    network (DMZ) is a small network that is set up separately from an organization's private network and the Internet. In a network, the hosts most vulnerable to attack are those that provide services to users outside of the LAN, such as e-mail, web, RD Gateway,
    RD Web Access and DNS servers. Because of the increased potential of these hosts being compromised, they are placed into their own sub-network called a perimeter network in order to protect the rest of the network if an intruder were to succeed. You can refer
    beneath article for more information.
    RD Gateway deployment in a perimeter network & Firewall rules
    http://blogs.msdn.com/b/rds/archive/2009/07/31/rd-gateway-deployment-in-a-perimeter-network-firewall-rules.aspx
    Hope it helps!
    Thanks.
    Dharmesh Solanki
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Best practice for the test environment  &  DBA plan Activities    Documents

    Dears,,
    In our company, we made sizing for hardware.
    we have Three environments ( Test/Development , Training , Production ).
    But, the test environment servers less than Production environment servers.
    My question is:
    How to make the best practice for the test environment?
    ( Is there any recommendations from Oracle related to this , any PDF files help me ............ )
    Also please , Can I have a detail document regarding the DBA plan activities?
    I appreciate your help and advise
    Thanks
    Edited by: user4520487 on Mar 3, 2009 11:08 PM

    Follow your build document for the same steps you used to build production.
    You should know where all your code is. You can use the deployment manager to export your configurations. Export customized files from MDS. Just follow the process again, and you will have a clean instance not containing production data.
    It only takes a lot of time if your client is lacking documentation or if you re not familiar with all the parts of the environment. What's 2-3 hours compared to all the issues you will run into if you copy databases or import/export schemas?
    -Kevin

  • Best practice for exporting from iMovie '08 to iDVD

    I am looking to find out what is the best practice for exporting from iMovie '08 to iDVD. I have read the other postings that give the basic howto (export to Media Browser then select the video in iDVD). However, my question is a little more technical. I have 1080i HD projects. I am interested in burning them to DVD in the best possible quality. What setting should I be using when I publish to Media Browser?
    I am wondering about quality loss due to more than one conversion/compression. I suspect that when I export to the Media Browser then this is occurring. If I am not mistaken iMovie is using something like H.264 for this. Then, when I run iDVD I suspect it will it do another conversion/compression, I think to get to MPEG2. Not only could this result in a loss of quality but also it will take extra time. I am interested to know what others think about this.
    Finally, I am looking to create DVDs for a lot of video. I am wondering if there are any USB or firewire hardware devices out there that could speed up the compression. I use the Elgato Turbo.264 when I want to encode to H.264 but I wonder if there is something similar for DVD creation.
    Thanks in advance.

    the standards for videoDVD are 720x480, and usually mpeg2 encoded..
    so, your HiDef project HAS to be 'downsampled' somehow..
    I would Export with Qucktime/apple intermediate => which is the 'format' your project is allready, and you avoid any useless 'inbetween encoding'..
    iDVD will 'swallow' this huge export file - don't mind: iDVD cares for length, not size.
    iDVD will then convert into DVD-standards..
    you can 'raise' quality, by using projects <60min - this sets iDVD automatically to highest technical possible bitrate
    hint: judge pic quality on a DVDplayer + TV.. not on your computer (DVDs are meant for TVdelivery)

  • Best Settings for Audigy 2 ZS Recording Out to S-VHS or DVD Recor

    I?m recording some very machine intensi've video & sound to outboard devices. I'm using these outboard devices so that the recording process itself does not impact the demonstration performance.
    Concerning the sound out, I?m using the line & 2 out of the card, and running it through a pro level audio board (Mackie) for any adjustments & monitoring before taking the line out of the board into the recording device. When returning the sound for monitoring, I?m also running from the external device directly to the board and directly from there to either headphones or control room monitors. Thus, I?m using the Audigy for initial sound generation. So far, I've not really made any ajustments at the board other than level or balance. I like to keep things pretty flat in recording.
    I?ve looked around for recommendations as to the best settings for the Audigy card from its internal software for this situation, but find no ready reference. My concern is that I have the ?hottest? signal from this device, while being very concerned with audio quality. I?m looking for both ?punch? and ?clarity? of signal.
    Is there a document or source that would give me this direction? While I can move from this SB card to other more studio based cards, I?d like to keep the demo product in line with what the actual target audience would have in their machines, and be able to replicate at their workstation.
    For example, does the card function best at 00% (0db) output, or 50% (-5db) output?
    Or some setting in between...most likely...
    Or, is there some other setting I?m not aware of?
    I have what I believe are the latest card drivers and software suite, XP Pro, with every possible OS and other machine hardware update available.
    Thanks,
    Rich

    I?m recording some very machine intensi've video & sound to outboard devices. I'm using these outboard devices so that the recording process itself does not impact the demonstration performance.
    Concerning the sound out, I?m using the line & 2 out of the card, and running it through a pro level audio board (Mackie) for any adjustments & monitoring before taking the line out of the board into the recording device. When returning the sound for monitoring, I?m also running from the external device directly to the board and directly from there to either headphones or control room monitors. Thus, I?m using the Audigy for initial sound generation. So far, I've not really made any ajustments at the board other than level or balance. I like to keep things pretty flat in recording.
    I?ve looked around for recommendations as to the best settings for the Audigy card from its internal software for this situation, but find no ready reference. My concern is that I have the ?hottest? signal from this device, while being very concerned with audio quality. I?m looking for both ?punch? and ?clarity? of signal.
    Is there a document or source that would give me this direction? While I can move from this SB card to other more studio based cards, I?d like to keep the demo product in line with what the actual target audience would have in their machines, and be able to replicate at their workstation.
    For example, does the card function best at 00% (0db) output, or 50% (-5db) output?
    Or some setting in between...most likely...
    Or, is there some other setting I?m not aware of?
    I have what I believe are the latest card drivers and software suite, XP Pro, with every possible OS and other machine hardware update available.
    Thanks,
    Rich

Maybe you are looking for

  • PDF Error in Performance Management

    Hi When launching the Performance Management link from Manager Self Service (Role) --> Talent Management --> Performance Management I am getting the PDF error. Seems the page is trying to open the form in PDF, instead I want it to be displayed in htm

  • Saving an output file in another location

    Hi friends, Is it possible that a specific concurrent report's output file be saved in a different location, other than $<PROD>_TOP/$APPLOUT when executed? Is it possible to configure that? Thanks in advance. Jose Melendez

  • Inserting imageicon in jtable

    Hi, I want to insert imageicon in Jtable using AbstractTableModel. Please help me out from this.

  • List the base tables of a view

    Hi All, I have a view but I do not have any information about the base tables and the query that was used to set up this view. Is there a way to atleast list the base tables on which the view was created! Thanks in advance! ~

  • WCS 6.0 Reporting

    Does anyone know what the definition is of a "Number of Unique User"  in the Client Summary report would be as opposed to the "Number of Users" ?