Gentoo Linux and Microsoft Failover Clusters / Hyper-V

Hello,
Hoping there are a few people on the boards familiar with running Gentoo Linux guests under Microsoft FailOver Cluster / Hyper-V hosts.
I have four Gentoo Linux guest VMs (running kernel 3.12.21-r1) running under the Microsoft Failover Cluster system with Hyper-V as the host. All of the Hyper-V drivers are built into the kernel (including the utilities and balloon drivers) and generally they
run without issue.
For several months now, however, I have been having strange issues with them. Essentially they stop responding to network requests after random intervals. However, these intervals aren't a few minutes or hours from each other; more like days or even weeks before
one of them will stop responding on the network side.
The funny thing is that the VMs themselves on the console side still responds. However, if I issue a reboot command on the externally non-responsive VM, the system will eventually get to a stage where all of the services are stopped and then hangs right after
the "mounting remaining system ro" line (or something like that).
The Failover Cluster Manager then reports that the system is "Stopping" but the system never reboots.
I have to completely restart the HOST system so that either (A) the VM in question transfers to another host and starts responding again or (B) when the HOST comes back up I can work with the VM again.
This *ONLY* happens on the Gentoo Linux guest VMs and not my Windows VMs.
Wondering if anyone has hints on this.
Thank you for your time.
Regards, Christopher K.

Hi ckoeber,
Hyper-V has support most of Linux distribution but not all because it’s have so many distribution, until now the Gentoo Linux not in the support list, you can refer the following
KB to know more about the detail:
Linux Virtual Machines on Hyper-V
http://technet.microsoft.com/en-US/library/dn531030
Hope this helps.
We
are trying to better understand customer views on social support experience, so your participation in this
interview project would be greatly appreciated if you have time.
Thanks for helping make community forums a great place.

Similar Messages

  • Data Guard Broker and Cold Failover clusters

    Hi,
    I wanted to use Data Guard Broker to control primary/standby systems on AIX clusters but found that without Oracle clusterware it is not supported on failover of a database to another node (unless you drop and recreate the configuration which in my opinion is not support at all!).
    The 11g documentation states that DG Broker offers support for single instance databases configured for HA using Oracle Clusterware and "cold failover clusters".
    Does anyone know whether this support for cold failover clusters in 11g means "proper support" i.e the configuration detects that the hostname has changed and automatically renames it so it continues to work?, or is the support in 11g the same as that in 10g?.
    Thanks,
    Andy

    Hi Jan,
    We already use virtual hostnames.
    When I set up the configuration the hostname for the databases default to the server name. The only way I know of changing this is to disable the configuration, then use the "edit database...set property" command but when I enable the configuration again it reverts back to its previous value.
    regards,
    Andy

  • J4400, linux and heartbeat failover

    Hi all
    I am contemplating getting a sun J4400, with two x4150 servers to make a nfs server which has failover to a standby server in case the primary fails, using Linux and heartbeat
    I have read that you can connect two servers to a J-series array, but I have the following questions:
    1. Does the J4400 have any disk management built-in or must all be handled by the raid adapters on the connected servers.
    2. If both servers are connected to the SAS connections, do they both have access simultaneously to the drives, or is one blocked from access at the physical level?
    3. If they are blocked, are the raid adapters in the two servers aware of each other to ensure that only one server can see the drives at any time.
    4. If they are blocked, if the primary server fails, how would its raid card 'release' the drives for the other one to be able to take them over.
    Anything else I should be aware of for this scenario?

    Hi ckoeber,
    Hyper-V has support most of Linux distribution but not all because it’s have so many distribution, until now the Gentoo Linux not in the support list, you can refer the following
    KB to know more about the detail:
    Linux Virtual Machines on Hyper-V
    http://technet.microsoft.com/en-US/library/dn531030
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Failover clustering Hyper-V + load balancing

    I have successfully setup a number of Hyper-V (5) on Server 2008 R2 fail-over cluster environment with two nodes connected to a SAN using MS iSCSI Target 3.3 with cluster shared volume and all is working well.
    Now each Hyper-v is run by node-A or node-B. is there a way to have both nodes process all Hyper-Vs at once? 
    Mike

    Yes you are right I was not using the proper term. I have Hyper-V installed on both nodes and added the Hyper-V service in the Fail-over Clustering service. I can run the VMs on either nodes and distribute the 5 VMs but instead of selecting which node to
    run each VM I would like to run all 5 VMs in one environment where both servers (nodes) can manage the processes of all 5 VMs. In other words each server have 64GB RAM and 16 core CPUs, I would like to have my 5 VMs processed by 128GBRAM and 32 core CPU. My
    understanding is I am looking at Load balancing, I have seen how to do this with a web application but couldn't find anything on Load Balancing VMs.
    You need to configure CSV on top of shared LUN, put your VM images there and run VMs on every host having access to LUN / CSV. See:
    http://technet.microsoft.com/en-us/library/dd630633(v=ws.10).aspx
    http://technet.microsoft.com/en-us/library/gg610692.aspx
    Load balancing is another thing, you can use Live Migration to move VMs from one host to another (no downtime). See:
    http://blogs.technet.com/b/iftekhar/archive/2010/01/18/microsoft-virtualization-hyper-v-live-migration-network-configuration.aspx
    You can configure HA so one VM would re-boot on another node if primary host is down (minimal downtime). See:
    http://blogs.technet.com/b/canitpro/archive/2011/04/18/creating-ha-vms-for-hyper-v-with-failover-clustering-using-free-microsoft-iscsi-target-3-3.aspx
    You can configure guest VM cluster so VM would transparently failover to another host if primary one is down (no downtime). See:
    http://blogs.technet.com/b/mghazai/archive/2009/12/12/hyper-v-guest-clustering-step-by-step-guide.aspx
    http://blogs.technet.com/b/keithmayer/archive/2013/03/21/virtual-machine-guest-clustering-with-windows-server-2012-become-a-virtualization-expert-in-20-days-part-14-of-20.aspx#.Uss8YvRdWSo
    Hope this helped :)
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • How to schedule Cluster logs to be generated for Microsoft Failover Clusters 2008 R2

    Hi, As per my understanding we always have to generate Cluster logs manually on the cluster nodes to get these generated.
    Is there any way we can schedule Cluster Logs to be generated every time so that it would be easy for us to analyze the issue?
    Kevin

    Hi Kevin-Steve,
    Are you trying to realize the cluster log generated automatically with schedule? As far as I know there don’t have this function in failover cluster, but you can use PowerShell
    cmdlet to create a schedule task.
    The related article:
    How to create the cluster.log in Windows Server 2008 Failover Clustering
    http://blogs.msdn.com/b/clustering/archive/2008/09/24/8962934.aspx
    Understanding the Cluster Debug Log in 2008
    http://blogs.technet.com/b/askcore/archive/2010/04/13/understanding-the-cluster-debug-log-in-2008.aspx
    The schedule task create related thread:
    Power Shell Script to create basic windows schedule task
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/e2b665a5-e505-43d4-8cea-0d992a2c3917/power-shell-script-to-create-basic-windows-schedule-task
    I’m glad to be of help to you!
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Handheld Computing with Linux and Microsoft.

    I was persuing Best Buy's web site and I saw this
    It's nice to see more commercial products with Linux on them.  This seems to be a cool idea, in an attempt to combat Microsoft's new handheld which has been rumored..oragiami.  http://www.origamiproject.com/1/

    And they're doing the right thing re GPL/OSS/etc - full details here.
    Wouldn't mind getting hold of one of these.....

  • Different behaviour of file vis under linux and microsoft

    I am porting software written under Windows labview to Linux Labview (6.1). When using the Open/Create/Replace vi in windows if I use a pattern of "*.abc" and enter "123" in the dialogue box a file called "123.abc" will be created which is what I would expect. The same operation under Linux will create a file called "123" without the extension. Is this normal and is there a way of making the linux version behave in the same way as the windows version?

    Sorry about the delay - vacation.
    There appears to be no check box under Linux to use native file dialogs although it does indeed exist under windows. In fact the linux version does use linux file dialogs by default. It seems that the only options are to force the windows version to behave like the the linux version - which is counter intuitive, or write a new file dialog vi for linux.

  • Hyper-v Failover clustering for SQL Server lab

    Hi,
    After 20yrs years of IT software experience, now I am realizing that I don't know much about hyperV clustering. :) I am trying to setup a lab. I badly need help on configuring network adapters
    to make the machines accessible from each other. Please see details below:
    I have a laptop with 8gb ram with windows 8.1 hyper-V. I created 4 VMs with Windows 2008r2 on it and further I intend having SQL Server Failover Clustering. VM details are as below:
    DC - Domain Controller
    SanStorage - MS iSCSI target
    Win1 - Primary node in the cluster that will have SQL Server instance
    Win2 - Secondary node in the cluster that will have SQL Server instance.
    The laptop has 1 Qualcomm network adapter (generally used for connecting Lan cable) and a Broadcomm (wifi) adapter. I intend setting up three virtual adapters
    First for regular communication
    Second for Data (SQL Server Data to be stored on SCSI)
    Third for HeartBit
    Normally, (on my laptop) I connect to internet using Broadcomm (Wi-Fi) and want to ensure that whenever needed:
    Run this lab even when Lan or wifi is NOT connected. If one of them is mandatory then even that could be the option.
    I should be able to use internet within all VMs as well- though this is good to have requirement.
    Internet needs to be available on laptop at least. Even if it is unavailable in VMs that is ok.
    Now the actual problem.
    I DON'T KNOW how many virtual internal/external/private switches to be
    defined and assigned to individual VMs. What should be the IPs, Subnets, DNS, Gateways defined on those?
    I was able to setup the stuff up-to certain extend and then it stopped working for unknown reasons. I have VMs ready. Next what should I do? Steps...
    setup network adapters in each VM
    setup DNS on DC
    Add remaining 3 VMs in domain
    Define Win1 and Win2 as cluster
    Configure Target
    Configure initiators and access the target
    e. Install SQL Server
    Is above correct?
    My queries probably demands a separate article. But I have already referred many articles and couldn't figure it out so posting it here. I sincerely request your help!
    PS – Even after disabling firewall (with and without having DNS setup) I am unable to ping VMs from each other.
    Regards,
    ThingToLearn
    Thanks, Pravin

    Hi,
    You need create another guest vm as the NAT device, attach the NAT external to the Wifi adapter or cable adapter, then you can set all your vms in internal network
    vswitch.
    The related third party article:
    Using Wireless with Hyper-V
    http://sqlblog.com/blogs/john_paul_cook/archive/2008/03/23/using-wireless-with-hyper-v.aspx
    Hope this helps.
    *** This response contains a reference to a third party World Wide Web site. Microsoft is providing this information as a convenience to you. Microsoft does not control
    these sites and has not tested any software or information found on these sites; therefore, Microsoft cannot make any representations regarding the quality, safety, or suitability of any software or information found there. There are inherent dangers in the
    use of any software found on the Internet, and Microsoft cautions you to make sure that you completely understand the risk before retrieving any software from the Internet. ***
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Hyper-V Failover Clustering MAC Address Question

    Hello,
    Can someone please explain how MAC address assignment works in a hyper-v cluster? I have a couple machines that need to have a static MAC address assigned and want to know how the cluster handles it when the VM is Live migrated to a new host. Thanks in advance.

    Hi ice2921,
    Yes, after export and import to another Hyper-V host, the dynamic MAC address of the VM changes
      after export and import to another Hyper-V host, the dynamic MAC address of the VM changes, this is by design.
    As Alexey offered article, to have a dedicated Mac Address for the VMs they should be set to static MAC address, it is normal the dynamic assign to vm when the all the Hyper-v
    host as there own MAC address range , this is by design of Hyper V and failover clustering when using Dynamic MAC addresses for VMs.
    If you are using 2008R2 Hyper-V failover and you found it often lost the network connection and configuration, please install the hotfix.
    Reinstalled the integration services
    The network connection of a running Hyper-V virtual machine may be lost under heavy outgoing network traffic on a computer that is running Windows Server 2008 R2 SP1
    http://support.microsoft.com/kb/2263829
    I’m glad to be of help to you!
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • 3 Node hyper-V 2012 R2 Failover Clustering with Storage spaces on one of the Hyper-V hosts

    Hi,
    We have 3x Dell R720s with 5x 600GB 3.5 15K SAS and 128 GB RAM each. Was wondering if I could setup a Fail-over Hyper-V 2012 R2 Clustering with these 3 with the shared storage for the CSV being provided by one of the Hyper-V hosts with storage spaces installed
    (Is storage spaces supported on Hyper-V?) Or I can use a 2-Node Failover clustering and the third one as a standalone Hyper-V or Server 2012 R2 with Hyper-V and storage spaces.  
    Each Server comes with QP 1G and a DP10G nics so that I can dedicate the 10G nics for iSCSI
    Dont have a SAN or a 10G switch so it would be a crossover cable connection between the servers.
    Most of the VMs would be Non-HA. Exchange 2010, Sharepoint 2010 and SQL Server 2008 R2 would be the only VMS running as HA-VMs. CSV for the Hyper-V Failover cluster would be provided by the storage spaces.

    I thought I was tying to do just that with 8x600 GB RAID-10 using the H/W RAID controller (on the 3rd Server) and creating CSVs out of that space so as to provide better storage performance for the HA-VMs.
    1. Storage Server : 8x 600GB RAID-10 (For CSVs to house all HA-VMs running on the other 2 Servers) It may also run some local VMs that have very little disk I/O
    2. Hyper-V-1 : Will act has primary HYPER-V hopst for 2x Exchange and Database Server HA-VMs (the VMDXs would be stored on the Storage Servers CSVs on top of the 8x600GB RAID-10). May also run some non-HA VMs using the local 2x600 GB in RAID-1 
    3. Hyper-V-2 : Will act as a Hyper-V host when the above HA-VMs fail-over to this one (when HYPER-V-1 is down for any reason). May also run some non-HA VMs using the local 2x600 GB in RAID-1 
    The single point of failure for the HA-VMs (non HA-VMs are non-HA so its OK if they are down for some time) is the Storage Server. The Exchange servers here are DAG peers to the Exchange Servers at the head office so in case the storage server mainboard
    goes down (disk failure is mitigated using RAID, other components such as RAM, mainboard may still go but their % failure is relatively low) the local exchange servers would be down but exchange clients will still be able to do their email related tasks using
    the HO Exchange servers.
    Also they are under 4hr mission critical support including entire server replacement within the 4 hour period. 
    If you're OK with your shared storage being a single point of failure then sure you can proceed the way you've listed. However you'll still route all VM-related I/O over Ethernet which is obviously slower then running VMs from DAS (with or without virtual SAN
    LUN-to-LUN replication layer) as DAS has higher bandwidth and smaller latency. Also with your scenario you exclude one host from your hypervisor cluster so running VMs on a pair of hosts instead of three would give you much worse performance and resilience:
    with 1 from 3 physical hosts lost cluster would be still operable and with 1 from 2 lost all VMs booted on a single node could give you inadequate performance. So make sure your hosts would be insanely underprovisioned as every single node should be able to
    handle ALL workload in case of disaster. Good luck and happy clustering :)
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Advice Requested - High Availability WITHOUT Failover Clustering

    We're creating an entirely new Hyper-V virtualized environment on Server 2012 R2.  My question is:  Can we accomplish high availability WITHOUT using failover clustering?
    So, I don't really have anything AGAINST failover clustering, and we will happily use it if it's the right solution for us, but to be honest, we really don't want ANYTHING to happen automatically when it comes to failover.  Here's what I mean:
    In this new environment, we have architected 2 identical, very capable Hyper-V physical hosts, each of which will run several VMs comprising the equivalent of a scaled-back version of our entire environment.  In other words, there is at least a domain
    controller, multiple web servers, and a (mirrored/HA/AlwaysOn) SQL Server 2012 VM running on each host, along with a few other miscellaneous one-off worker-bee VMs doing things like system monitoring.  The SQL Server VM on each host has about 75% of the
    physical memory resources dedicated to it (for performance reasons).  We need pretty much the full horsepower of both machines up and going at all times under normal conditions.
    So now, to high availability.  The standard approach is to use failover clustering, but I am concerned that if these hosts are clustered, we'll have the equivalent of just 50% hardware capacity going at all times, with full failover in place of course
    (we are using an iSCSI SAN for storage).
    BUT, if these hosts are NOT clustered, and one of them is suddenly switched off, experiences some kind of catastrophic failure, or simply needs to be rebooted while applying WSUS patches, the SQL Server HA will fail over (so all databases will remain up
    and going on the surviving VM), and the environment would continue functioning at somewhat reduced capacity until the failed host is restarted.  With this approach, it seems to me that we would be running at 100% for the most part, and running at 50%
    or so only in the event of a major failure, rather than running at 50% ALL the time.
    Of course, in the event of a catastrophic failure, I'm also thinking that the one-off worker-bee VMs could be replicated to the alternate host so they could be started on the surviving host if needed during a long-term outage.
    So basically, I am very interested in the thoughts of others with experience regarding taking this approach to Hyper-V architecture, as it seems as if failover clustering is almost a given when it comes to best practices and high availability.  I guess
    I'm looking for validation on my thinking.
    So what do you think?  What am I missing or forgetting?  What will we LOSE if we go with a NON-clustered high-availability environment as I've described it?
    Thanks in advance for your thoughts!

    Udo -
    Yes your responses are very helpful.
    Can we use the built-in Server 2012 iSCSI Target Server role to convert the local RAID disks into an iSCSI LUN that the VMs could access?  Or can that not run on the same physical box as the Hyper-V host?  I guess if the physical box goes down
    the LUN would go down anyway, huh?  Or can I cluster that role (iSCSI target) as well?  If not, do you have any other specific product suggestions I can research, or do I just end up wasting this 12TB of local disk storage?
    - Morgan
    That's a bad idea. First of all Microsoft iSCSI target is slow (it's non-cached @ server side). So if you really decided to use dedicated hardware for storage (maybe you do have a reason I don't know...) and if you're fine with your storage being a single
    point of failure (OK, maybe your RTOs and RPOs are fair enough) then at least use SMB share. SMB at least does cache I/O on both client and server sides and also you can use Storage Spaces as a back end of it (non-clustered) so read "write back flash cache
    for cheap". See:
    What's new in iSCSI target with Windows Server 2012 R2
    http://technet.microsoft.com/en-us/library/dn305893.aspx
    Improved optimization to allow disk-level caching
    Updated
    iSCSI Target Server now sets the disk cache bypass flag on a hosting disk I/O, through Force Unit Access (FUA), only when the issuing initiator explicitly requests it. This change can potentially improve performance.
    Previously, iSCSI Target Server would always set the disk cache bypass flag on all I/O’s. System cache bypass functionality remains unchanged in iSCSI Target Server; for instance, the file system cache on the target server is always bypassed.
    Yes you can cluster iSCSI target from Microsoft but a) it would be SLOW as there would be only active-passive I/O model (no real use from MPIO between multiple hosts) and b) that would require a shared storage for Windows Cluster. What for? Scenario was
    usable with a) there was no virtual FC so guest VM cluster could not use FC LUs and b) there was no shared VHDX so SAS could not be used for guest VM cluster as well. Now both are present so scenario is useless: just export your existing shared storage without
    any Microsoft iSCSI target and you'll be happy. For references see:
    MSFT iSCSI Target in HA mode
    http://technet.microsoft.com/en-us/library/gg232621(v=ws.10).aspx
    Cluster MSFT iSCSI Target with SAS back end
    http://techontip.wordpress.com/2011/05/03/microsoft-iscsi-target-cluster-building-walkthrough/
    Guest
    VM Cluster Storage Options
    http://technet.microsoft.com/en-us/library/dn440540.aspx
    Storage options
    The following tables lists the storage types that you can use to provide shared storage for a guest cluster.
    Storage Type
    Description
    Shared virtual hard disk
    New in Windows Server 2012 R2, you can configure multiple virtual machines to connect to and use a single virtual hard disk (.vhdx) file. Each virtual machine can access the virtual hard disk just like servers
    would connect to the same LUN in a storage area network (SAN). For more information, see Deploy a Guest Cluster Using a Shared Virtual Hard Disk.
    Virtual Fibre Channel
    Introduced in Windows Server 2012, virtual Fibre Channel enables you to connect virtual machines to LUNs on a Fibre Channel SAN. For more information, see Hyper-V
    Virtual Fibre Channel Overview.
    iSCSI
    The iSCSI initiator inside a virtual machine enables you to connect over the network to an iSCSI target. For more information, see iSCSI
    Target Block Storage Overviewand the blog post Introduction of iSCSI Target in Windows
    Server 2012.
    Storage requirements depend on the clustered roles that run on the cluster. Most clustered roles use clustered storage, where the storage is available on any cluster node that runs a clustered
    role. Examples of clustered storage include Physical Disk resources and Cluster Shared Volumes (CSV). Some roles do not require storage that is managed by the cluster. For example, you can configure Microsoft SQL Server to use availability groups that replicate
    the data between nodes. Other clustered roles may use Server Message Block (SMB) shares or Network File System (NFS) shares as data stores that any cluster node can access.
    Sure you can use third-party software to replicate 12TB of your storage between just a pair of nodes to create a fully fault-tolerant cluster. See (there's also a free offering):
    StarWind VSAN [Virtual SAN] for Hyper-V
    http://www.starwindsoftware.com/native-san-for-hyper-v-free-edition
    Product is similar to what VMware had just released for ESXi except it's selling for ~2 years so is mature :)
    There are other guys doing this say DataCore (more playing for Windows-based FC) and SteelEye (more about geo-cluster & replication). But you may want to give them a try.
    Hope this helped a bit :) 
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Financial Reporting and Microsoft OLAP - issues with MDX

    We have installed Oracle/Hyperion Financial Reporting 11 for a client. They are using a Microsoft Analysis Services 2000 'cube' as the data source. The cube has an Account Dimension with 9 levels in its hierarchy. We have found Financial Reporting truncates the MDX when querying Accounts at the lowest level (around 100 characters in the length of the MDX string) or when querying multiple account groups at a higher level (and displaying descendants). Does anyone have any workarounds or solutions? Is there a better way of doing things? Is there any documentaion on the limitations of using FR over MSOLAP? We have have built the cube with an Account No as the Short Name (Key) and the Account Description as the Name but FR insists on only displaying (and using in the MDX) the long Account Descriiption - whereas using Essabase you can drill down on the Member Names [Account No] but display the Alias [Account Name]

    Hi,
    I don’t found the similar issue, please confirm your hardware compliant with the server 2012, update your BIOS at last release.
    Please rerun the validation test then post the error or warning part, that report will quickly locate the issue.
    Please try to install the recommended hotfixes and updates for Windows Server 2012-based failover clusters then try again.
    Recommended hotfixes and updates for Windows Server 2012-based failover clusters
    http://support.microsoft.com/kb/2784261/en-us
    More information:
    Windows Server Catalog
    http://www.windowsservercatalog.com/default.aspx
    Hyper-V Virtual Fibre Channel Overview
    http://technet.microsoft.com/en-us/library/hh831413.aspx
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Failover Cluster Hyper-V Storage Choice

    I am trying to deploy a 2 nodes Hyper-V Failover cluster in a closed environment.  My current setup is 2 servers as hypervisors and 1 server as AD DC + Storage server.  All 3 are running Windows Server 2012 R2.
    Since everything is running on Ethernet, my choice of storage is between iSCSI and SMB3.0 ?
    I am more inclined to use SMB3.0 and I did find some instructions online about setting up a Hyper-V cluster connecting to a SMB3.0 File server cluster.  However, I only have budget for 1 storage Server.  Is it a good idea to choice SMB over iSCSI
    in this scenario?  ( where there is only 1 storage for the Hyper-V Cluster ). 
    What do I need to pay attention to in this setup?  Except some unavoidable single points of failures. 
    In the SMB3.0 File server cluster scenario that I mentioned above, they had to use SAS drives for the file server cluster (CSV).  I am guessing in my scenario, SATA drives should be fine, right?

    "I suspect that Starwind solution achieves FT by running shadow copies of VMs on the partner Hypervisor"
    No, it does not run shadow VMs on the partner hypervisor.  Starwind is a product in a family known as 'software defined storage'.  There are a number of solutions on the market.  They all provide a similar service in that they allow for the
    use of local storage, also known as Direct Attached Storage (DAS), instead of external, shared storage for clustering.  Each of these products provides some method to mirror or 'RAID' the storage among the nodes in the software defined storage. 
    So, yes, there is some overhead to ensure data redundancy, but none of this class of product will 'shadow' VMs on another node.  Products like Starwind, Datacore, and others are nice entry points to HA without the expense of purchasing an external storage
    shelf/array of some sort because DAS is used instead.
    1) "Software Defined Storage" is a VERY wide term. Many companies use it for solutions that DO require actual hardware to run on. Say Nexenta claims they do SDS and they need a separate physical servers running Solaris and their (Nexenta) storage app. Microsoft
    we all love so much because they give us infrastructure we use to make our living also has Clustered Storage Spaces MSFT tells is a "Software Defined Storage" but they need physical SAS JBODs, SAS controllers and fabric to operate. These are hybrid software-hardware
    solutions. More pure ones don't need any hardware but they still share actual server hardware with hypervisor (HP VSA, VMware Virtual SAN, oh, BTW, it does require flash to operate so it's again not pure software thing). 
    2) Yes there are number of solutions but devil is in details. Technically all virtualization world is sliding away from ancient way of VM-running storage virtualization stacks to ones being part of a hypervisor (VMware Virtual Storage Appliance replaced
    with VMware Virtual SAN is an excellent example). So talking about Hyper-V there are not so many companies who have implemented VM-less solutions. Except the ones you've named it's also SteelEye and that's all probably (Double-Take cannot replicate running
    VMs effectively so cannot be counted). Running storage virtualization stack as part of a Hyper-V has many benefits compared to VM-running stuff:
    - Performance. Obviously kernel-space running DMA engines (StarWind) and polling driver model (DataCore) are faster in terms of latency and IOPS compared to VM-running I/O all routed over VMBus and emulated storage and network hardware.
    - Simplicity. With native apps it's click and install. With VMs it's UNIX management burden (BTW, who will update forked-out Solaris VSA is running on top of? Sun? Out of business. Oracle? You did not get your ZFS VSA from Oracle. Who?) and always "hen and
    chicken" issue. Cluster starts, it needs access to shared storage to spawn a VMs but VMs are inside a VM VSA that need to be spawned. So first you start storage VMs, then make them sync (few terabytes, maybe couple of hours to check access bitmaps for volumes)
    and only after that you can start your other production VMs. Very nice!
    - Scenario limitations. You want to implement a SCV for Scale-Out File Servers? You canont use HP VSA or StorMagic because SoFS and Hyper-V roles cannot mix on the same hardware. To surf SMB 3.0 tide you need native apps or physical hardware behind. 
    That's why current virtualization leader VMware had clearly pointed where these types of things need to run - side-by-side with hypervisor kernel.
    3) DAS is not only cheaper but also faster then SAN and NAS (obviously). So sure there's no "one size fits all" but unless somebody needs a a) very high LUN density (Oracle or huge SQL database or maybe SAP) and b) very strict SLAs (friendly telecom company
    we provide Tier2 infrastructure for runs cell phone stats on EMC, $1M for a few terabytes. Reason? EMC people have FOUR units like that marked as a "spare" and have requirement to replace failed one in less then 15 minutes) there's no point to deploy hardware
    SAN / NAS for shared storage. SAN / NAS is an sustained innovation and Virtual SAN is disruptive. Disruptive comes to replace sustained for 80-90% of business cases to allow sustained live in a niche deployments. Clayton Christiansen\s "Innovator's Dilemma".
    Classic. More here:
    Disruptive Innovation
    http://en.wikipedia.org/wiki/Disruptive_innovation
    So I would not consider Software Defined Storage as a poor-mans HA or usable to Test & Development only. Thing is ready for prime time long time ago. Talk to hardware SAN VARs if you have connections: How many stand-alone units did they sell to SMBs
    & ROBO deployments last year?
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Microsoft Teredo Tunneling Adapter and Microsoft 6to4 Adapter

    Hi ,
    as per best practice of Cluster Hyper-v All drives on all nodes should be the same version
    But I found network drivers "Microsoft Teredo Tunneling Adapter and Microsoft 6to4 Adapter" are installed one some nodes and not installed on other nodes
    IP Ver 6 is enabled on all network drivers
    So What should I do and How ?  Install the drivers or uninstall it  and how to do that ?
    thanks in advance.
    Ramy

    Hi Ramy,
    "So What should I do and How ?  Install the drivers or uninstall it  and how to do that ?"
    Teredo is an IPv6 transition technology that provides address assignment and host-to-host automatic tunneling for unicast IPv6 traffic when IPv6/IPv4 hosts are located behind one or multiple IPv4 network address translators (NATs).
    You may keep the default setting .
    If you want to create a hyper-v cluster , please refer to the step-by-step guide :
    http://technet.microsoft.com/en-us/library/cc732181(v=WS.10).aspx
    "Before you create the cluster, we strongly recommend that you run a full validation test of your configuration. Validation helps you to confirm that the configuration of your servers, network, and storage meets a set of specific requirements for failover
    clusters. You can validate either an existing cluster or one or more servers that are not yet clustered."
    This step will help you to check cluster configuration .
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • SCCM 2007 Distribution Point Share on a Failover Clustered File Server.

    Is SCCM 2007 Distribution Point Share supported on a Windows 2008 R2 Failover Clustered File Server? I can get the packages to copy to the DP Share. But if I create them on a non clustered server it works fine. Are there any special settings that need top
    be configured for clustered shares?
    Thanks,
    Steve
    Steve

    Hi Steve,
    As far as I know this scenario is not supported scenario, In Windows Server® 2008 R2, the Cluster Shared Volumes feature included in failover clustering is only supported for
    use with the Hyper-V server role.
    The creation, reproduction, and storage of files on Cluster Shared Volumes that were not created for the Hyper-V role, including any user or application data stored under the
    ClusterStorage folder of the system drive on every node, are not supported and may result in unpredictable behavior, including data corruption or data loss on these shared volumes. Only files that are created for the Hyper-V role can be stored on Cluster Shared
    Volumes. An example of a file type that is created for the Hyper-V role is a Virtual Hard Disk (VHD) file.
    More information:
    Deploying Cluster Shared Volumes (CSV) in Windows Server 2008 R2 Failover Clustering
    http://blogs.msdn.com/b/clustering/archive/2009/02/19/9433146.aspx
    Backing Up Cluster Shared Volumes in a Failover Cluster in Windows Server 2008 R2
    http://technet.microsoft.com/en-us/library/ff182356(WS.10).aspx
    I’m glad to be of help to you!
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

Maybe you are looking for

  • Waiting for Task Completion from multiple users

    iam using Weblogic8.1 Integration BPM to design a worklist application. I need to create a number of Tasks(number know only at runtime), assign to users and wait for the completion of each of these task. I've tried using the Task Control factory to a

  • Fill in pdf form fields from html form

    hi i went to this site http://accesspdf.com/html_pdf_form/ and saw that it worked great so i downloaded their code, and uploaded it to my site (php friendly) and it gives me an error that the pdf file is corrupted http://inovasphere.com/pdf/pdf_form_

  • How to i get my itunes library to another computer

    I have two computers and one itunes account. i want to have my playlists and library on my other computer. how do i go about doing that?

  • Alt text and doctype XHTML

    Just a question to know if my thoughts are right... I have made my website and at the top DW has put: <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"><html xmlns="http://www.w3.

  • Trying to reboot iphone 4. pushed sleep/wake and home buttons for 20 sec. phone didn't reboot. itunes won't recognize phone.

    Trying to reboot iphone 4. Pushed the sleep/wake and home buttons together for 20+ sec. nothing happens. i tunes does not recognize the phone. The phone was working fine until the S/W and home buttons were pushed for aprox. 30 sec. it shut down and w