Guest Cluster error in Hyper-V Cluster

Hello everybody,
in my environment I do have an issue with failover clusters (Exchange, Fileserver) while performing a live migration of one virtual clusternode. The clustergroup is going offline.
The environment is the following:
2x Hyper-V Clusters: Hyper-V-Cluster1 and Hyper-V-Cluster2 (Windows Server 2012 R2) with 5 Nodes per Cluster
1x Scaleout Fileserver (Windows Server 2012 R2) with 2 Nodes
1x Exchange Cluster (Windows Server 2012 R2) with EX01 VM running on Hyper-V-Cluster1 and EX02 VM running on Hyper-V-Cluster2
1x Fileserver Failover Cluster (Windows Server 2012 R2) with FS01 VM running on Hyper-V-Cluster1 and FS02 VM running on Hyper-V-Cluster2
The physical networks on the Hyper-V Nodes are redundant with 2x 10Gb/s uplinks to 2x physical switches for VMs in a LBFO Team:
New-NetLbfoTeam
-Name 10Gbit_TEAM -TeamMembers 10Gbit_01,10Gbit_02
-TeamingMode SwitchIndependent -LoadBalancingAlgorithm HyperVPort
The SMB 3 traffic runs on 2x 10Gb/s NIC without NIC-Teaming (SMB-Multichannel).
SMB is used for livemigrations.
The VMs for clustering were installed according to the technet guideline:
http://technet.microsoft.com/en-us/library/dn265980.aspx
Because my Hyper-V Uplinks are allready redundant, I am using one NIC inside the VM.
As I understand, there is no advantage of using two NICs inside the VM as long they are connected to the same vSwitch.
Now, when I want to perform a hardware maintenance, I have to livemigrate the EX01 VM from Hyper-V-Cluster1-Node-1 to Hyper-V-Cluster1-Node-2.
EX02 VM still runs untouched on Hyper-V-Cluster2-Node-1.
At the end of the livemigration I see error 1135 (source: FailoverClustering) on EX01 VM, which says that EX02 VM was removed from Failover Cluster and I have to check my network.
The clustergroup of exchange is offline after that event and I have to bring it online again manually.
Any ideas what can cause this behavior?
Thanks.
Greetings,
torsten

Hello again,
I found the cause and the solution :-)
In the article here: http://technet.microsoft.com/en-us/library/dn440540.aspx
is the description of my cluster failure:
########## relevant part from article #######################
Protect against short-term network interruptions
Failover cluster nodes use the network to send heartbeat packets to other nodes of the cluster. If a node does not receive a response from another node for a specified period of time, the cluster removes the node from cluster membership. By default, a guest
cluster node is considered down if it does not respond within 5 seconds. Other nodes that are members of the cluster will take over any clustered roles that were running on the removed node.
Typically, during the live migration of a virtual machine there is a fast final transition when the virtual machine is stopped on the source node and is running on the destination node. However, if something causes the final transition to take longer than
the configured heartbeat threshold settings, the guest cluster considers the node to be down even though the live migration eventually succeeds. If the live migration final transition is completed within the TCP time-out interval (typically around 20 seconds),
clients that are connected through the network to the virtual machine seamlessly reconnect.
To make the cluster heartbeat time-out more consistent with the TCP time-out interval, you can change the
SameSubnetThreshold and CrossSubnetThreshold cluster properties from the default of 5 seconds to 20 seconds. By default, the cluster sends a heartbeat every 1 second. The threshold specifies how many heartbeats to miss in succession
before the cluster considers the cluster node to be down.
After changing both parameters in failover cluster as described the error is gone.
Greetings,
torsten

Similar Messages

  • Cluster Quorum Disk failing inside Guest cluster VMs in Hyper-V Cluster using Virtual Disk Sharing Windows Server 2012 R2

    Hi, I'm having a problem in a VM Guest cluster using Windows Server 2012 R2 and virtual disk sharing enabled. 
    It's a SQL 2012 cluster, which has around 10 vhdx disks shared this way. all the VHDX files are inside LUNs on a SAN. These LUNs are presented to all clustered members of the Windows Server 2012 R2 Hyper-V cluster, via Cluster Shared Volumes.
    Yesterday happened a very strange problem, both the Quorum Disk and the DTC disks got the information completetly erased. The vhdx disks themselves where there, but the info inside was gone.
    The SQL admin had to recreated both disks, but now we don't know if this issue was related to the virtualization platform or another event inside the cluster itself.
    Right now I'm seen this errors on one of the VM Guest:
     Log Name:      System
    Source:        Microsoft-Windows-FailoverClustering
    Date:          3/4/2014 11:54:55 AM
    Event ID:      1069
    Task Category: Resource Control Manager
    Level:         Error
    Keywords:      
    User:          SYSTEM
    Computer:      ServerDB02.domain.com
    Description:
    Cluster resource 'Quorum-HDD' of type 'Physical Disk' in clustered role 'Cluster Group' failed.
    Based on the failure policies for the resource and role, the cluster service may try to bring the resource online on this node or move the group to another node of the cluster and then restart it.  Check the resource and group state using Failover Cluster
    Manager or the Get-ClusterResource Windows PowerShell cmdlet.
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="Microsoft-Windows-FailoverClustering" Guid="{BAF908EA-3421-4CA9-9B84-6689B8C6F85F}" />
        <EventID>1069</EventID>
        <Version>1</Version>
        <Level>2</Level>
        <Task>3</Task>
        <Opcode>0</Opcode>
        <Keywords>0x8000000000000000</Keywords>
        <TimeCreated SystemTime="2014-03-04T17:54:55.498842300Z" />
        <EventRecordID>14140</EventRecordID>
        <Correlation />
        <Execution ProcessID="1684" ThreadID="2180" />
        <Channel>System</Channel>
        <Computer>ServerDB02.domain.com</Computer>
        <Security UserID="S-1-5-18" />
      </System>
      <EventData>
        <Data Name="ResourceName">Quorum-HDD</Data>
        <Data Name="ResourceGroup">Cluster Group</Data>
        <Data Name="ResTypeDll">Physical Disk</Data>
      </EventData>
    </Event>
    Log Name:      System
    Source:        Microsoft-Windows-FailoverClustering
    Date:          3/4/2014 11:54:55 AM
    Event ID:      1558
    Task Category: Quorum Manager
    Level:         Warning
    Keywords:      
    User:          SYSTEM
    Computer:      ServerDB02.domain.com
    Description:
    The cluster service detected a problem with the witness resource. The witness resource will be failed over to another node within the cluster in an attempt to reestablish access to cluster configuration data.
    Event Xml:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
      <System>
        <Provider Name="Microsoft-Windows-FailoverClustering" Guid="{BAF908EA-3421-4CA9-9B84-6689B8C6F85F}" />
        <EventID>1558</EventID>
        <Version>0</Version>
        <Level>3</Level>
        <Task>42</Task>
        <Opcode>0</Opcode>
        <Keywords>0x8000000000000000</Keywords>
        <TimeCreated SystemTime="2014-03-04T17:54:55.498842300Z" />
        <EventRecordID>14139</EventRecordID>
        <Correlation />
        <Execution ProcessID="1684" ThreadID="2180" />
        <Channel>System</Channel>
        <Computer>ServerDB02.domain.com</Computer>
        <Security UserID="S-1-5-18" />
      </System>
      <EventData>
        <Data Name="NodeName">ServerDB02</Data>
      </EventData>
    </Event>
    We don't know if this can happen again, what if this happens on disk with data?! We don't know if this is related to the virtual disk sharing technology or anything related to virtualization, but I'm asking here to find out if it is a possibility.
    Any ideas are appreciated.
    Thanks.
    Eduardo Rojas

    Hi,
    Please refer to the following link:
    http://blogs.technet.com/b/keithmayer/archive/2013/03/21/virtual-machine-guest-clustering-with-windows-server-2012-become-a-virtualization-expert-in-20-days-part-14-of-20.aspx#.Ux172HnxtNA
    Best Regards,
    Vincent Wu
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Win Server 2012 Failover Cluster - Error: Failed to bring cluster disk online

    Hi Technet
    I'm currently running running 2 VMs Win Server 2012 and would like to test Failover Clustering for one of our FTP server
    I've added on both servers an additional partition, formatted and Online
    One of the drives comes Online but cannot bring the 2nd disk online from the cluster manager
    Error: failed to bring resource online - clustered storage is not connected to the node
    Assistance would be greatly appreciated
    Thank you
    Jabu

    Hi jsibeko,
    Since the VMware offer the MSCS shared storage solution themselves, I suggest you first ask VMware whether that shared storage is supported with your vSphere edition, I found
    some VMware KB about the VMware shared storage for MSCS, may you can get more tips.
    The VMware KB:
    Microsoft Clustering on VMware vSphere: Guidelines for supported configurations (1037959)
    http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1037959
    Microsoft Cluster Service (MSCS) support on ESXi/ESX (1004617)
    http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1004617
    I’m glad to be of help to you!
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Very Strange Network Issue With Two Guests on 2012 R2 Hyper-V Failover Cluster

    Hi all.  We're having a odd issue with two guests on our 2012 R2 failover cluster.  
    In a nutshell, if we shutdown a particular server (I'll call it Server A) another totally different server (Server B) on the same node loses it's network connectivity to the domain. If we start server A back up, network connectivity returns on server B.
    At first I thought server A might be running a service that was somehow linked to server B, so I decided to disable server A's NIC.  Interestingly, that had no affect on server B's connectivity.  
    The next step I tried was pausing server A and again, no adverse affect on server B's connectivity.  
    Next step was to live migrate server A to another node.  This action did
    cause server B to lose its network connection. 
    One other clue is that if I ping server B from either of the Hyper-V hosts in the cluster, I never lose network connection to server B.
    So I would suspect this is some network issue on the cluster, but I'm kind of at a loss where to go from here.  
    Has anyone seen this behavior before or does anyone have any troubleshooting suggestions I can try?
    Thanks! 
    George Moore

    Hi Sir,
    I'v never seen this before .
    >>Next step was to live migrate server A to another node.  This action did
    cause server B to lose its network connection. 
    They are connecting to same virtual switch ?
    First please run cluster validation to check if there is any error .
    If it is ok , please try the following items for troubleshooting :
    1. shutdown  serverA   serverB
    2. then add another virtual NIC for serverB
    3. start server B  check if the issue happens to both "old" and "new" virtual NIC .
    In addition , you can live migrate both A and B to another node , then try to live migrate A to the original node .
    If the issue persists , I would suggest you to remove that virtual switch on both nodes then re-create them .
    Best Regards,
    Elton Ji
    If it is not the answer please unmark it to continue
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected] .

  • Hyper-V cluster: Unable to fail VM over to secondary host

    I am working on a Server 2012 Hyper-V Cluster. I am unable to fail my VMs from one node to the other using either LIVE or Quick migration.
    A force shutdown of VMHost01 will force a migration to VMHost02. And once we are on VMHost02 we can migrate back to VMHost01, but once that is done we can't move the VMs back to VMHost02 without a force shutdown.
    The following error pops up:
    Event ID: 21502 The Virtual Machine Management Service failed to establish a connection for a Virtual machine migration with host.... The connection attempt failed because the connected party did not properly respond after a period of time, or the established
    connection failed because connected host has failed to respond (0X8007274C)
    Here's what I noticed:
    VMMS.exe is running on VMHost02 however it is not listening on Port 6600. Confirmed this after a reboot by running netstat -a. We have tried setting this service to a delayed start.
    I have checked Firewall rules and Anti-Virus exclusions, and they are correct. I have not run the cluster validation test yet, because I'll need to schedule a period of downtime to do so.
    We can start/stop the VMMS.exe service just fine and without errors, but I am puzzled as to why it will not listen on Port 6600 anywhere. Anyone have any suggestions on how to troubleshoot this particular issue? 
    Thanks,
    Tho H. Le

    Just ran into the same issue in a 16-node cluster being managed by VMM. When trying to live migrate VMs using the VMM console the migration would fail with the following: Error 10698. Failover Cluster manager would report the following error code: Error
    (0x8007274C).
    + Validated Live Migration and Cluster networks. Everything checked out.
    + Looking in Hyper-V manager and migrations are enabled and correct networks displayed.
    + Found this particular Blog that mentions that the Virtual Machine Management service is not listening to port 6600
    http://blogs.technet.com/b/roplatforms/archive/2012/10/16/shared-nothing-migration-fails-0x8007274c.aspx
    Ran the following from an elivated command line:
    Netstat -ano | findstr 6600
    Node 2 did not return anything
    Node 1 returned correct output:
    TCP
    10.xxx.251.xxx:6600
    0.0.0.0:0
    LISTENING
    4540
    TCP
    10.xxx.252.xxx:6600
    0.0.0.0:0
    LISTENING
    4560
    Set Hyper-V Virtual Machine Service to delayed start.
    Restarted the service; no change.
    Checked the Event Logs for Hyper-V VMMS and noted the following events - VMMS Listener started
    for Live Migration networks, and then shortly after listener stopped.
    Removed the system from the cluster and restarted - No change
    Checked this host by running gpedit.msc - could not open console: Permission Error
    Tried to run a GPO refresh (gpupdate /force), but error returned that LocalGPO could not apply registry settings. Group Policy
    processing would not continue until this was resolved.
    Checked the local group policy folder on node 2 and it was corrupt:
    C:\Windows\System32\GroupPolicy\Machine\reg.pol showed 0K for the size.
    Copied local policy folders from Node 1 to 2, and then was able to refresh the GPOs.
    Restarting the VMMS service did not change the status of the ports.
    Restarted Server, added Live Migration networks back into Hyper-V manager and now netstat output reports that VMMS service
    is listening on 6600.

  • Hyper-V cluster Backup causes virtual machine reboots for common Cluster Shared Volumes members.

    I am having a problem where my VMs are rebooting while other VMs that share the same CSV are being backed up. I have provided all the information that I have gather to this point below. If I have missed anything, please let me know.
    My HyperV Cluster configuration:
    5 Node Cluster running 2008R2 Core DataCenter w/SP1. All updates as released by WSUS that will install on a Core installation
    Each Node has 8 NICs configured as follows:
     NIC1 - Management/Campus access (26.x VLAN)
     NIC2 - iSCSI dedicated (22.x VLAN)
     NIC3 - Live Migration (28.x VLAN)
     NIC4 - Heartbeat (20.x VLAN)
     NIC5 - VSwitch (26.x VLAN)
     NIC6 - VSwitch (18.x VLAN)
     NIC7 - VSwitch (27.x VLAN)
     NIC8 - VSwitch (22.x VLAN)
    Following hotfixes additional installed by MS guidance (either while build or when troubleshooting stability issue in Jan 2013)
     KB2531907 - Was installed during original building of cluster
     KB2705759 - Installed during troubleshooting in early Jan2013
     KB2684681 - Installed during troubleshooting in early Jan2013
     KB2685891 - Installed during troubleshooting in early Jan2013
     KB2639032 - Installed during troubleshooting in early Jan2013
    Original cluster build was two hosts with quorum drive. Initial two hosts were HST1 and HST5
    Next host added was HST3, then HST6 and finally HST2.
    NOTE: HST4 hardware was used in different project and HST6 will eventually become HST4
    Validation of cluster comes with warning for following things:
     Updates inconsistent across hosts
      I have tried to manually install "missing" updates and they were not applicable
      Most likely cause is different build times for each machine in cluster
       HST1 and HST5 are both the same level because they were built at same time
       HST3 was not rebuilt from scratch due to time constraints and it actually goes back to Pre-SP1 and has a larger list of updates that others are lacking and hence the inconsistency
       HST6 was built from scratch but has more updates missing than 1 or 5 (10 missing instead of 7)
       HST2 was most recently built and it has the most missing updates (15)
     Storage - List Potential Cluster Disks
      It says there are Persistent Reservations on all 14 of my CSV volumes and thinks they are from another cluster.
      They are removed from the validation set for this reason. These iSCSI volumes/disks were all created new for
      this cluster and have never been a part of any other cluster.
     When I run the Cluster Validation wizard, I get a slew of Event ID 5120 from FailoverClustering. Wording of error:
      Cluster Shared Volume 'Volume12' ('Cluster Disk 13') is no longer available on this node because of
      'STATUS_MEDIA_WRITE_PROTECTED(c00000a2)'. All I/O will temporarily be queued until a path to the
      volume is reestablished.
     Under Storage and Cluster Shared VOlumes in Failover Cluster Manager, all disks show online and there is no negative effect of the errors.
    Cluster Shared Volumes
     We have 14 CSVs that are all iSCSI attached to all 5 hosts. They are housed on an HP P4500G2 (LeftHand) SAN.
     I have limited the number of VMs to no more than 7 per CSV as per best practices documentation from HP/Lefthand
     VMs in each CSV are spread out amonst all 5 hosts (as you would expect)
    Backup software we use is BackupChain from BackupChain.com.
    Problem we are having:
     When backup kicks off for a VM, all VMs on same CSV reboot without warning. This normally happens within seconds of the backup starting
    What have to done to troubleshoot this:
     We have tried rebalancing our backups
      Originally, I had backup jobs scheduled to kick off on Friday or Saturday evening after 9pm
      2 or 3 hosts would be backing up VMs (Serially; one VM per host at a time) each night.
      I changed my backup scheduled so that of my 90 VMs, only one per CSV is backing up at the same time
       I mapped out my Hosts and CSVs and scheduled my backups to run on week nights where each night, there
       is only one VM backed up per CSV. All VMs can be backed up over 5 nights (there are some VMs that don't
       get backed up). I also staggered the start times for each Host so that only one Host would be starting
       in the same timeframe. There was some overlap for Hosts that had backups that ran longer than 1 hour.
      Testing this new schedule did not fix my problem. It only made it more clear. As each backup timeframe
      started, whichever CSV the first VM to start was on would have all of their VMs reboot and come back up.
     I then thought maybe I was overloading the network still so I decided to disable all of the scheduled backup
     and run it manually. Kicking off a backup on a single VM, in most cases, will cause the reboot of common
     CSV members.
     Ok, maybe there is something wrong with my backup software.
      Downloaded a Demo of Veeam and installed it onto my cluster.
      Did a test backup of one VM and I had not problems.
      Did a test backup of a second VM and I had the same problem. All VMs on same CSV rebooted
     Ok, it is not my backup software. Apparently it is VSS. I have looked through various websites. The best troubleshooting
     site I have found for VSS in one place it on BackupChain.com (http://backupchain.com/hyper-v-backup/Troubleshooting.html)
     I have tested almost every process on there list and I will lay out results below:
      1. I have rebooted HST6 and problems still persist
      2. When I run VSSADMIN delete shadows /all, I have no shadows to delete on any of my 5 nodes
       When I run VSSADMIN list writers, I have no error messages on any writers on any node...
      3. When I check the listed registry key, I only have the build in MS VSS writer listed (I am using software VSS)
      4. When I run VSSADMIN Resize ShadowStorge command, there is no shadow storage on any node
      5. I have completed the registration and service cycling on HST6 as laid out here and most of the stuff "errors"
       Only a few of the DLL's actually register.
      6. HyperV Integration Services were reconciled when I worked with MS in early January and I have no indication of
       further issue here.
      7. I did not complete the step to delete the Subscriptions because, again, I have no error messages when I list writers
      8. I removed the Veeam software that I had installed to test (it hadn't added any VSS Writer anyway though)
      9. I can't realistically uninstall my HyperV and test VSS
      10. Already have latest SPs and Updates
      11. This is part of step 5 so I already did this. This seems to be a rehash of various other stratgies
     I have used the VSS Troubleshooter that is part of BackupChain (Ctrl-T) and I get the following error:
      ERROR: Selected writer 'Microsoft Hyper-V VSS Writer' is in failed state!
      - Status: 8 (VSS_WS_FAILED_AT_PREPARE_SNAPSHOT)
      - Writer Failure code: 0x800423f0 (<Unknown error code>)
      - Writer ID: {66841cd4-6ded-4f4b-8f17-fd23f8ddc3de}
      - Instance ID: {d55b6934-1c8d-46ab-a43f-4f997f18dc71}
      VSS snapshot creation failed with result: 8000FFFF
    VSS errors in event viewer. Below are representative errors I have received from various Nodes of my cluster:
    I have various of the below spread out over all hosts except for HST6
    Source: VolSnap, Event ID 10, The shadow copy of volume took too long to install
    Source: VolSnap, Event ID 16, The shadow copies of volume x were aborted because volume y, which contains shadow copy storage for this shadow copy, wa force dismounted.
    Source: VolSnap, Event ID 27, The shadow copies of volume x were aborted during detection because a critical control file could not be opened.
    I only have one instance of each of these and both of the below are from HST3
    Source: VSS, Event ID 12293, Volume Shadow Copy Service error: Error calling a routine on a Shadow Copy Provider {b5946137-7b9f-4925-af80-51abd60b20d5}. Routine details RevertToSnashot [hr = 0x80042302, A Volume Shadow Copy Service component encountered an
    unexpected error.
    Source: VSS, Event ID 8193, Volume Shadow Copy Service error: Unexpected error calling routine GetOverlappedResult.  hr = 0x80070057, The parameter is incorrect.
    So, basically, everything I have tried has resulted in no success towards solving this problem.
    I would appreciate anything assistance that can be provided.
    Thanks,
    Charles J. Palmer
    Wright Flood

    Tim,
    Thanks for the reply. I ran the first two commands and got this:
    Name                                                            
    Role Metric
    Cluster Network 1                                              
    3  10000
    Cluster Network 2 - HeartBeat                              1   1300
    Cluster Network 3 - iSCSI                                    0  10100
    Cluster Network 4 - LiveMigration                         1   1200
    When you look at the properties of each network, this is how I have it configured:
    Cluster Network 1 - Allow cluster network communications on this network and Allow clients to connect through this network (26.x subnet)
    Cluster Network 2 - Allow cluster network communications on this network. New network added while working with Microsoft support last month. (28.x subnet)
    Cluster Network 3 - Do not allow cluster network communications on this network. (22.x subnet)
    Cluster Network 4 - Allow cluster network communications on this network. Existing but not configured to be used by VMs for Live Migration until MS corrected. (20.x subnet)
    Should I modify my metrics further or are the current values sufficient.
    I worked with an MS support rep because my cluster (once I added the 5th host) stopped being able to live migrate VMs and I had VMs host jumping on startup. It was a mess for a couple of days. They had me add the Heartbeat network as part of the solution
    to my problem. There doesn't seem to be anywhere to configure a network specifically for CSV so I would assume it would use (based on my metrics above) Cluster Network 4 and then Cluster Network 2 for CSV communications and would fail back to the Cluster Network
    1 if both 2 and 4 were down/inaccessible.
    As to the iSCSI getting a second NIC, I would love to but management wants separation of our VMs by subnet and role and hence why I need the 4 VSwitch NICs. I would have to look at adding an additional quad port NIC to my servers and I would be having to
    use half height cards for 2 of my 5 servers for that to work.
    But, on that note, it doesn't appear to actually be a bandwidth issue. I can run a backup for a single VM and get nothing on the network card (It caused the reboots before any real data has even started to pass apparently) and still the problem occurs.
    As to Backup Chain, I have been working with the vendor and they are telling my the issue is with VSS. They also say they support CSV as well. If you go to this page (http://backupchain.com/Hyper-V-Backup-Software.html)
    they say they support CSVs. Their tech support has been very helpful but unfortunately, nothing has fixed the problem.
    What is annoying is that every backup doesn't cause a problem. I have a daily backup of one of our machines that runs fine without initiating any additional reboots. But most every other backup job will trigger the VMs on the common CSV to reboot.
    I understood about the updates but I had to "prove" it to the MS tech I was on the phone with and hence I brought it up. I understand on the storage as well. Why give a warning for something that is working though... I think that is just a poor indicator
    that it doesn't explain that in the report.
    At a loss for what else I can do,
    Charles J. Palmer

  • Cluster Network Randomly Failing on Hyper-V Cluster

    Please let me know if there is a more appropriate forum. I am having a really strange issue that is seemingly random. I have a 3 host cluster that are all identical hardware and running Hyper-V Server 2012 R2. The networking is as follows and each network
    is a different VLAN/Subnet:
    3 Cluster networks for virtual machines
    1 Cluster network for cluster traffic/management
    1 Heartbeat network
    2 iSCSI networks for storage
    All of the networks are perfectly fine except for one which seems to fail on a random node at a random time during the day (so far, a maximum of once per day).
    If I start to live migrate virtual machines that are on the failed network, the cluster network comes back up. The cluster networks are teamed using SCVMM and they are switch independent and running the Dynamic teaming algorithm. We have tried changing the
    network switches to see if it was faulty network hardware and things ran fine for one day and then just happened again today so we've ruled that out. The only error message I get is 1127 which is the error stating that the cluster network has gone into a failed
    state which doesn't help much. I've run the cluster validation tool for networking several times and it always passes 100%. What I am worried about is hardware incompatibilities as I am using Dell servers (PowerEdge R720) that have Broadcom NIC's in them.
    We have 12 Ethernet ports in each server and they are all identical hardware. Four of them are integrated Broadcom, another four that are from a Broadcom quad add-on NIC, and another 4 that are from an Intel quad add-on NIC. All are server grade NIC's. The
    only problem I've had in the past is with VMQ which we've had to disable as a workaround but that has always stabilized our virtual networks. In any case all of the cluster networks for virtual machines are set up identically and only this particular one randomly
    fails on any one of the three hosts (it has happened at least once on each node now).
    I am wondering if anyone has had this experience before. I have read that there are some nasty compatibility issues between Broadcom and Hyper-V  but I am wondering if someone could give me some ideas to find out how to narrow this down since the event
    logs don't seem to be speaking in obvious terms to me.
    Please let me know if you have any suggestions on how to narrow down what's causing this or if there is more information that I could provide. In the meantime, I'm going to try and take note of which virtual machines are running on the host that has the
    network fail just in case there's some correlation there but that could take a while to accrue any useful data and our users aren't too happy with the instability...
    Thank you in advance for your time and sorry for the lengthy post!

    Since I made the change last Friday evening, 4/10, I haven't experienced the issue. I won't be completely convinced that this resolved it until I monitor for at least one more week since it didn't actually present itself for the first time until I was already
    one week into live deployment. Also, this link below is much more eloquent than how I put it and describes my issue exactly. Coupled with the KB article that someone posted within this comments section (the same that I posted earlier here) of the article,
    this is what led me to check the VMQ status through PowerShell which is much better than going through the registry to do it (I'm running Hyper-V Server 2012 R2 which is like core so I don't have the GUI options shown in the article).
    http://alexappleton.net/post/77116755157/hyper-v-virtual-machines-losing-network
    I could try updating the driver but there is mention in the comments of this post that driver updates have yet to resolve this issue so we may still be waiting on Broadcom for a fix. Please confirm otherwise if anyone has any information.

  • Hyper-V Cluster Name offline

    We have a 2012 Hyper-V cluster that isn't online and we can't migrate VMs to the other Hyper-V host.  We see event errors in the Failover Cluster Manager:
    The description for Event ID 1069 from source Microsoft-Windows-FailoverClustering cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component
    on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    Cluster Name
    Cluster Group
    Network Name
    The description for Event ID 1254 from source Microsoft-Windows-FailoverClustering cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component
    on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    Cluster Group
    The description for Event ID 1155 from source Microsoft-Windows-FailoverClustering cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component
    on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    ACMAIL
    3604536
    Any help or info is appreciated.
    Thank you!

    Here is the network validation.  Any thoughts?
    Failover Cluster Validation Report
    Failover Cluster Validation Report
          Node: ACHV01.AshtaChemicals.localValidated
          Node: ACHV02.AshtaChemicals.localValidated
          Started8/6/2014 5:04:47 PM
          Completed8/6/2014 5:05:22 PM
    The Validate a Configuration Wizard must be run after any change is made to the
    configuration of the cluster or hardware. For more information, see
    Results by Category
          NameResult SummaryDescription
          NetworkWarning
    Network
          NameResultDescription
          List Network Binding OrderSuccess
          Validate Cluster Network ConfigurationSuccess
          Validate IP ConfigurationWarning
          Validate Multiple Subnet PropertiesSuccess
          Validate Network CommunicationSuccess
          Validate Windows Firewall ConfigurationSuccess
    Overall Result
      Testing has completed for the tests you selected. You should review the
      warnings in the Report. A cluster solution is supported by Microsoft only if
      it passes all cluster validation tests.
    List Network Binding Order
      Description: List the order in which networks are bound to the adapters on
      each node.
      ACHV01.AshtaChemicals.local
            Binding OrderAdapterSpeed
            iSCSI3Intel(R) PRO/1000 PT Quad Port LP Server Adapter #31000 Mbit/s
            Ethernet 3Intel(R) PRO/1000 PT Quad Port LP Server AdapterUnavailable
            Mgt - HeartbeatMicrosoft Network Adapter Multiplexor Driver #42000
    Mbit/s
            Mgt - LiveMigrationMicrosoft Network Adapter Multiplexor Driver #32000
            Mbit/s
            MgtMicrosoft Network Adapter Multiplexor Driver2000 Mbit/s
            iSCSI2Broadcom BCM5709C NetXtreme II GigE (NDIS VBD Client) #371000
            Mbit/s
            3Broadcom BCM5709C NetXtreme II GigE (NDIS VBD Client)Unavailable
      ACHV02.AshtaChemicals.local
            Binding OrderAdapterSpeed
            Mgt - HeartbeatMicrosoft Network Adapter Multiplexor Driver #42000
    Mbit/s
            Mgt - LiveMigrationMicrosoft Network Adapter Multiplexor Driver #32000
            Mbit/s
            MgtMicrosoft Network Adapter Multiplexor Driver #22000 Mbit/s
            iSCSI1Broadcom NetXtreme Gigabit Ethernet #71000 Mbit/s
            NIC2Broadcom NetXtreme Gigabit EthernetUnavailable
            SLOT 5 2Broadcom NetXtreme Gigabit EthernetUnavailable
            iSCSI2Broadcom NetXtreme Gigabit Ethernet1000 Mbit/s
    Back to Summary
    Back to Top
    Validate Cluster Network Configuration
      Description: Validate the cluster networks that would be created for these
      servers.
      Network: Cluster Network 1
      DHCP Enabled: False
      Network Role: Disabled
      One or more interfaces on this network are connected to an iSCSI Target. This
      network will not be used for cluster communication.
            PrefixPrefix Length
            192.168.131.024
            ItemValue
            Network InterfaceACHV01.AshtaChemicals.local - iSCSI3
            DHCP EnabledFalse
            Connected to iSCSI targetTrue
            IP Address192.168.131.113
            Prefix Length24
            ItemValue
            Network InterfaceACHV02.AshtaChemicals.local - iSCSI2
            DHCP EnabledFalse
            Connected to iSCSI targetTrue
            IP Address192.168.131.121
            Prefix Length24
      Network: Cluster Network 2
      DHCP Enabled: False
      Network Role: Internal
            PrefixPrefix Length
            192.168.141.024
            ItemValue
            Network InterfaceACHV01.AshtaChemicals.local - Mgt - Heartbeat
            DHCP EnabledFalse
            Connected to iSCSI targetFalse
            IP Address192.168.141.10
            Prefix Length24
            ItemValue
            Network InterfaceACHV02.AshtaChemicals.local - Mgt - Heartbeat
            DHCP EnabledFalse
            Connected to iSCSI targetFalse
            IP Address192.168.141.12
            Prefix Length24
      Network: Cluster Network 3
      DHCP Enabled: False
      Network Role: Internal
            PrefixPrefix Length
            192.168.140.024
            ItemValue
            Network InterfaceACHV01.AshtaChemicals.local - Mgt - LiveMigration
            DHCP EnabledFalse
            Connected to iSCSI targetFalse
            IP Address192.168.140.10
            Prefix Length24
            ItemValue
            Network InterfaceACHV02.AshtaChemicals.local - Mgt - LiveMigration
            DHCP EnabledFalse
            Connected to iSCSI targetFalse
            IP Address192.168.140.12
            Prefix Length24
      Network: Cluster Network 4
      DHCP Enabled: False
      Network Role: Enabled
            PrefixPrefix Length
            10.1.1.024
            ItemValue
            Network InterfaceACHV01.AshtaChemicals.local - Mgt
            DHCP EnabledFalse
            Connected to iSCSI targetFalse
            IP Address10.1.1.4
            Prefix Length24
            ItemValue
            Network InterfaceACHV02.AshtaChemicals.local - Mgt
            DHCP EnabledFalse
            Connected to iSCSI targetFalse
            IP Address10.1.1.5
            Prefix Length24
      Network: Cluster Network 5
      DHCP Enabled: False
      Network Role: Disabled
      One or more interfaces on this network are connected to an iSCSI Target. This
      network will not be used for cluster communication.
            PrefixPrefix Length
            192.168.130.024
            ItemValue
            Network InterfaceACHV01.AshtaChemicals.local - iSCSI2
            DHCP EnabledFalse
            Connected to iSCSI targetTrue
            IP Address192.168.130.112
            Prefix Length24
            ItemValue
            Network InterfaceACHV02.AshtaChemicals.local - iSCSI1
            DHCP EnabledFalse
            Connected to iSCSI targetTrue
            IP Address192.168.130.121
            Prefix Length24
      Verifying that each cluster network interface within a cluster network is
      configured with the same IP subnets.
      Examining network Cluster Network 1.
      Network interface ACHV01.AshtaChemicals.local - iSCSI3 has addresses on all
      the subnet prefixes of network Cluster Network 1.
      Network interface ACHV02.AshtaChemicals.local - iSCSI2 has addresses on all
      the subnet prefixes of network Cluster Network 1.
      Examining network Cluster Network 2.
      Network interface ACHV01.AshtaChemicals.local - Mgt - Heartbeat has addresses
      on all the subnet prefixes of network Cluster Network 2.
      Network interface ACHV02.AshtaChemicals.local - Mgt - Heartbeat has addresses
      on all the subnet prefixes of network Cluster Network 2.
      Examining network Cluster Network 3.
      Network interface ACHV01.AshtaChemicals.local - Mgt - LiveMigration has
      addresses on all the subnet prefixes of network Cluster Network 3.
      Network interface ACHV02.AshtaChemicals.local - Mgt - LiveMigration has
      addresses on all the subnet prefixes of network Cluster Network 3.
      Examining network Cluster Network 4.
      Network interface ACHV01.AshtaChemicals.local - Mgt has addresses on all the
      subnet prefixes of network Cluster Network 4.
      Network interface ACHV02.AshtaChemicals.local - Mgt has addresses on all the
      subnet prefixes of network Cluster Network 4.
      Examining network Cluster Network 5.
      Network interface ACHV01.AshtaChemicals.local - iSCSI2 has addresses on all
      the subnet prefixes of network Cluster Network 5.
      Network interface ACHV02.AshtaChemicals.local - iSCSI1 has addresses on all
      the subnet prefixes of network Cluster Network 5.
      Verifying that, for each cluster network, all adapters are consistently
      configured with either DHCP or static IP addresses.
      Checking DHCP consistency for network: Cluster Network 1. Network DHCP status
      is disabled.
      DHCP status (disabled) for network interface ACHV01.AshtaChemicals.local -
      iSCSI3 matches network Cluster Network 1.
      DHCP status (disabled) for network interface ACHV02.AshtaChemicals.local -
      iSCSI2 matches network Cluster Network 1.
      Checking DHCP consistency for network: Cluster Network 2. Network DHCP status
      is disabled.
      DHCP status (disabled) for network interface ACHV01.AshtaChemicals.local - Mgt
      - Heartbeat matches network Cluster Network 2.
      DHCP status (disabled) for network interface ACHV02.AshtaChemicals.local - Mgt
      - Heartbeat matches network Cluster Network 2.
      Checking DHCP consistency for network: Cluster Network 3. Network DHCP status
      is disabled.
      DHCP status (disabled) for network interface ACHV01.AshtaChemicals.local - Mgt
      - LiveMigration matches network Cluster Network 3.
      DHCP status (disabled) for network interface ACHV02.AshtaChemicals.local - Mgt
      - LiveMigration matches network Cluster Network 3.
      Checking DHCP consistency for network: Cluster Network 4. Network DHCP status
      is disabled.
      DHCP status (disabled) for network interface ACHV01.AshtaChemicals.local - Mgt
      matches network Cluster Network 4.
      DHCP status (disabled) for network interface ACHV02.AshtaChemicals.local - Mgt
      matches network Cluster Network 4.
      Checking DHCP consistency for network: Cluster Network 5. Network DHCP status
      is disabled.
      DHCP status (disabled) for network interface ACHV01.AshtaChemicals.local -
      iSCSI2 matches network Cluster Network 5.
      DHCP status (disabled) for network interface ACHV02.AshtaChemicals.local -
      iSCSI1 matches network Cluster Network 5.
    Back to Summary
    Back to Top
    Validate IP Configuration
      Description: Validate that IP addresses are unique and subnets configured
      correctly.
      ACHV01.AshtaChemicals.local
            ItemName
            Adapter NameiSCSI3
            Adapter DescriptionIntel(R) PRO/1000 PT Quad Port LP Server Adapter #3
            Physical Address00-26-55-DB-CF-73
            StatusOperational
            DNS Servers
            IP Address192.168.131.113
            Prefix Length24
            ItemName
            Adapter NameMgt - Heartbeat
            Adapter DescriptionMicrosoft Network Adapter Multiplexor Driver #4
            Physical Address78-2B-CB-3C-DC-F5
            StatusOperational
            DNS Servers10.1.1.2, 10.1.1.8
            IP Address192.168.141.10
            Prefix Length24
            ItemName
            Adapter NameMgt - LiveMigration
            Adapter DescriptionMicrosoft Network Adapter Multiplexor Driver #3
            Physical Address78-2B-CB-3C-DC-F5
            StatusOperational
            DNS Servers10.1.1.2, 10.1.1.8
            IP Address192.168.140.10
            Prefix Length24
            ItemName
            Adapter NameMgt
            Adapter DescriptionMicrosoft Network Adapter Multiplexor Driver
            Physical Address78-2B-CB-3C-DC-F5
            StatusOperational
            DNS Servers10.1.1.2, 10.1.1.8
            IP Address10.1.1.4
            Prefix Length24
            ItemName
            Adapter NameiSCSI2
            Adapter DescriptionBroadcom BCM5709C NetXtreme II GigE (NDIS VBD Client)
            #37
            Physical Address78-2B-CB-3C-DC-F7
            StatusOperational
            DNS Servers
            IP Address192.168.130.112
            Prefix Length24
            ItemName
            Adapter NameLocal Area Connection* 12
            Adapter DescriptionMicrosoft Failover Cluster Virtual Adapter
            Physical Address02-61-1E-49-32-8F
            StatusOperational
            DNS Servers
            IP Addressfe80::cc2f:d769:fe24:3d04%23
            Prefix Length64
            IP Address169.254.2.195
            Prefix Length16
            ItemName
            Adapter NameLoopback Pseudo-Interface 1
            Adapter DescriptionSoftware Loopback Interface 1
            Physical Address
            StatusOperational
            DNS Servers
            IP Address::1
            Prefix Length128
            IP Address127.0.0.1
            Prefix Length8
            ItemName
            Adapter Nameisatap.{96B6424D-DB32-480F-8B46-056A11A0A6A8}
            Adapter DescriptionMicrosoft ISATAP Adapter
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.131.113%16
            Prefix Length128
            ItemName
            Adapter Nameisatap.{A0353AF4-CE7F-4811-B4FC-35273C2F2C6E}
            Adapter DescriptionMicrosoft ISATAP Adapter #3
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.130.112%18
            Prefix Length128
            ItemName
            Adapter Nameisatap.{FAAF4D6A-5A41-4725-9E83-689D8E6682EE}
            Adapter DescriptionMicrosoft ISATAP Adapter #4
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.141.10%22
            Prefix Length128
            ItemName
            Adapter Nameisatap.{C66443C2-DC5F-4C2A-A674-2191F76E33E1}
            Adapter DescriptionMicrosoft ISATAP Adapter #5
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:10.1.1.4%27
            Prefix Length128
            ItemName
            Adapter Nameisatap.{B3A95E1D-CB95-4111-89E5-276497D7EF42}
            Adapter DescriptionMicrosoft ISATAP Adapter #6
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.140.10%29
            Prefix Length128
            ItemName
            Adapter Nameisatap.{7705D42A-1988-463E-9DA3-98D8BD74337E}
            Adapter DescriptionMicrosoft ISATAP Adapter #7
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:169.254.2.195%30
            Prefix Length128
      ACHV02.AshtaChemicals.local
            ItemName
            Adapter NameMgt - Heartbeat
            Adapter DescriptionMicrosoft Network Adapter Multiplexor Driver #4
            Physical Address74-86-7A-D4-C9-8B
            StatusOperational
            DNS Servers10.1.1.8, 10.1.1.2
            IP Address192.168.141.12
            Prefix Length24
            ItemName
            Adapter NameMgt - LiveMigration
            Adapter DescriptionMicrosoft Network Adapter Multiplexor Driver #3
            Physical Address74-86-7A-D4-C9-8B
            StatusOperational
            DNS Servers10.1.1.8, 10.1.1.2
            IP Address192.168.140.12
            Prefix Length24
            ItemName
            Adapter NameMgt
            Adapter DescriptionMicrosoft Network Adapter Multiplexor Driver #2
            Physical Address74-86-7A-D4-C9-8B
            StatusOperational
            DNS Servers10.1.1.8, 10.1.1.2
            IP Address10.1.1.5
            Prefix Length24
            IP Address10.1.1.248
            Prefix Length24
            ItemName
            Adapter NameiSCSI1
            Adapter DescriptionBroadcom NetXtreme Gigabit Ethernet #7
            Physical Address74-86-7A-D4-C9-8A
            StatusOperational
            DNS Servers
            IP Address192.168.130.121
            Prefix Length24
            ItemName
            Adapter NameiSCSI2
            Adapter DescriptionBroadcom NetXtreme Gigabit Ethernet
            Physical Address00-10-18-F5-08-9C
            StatusOperational
            DNS Servers
            IP Address192.168.131.121
            Prefix Length24
            ItemName
            Adapter NameLocal Area Connection* 11
            Adapter DescriptionMicrosoft Failover Cluster Virtual Adapter
            Physical Address02-8F-46-67-27-51
            StatusOperational
            DNS Servers
            IP Addressfe80::3471:c9bf:29ad:99db%25
            Prefix Length64
            IP Address169.254.1.193
            Prefix Length16
            ItemName
            Adapter NameLoopback Pseudo-Interface 1
            Adapter DescriptionSoftware Loopback Interface 1
            Physical Address
            StatusOperational
            DNS Servers
            IP Address::1
            Prefix Length128
            IP Address127.0.0.1
            Prefix Length8
            ItemName
            Adapter Nameisatap.{8D7DF16A-1D5F-43D9-B2D6-81143A7225D2}
            Adapter DescriptionMicrosoft ISATAP Adapter #2
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.131.121%21
            Prefix Length128
            ItemName
            Adapter Nameisatap.{82E35DBD-52BE-4BCF-BC74-E97BB10BF4B0}
            Adapter DescriptionMicrosoft ISATAP Adapter #3
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.130.121%22
            Prefix Length128
            ItemName
            Adapter Nameisatap.{5A315B7D-D94E-492B-8065-D760234BA42E}
            Adapter DescriptionMicrosoft ISATAP Adapter #4
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.141.12%23
            Prefix Length128
            ItemName
            Adapter Nameisatap.{2182B37C-B674-4E65-9F78-19D93E78FECB}
            Adapter DescriptionMicrosoft ISATAP Adapter #5
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:192.168.140.12%24
            Prefix Length128
            ItemName
            Adapter Nameisatap.{104DC629-D13A-4A36-8845-0726AC9AE25E}
            Adapter DescriptionMicrosoft ISATAP Adapter #6
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:10.1.1.5%33
            Prefix Length128
            ItemName
            Adapter Nameisatap.{483266DF-7620-4427-BE5D-3585C8D92A12}
            Adapter DescriptionMicrosoft ISATAP Adapter #7
            Physical Address00-00-00-00-00-00-00-E0
            StatusNot Operational
            DNS Servers
            IP Addressfe80::5efe:169.254.1.193%34
            Prefix Length128
      Verifying that a node does not have multiple adapters connected to the same
      subnet.
      Verifying that each node has at least one adapter with a defined default
      gateway.
      Verifying that there are no node adapters with the same MAC physical address.
      Found duplicate physical address 78-2B-CB-3C-DC-F5 on node
      ACHV01.AshtaChemicals.local adapter Mgt - Heartbeat and node
      ACHV01.AshtaChemicals.local adapter Mgt - LiveMigration.
      Found duplicate physical address 78-2B-CB-3C-DC-F5 on node
      ACHV01.AshtaChemicals.local adapter Mgt - Heartbeat and node
      ACHV01.AshtaChemicals.local adapter Mgt.
      Found duplicate physical address 78-2B-CB-3C-DC-F5 on node
      ACHV01.AshtaChemicals.local adapter Mgt - LiveMigration and node
      ACHV01.AshtaChemicals.local adapter Mgt.
      Found duplicate physical address 74-86-7A-D4-C9-8B on node
      ACHV02.AshtaChemicals.local adapter Mgt - Heartbeat and node
      ACHV02.AshtaChemicals.local adapter Mgt - LiveMigration.
      Found duplicate physical address 74-86-7A-D4-C9-8B on node
      ACHV02.AshtaChemicals.local adapter Mgt - Heartbeat and node
      ACHV02.AshtaChemicals.local adapter Mgt.
      Found duplicate physical address 74-86-7A-D4-C9-8B on node
      ACHV02.AshtaChemicals.local adapter Mgt - LiveMigration and node
      ACHV02.AshtaChemicals.local adapter Mgt.
      Verifying that there are no duplicate IP addresses between any pair of nodes.
      Checking that nodes are consistently configured with IPv4 and/or IPv6
      addresses.
      Verifying that all nodes IPv4 networks are not configured using Automatic
      Private IP Addresses (APIPA).
    Back to Summary
    Back to Top
    Validate Multiple Subnet Properties
      Description: For clusters using multiple subnets, validate the network
      properties.
      Testing that the HostRecordTTL property for network name Name: Cluster1 is set
      to the optimal value for the current cluster configuration.
      HostRecordTTL property for network name Name: Cluster1 has a value of 1200.
      Testing that the RegisterAllProvidersIP property for network name Name:
      Cluster1 is set to the optimal value for the current cluster configuration.
      RegisterAllProvidersIP property for network name Name: Cluster1 has a value of
      0.
      Testing that the PublishPTRRecords property for network name Name: Cluster1 is
      set to the optimal value for the current cluster configuration.
      The PublishPTRRecords property forces the network name to register a PTR in
      DNS reverse lookup record IP address to name mapping.
    Back to Summary
    Back to Top
    Validate Network Communication
      Description: Validate that servers can communicate, with acceptable latency,
      on all networks.
      Analyzing connectivity results ...
      Multiple communication paths were detected between each pair of nodes.
    Back to Summary
    Back to Top
    Validate Windows Firewall Configuration
      Description: Validate that the Windows Firewall is properly configured to
      allow failover cluster network communication.
      The Windows Firewall on node 'ACHV01.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV01.AshtaChemicals.local - Mgt - LiveMigration'.
      The Windows Firewall on node 'ACHV01.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV01.AshtaChemicals.local - iSCSI3'.
      The Windows Firewall on node 'ACHV01.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV01.AshtaChemicals.local - Mgt - Heartbeat'.
      The Windows Firewall on node 'ACHV01.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV01.AshtaChemicals.local - Mgt'.
      The Windows Firewall on node 'ACHV01.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV01.AshtaChemicals.local - iSCSI2'.
      The Windows Firewall on node ACHV01.AshtaChemicals.local is configured to
      allow network communication between cluster nodes.
      The Windows Firewall on node 'ACHV02.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV02.AshtaChemicals.local - Mgt'.
      The Windows Firewall on node 'ACHV02.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV02.AshtaChemicals.local - iSCSI2'.
      The Windows Firewall on node 'ACHV02.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV02.AshtaChemicals.local - Mgt - LiveMigration'.
      The Windows Firewall on node 'ACHV02.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV02.AshtaChemicals.local - Mgt - Heartbeat'.
      The Windows Firewall on node 'ACHV02.AshtaChemicals.local' is configured to
      allow network communication between cluster nodes over adapter
      'ACHV02.AshtaChemicals.local - iSCSI1'.
      The Windows Firewall on node ACHV02.AshtaChemicals.local is configured to
      allow network communication between cluster nodes.
    Back to Summary
    Back to Top

  • Add Node to Hyper-V Cluster running Server 2012 R2

    Hi All,
    I am in the process to upgrade our Hyper-V Cluster to Server 2012 R2 but I am not sure about the required Validation test.
    The Situation at the Moment-> 1 Node Cluster running Server 2012 R2 with 2 CSVs and Quorum. Addtional Server prepared to add to the Cluster. One CSV is empty and could be used for the Validation Test. On the Other CSV are running 10 VMs in production.
    So when I start the Validation wizard I can select specific CSVs to test, which makes sense;-) But the Warning message is not clear for me "TO AVOID ROLE FAILURES, IT IS RECOMMENDED THAT ALL ROLES USING CLUSTER SHARED VOLUMES BE STOPPED BEFORE THE STORAGE
    IS VALIDATED". Does it mean that ALL CSVs will be testest and Switched offline during the test or just the CSV that i have selected in the Options? I have to avoid definitly that the CSV where all the VMs are running will be switched offline and also
    that the configuration will be corputed after loosing the CSV where the VMs are running.
    Can someone confirm that ONLY the selected CSV will be used for the Validation test ???
    Many thanks
    Markus

    Hi,
    The validation will test the select the CSV storage, if you have guest vm running this CSV it must shutdown or saved before you validate the CSV.
    Several tests will actually trigger failovers and move the disks and groups to different cluster nodes which will cause downtime, and these include Validating Disk Arbitration,
    Disk Failover, Multiple Arbitration, SCSI-3 Persistent Reservation, and Simultaneous Failover. 
    So if you want to test a majority of the functionality of your cluster without impacting availability, exclude these tests.
    The related information:
    Validating a Cluster with Zero Downtime
    http://blogs.msdn.com/b/clustering/archive/2011/06/28/10180803.aspx
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Best design for HA Fileshare on existing Hyper-V Cluster?

    Have a three node 2012 R2 Hyper-V Cluster. The storage is a HP MSA 2000 G3 SAS Block Storage with CSV's. 
    We have a fileserver for all users running as VM on the cluster. Fileserver availability is important and it's difficult to take this fileserver down for the monthly patching. So we want to make these file services HA. Nearly all clients are Windows 8.1,
    so SMB 3 can be used. 
    What is the best way to make these file services HA?
    1. The easiest way would probably be to migrate these fileserver ressources to a dedicated LUN on the MSA 2000, and to add a "general fileserver role" to the existing hyper-V cluster. But is it supported and a good solution to provide Hyper-V VM's
    and HA file services on the same cluster (even when the performance requirements for file services are not high)? Or does this configuration affect the Hyper-V VM performance too much?
    2. Is it better to create a two node guest cluster with "Shared VHDX" for the file services? I'm not sure if this would even work. Because we had "Persistent Reservation" warnings when creating the Hyper-V cluster with the MSA 2000. According "http://blogs.msdn.com/b/clustering/archive/2013/05/24/10421247.aspx",
    these warnings are normal with block storage and can be ignored when we never want to create Windows storage pools or storage spaces. But the Hyper-V MMC shows that "shared VHDX" work with "persistent reservations". 
    3. Are there other possibilities to provide HA file services with this configuration without buying new HW? (Remark: DFSR with two independet Fileservers is probably not a good solution, we have a lot of data that change frequently).
    Thank you in advance for any advice and recommedations!
    Franz

    Hi Franz,
    If you are not going to be using Storage Spaces in the Cluster, this is a warning that you can safely ignore. 
    It passes the normal SCSI3 Persistent Reservation tests, so you are good with those. Additional, when we use the cluster we can install the cluster CAU it will automatically install the cluster updates.
    The related KB:
    Requirements and Best Practices for Cluster-Aware Updating
    https://technet.microsoft.com/en-us/library/jj134234.aspx
    Cluster-Aware Updating: Frequently Asked Questions
    https://technet.microsoft.com/en-us/library/hh831367.aspx
    I’m glad to be of help to you!
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Hyper-V cluster validation report "Found duplicate physical address" on nic team interfaces.

    I recently built a Windows 2012 Hyper-V cluster with 5 nodes. The validation report shows “duplicate physical address” error (error text pasted below).
    The hardware: HP BladeSystem – servers are BL460c blades, in a c7000 enclosure, connected to HP Virtual Connect switches.
    Each server has 2 physcal nics, teamed in Windows. In the NIC Teaming console, I created the following Team Interfaces and assigned each a VLAN ID:
    “Team1” (the default team)
    “Team1 - VLAN 204 – Management”
    “Team1 - VLAN 212 - 2012HB”
    “Team1 - VLAN 211 -Exchange DAG Replication”
    I also created 2 HV Virtual Switches. Neither one allows management interface to share. They are assigned to “Team1” and “Team1 - VLAN 211 -Exchange DAG Replication” respectively.
    Therefore, in Network Connection, I see the 2 physical Ethernet nics, and 4 “virtual” nics. Only 2 of them have IP addresses assigned: Management and HB. These are the two that the validation wizard complains
    about.
    The MAC address is not configurable in the NIC Teaming console, so I don’t see a way to resolve this error, except to use separate physical nics. I don’t want to do that because a) I would lose the benefits of
    the bandwidth aggregation that Virtual Connect provides, and b) When creating an Interface on a Team in Windows, it looks like it ALWAYS gives it the same MAC address, so that should be a supported configuration.
    Everything works just fine, and there are no other errors or IP conflicts or anything else. But I really want to fix it because I don’t know what unknown problems this may be causing.
    From the Cluster Validation report:
    Found duplicate physical address 10-60-4B-A9-4A-30 on node Cluster201.OurDomain.local adapter
    Team1 - VLAN 212 - 2012HB and node Cluster201.OurDomain.local adapter
    Team1 - VLAN 204 - Management.
    Found duplicate physical address F0-92-1C-13-3C-2C on node Cluster202.OurDomain.local adapter
    Team1 - VLAN 212 - 2012HB and node Cluster202.OurDomain.local adapter
    Team1 - VLAN 204 - Management.
    Found duplicate physical address 68-B5-99-C1-7E-9C on node Cluster210.OurDomain.local adapter
    Team1 - VLAN 212 - 2012HB and node Cluster210.OurDomain.local adapter
    Team1 - VLAN 204 - Management.
    Found duplicate physical address 3C-4A-92-DE-1E-74 on node Cluster211.OurDomain.local adapter VC-Team - VLAN 212 - 2012HB and node Cluster211.OurDomain.local adapter
    VC-Team - VLAN 204 - Management.
    Found duplicate physical address 68-B5-99-C0-3D-50 on node Cluster212.OurDomain.local adapter
    Team1 - VLAN 212 - 2012HB and node Cluster212.OurDomain.local adapter
    Team1 - VLAN 204 - Management.
    Thanks!
    Dan

    Hi Dan,
    "It turns out that both hosts had the same default MAC address ranges for their virtual switches. Since the host vNICs were attached to the virtual switch on each host they received the first couple of MAC addresses from the switches.
    For details please refer to following link:
    http://www.jefflafr.com/blog/4/19/2013/conflicting-mac-addresses-when-building-a-hyper-v-cluster-with-converged-networking
    Hope this helps
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Windows server 2012 Datacenter Hyper-V Cluster -- Failed to validate Operating System Installation Option?

    Hi I have a 4 node Windows server 2012 Hyper-V cluster. When I try to run a cluster validation report, everything else is fine but it fails at validate the Operating System Installation Option step. I did some research but couldn't really find any solution.
    Anyone knows how to pass this test? Thanks.
    Here's the error I get when run the test:
    An error occurred while executing the test.
    The operation has failed. An error occurred while getting the operating system installation option for node "server1"

    Hi JasonLiu2002,
    Please post the original error information, the current information is so wide that difficult to determine where may have issue and please offer more information about your
    server configuration, you can refer the following article to prepare your cluster environment first.
    Windows Server 2012 Hyper-V Best Practices (In Easy Checklist Form)
    http://blogs.technet.com/b/askpfeplat/archive/2013/03/10/windows-server-2012-hyper-v-best-practices-in-easy-checklist-form.aspx
    When you preparing the new cluster on Server 2012 please install the Recommended hotfixes and updates for Windows Server 2012-based failover clusters updates.
    http://support.microsoft.com/kb/2784261
    I’m glad to be of help to you!
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • SMB3 for Hyper-V Cluster

    I'm contemplating using the much hyped SMB3 backed Hyper-V cluster. I just have a few questions.
    1. Is there a way for the SMB3 share to be HA?
    2. Is it easily scalable / can I add storage live without downtime?
    3. Is there any performance or reliability advantage over iSCSI attached storage?
    This is assuming the data within the VHDs is OS and general company data, not large CAD or multimedia data.
    I can probably google most of this knowledge but I'm looking for the confirmation from someone who has done / is doing it. Whitepapers can be not so helpful sometimes and sales guys usually have to refer me to their sales engineer. Thanks in advance.
    This topic first appeared in the Spiceworks Community

    Are there any ways around this limitation without having to install 3rd party software?
    I'm surprised I wasn't able to find much about this on any of my searches.
    Run your workload inside a virtual machine. Configure guest VM cluster between a pair of VMs running Windows Server 2012 R2 and make built-in MSFT iSCSI target to do a failover. See for reference:
    Configure MSFT iSCSI Target for HA
    http://technet.microsoft.com/en-us/library/gg232621(v=ws.10).aspx
    (yes it would bring up virtualization overhead as all I/O would be routed over VMbus and also you'll be still active-passive as MSFT cannot do active-active but if you don't want to use third-party software and don't care much about performance that's the
    viable way to go)
    Good luck!
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Hyper-V Cluster in VMM

    I am trying to build a Hyper-V cluster with VMM 2012 R2 but require some advice as it is not working how I want it too.
    I have 2 Hyper-V servers, both with their own local storage and 1 iSCSI disk shared between them. I am trying to cluster the servers so that the shared iSCSI disk becomes a shared volume while maintaining the ability to use the local storage as well - some
    VMs will run from local storage while others will run from the CSV.
    The issue I'm having is that when I cluster the 2 servers the iSCSI disk does not show up in VMM as a shared volume. In Windows Explorer the disk has the cluster icon but in VMM there is nothing. In the cluster properties I can add a shared volume... but
    it asks for a logical node which I cannot create because I have no storage pools (server manager says no groups of disks are available to pool).
    I also noticed when I clustered the servers my 2 file shares to their local storage disappeared from VMM which isn't what I want.
    Can someone please advise, or link to, a way to achieve my desired configuration?
    Cheers,
    MrGoodBytes
    Note: Posts are provided “AS IS” without warranty of any kind, either expressed or implied, including but not limited to the implied warranties of merchantability and/or fitness for a particular purpose.

    Hi MrGoodBytes,
    Hi,
    Unfortunately, the available information is not enough to have a clear view of the occurred behavior. Could you provide more information about your environment. For example,
    the server version of the problem on, when this problem occurs the system log record information, screenshots is the best information.
    Before you create the cluster we strongly recommend you run the cluster validation, If you are considering the cluster may have some issue please rerun the validation, then
    post the validation report warning and error part information, this report will quickly locate the cluster potential issue.
    A disk witness is a disk in the cluster storage that is designated to hold a copy of the cluster configuration database. A failover cluster has a disk witness only if this
    is specified as part of the quorum configuration.
    Configure and Manage the Quorum in a Windows Server 2012 Failover Cluster
    http://technet.microsoft.com/zh-cn/library/jj612870.aspx
    I am not familiar with SVCMM so please refer the following related KB to confirm your shared storage add steps is correct.
    How to Configure Storage on a Hyper-V Host Cluster in VMM
    http://technet.microsoft.com/en-us/library/gg610692.aspx
    Configuring Storage in VMM
    http://technet.microsoft.com/en-us/library/gg610600.aspx
    More information:
    How to add storage to Clustered Shared Volumes in Windows Server 2012
    http://blogs.msdn.com/b/clustering/archive/2012/04/06/10291490.aspx
    Configure and Manage the Quorum in a Windows Server 2012 Failover Cluster
    http://technet.microsoft.com/zh-cn/library/jj612870.aspx
    Event Logs
    http://technet.microsoft.com/en-us/library/cc722404.aspx
    I’m glad to be of help to you!
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Moving Physical SQL Cluster into Virtual SQL Cluster based on Hyper-V Failover

    Hello All............I have a SQL Cluster based on Physical Hardware that has Three-3 instances as well. I have setup a Hyper-V Failover Cluster (2012 R2) and have built Virtual/Guest SQL Cluster (2012 R2) upon it.  Now, I intend to move
    the instances/databases from Physical SQL Cluster to Virtual SQL Cluster.
    1.  Is this supported?  If so, I would appreciate any guidance on it?
    2.  Is P2V of SQL Cluster supported in Hyper-V Failover Cluster based on Windows Server 2012 R2?

    Hi Sir,
    Please refer to the following blog regarding moving SCCVM SQL database to another SQL server:
    http://blogs.technet.com/b/configurationmgr/archive/2013/04/02/how-to-move-the-configmgr-2012-site-database-to-a-new-sql-server.aspx
    It is quoted from the  similar thread :
    https://social.technet.microsoft.com/Forums/en-US/a1558842-cdf5-4e5f-8f10-d660e96eae1b/migration-sql-for-sccm-2012?forum=configmanagermigration
    But it seems that is a migration of system center production , I would suggest you to post the question to system center forum :
    https://social.technet.microsoft.com/Forums/en-US/home?forum=configmanagermigration%2Coperationsmanagergeneral&filter=alltypes&sort=lastpostdesc
    Best Regards,
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

Maybe you are looking for

  • Please advise - new iMac i7, old iMac G5 PPC, and migration assistant

    I just received my new iMac i7 and am getting ready to use Migration Assistant. However, I have some questions: The logic board on my old machine died, so I pulled out the Hard Drive and put it in an enclosure. This is what I will be using to make th

  • How to monitor the alerts(error handling) in seerburger adapter?

    hi, is it like normal moniotoring for any scenario..i have to do the monitoring in seeburger adapter. any other specific thing i have to take care to do monitoring the alerts (error handling) in seeburger adapter for AS2 and FTP adapters.. plz give o

  • Is it have some User exits for Vendor master  trigger when click some field

    Dear Experts,      I would like to know Is it have User exit for Vendor master  trigger when click some field in Vendor master? not just User exit for Prior Save . Please kindly let me know some solution for this case.      Many thank.

  • Upgrading to CS5.5

    Is there any benefit to upgrading via the download from the adobe store compared to the boxed CD Set. I will also post this in the Premiere Forum Thank You

  • HELP  iPhoto database stuck

    iPhoto & Aperture database stuck on 'processing' for days. Measage preparing thumbnails.  Tried to cancel  but just keeps processing and I can't use either app. Any ideas?  Thanks