Hyper-V NIC Team Load Balancing Algorithm: TranportPorts vs Hyper-VPorts

Hi, 
I'm going to need to configure a NIC team for the LAN traffic for a Hyper-V 2012 R2 environment. What is the recommended load balancing algorithm? 
Some background:
- The NIC team will deal with LAN traffic (NOT iSCSI storage traffic)
- I'll set up a converged network. So there'll be a virtual switch on top of this team, which will have vNICs configured for each cluster, live migration and management
- I'll implement QOS at the virtual switch level (using option -DefaultFlowMinimumBandwidthWeight) and at the vNIC level (using option -MinimumBandwidthWeight)
- The CSV is set up on an Equallogics cluster. I know that this team is for the LAN so it has nothing to do with the SAN, but this reference will become clear in the next paragraph. 
Here's where it gets a little confusing. I've checked some of the Equallogics documentation to ensure this environment complies with their requirements as far as storage networking is concerned. However, as part of their presentation the Dell publication
TR1098-4, recommends creating the LAN NIC team with the TrasportPorts Load Balancing Algorithm. However, in some of the Microsoft resources (i.e. http://technet.microsoft.com/en-us/library/dn550728.aspx), the recommended load balancing algorithm is HyperVPorts.
Just to add to the confusion, in this Microsoft TechEd presentation, http://www.youtube.com/watch?v=ed7HThAvp7o, the recommendation (at around minute 8:06) is to use dynamic ports algorithm mode. So obviously there are many ways to do this, but which one is
correct? I spoke with Equallogics support and the rep said that their documentation recommends TransportPorts LB algorithm because that's what they've tested and works. I'm wondering what the response from a Hyper-V expert would be to this question. Anyway,
any input on this last point would be appreciated.

Gleb,
>>See Windows Server 2012 R2 NIC Teaming (LBFO) Deployment and Management  for more
info
Thanks for this reference. It seems that I have an older version of this document where there's absolutely
no mention of the dynamic LBA. Hence my confusion when in the Microsoft TechEd presentation the
recommendation was to use Dynamic. I almost implemented this environment with switch dependent and Address Hash Distribution because, based on the older version of the document, this combination offered: 
a) Native teaming for maximum performance and switch diversity is not required; or
b) Teaming under the Hyper-V switch when an individual VM needs to be able to transmit at rates in excess of what one team member can deliver
The new version of the document recommends Dynamic over the other two LBA. The analogy that the document
makes of TCP flows with human speech was really helpful for me to understand what this algorithm is doing. For those who will never read the document, I'm referring to this: 
"The outbound loads in this mode are dynamically balanced based on the concept of
flowlets.  Just as human speech has natural breaks at the ends of words and sentences, TCP flows (TCP communication streams) also have naturally
occurring breaks.  The portion of a TCP flow between two such breaks is referred to as a flowlet.  When the dynamic mode algorithm detects that a flowlet boundary has been encountered, i.e., a break of sufficient length has occurred in the TCP flow,
the algorithm will opportunistically rebalance the flow to another team member if apropriate.  The algorithm may also periodically rebalance flows that do not contain any flowlets if circumstances require it.    As a result the affinity
between TCP flow and team member can change at any time as the dynamic balancing algorithm works to balance the workload of the team members. "
Anyway, this post made my week. You sir are deserving of a beer!

Similar Messages

  • ESXi 4.1 NIC Teaming's Load-Balancing Algorithm,Nexus 7000 and UCS

    Hi, Cisco Gurus:
    Please help me in answering the following questions (UCSM 1.4(xx), 2 UCS 6140XP, 2 Nexus 7000, M81KR in B200-M2, No Nexus 1000V, using VMware Distributed Switch:
    Q1. For me to configure vPC on a pair of Nexus 7000, do I have to connect Ethernet Uplink from each Cisco Fabric Interconnect to the 2 Nexus 7000 in a bow-tie fashion? If I connect, say 2 10G ports from Fabric Interconnect 1 to 1 Nexus 7000 and similar connection from FInterconnect 2 to the other Nexus 7000, in this case can I still configure vPC or is it a validated design? If it is, what is the pro and con versus having 2 connections from each FInterconnect to 2 separate Nexus 7000?
    Q2. If vPC is to be configured in Nexus 7000, is it COMPULSORY to configure Port Channel for the 2 Fabric Interconnects using UCSM? I believe it is not. But what is the pro and con of HAVING NO Port Channel within UCS versus HAVING Port Channel when vPC is concerned?
    Q3. if vPC is to be configured in Nexus 7000, I understand there is a limitation on confining to ONLY 1 vSphere NIC Teaming's Load-Balancing Algorithm i.e. Route Based on IP Hash. Is it correct?
    Again, what is the pro and con here with regard to application behaviours when Layer 2 or 3 is concerned? Or what is the BEST PRACTICES?
    I would really appreciate if someone can help me clear these lingering doubts of mine.
    God Bless.
    SiM

    Sim,
    Here are my thoughts without a 1000v in place,
    Q1. For me to configure vPC on a pair of Nexus 7000, do I have to connect Ethernet Uplink from each Cisco Fabric Interconnect to the 2 Nexus 7000 in a bow-tie fashion? If I connect, say 2 10G ports from Fabric Interconnect 1 to 1 Nexus 7000 and similar connection from FInterconnect 2 to the other Nexus 7000, in this case can I still configure vPC or is it a validated design? If it is, what is the pro and con versus having 2 connections from each FInterconnect to 2 separate Nexus 7000?   //Yes, for vPC to UCS the best practice is to bowtie uplink to (2) 7K or 5Ks.
    Q2. If vPC is to be configured in Nexus 7000, is it COMPULSORY to configure Port Channel for the 2 Fabric Interconnects using UCSM? I believe it is not. But what is the pro and con of HAVING NO Port Channel within UCS versus HAVING Port Channel when vPC is concerned? //The port channel will be configured on both the UCSM and the 7K. The pro of a port channel would be both bandwidth and redundancy. vPC would be prefered.
    Q3. if vPC is to be configured in Nexus 7000, I understand there is a limitation on confining to ONLY 1 vSphere NIC Teaming's Load-Balancing Algorithm i.e. Route Based on IP Hash. Is it correct? //Without the 1000v, I always tend to leave to dvSwitch load balence behavior at the default of "route by portID". 
    Again, what is the pro and con here with regard to application behaviours when Layer 2 or 3 is concerned? Or what is the BEST PRACTICES? UCS can perform L2 but Northbound should be performing L3.
    Cheers,
    David Jarzynka

  • Windows Server 2012 R2 - Hyper-V NIC Teaming Issue

    Hi All,
    I have cluster windows server 2012 R2 with hyper-v role installed. I have an issue with one of my windows 2012 R2 hyper-v host. 
    The virtual machine network adapter show status connected but it stop transmit data, so the vm that using that NIC cannot connect to external network.
    The virtual machine network adapter using Teamed NIC, with this configuration:
    Teaming Mode : Switch Independent
    Load Balance Algorithm : Hyper-V Port
    NIC Adapter : Broadcom 5720 Quad Port 1Gbps
    I already using the latest NIC driver from broadcom.
    I found a little trick for this issue by disable one of the teamed NIC, but it will happen again.
    Anyone have the same issue with me, and any workaround for this issue?
    Please Advise
    Thanks,

    Hi epenx,
    Thanks for the information .
    Best Regards,
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • ACE Load Balancing algorithm

    Team,
    I was reading Designing Content Switching Solutions last night. I came across a page that suggested Round Robin for HTTP connections, Least Conns for FTP connections, dst Hash for caching connections and so on.
    Could someone please provide information or a link on which load balancing algorithm to use based on the application, is there some form of best practice for this?
    Thank you,
    John...

    John,
    there is no best practices.
    It depends on your applications and needs.
    For example, for caching, some people prefer to optimize the disk space, and other the response time.
    So, if you do destination hash, you guarantee that all traffic for one site is always handled by the same cache.
    Therefore you optimize the disk space since you will not find the same object on all caches.
    BUT if one site attracts a lot of connections, the cache device that handles that site will be overloaded (for example youtube.com)
    Leastconn is a good option in theory.
    The device that has less connections should receive the next one.
    The problem is if you have flapping links or servers crashing or if you do a lot of maintenance and add/remove servers frequently.
    This confuses the algorithm and is the source of a lot of bugs.
    My recommendation is to go with roundrobin unless you have identify that you really need another algorithm.
    And you can always start with roundrobin and see what happens...
    Gilles.

  • SG300-28: LAG Load Balance Algorithm

    The SG300-28 LAG Load Balance Algorithm options are:
    MAC Address
    IP/MAC Address
    This switch is connected to two ESXi hosts running version 5.5 each with quad port 1 gbit nic.
    These ESXI hosts supports > 20 algorithms.
    Is there in a upcoming firmware for the SG300-28 supporting more algorithms?

    The SG300-28 LAG Load Balance Algorithm options are:
    MAC Address
    IP/MAC Address
    This switch is connected to two ESXi hosts running version 5.5 each with quad port 1 gbit nic.
    These ESXI hosts supports > 20 algorithms.
    Is there in a upcoming firmware for the SG300-28 supporting more algorithms?

  • Load balancing algorithm for groups in RZ12

    Hello,
    I would like to know the load balancing algorithm for groups defined in RZ12.
    I know that log on groups for external connections are administered via SMLG and table RZLLICLASS.
    I also know that RFC resources can be managed for RFC logon groups via RZ12.
    Kind regards,
    Peter
    <removed_by_moderator>
    Point awarding is at your discretion, but read and follow the "Rules of Engagement"
    Edited by: Juan Reyes on Dec 3, 2010 10:21 AM

    Hello!
    Found this post while searching information about RFC and Logon Groups...
    I have some mess in my head with SMLG functionality and RZ12. As I know SMLG we can use to distribute users to application server instances, it gives us good achievement in performance. With RZ12 we can distribute RFC connection of particular job for parallel execution on predefined application server. With SMQS and SMQR transaction we can set "Name of AS Group" to route RFC-execution on certain server or servers. But I have troubles with understanding. Imagine, we set up group 1 with 2 servers (name it RFC_GR1), and group 2 with another 2 servers (name it RFC_GR2). How could qRFC scheduler decide on which RFC server group (RFC_GR1 or RFC_GR2) distribute RFC-execution? How to interact "Name of AS Group" with RFC groups if we can set only one group?   How could we distribute RFC-execution depending on our logon groups (smlg)? We would like to distribute RFC depending on SAP logon groups. Is it possible? Or do I compare apple and orange?
    Regards,
    Artem Ivashkin

  • Port Channel Load-Balancing Algorithm (North Bound)

    I'm trying to figure out what the load balancing algorithm for the 6100 and 6200 FIs for the Northbound connections. I can't find any documentation on how to change it.
    The Nexus 7000s use an 8-bit hash, making it very easy to do something other than 2, 4, or 8 link port channel and get even (at least algorithmically) distribution.
    Catalyst switches (not sure about Sup2T though) would use a 1, 2, or 3-bit index, which would skew traffic algorthmically if you used a non-power of 2.
    Looking at the 5K documentation, it seems to use the Catalyst style (though haven't been able to confirm). My guess is that whatever is used for the 5Ks is used for the 6100/6200.
    Design wise, this would mean you would want to use powers of 2 for your NB uplinks.

    Hello Tony,
    On UCS FI, it uses " sr-cdest-ip " as the load balancing algorithm and uses 8 parameters for hashing
    6248-01-B(nxos)# sh port-channel load-balance
    Port Channel Load-Balancing Configuration:
    System: source-dest-ip
    Port Channel Load-Balancing Addresses Used Per-Protocol:
    Non-IP: source-dest-mac
    IP: source-dest-ip source-dest-mac
    6248-01-B(nxos)# show platform fwm info pc port-channel 1
    dump pc info: if_index 369098752 dump_all 0 verbose 1
    Po1: state 0x0  #pifs 1  fwimpd ctx 0x9666c1c
    Po1: hash params - l2_da 1 l2_sa 1 l3_da 1 l3_sa 1
    Po1: hash params - l4_da 1 l4_sa 1 xor_sa_da 1 hash_elect 1
    I could not find an option to change these values.
    Padma

  • Load balancing algorithms

    hello
    can any one give URL about load balancing algorithms in 11500 series?
    sincerally

    Hi,
    Here is a link.
    http://www.cisco.com/univercd/cc/td/doc/product/webscale/css/css_730/contlbgd/contrule.htm#1038118
    Hope this helps,
    Regards,
    Sagar

  • Hyper-V, NIC Teaming and 2 hosts getting in the way of each other

    Hey TechNet,
    After my initial build of 2 Hyper-V Core server which took me a bit of time without a domain, I started building 2 more for another site. After the initial two, setting up the new ones went very fast until I ran into a very funny issue. And I am willing
    to bet it is just my luck but I am wondering if any other out there ended up with it.
    So, I build these 2 new servers, create a NIC teaming on each host, add the management OS adapter, give it an IP and I can ping the world. So I went back to my station and tried to start working on these hosts but I kept getting DCed especially from one
    of them. Reinstalled it and remade the NIC teaming config, just in case. Same issue
    So I started pinging both of the servers and I remarked that when one was pinging, the other one tended to not answer ping anymore and vice versa. After testing the firewall and the switch and even trying to put the 2 machines on different switches, did
    not help. So I thought, what the heck, let's just remove all the network config from both machine, reboot, and redo the network config. Since then no issue.
    I only forgot to do one thing before removing the network configuration, I forgot to check if the MAC address on the Management OS adapters were the same. Even if it is a small chance, it can still happen (1 in 256^4 i'd say).
    So to get to my question, am I that unlucky or might it have been something else ?
    Enjoy your weekends

    I raised this bug long ago (one year ago in fact) and it still happens today.
    If you create a virtual switch, then add a management vNIC to it - there are times when you will get two hosts with the same MAC on the vNIC that was added for management.
    I have seen this in my lab (and I can reproduce it at will).
    Modify the entire Hyper-V MAC address pool.  Or else you will have the same issue with VMs.  This is the only workaround.
    But yes, it is a very confusing issue.
    Brian Ehlert
    http://ITProctology.blogspot.com
    Learn. Apply. Repeat.

  • Can you use NIC Teaming for Replica Traffic in a Hyper-V 2012 R2 Cluster

    We are in the process of setting up a two node 2012 R2 Hyper-V Cluster and will be using the Replica feature to make copies of some of the hosted VM's to an off-site, standalone Hyper-V server.
    We have planned to use two physical NIC's in an LBFO Team on the Cluster Nodes to use for the Replica traffic but wanted to confirm that this is supported before we continue?
    Cheers for now
    Russell

    Sam,
    Thanks for the prompt response, presumably the same is true of the other types of cluster traffic (Live Migration, Management, etc.)
    Cheers for now
    Russell
    Yep.
    In our practice we actually use converged networking, which basically NIC-teams all physical NICs into one pipe (switch independent/dynamic/active-active), on top of which we provision vNICs for the parent partition (host OS), as well as guest VMs. 
    Sam Boutros, Senior Consultant, Software Logic, KOP, PA http://superwidgets.wordpress.com (Please take a moment to Vote as Helpful and/or Mark as Answer, where applicable) _________________________________________________________________________________
    Powershell: Learn it before it's an emergency http://technet.microsoft.com/en-us/scriptcenter/powershell.aspx http://technet.microsoft.com/en-us/scriptcenter/dd793612.aspx

  • Load-balancing Algorithm for NX-OS Port Channels

    Hi, all
    I do not understand description of port-channel load-balance ethernet command.
    switch(config)# port-channel load-balance ethernet ?
      destination-ip         Destination IP address
      destination-mac        Destination MAC address
      destination-port       Destination TCP/UDP port
      source-dest-ip         Source & Destination IP address (includes l2)
      source-dest-ip-only    Source & Destination IP addresses only
      source-dest-mac        Source & Destination MAC address
      source-dest-port       Source & Destination TCP/UDP port (includes l2 and l3)
      source-dest-port-only  Source & Destination TCP/UDP port only
      source-ip              Source IP address
      source-mac             Source MAC address
      source-port            Source TCP/UDP port
    Please tell me what the following descriptions mean.
      Source & Destination IP address (includes l2)
      Source & Destination TCP/UDP port (includes l2 and l3)
    What are the meaning of "includes l2" and "includes l2 and l3" ?
    Thank you for your cooperation in advance.

    Hi Satoru,
    On the Nexus 5000/6000 platforms, all FEXs will inherit the global hashing algorithm from the parent device.
    On the Nexus 7000 platform, hashing algorithms can be assigned on a per FEX basis (all load balancing changes must be made from the Admin VDC):
    N7K-A(config)# port-channel load-balance src-dst ip-l4port fex 134
    Any FEX without a hashing algorithm configured with inherit the global hash. Making changes to the modular/global hash will not alter FEX specific hashing algorithms.
    To verify the configuration applied you can use this command:
    N5K_A# show port-channel load-balance
    On the Nexus 7000, the per FEX algorithm can be checked by appending the ‘fex <#>’ to the end of the command in the Admin VDC or the FEX’s respective VDC:
    N7K-A(config)# show port-channel load-balance fex 134
    Regards,
    Richard

  • Hyper-V Nic Teaming (reserve a nic for host OS)

    Whilst setting up nic teaming on my host (server 2012 r2) the OS recommends leaving one nic for host management(access). IS this best practice?  Seems like a waste for a nic as the host would hardly ever be accessed after initial setup.
    I have 4 nics in total. What is the best practice in this situation?

    Depending on if it is a single and the one and only or you build a Cluster you need some networks on your Hyper-V
    at least one connection for the Host to do Management.
    so in case of a single node with local disks you would create a Team with the 4 Nics and create a Hyper-V Switch with the Option checked for creating that Management OS Adapter what is a so called vNIC on that vSwitch and configure that vNIC with the needed
    IP Setting etc...
    If you plan a Cluster and also ISCSI/SMB for Storage Access take a look here
    http://www.thomasmaurer.ch/2012/07/windows-server-2012-hyper-v-converged-fabric/
    You find a few possible ways for teaming and the Switch Settings and also all needed Steps for doing a fully converged Setup via PowerShell.
    If you share more Informations on you setup we can give more Details on that.

  • Hyper-v NIC teaming, external network

    http://blogs.technet.com/b/keithmayer/archive/2012/11/26/configuring-hyper-v-virtual-networking-in-w...Hello Yall's, Wanted to get some advise. I am running Server 2012r2 on the Host OS with 4 NICs... I team two of the NICs The team comes up active.1. I connect the team to a Virtual Switch in Hyper-v2 . I uncheck "Allow management operating system to share this network adapter"3. I connect a VM to this V-switchI feel like I am missing somethingProblem. Can't get no external network on the VM. Please advise.Thank u!
    This topic first appeared in the Spiceworks Community

    I raised this bug long ago (one year ago in fact) and it still happens today.
    If you create a virtual switch, then add a management vNIC to it - there are times when you will get two hosts with the same MAC on the vNIC that was added for management.
    I have seen this in my lab (and I can reproduce it at will).
    Modify the entire Hyper-V MAC address pool.  Or else you will have the same issue with VMs.  This is the only workaround.
    But yes, it is a very confusing issue.
    Brian Ehlert
    http://ITProctology.blogspot.com
    Learn. Apply. Repeat.

  • Mod_weblogic and load balancing algorithm

    Hi,
    we have a problem where we have 2 boxes, each with Apache (doing some header and url rewrites) and configured with mod_weblogic, and weblogic servers with our stateless applications.
    We have a HW loadbalancer in front, doing round-robin balancing, and we would like to configure load balancing feature in the Apaches so that each Apache always tries the request to the weblogic located in the same box, and only if that fails, try to the weblogic in the other box.
    We also have a web applications with sessions, so we would like balancing feature for sticky sessions to apply.
    Can somebody help?

    The loadbalancing section in this post (http://middlewaremagic.com/weblogic/?p=7819) might help you
    You can configure the weblogic module by using, for example,
    LoadModule weblogic_module   "${ORACLE_HOME}/ohs/modules/mod_wl_ohs.so"
    <IfModule weblogic_module>
         ConnectTimeoutSecs 10
         ConnectRetrySecs 2
         DebugConfigInfo ON
         WLSocketTimeoutSecs 2
         WLIOTimeoutSecs 300
         Idempotent ON
         FileCaching ON
         KeepAliveSecs 20
         KeepAliveEnabled ON
         DynamicServerList ON
         WLProxySSL OFF
    </IfModule>
    <Location /b2bconsole>
         SetHandler weblogic-handler
         WebLogicCluster 192.168.1.66:8001,192.168.1.67:9001
    </Location>
    <Location /integration>
         SetHandler weblogic-handler
         WebLogicCluster 192.168.1.66:8001,192.168.1.67:9001
    </Location>
    <Location /soa-infra>
         SetHandler weblogic-handler
         WebLogicCluster 192.168.1.66:8001,192.168.1.67:9001
    </Location>

  • Load balancing algorithm problems with servlets

    All,
    We have a simple servlet which looks up an RMI object from JNDI and invokes
    one of its methods in a loop. The RMI object is the HelloClusterImpl
    example provided by WebLogic. The servlet basically copies the code from
    the HelloClusterClient. In the cluster property file, our load algorithm is
    setup to be round-robin. We are using IIS as our web server. If we start
    up both servers in the cluster and then use another machine as our client to
    call the HelloClusterClient, we see that the calls to HelloClusterImpl
    alternate between the two clustered servers. In fact, it is done perfectly
    at 50% for each server. When we run the servlet from with the client's
    browser, however, it seems as if which ever server in the cluster that
    receives the servlet request then takes 100% of the calls to
    HelloClusterImpl and there is never any alternating. Although we can verify
    that different servers in the cluster receive the servlet request, it seems
    as if there is never any altering between the servers during the method
    invocations on the HelloClusterImpl servant. Does this seem right?
    Shouldn't there be alternating between the servers? Any help would be
    greatly appreciated.
    Thanks,
    -Jon

    I have to try this and I will let you.
    Thanks
    Jon Eagles wrote:
    All,
    We have a simple servlet which looks up an RMI object from JNDI and invokes
    one of its methods in a loop. The RMI object is the HelloClusterImpl
    example provided by WebLogic. The servlet basically copies the code from
    the HelloClusterClient. In the cluster property file, our load algorithm is
    setup to be round-robin. We are using IIS as our web server. If we start
    up both servers in the cluster and then use another machine as our client to
    call the HelloClusterClient, we see that the calls to HelloClusterImpl
    alternate between the two clustered servers. In fact, it is done perfectly
    at 50% for each server. When we run the servlet from with the client's
    browser, however, it seems as if which ever server in the cluster that
    receives the servlet request then takes 100% of the calls to
    HelloClusterImpl and there is never any alternating. Although we can verify
    that different servers in the cluster receive the servlet request, it seems
    as if there is never any altering between the servers during the method
    invocations on the HelloClusterImpl servant. Does this seem right?
    Shouldn't there be alternating between the servers? Any help would be
    greatly appreciated.
    Thanks,
    -Jon

Maybe you are looking for

  • Printing to networked HP7210 problems

    When any of the Macs in our house prints to an HP7210 that is networked through a PC a strange thing happens. It just intakes paper and spits them out blank until the paper tray is empty and then says it is out of paper. Any suggestions? The PCs all

  • Cash flow s_alr_87012271

    Hi friends i have configured the cashflow and forcasting as per the standerd 1 source symbols 2 planning levels 3 Planning groups 4 assigning internal code 5 cash management group 6 cash management grouping structure 7 activating cash management in g

  • Why do I lose signal since updating to iOS 5.1?

    Hi, Since I updated my iOS from 5.0.1 to 5.1, my iPhone: 1) Takes a lot more time finding the network when I turn it on. 2) Frequently loses reception and searches for network for 10-30 seconds before going back on the network full strength. The weir

  • Can't see the update on my App Store

    Can't see the update on my App Store

  • A buzz every few minutes

    Every minute or two I get a buzz from what I'm thinking is the back left side of my Macbook Pro Retina. I even get this buzz when I mute sound (I figured this out when I put volume to max and it didn't get louder). Not sure what it is. I've turned al