NIC Teaming question

Hello All,
I'm setting up a new 2012 R2 server. It will be our new AD/DNS/DCHP server. Question is if I setup NIC teaming, Should I point the client computers DNS to the virtual IP Team NIC address?
Thanks!
JnB

Hi!
If you setup NIC teaming, you get presented a "teamed adapter". You do all your IP config on this thing.
You can't do any config on the members of the team (e.g. Ethernet 1, Ethernet 2). You'll see they just have the Multiplexor Protocol configured.
So yep, you'll configure the team, do everything on the teamed adapter. Point your computers to that IP.

Similar Messages

  • Hyper-V, NIC Teaming and 2 hosts getting in the way of each other

    Hey TechNet,
    After my initial build of 2 Hyper-V Core server which took me a bit of time without a domain, I started building 2 more for another site. After the initial two, setting up the new ones went very fast until I ran into a very funny issue. And I am willing
    to bet it is just my luck but I am wondering if any other out there ended up with it.
    So, I build these 2 new servers, create a NIC teaming on each host, add the management OS adapter, give it an IP and I can ping the world. So I went back to my station and tried to start working on these hosts but I kept getting DCed especially from one
    of them. Reinstalled it and remade the NIC teaming config, just in case. Same issue
    So I started pinging both of the servers and I remarked that when one was pinging, the other one tended to not answer ping anymore and vice versa. After testing the firewall and the switch and even trying to put the 2 machines on different switches, did
    not help. So I thought, what the heck, let's just remove all the network config from both machine, reboot, and redo the network config. Since then no issue.
    I only forgot to do one thing before removing the network configuration, I forgot to check if the MAC address on the Management OS adapters were the same. Even if it is a small chance, it can still happen (1 in 256^4 i'd say).
    So to get to my question, am I that unlucky or might it have been something else ?
    Enjoy your weekends

    I raised this bug long ago (one year ago in fact) and it still happens today.
    If you create a virtual switch, then add a management vNIC to it - there are times when you will get two hosts with the same MAC on the vNIC that was added for management.
    I have seen this in my lab (and I can reproduce it at will).
    Modify the entire Hyper-V MAC address pool.  Or else you will have the same issue with VMs.  This is the only workaround.
    But yes, it is a very confusing issue.
    Brian Ehlert
    http://ITProctology.blogspot.com
    Learn. Apply. Repeat.

  • ESXi 4.1 NIC Teaming's Load-Balancing Algorithm,Nexus 7000 and UCS

    Hi, Cisco Gurus:
    Please help me in answering the following questions (UCSM 1.4(xx), 2 UCS 6140XP, 2 Nexus 7000, M81KR in B200-M2, No Nexus 1000V, using VMware Distributed Switch:
    Q1. For me to configure vPC on a pair of Nexus 7000, do I have to connect Ethernet Uplink from each Cisco Fabric Interconnect to the 2 Nexus 7000 in a bow-tie fashion? If I connect, say 2 10G ports from Fabric Interconnect 1 to 1 Nexus 7000 and similar connection from FInterconnect 2 to the other Nexus 7000, in this case can I still configure vPC or is it a validated design? If it is, what is the pro and con versus having 2 connections from each FInterconnect to 2 separate Nexus 7000?
    Q2. If vPC is to be configured in Nexus 7000, is it COMPULSORY to configure Port Channel for the 2 Fabric Interconnects using UCSM? I believe it is not. But what is the pro and con of HAVING NO Port Channel within UCS versus HAVING Port Channel when vPC is concerned?
    Q3. if vPC is to be configured in Nexus 7000, I understand there is a limitation on confining to ONLY 1 vSphere NIC Teaming's Load-Balancing Algorithm i.e. Route Based on IP Hash. Is it correct?
    Again, what is the pro and con here with regard to application behaviours when Layer 2 or 3 is concerned? Or what is the BEST PRACTICES?
    I would really appreciate if someone can help me clear these lingering doubts of mine.
    God Bless.
    SiM

    Sim,
    Here are my thoughts without a 1000v in place,
    Q1. For me to configure vPC on a pair of Nexus 7000, do I have to connect Ethernet Uplink from each Cisco Fabric Interconnect to the 2 Nexus 7000 in a bow-tie fashion? If I connect, say 2 10G ports from Fabric Interconnect 1 to 1 Nexus 7000 and similar connection from FInterconnect 2 to the other Nexus 7000, in this case can I still configure vPC or is it a validated design? If it is, what is the pro and con versus having 2 connections from each FInterconnect to 2 separate Nexus 7000?   //Yes, for vPC to UCS the best practice is to bowtie uplink to (2) 7K or 5Ks.
    Q2. If vPC is to be configured in Nexus 7000, is it COMPULSORY to configure Port Channel for the 2 Fabric Interconnects using UCSM? I believe it is not. But what is the pro and con of HAVING NO Port Channel within UCS versus HAVING Port Channel when vPC is concerned? //The port channel will be configured on both the UCSM and the 7K. The pro of a port channel would be both bandwidth and redundancy. vPC would be prefered.
    Q3. if vPC is to be configured in Nexus 7000, I understand there is a limitation on confining to ONLY 1 vSphere NIC Teaming's Load-Balancing Algorithm i.e. Route Based on IP Hash. Is it correct? //Without the 1000v, I always tend to leave to dvSwitch load balence behavior at the default of "route by portID". 
    Again, what is the pro and con here with regard to application behaviours when Layer 2 or 3 is concerned? Or what is the BEST PRACTICES? UCS can perform L2 but Northbound should be performing L3.
    Cheers,
    David Jarzynka

  • Load Balancing and NIC Teaming

    Hi! i have been looking through lots of links and none of them actually can fully answer my queries.
    I am to do a writeup on load balancing and NIC Teaming, is there any1 that knows what are the commonly used load balancing and NIC Teaming methods, when to use each method, and the advantages and disadvantages of each method and the configuration for each
    method!
    Sorry its lots of questions but i have to do a detailed writeup!
    Many thanks in advance :D

    HI
    NIC Teaming - On a single server, you will have mutiple NIC. You can Team the NIC so that both NIC will act togather to provide better bandwidth and High avaliblity.
    Example : NIC 1 - 1 GB and NIC -2 1 GB so in Team it can act a 2 GB single NIC, If one fails speed will be reduced but it will have HA
    Loadbalancing : Two servers hosting same content:
    Example : Microsoft.com can be hosted in two or even more servers and a loadbalancer will be used to split load to each server based of the current load and traffic.
    No disadvantages

  • NIC teaming - Server 2008 R2 DC combined with other Software

    Hello!
    I've been searching all morning for an answer of what we have in mind to do at work....
    We've got a server installed with Windows Server 2008 R2 and have 4 NICs on it. We want to make it a DC (with DNS, DHCP and print services) and also want to install our Backup Solution (from Veeam) for our VMs. This server will be the only physical Microsoft
    server next to our 3 ESX servers at the end.
    I read here (http://markparris.co.uk/2010/02/09/top-tipactive-directory-domain-controllers-and-teamed-network-cards/) that there is a statement that a DC with NIC teaming is only using the FO (Fail-Over) feature of the teaming. Since there is also the backup
    solution on this server, it would be great also to use the LB (Load-Balancing) feature. My question is, when I active NIC teaming and install the DC roles, does the roles just use the FO feature and neglect the LB feature or does it enable/disable those modes/features
    of NIC teaming? Cause it would be nice if the backup solution could use the LB for bigger bandwidth for backup and restores and I wouldn't really care about the FO for the DC role.
    cheers
    Ivo

    Hi,
    I think the issue is related to the third party NIC teaming solution. You can refer to the third party manufacture.
    Here I should remind you something else, a DC with multiple NICs will cause many problems. So I would recommend you run a dedicated
    Hyper-v server and promote a DC on one of the virtual machine.
    Hope this helps.

  • Relationship between coherence and NIC teaming

    Hi,
    We are using Tangosol coherence for clustering purpose in our product Webmethods Integration server.
    When our server starts up it tries to jojn tne cluster.
    Our scenario is this :-
    We have 2 servers running on 2 separate boxes A&B.
    They are on same network segment.
    Multicast test is working properly .
    The issue is only one of the nodes(which is started first) in becoming the part of the cluster and other one remain disabled.
    We found out that the NIC teaming was disabled in the boxes.
    When we enabled NIC teaming with smart load balancing then both the nodes are able to join the cluster.
    My specific question is,
    Is there any relationship between Tangosol coherence and NIC teaming? If yes, what's the relationship.
    Regards,
    Ritwik Bhattacharyya

    I did some tinkering a while back trying to get 4Gb/s bonded etherchannels going on linux boxes but I had issues with out of order and missing packets:
    4Gb/s bonded ethernet test results - finally...
    But to answer your question there is no reason that you would need NIC teaming on in order to make Coherence work. It sounds like something is not configured correctly with your NIC or switch. Maybe try connecting the machines with a crossover cable instead of a switch just to eliminate the switch as a possible problem. It sounds like maybe you're just using the wrong ethernet port on a server or something.
    -Andrew

  • Windows Server 2012/2012R2 NIC Teaming Mode

    Hi,
    Question 1:
    In Windows Server 2012 the following teaming mode was recommended for Hyper-V NIC teams:
    Teaming mode: Switch Independent
    Load balancing mode: Hyper-V Port
    All Adapers Active
    In a session at TechEd 2014 it was stated that Dynamic is the new recommendation for Windows Server 2012 R2. However, a Microsoft PFE stated a few weeks ago that he would still recommend Hyper-V Port for Windows Server 2012 R2. What is your opinions around
    this?
    Question 2:
    We have a Hyper-V Failover Cluster which isn`t migrated to 2012 R2 yet, it`s running 2012. In this cluster we use Switch Independent/Hyper-V Port for the team. We also use converged networking, having 2 physical adapters bound to the NIC team, as well as
    3 virtual adapters in the management OS for management, CSV and Live Migration. Recently one of the team NICs failed, and this incident also caused the cluster membership on the affected node to go offline even though the other team NIC was
    connected. Is this expected behaviour? Would the behaviour be different if 2012 R2 with Dynamic mode was being used?

    Hello,
    As for question number 1:
    For Hyper-V workload it's recommended to use Dynamic with
    Switch Independent mode. Why?
    This configuration will distribute the load based on the TCP Ports address hash as modified by the Dynamic load balancing algorithm. The Dynamic load balancing algorithm will redistribute flows to optimize team member bandwidth utilization so individual
    flow transmissions may move from one active team member to another.  The algorithm takes into account the small possibility that redistributing traffic could cause out-of-order delivery of packets so it takes steps to minimize that possibility.
    The receive side, however, will look identical to Hyper-V Port distribution.  Each Hyper-V switch port’s traffic, whether bound for a virtual NIC in a VM (vmNIC) or a virtual NIC in the host (vNIC), will see all its inbound traffic arriving on a single
    NIC.
    This mode is best used for teaming in both native and Hyper-V environments except when:
    1) Teaming is being performed in a VM,
    2) Switch dependent teaming (e.g., LACP) is required by policy, or
    3) Operation of a two-member Active/Standby team is required by policy. 
    As for question number 2:
    The Switch Independent/Hyper-V Port will send packets using all active team members distributing the load based on the Hyper-V switch port number.  Each Hyper-V port will be bandwidth limited to not more than one team member’s bandwidth because the port
    is affinitized to exactly one team member at any point in time. 
    In all cases where this configuration was recommended back in Windows Server 2012 the new configuration in 2012 R2, Switch Independent/Dynamic, will provide better performance.
    Microsoft recommend for a clustered Hyper-V deployment
    in Windows server 2012 to use Switch Independent/Hyper-V Port as you mentioned and to configure
    Hyper-V QoS that applies to the virtual switch. (Configure minimum bandwidth in
    weight mode instead of in bits per second and Enable and configure QoS
    for all virtual network adapters 
    Did you apply QoS on the Converged vSwitch after you
    created the team?? However Nodes are considered down if they do not respond to 5 heartbeats. The Switch Independent/Hyper-V Port does not cause the cluster to goes down if one NIC failed. The issue is somewhere else and not in the teaming mode
    that you choose.
    Hope this help.
    Regards,
    Charbel Nemnom
    MCSA, MCSE, MCS, MCITP
    Blog: www.charbelnemnom.com
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if
    a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Hyper-V NIC Team Load Balancing Algorithm: TranportPorts vs Hyper-VPorts

    Hi, 
    I'm going to need to configure a NIC team for the LAN traffic for a Hyper-V 2012 R2 environment. What is the recommended load balancing algorithm? 
    Some background:
    - The NIC team will deal with LAN traffic (NOT iSCSI storage traffic)
    - I'll set up a converged network. So there'll be a virtual switch on top of this team, which will have vNICs configured for each cluster, live migration and management
    - I'll implement QOS at the virtual switch level (using option -DefaultFlowMinimumBandwidthWeight) and at the vNIC level (using option -MinimumBandwidthWeight)
    - The CSV is set up on an Equallogics cluster. I know that this team is for the LAN so it has nothing to do with the SAN, but this reference will become clear in the next paragraph. 
    Here's where it gets a little confusing. I've checked some of the Equallogics documentation to ensure this environment complies with their requirements as far as storage networking is concerned. However, as part of their presentation the Dell publication
    TR1098-4, recommends creating the LAN NIC team with the TrasportPorts Load Balancing Algorithm. However, in some of the Microsoft resources (i.e. http://technet.microsoft.com/en-us/library/dn550728.aspx), the recommended load balancing algorithm is HyperVPorts.
    Just to add to the confusion, in this Microsoft TechEd presentation, http://www.youtube.com/watch?v=ed7HThAvp7o, the recommendation (at around minute 8:06) is to use dynamic ports algorithm mode. So obviously there are many ways to do this, but which one is
    correct? I spoke with Equallogics support and the rep said that their documentation recommends TransportPorts LB algorithm because that's what they've tested and works. I'm wondering what the response from a Hyper-V expert would be to this question. Anyway,
    any input on this last point would be appreciated.

    Gleb,
    >>See Windows Server 2012 R2 NIC Teaming (LBFO) Deployment and Management  for more
    info
    Thanks for this reference. It seems that I have an older version of this document where there's absolutely
    no mention of the dynamic LBA. Hence my confusion when in the Microsoft TechEd presentation the
    recommendation was to use Dynamic. I almost implemented this environment with switch dependent and Address Hash Distribution because, based on the older version of the document, this combination offered: 
    a) Native teaming for maximum performance and switch diversity is not required; or
    b) Teaming under the Hyper-V switch when an individual VM needs to be able to transmit at rates in excess of what one team member can deliver
    The new version of the document recommends Dynamic over the other two LBA. The analogy that the document
    makes of TCP flows with human speech was really helpful for me to understand what this algorithm is doing. For those who will never read the document, I'm referring to this: 
    "The outbound loads in this mode are dynamically balanced based on the concept of
    flowlets.  Just as human speech has natural breaks at the ends of words and sentences, TCP flows (TCP communication streams) also have naturally
    occurring breaks.  The portion of a TCP flow between two such breaks is referred to as a flowlet.  When the dynamic mode algorithm detects that a flowlet boundary has been encountered, i.e., a break of sufficient length has occurred in the TCP flow,
    the algorithm will opportunistically rebalance the flow to another team member if apropriate.  The algorithm may also periodically rebalance flows that do not contain any flowlets if circumstances require it.    As a result the affinity
    between TCP flow and team member can change at any time as the dynamic balancing algorithm works to balance the workload of the team members. "
    Anyway, this post made my week. You sir are deserving of a beer!

  • IBM MCS Server RUnning Unity with NIC Teaming

    All,
    has anyone ever run NIC teaming on and IBM MCS Server with Unity before? At question is the fact that many severs create a virtual MAC address that is different from either of the actual MACs when you team NICs. If this is the case on the IBM Servers then we may need to request an updated license. The servers are branded as Cisco MCS-7835-I1
    units.
    Thanks in advance. All replies rated!

    Hi
    According to this doc: http://www.cisco.com/en/US/prod/collateral/voicesw/ps6790/ps5748/ps378/product_solution_overview0900aecd80091615.html
    The 7825-I4 is an IBM IBM x3250-M2. You can use that model number to search the IBM drivers page for the things you need:
    http://www-933.ibm.com/support/fixcentral/systemx/selectFixes
    I think the NIC teaming comes with the Network Drivers.
    Regards
    Aaron
    Please rate helpful posts...

  • NIC teaming creates packet loss (Windows 2008 R2)?

    I'm experiencing some packet loss to all of our VMs that we didn't have before we made some changes to our Hyper-V implementation (Windows 2008 R2). Most of the VMs also run 2008 R2 - with 3 that run Server 2003.
    The host server is a Dell R610 with three 4 port NICS - two Intel quad port gigabit and a quad port Broadcom. 
    We us the individual ports of the Broadcom for host management and live migration - no problems here. We use the Intel cards for both iSCSI and VM networks. Calling the two intel cards “A” and “B”, and the ports P1-4 we've used AP1, AP2, BP1, BP2 (ports
    1 & 2 of both Intel NICs) for iSCSI connections, and we've created a NIC Team with AP3, AP4, BP3, and BP4 (ports 3 and 4 of both Intel NICs). The team type is "Virtual Machine Load Balancing". We then created a Hyper-V switch based on this team
    for use with all of the VMs created on the host. (as a side note: prior to implementing the NIC team, we just had 4 Hyper-V switches, one associated with each of these 4 ports.)
    The 4 ports of the NIC team are connected to two different Cisco SG200 switches - AP3 and BP3 are connected to switch1, and AP4 and BP4 are connected to switch2 (in an attempt to maximize redundancy). The two Cisco SG200s are simply connected to the rest
    of our network - each to a different switch within the subnet. There is minimal configuration done to the SG200s (for example NO
     link aggregation); spanning tree is enable however.
    My question is: can the network cables be connected to different switches (as they currently are) and if so is there some configuration piece (either on the switch or within Windows) that I'm missing? 
    What are the options here if this configuration is incorrect? The packet loss is in the range of 0.1%, but we've had odd spikes where a VM was essentially unavailable for a brief period (a few minutes) then returned to "normal" (0,1% loss). 
    Pinging a device (like the SG200 itself) or another physical server (for example our domain controller or the hyper-v host itself) results in essentially 0 loss; maybe one or two packets during the course of a 12 hour ping (this was the “normal” ping
    response to VMs before we created the NIC team, so I’m quite sure this has something to do with it).
    Thanks in advance!

    I believe when utilizing the Virtual Machine Load Balancing the ports must be connected to the same switch, stack, or chassis as the arp for the MAC could move.  I believe, although I could be wrong, that the outages you see is when the machine "moves"
    between ports and the arp being updated between the two switches. 
    I believe you are looking for switch fault tolerance teaming which will allow for the failure of adapter, cabling, or switch which will achieve your goal of maximum redundancy.  This is achieved via spanning tree on the switches, which you indicated
    is already configured.
     

  • ProLiant DL380 Gen9 - NIC Teaming

    Hi community,
    since one week we have new DL380 Gen9 servers. Now we want to install them as a Win 2008 R2 server, but we cannot create a NIC teaming.
    If I install the HP NCO (Network Configuration Utility), cp023339.exe Version 10.90, I get
    "The software will bot be installed on the system because the required hardware is not present in the system or the software/firmware doesn't apply to the system"
    I updated driver / firmware of the NIC, searched any new HP Network Configuration Tools, nothing.
    NIC: HP Ethernet 1GB 4-port 331i Adapter
    Can someone help?
    Many thanks
    /Hugo
    This question was solved.
    View Solution.

    Hi:
    You may also want to post your question on the HP Business Support Forum -- DL Servers section.
    http://h30499.www3.hp.com/t5/ProLiant-Servers-ML-DL-SL/bd-p/itrc-264#.VHM-jHktC9I

  • HP NIC Team with MDT

    I am trying to deploy servers with MDT and as a requirement would need to team NIC, We use HP servers and they provide a utility
    CQNICCMD to get that done.
    Have referred the below link for the command/switches.
    http://h20564.www2.hp.com/hpsc/doc/public/display?docId=c04024934
    I am not sure where in the task sequence should I get the NIC team done cause they keep failing where every I put it.Any documentation on how to get this done will be of great help.

    I would place the command during the State Restore phase, somewhere during custom steps. Assuming that the Teaming can be done in the regular OS.
    The question is: What are the failures?
    Please note that if you are configuring the only NIC's on the machine *AND* running the scripts from *over* the network, your script could suddenly find it self running without access to it's own source. I would devise a method to copy the scripts/install
    program/drivers locally, kick off the script locally, and then wait for the network to be available.
    Keith Garner - Principal Consultant [owner] -
    http://DeploymentLive.com

  • NIC Teaming for hyper-v serve.

    I have installed windows server 2012 r2 on server. Server is having network adapters i have given static ip address to both nic's
    LAN1: - 192.168.0.100 & LAN2: - 192.168.0.101 after enabling NIC Teaming server have added one more adapter called
    "Network Adapter Multiplexor" after this above mentioned ip address are not responding to PING or any requests. Then i have given
    192.168.0. 102 ip address to Multiplexor and its started working.
    So my question do i need to give ip address to LAN1 &
    LAN2 or i can just create team and give ip address to Multiplexor
    Also if i installed hyper-v server on it will it give me failover thing for this machine.????
    Akshay Pate

    Hello Akshay,
    In brief, after creating the Teaming adapter (Multiplexor) you'll use it's address for future networking purposes.
    Regarding the lack of ping, I had the same "issue" and it seems like is block by the Microsoft code itself. Still couldn't find how to allow it.
    When digging into W2k12(&R2) NIC Teaming this two pages were very explanative and usefull:
    Geek of All Trades: The Availability Answer (by Greg Shields)
    Windows server 2012 Hyper-V 3.0 network virtualization (if you need more technical detail)
    Hope it helps!

  • Server 2012 R2 Crashes with NIC Team

    Server 2012 R2 Core configured for Hyper-V. Using 2-port 10Gbe Brocades, we want to use NIC teaming for guest traffic. Create the team... seems fine. Create the virtual switch in Hyper-V, and assign it to the NIC team... seems fine. Create
    a VM, assign the network card to the Virtual switch... still doing okay. Power on the VM... POOF! The host BSOD's. If I remove the switch from the VM, I can run the VM from the console, install the OS, etc... but as soon as I reassign the virtual
    NIC to the switch, POOF! Bye-bye again. Any ideas here?
    Thank you in advance!
    EDIT: A little more info... Two 2-port Brocades and two Nexus 5k's. Running one port on NIC1 to one 5k, and one port on NIC2 to the other 5k. NIC team is using Switch Independent Mode, Address Hash load balancing, and all adapters active.

    Hi,
    Have you updated the NIC driver to latest?
    If issue persists after updating the driver, we can use WinDbg to analyze a crash dump.
    If the NIC driver cause the BSOD, please consult the NIC manufacture about this issue.
    For detailed information about how to analyze a crash dump, please refer to the link below,
    http://blogs.technet.com/b/juanand/archive/2011/03/20/analyzing-a-crash-dump-aka-bsod.aspx
    Best Regards.
    Steven Lee
    TechNet Community Support

  • Windows 7/8.0/8.1 NIC teaming issue

    Hello,
    I'm having an issue with Teaming network adapters in all recent Windows client OSs.
    I'm using Intel Pro Dual Port or Broadcom NetExtreme II GigaBit adapters with the appropriate drivers/applications from the vendors.
    I am able to set up teaming and fail-over works flawlessly, but the connection will not use the entire advertised bandwidth of 2Gbps. Basically it will use either one port or the other.
    I'm doing the testing with the iperf tool and am communicating with a unix based server.
    I have the following setup:
    Dell R210 II server with 2 Broadcom NetEtreme II adapters and a DualPort Intel Pro adapter - Centos 6.5 installed bonding configured and working wile communicating with other unix based systems.
    Zyxel GS2200-48 switch - Link Aggregation configured and working
    Dell R210 II with Windows 8.1 with Broadcom NetExtreme II cards or Intel Pro dualport cards.
    For the Windows machine I have also tried Windows 7 and Windows 8, also non server type hardware with identical results.
    so.. Why am I not getting > 1 Gbps throughput on the created team? although load balancing is activated, team adapter says the connection type is 2 Gbps, a the same setup with 2 unix machines works flawlessly.
    Am I to understand that Link Aggregation (802.3ad) under Microsoft OS does not support load balancing if connection is only towards one IP?
    To make it clear, I need client version of Windows OS to communicate unix based OS over a higher then 1Gbps bandwidth (as close to 2 Gbps as possible). Without the use of 10 Gbps network adapters.
    Thanks in advance,
    Endre

    As v-yamliu has mentioned, NIC teaming through the operating system is
    only available in Windows Server 2012 and Windows Server 2012 R2. For Windows Client or for previous versions of Windows Server you will need to create the team via the network driver. For Broadcom this is accomplished
    using the Broadcom Advanced Server Program (BASP) as documented here and
    for Intel via Advanced Network Services as documented here.
    If you have configured the team via the drivers, you may need to ensure the driver is properly installed and updated. You may also want to ensure that the adapters are configured for aggregation (802.3ad/802.1ax/LACP), rather than fault tolerance or load
    balancing and that the teaming configuration on the switch matches and is compatible with the server configuration. Also ensure that all of the links are connecting at full duplex as this is a requirement.
    Brandon
    Windows Outreach Team- IT Pro
    The Springboard Series on TechNet

Maybe you are looking for