Cisco Nexus 1000v Virtual Switch for Hyper-V Availability

Hi,
Does anyone have any information on the availability of the Cisco Nexus 1000v virtual switch for Hyper-V. Is it available to download from Cisco yet? If not when will it be released? Are there any Beta programs etc?
I can download the 1000v for VmWare but cannot find any downloads for the Hyper-V version.
Microsoft Partner

Any updates on the Cisco Nexus 1000v virtual switch for Hyper-V? Just checked on the Cisco site, however still only the download for VMware and no trace of any beta version. Also posted the same question at:
http://blogs.technet.com/b/schadinio/archive/2012/06/09/windows-server-2012-hyper-v-extensible-switch-cisco-nexus-1000v.aspx
"Hyper-V support isn't out yet. We are looking at a beta for Hyper-V starting at the end of February or the begining of March. "
-Ian @ Cisco Community
|| MCITP: EA, VA, EMA, Lync SA, makes a killer sandwich. ||

Similar Messages

  • Cisco Nexus 1000v on Hyper-v 2012 R2

    Dears;
    I have deployed Cisco Nexus 1000v on Hyper-v hosts 2012 R2, and I'm in phase of testing and exploring feature, while doing this I removed the Nexus Virtual Switch {VEM} from HOST, it disappeared from host but I couldn't use the uplink attached previously with the switch as it sees it still attached on Nexus 1000v. I tried to remove it by several ways finally the host gets unusable and I had to setup the host again.
    the question here; there is no mention on cisco documents for how to uninstall or remove the VEM attached to a host, can any one help in this ?
    Thanks
    Regards

    Zoning is generally a term used with fibre channel, but I think I understand what you mean.
    Microsoft Failover Clusters rely on shared storage.  So you would configure your storage so that it is accessible from all three nodes of the cluster.  Any LUN you want to be part of the cluster should be presented to all nodes.  With iSCSI,
    it is recommended to use two different IP subnets and configure MPIO.  The LUNs have to be formatted as NTFS volumes.  Run the cluster validation wizard once you think you have things configured correctly.  It will help you find any potential
    configuration issues.
    After you have run a cluster validation and there aren't any warnings left that you can't resolve, build the cluster.  The cluster will form with the available LUNs as storage to the cluster.  Configure the storage to be Cluster Shared Volumes
    for the VMs, and left the witness as the witness.  By default, the cluster will take the smallest LUN to be the witness disk.  If you are just using the cluster for Hyper-V (recommended) you do not need to assign drive letters to any of the disks. 
    You do not need, nor is it recommended to use, pass-through disks.  There are many downsides to using pass through disks, and maybe one benefit, and that one is very iffy.
    . : | : . : | : . tim

  • Using NIC Teaming and a virtual switch for Windows Server 2012 host networking and Hyper-V.

    Using NIC Teaming and a virtual switch for Windows Server 2012 host networking!
    http://www.youtube.com/watch?v=8mOuoIWzmdE
    Hi thanks for reading. Now I may well have my terminology incorrect here so I will try to explain  as best I can and apologies from the start.
    It’s a bit of both Hyper-v and Server 2012R2. 
    I am setting up a lab with Server 2012 R2. I have several physical network cards that I have teamed called “HostSwitchTeam” from those I have made several Virtual Network Adaptors such as below
    examples.
    New-VMSwitch "MgmtSwitch" -MinimumBandwidthMode weight -NetAdaptername "HostSwitchTeam" -AllowManagement $false
    Add-VMNetworkAdapter -ManagementOS -Name "Vswitch" -SwitchName "MgmtSwitch"
    Add-VMNetworkAdapter -ManagementOS -Name "Cluster" -SwitchName "MgmtSwitch"
    When I install Hyper-V and it comes to adding a virtual switch during installation it only shows the individual physical network cards and the
    HostSwitchTeam for selection.  When installed it shows the Microsoft Network Multiplexor Driver as the only option. 
    Is this correct or how does one use the Vswitch made above and incorporate into the Hyper-V so a weight can be put against it.
    Still trying to get my head around Vswitches,VMNetworkadapters etc so somewhat confused as to the way forward at this time so I may have missed the plot altogether!
    Any help would be much appreciated.
    Paul
    Paul Edwards

    Hi P.J.E,
    >>I have teams so a bit confused as to the adapter bindings and if the teams need to be added or just the vEthernet Nics?.
    Nic 1,2 
    HostVMSwitchTeam
    Nic 3,4,5
             HostMgmtSwitchTeam
    >>The adapter Binding settings are:
    HostMgmtSwitchTeam
    V-Curric
    Nic 3
    Nic 4
    Nic 5
    V-Livemigration
    HostVMSwitch
    Nic 1
    Nic 2
    V-iSCSI
    V-HeartBeat
    Based on my understanding of the description , "HostMgmtSwitchTeam and
    HostVMSwitch " are teamed NIC .
    You can think of them as two physical NICs (do not use NIC 1,2,3,4,5 any more , there are just two NICs "HostMgmtSwitchTeam and
    HostVMSwitch").
    V-Curric,
    V-Livemigration , V-iSCSI ,
    V-HeartBeat are just VNICs of host  (you can change their name then check if the virtual switch name will be changed )
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Ask the Expert: Different Flavors and Design with vPC on Cisco Nexus 5000 Series Switches

    Welcome to the Cisco® Support Community Ask the Expert conversation.  This is an opportunity to learn and ask questions about Cisco® NX-OS.
    The biggest limitation to a classic port channel communication is that the port channel operates only between two devices. To overcome this limitation, Cisco NX-OS has a technology called virtual port channel (vPC). A pair of switches acting as a vPC peer endpoint looks like a single logical entity to port channel attached devices. The two devices that act as the logical port channel endpoint are actually two separate devices. This setup has the benefits of hardware redundancy combined with the benefits offered by a port channel, for example, loop management.
    vPC technology is the main factor for success of Cisco Nexus® data center switches such as the Cisco Nexus 5000 Series, Nexus 7000 Series, and Nexus 2000 Series Switches.
    This event is focused on discussing all possible types of vPC along-with best practices, failure scenarios, Cisco Technical Assistance Center (TAC) recommendations and troubleshooting
    Vishal Mehta is a customer support engineer for the Cisco Data Center Server Virtualization Technical Assistance Center (TAC) team based in San Jose, California. He has been working in TAC for the past 3 years with a primary focus on data center technologies, such as the Cisco Nexus 5000 Series Switches, Cisco Unified Computing System™ (Cisco UCS®), Cisco Nexus 1000V Switch, and virtualization. He presented at Cisco Live in Orlando 2013 and will present at Cisco Live Milan 2014 (BRKCOM-3003, BRKDCT-3444, and LABDCT-2333). He holds a master’s degree from Rutgers University in electrical and computer engineering and has CCIE® certification (number 37139) in routing and switching, and service provider.
    Nimit Pathak is a customer support engineer for the Cisco Data Center Server Virtualization TAC team based in San Jose, California, with primary focus on data center technologies, such as Cisco UCS, the Cisco Nexus 1000v Switch, and virtualization. Nimit holds a master's degree in electrical engineering from Bridgeport University, has CCNA® and CCNP® Nimit is also working on a Cisco data center CCIE® certification While also pursuing an MBA degree from Santa Clara University.
    Remember to use the rating system to let Vishal and Nimit know if you have received an adequate response. 
    Because of the volume expected during this event, Vishal and Nimit might not be able to answer every question. Remember that you can continue the conversation in the Network Infrastructure Community, under the subcommunity LAN, Switching & Routing, shortly after the event. This event lasts through August 29, 2014. Visit this forum often to view responses to your questions and the questions of other Cisco Support Community members.

    Hello Gustavo
    Please see my responses to your questions:
    Yes almost all routing protocols use Multicast to establish adjacencies. We are dealing with two different type of traffic –Control Plane and Data Plane.
    Control Plane: To establish Routing adjacency, the first packet (hello) is punted to CPU. So in the case of triangle routed VPC topology as specified on the Operations Guide Link, multicast for routing adjacencies will work. The hellos packets will be exchanged across all 3 routers and adjacency will be formed over VPC links
    http://www.cisco.com/c/en/us/td/docs/switches/datacenter/nexus5000/sw/operations/n5k_L3_w_vpc_5500platform.html#wp999181
    Now for Data Plane we have two types of traffic – Unicast and Multicast.
    The Unicast traffic will not have any forwarding issues, but because the Layer 3 ECMP and port channel run independent hash calculations there is a possibility that when the Layer 3 ECMP chooses N5k-1 as the Layer 3 next hop for a destination address while the port channel hashing chooses the physical link toward N5k-2. In this scenario,N5k-2 receives packets from R with the N5k-1 MAC as the destination MAC.
    Sending traffic over the peer-link to the correct gateway is acceptable for data forwarding, but it is suboptimal because it makes traffic cross the peer link when the traffic could be routed directly.
    For that topology, Multicast Traffic might have complete traffic loss due to the fact that when a PIM router is connected to Cisco Nexus 5500 Platform switches in a vPC topology, the PIM join messages are received only by one switch. The multicast data might be received by the other switch.
    The Loop avoidance works little different across Nexus 5000 and Nexus 7000.
    Similarity: For both products, loop avoidance is possible due to VSL bit
    The VSL bit is set in the DBUS header internal to the Nexus.
    It is not something that is set in the ethernet packet that can be identified. The VSL bit is set on the port asic for the port used for the vPC peer link, so if you have Nexus A and Nexus B configured for vPC and a packet leaves Nexus A towards Nexus B, Nexus B will set the VSL bit on the ingress port ASIC. This is not something that would traverse the peer link.
    This mechanism is used for loop prevention within the chassis.
    The idea being that if the port came in the peer link from the vPC peer, the system makes the assumption that the vPC peer would have forwarded this packet out the vPC-enabled port-channels towards the end device, so the egress vpc interface's port-asic will filter the packet on egress.
    Differences:  In Nexus 5000 when it has to do L3-to-L2 lookup for forwarding traffic, the VSL bit is cleared and so the traffic is not dropped as compared to Nexus 7000 and Nexus 3000.
    It still does loop prevention but the L3-to-L2 lookup is different in Nexus 5000 and Nexus 7000.
    For more details please see below presentation:
    https://supportforums.cisco.com/sites/default/files/session_14-_nexus.pdf
    DCI Scenario:  If 2 pairs are of Nexus 5000 then separation of L3/L2 links is not needed.
    But in most scenarios I have seen pair of Nexus 5000 with pair of Nexus 7000 over DCI or 2 pairs of Nexus 7000 over DCI. If Nexus 7000 are used then L3 and L2 links are required for sure as mentioned on above presentation link.
    Let us know if you have further questions.
    Thanks,
    Vishal

  • VSM and Cisco nexus 1000v

    Hi,
    We are planning to install Cisco Nexus 1000v in our environment. Before we want to install we want to explore little bit about Cisco Nexus 1000v
    •  I know there is 2 elements for Cisco 1k, VEM and VSM. Does VSM is required? Can we configure VEM individually?
    •   How does Nexus 1k integrated with vCenter. Can we do all Nexus 1000v configuration from vCenter without going to VEM or VSM?
    •   In term of alarming and reporting, does we need to get SNMP trap and get from individual VEM or can be use VSM to do that. OR can we   get    Cisco Nexus 1000v alarming and reporting form VMware vCenter.
    •  Apart from using Nexus 1010 can what’s the recommended hosting location for VSM, (same Host as VEM, different VM, and different physical server)
    Foyez Ahammed

    Hi Foyez,
    Here is a brief on the Nexus1000v and I'll answer some of your questions in that:
    The Nexus1000v is a Virtual Distributed Switch (software based) from Cisco which integrated with the vSphere environment to provide uniform networking across your vmware environment for the host as well as the VMs. There are two components to the N1K infrastructure 1) VSM 2) VEM.
    VSM - Virtual supervisor module is the one which controls the entire N1K setup and is from where the configuration is done for the VEM modules, interfaces, security, monitoring etc. VSM is the one which interacts with the VC.
    VEM - Virtual ethernet module are simply the module or virtual linecards which provide the connectivity option or virtual ports for the VMs and other virtaul interfaces. Each ESX host today can only have one VEM. These VEMs recieve their configuration / programing from the VSM.
    If you are aware of any other switching products from Cisco like the Cat 6k switches, the n1k behaves the same way but in a software / virtual environment. Where the VSM are equal of a SUPs and the VEM are similar to the line cards. The control and the packet VLANs in the n1k provide the same kind of AIPC and Inband connectivity as the 6k backplane would for the communication between the modules and the SUP (VSM in this case).
    *The n1k configuration is done only from the VSM and is visible in the VC.However the port-profiles created from the VSM are pushed from the VSM to the VC and have to be assigned to the virtual / physical ports from the VC.
    *You can run the VSM either on the Nexus1010 as a Virtual service blade (VSB) or as a normal VM on any of the ESX/ESXi server. The VSM and the VEM on the same server are fully supported.
    You can refer the following deployment guide for some more details: http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/guide_c07-556626.html
    Hope this answers your queries!
    ./Abhinav

  • Cisco Nexus 1000V InterCloud

    Need download link for Cisco Nexus 1000V InterCloud

    We had a simliar issue with 5.2(1)SV3(1.3) and found this in the release notes:
    ERSPAN
    If the ERSPAN source and destination are in different subnets, and if the ERSPAN source is an L3 control VM kernel NIC attached to a Cisco Nexus 1000V VEM, you must enable proxy-ARP on the upstream switch.
    If you do not enable proxy-ARP on the upstream switch (or router, if there is no default gateway), ERSPAN packets are not sent to the destination.
    http://www.cisco.com/c/en/us/td/docs/switches/datacenter/nexus1000/sw/5_x/release_notes/b_Cisco_N1KV_VMware_521SV313_ReleaseNotes.html#concept_652D9BADC4B04C0997E7F6C29A2C8B1F
    After enabling 'ip proxy-arp' on the upstream SVI it started working properly.

  • Cisco Nexus 1000V - DMZ - ARP

    Hi there,
    Thanks for reading.
    I have a VM (VM1) connected to a Nexus 1000V distributed switch. The 1000V has a connection to our DMZ (physically, an interface on our Cisco ASA 5520) which has 3 other VMs that are successfully serving up in the DMZ. The problem is that a SHOW ARP run on the ASA shows the other VM's MAC addresses but not VM1.
    The vSphere properties for all VMs (including VM1) participating in the DMZ are the same:
    Network label
    VLAN ID
    Port Group
    State - Link Up
    DirectPath I/O - Inactive "Direct Path I/O has been explicitly disabled for this port"
    The one major difference between VM1 and the others is they are multihomed and have a foot in our private network space. I think the absence of a private IP on VM1 is not the source of the problem. All the VMs recognized as directly connected to the ASA (except VM1).
    Have you ever seen this kind of thing before?
    Thanks again for reading!
    Bob

    FYI: we solved this problem on the VM side.  We removed the network object with VMWare and recreated it.  Once that delete-recreate was complete, I saw the VM1 mac in the firewall.

  • Cisco Nexus 1000v VXLAN don't work

    Hi to all,
    I configured VXLAN configuration by the book (Cisco Nexus 1000V VXLAN Configuration Guide, Release 4.2(1)SV1(5.1)), but there is some problem.
    There are two ESXs with four VMs (two VMs on each ESX). Each VM has one NIC and that NIC is assigned to a port-profile configured for same VXLAN bridge-domain access. There is connectivity between VMs on same ESX but there is no connectivity between VMs hosted on different ESXs. In other words, L2 connectivity works between VMs on same ESX but not between VMs on different ESXs.
    Nexus 1000V VSM is installed on Nexus 1010 Appliance and manages two VEMs through L3 control interfaces.
    VSM version is 4.2(1)SV1(5.1) and VEM feature level is 4.2(1)SV1(5.1).
    Bridge-domain is VXLAN-5001 with segment id 5001 and group address 239.1.1.1
    Port-profile for VMK VXLAN interface is properly configured for access to VLAN 588 ("transport" VLAN for VXLAN) and capability vxlan.
    VLAN 588 is allowed on all uplinks on both sides (Nexus and physical switch).
    Port profile for VMs if properly configured for access to bridge-domain.
    I was create a monitor session for VLAN 588 on upstrean switch (Cisco 6513 with 12.2(18)SXF14 IOS) and  did't see any multicast, unicast or any other traffic. According to documentation, first I shuld to see IGMP join, after that multicast and after that unicast traffic between two VMK interfaces.
    Here is MAC address table for bridge-domain VXLAN-5001:
    Nexus1000V-VSM-1# sh mac address-table bridge-domain VXLAN-5001
    Bridge-domain: VXLAN-5001
              MAC Address       Type    Age       Port            IP Address     Mod
    --------------------------+-------+---------+---------------+---------------+---
              0050.56a3.0009    static  0         Veth6           0.0.0.0         3 
              0050.56a3.000a    static  0         Veth7           0.0.0.0         3 
              0050.56a3.0007    static  0         Veth4           0.0.0.0         4 
              0050.56a3.0008    static  0         Veth5           0.0.0.0         4 
    Total MAC Addresses: 4
    As you can see, there is no proper destination IP addresses.
    Can somebody help me?

    Good hint, but it seems that is not the problem...
    Cat ports connecting VEMs support jumbo frames and their MTU is set to 9216B.
    I saw that MTU on Ethernet interfaces of VEMs is set to 1500B, I changed uplink port-profile and set MTU to first to 1550B, and after that to 9000B (max), but thing still isn't working.
    I'm not using vCloud director, just VMware vSphere 4.1 (vCenter Server with VUM, vCenter Client and two ESX hosts).
    Message was edited by: Mate Grbavac
    After little research I found something strange... I setted up SVI on Cat in Vlan 588 ("transport" VLAN for VXLAN) and when I ping VMKernel interface (with capabilitiy vxlan) with packet size more than 1500B and df bit set I have no reply. My Cat ports and UpLink port profiles are configured for jumbo frames. Is it possible to change MTU of VMKernel interface?

  • Cisco Nexus 1000v stops inheriting

    Guys,
    I have an issue with the Nexus 1000v, basically the trunk ports on the ESXi hosts stop inheriting from the main DATA-UP link port profile, which means that not all VLANS get presented down that given trunk port, its like it gets completey out of sync somehow. An example is below,
    THIS IS A PC CONFIG THAT'S NOT WOKRING CORRECTLY
    show int trunk
    Po9        100,400-401,405-406,412,430,434,438-439,446,449-450,591,850
    sh run int po9
    interface port-channel9
      inherit port-profile DATA-UP
      switchport trunk allowed vlan add 438-439,446,449-450,591,850 (the system as added this not user)
    THIS IS A PC CONFIG THAT IS WORKING CORRECTLY
    show int trunk
    Po2        100,292,300,313,400-401,405-406,412,429-430,434,438-439,446,449-450,582,591,850
    sh run int po2
    interface port-channel2
        inherit port-profile DATA-UP
    I have no idea why this keeps happening, when i remove the manual static trunk configuration on po9, everything is fine, few days later, it happens again, its not just po9, there is at least 3 port-channel that it affects.
    My DATA-UP link port-profile configuration looks like this and all port channels should reflect the VLANs allowed but some are way out.
    port-profile type ethernet DATA-UP
      vmware port-group
      switchport mode trunk
      switchport trunk allowed vlan 100,292,300,313,400-401,405-406,412,429-430,434,438-439,446,449-450,5
    82,591,850
      channel-group auto mode on sub-group cdp
      no shutdown
      state enabled
    The upstream switches match the same VLANs allowed and the VLAN database is a mirror image between Nexus and Upstream switches.
    The Cisco Nexus version is 4.2.1
    Anyone seen this problem?
    Cheers

    Using vMotion you can perform the entire upgrade with no disruption to your virtual infrastructure. 
    If this is your first upgrade, I highly recommend you go through the upgrade guides in detail.
    There are two main guides.  One details the VSM and overall process, the other covers the VEM (ESX) side of the upgrade.  They're not very long guides, and should be easy to follow.
    1000v Upgrade Guide:
    http://www.cisco.com/en/US/docs/switches/datacenter/nexus1000/sw/4_2_1_s_v_1_4_a/upgrade/software/guide/n1000v_upgrade_software.html
    VEM Upgrade Guides:
    http://www.cisco.com/en/US/docs/switches/datacenter/nexus1000/sw/4_2_1_s_v_1_4_a/install/vem/guide/n1000v_vem_install.html
    In a nutshell the procedure looks like this:
    -Backup of VSM Config
    -Run pre-upgrade check script (which will identify any config issues & ensures validation of new version with old config)
    -Upgrade standby VSM
    -Perform switchover
    -Upgrade image on old active (current standby)
    -Upgrade VEM modules
    One decision you'll need to make is whether to use Update Manager or not for the VEM upgrades.  If you don't have many hosts, the manual method is a nice way to maintain control on exactly what's being upgrade & when.  It will allow you to migrate VMs off the host, upgrade it, and then continue in this manner for all remaining hosts.  The alternate is Update Manager, which can be a little sticky if it runs into issues.  This method will automatically put hosts in Maintenance Mode, migrate VMs off, and then upgrade each VEM one by one.  This is a non-stop process so there's a little less control from that perspective.   My own preference is any environment with 10 or less hosts, I use manual, for more than that let VUM do the work.
    Let me know if you have any other questions.
    Regards,
    Robert

  • Server loses internet connection after creating virtual switch in hyper-v

    On a fresh Server 2012R2 Essentials install, I set up Hyper-v and created an external virtual switch. Connection to the vm is good but Health Report no shows loss of internet connectivity and that the router is incorrectly setup. 
    However, I still have a working internet connection? I worry that the server may not update or allow Anywhere Access at some time. Any suggestions? Thanks

    If Hyper-V virtual switch behaves strange and server is destined for testing, then it's correct and I recomend delete this virtual switch and create it again.
    Here is some issues with internet connection and Hyper-V switch, none of them does not match your problem, but can be inspirative for troubleshooting:
    Windows 8 Hyper-V - how to give VM internet access?
    Hyper-V kills internet connection when bridging
    Configuring Hyper-V for multiple subnets with only one NIC (Server 2008 R2 Edition)
    How does basic networking work in Hyper-V?
    Best Regards,
    thennet

  • Cisco Nexus 5K + Micrososft Radius for Admin Authentication

    Hi,
    I have cisco 3750 switches configured to use MS radius for administrator authention. however, now I would like to add our cisco nexus switches to MS radius as well so that administrators are authenticated against the Microsoft radius for admin authention.
    I tried it earlier but it won't accept 3750 commands.. Can you please help with me with a configuration example please that I can follow?
    the commands I have used on 3750 are as follows:
    aaa new-model
    aaa authentication login vtylogin group radius local
    aaa authentication login conlogin group radius local
    aaa authentication enable default group radius enable
    aaa authorization console
    aaa authorization exec vtylogin group radius local
    aaa authorization exec conlogin group radius local
    radius-server host x.x.x.x key SECRETE
    line con 0
    exec-timeout 5 0
    authorization exec conlogin
    logging synchronous
    login authentication conlogin
    line vty 0 4
    exec-timeout 0 0
    authorization exec vtylogin
    login authentication vtylogin
    transport input ssh
    line vty 5 15
    exec-timeout 0 0
    authorization exec vtylogin
    login authentication vtylogin
    transport input ssh

    I have never done this before with ACS but not with NPS. However, you are in the right path. Nexus uses NX-OS which is different in some regards to regular IOS. One of those differences is the AAA setup. In NX-OS you assign users to roles. So for full access you will need to return the following attributes from your Radius server:
    Attribute: cisco-av-pair
    Requirement: Mandatory
    Value: shell:roles*"network-admin vdc-admin"
    For more information take a look at this link:
    http://www.cisco.com/c/en/us/support/docs/security/secure-access-control-system/115925-nexus-integration-acs-00.html
    Hope this helps
    Thank you for rating helpful posts!

  • Install Cisco Nexus 1000V without vCenter?

    Hi guys,
    Is this possible that we install nexus 1000v without using vcenter?
    I have vmware enterprise plus license, but I don't need to install vcenter.
    Regards,

    No, one of the features of the Enterprise Plus licenses provides are VMware & 3rd Party Distributed Virtual Switch functionality.  This requires vCenter to manage them.
    Regards,
    Robert

  • Poor network when create virtual switch in hyper-v 2012

    Hello,
    i installed on DELL R620 windows server 2012 STD and add the hyper-v role
    when i created a virtual switch , the speed of network in the HOST( the server itself) is ver poor ( copy file from fileserver is 300Kbs)
    when i remove the virtual switch , the speed of the network is better(  copy file from fileserver is 10-20Mbps)
    i spoke with DELL support center , and after check up , no problem has found.
    i tried all the chnges with the offload checksum etc but nothing helped
    Please your comment about this issue
    Thanks
    Dotan

    Need some additional information
    Which Team type are you utilizing? 
    Which Raid Controller are you using on the hosts?
    I would suggest you try this as well while its specific to 10Gb CNA's it can help with 1Gb in some cases.  SMB
    FILE COPY PERFORMANCE
    Also utilizing the low latency Bios settings can often offer additional NIC performance 
    Dell 12th Gen Low Latency Bios Settings
    -Sean

  • Cisco Nexus 9300 Virtual Port Channel Support

    Hi,
    As I am new to Nexus 9300 and I was wondering if the switch may support virtual port channel (VPC)? I was wondering if there may be any feature matrix to compare it against the 9500/N7Ks/N5Ks
    Any suggestion is appreciated.
    Thanks.

    Hi Marvin,
    Appreciate for the information. I believe the customer is in the process of moving to ACI gradually however the setup I will be making will be classical ethernet setup which is similar to VSS setup whereby the N9300 would make use of VPC to connect to catalyst 3K/6K switch. I have illustrated it below for reference:
          Catalyst 6500 Non-VSS (Core)
          |----(VPC)---|
    N9300 ----- N9300   (Aggregation/Distribution)
          |----(VPC)---|
     Catalyst 3K switch (Access)
    Appreciate for clarrifiying the support for VPC for Nexus 9300 as I made a search on feature navigator for "Virtual Port Channel" which only listed the N7K platforms.
    Thanks.

  • Nexus 1000V - port-profile for VSM and VEM on same host

    Hello,
    I'm trying to run the VSM and VEM on the same host. At first I connected the VSM to the vSwitch. This works fine. Then I tried to move the VSM behind its own VEM.
    I have 3 different VLANs for management, control and packet traffic. Is it possible to connect the three VSM interfaces (management, control and packet) to only one port-profile, in which these three different VLANs are configured in a trunk? Or do I have to configure for each interface a separate port-profile, if I am using three different VLANs?
    Thanks
    Tobias

    Hello,
        The VSM does not emit VLAN-tagged frames. Therefore, its ports need to be configured in access mode. If your Control and Packet VLANs are the same, you could use the same port profile for both.
    Please make sure that the access VLAN is a system VLAN for all three ports.
    On a related note, if the VSM's storage is on NFS or iSCSI, the storage vmknic needs a system port profile as well. More info at:
    http://www.cisco.com/en/US/docs/switches/datacenter/nexus1000/sw/4_0_4_s_v_1_3/getting_started/configuration/guide/n1000v_gsg_5vsm_behind_vem.html

Maybe you are looking for