Cisco Nexus 6004 | Fibre Tx and Rx Levels

Hi All,
On the Nexus 6004 series, does anybody know how to see the light levels for the optics.
I am using the command: show interface ethernet 2/5/4 transceiver details which outputs:
Ethernet2/5/4
    transceiver is present
    type is QSFP-40G-CSR4
    name is CISCO
    part number is AFBR-79EEPZ-CS2
    revision is 02
    serial number is XXXXXXXX
    nominal bitrate is 10300 MBit/sec
    Link length supported for 50/125um OM3 fiber is 300 m
    Link length supported for 50/125um OM2 fiber is 82 m
    cisco id is --
    cisco extended id number is 16
           SFP Detail Diagnostics Information (internal calibration)
                Current              Alarms                  Warnings
                Measurement     High        Low         High          Low
  Temperature   47.63 C        75.00 C     -5.00 C     70.00 C        0.00 C
  Voltage        3.32 V         3.63 V      2.97 V      3.46 V        3.09 V
  Current        0.00 mA        0.00 mA     0.00 mA     0.00 mA       0.00 mA
  Tx Power        N/A            N/A         N/A         N/A           N/A
  Rx Power        N/A            N/A         N/A         N/A           N/A
  Note: ++  high-alarm; +  high-warning; --  low-alarm; -  low-warning
Is it a case that the SFP's in this instance do not support the light level information? the controller, optics or wdm commands are not available on the Nexus.
Thanks all for your support.

Hi Marvin,
The System version: 6.0(2)N1(2)
Software
  BIOS:      version 2.6.0
  loader:    version N/A
  kickstart: version 6.0(2)N2(2)
  system:    version 6.0(2)N2(2)
  Power Sequencer Firmware:
             Module 0: version v5.0
             Module 1: version v2.0
             Module 2: version v2.0
             Module 3: version v2.0
             Module 4: version v2.0
  Fabric Power Sequencer Firmware: Module 0: version v3.0
  Microcontroller Firmware:        version v1.1.0.3
  QSFP Microcontroller Firmware:
             Module 1: v1.3.0.0
             Module 2: v1.3.0.0
             Module 3: v1.3.0.0
             Module 4: v1.3.0.0

Similar Messages

  • Ask the Expert: Basic Introduction and Troubleshooting on Cisco Nexus 7000 NX-OS Virtual Device Context

    With Vignesh R. P.
    Welcome to the Cisco Support Community Ask the Expert conversation.This is an opportunity to learn and ask questions of Cisco expert Vignesh R. P. about the Cisco® Nexus 7000 Series Switches and support for the Cisco NX-OS Software platform .
    The Cisco® Nexus 7000 Series Switches introduce support for the Cisco NX-OS Software platform, a new class of operating system designed for data centers. Based on the Cisco MDS 9000 SAN-OS platform, Cisco NX-OS introduces support for virtual device contexts (VDCs), which allows the switches to be virtualized at the device level. Each configured VDC presents itself as a unique device to connected users within the framework of that physical switch. The VDC runs as a separate logical entity within the switch, maintaining its own unique set of running software processes, having its own configuration, and being managed by a separate administrator.
    Vignesh R. P. is a customer support engineer in the Cisco High Touch Technical Support center in Bangalore, India, supporting Cisco's major service provider customers in routing and MPLS technologies. His areas of expertise include routing, switching, and MPLS. Previously at Cisco he worked as a network consulting engineer for enterprise customers. He has been in the networking industry for 8 years and holds CCIE certification in the Routing & Switching and Service Provider tracks.
    Remember to use the rating system to let Vignesh know if you have received an adequate response. 
    Vignesh might not be able to answer each question due to the volume expected during this event. Remember that you can continue the conversation on the  Data Center sub-community discussion forum shortly after the event. This event lasts through through January 18, 2013. Visit this forum often to view responses to your questions and the questions of other community members.

    Hi Vignesh
    Is there is any limitation to connect a N2K directly to the N7K?
    if i have a an F2 card 10G and another F2 card 1G and i want to creat 3 VDC'S
    VDC1=DC-Core
    VDC2=Aggregation
    VDC3=Campus core
    do we need to add a link between the different VDC's
    thanks

  • Routing issue between Cisco Nexus and Cisco 4510 R+E Chassis

    We have configured Cisco Nexus 7K9 as core and Cisco 4510 R+E as access switches for Server connectivity.
    We are experiencing problem in terms of ARP learning and Ping issues between Cisco Nexus and end hosts.

    Hi,
    So you have N7k acting as L3 with servers connected to 4510?.
    Do you see the MAC associated with failing ARP in 4510?. Is it happening with all or few servers?. Just to verify if it is connectivity issue between N7k and 4510, you can configure an SVI on 4510 and assign address from same raneg (server/core range) and perform a ping.
    This will help narrow down if issue is between server to 4510 or 4510 to N7k.
    Thanks,
    Nagendra

  • Nexus 6004 EIGRP Relationship between the two switches

    Hi All,
    I will try to explain this as best as I can. In our current TEST LAB we have a Pair of Cisco ASA5585x running in Active/Passive mode. We use a VRF transit to connect the 10 GB interface to a Pair of Cisco Nexus 6004 (L3) switches running vPC between them. Downstream we also have a pair of Cisco 9372 switches (L2) also running vPC between the two.
    As of right now we have EIGRP neighbor relationship formed between the two N6K's and the ASA.
    ASA
    ciscoasa# sh eigrp neighbors
    EIGRP-IPv4 neighbors for process 100
    H Address Interface Hold Uptime SRTT RTO Q Seq
    (sec) (ms) Cnt Num
    1 172.16.230.9 Te0/8.451 12 01:30:25 1 200 0 52
    0 172.16.230.10 Te0/8.451 12 01:30:25 1 200 0 48
    The ASA formed relationship with both N6K's
    SWITCH1
    Nexus6-1# sh ip eigrp neighbors vrf inside
    IP-EIGRP neighbors for process 100 VRF Inside
    H Address Interface Hold Uptime SRTT RTO Q Seq
    (sec) (ms) Cnt Num
    0 172.16.8.3 Vlan680 11 01:28:28 1 50 0 45
    1 172.16.230.10 Vlan451 13 01:28:28 1 50 0 46
    2 172.16.230.11 Vlan451 10 01:28:00 4 50 0 13
    Nexus6-1#
    SWITCH2
    Nexus6-2# sh ip eigrp neighbors vrf Inside
    IP-EIGRP neighbors for process 100 VRF Inside
    H Address Interface Hold Uptime SRTT RTO Q Seq
    (sec) (ms) Cnt Num
    2 172.16.8.2 Vlan680 14 01:30:11 23 138 0 48
    0 172.16.230.9 Vlan451 13 01:30:11 480 2880 0 50
    1 172.16.230.11 Vlan451 13 01:29:48 1598 5000 0 13
    Nexus6-2#
    Both Nexus Switches formed EIGRP neighbors using the vPC Peer-Link. There is enough documentation out there that strongly suggest not to use vPC Peer-Links for EIGRP anything.
    We do have additional interfaces available on the 6K's that we can use as a cross connect for EIGRP. What we are having trouble understanding how we can force EIGRP traffic over those ports?
    Here is a complete Switch config:
    Switch1
    Nexus6-1# sh run
    feature telnet
    cfs eth distribute
    feature eigrp
    feature interface-vlan
    feature lacp
    feature vpc
    feature lldp
    vlan 1
    vlan 451
    name P2P_VRF_SVI
    vlan 652
    name Management
    vlan 680
    name Inside
    vrf context Inside
    vrf context management
    ip route 0.0.0.0/0 172.16.52.1
    vrf context peer-keepalive
    vpc domain 99
    role priority 1
    peer-keepalive destination 10.200.50.2 source 10.200.50.1 vrf peer-keepalive
    delay restore 120
    interface Vlan1
    interface Vlan451
    description Inside p2p to ASA
    no shutdown
    vrf member Inside
    ip address 172.16.230.9/29
    ip router eigrp 100
    no ip passive-interface eigrp 100
    interface Vlan651
    interface Vlan680
    description Inside Network
    no shutdown
    vrf member Inside
    ip address 172.16.8.2/22
    ip router eigrp 100
    interface port-channel99
    switchport mode trunk
    spanning-tree port type network
    vpc peer-link
    interface port-channel102
    switchport mode trunk
    vpc 102
    interface Ethernet1/1
    description vPC Peer Link 1.1
    switchport mode trunk
    speed auto
    channel-group 99
    interface Ethernet1/6
    interface Ethernet1/7
    description vPC Peer Link 1.7 to Nexus 9372 PRI
    switchport mode trunk
    speed auto
    channel-group 102 mode active
    interface Ethernet1/8
    interface Ethernet1/9
    interface Ethernet2/1
    description vPC Peer Link 2.1
    switchport mode trunk
    speed auto
    channel-group 99
    interface Ethernet2/2
    interface Ethernet2/7
    description vPC Peer Link 2.1 to Nexus SEC
    switchport mode trunk
    speed auto
    channel-group 102 mode active
    interface Ethernet2/8
    interface Ethernet8/1
    description keep-alive peer-link to ALNSWI02
    no switchport
    vrf member peer-keepalive
    ip address 10.200.50.1/30
    interface Ethernet8/2
    description Uplink to ASA
    switchport mode trunk
    interface Ethernet8/3
    interface mgmt0
    vrf member management
    ip address 172.16.52.3/23
    line console
    line vty
    boot kickstart bootflash:/n6000-uk9-kickstart.7.0.1.N1.1.bin
    boot system bootflash:/n6000-uk9.7.0.1.N1.1.bin
    router eigrp 100
    passive-interface default
    default-information originate
    vrf Inside
    autonomous-system 100
    default-information originate
    poap transit
    Nexus6-1#
    Nexus6-1# sh ip eigrp neighbors vrf inside
    IP-EIGRP neighbors for process 100 VRF Inside
    H Address Interface Hold Uptime SRTT RTO Q Seq
    (sec) (ms) Cnt Num
    0 172.16.8.3 Vlan680 11 01:28:28 1 50 0 45
    1 172.16.230.10 Vlan451 13 01:28:28 1 50 0 46
    2 172.16.230.11 Vlan451 10 01:28:00 4 50 0 13
    Nexus6-1#
    Nexus6-1# sh ip eigrp topology vrf Inside
    IP-EIGRP Topology Table for AS(100)/ID(172.16.8.2) VRF Inside
    Codes: P - Passive, A - Active, U - Update, Q - Query, R - Reply,
    r - reply Status, s - sia Status
    P 172.16.8.0/22, 1 successors, FD is 2816
    via Connected, Vlan680
    P 172.16.230.8/29, 1 successors, FD is 2816
    via Connected, Vlan451
    Nexus6-1# sh vpc
    Legend:
    (*) - local vPC is down, forwarding via vPC peer-link
    vPC domain id : 99
    Peer status : peer adjacency formed ok
    vPC keep-alive status : peer is alive
    Configuration consistency status : success
    Per-vlan consistency status : success
    Type-2 consistency status : success
    vPC role : primary
    Number of vPCs configured : 1
    Peer Gateway : Disabled
    Dual-active excluded VLANs : -
    Graceful Consistency Check : Enabled
    Auto-recovery status : Disabled
    vPC Peer-link status
    id Port Status Active vlans
    1 Po99 up 1,451,652,680
    vPC status
    id Port Status Consistency Reason Active vlans
    102 Po102 up success success 1,451,652,6
    80
    Nexus6-1# sh spanning-tree
    VLAN0001
    Spanning tree enabled protocol rstp
    Root ID Priority 32769
    Address 1005.caf5.88ff
    Cost 2
    Port 4197 (port-channel102)
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Bridge ID Priority 32769 (priority 32768 sys-id-ext 1)
    Address 8c60.4f2d.2ffc
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Interface Role Sts Cost Prio.Nbr Type
    Po99 Desg FWD 1 128.4194 (vPC peer-link) Network P2p
    Po102 Root FWD 1 128.4197 (vPC) P2p
    Eth8/2 Desg FWD 2 128.1026 P2p
    Eth8/3 Desg FWD 2 128.1027 P2p
    VLAN0451
    Spanning tree enabled protocol rstp
    Root ID Priority 33219
    Address 8c60.4f2d.2ffc
    This bridge is the root
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Bridge ID Priority 33219 (priority 32768 sys-id-ext 451)
    Address 8c60.4f2d.2ffc
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Interface Role Sts Cost Prio.Nbr Type
    Po99 Desg FWD 1 128.4194 (vPC peer-link) Network P2p
    Po102 Desg FWD 1 128.4197 (vPC) P2p
    Eth8/2 Desg FWD 2 128.1026 P2p
    VLAN0652
    Spanning tree enabled protocol rstp
    Root ID Priority 33420
    Address 1005.caf5.88ff
    Cost 2
    Port 4197 (port-channel102)
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Bridge ID Priority 33420 (priority 32768 sys-id-ext 652)
    Address 8c60.4f2d.2ffc
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Interface Role Sts Cost Prio.Nbr Type
    Po99 Desg FWD 1 128.4194 (vPC peer-link) Network P2p
    Po102 Root FWD 1 128.4197 (vPC) P2p
    Eth8/2 Desg FWD 2 128.1026 P2p
    VLAN0680
    Spanning tree enabled protocol rstp
    Root ID Priority 33448
    Address 1005.caf5.88ff
    Cost 2
    Port 4197 (port-channel102)
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Bridge ID Priority 33448 (priority 32768 sys-id-ext 680)
    Address 8c60.4f2d.2ffc
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Interface Role Sts Cost Prio.Nbr Type
    Po99 Desg FWD 1 128.4194 (vPC peer-link) Network P2p
    Po102 Root FWD 1 128.4197 (vPC) P2p
    Eth8/2 Desg FWD 2 128.1026 P2p
    Nexus6-1#
    Switch2
    Nexus6-2# sh run
    !Command: show running-config
    !Time: Sat Feb 12 19:02:44 2011
    version 7.0(1)N1(1)
    hostname Nexus6-2
    feature telnet
    cfs eth distribute
    feature eigrp
    feature interface-vlan
    feature lacp
    feature vpc
    feature lldp
    vlan 1
    vlan 451
    name P2P_VRF_SVI
    vlan 652
    name Management
    vlan 680
    name Inside
    vrf context Inside
    vrf context P2P_Inside_VRF
    vrf context management
    ip route 0.0.0.0/0 172.16.52.1
    vrf context peer-keepalive
    vpc domain 99
    role priority 2
    peer-keepalive destination 10.200.50.1 source 10.200.50.2 vrf peer-keepalive
    delay restore 120
    interface Vlan1
    interface Vlan451
    description Inside p2p to ASA
    no shutdown
    vrf member Inside
    ip address 172.16.230.10/29
    ip router eigrp 100
    no ip passive-interface eigrp 100
    interface Vlan680
    description Inside Network
    no shutdown
    vrf member Inside
    ip address 172.16.8.3/22
    ip router eigrp 100
    interface port-channel99
    switchport mode trunk
    spanning-tree port type network
    vpc peer-link
    interface port-channel102
    switchport mode trunk
    vpc 102
    interface Ethernet1/1
    description vPC Peer Link 1.1
    switchport mode trunk
    speed auto
    channel-group 99
    interface Ethernet1/2
    interface Ethernet1/6
    interface Ethernet1/7
    description vPC Link 1.7 to Nexus 9372 SEC
    switchport mode trunk
    speed auto
    channel-group 102 mode active
    interface Ethernet1/8
    interface Ethernet1/12
    interface Ethernet2/1
    description vPC Peer Link 2.1
    switchport mode trunk
    speed auto
    channel-group 99
    interface Ethernet2/2
    interface Ethernet2/6
    interface Ethernet2/7
    description vPC Link 2.1 to Nexus PRI
    switchport mode trunk
    speed auto
    channel-group 102 mode active
    interface Ethernet2/8
    interface Ethernet2/12
    interface Ethernet8/1
    description keep-alive peer-link to ALNSWI01
    no switchport
    vrf member peer-keepalive
    ip address 10.200.50.2/30
    interface Ethernet8/2
    description Uplink to ASA
    switchport mode trunk
    switchport trunk allowed vlan 1,451,652,680
    interface Ethernet8/3
    interface Ethernet8/20
    interface mgmt0
    vrf member management
    ip address 172.16.52.4/23
    line console
    line vty
    boot kickstart bootflash:/n6000-uk9-kickstart.7.0.1.N1.1.bin
    boot system bootflash:/n6000-uk9.7.0.1.N1.1.bin
    router eigrp 100
    vrf Inside
    autonomous-system 100
    default-information originate
    poap transit
    logging logfile messages 6
    Nexus6-2#
    Nexus6-2#
    Nexus6-2# sh ip eigrp neighbors vrf Inside
    IP-EIGRP neighbors for process 100 VRF Inside
    H Address Interface Hold Uptime SRTT RTO Q Seq
    (sec) (ms) Cnt Num
    2 172.16.8.2 Vlan680 14 01:30:11 23 138 0 48
    0 172.16.230.9 Vlan451 13 01:30:11 480 2880 0 50
    1 172.16.230.11 Vlan451 13 01:29:48 1598 5000 0 13
    Nexus6-2#
    Nexus6-2# sh ip eigrp topology vrf Inside
    IP-EIGRP Topology Table for AS(100)/ID(172.16.8.3) VRF Inside
    Codes: P - Passive, A - Active, U - Update, Q - Query, R - Reply,
    r - reply Status, s - sia Status
    P 172.16.8.0/22, 1 successors, FD is 2816
    via Connected, Vlan680
    P 172.16.230.8/29, 1 successors, FD is 2816
    via Connected, Vlan451
    Nexus6-2#
    Nexus6-2#
    Nexus6-2# sh vpc
    Legend:
    (*) - local vPC is down, forwarding via vPC peer-link
    vPC domain id : 99
    Peer status : peer adjacency formed ok
    vPC keep-alive status : peer is alive
    Configuration consistency status : success
    Per-vlan consistency status : success
    Type-2 consistency status : success
    vPC role : secondary
    Number of vPCs configured : 1
    Peer Gateway : Disabled
    Dual-active excluded VLANs : -
    Graceful Consistency Check : Enabled
    Auto-recovery status : Disabled
    vPC Peer-link status
    id Port Status Active vlans
    1 Po99 up 1,451,652,680
    vPC status
    id Port Status Consistency Reason Active vlans
    102 Po102 up success success 1,451,652,6
    80
    Nexus6-2#
    Nexus6-2#
    Nexus6-2# sh spanning-tree
    VLAN0001
    Spanning tree enabled protocol rstp
    Root ID Priority 32769
    Address 1005.caf5.88ff
    Cost 3
    Port 4194 (port-channel99)
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Bridge ID Priority 32769 (priority 32768 sys-id-ext 1)
    Address 8c60.4f2d.777c
    Hello Time 2 sec Max Age 20 sec Forward Delay 15 sec
    Interface Role Sts Cost Prio.Nbr Type
    Po99 Root FWD 1 128.4194 (vPC peer-link) Network P2p
    Po102 Root FWD 1 128.4197 (vPC) P2p
    Eth8/2 Desg FWD 2 128.1026 P2p
    Eth8/3 Desg FWD 2 128.1027 P2p
    VLAN0451
    Spanning tree enabled protocol rstp
    Root ID Priority 33219
    Address 8c

    Jon,
    Are you ready for the mass confusion?
    when Looking at the ASA EIGRP neighbors output here is what I see.
    ASA# sh eigrp neighbors
    EIGRP-IPv4 neighbors for process 100
    H   Address                 Interface       Hold Uptime   SRTT   RTO  Q  Seq
                                                (sec)         (ms)       Cnt Num
    3   172.16.230.1            Te0/8.450        13  16:45:14 1    200   0   64
    2   172.16.230.2            Te0/8.450        11  16:45:14 1    200   0   84
    1   172.16.230.10           Te0/8.451        11  16:45:20 1    200   0   178
    0   172.16.230.9            Te0/8.451        13  16:45:20 1    200   0   148
    For simplicity sake lets just concetrate on Interface TenGigabit0/8.451 which is the SVI on the Nexus switch that is VLAN451
    From the Nexus Switch 6004 that is directly connected to the ASA here is what I see
    SWI01# sh ip eigrp neighbors vrf Inside
    IP-EIGRP neighbors for process 100 VRF Inside
    H   Address                 Interface       Hold  Uptime  SRTT   RTO  Q  Seq
                                                (sec)         (ms)       Cnt Num
    0   172.16.8.3              Vlan680         10   17:04:30  54   324   0   177
    1   172.16.230.10           Vlan451         11   16:59:10  819  4914  0   178
    2   172.16.230.11           Vlan451         14   16:53:48  24   144   0   20
    The Inside VRF that is tied to both SVI's on the Switch vlans 451 and 680 is in EIGRP 100 on the switch
    SWI01# sh run int vlan 451
    interface Vlan451
      description Inside p2p to ASA
      no shutdown
      vrf member Inside
      ip address 172.16.230.9/29
      ip router eigrp 100
      no ip passive-interface eigrp 100
    SWI01# sh run int vlan 680
    interface Vlan680
      description Inside Network
      no shutdown
      vrf member Inside
      ip address 172.16.8.2/22
      ip router eigrp 100
      hsrp 1
        authentication text test
        preempt
        priority 250
        ip 172.16.8.1
    so you with me so far?
    If you are you have noticed that on the ASA neighbors the ASA sees 172.16.230.11 as a neighbor which is the Secondary Nexus SW. That is becauise they all share the same subnet.
    172.16.230.8/29
    Brakedown:
    PRI Nexus 6004 - 172.16.230.9
    SEC NEXUS 6004 - 172.16.230.10
    PRI ASA 5585x  - 172.16.230.11
    SEC ASA 5585x  - 172.16.230.12
    Because the ASA EIGRP network is a /29 it learns the Secondary Nexus via the Primary Nexus.
    I am not sure that the link we created between the two Nexus Switches is doing anything but consuming ports right now.
    SWI01# sh run int ethernet 8/9
    interface Ethernet8/9
      description EIGRP PORT to Secondary Nexus
      switchport mode trunk
      switchport trunk allowed vlan 450-451
    SWI02# sh run int ethernet 8/9
    interface Ethernet8/9
      description EIGRP PORT to Primary Nexus
      switchport mode trunk
      switchport trunk allowed vlan 450-451
    So the SVI's that go up to the ASA for inspection are 450 and 451. The network SVI's are 600 and 680 all of them live on the switch, and 680, and 600 are extended over the peer links down to the 9372's.
    I think that we are breaking the golden rule of vPC BUT.. I am not 100% sure. Some of the documents read that we should not be allowing network vlans over peer links, but then how do you extend the vlans down to the leaf switch?
    This is giving me nightmares at the moment…
    does this make sense? 

  • FCoE with Cisco Nexus 5548 switches and VMware ESXi 4.1

    Can someone share with me what needs to be setup on the Cisco Nexus side to work with VMware in the following scenario?
    Two servers with two cards dual port FCoE cards with two ports connected to two Nexus 5548 switches that are clusterd together.  We want to team the ports together on the VMware side using IP Hash so what should be done on the cisco side for this to work? 
    Thanks...

    Andres,
    The Cisco Road Map for the 5010 and 5020 doesn't include extending the current total (12) FEX capabities.  The 5548 and 5596 will support more (16) per 55xxk, and with the 7K will support upto 32 FEX's.
    Documentation has been spotty on this subject, because the term 5k indicates that all 5000 series switches will support extended FEX's which is not the case only the 55xx will support more than 12 FEX.  Maybe in the future the terminology for the 5k series should be term 5000 series and 5500 series Nexus, there are several differences and advancements between the two series.

  • Ask the Expert: Cisco Nexus 2000, 5000, and 6000 Series Switches

    with Cisco Expert Vinayak Sudame
    Welcome to the Cisco Support Community Ask the Expert conversation. This is an opportunity to learn and ask questions how to configure and troubleshoot the Cisco Nexus 2000, 5000 and 6000 Series Switches with Cisco subject matter expert Vinayak Sudame. You can ask any question on configuration, troubleshooting, features, design and Fiber Channel over Ethernet (FCoE).
    Vinayak Sudame is a Technical Lead in Data Center Switching Support Team within Cisco's Technical Services in RTP, North Carolina. His current responsibilities include but are not limited to Troubleshooting Technical support problems and Escalations in the areas of Nexus 5000, Nexus 2000, FCoE. Vinayak is also involved in developing technical content for Cisco Internal as well as external. eg, Nexus 5000 Troubleshooting Guide (CCO), Nexus 5000 portal (partners), etc. This involves cross team collaboration and working with multiple different teams within Cisco. Vinayak has also contributed to training account teams and partners in CAE (Customer Assurance Engineering) bootcamp dealing with Nexus 5000 technologies. In the past, Vinayak's responsibilities included supporting MDS platform (Fiber Channel Technologies) and work with EMC support on Escalated MDS cases. Vinayak was the Subject Matter Expert for Santap Technologies before moving to Nexus 5000 support. Vinayak holds a Masters in Electrical Engineering with Specialization in Networking from Wichita State University, Kansas. He also holds Cisco Certification CCIE (#20672) in Routing and Switching.
    Remember to use the rating system to let Vinayak know if you have received an adequate response.
    Vinayak might not be able to answer each question due to the volume expected during this event. Remember that you can continue the conversation on the  Data Center sub-community, Other Data Center Topics discussion forum shortly after the event.
    This event last through Friday July 12, 2013. Visit the community often to view responses to youe questions of other community members.

    Hi Vinayak,
    Output of "show cfs internal ethernet-peer database"
    Switch 1
    ETH Fabric
    Switch WWN              logical-if_index
    20:00:54:7f:ee:b7:c2:80 [Local]
    20:00:54:7f:ee:b6:3f:80 16000005
    Total number of entries = 2
    Switch 2
    ETH Fabric
    Switch WWN              logical-if_index
    20:00:54:7f:ee:b6:3f:80 [Local]
    20:00:54:7f:ee:b7:c2:80 16000005
    Total number of entries = 2
    Output of "show system internal csm info trace"
    Switch 1 in which "show cfs peers" show proper output
    Mon Jul  1 05:46:19.145339  (CSM_T) csm_sp_buf_cmd_tbl_expand_range(8604): No range command in buf_cmd_tbl.
    Mon Jul  1 05:46:19.145280  (CSM_T) csm_set_sync_status(6257): Peer RT status PSSed
    Mon Jul  1 05:46:19.145188  (CSM_T) csm_sp_handle_local_verify_commit(4291):
    Mon Jul  1 05:46:19.145131  csm_continue_verify_ac[597]: peer is not reachable over CFS so continuing with local verify/commit
    Mon Jul  1 05:46:19.145071  csm_tl_lock(766): Peer information not found for IP address: '172.16.1.54'
    Mon Jul  1 05:46:19.145011  csm_tl_lock(737):
    Mon Jul  1 05:46:19.144955  (CSM_EV) csm_sp_build_tl_lock_req_n_send(941): sending lock-request for CONF_SYNC_TL_SESSION_TYPE_VERIFY subtype 0 to Peer ip = (172.16.1.54)
    Mon Jul  1 05:46:19.143819  (CSM_T) csm_copy_image_and_internal_versions(788): sw_img_ver: 5.2(1)N1(2a), int_rev: 1
    Mon Jul  1 05:46:19.143761  (CSM_T) csm_sp_get_peer_sync_rev(329): found the peer with address=172.16.1.54 and sync_rev=78
    Mon Jul  1 05:46:19.143699  (CSM_T) csm_sp_get_peer_sync_rev(315):
    Mon Jul  1 05:46:19.143641  (CSM_EV) csm_sp_build_tl_lock_req_n_send(838): Entered fn
    Mon Jul  1 05:46:19.143582  (CSM_T) csm_set_sync_status(6257): Peer RT status PSSed
    Switch 2 in which "show cfs peers" does not show proper output
    Mon Jul  1 06:13:11.885354  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport mode trunk, cmd pseq 77 seq 482
    Mon Jul  1 06:13:11.884992  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd channel-group 51 mode active, cmd pseq 357 seq 369
    Mon Jul  1 06:13:11.884932  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport trunk allowed vlan 2, 11, cmd pseq 357 seq 368
    Mon Jul  1 06:13:11.884872  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport mode trunk, cmd pseq 357 seq 367
    Mon Jul  1 06:13:11.884811  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd description process_vpc, cmd pseq 357 seq 366
    Mon Jul  1 06:13:11.884750  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd channel-group 51 mode active, cmd pseq 352 seq 365
    Mon Jul  1 06:13:11.884690  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport trunk allowed vlan 2, 11, cmd pseq 352 seq 364
    Mon Jul  1 06:13:11.884630  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport mode trunk, cmd pseq 352 seq 363
    Mon Jul  1 06:13:11.884568  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd description process_vpc, cmd pseq 352 seq 362
    Mon Jul  1 06:13:11.884207  (CSM_EV) csm_sp_acfg_gen_handler(3011):  Preparing config into /tmp/csm_sp_acfg_1733916569.txt
    Mon Jul  1 06:13:11.878695  csm_get_locked_ssn_ctxt[539]: Lock not yet taken.
    Mon Jul  1 06:13:11.878638  (CSM_EV) csm_sp_acfg_gen_handler(2937): Recieved sp acfg merge request for type: running cfg
    Mon Jul  1 06:12:29.527840  (CSM_T) csm_pss_del_seq_tbl(1989): Freeing seq tbl data
    Mon Jul  1 06:12:29.513255  (CSM_T) csm_sp_acfg_gen_handler(3106): Done acfg file write
    Mon Jul  1 06:12:29.513179  (CSM_EV) csm_sp_acfg_gen_handler(3011):  Preparing config into /tmp/csm_sp_acfg_1733911262.txt
    Mon Jul  1 06:12:29.508859  csm_get_locked_ssn_ctxt[539]: Lock not yet taken.
    Mon Jul  1 06:12:29.508803  (CSM_EV) csm_sp_acfg_gen_handler(2937): Recieved sp acfg merge request for type: running cfg
    Mon Jul  1 05:53:17.651236  Collecting peer info
    Mon Jul  1 05:53:17.651181  Failed to get the argumentvalue for 'ip-address'
    Mon Jul  1 05:40:59.262736  DB Unlocked Successfully
    Mon Jul  1 05:40:59.262654  Unlocking DB, Lock Owner Details:Client:1 ID:1
    Mon Jul  1 05:40:59.262570  (CSM_T) csm_sp_del_buf_cmd(1713): Deleting comand with Id = 1
    Mon Jul  1 05:40:59.262513  DB Lock Successful by Client:1 ID:1
    Mon Jul  1 05:40:59.262435  Recieved lock request by Client:1 ID:1
    Mon Jul  1 05:40:41.741224  ssnmgr_ssn_handle_create_get: Session FSM already present, ID:1
    Mon Jul  1 05:40:41.741167  ssnmgr_handle_mgmt_request: Create/Get request received for session[process_n5kprof]
    show cfs lock gives no output.
    Just to further clarify, we have 4 5548UP switches in the same management vlan. 2 switches are in one location lets say location A and they are CFS peers and are working fine.
    These two switches which are having problem are in location B. All the switches are in the same vlan. Essentially the all CFS multicast messages will be seen by all 5548 switches as they are in the same vlan. I am assuming that this might not create any problems as we specify the peers in the respective configurations. Or do we have to change the CFSoIPv4 multicast addresses in location B or may be configure a different region.
    Regards.

  • Ask the Expert: Configuration, Design, and Troubleshooting of Cisco Nexus 1000

    With Louis Watta
    Welcome to the Cisco Support Community Ask the Expert conversation. This is an opportunity to learn and ask questions about design, configuration, and troubleshooting of Cisco Nexus 1000V Series Switches operating inside VMware ESXi and Hyper-V with Cisco expert Louis Watta. Cisco Nexus 1000V Series Switches deliver highly secure, multitenant services by adding virtualization intelligence to the data center network. With Cisco Nexus 1000V Series Switches, you can have a consistent networking feature set and provisioning process all the way from the virtual machine access layer to the core of the data center network infrastructure.
    This is a continuation of the live Webcast.
    Louis Watta is a technical leader in the services organization for Cisco. Watta's primary background is in data center technologies: servers (UNIX, Windows, Linux), switches (MDS, Brocade), storage arrays (EMC, NetApp, HP), network switches (Cisco Catalyst and Cisco Nexus), and enterprise service hypervisors (VMware ESX, Hyper-V, KVM, XEN). As a Technical Leader in Technical Services, Louis currently supports beta and early field trials (EFTs) on new Cisco software and hardware. He has more than 15 years of experience in a wide variety of data center applications and is interested in data center technologies oriented toward data center virtualization and orchestration. Prior to Cisco, Louis was a system administrator for GTE Government Systems. He has a bachelor of science degree in computer science from North Carolina State University. .
    Remember to use the rating system to let Louis know if you have received an adequate response.
    Louis might not be able to answer each question because of the volume expected during this event. Remember that you can continue the conversation on the Data Center community Unified Computing shortly after the event.
    This event lasts through Friday, JUne 14, 2013. Visit this forum often to view responses to your questions and the questions of other Cisco Support Community members.
    Webcast related links:
    Slides
    FAQ
    Webcast Video Recording

    Right now there is only a few features that are not supported on N1Kv on Hyper-V
    They are VXLAN and QOS Fair Weighted Queuing. We are currently demoing VXLAN functionality at Microsoft TechEd Conference this week in New Orleans. So VXLAN support should be coming soon. I can't give you a specific timeline.
    For Fair Weighted Queuing I'm not sure. In the VMware world we take advantage of NETIOC infrastructure. In the MS world they do not have a NETIOC infrastructure that we can use to create a similar feature.
    Code base parity (as in VMware and Hyper-V VSMs running NXOS 5.x) will happen with the next major N1KV release for ESX.
    Let me know if that doesn't answer your question.
    thanks
    louis

  • Ask the Expert: Different Flavors and Design with vPC on Cisco Nexus 5000 Series Switches

    Welcome to the Cisco® Support Community Ask the Expert conversation.  This is an opportunity to learn and ask questions about Cisco® NX-OS.
    The biggest limitation to a classic port channel communication is that the port channel operates only between two devices. To overcome this limitation, Cisco NX-OS has a technology called virtual port channel (vPC). A pair of switches acting as a vPC peer endpoint looks like a single logical entity to port channel attached devices. The two devices that act as the logical port channel endpoint are actually two separate devices. This setup has the benefits of hardware redundancy combined with the benefits offered by a port channel, for example, loop management.
    vPC technology is the main factor for success of Cisco Nexus® data center switches such as the Cisco Nexus 5000 Series, Nexus 7000 Series, and Nexus 2000 Series Switches.
    This event is focused on discussing all possible types of vPC along-with best practices, failure scenarios, Cisco Technical Assistance Center (TAC) recommendations and troubleshooting
    Vishal Mehta is a customer support engineer for the Cisco Data Center Server Virtualization Technical Assistance Center (TAC) team based in San Jose, California. He has been working in TAC for the past 3 years with a primary focus on data center technologies, such as the Cisco Nexus 5000 Series Switches, Cisco Unified Computing System™ (Cisco UCS®), Cisco Nexus 1000V Switch, and virtualization. He presented at Cisco Live in Orlando 2013 and will present at Cisco Live Milan 2014 (BRKCOM-3003, BRKDCT-3444, and LABDCT-2333). He holds a master’s degree from Rutgers University in electrical and computer engineering and has CCIE® certification (number 37139) in routing and switching, and service provider.
    Nimit Pathak is a customer support engineer for the Cisco Data Center Server Virtualization TAC team based in San Jose, California, with primary focus on data center technologies, such as Cisco UCS, the Cisco Nexus 1000v Switch, and virtualization. Nimit holds a master's degree in electrical engineering from Bridgeport University, has CCNA® and CCNP® Nimit is also working on a Cisco data center CCIE® certification While also pursuing an MBA degree from Santa Clara University.
    Remember to use the rating system to let Vishal and Nimit know if you have received an adequate response. 
    Because of the volume expected during this event, Vishal and Nimit might not be able to answer every question. Remember that you can continue the conversation in the Network Infrastructure Community, under the subcommunity LAN, Switching & Routing, shortly after the event. This event lasts through August 29, 2014. Visit this forum often to view responses to your questions and the questions of other Cisco Support Community members.

    Hello Gustavo
    Please see my responses to your questions:
    Yes almost all routing protocols use Multicast to establish adjacencies. We are dealing with two different type of traffic –Control Plane and Data Plane.
    Control Plane: To establish Routing adjacency, the first packet (hello) is punted to CPU. So in the case of triangle routed VPC topology as specified on the Operations Guide Link, multicast for routing adjacencies will work. The hellos packets will be exchanged across all 3 routers and adjacency will be formed over VPC links
    http://www.cisco.com/c/en/us/td/docs/switches/datacenter/nexus5000/sw/operations/n5k_L3_w_vpc_5500platform.html#wp999181
    Now for Data Plane we have two types of traffic – Unicast and Multicast.
    The Unicast traffic will not have any forwarding issues, but because the Layer 3 ECMP and port channel run independent hash calculations there is a possibility that when the Layer 3 ECMP chooses N5k-1 as the Layer 3 next hop for a destination address while the port channel hashing chooses the physical link toward N5k-2. In this scenario,N5k-2 receives packets from R with the N5k-1 MAC as the destination MAC.
    Sending traffic over the peer-link to the correct gateway is acceptable for data forwarding, but it is suboptimal because it makes traffic cross the peer link when the traffic could be routed directly.
    For that topology, Multicast Traffic might have complete traffic loss due to the fact that when a PIM router is connected to Cisco Nexus 5500 Platform switches in a vPC topology, the PIM join messages are received only by one switch. The multicast data might be received by the other switch.
    The Loop avoidance works little different across Nexus 5000 and Nexus 7000.
    Similarity: For both products, loop avoidance is possible due to VSL bit
    The VSL bit is set in the DBUS header internal to the Nexus.
    It is not something that is set in the ethernet packet that can be identified. The VSL bit is set on the port asic for the port used for the vPC peer link, so if you have Nexus A and Nexus B configured for vPC and a packet leaves Nexus A towards Nexus B, Nexus B will set the VSL bit on the ingress port ASIC. This is not something that would traverse the peer link.
    This mechanism is used for loop prevention within the chassis.
    The idea being that if the port came in the peer link from the vPC peer, the system makes the assumption that the vPC peer would have forwarded this packet out the vPC-enabled port-channels towards the end device, so the egress vpc interface's port-asic will filter the packet on egress.
    Differences:  In Nexus 5000 when it has to do L3-to-L2 lookup for forwarding traffic, the VSL bit is cleared and so the traffic is not dropped as compared to Nexus 7000 and Nexus 3000.
    It still does loop prevention but the L3-to-L2 lookup is different in Nexus 5000 and Nexus 7000.
    For more details please see below presentation:
    https://supportforums.cisco.com/sites/default/files/session_14-_nexus.pdf
    DCI Scenario:  If 2 pairs are of Nexus 5000 then separation of L3/L2 links is not needed.
    But in most scenarios I have seen pair of Nexus 5000 with pair of Nexus 7000 over DCI or 2 pairs of Nexus 7000 over DCI. If Nexus 7000 are used then L3 and L2 links are required for sure as mentioned on above presentation link.
    Let us know if you have further questions.
    Thanks,
    Vishal

  • Is Cisco Nexus 5596UP support vlan base Policing and traffic shaping on code NX OS version: 5.1(3)N1(1)

    Is Cisco Nexus 5596UP support vlan base Policing and traffic shaping on code NX OS version: 5.1(3)N1(1)
    where i couldn't see any police command under the policy map 

    I have tested this issue on another 5548UP with L3 running the same NX-OS version and get the same problem. Show CDP from the switch is not discovering devices, but the neightbors can see the 5K in question. Reboot sometimes will fix it, but not always. I suspect a problem with the software since that doesn't happen in NX-OS 5.2. The one I am using is
    Software
      BIOS:      version 3.6.0
      loader:    version N/A
      kickstart: version 5.1(3)N2(1)
      system:    version 5.1(3)N2(1)

  • VSM and Cisco nexus 1000v

    Hi,
    We are planning to install Cisco Nexus 1000v in our environment. Before we want to install we want to explore little bit about Cisco Nexus 1000v
    •  I know there is 2 elements for Cisco 1k, VEM and VSM. Does VSM is required? Can we configure VEM individually?
    •   How does Nexus 1k integrated with vCenter. Can we do all Nexus 1000v configuration from vCenter without going to VEM or VSM?
    •   In term of alarming and reporting, does we need to get SNMP trap and get from individual VEM or can be use VSM to do that. OR can we   get    Cisco Nexus 1000v alarming and reporting form VMware vCenter.
    •  Apart from using Nexus 1010 can what’s the recommended hosting location for VSM, (same Host as VEM, different VM, and different physical server)
    Foyez Ahammed

    Hi Foyez,
    Here is a brief on the Nexus1000v and I'll answer some of your questions in that:
    The Nexus1000v is a Virtual Distributed Switch (software based) from Cisco which integrated with the vSphere environment to provide uniform networking across your vmware environment for the host as well as the VMs. There are two components to the N1K infrastructure 1) VSM 2) VEM.
    VSM - Virtual supervisor module is the one which controls the entire N1K setup and is from where the configuration is done for the VEM modules, interfaces, security, monitoring etc. VSM is the one which interacts with the VC.
    VEM - Virtual ethernet module are simply the module or virtual linecards which provide the connectivity option or virtual ports for the VMs and other virtaul interfaces. Each ESX host today can only have one VEM. These VEMs recieve their configuration / programing from the VSM.
    If you are aware of any other switching products from Cisco like the Cat 6k switches, the n1k behaves the same way but in a software / virtual environment. Where the VSM are equal of a SUPs and the VEM are similar to the line cards. The control and the packet VLANs in the n1k provide the same kind of AIPC and Inband connectivity as the 6k backplane would for the communication between the modules and the SUP (VSM in this case).
    *The n1k configuration is done only from the VSM and is visible in the VC.However the port-profiles created from the VSM are pushed from the VSM to the VC and have to be assigned to the virtual / physical ports from the VC.
    *You can run the VSM either on the Nexus1010 as a Virtual service blade (VSB) or as a normal VM on any of the ESX/ESXi server. The VSM and the VEM on the same server are fully supported.
    You can refer the following deployment guide for some more details: http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/guide_c07-556626.html
    Hope this answers your queries!
    ./Abhinav

  • Nexus 1000v UCS Manager and Cisco UCS M81KR

    Hello everyone
    I am confused about how works the integration between N1K and UCS Manager:
    First question:
    If two VMs on different ESXi and different VEM but in the same VLAN,would like to talk each other, the data flow between them is managed from the upstream switch( in this case UCS Fabric Inteconnect), isn'it?
    I created a Ethernet uplink port-profile on N1K in switch port mode access(100), I created a vEthernet port-profile for the VM in switchport mode access(100) as well. In the Fabric Interconnect I created a vNIC profile for the physical NICs of ESXi(where there are the VMs). Also I created the vlan 100(the same in N1K)
    Second question: With the configuration above, if I include in the vNIC profile the vlan 100 (not as native vlan) only, the two VMs can not ping each other. Instead if I include in the vNIC profile only the defaul vlan(I think it is the vlan 1) as native vlan evereything works fine. WHY????
    Third question: How it works the tagging vlan on Fabric interconnectr and also in N1K.
    I tried to read differnt documents, but I did not understand.
    Thanks                 

    This document may help...
    Best Practices in Deploying Cisco Nexus 1000V Series Switches on Cisco UCS B and C Series Cisco UCS Manager Servers
    http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/white_paper_c11-558242.html
    If two VMs on different ESXi and different VEM but in the same  VLAN,would like to talk each other, the data flow between them is  managed from the upstream switch( in this case UCS Fabric Inteconnect),  isn'it?
    -Yes.  Each ESX host with the VEM will have one or more dedicated NICs for the VEMs to communicate with the upstream network.  These would be your 'type ethernet' port-profiles.  The ustream network would need to bridge the vlan between the two physicall nics.
    Second question: With the configuration above, if I include in the vNIC  profile the vlan 100 (not as native vlan) only, the two VMs can not ping  each other. Instead if I include in the vNIC profile only the defaul  vlan(I think it is the vlan 1) as native vlan evereything works fine.  WHY????
    -  The N1K port profiles are switchport access making them untagged.  This would be the native vlan in ucs.  If there is no native vlan in the UCS configuration, we do not have the upstream networking bridging the vlan.
    Third question: How it works the tagging vlan on Fabric interconnectr and also in N1K.
    -  All ports on the UCS are effectively trunks and you can define what vlans are allowed on the trunk as well as what vlan is passed natively or untagged.  In N1K, you will want to leave your vEthernet port profiles as 'switchport mode access'.  For your Ethernet profiles, you will want them to be 'switchport mode trunk'.  Use an used used vlan as the native vlan.  All production vlans will be passed from N1K to UCS as tagged vlans.
    Thank You,
    Dan Laden
    PDI Helpdesk
    http://www.cisco.com/go/pdihelpdesk

  • Cisco LAN Management Solution is required to support Cisco Nexus 5548P and 5596UP switches?

    Hi,
    Could someone help to know what Cisco LAN Management Solution is required to support Cisco Nexus 5548P switches and Cisco Nexus 5596UP switches?
    These new Cisco switches are being implementing on customer network and he ask us that he requires these equipments be supported on a LMS solution (customer currently is using LMS 3.2.1)
    Can someone help?
    Thanks in advanced,
    guruiz

    Some very limited Nexus support is present in LMS 3.2.1 - see the supported device table here.
    To get more complete support, including the 5596UP, they need to upgrade to LMS 4.x (e.g.  LMS 4.2.2 is the latest and is sold under the Cisco Prime Infrastructure 1.2 umbrella). The major upgrade from 3.x to 4.x requires purchasing an upgrade license.
    Some functions (namely User Tracking ) will not be available on the 5k due to non-support of the requisite MIB on the device. I believe LMS still doesn't let you do VLAN management on 5k's - you need to use DCNM for that if you want to do it from a GUI.
    See the table here for LMS 4.2 device support.

  • Interoperability between Cisco Nexus 5548UP and IBM FC3171

    Dear Techies,
    We have a Cisco Nexus Switch 5548UP already in our environment with DS-SFP-FC8G-SW installed. We are planning to procure IBM Pure Flex Chassis without FC Switch, since we already have Nexus that can provide this facility. But in order to connect IBM Flex with Cisco Nexus we are planning to use IBM FC3171 Switch with IBM 44X1964 Transceivers which will act as a Mediator/Pass Through Switch for connecting IBM and Cisco.
    If any expert knows OR have used this scenario, please let me know since I am interested to know if it will work in this scenario or not.
    Regards,
    Farhan.

    Dear prkrishn
    I am working on the Data Center Solution between two Data Center, details underneath
    DC 1 Site
    1. 2 x DC Core Switch (Nexus 7009) - will be used for Servers, Application, IBM V7000, Database, etc.
    2. 2 x Network Core Switch (Cisco 6509) - Handle Campus and Inter Building connection.
    3. IBM V7000 (SAN)
    DC 2 Site
    1. 2 x DC Core Switch (Nexus 7009) - will be used for Servers, Application, IBM V7000, Database, etc.
    2. 2 x Network Core Switch (Cisco 6509) - Handle Campus and Inter Building connection.
    3. IBM V7000 (SAN)
    With the above mention setup, can I configure FCIP between DC1 & DC2 using Nexus 7009? or I need FCIP capable Switch such as IBM SAN Switch (SAN06-BR), I was wondering if I can configure FCIP on Nexus 7009 DC Switch.
    Hoping for your kind response at earliest.
    Kind Regards,
    Arnold

  • Cisco Nexus and 10GbE Autonegotiation not supported?

    Hi,
    I'm currently reading through the Interfaces Config Guide for Nexus 5600 NX-OS 7.x and have found the following statement:
    Autonegotiation configuration is not applicable on 10-Gigabit Ethernet ports. When autonegotiation is
    configured on a 10-Gigabit port the following error message is displayed:
    ERROR: Ethernet1/40: Configuration does not match the port capability
    Is it true that autonegotiation is not supported for 10GbE Interfaces on Cisco Nexus? I thought autonegotation was supported by 10GbE?! So would it be best practice so set a fixed speed on all Interfaces (servers, other switches) connected to the 10GbE ports on the Nexus switches?
    Thanks,
    Michael

    Hi Bilal,
    thanks. Yes, I know that 10GbE can run over both media types, the problem here is that the documentation starts with stating that auto negotiation is not supported for 10GbE in beneral but later references only 10-GBASE-T, so I was not sure what is true - whether auto-negotiation only not works with copper or with all ports.
    But I guess then it is true for all 10GbE ports, no matter what media-type. It's somewhat stange however, since I have read that auto-negotiation is generally supported for 10GbE (according to Wikipedia, at least). Maybe only Cisco doesn't supported auto-negotiation for 10GbE.
    Thank you again,
    Michael

  • Cisco Nexus 5548UP and FI6248UP compatibility with FC SFP

    Cisco Nexus 5548UP and FI 6248UP comes with Unified Ports. What are the SFP types this port can take? 1Gig, 10Gig and 2/4/8FC. Could you please clarify?  
    Thanks,
    Cheriyan

    Hi Cheriyan,
    Here is te URL to the 6200 series FI data sheet:
    http://www.cisco.com/en/US/prod/collateral/ps10265/ps11544/data_sheet_c78-675245.pdf
    Check for the table to supported SFPs.
    Same for the 5500 switches
    http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9670/data_sheet_c78-618603.html.
    Hope this helps!
    ./Abhinav

Maybe you are looking for