Nexus 7000, 2000, FCOE and Fabric Path

Hello,
I have a couple of design questions that I am hoping some of you can help me with.
I am working on a Dual DC Upgrade. It is pretty standard design, customer requires a L2 extension between the DC for Vmotion etc. Customer would like to leverage certain features of the Nexus product suite, including:
Trust Sec
VDC
VPC
High Bandwidth Scalability
Unified I/O
As always cost is a major issue and consolidation is encouraged where possible. I have worked on a couple of Nexus designs in the past and have levergaed the 7000, 5000, 2000 and 1000 in the DC.
The feedback that I am getting back from Customer seems to be mirrored in Cisco's technology roadmap. This relates specifically to the features supported in the Nexus 7000 and Nexus 5000.
Many large enterprise Customers ask the question of why they need to have the 7000 and 5000 in their topologies as many of the features they need are supported in both platforms and their environments will never scale to meet such a modular, tiered design.
I have a few specific questions that I am hoping can be answered:
The Nexus 7000 only supports the 2000 on the M series I/O Modules; can FCOE be implemented on a 2000 connected to a 7000 using the M series I/O Module?
Is the F Series I/O Module the only I/O Module that supports FCOE?
Are there any plans to introduce the native FC support on the Nexus 7000?
Are there any plans to introduce full fabric support (230 Gbps) to the M series I/O module?
Are there any plans to introduce Fabric path to the M series I/O module?
Are there any plans to introduce L3 support to the F series I/O Module?
Is the entire 2000 series allocated to a single VDC or can individual 2000 series ports be allocated to a VDC?
Is Trust Sec only support on multi hop DCI links when using the ASR on EoMPLS pwire?
Are there any plans to inroduce Trust Sec and VDC to the Nexus 5500?
Thanks,
Colm

Hello Allan
The only IO card which cannot co-exist with other cards in the same VDC is F2 due to specific hardware realisation.
All other cards can be mixed.
Regarding the Fabric versions - Fabric-2 gives much bigger throughoutput in comparing with Fabric-1
So in order to get full speed from F2/M2 modules you will need Fab-2 modules.
Fab2 modules won't give any advantages to M1/F1 modules.
http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9402/data_sheet_c78-685394.html
http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9402/prodcut_bulletin_c25-688075.html
HTH,
Alex

Similar Messages

  • Nexus 7000 with VPC and HSRP Configuration

    Hi Guys,
    I would like to know how to implement HSRP with the following setup:
    There are 2 Nexus 7000 connected with VPC Peer link. Each of the Nexus 7000 has a FEX attached to it.
    The server has two connections going to the FEX on each Nexus 7k (VPC). FEX's are not dual homed as far as I now they are not supported currently.
    R(A)              R(S)
    |                     |
    7K Peer Link 7K
    |                     |
    FEX              FEX
    Server connected to both FEX
    The question is we have two routers connected to each of the Nexus 7k in HSRP (active and one is standby). How can I configure HSRP on the nexus switches and how the traffic will routed from the Standby Nexus switch to Active Nexus switch (I know HSRP works differently here as both of them can forward packets). Will the traffic go to the secondary switch and then via the peer link to the active switch and then to the active router ? (From what I read the packet from end hosts which will go via the peer link will get dropped)
    Has anyone implemented this before ?
    Thanks

    Hi Kuldeep,
    If you intend to put those routers on a non-vpc vlan, you  may create  a new inter-switch trunk between the N7K and allow that non-vpc vlan . However if those will be on a VPC vlan, best to create two links to the N7K pair and create a VPC, otherwise configure those ports as orphan ports which will leverage the VPC peer link .
    HTH
    Jay Ocampo

  • Nexus 7000 and 2000. Is FEX supported with vPC?

    I know this was not supported a few months ago, curious if anything has changed?

    Hi Jenny,
    I think the answer will depend on what you mean by is FEX supported with vPC?
    When connecting a FEX to the Nexus 7000 you're able to run vPC from the Host Interfaces of a pair of FEX to an end system running IEEE 802.1AX (802.3ad) Link Aggregation. This is shown is illustration 7 of the diagram shown on the post Nexus 7000 Fex Supported/Not Supported Topologies.
    What you're not able to do is run vPC on the FEX Network Interface that connect up to the Nexus 7000 i.e., dual-homing the FEX to two Nexus 7000. This is shown in illustrations 8 and 9 of under the FEX topologies not supported on the same page.
    There's some discussion on this in the forum post DualHoming 2248TP-E to N7K that explains why it's not supported, but essentially it offers no additional resilience.
    From that post:
    The view is that when connecting FEX to the Nexus 7000, dual-homing does not add any level of resilience to the design. A server with dual NIC can attach to two FEX  so there is no need to connect the FEX to two parent switches. A server with only a single NIC can only attach to a single FEX, but given that FEX is supported by a fully redundant Nexus 7000 i.e., SE, fabrics, power, I/O modules etc., the availability is limited by the single FEX and so dual-homing does not increase availability.
    Regards

  • Ask the Expert: Basic Introduction and Troubleshooting on Cisco Nexus 7000 NX-OS Virtual Device Context

    With Vignesh R. P.
    Welcome to the Cisco Support Community Ask the Expert conversation.This is an opportunity to learn and ask questions of Cisco expert Vignesh R. P. about the Cisco® Nexus 7000 Series Switches and support for the Cisco NX-OS Software platform .
    The Cisco® Nexus 7000 Series Switches introduce support for the Cisco NX-OS Software platform, a new class of operating system designed for data centers. Based on the Cisco MDS 9000 SAN-OS platform, Cisco NX-OS introduces support for virtual device contexts (VDCs), which allows the switches to be virtualized at the device level. Each configured VDC presents itself as a unique device to connected users within the framework of that physical switch. The VDC runs as a separate logical entity within the switch, maintaining its own unique set of running software processes, having its own configuration, and being managed by a separate administrator.
    Vignesh R. P. is a customer support engineer in the Cisco High Touch Technical Support center in Bangalore, India, supporting Cisco's major service provider customers in routing and MPLS technologies. His areas of expertise include routing, switching, and MPLS. Previously at Cisco he worked as a network consulting engineer for enterprise customers. He has been in the networking industry for 8 years and holds CCIE certification in the Routing & Switching and Service Provider tracks.
    Remember to use the rating system to let Vignesh know if you have received an adequate response. 
    Vignesh might not be able to answer each question due to the volume expected during this event. Remember that you can continue the conversation on the  Data Center sub-community discussion forum shortly after the event. This event lasts through through January 18, 2013. Visit this forum often to view responses to your questions and the questions of other community members.

    Hi Vignesh
    Is there is any limitation to connect a N2K directly to the N7K?
    if i have a an F2 card 10G and another F2 card 1G and i want to creat 3 VDC'S
    VDC1=DC-Core
    VDC2=Aggregation
    VDC3=Campus core
    do we need to add a link between the different VDC's
    thanks

  • FCoE and multiple Nexus 5000s and a 7000 core

    Hi
    I have a customer who is looking at four Nexus 5020s to start with and more in the future uplinked to a Nexus 7000 core.
    Am I right in thinking that whilst data traffic will be able to reach hosts ocnnected to a different Nexus 5002 via the 7000 core FCoE traffic will not ?
    If so what is teh recommended way of rolling out pods of 5020s for VMware servers with converged network adaptors so that they can all access the SAN ?
    Regards
    Pat

    Don't use corecenter for fan speed control.  Instead, use the BIOS:
    1) Open Corecenter from the administrator account or a user account that has admin privileges.  Click on the top center logo which should open the fan speed control window.  There should be two items, CoolnQuiet and User/Manual or something like.  Put it in manual and move the slider all the way to the right. Ignore the CoolnQuiet, it's misnamed here - MSI's corecenter does not control CoolnQuiet.  What MSI is calling CoolnQuiet is just fan speed control.  Anyway, put it in manual.  Close Corecenter and you should never have to open it again unless you want to monitor fan RPM's.
    2) Next, reboot and enter the BIOS.  Go to the H/W  Monitor settings.  Turn on Smart CPU Fan Speed.  Set to 40 C +/- 1.    This will allow the motherboard to control the fan speed.  On the other hand, if you want your fan always at maximum, then disable Smart Fan Speen control.   Don't worry about the Smart NB Fan speed, leave it disabled.
    Also, keep in mind that when you first open Corecenter, the immediate fan RPM's reported are not correct.  It takes it a few seconds to get the readings.  So wait until Corecenter minimizes itself to the system tray, then open it from there, and you will see the correct fan RPM's.

  • Ask the Expert: Cisco Nexus 2000, 5000, and 6000 Series Switches

    with Cisco Expert Vinayak Sudame
    Welcome to the Cisco Support Community Ask the Expert conversation. This is an opportunity to learn and ask questions how to configure and troubleshoot the Cisco Nexus 2000, 5000 and 6000 Series Switches with Cisco subject matter expert Vinayak Sudame. You can ask any question on configuration, troubleshooting, features, design and Fiber Channel over Ethernet (FCoE).
    Vinayak Sudame is a Technical Lead in Data Center Switching Support Team within Cisco's Technical Services in RTP, North Carolina. His current responsibilities include but are not limited to Troubleshooting Technical support problems and Escalations in the areas of Nexus 5000, Nexus 2000, FCoE. Vinayak is also involved in developing technical content for Cisco Internal as well as external. eg, Nexus 5000 Troubleshooting Guide (CCO), Nexus 5000 portal (partners), etc. This involves cross team collaboration and working with multiple different teams within Cisco. Vinayak has also contributed to training account teams and partners in CAE (Customer Assurance Engineering) bootcamp dealing with Nexus 5000 technologies. In the past, Vinayak's responsibilities included supporting MDS platform (Fiber Channel Technologies) and work with EMC support on Escalated MDS cases. Vinayak was the Subject Matter Expert for Santap Technologies before moving to Nexus 5000 support. Vinayak holds a Masters in Electrical Engineering with Specialization in Networking from Wichita State University, Kansas. He also holds Cisco Certification CCIE (#20672) in Routing and Switching.
    Remember to use the rating system to let Vinayak know if you have received an adequate response.
    Vinayak might not be able to answer each question due to the volume expected during this event. Remember that you can continue the conversation on the  Data Center sub-community, Other Data Center Topics discussion forum shortly after the event.
    This event last through Friday July 12, 2013. Visit the community often to view responses to youe questions of other community members.

    Hi Vinayak,
    Output of "show cfs internal ethernet-peer database"
    Switch 1
    ETH Fabric
    Switch WWN              logical-if_index
    20:00:54:7f:ee:b7:c2:80 [Local]
    20:00:54:7f:ee:b6:3f:80 16000005
    Total number of entries = 2
    Switch 2
    ETH Fabric
    Switch WWN              logical-if_index
    20:00:54:7f:ee:b6:3f:80 [Local]
    20:00:54:7f:ee:b7:c2:80 16000005
    Total number of entries = 2
    Output of "show system internal csm info trace"
    Switch 1 in which "show cfs peers" show proper output
    Mon Jul  1 05:46:19.145339  (CSM_T) csm_sp_buf_cmd_tbl_expand_range(8604): No range command in buf_cmd_tbl.
    Mon Jul  1 05:46:19.145280  (CSM_T) csm_set_sync_status(6257): Peer RT status PSSed
    Mon Jul  1 05:46:19.145188  (CSM_T) csm_sp_handle_local_verify_commit(4291):
    Mon Jul  1 05:46:19.145131  csm_continue_verify_ac[597]: peer is not reachable over CFS so continuing with local verify/commit
    Mon Jul  1 05:46:19.145071  csm_tl_lock(766): Peer information not found for IP address: '172.16.1.54'
    Mon Jul  1 05:46:19.145011  csm_tl_lock(737):
    Mon Jul  1 05:46:19.144955  (CSM_EV) csm_sp_build_tl_lock_req_n_send(941): sending lock-request for CONF_SYNC_TL_SESSION_TYPE_VERIFY subtype 0 to Peer ip = (172.16.1.54)
    Mon Jul  1 05:46:19.143819  (CSM_T) csm_copy_image_and_internal_versions(788): sw_img_ver: 5.2(1)N1(2a), int_rev: 1
    Mon Jul  1 05:46:19.143761  (CSM_T) csm_sp_get_peer_sync_rev(329): found the peer with address=172.16.1.54 and sync_rev=78
    Mon Jul  1 05:46:19.143699  (CSM_T) csm_sp_get_peer_sync_rev(315):
    Mon Jul  1 05:46:19.143641  (CSM_EV) csm_sp_build_tl_lock_req_n_send(838): Entered fn
    Mon Jul  1 05:46:19.143582  (CSM_T) csm_set_sync_status(6257): Peer RT status PSSed
    Switch 2 in which "show cfs peers" does not show proper output
    Mon Jul  1 06:13:11.885354  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport mode trunk, cmd pseq 77 seq 482
    Mon Jul  1 06:13:11.884992  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd channel-group 51 mode active, cmd pseq 357 seq 369
    Mon Jul  1 06:13:11.884932  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport trunk allowed vlan 2, 11, cmd pseq 357 seq 368
    Mon Jul  1 06:13:11.884872  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport mode trunk, cmd pseq 357 seq 367
    Mon Jul  1 06:13:11.884811  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd description process_vpc, cmd pseq 357 seq 366
    Mon Jul  1 06:13:11.884750  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd channel-group 51 mode active, cmd pseq 352 seq 365
    Mon Jul  1 06:13:11.884690  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport trunk allowed vlan 2, 11, cmd pseq 352 seq 364
    Mon Jul  1 06:13:11.884630  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd switchport mode trunk, cmd pseq 352 seq 363
    Mon Jul  1 06:13:11.884568  (CSM_ERR) csm_pss_cmd_tree_walk_cb(2057): Parent command not found for cmd description process_vpc, cmd pseq 352 seq 362
    Mon Jul  1 06:13:11.884207  (CSM_EV) csm_sp_acfg_gen_handler(3011):  Preparing config into /tmp/csm_sp_acfg_1733916569.txt
    Mon Jul  1 06:13:11.878695  csm_get_locked_ssn_ctxt[539]: Lock not yet taken.
    Mon Jul  1 06:13:11.878638  (CSM_EV) csm_sp_acfg_gen_handler(2937): Recieved sp acfg merge request for type: running cfg
    Mon Jul  1 06:12:29.527840  (CSM_T) csm_pss_del_seq_tbl(1989): Freeing seq tbl data
    Mon Jul  1 06:12:29.513255  (CSM_T) csm_sp_acfg_gen_handler(3106): Done acfg file write
    Mon Jul  1 06:12:29.513179  (CSM_EV) csm_sp_acfg_gen_handler(3011):  Preparing config into /tmp/csm_sp_acfg_1733911262.txt
    Mon Jul  1 06:12:29.508859  csm_get_locked_ssn_ctxt[539]: Lock not yet taken.
    Mon Jul  1 06:12:29.508803  (CSM_EV) csm_sp_acfg_gen_handler(2937): Recieved sp acfg merge request for type: running cfg
    Mon Jul  1 05:53:17.651236  Collecting peer info
    Mon Jul  1 05:53:17.651181  Failed to get the argumentvalue for 'ip-address'
    Mon Jul  1 05:40:59.262736  DB Unlocked Successfully
    Mon Jul  1 05:40:59.262654  Unlocking DB, Lock Owner Details:Client:1 ID:1
    Mon Jul  1 05:40:59.262570  (CSM_T) csm_sp_del_buf_cmd(1713): Deleting comand with Id = 1
    Mon Jul  1 05:40:59.262513  DB Lock Successful by Client:1 ID:1
    Mon Jul  1 05:40:59.262435  Recieved lock request by Client:1 ID:1
    Mon Jul  1 05:40:41.741224  ssnmgr_ssn_handle_create_get: Session FSM already present, ID:1
    Mon Jul  1 05:40:41.741167  ssnmgr_handle_mgmt_request: Create/Get request received for session[process_n5kprof]
    show cfs lock gives no output.
    Just to further clarify, we have 4 5548UP switches in the same management vlan. 2 switches are in one location lets say location A and they are CFS peers and are working fine.
    These two switches which are having problem are in location B. All the switches are in the same vlan. Essentially the all CFS multicast messages will be seen by all 5548 switches as they are in the same vlan. I am assuming that this might not create any problems as we specify the peers in the respective configurations. Or do we have to change the CFSoIPv4 multicast addresses in location B or may be configure a different region.
    Regards.

  • Dell Servers with Nexus 7000 + Nexus 2000 extenders

    << Original post by smunzani. Answered by Robert. Moving from Document section to Discussions>>
    Team,
    I would like to use some of the existing Dell Servers for new network design of Nexus 7000 + Nexus 2000 extenders. What are my options for FEC to the hosts? All references of M81KR I found on CCO are related to UCS product only.
    What's best option for following setup?
    N7K(Aggregation Layer) -- N2K(Extenders) -- Dell servers
    Need 10G to the servers due to dense population of the VMs. The customer is not up for dumping recently purchased dell boxes in favor of UCS. Customer VMware license is Enterprise Edition.
    Thanks in advance.

    To answer your question, the M81KR-VIC is a Mezz card for UCS blades only.  For Cisco rack there is a PCIe version which is called the P81.  These are both made for Cisco servers only due to the integration with server management and virtual interface functionality.
    http://www.cisco.com/en/US/prod/collateral/ps10265/ps10493/data_sheet_c78-558230.html
    More information on it here:
    Regards,
    Robert

  • ESXi 4.1 NIC Teaming's Load-Balancing Algorithm,Nexus 7000 and UCS

    Hi, Cisco Gurus:
    Please help me in answering the following questions (UCSM 1.4(xx), 2 UCS 6140XP, 2 Nexus 7000, M81KR in B200-M2, No Nexus 1000V, using VMware Distributed Switch:
    Q1. For me to configure vPC on a pair of Nexus 7000, do I have to connect Ethernet Uplink from each Cisco Fabric Interconnect to the 2 Nexus 7000 in a bow-tie fashion? If I connect, say 2 10G ports from Fabric Interconnect 1 to 1 Nexus 7000 and similar connection from FInterconnect 2 to the other Nexus 7000, in this case can I still configure vPC or is it a validated design? If it is, what is the pro and con versus having 2 connections from each FInterconnect to 2 separate Nexus 7000?
    Q2. If vPC is to be configured in Nexus 7000, is it COMPULSORY to configure Port Channel for the 2 Fabric Interconnects using UCSM? I believe it is not. But what is the pro and con of HAVING NO Port Channel within UCS versus HAVING Port Channel when vPC is concerned?
    Q3. if vPC is to be configured in Nexus 7000, I understand there is a limitation on confining to ONLY 1 vSphere NIC Teaming's Load-Balancing Algorithm i.e. Route Based on IP Hash. Is it correct?
    Again, what is the pro and con here with regard to application behaviours when Layer 2 or 3 is concerned? Or what is the BEST PRACTICES?
    I would really appreciate if someone can help me clear these lingering doubts of mine.
    God Bless.
    SiM

    Sim,
    Here are my thoughts without a 1000v in place,
    Q1. For me to configure vPC on a pair of Nexus 7000, do I have to connect Ethernet Uplink from each Cisco Fabric Interconnect to the 2 Nexus 7000 in a bow-tie fashion? If I connect, say 2 10G ports from Fabric Interconnect 1 to 1 Nexus 7000 and similar connection from FInterconnect 2 to the other Nexus 7000, in this case can I still configure vPC or is it a validated design? If it is, what is the pro and con versus having 2 connections from each FInterconnect to 2 separate Nexus 7000?   //Yes, for vPC to UCS the best practice is to bowtie uplink to (2) 7K or 5Ks.
    Q2. If vPC is to be configured in Nexus 7000, is it COMPULSORY to configure Port Channel for the 2 Fabric Interconnects using UCSM? I believe it is not. But what is the pro and con of HAVING NO Port Channel within UCS versus HAVING Port Channel when vPC is concerned? //The port channel will be configured on both the UCSM and the 7K. The pro of a port channel would be both bandwidth and redundancy. vPC would be prefered.
    Q3. if vPC is to be configured in Nexus 7000, I understand there is a limitation on confining to ONLY 1 vSphere NIC Teaming's Load-Balancing Algorithm i.e. Route Based on IP Hash. Is it correct? //Without the 1000v, I always tend to leave to dvSwitch load balence behavior at the default of "route by portID". 
    Again, what is the pro and con here with regard to application behaviours when Layer 2 or 3 is concerned? Or what is the BEST PRACTICES? UCS can perform L2 but Northbound should be performing L3.
    Cheers,
    David Jarzynka

  • Layer 3 config design on Nexus 5500 with Fabric Path

    I trying to Network deisgn for new data Center ,  i am new to DataCenter desgin, i attached the network diagram
    i would like to know if can configure my layer3 on 5500 and configure Fabric path to uplink switch
    please help give your suggestions on this design 

    You can configure layer-3 on the 5500 series, but you need to install a daughter cards in each 5500.
    See this link:
    Layer 3 Daughter Card and Expansion Module Options for Cisco Nexus 5548P, 5548UP, 5596UP, and 5596T Switches
    http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9670/data_sheet_c78-618603.html
    HTH

  • Nexus 7000 fcoe expert advice

    Hi,
    I have one Nexus 7000 with a fcoe supported blade N7K-F132XP-15.
    Want to deply this in the LAB. I created a fcoe port for test purpose which is working. But some things are not very clear to me. I am pointing them below.
    How to create the interface membership for storage VDC. Right now i created shared interfcae with the default VDC. But can i create the interfcae dedicated? What is the difference between when the interface is shared and dedicated ?
    Also, on my shared interfcae while configuring FCoE port, it did not allow me to create priority flow control settings. See the error below.
    fcoe-dvt(config-if)# priority-flow-control mode auto
    ERROR: pfc config not allowed on shared interface (0x1a000000)
    fcoe-dvt(config-if)#
    On Nexus 5020, we have priority flow control settings as "auto" for eahc fcoe interface. In the nexus 7000 case, i dont know what and how priority flow control is taking place.
    Any white paper on Nexus 7000 fcoe will be greatly appreciated.
    Thanks,

    Marko,
    Yes it is supported, here is a link to the interoperability matrix for storage, also  you can find CNA compatibility
    here as well. It is a difficult to find on cisco.com because it has been added in with the MDS.
    http://www.cisco.com/en/US/docs/switches/datacenter/mds9000/interoperability/matrix/Matrix1.html
    Thanks,
    Bill

  • Requirements about delay and bandwith for using OTV in Nexus 7000 between two data centers separated 25 miles?

    We have two Nexus 7000, and I need use them with OTV between two data Centers separated 25 miles, but I don´t know what are the optimal values about bandwidth and delay (ms) for extended VLANs IDs (production and DAG replication) for Microsoft Exchange environment. Can somebody tell me please which are the values required for operate OTV in optimal conditions in this case? We have about 35 000 users that will use that platform of email. Thanks a lot for your comments. Regards.

    We have two Nexus 7000, and I need use them with OTV between two data Centers separated 25 miles, but I don´t know what are the optimal values about bandwidth and delay (ms) for extended VLANs IDs (production and DAG replication) for Microsoft Exchange environment. Can somebody tell me please which are the values required for operate OTV in optimal conditions in this case? We have about 35 000 users that will use that platform of email. Thanks a lot for your comments. Regards.

  • Ciscoworks 2.6 and Nexus 7000 issues

    Running LMS 2.6 with RME version 4.0.6, and DFM 2.0.13.
    We keep getting false alerts in DFM on the temperature in our Nexus 7000 switches. The alert says that the high temp threshold is 45C, and it's being exceeded at 46C. The thing that bothers me is that the actual switch reads that the threshold is around 100C or more. Any ideas as to why DFM would be picking up a temperature so far off the mark?
    Also, in regards to RME, I cannot pull configs from the Nexus 7000's. The check box in "archive config" is blanked out to where I can't check it. I download the device packages for the 7000 into RME but it will not pull configs. Is this not supported under our version of RME, or would there be some other reason that I can't do this?
    Thanks for any assistance with these issues!

    UPDATE:
    I fixed the RMA config pull issue. I thought I had previously downloaded the Nexus device packages so that RMA could work with them, but upon checking again, it looks like I just didn't have them installed. Got that piece fixed and now I can pull configs from the switches just fine.
    Still having problems with the temperature reading in DFM not accurately reflecting what is actually on the switches. Any suggestions as to where to start hunting down the issue for this are greatly appreciated. Thanks!

  • Nexus 5596 to join Brocade fabric

    Hi all, I tried to search but didn't find a "real" answer, it is not clear if it is possible or not...
    Problem: I always worked with homogeneous Brocade fabrics. Some days ago my company asked me if it is possible to add a Nexus 5596UB to an existing (6 switches) production Brocade fabric. I underlined production as we should not interrupt SAN traffic on joining the fabric nor we should reboot and/or reconfigure existing switches.
    I found a lot of info about "interoperability mode" on MDS switches but nothing on Nexus.
    Can someone help me? Thanks.

    According to the datasheet
    http://www.cisco.com/c/en/us/products/collateral/switches/nexus-5000-series-switches/data_sheet_c78-618603.html
    Native interop 1,2,3 and 4 is supported !
    Therefore you are ok, and have a TAC supported configuration.
    Fibre Channel and FCoE Features (Requires Storage Services License)
    ●  T11 standards-compliant FCoE (FC-BB-5)
    ●  T11 FCoE Initialization Protocol (FIP) (FC-BB-5)
    ●  Any 10 Gigabit Ethernet port configurable as FCoE
    ●  SAN administration separate from LAN administration
    ●  FCP
    ●  Fibre Channel forwarding (FCF)
    ●  Fibre Channel standard port types: E, F, and NP
    ●  Fibre Channel enhanced port types: VE, TE, and VF
    ●  F-port trunking
    ●  F-port channeling
    ●  Direct attachment of FCoE and Fibre Channel targets
    ●  Up to 240 buffer credits per native Fibre Channel port
    ●  Up to 32 VSANs per switch
    ●  Fibre Channel (SAN) PortChannel
    ●  Native Interop Mode 1
    ●  Native Interop Mode 2
    ●  Native Interop Mode 3
    ●  Native Interop Mode 4
    ●  VSAN trunking
    ●  Fabric Device Management Interface (FDMI)
    ●  Fibre Channel ID (FCID) persistence
    ●  Distributed device alias services
    ●  In-order delivery
    ●  Port tracking
    ●  Cisco N-Port Virtualization (NPV) technology
    ●  N-port identifier virtualization (NPIV)
    ●  Fabric services: Name server, registered state change notification (RSCN), login services, and name-server zoning
    ●  Per-VSAN fabric services
    ●  Cisco Fabric Services
    ●  Diffie-Hellman Challenge Handshake Authentication Protocol (DH-CHAP) and Fibre Channel Security Protocol (FC-SP)
    ●  Distributed device alias services
    ●  Host-to-switch and switch-to-switch FC-SP authentication
    ●  Fabric Shortest Path First (FSPF)
    ●  Fabric binding for Fibre Channel
    ●  Standard zoning
    ●  Port security
    ●  Domain and port
    ●  Enhanced zoning
    ●  SAN PortChannels
    ●  Cisco Fabric Analyzer
    ●  Fibre Channel traceroute
    ●  Fibre Channel ping
    ●  Fibre Channel debugging
    ●  Cisco Fabric Manager support
    ●  Storage Management Initiative Specification (SMI-S)
    ●  Boot from SAN over VPC/EVPC

  • NEXUS 7000 xcvrInval

    Hi,
    I have some Nexus 7000 with FET-10G with xcvrInval status
    Eth7/33          N5k-S1-3T-1/3   xcvrInval trunk     auto    auto    Fabric Exte
    and some other FET-10G with notconn status
    Eth7/8           FEX-101         notconnec 1         auto    auto    Fabric Exte
    If I inter exchange the position of both FET-10G the status port doesn´t change
    FET-10G from 7/8 to 7/33
    FET-10G from 7/33 to 7/8
    7/33 holds xcvrInval status
    7/8 holds notconnec status
    I have reconfigured from default interface with same results
    Next you´ll find the same serial number in deferent port, the diference is the current
    when is xcvrInva or when is notconnec
    What can I do to get FET10G in e7/33 validated?
    sh interface e7/33   transceiver details
    Ethernet7/33
          transceiver is present
          type is Fabric Extender Transceiver
          name is CISCO-FINISAR  
          part number is FTLX8570D3BCL-C2
          revision is A  
          serial number is FNS17201TE5    
          nominal bitrate is 10300 MBit/sec
          Link length supported for 62.5/125um fiber is 10 m
          Link length supported for 50/125um OM3 fiber is 100 m
          cisco id is --
          cisco extended id number is 4
          cisco part number is 10-2566-02
          cisco product id is FET-10G            
          cisco vendor id is V02
          number of lanes 1
                 SFP Detail Diagnostics Information (internal calibration)
                    Current              Alarms                  Warnings
                    Measurement     High        Low         High          Low
        Temperature   19.30 C        75.00 C      5.00 C     70.00 C       10.00 C
    [7m--More-- [m
        Voltage        3.31 V         3.63 V      2.97 V      3.46 V        3.13 V
        Current        0.06 mA  --     11.80 mA     4.00 mA    10.80 mA       5.00 mA
        Tx Power          N/A        22.69 dBm    8.69 dBm     18.69 dBm     12.69 dBm
        Rx Power          N/A        22.99 dBm    6.09 dBm     18.99 dBm     10.09 dBm
        Transmit Fault Count = 0
        Note: ++  high-alarm; +  high-warning; --  low-alarm; -  low-warning
    now in slot 7/8
    Ethernet7/8
          transceiver is present
          type is Fabric Extender Transceiver
          name is CISCO-FINISAR  
          part number is FTLX8570D3BCL-C2
          revision is A  
          serial number is FNS17201TE5    
          nominal bitrate is 10300 MBit/sec
          Link length supported for 62.5/125um fiber is 10 m
          Link length supported for 50/125um OM3 fiber is 100 m
          cisco id is --
          cisco extended id number is 4
          cisco part number is 10-2566-02
          cisco product id is FET-10G            
          cisco vendor id is V02
          number of lanes 1
                 SFP Detail Diagnostics Information (internal calibration)
                    Current              Alarms                  Warnings
                    Measurement     High        Low         High          Low
        Temperature   23.17 C        75.00 C      5.00 C     70.00 C       10.00 C
    [7m--More-- [m
        Voltage        3.30 V         3.63 V      2.97 V      3.46 V        3.13 V
        Current        7.50 mA       11.80 mA     4.00 mA    10.80 mA       5.00 mA
        Tx Power      17.65 dBm      22.69 dBm    8.69 dBm     18.69 dBm     12.69 dBm
        Rx Power     -12.21 dBm --   22.99 dBm    6.09 dBm     18.99 dBm     10.09 dBm
        Transmit Fault Count = 0
        Note: ++  high-alarm; +  high-warning; --  low-alarm; -  low-warning
    NX7K-1-VDC-3T-S1-L2FP# sh int e7/33
    Ethernet7/33 is down (Transceiver validation failed)
    admin state is up, Dedicated Interface
      Belongs to Po51
      Hardware: 1000/10000 Ethernet, address: 8478.ac23.6cec (bia 8478.ac23.6cec)
      Description: N5k-S1-3T-1/3
      MTU bytes (CoS values):  MTU  1500(0-2,4-7) bytes  MTU  2112(3) bytes
      BW 10000000 Kbit, DLY 10 usec, reliability 255/255, txload 1/255, rxload 1/255
      Encapsulation ARPA, medium is broadcast
      Port mode is trunk
    auto-speed  auto-duplex,, media type is 10G
      Beacon is turned off
      Auto-Negotiation is turned on
      Input flow-control is off, output flow-control is off
      Auto-mdix is turned on
      Rate mode is dedicated
      Switchport monitor is off
      EtherType is 0x8100
      EEE (efficient-ethernet) : n/a
      Last link flapped never
      Last clearing of "show interface" counters 07:22:09
      0 interface resets
      Load-Interval #1: 30 seconds
        30 seconds input rate 0 bits/sec, 0 packets/sec
        30 seconds output rate 0 bits/sec, 0 packets/sec
      Load-Interval #2: 5 minute (300 seconds)
        300 seconds input rate 0 bits/sec, 0 packets/sec
        300 seconds output rate 0 bits/sec, 0 packets/sec
      RX
        88 unicast packets  0 multicast packets  0 broadcast packets
        0 input packets  0 bytes
        0 jumbo packets  0 storm suppression packets
        0 runts  0 giants  0 CRC/FCS  0 no buffer
        0 input error  0 short frame  0 overrun   0 underrun  0 ignored
        0 watchdog  0 bad etype drop  0 bad proto drop  0 if down drop
        0 input with dribble  0 input discard
        0 Rx pause
      TX
        88 unicast packets  0 multicast packets  0 broadcast packets
        0 output packets  0 bytes
        0 jumbo packets
        0 output error  0 collision  0 deferred  0 late collision
        0 lost carrier  0 no carrier  0 babble  0 output discard
        0 Tx pause
    NX7K-1-VDC-3T-S1-L2FP# sh int e7/33
    Ethernet7/8 is down (Link not connected)
    admin state is up, Dedicated Interface
      Belongs to Po101
      Hardware: 1000/10000 Ethernet, address: 8478.ac23.6cd3 (bia 8478.ac23.6cd3)
      Description: FEX-101
      MTU bytes (CoS values):  MTU  1500(0-2,4-7) bytes  MTU  2112(3) bytes
      BW 10000000 Kbit, DLY 10 usec, reliability 255/255, txload 1/255, rxload 1/255
      Encapsulation ARPA, medium is p2p
      Port mode is fex-fabric
    auto-speed  auto-duplex,, media type is 10G
      Beacon is turned off
      Auto-Negotiation is turned on
      Input flow-control is off, output flow-control is off
      Auto-mdix is turned on
      Rate mode is dedicated
      Switchport monitor is off
      EtherType is 0x8100
      EEE (efficient-ethernet) : n/a
      Last link flapped 5week(s) 1day(s)
      Last clearing of "show interface" counters never
      0 interface resets
      Load-Interval #1: 30 seconds
        30 seconds input rate 0 bits/sec, 0 packets/sec
    [7m--More-- [m
        30 seconds output rate 0 bits/sec, 0 packets/sec
      Load-Interval #2: 5 minute (300 seconds)
        300 seconds input rate 0 bits/sec, 0 packets/sec
        300 seconds output rate 0 bits/sec, 0 packets/sec
      RX
        10588 unicast packets  0 multicast packets  0 broadcast packets
        4 input packets  0 bytes
        0 jumbo packets  0 storm suppression packets
        0 runts  0 giants  0 CRC/FCS  0 no buffer
        0 input error  0 short frame  0 overrun   0 underrun  0 ignored
        0 watchdog  0 bad etype drop  0 bad proto drop  0 if down drop
        0 input with dribble  0 input discard
        0 Rx pause
      TX
        10588 unicast packets  1 multicast packets  0 broadcast packets
        4 output packets  5688 bytes
        0 jumbo packets
        0 output error  0 collision  0 deferred  0 late collision
        0 lost carrier  0 no carrier  0 babble  0 output discard
        0 Tx pause

    Hi Ans,
    You are rigth, I have defaulted againt the port, now configured with switchport mode FEX, and now the FET-10G is validated
    NX7K-1-VDC-3T-S1-L2FP(config-if)#  description FEX-101
    NX7K-1-VDC-3T-S1-L2FP(config-if)#   switchport
    NX7K-1-VDC-3T-S1-L2FP(config-if)#   switchport mode fex-fabric
    NX7K-1-VDC-3T-S1-L2FP(config-if)#   fex associate 101
    NX7K-1-VDC-3T-S1-L2FP(config-if)#   medium p2p
    NX7K-1-VDC-3T-S1-L2FP(config-if)#   channel-group 101
    NX7K-1-VDC-3T-S1-L2FP(config-if)#   no shutdown
    NX7K-1-VDC-3T-S1-L2FP(config-if)#
    NX7K-1-VDC-3T-S1-L2FP(config-if)# sh int e7/33 status
    Port             Name            Status    Vlan      Duplex  Speed   Type
    Eth7/33          FEX-101         notconnec 1         auto    auto    Fabric Exte
    NX7K-1-VDC-3T-S1-L2FP(config-if)#
    Thanks for your help, and have a nice weekend.
    Atte,
    EF

  • SFP with Nexus 5500/2000

    We ordered two 5548P and several 2000. They came with two types of SFPs. One is SFP-10G-SR, the other FET-10G.
    What is FET-10G used for? between 2000 FEX and computer servers?
    thanks,
    Han

    It will show invalid SFP until you enable the fex feature and configure the interfaces/port-channel as fex fabric. You have to use the same on both ends, either FET or 10G-SR.
    example
    feature fex
    interface port-channel101
      description Port channel, vPC 101 to FEX 101 Nexus 2248
      switchport mode fex-fabric
      fex associate 101
      vpc 101
    interface Ethernet1/32
      description Connection to FEX 101 Nexus 2248 Fabric port 1/1
      switchport mode fex-fabric
      fex associate 101
      channel-group 101

Maybe you are looking for