Note: This is the second of a three-part series on Next Generation Data Center Design with MDS 9700; learn how customers can deploy scalable SAN networks that allow them to Scale Up or Scale Out in a non disruptive way. Part 1 | Part 3 ]
EMC World was wonderful. It was gratifying to meet industry professionals, listen in on great presentations and watch the demos for key business enabling technologies that Cisco, EMC and others have brought to fruition. Its fascinating to see the transition of DC from cost center to a strategic business driver . The same repeated all over again at Cisco Live. More than 25000 attendees, hundreds of demos and sessions. Lot of interesting customer meetings and MDS continues to resonate. We are excited about the MDS hardware that was on the display on show floor and interesting Multiprotocol demo and a lot of interesting SAN sessions.
Outside these we recently did a webinar on how Cisco MDS 9710 is enabling High Performance DC design with customer case studies. You can listen to that here.
So let’s continue our discussion. There is no doubt when it comes to High Performance SAN switches there is no comparable to Cisco MDS 9710. Another component that is paramount to a good data center design is high availability. Massive virtualization, DC consolidation and ability to deploy more and more applications on powerful multi core CPUs has increased the risk profile within DC. These DC trends requires renewed focus on availability. MDS 9710 is leading the innovation there again. Hardware design and architecture has to guarantee high availability. At the same time, it’s not just about hardware but it’s a holistic approach with hardware, software, management and right architecture. Let me give you some just few examples of the first three pillars for high reliability and availability.
MDS 9710 is the only director in the industry that provides Hardware Redundancy on all critical components of the switch, including fabric cards. Cisco Director Switches provide not only CRC checks but ability to drop corrupted frames. Without that ability network infrastructure exposes the end devices to the corrupted frames. Having ability to drop the CRC frames and quickly isolate the failing links outside as well as inside of the director provides Data Integrity and fault resiliency. VSAN allows fault isolation, Port Channel provides smaller failure domains, DCNM provides rich feature set for higher availability and redundancy. All of these are but a subset of examples which provides high resiliency and reliability.
We are proud of the 9500 family and strong foundation for reliability and availability that we stand on. We have taken that to a completely new level with 9710. For any design within Data center high availability has to go hand in hand with consistent performance. One without the other doesn’t make sense. Right design and architecture with DC as is important as components that power the connectivity. As an example Cisco recommend customers to distribute the ISL ports of an Port Channel across multiple line cards and multiple ASICs. This spreads the failure domain such that any ASIC or even line card failures will not impact the port channel connectivity between switches and no need to reinitiate all the hosts logins. You can see white paper on Next generation Cisco MDS here. At part of writing this white paper ESG tested the Fabric Card redundancy (Page 9) in addition to other features of the platform. Remember that a chain is only as strong as its weakest link.
The most important aspect for all of this is for customer is to be educated.
Ask the right questions. Have in depth discussions to achieve higher availability and consistent performance. Most importantly selecting the right equipment, right architecture and best practices means no surprises.
We will continue our discussion for the Flexibility aspect of MDS 9710.
-We are what we repeatedly do. Excellence, then, is not an act, but a habit (Aristotle)
Cisco Prime Network Services Controller 3.2.2 is available for download, follow this link to download the software and documentation. The 3.2.2 release incorporates a number of new features on top of 3.2.1 release. Refer to an earlier blog for the features introduced in 3.2.1 release. Following video provides a brief introduction to Prime Network Services Controller.
Following are some of the key capabilities introduced in Cisco Prime Network Services Controller 3.2.2 release:
Prime Network Services Controller operating in OpenStack environment supporting Edge Router and Load Balancer network services
License automation for CSR 1000V, Citrix NetScaler VPX and Citrix NetScaler 1000V
Automatic instantiation of Compute Firewall, Edge Router and Load Balancer network services with Dynamic Fabric Automation
The Cisco Prime Network Services Controller 3.2.1 supports Microsoft Hyper-V and VMware vSphere environments. Prime Network Services Controller 3.2.2 extends multi-hypervisor capabilities by introducing the support for OpenStack (KVM) in addition to Microsoft Hyper-V and VMware vSphere.
Note: This is the first of a three-part series on Next Generation Data Center Design with MDS 9700; learn how customers can deploy scalable SAN networks that allow them to Scale Up or Scale Out in a non disruptive way. Part 2 | Part 3 ]
Data centers are undergoing a major transition to meet higher performance, scalability, and resiliency requirements with fewer resources, smaller footprint, and simplified designs. These rigorous requirements coupled with major data center trends, such as virtualization, data center consolidation and data growth, are putting a tremendous amount of strain on the existing infrastructure and adding complexity. MDS 9710 is designed to surpass these requirements without a forklift upgrade for the decade ahead.
MDS 9700 provides unprecedented
Performance – 24 Tbps Switching capacity
Reliability – Redundancy for every critical component in the chassis including Fabric Card
Flexibility – Speed, Protocol, DC Architecture
In addition to these unique capabilities MDS 9710 provides the rich feature set and investment protection to customers.
In this series of blogs I plan to focus on design requirements of the next generation DC with MDS 9710. We will review one aspect of the DC design requirements in each. Let us look at performance today. A lot of customers how MDS 9710 delivers highest performance today. The performance that application delivers depend
Extensive Message Protocol (XMPP) is an open standard protocol based on XML (Extensible Markup Language). XMPP is designed to transport instant messages (IM) between entities and to detect online presence. It supports authentication of IM application and secure transport of messages over SSL/TLS. In XMPP entities can be bots, physical users, servers, devices or components. It’s really a powerful tool that has great potential for system administrators to add to their toolbox because:
XMPP is powerful
XMPP with Python is only 12 lines of code – trust me, it’s easy!
XMPP only requires a single query for multiple nodes
Status message can be used to track host presence
The Power of XMPP
For those of you that are not familiar with XMPP, it not only supports one-to-one messaging between entities but it also supports multi-party messaging (which enables an entity to join a chat room for the exchange of messages with several participants). The messages can be text messages embedded in XML format but XML can also be used to send control messages between entities as we will see with the presence stanza in a bit.
XMPP is widely used; Google uses it (for its Hangout application – formerly google chat) and so does Yahoo and MSN. At Cisco, we use Cisco Jabber extensively to communicate internally. The XMPP client function is now integrated in the Cisco Nexus 5000 series with the release 5.2(1)N1(7) and the Nexus 6000 series with the release of 7.0(0)N1(1). XMPP is an integral part of the single console access for Dynamic Fabric Automation (DFA) which is a powerful framework described in my previous blog.
The new Data Center Network Manager (DCNM) 7.0(1) is delivered as an OVA file that can be deployed quickly on an existing VMware-enabled server. Although DCNM comes with a lot of features that simplify the deployment of the Data Center fabric, we can pick and choose any service we want to use independently – which is great since DCNM comes with Cisco Jabber XCP and is license free. If you already have a XMPP service installed (like Openfire or ejabberd), it will not be a problem because everything discussed here is valid on any standard XMPP implementation.
On NX-OS devices, the XMPP feature is activated by configuring ‘feature fabric access’ and is part of the Enhanced L2 license (ENHANCED_LAYER2_PKG). Once activated, the switch becomes a XMPP client that needs to be registered on the server. In order to register it, XMPP requires the use of fully qualified domain names (FQDNs) to identify the domain server. If the switch does not have access to a DNS service, I recommend that you use the switch management network for messaging and a static host–to–IP address mapping in the switch configuration.
The switch will use its hostname to login to the XMPP service. If your XMPP server does not support auto-registration, you will need to register the switch and the rooms in the XMPP database beforehand. The DCNM OVA requires users and groups to be created via the CLI, and example of this user and group creation is:
[root@dcnm-ova ~]# appmgr add_user xmpp -u leaf0 -p cisco123
[root@dcnm-ova ~]# appmgr add_user xmpp -u leaf1 -p cisco123
User added. Read More »
We at Cisco are really excited to introduce the new Cisco Nexus 5600 platform! It is the third generation of industry’s leading Data Center Server-Access Nexus 5000 series of switches. Cisco Nexus 5600 is the successor of industry’s most widely adopted Cisco Nexus 5500 series switches (with over 20,000 customers and 25 million ports shipped) that maintain all the existing Nexus 5500 features such as LAN/SAN convergence, Fabric Extenders (FEX) and Fabric Path.
The new Nexus 5600 was unveiled at CiscoLive Milan in January 2014 with quite a bit of interest
Nuts and Bolts
We are introducing 2 models under the 5600 platform:
Cisco Nexus 5672UP – A 1 RU 10/40G Ethernet switch offering wire-speed performance for up to 48 10G Ethernet ports (16 of which are Unified Ports) and 6 true 40G ports.
Cisco Nexus 56128P – A 2 RU 10/40G Ethernet switch offering wire-speed performance for up to 96 10G Ethernet ports (48 of which are Unified Ports) and 8 true 40G ports..
In addition to the existing features of the Nexus 5000, 5600 platform brings new features such as True 40 GE support, VXLAN bridging and routing and Cisco Dynamic Fabric Automation (DFA) innovation. With a latency of about 1 µsec, the 5600 platform is ideal for applications which need low latency. For those of you who need network programmability, the Nexus 5600 supports Cisco OnePK and Openflow.
Why these new features matter
Extensibility with VXLAN support
The Cisco Nexus 5600 with its VXLAN support is very well suited for multi-tenant cloud deployments. In large scale, multi-tenant cloud deployments, there is a need for VMs to migrate across layer-3 boundaries. Traditional VLANs only support about 4000 VLANs which are insufficient in the deployments of thousands of VMs. With the migration need across layer-3 boundaries complexities of layer-3 routers are introduced. To solve the scalability as well as the migration issues VXLAN was developed. For more details on VXLAN, watch the video: http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/vidoe_fundamentals_vxlan.html
Ease of management with Cisco Dynamic Fabric Automation (DFA)
Our customers are also faced with complex, manual network configurations and have a hard time keeping up with application requirements. To solve these challenges, Cisco has developed an architecture called Dynamic Fabric Automation (DFA) which simplifies management and automation – such as automatic device and fabric configuration, automatic VM deployment, migration and seamless integration of bare-metal and virtualized resources in the data center. The Cisco Nexus 5600 platform with DFA implemented in the hardware as well as software is ideal for the multi-tenant and mixed (physical and virtual) cloud infrastructure.
The difference between a 40G and True 40G port is that you can push an entire 40G flow in the True 40G port and in the normal 40G port, the port really is divided into 4 10G ports via Etherchannels, Thus in the True 40G, you get full 40G bandwidth. The Cisco Nexus 5600 platform switches has True 40G ports, which help in servicing the full 40G flows.
The Big Picture
Cisco has one of the most comprehensive portfolios for the Data Center and Cloud Networking and this Nexus 5600 platform is but one of the pieces of this portfolio. You may want to read this excellent blog, which explains Cisco’s Data Center and Cloud Networking portfolio.