Interesting analogy I saw today in story from Byte & Switch blogger Frank Berry. He used the analogy of SUVs and their rather spartan initial models (can you say “early Ford Bronco” or “Cash for Clunkers”?) to the recent unified networking (in Cisco-speak “unified fabric“) offerings — the converging of traditional data and storage traffic over a single low latency and high performance 10Gb Ethernet pipe. I’m not sure if today’s unified networking offerings are quite as stripped down as a 1980s Bronco, and I’m sure the energy efficiency is much better. In fact, they may be much closer to an early 2000s Toyota Prius…
Pretty much any major trade show is an exercise in barely controlled chaos and VMworld 2009 is no exception, but all the pieces are falling into place and I am looking forward to a good show next week. Last year, with the announcement of the Cisco Nexus 1000V we focused on what was possible. This year, with both the Nexus 1000V shipping and Cisco UCS shipping, we will be talking more about what is doable. The common theme across our demos, sessions and labs is practical solutions you can take home today. So, in terms of the must see list:
- Register for the Nexus 1000V Self-Paced lab and get hands on experience with basic set up and more advanced things like troubleshooting and security features.
- We have a number of demos in the booth covering UCS, Unified Fabric, Nexus 1000V, Accelerating VMware View, and one that I will be writing more about next week: inter-data center VMotion
- You should also go check out the big honking UCS deployment 16 racks, 512 blades, humming along--I’ll be writing more about that too next week pretty impressive stuff
In addition, we have a strong collection of speaking sessions:
Since I am apparently feeling a bit nostalgic about VMworld and all the frenetic activity we had about this time last year, getting ready for the announcement of the Cisco Nexus 1000V, I caught up with some of the original players that brought our first softswitch to market.
Saravan is a Director of Engineering within the Server Access and Virtualization Business Unit at Cisco and has been leading the Nexus 1000V engineering organization and product strategy from its inception. In addition to Nexus 1000V, Saravan is currently focused on Cisco’s Data Center, Virtualization and Cloud networking solutions.
Michael is a Distinguished Engineer within the Server Access and Virtualization Business Unit at Cisco and was one of the inventors of the original Nexus 1000V concept. His current focus is on Cisco’s efforts related to data center, server virtualization, and cloud computing.
Paul Fazzone is a Senior Manager, Product Management in Cisco’s Server Access and Virtualization Business Unit and one of the original developers of the Nexus 1000V concept. Paul currently manages all of Cisco’s data center access layer software strategy across the Nexus portfolio.
The interview provides some intersting insight into how we moved from customer “asks” to a shipping product:
OS: What was the initial driver behind the Nexus 1000V? SR: We noticed that the edge of the network was moving from a traditional access layer switch to blade switches with the introduction of blade servers and with the introduction of virtualization, it was moving to the virtual switches in the virtualized servers. To provide rich end to end networking solutions, we wanted to develop a presence in the new “edge” of the network and hence started working on Swordfish (later renamed to be Nexus 1000V). MS: We originally ran into this problem when discussing security solutions with customers. With current virtualization solutions, traffic can flow between virtual machines without ever touching the physical network. With the network access layer blending into the server, we realized we wouldn’t be able to offer a true pervasive security solution without having a presence within the hypervisor. PF: We noticed in 2005/2006 that customers were starting to embrace server virtualization in small pockets for non-production applications. The server teams complained about having to get the network team to trunk vlans to the ESX hosts. The network teams complained about lack of visibility and management to perform troubleshooting when the VM couldn’t be accessed. The security teams were raising red flags because the virtual network infrastructure couldn’t be secured like the physical. We saw these 3 items really impacting customer’s ability to virtualize large portions of their server workloads and we thought a more intelligent and feature rich software switch implementation could address the problem.
OS: How long did we spend developing the switch?
So, its coming up to our one year anniversary of the initial Nexus 1000V announcement and its kind cool to see the progression from the announcement to a 2,000+ user beta to customers starting to deploy the product. First off, we have a couple of really helpful new docs posted to the web site. Check out the new Cisco Nexus 1000V Deployment Guide and the second is a new expanded FAQ document. Definitely check them out. Beyond that, we have VMworld coming up next week--if you are going to the show, definitely stop by and say hello. We have a solid selection of sessions around virtualization and network plus the hands-on lab:
The first surprise you will get after you consolidate your branch office servers in the data center is that application performance over the WAN falls off and user productivity suffers. To fix this you deploy WAN Optimization Controllers (WOCs) so your applications go much faster and you recover bandwidth that was lost to redundant traffic. After the WOC’s are installed a strange thing happens to your performance reports. Application response appears to be instantaneous. It’s as if the distance between your data center and your branch office shrunk to nothing. We know that you can’t overcome the speed of light and there will always be some latency on WAN links. So what happened?
What happened is that the WOC distortion effect has come in to play. WOCs proxy TCP at each end and send acknowledgments immediately. As result performance monitoring systems lose visibility to the actual round-trip time. WOC distortion poses a challenge to getting accurate end-to-end performance analysis. To understand how your applications are performing you need to know the average response time between client and server, the Network round trip time to move data from one node to another as well as the server’s response time to a request and the data rate that was achieved.
Read More »