Cisco Blogs

Cisco Blog > Data Center

Introducing Cisco Algo Boost and Nexus 3548 – Breaking 200 ns Latency Barrier!

… While Delivering Superior Fabric Visibility!

Today, at the High Performance Computing for Wall Street event, we announced Cisco Algorithm Boost or Algo Boost technology, a groundbreaking networking innovation with numerous patents pending, that offers the highest speed, visibility and monitoring capabilities in the networking industry.  A true game changer delivering competitive advantage to our customers!

Ideal for high performance trading, big data and high performance computing environments, this new technology offers network access performance as low as 190 nanoseconds, more than 60% faster than other full featured Ethernet switches.  When your business success is determined by nanoseconds, this is a huge gain!

The first switch to integrate the Cisco Algo Boost technology is the new Cisco Nexus 3548 full-featured switch which extends Cisco’s leadership in networking by pairing performance and low latency with innovations in visibility, automation, and time synchronization.  And it is tightly integrated with the rich feature set of our Nexus Operating System, a proven operating system used in many of the world’s leading data centers, creating a truly differentiated offering.


Algo boost infographic from Cisco Data Center

So you may ask how we deliver this breakthrough offering that will change the game.  Here is how…

Read More »

Tags: , , , , , , , , , , , , , , , ,

MPI_Ibarrier: Crazy?

Most people’s reactions to hearing about the new MPI-3 non-blocking “barrier” collective think: huh?

Why on earth would you have a non-blocking barrier?  The whole point of a barrier is to synchronize — how does it make sense not to block while waiting?

The key is re-phrasing that previous question: why would you block while waiting?

Read More »

Tags: , ,

Process Affinity in OMPI v1.7 (part 2)

In my last post, I described the Simple mode of Open MPI v1.7’s process affinity system.

The Simple mode is actually quite flexible, and we anticipate that it will meet most users’ needs. However, some users will need more flexibility. That’s what the Expert mode is for.

Before jumping in to the Expert mode, though, let me describe two more features of the revamped v1.7 affinity system.

Read More »

Tags: , , , , ,

Process Affinity in OMPI v1.7 (part 1)

In my last post, I mentioned that we just finished a complete revamp of the Open MPI process affinity system, and provided only a few details as to what we actually did.

I did link to a SVN commit message, but I’ll wager that few readers — if anyone — actually read it.  :-)

Much of what is in the Open MPI v1.6.x series is the same as what Ralph Castain described in a prior blog post.  I’ll describe below what we changed for the v1.7 series.

Read More »

Tags: , , , ,

Taking MPI Process Affinity to the Next Level

Process affinity is a hot topic.  With commodity servers getting more and more complex internally (think: NUMA and NUNA), placing and binding individual MPI processes to specific processor, cache, and memory resources is becoming quite important in terms of delivered application performance.

MPI implementations have long offered options for laying out MPI processes across the resources allocated for the job.  Such options typically included round-robin schemes by core or by server node.  Additionally,  MPI processes can be bound to individual processor cores (and even sockets).

Today caps a long-standing effort between Josh Hursey, Terry Dontje, Ralph Castain, and myself (all developers in the Open MPI community) to revamp the processor affinity system in Open MPI.

The first implementation of the Location Aware Mapping Algorithm (LAMA) for process mapping, binding, and ordering has been committed to the Open MPI SVN trunk.  LAMA provides a whole new level of processor affinity control to the end user.

Read More »

Tags: , , , , , ,