Cisco Blogs
Voice Search is currently unavailable
Powered by Google Web Speech API
We didn't hear that. Try again.
When autocomplete results are available use up and down arrows to review and enter to select
Avatar

Jeff Squyres

The MPI Guy

UCS Platform Software

Dr. Jeff Squyres is Cisco's representative to the MPI Forum standards body and is Cisco's core software developer in the open source Open MPI project. He has worked in the High Performance Computing (HPC) field since his early graduate-student days in the mid-1990's, and is a chapter author of the MPI-2 and MPI-3 standards.

Jeff received both a BS in Computer Engineering and a BA in English Literature from the University of Notre Dame in 1994; he received a MS in Computer Science and Engineering from Notre Dame two years later in 1996. After some active duty tours in the military, Jeff received his Ph.D. in Computer Science and Engineering from Notre Dame in 2004. Jeff then worked as a Post-Doctoral research associate at Indiana University, until he joined Cisco in 2006.

In Cisco, Jeff is part of the VIC group (Virtual Interface Card, Cisco's virtualized server NIC) in the larger UCS server group. He works in designing and writing systems-level software for optimized network IO in HPC and other high-performance types of applications. Jeff also represents Cisco to several open source software communities and the MPI Forum standards body.

Articles

What’s Next for MPI?

1 min read

MPI-3 has been out for over a year and a half.  MPICH supports all of the mandatory MPI-3 behavior and some of its optional semantics.  Open MPI supports all of MPI-3 except the new one-sided semantics.  New functionality is becoming mature in both, and that maturity is trickling down to the implementations that are derived […]

Open MPI 1.7.4 released!

1 min read

It took us longer than we intended, but we finally released Open MPI v1.7.4.   Woo hoo!  (we got nice coverage from El Reg, too) This is a monster release; it represents hundreds (thousands? millions?) of person-hours of work.  Consider this a ginormous “thank you!” to the entire Open MPI community! Special thanks goes to […]

More Network Locality (Netloc) progress

1 min read

We announced the Network locality project at SC’13, and generated a LOT of interest (far more than I even anticipated!).  As a refresher, here’s a link to a a blog entry we wrote about Netloc back in November. There is still much work to be done; we’re actively continuing work in multiple areas:

InsideHPC podcast: MPI collaboration with OpenFabrics

1 min read

In my last blog post, I described a new collaboration between the MPI community and the OpenFabrics verbs community. The collaboration started with the OpenFrameworks group asking the MPI community to list its requirements for a lower layer network API to the OpenFabrics OpenFrameworks working group. In that last blog post, I posted an abbreviated […]

A fun thing happened on the way to the OpenFrameworks discussion today…

1 min read

A few months ago, Sean Hefty from Intel started an effort to design a new low-level network API to replace libibverbs. That is, it’s not libibverbs 2.0 — it’s a new API that aims to both expand the scope of what libibverbs did, and also to address many of its much-criticized shortcomings.  Sean and Paul […]

Process affinity: Hop on the bus, Gus!

7 min read

Today’s blog post is written by Joshua Ladd, Open MPI developer and HPC Algorithms Engineer at Mellanox Technologies. At some point in the process of pondering this blog post I noticed that my subconscious had, much to my annoyance, registered a snippet of the chorus to Paul Simon’s timeless classic “50 Ways to Leave Your […]

MPI_FESTIVUS(3)

2 min read

NAME MPI_Festivus – An MPI function for the rest of us

Call for Workshops: EuroMPI/Asia 2014

2 min read

The 21st European MPI Users’ Group Meeting, EuroMPI/AISA 2014, will be held in Kyoto, Japan, 9th – 12th September, 2014. In addition to the main conference’s technical program, EuroMPI/ASIA 2014 is soliciting proposals for one-day or half-day workshops to be held in conjunction with the main conference.  It is intended that those workshops are aim to discuss on […]

Open MPI: Binding to core by default

2 min read

After years of discussion, the upcoming release of Open MPI 1.7.4 will change how processes are laid out (“mapped”) and bound by default.  Here’s the specifics: If the number of processes is <= 2, processes will be mapped by core If the number of processes is > 2, processes will be mapped by socket Processes […]