High Performance Computing Networking

Open MPI: Binding to core by default

After years of discussion, the upcoming release of Open MPI 1.7.4 will change how processes are laid out (“mapped”) and bound by default.  Here’s the specifics: If the number of processes is <= 2, processes will be mapped by core If the number of processes is > 2, processes will be mapped by socket Processes […]

10 Years of Open MPI

Today’s the day. Today marks 10 years since the first commit in the original Open MPI CVS source code repository (which was later converted to Subversion): $ svn log -r 1 http://svn.open-mpi.org/svn/ompi ------------------------------------------------------------ r1 | jsquyres | 2003-11-22 11:36:58 -0500 (Sat, 22 Nov 2003) First commit ------------------------------------------------------------

The Network Locality Project (netloc)

Today’s guest post comes from Dr. Joshua Hursey, an Assistant Professor in the Computer Science Department at the University of Wisconsin, La Crosse. For a number of years, developers tuning High Performance Computing (HPC) applications and libraries have been harnessing server topology information to significantly optimize performance on servers with increasingly complex memory hierarchies and […]

At SC’13 next week? Come say hi!

Are you going to be in Denver at SC’13 next week? Good! You need to stop by the Cisco booth (#2535) and say hi to your friendly neighborhood Open MPI developers: Dave Goodell, Reese Faucette, and Jeff Squyres.

Lawrence Berkeley Labs talk: Cisco Userspace NIC (usNIC)

Here’s the slides from my second talk, which is a deep technical dive into both how the usNIC technology works, and how we use that technology in the BTL plugin that we wrote for Open MPI (which is upstream starting with Open MPI v1.7.3).

Lawrence Berkeley Labs talk: (Open) MPI, Parallel Computing, Life, the Universe, and Everything

Many thanks to the crew at LBL for hosting my talks yesterday.  There were many insightful questions and comments throughout both talks. Here’s the slides from my first talk, entitled “(Open) MPI, Parallel Computing, Life, the Universe, and Everything.”  This is a general MPI/Open MPI talk, where I discussed the current state of Open MPI, […]

My new favorite Open MPI mpirun feature: tab completion

Today’s guest author is Nathan Hjelm, a Scientist 2 at Los Alamos National Laboratory. We recently added scripts to support tab completion of mpirun flags and run-time MCA configuration variables to the Open MPI trunk development. The scripts support both bash and zsh and have a number of useful features (depending on the shell). Can’t […]

Speaking at Lawrence Berkeley National Lab next week

Are you in the Northern California Bay Area and want to hear about Open MPI and/or Cisco’s usNIC technology next week? If so, you’re in luck! I’ll be speaking at Lawrence Berkeley Lab (LBL) next Thursday, November 7, 2013, at 2:30pm.  Click through to see the location and directions and whatnot (LBL requests that you […]

Hardware and software queuing

I’ve talked before about how getting high performance in MPI is all about offloading to dedicated hardware.  You want to get software out of the way as soon as possible and let the underlying hardware progress the message passing at max speed. But the funny thing about networking hardware: it tends to have limited resources. […]