10 Years of Open MPI
Today’s the day. Today marks 10 years since the first commit in the original Open MPI CVS source code repository (which was later converted to Subversion): $ svn log -r 1 http://svn.open-mpi.org/svn/ompi
Today’s the day. Today marks 10 years since the first commit in the original Open MPI CVS source code repository (which was later converted to Subversion): $ svn log -r 1 http://svn.open-mpi.org/svn/ompi
Are you going to be in Denver at SC’13 next week? Good! You need to stop by the Cisco booth (#2535) and say hi to your friendly neighborhood Open MPI developers: Dave Goodell, Reese Faucette, and Jeff Squyres.
Here’s the slides from my second talk, which is a deep technical dive into both how the usNIC technology works, and how we use that technology in the BTL plugin that we wrote for Open MPI (which is upstream starting with Open MPI v1.7.3).
Many thanks to the crew at LBL for hosting my talks yesterday. There were many insightful questions and comments throughout both talks. Here’s the slides from my first talk, entitled “(Open) MPI, Parallel Computing, Life, the Universe
Today’s guest author is Nathan Hjelm, a Scientist 2 at Los Alamos National Laboratory. We recently added scripts to support tab completion of mpirun flags and run-time MCA configuration variables to the Open MPI trunk development. The scripts
Are you in the Northern California Bay Area and want to hear about Open MPI and/or Cisco’s usNIC technology next week? If so, you’re in luck! I’ll be speaking at Lawrence Berkeley Lab (LBL) next Thursday, November 7, 2013, at 2:30pm.
I’ve talked before about how getting high performance in MPI is all about offloading to dedicated hardware. You want to get software out of the way as soon as possible and let the underlying hardware progress the message passing at max speed.
In a previous post, I gave some (very) general requirements for how to setup / install an MPI installation. This is post #2 in the series: now that you’ve got a shiny new computational cluster, and you’ve got one or more MPI
The slides below are from my presentation at EuroMPI’13 about Open MPI’s flexible process affinity interface (in OMPI 1.7.2 and later). I described this system in a prior blog entries (one, two, three), but many people keep asking me