HPC

A supercomputer in your browser

Cisco is pleased to announce the “Supercomputer in your browser” (SiYB) project, designed to bring the rich High Performance Computing (HPC) ecosystem to the world’s most popular software: web browsers. The free SiYB software is a web browser plugin that is easily installed on any desktop or laptop computer running Windows, OS X, or Linux. “I’ve […]

Euro MPI 2015 CFP (now featuring 100% more Europe)

Euro MPI 2015 has returned to Europe!  (recall that Euro MPI/Asia 2014 was in beautiful Kyoto, Japan) Euro MPI 2015 will be September 21-24 in Bordeaux, France. That means that it’s that time of year again: get your papers ready for submission!  The CFP was just recently published:

The state of libfabric in Open MPI

Yesterday morning, I gave a presentation at the 2015 OpenFabrics Software Developers’ Workshop.  I discussed the status of libfabric support in Open MPI. Here’s a copy of my slides:

Cisco usNIC libfabric provider presentation

Earlier this morning, I gave a presentation at the 2015 OpenFabrics Software Developers’ Workshop.  I discussed Cisco’s experiences with writing providers for both the Linux Verbs API and the Libfabric API. Here’s a copy of my slides:

MPI-3.1! …not quite yet

The MPI Forum met for our quarterly meeting last week in Portland, Oregon. The main goal of the meeting was to pass the MPI-3.1 standard into law.  MPI-3.1 contains a bunch of errata from MPI-3.0, and a small number of new things.

A Farewell to LAM/MPI

With a little sadness, I note that LAM/MPI was officially retired recently. LAM/MPI’s hosting provider, Indiana University, made the decision not to renew the lam-mpi.org domain any more.  As of a few weeks ago, LAM/MPI’s web site is no more, and its domain is in the process of expiring. LAM/MPI was a highly popular implementation […]

Open MPI: behind the scenes

Working on an MPI implementation isn’t always sexy.  There’s a lot of grubby, grubby work that needs to happen on a continual basis to produce a production-quality MPI implementation that can be used for real-world HPC applications. Sure, we always need to work on optimizing short message latency. Sure, we need to keep driving MPI’s […]

MPI 3.1: coming soon to an implementation near you

The next MPI Forum meeting will be in Portland, OR, USA, in early March. One of the major topics on the agenda will be voting on the MPI 3.1 standard. You might be wondering what’s new in MPI-3.1. I’m glad you asked.

Tree-based launch in Open MPI (part 2)

In my prior blog entry, I described the basics of Open MPI’s tree-based launching system over ssh (yes, there are still some valid / good reasons for using ssh over a native job scheduler / resource manager’s parallel launch mechanisms…). That entry got a little long, so I split the rest of the discussion into […]