I was cleaning up some files and I found this time lapse that I did when we were building Kelvin, it’s a few years old now. Even by current standards it’s still pretty respectable.
The timelapse
We had to unpack and install all the infiniband cards ourselves, cabled, racked, installed, configured and burnt it in for production usage. The cluster has 100 nodes, each node has 2 sockets, each socket has 6 cores and 24gb of ram.
It was clear at this year’s supercomputing conference that there wasn’t as much excitement as previous years. There wasn’t much surprise as nothing too revolutionary and radical was announced. In the past when Bluegene/L and P arrived after the earth simulator there was an arms race to being number 1 in the top500 list. Even things like GPGPU’s aren’t as cool anymore, everyone is selling effectively the same systems when it comes to clusters.
I shall be going to SC2012 next month, I plan on hitting a few of the storage vendors for possible collaborations and flagging to them that we’re on the look out for storage systems. One of the first observation that the reader will note is “where is that link between HPC and Digital Preservation and Archiving”. It’s probably not obvious to most people, one of the big problems in the area of preservation and archiving is the the amount of data involved and the varied types of data.
A co-worker of mine (Paddy Doyle) had originally hacked at a perl script for reporting balances from SLURM’s accounting system a year or two ago and he had figured out that it might be possible to do some minimalistic ‘configuration’ and scripting to get a system that’s very basic but functional.
It was just one of those things that funding agencies wanted to justify how the system was being used, GOLD was clunky and obtrusive and complicated for what we wanted.
The Top 500 List is out! It’s been no surprise that LLNL and IBM has reclaimed the top spot. There is some pretty cool tech that’s going to come out of these systems which the HPC community should be rubbing their hands over.
Slurm will be moving along development for these big systems as well as Lustre and zfs on linux.
This time the usergroup meeting is in BSC in Spain (this place is cool, I’ve been at BSC for a few conferences and meetings before). I wish I was doing some HPC right now, I’m very tempted to submit an abstract or two on the last two things with slurm that I was involved with.
The php-slurm package needs some love, its somewhat in use at work. I just wish I had more time to fix the memory leaks.
I’ve talked about cports in the past, it’s basically a collection of makefiles which mostly automates the process of downloading, configuring, building and installing applications and libraries for High Performance Computing systems that use environment-modules.
One of the key-features that cports offers is the automated modulefile generation, and the fact that the makefiles acts as documentation to how software is configured, built and installed. It’s currently being used on the clusters at my work place, it has been a boost to the productivity of the systems admin team.