December 19th, 2012
(written by lawrence krubner, however indented passages are often quotes). You can contact lawrence at: email@example.com
In this article Pieter Hintjens and Martin Sustrik examine the difficulties of building concurrent (multithreaded) applications and what this means for enterprise computing. The authors argue that a lack of good tools for software designers means that neither chip vendors nor large businesses will be able to fully benefit from more than 16 cores per CPU, let alone 64 or more. They then examine an ideal solution, and explain how the ØMQ framework for concurrent software design is becoming this ideal solution. Finally they explain ØMQ’s origins, and the team behind it.
Until a few years ago, concurrent programming was synonymous with high-performance computing (HPC) and multithreading was what a word processor did to re-paginate a document while also letting you edit it. Multi-core CPUs were expensive and rare, and limited to higher-end servers. We achieved speed by getting more and more clock cycles out of single cores, which ran hotter and hotter.
Today, multi-core CPUs have become commodity items. While clock speeds are stable at around 2-3GHz, the number of cores per chip is doubling every 18-24 months. Moore’s Law still applies. The spread of multi-core CPUs out from the data centre will continue so that netbooks and portable devices come with 2 or 4 cores and top-end server CPUs come with 64 cores. And this growth will continue, indefinitely.
Several factors drive this evolution. First, the need for CPU producers to compete. Whether or not we can use the power, we prefer to buy more capacity. Second, hitting the clock cycles ceiling, CPU designers have found multi-core to be the next way of scaling their architectures and offering more competitive products. Third, at the low end, the spread of multitasking operating systems like Android mean that those additional cores can immediately translate into performance. And lastly, at the high end, the high cost of a data centre slot for a blade computer (estimated by one investment bank at $50,000 per year) pushes users to demand more cores per blade.
It has been clear for some years that the future is massively multi-core, but the software industry is lagging behind. At the “High Performance on Wall Street 2008″ event, speaker after speaker said the same thing: our software tools are not keeping up with hardware evolution.
The most widely used languages, C and C++, do not offer any support for concurrency. Programmers roll their own by using threading APIs. Languages that do support concurrency, such as Java, Python, .NET, and Ruby, do it in a brute-force manner. Depending on the implementation – there are over a dozen Ruby interpreters, for example – they may offer “green threading” or true multithreading. Neither approach scales, due to reliance on locks. It is left to exotic languages like Erlang to do it right. We’ll see what this means in more detail later.
I could end this article by telling everyone to just switch to Erlang but that’s not a realistic answer. Like many clever tools, it works only for very clever developers. But most developers – including those who more and more need to produce multithreaded applications – are just average.