- 1 Some Notes on the APL2010 Conference in Berlin
Some Notes on the APL2010 Conference in Berlin
Dr. habil. Sven-Bodo Scholz (University of Hertfordshire, UK): Multi-/ Many-Cores: array programming at the heart of the hype!
Sven-Bodo gave an enthusiastic talk about his work with the functional array-programming language "Single Assignment C" - see www.sac-home.org . Unfortunately for we of the Windows community, these are the only offerings there:
The current SAC compiler, called sac2c is developed and runs on UNIX systems. Binary distributions are available for the following platforms:
- Solaris (Sparc,x86)
- Linux (x86 32bit, x86 64bit)
- DEC (alpha),
- Mac OS X (ppc,x86),
- netBSD (under preparation).
He had a somewhat amusing story to tell about one of his students who had a paper rejected by a parallel-programming conference because it was based on using an array notation and this was said by one reviewer to be too unconventional - a more proper paper would be to have written about discovering parallelism within loops. It was considered a kind of "cheating" to avoid this problem by using a notation that renders this consideration moot.
Sven-Bodo's work on SAC targets numerically-intensive programming, particularly those involving multi-dimensional arrays. He claims to get performance comparable to that of hand-optimized Fortran. Some of his exciting recent work targets GPGPUs - see http://www.sac-home.org/publications/RollSchoJoslStav.pdf
Keynote on “Multi-Core and Hybrid Systems – New Computing Trends”
IBM's Dr. Helmut Weber gave a very thoughtful, informed, and forward-looking talk exploring contemporary trends in performance improvements with the current generation of multi-core and hybrid systems. He started by outlining the physical limitations driving the new directions of chip architecture, which are
- higher n-way,
- more multi-threading, and
- memory access optimization.
Once the labels were revealed, we saw this was not the case. However, Dr. Weber was making a point about the natural evolution of performance of any technology over time – that there will eventually be a plateau after a period of exponential growth for a given technology without some fundamental change.
Some of the technology trends he sees coming up are
- increasing CMOS density - from current 22 nm to 15 nm or better,
- more instruction-level parallelism,
- more cores per die and threads per core,
- perhaps leading to many small cores - microservers,
- I/O link bandwidth growing to 100 Gb,
- optical links becoming a commodity.
Other trends include power consumption limiting future DRAM performance growth and the increasing prevalance of solid-state disks. Further in the future, he sees increasing use of "storage class memory" (large non-volatile memory) which could lead to a new memory tier. Some recommendations on how to compute more efficiently from a perspective of power consumption:
- Run many threads more slowly -> constant power, peak performance.
- SIMD power benefit proportional to width but diminishes for very wide.
- Data movement management using reduced cache and simpler core design -> multiple, homogenous processors.
- Add hardwired function specialization to allow bandwidth targeting.
He also had some ideas about how software can improve to be better integrated with hardware and work within a power budget. Part of this improvement will make use of hardware function accelerators, perhaps as part of special purpose appliances. A suggested plan for building these appliances includes adding open-source technology to off-the-shelf hardware and using "builder tools" (e.g. rPath) to tailor software stacks to take advantage of this, along with FPGA and "IP-cores" to improve performance for key functionality. Example appliances include Tivo, the IBM SAN volume controller, and the Google Search Appliance. Some suggestions for taking advantage of these trends:
Programming models need to
- support efficient parallelism
- be at higher level of abstraction
- match the mental model of the programmer.
This leads to the conclusion that the time is ripe for languages to adapt to the multicore future. Dr. Weber suggests that most of these efforts aim at C derivatives which restricts the number of programmers able to exploit parallelism by losing the productivity gains of languages like Java. This prompted IBM to develop X10 which is an adaptation of Java for concurrency.
In addition to the better known mature parallel models of MPI and OpenMP, he mentions OpenCL, which has support for accelerators like GPGPUs. He also showed off some new IBM processor technology: a 5.2 GHz Quad-core processor with 100 new instructions to provide support for CPU intensive Java and C++ applications. It also includes data compression and cryptographic processors on-chip.
Awards to Winners of Dyalog Programming Competition
The talks by two of the winners were refreshing and valuable to see how novices view their first experiences with the language. One thing that puzzled me was that both speakers criticized the lack of type-handling in APL – to me, this has always seemed to be a positive feature of the language. I don’t understand how this kind of restriction provides a benefit that outweighs the limitations it imposes. Can anyone explain? Or better yet, provide a short program illustrating the desired behavior and expound on why this would be helpful?