March 31, 2011
The HPC conference season started in earnest this Tuesday with the HPCC event, a two-and-a-half day conference that attracts a rather elite lineup of speakers and attendees. HPCC, aka the Newport conference (after its Rhode Island locale), is celebrating its 25th anniversary this year, and as someone pointed out, that makes it a couple of years older than the much better-known (and much larger) SC conference of November.
HPCC is geared for the government supercomputing crowd, but topic coverage is broadly spread across the HPC application spectrum. The first day was no exception; it included everything from climate modeling and computational engineering to exascale computing.
Two of the sessions on Tuesday had to do with the infamous "missing middle" problem in HPC. That termed, coined by the Council on Competitiveness, refers to the group of HPC users between low-end, mostly workstation-bound HPC users, and the kind of high-end HPC typically performed at national labs and some universities. The problem is that transitioning from desktop HPC to server-based HPC is filled with roadblocks, especially for commercial users looking to make the leap to small clusters.
At HPCC, we got two perspectives on the problem -- one from a vendor, the other from a user. The user in this case was Caterpillar, a company that is fairly well along in the adoption of HPC. In that sense, they're no longer missing, but they're definitely in the middle. According to Keven Hoffstetter, Caterpillar's research program manager for their Virtual Product Development group, they use HPC-based virtual tools to drive much of their product development.
Caterpillar is the largest manufacturer of construction and mining equipment, which may make it seem like they're not a "middle" HPC user at all. But keep in mind that the term refers to a capability level, not to company size or R&D budget, which in the case of Caterpillar is significant on both counts (93K employees and $1 billion-plus for R&D). Even though they use a number of virtual tools to model things like engines, cooling systems, and linkage systems, they do so at a relatively low capability level.
For example, Caterpillar does modeling of a differential system using a machine with just 64 cores. That simulation takes 4 weeks. According to Hoffstetter, they would like to increase the performance by at least an order of magnitude so that they can run that model much faster and do hundreds of them at a time. The idea is to go through many iterations as possible in order to optimize the design.
Despite the modest level of use, they've managed to realize significant gains from these tools. They're no longer tied to a 10-year product development cycle, which used to be the accepted timeline at Caterpillar. The virtual modeling speeds up all development phases: concept, design, and testing.
But for Caterpillar, there are still major hurdles to get to greater levels of HPC including getting software to scale to greater level, software licensing costs (which Hoffstetter said are already greater than hardware costs), software usability, multi-physics integration and dealing with increasing hardware complexity (including divergent technologies like CPUs and GPUs).
But Hoffstetter contends the biggest problem there is the cultural change required to move to a complete virtual design paradigm. Basically the fear of doing something differently and the concern that the virtual models do not accurately represent the physical product hinder the acceptance of these tools internally. Besides, the models don't yet simulate everything, so the engineers end up having to build a physical prototype anyway. In such an environment, the adoption of this technology is by degrees, rather than en masse.
From the vendor side, we had Intel's Stephen Wheat talking up the democratization of HPC. Intel, of course, has a big stake in broadening the HPC market as far and as fast as it can go. Wheat said HPC currently represents north of 25 percent of their processor revenue and volume. Obviously, expanding that market means big bucks for the chipmaker.
To Intel, this missing middle appears to be low-hanging fruit. A specific focus lately has been the manufacturing industry, which represents a particularly large group of users who have a quantified need for HPC-level simulation and modeling. Intel estimates of that there are about 280,000 small and mid-sized manufacturers in the US alone, and nearly half would use advanced simulation and modeling. That represents a market nearly as large as the entire global HPC segment today.
To that end, Intel has been a big driver behind the newfound Alliance for High Performance Digital Manufacturing (AHPDM), a group devoted to bringing high performance computing to manufacturers. It is especially targeted to these smaller players without the resources or expertise to swallow HPC whole. The alliance consists of industry players, ISVs, academia, and HPC centers of various stripes. The idea is to leverage the existing HPC expertise and infrastructure in order to connect to dots for the manufacturing companies.
Intel is not the only vendor in this group. Microsoft, HP, Dell, NVIDIA, ANSYS and Cray (to name a few) are also part of this. (Disclaimer: HPCwire's publisher, Tabor Communications, is an Alliance member too.) In any case, I suspect we'll be hearing a lot more about AHPDM in the future as some of the efforts get going.
As Wheat said, the definition of success if fairly simple: "When the middle is no longer missing."
Posted by Michael Feldman - March 31, 2011 @ 9:37 AM, Pacific Daylight Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
10/30/2013 | Cray, DDN, Mellanox, NetApp, ScaleMP, Supermicro, Xyratex | Creating data is easy… the challenge is getting it to the right place to make use of it. This paper discusses fresh solutions that can directly increase I/O efficiency, and the applications of these solutions to current, and new technology infrastructures.
10/01/2013 | IBM | A new trend is developing in the HPC space that is also affecting enterprise computing productivity with the arrival of “ultra-dense” hyper-scale servers.
Ken Claffey, SVP and General Manager at Xyratex, presents ClusterStor at the Vendor Showdown at ISC13 in Leipzig, Germany.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?