November 15, 2011
Innovative system also ranks as most powerful NSF system for IOPS
SAN DIEGO, Calif.,Nov. 15 -- Gordon, a unique data-intensive supercomputer using flash-based memory that will enter production in January at the San Diego Supercomputer Center (SDSC) at the University of California, San Diego, made its debut as the 48th fastest supercomputer in the world, according to the latest Top500 list.
SDSC researchers submitted an entry for Gordon using 218 teraflops per second (Tflop/s) and 12,608 cores – about 75 percent of the system. Built by SDSC researchers and Appro, a leading developer of supercomputing solutions, Gordon, the next generation Appro Xtreme-X™ Supercomputer is currently undergoing acceptance testing and when made available to the research community on January 1, 2012, it will have 16,384 cores and achieve over 280 Tflop/s.
The latest Top500 rankings were announced during the SC11 (Supercomputing 2011) conference in Seattle this week. In related acceptance runs, Gordon was tested as to its ability to do I/O (input/output) operations, and achieved an unprecedented 35 million IOPS (input/output operations per second), making it the most powerful supercomputer ever commissioned by the National Science Foundation (NSF) for doing IO.
IOPS is an important measure for data intensive computing since it indicates the ability of a storage system to perform I/O operations on small transfer sizes of randomly organized data – something prevalent in database and data mining applications.
Gordon’s Top500 result is notable in that the Tflop/s ranking was achieved using about half the number of cores compared to most other systems. That’s because Gordon is among the first systems – and the first one commissioned by the NSF – to use Intel® Xeon® processor E5 Family, which perform twice as many operations per clock (eight versus four) of any system currently in use.
“These are truly remarkable results not only for us, but for countless researchers who are moving along with us into an era of data-intensive computing, where supercomputers using innovative new architectures will become a vital tool for them in accelerating new discoveries across a wide range of disciplines,” said Michael Norman, director of SDSC and a co-principal investigator of the Gordon project, the result of a five-year, $20 million NSF award.
“Taken together, these two results mean that Gordon will be the defacto leader in NSF systems in terms of ‘big data’ analysis computing,” said Allan Snavely, SDSC’s associate director and a co-PI for the Gordon system. “Gordon was designed from the outset to be a balanced system between compute and I/O to take on some of the most data-intensive challenges in the real world, not just to score well in terms of FLOPS.”
Gordon is the first high-performance supercomputer to use large amounts of flash-based SSD (solid state drive) memory – think of it as the world’s largest thumb drive. Flash memory is more common in smaller devices such as mobile phones and laptop computers, but unique for supercomputers, which generally use slower spinning disk technology. Gordon uses 1,024 of Intel’s latest high endurance SSD’s, providing the significantly higher performance and endurance required for data intensive applications.
“Gordon, the Appro Xtreme-X Supercomputer, rises to the challenge of the San Diego Supercomputer Center’s computational demands with delivery of a data-intensive computing solution that blends cutting-edge for compute, network, and storage,” said Steve Lyness, vice president of Appro HPC Solutions Group. “The Appro Xtreme-X system based on the Future Intel Xeon processor E5 Family and the new Intel SSD 710 Series will maximize I/O capabilities and flash memory support to analyze large data sets used by SDSC to answer modern science’s critical problems.”
Gordon is capable of handling massive data bases while providing up to 100 times faster speeds when compared to hard drive disk systems for some queries. With almost 5 TB (terabytes) of high-performance flash memory in each I/O node, Gordon will provide a new level of performance and capacity for solving the most challenging data-intensive challenges, such as large graph problems, data mining operations, de novo genome assembly, database applications, and quantum chemistry. Gordon’s unique architecture includes:
The Top500 list of the world’s most powerful supercomputers, now in its 19th year, has been joined by another supercomputer ranking called the Graph500. While developers of the new ranking complements the Top500, industry experts say that the Graph500 ranking seeks to quantify how much work a supercomputer can do based on its ability to analyze very large graph-based datasets that have millions of disparate points, while the Top500 ranking uses LINPACK software to determine sheer speed – or how fast a supercomputer can perform linear algebraic calculations.
Source: University of California, San Diego
10/30/2013 | Cray, DDN, Mellanox, NetApp, ScaleMP, Supermicro, Xyratex | Creating data is easy… the challenge is getting it to the right place to make use of it. This paper discusses fresh solutions that can directly increase I/O efficiency, and the applications of these solutions to current, and new technology infrastructures.
10/01/2013 | IBM | A new trend is developing in the HPC space that is also affecting enterprise computing productivity with the arrival of “ultra-dense” hyper-scale servers.
Ken Claffey, SVP and General Manager at Xyratex, presents ClusterStor at the Vendor Showdown at ISC13 in Leipzig, Germany.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?