December 15, 2010
Dec. 15 -- Steven Brandt, a computational scientist in the LSU Center for Computation & Technology, delivered the first specifications of the "chess benchmark" to be used in the Defense Advanced Research Projects Agency's (DARPA) Ubiquitous High Performance Computing (UHPC) program.
The UHPC Program is DARPA's most recent and aggressive research initiative in high performance computing, to create a revolutionary new generation of computing systems that overcomes the current limitations of power consumption and programming. The UHPC objective is to deliver a single rack capable of 1 Petaflops within a power envelope of 57 Kwatts including cooling.
LSU is participating in the "TA2" project led by the Georgia Tech Research Institute (GTRI) for the development of applications, benchmarks, and metrics in support of DARPA evaluations of four TA1 team system designs.
The chess benchmark serves as an exemplar of graph-based decision problems, designed to stress candidate UHPC machines in atypical and important ways. In particular, an efficient chess program is a dynamic graph-based algorithm that must be able to distribute or redistribute large amounts of parallel work, synchronize and communicate state information across the machine, and selectively abort entire groups of running calculations.
"These capabilities are of use to a number of mission-critical Department of Defense applications related to search or tactical analysis," said Thomas Sterling, LSU CCT and Computer Science professor.
Work on this project is being carried out by the LSU Center for Computation & Technology's ParalleX Group led by Professor Thomas Sterling and includes Dr. Steven Brandt, Chirag Dekate, and Phillip LeBlanc. For more information on the LSU CCT ParalleX group projects, visit http://px.cct.lsu.edu.
Source: LSU Center for Computation & Technology
10/30/2013 | Cray, DDN, Mellanox, NetApp, ScaleMP, Supermicro, Xyratex | Creating data is easy… the challenge is getting it to the right place to make use of it. This paper discusses fresh solutions that can directly increase I/O efficiency, and the applications of these solutions to current, and new technology infrastructures.
10/01/2013 | IBM | A new trend is developing in the HPC space that is also affecting enterprise computing productivity with the arrival of “ultra-dense” hyper-scale servers.
Ken Claffey, SVP and General Manager at Xyratex, presents ClusterStor at the Vendor Showdown at ISC13 in Leipzig, Germany.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?