The Portland Group
CSCS Top Right Frontpage

Since 1986 - Covering the Fastest Computers
in the World and the People Who Run Them

Language Flags

Visit additional Tabor Communication Publications

Enterprise Tech
HPCwire Japan

Researchers Invited to Apply for Computational Time on TeraGrid by January 15

Jan. 6 -- Scientists, engineers and other U.S. researchers may apply until January 15, 2011 (12:00 midnight submitter's local time) for the next quarterly review of requests for free allocations of high-performance computer time, advanced user support,  data, and visualization resources that are available through the National Science Foundation (NSF) Office of Cyberinfrastructure's (OCI) TeraGrid. To apply for an allocation of any size, please visit TeraGrid's online submission system

Each quarter, a panel of computational experts known as the TeraGrid Resource Allocations Committee (TRAC) evaluates requests primarily on the appropriateness and technical aspects of using TeraGrid resources. Applications received by the January 15 deadline will be considered at the March 2011 TRAC meeting and awards will be available April 1, 2011 through March 31, 2012.  Multi-year allocations may be requested with an appropriate justification for use over an extended period of time.
TeraGrid allocates more than 1.5 billion processor hours to meritorious requests each year. Resources currently exceed 2.5 petaflops of combined computing capability and approximately 50 petabytes of online and archival data storage from 11 resource provider sites across the nation. At the December 2010 TRAC, 125 requests for computational time and storage were reviewed, and 290 million service units of computational time were awarded.

For the next cycle, researchers can request time on 24 systems, including eight that were featured on the November 2010 Top500 list. Among the diverse resources available through TeraGrid are:

*Kraken at the National Institute for Computational Sciences (NICS)--TeraGrid's largest system is a 1030 teraflops Cray XT5, recently upgraded by 144 teraflops with an additional 1,152 nodes for a peak performance of 1174 teraflops. Users may also request time on NICS' Athena, a 166 teraflops Cray XT4 system, which is co-allocated with Kraken.

*Ranger at the Texas Advanced Computing Center (TACC)--TeraGrid's second largest system is a 579 teraflops Sun Constellation Cluster.

*Nautilus (NICS) and Longhorn (TACC) remote visualization and data analysis systems are the first resources to be awarded under the eXtreme Digital, or XD phase of NSF-supported cyberinfrastructure.

*Wispy, a production cloud environment is available at Purdue, featuring a cluster running KVM and the Nimbus cloud software. Users can submit disk images to run a virtual machine with up to 4 CPUs and 16GB of memory.

*Ember, a new shared memory supercomputer at the National Center for Supercomputing Applications (NCSA), features a peak performance of 16 teraflops, double the performance of its predecessor, Cobalt. 

New resources will be available this cycle. The new Lonestar at TACC will be placed into production February 1, 2011. This 302 teraflops system, coupled with a 1 petabyte Luster file system, will be allocated separate from the Abe/QueenBee/Steele/Lonestar Dell PowerEdge Cluster pool. The San Diego Supercomputer Center (SDSC) will soon offer a 324-node/100 Teraflops system with 38 terabytes of flash memory, called Trestles.  The system will work with, and span the deployments of, SDSC's recently introduced Dash system and its larger Gordon data-intensive system, which will be operational in mid-2011. The Pittsburgh Supercomputing Center (PSC) will introduce Blacklight, an Altix UV system, featuring 512 Nehalem 2.226 gigahertz 8-core processors--4,096 cores total.

For the most current information about TeraGrid resources, visit the resource Tips for writing a successful resource request are included at

TeraGrid and Indiana University Introduce FutureGrid, an Experimental Cloud Test Bed

FutureGrid, an experimental grid test bed, is available (for early adopter use) to researchers who are developing new grid and cloud application frameworks. FutureGrid is a proving ground on which to test applications in a variety of different environments (e.g. CTSS vs. clouds). Access and accounts are managed by Indiana University via an online process. For more information, please visit or

TeraGrid's Transition to XD--The Future is Now!

The NSF has already begun to introduce elements of the next phase of their investment in high-end computing: the eXtreme Digital, or XD, program, which will supplant the TeraGrid program. In 2010, new services were announced, including the Technology Audit Services (TAS) and Technology Insertion Services (TIS). TAS, managed by the Center for Computational Research at the University at Buffalo, State Universities of New York (SUNY), will have systemic tools in place to benchmark user satisfaction and resource utilization across the board. TIS, awarded to NCSA, PSC, NICS, and TACC, will feature a means by which to identify the most promising technologies to meet evolving demands. These services will allow us to be better stewards of XD resources and optimize their use by an increasing number of users and new fields of research. A competition for the management of XD is under review by NSF and is expected to be announced in late spring of 2011.  For more information about XD, visit the NSF/OCI web site

For the most up-to-date information, including an XD transition schedule and answers to frequently asked questions, visit: Contact TeraGrid with specific questions by email:, via the TeraGrid User Portal web form (select the consulting tab), or by calling 1-866-907-2383.

About TeraGrid

TeraGrid, sponsored by the National Science Foundation Office of Cyberinfrastructure, is a partnership of people, resources, and services that enables discovery in U.S. science and engineering. Through coordinated policy, grid software, and high-performance network connections, TeraGrid integrates a distributed set of high-capability computational, data-management and visualization resources to make research more productive. With Science Gateway collaborations and education, outreach, and training programs, TeraGrid also connects and broadens scientific communities.


Source TeraGrid

Most Read Features

Most Read Around the Web

Most Read This Just In

Most Read Blogs

Sponsored Whitepapers

Breaking I/O Bottlenecks

10/30/2013 | Cray, DDN, Mellanox, NetApp, ScaleMP, Supermicro, Xyratex | Creating data is easy… the challenge is getting it to the right place to make use of it. This paper discusses fresh solutions that can directly increase I/O efficiency, and the applications of these solutions to current, and new technology infrastructures.

A New Ultra-Dense Hyper-Scale x86 Server Design

10/01/2013 | IBM | A new trend is developing in the HPC space that is also affecting enterprise computing productivity with the arrival of “ultra-dense” hyper-scale servers.

Sponsored Multimedia

Xyratex, presents ClusterStor at the Vendor Showdown at ISC13

Ken Claffey, SVP and General Manager at Xyratex, presents ClusterStor at the Vendor Showdown at ISC13 in Leipzig, Germany.

HPCwire Live! Atlanta's Big Data Kick Off Week Meets HPC

Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?


Stay informed! Subscribe to HPCwire email Newsletters.

HPCwire Weekly Update
HPC in the Cloud Update
Digital Manufacturing Report
HPCwire Conferences & Events
Job Bank
HPCwire Product Showcases


HPC Job Bank

Featured Events

HPCwire Events