Sunday, February 24th 2008

Ranger Supercomputer Marks New Era for Petascale Science

Ranger, the most powerful supercomputing system in the world for open science research, today will be dedicated by the National Science Foundation (NSF) and the Texas Advanced Computing Center (TACC) at The University of Texas at Austin. This first-of-its-kind system entered full production on Feb. 4.

Ranger's deployment marks the beginning of the Petascale Era in high-performance computing (HPC) where systems will approach a thousand trillion floating point operations per second and manage a thousand trillion bytes of data.

Ranger is the largest HPC computing resource on the NSF TeraGrid, a nationwide network of academic HPC centers that provides scientists and researchers access to large-scale computing power and resources. Ranger will provide more than 500 million processor hours of computing time to the science community, performing more than 200,000 years of computational work over its four-year lifetime.

"Ranger is the first of the new 'Path to Petascale' systems that NSF provides to open science. It is out in front on the pathway to sustained petascale performance," said Daniel Atkins, director of the NSF's Office of Cyberinfrastructure. "This system and others to come underscore NSF's commitment to world-class, high-performance computing ensuring that the U.S. is a leader in computational science. No longer used by a handful of elite researchers in a few research communities on select problems, advanced computing has become essential to the way science and engineering research and education are accomplished." Ranger is a collaboration among TACC, The University of Texas at Austin's Institute for Computational and Engineering Sciences (ICES), Sun Microsystems, Advanced Micro Devices, Arizona State University and Cornell University. The $59 million award covers the system and four years of operating costs.

TACC Director Jay Boisseau said, "Ranger provides incredible new capabilities for computational researchers across the nation and world. Its computational power, memory and storage capacity greatly exceed anything the open science community has had access to. It takes tremendous expertise to deploy and support research on such a system as well as to use it effectively, but it is an awesome honor and responsibility for us at TACC. Together with our partners, we are excited about fulfilling the promise of Ranger by helping researchers achieve breakthrough science across domains and disciplines -discoveries that will really change the world as well as our understanding of it."

At more than one-half a petaflop of peak performance (504 teraflops), Ranger is up to 50,000 times more powerful than today's PCs, and five times more capable than any open-science computer available to the national science community. Ranger is built on the Sun Constellation System which combines ultra-dense, high-performance compute, networking, storage and software into an integrated general purpose system. Ranger comprises 3,936 compute nodes in a Sun Blade 6048 Modular System with 15,744 Quad-Core AMD Opteron processors, and Sun Fire x4500 servers providing 1.7 petabytes of storage. More information on the Sun Constellation System is available at: www.sun.com.

Ranger offers more than six times the performance of the previous largest system for open science research. The boost in performance offered by Ranger relative to the previously largest open science machine is comparable to reducing the flight time from New York to London to just one hour.

Ranger and other petascale systems to follow will address many of society's most pervasive grand challenges including global climate change, water resource management, new energy sources, natural disasters, new materials and manufacturing processes, tissue and organ engineering, patient-specific medical therapies, and drug design. These issues cannot be addressed or overcome without modeling and simulation.

"Our world is facing great challenges and grappling with big questions across a broad spectrum," said William Powers Jr., president of The University of Texas at Austin. "Advances in computer technology, like the new Ranger supercomputer at UT Austin, will help us manage and understand the vast streams of data flowing into our research. Ranger will attract the nation's leading researchers and accelerate their work to produce faster, more probing analyses of the information they generate. There is no question that Ranger's massive computing power will lead to some of the most significant discoveries of our time. We are pleased that UT Austin and the Texas Advanced Computing Center are playing a leadership role in this endeavor."

The award for Ranger represents the largest NSF grant given to The University of Texas at Austin.

Omar Ghattas, professor of geological sciences and mechanical engineering and director of the Center for Computational Geosciences at ICES, said computational scientists have been making the case for petascale systems for more than a decade.

"NSF has responded with an aggressive campaign to provide U.S. computational scientists with the most powerful open science systems in the world," Ghattas said. "The age of petascale computing - which began when Ranger went live on Feb. 4 - brings with it tremendous opportunities for addressing societal problems and boosting national competitiveness, as well as a tremendous responsibility to society to fulfill the promise of petascale computing. This is undoubtedly the most exciting time in history to be working as a computational scientist."

Ghattas leads a team that will produce the highest resolution models of convection in the Earth's mantle to date, enabling a better understanding of the evolution of tectonic deformation. Their work is emblematic of how larger HPC systems allow for more accurate simulations, finer-grained models, shorter time to results, better statistical analysis, higher-resolution visualization - in other words, bigger, better science.

The additional power, memory and storage of Ranger will enable new research in at least three ways: First, it will allow finer grids for resolution-starved problems, leading to more accurate solution of models. Second, it will permit additional physics to be incorporated into these models, leading to higher-fidelity simulations. Third, it will allow scientists to conduct parameter sweeps for very large-scale models, which are essential for data assimilation and uncertainty quantification, leading to better tools to support decision making.

Ninety percent of Ranger is dedicated to the TeraGrid (www.teragrid.org). Ten percent of Ranger's time is allocated by TACC, with five percent going to research projects at Texas higher education institutions, and five percent going to help industrial partners develop more advanced computational practices.

Any researcher at a U.S. institution can submit a proposal to request an allocation of cycles on the system. The request must describe the research, justify the need for such a powerful system to achieve new scientific discoveries and demonstrate that the proposer's team has the expertise to use the resource effectively. To submit a proposal to request an allocation, please visit the TeraGrid website.
Source: R A N G E R
Add your own comment

27 Comments on Ranger Supercomputer Marks New Era for Petascale Science

#1
spud107
another supercomputer filled with opterons, :rockout:
Posted on Reply
#2
Solaris17
Super Dainty Moderator
that is just lovely. no fold for TPU!!
Posted on Reply
#3
Basard
They really just use this thing to play "Crysis," right?
Posted on Reply
#4
cdawall
where the hell are my stars
BasardThey really just use this thing to play "Crysis," right?
nah they are trying to play dirt they still probably aren't getting a good framerate :roll:
Posted on Reply
#5
Basard
They should stick Quad-Crossfire in each of the nodes. Can't they use GPU's now to compute data too?
Posted on Reply
#6
kwchang007
BasardThey should stick Quad-Crossfire in each of the nodes. Can't they use GPU's now to compute data too?
It's hard to program I believe. Also you have to go through the cpu anyways to "translate" the information from x86 (or 64) to w/e the gpus use. I think it's just more cost effective to stick a bunch of cpus in there.
Posted on Reply
#7
AphexDreamer
What ever happen to TRIPS, the 1,000,000,000,000 calculations per second processor by 2012? I thought thats what this would be about.
Posted on Reply
#8
btarunr
Editor & Senior Moderator
Almost 16K AMD Barcelona units :rockout:
.
.
.
.
.
.
.
They could've done it with 8K Harpertown units :wtf: but anyway, an impressive feat at the 0.5 PFlops monstrosity.
Posted on Reply
#10
jbunch07
can i buy one?
:respect: AMD
Posted on Reply
#11
russianboy
Wait, did AMD charge for those processors?

I'd like to see how much money AMD makes off of one of these, if any.
Posted on Reply
#12
jbunch07
im sure AMD gave them an insane discount for buying that many!
Posted on Reply
#13
Basard
Well, the sure cant sell em... they may as well give em away... hahaha.
Posted on Reply
#14
kwchang007
btarunrAlmost 16K AMD Barcelona units :rockout:
.
.
.
.
.
.
.
They could've done it with 8K Harpertown units :wtf: but anyway, an impressive feat at the 0.5 PFlops monstrosity.
AMD is better than core 2 at server stuff because of HTT and it's better at floating point per clock than core 2.
Posted on Reply
#15
hat
Enthusiast
Yeah but Intel loads thier processors up with cache to bring thier processor bus around the same level of performance.

It's like having one really fast cargo boat (AMD) that can make like 10 trips in the same time 10 normal cargo boats (Intel) can, but in the end the same amount of stuff gets shipped in the same amount of time.
Posted on Reply
#16
ChillyMyst
russianboyWait, did AMD charge for those processors?

I'd like to see how much money AMD makes off of one of these, if any.
BasardWell, the sure cant sell em... they may as well give em away... hahaha.
id..ts, at server related stuff, or some common server tasks per clock amds barc chips are BETTER then intels versions, also note that for this kinda app the bugg has ZERO effect, same as with most servers, it dosnt need a bios patch, because they tend to use linux/unix based os's on them, those os's are compiled with the patch built in and it has zero effect on performance.

stop thinking like overclocking fanbois' and start thinking about the use they are being put to........


its like saying a celeron is better then a semperon because its core2 not k8/k10, when the benchmarks clearly show thats not true(look in the asus buying large order of sempy chips new thred, theres a review link)


as to the main artical, sweet, amd makes alot off of these server/super puter deals, thats why they are making $ off the barc chips dispite the desktop side not having hot sales to enthusists.

people need to stop ASSuming that whats true for an overclocking enthusist is true for every sagment of the market, its just not true, fact is that most of the market dosnt overclock, and isnt effected the same way but things that make us choose one chip over another.
Posted on Reply
#17
russianboy
I didn't say anything about overclocking and whatnot, I was asking about how much AMD was charging for the processors!

Man, don't call someone an idiot until you absolutely know they are one.
Posted on Reply
#18
ChillyMyst
i didnt say idiot i said id..ot ;)

and sorry but what you said implyed they shouldnt be changing anything for them.

sorry if i was mistaken :)
Posted on Reply
#19
ChillyMyst
malwareRanger, the most powerful supercomputing system in the world for open science research, today will be dedicated by the National Science Foundation (NSF) and the Texas Advanced Computing Center (TACC) at The University of Texas at Austin. This first-of-its-kind system entered full production on Feb. 4.
[---]

Ranger’s deployment marks the beginning of the Petascale Era in high-performance computing (HPC) where systems will approach a thousand trillion floating point operations per second and manage a thousand trillion bytes of data.

Ranger is the largest HPC computing resource on the NSF TeraGrid, a nationwide network of academic HPC centers that provides scientists and researchers access to large-scale computing power and resources. Ranger will provide more than 500 million processor hours of computing time to the science community, performing more than 200,000 years of computational work over its four-year lifetime.

"Ranger is the first of the new 'Path to Petascale’ systems that NSF provides to open science. It is out in front on the pathway to sustained petascale performance," said Daniel Atkins, director of the NSF’s Office of Cyberinfrastructure. "This system and others to come underscore NSF's commitment to world-class, high-performance computing ensuring that the U.S. is a leader in computational science. No longer used by a handful of elite researchers in a few research communities on select problems, advanced computing has become essential to the way science and engineering research and education are accomplished." Ranger is a collaboration among TACC, The University of Texas at Austin’s Institute for Computational and Engineering Sciences (ICES), Sun Microsystems, Advanced Micro Devices, Arizona State University and Cornell University. The $59 million award covers the system and four years of operating costs.

TACC Director Jay Boisseau said, “Ranger provides incredible new capabilities for computational researchers across the nation and world. Its computational power, memory and storage capacity greatly exceed anything the open science community has had access to. It takes tremendous expertise to deploy and support research on such a system as well as to use it effectively, but it is an awesome honor and responsibility for us at TACC. Together with our partners, we are excited about fulfilling the promise of Ranger by helping researchers achieve breakthrough science across domains and disciplines —discoveries that will really change the world as well as our understanding of it.”

At more than one-half a petaflop of peak performance (504 teraflops), Ranger is up to 50,000 times more powerful than today’s PCs, and five times more capable than any open-science computer available to the national science community. Ranger is built on the Sun™ Constellation System which combines ultra-dense, high-performance compute, networking, storage and software into an integrated general purpose system. Ranger comprises 3,936 compute nodes in a Sun Blade™ 6048 Modular System with 15,744 Quad-Core AMD Opteron™ processors, and Sun Fire™ x4500 servers providing 1.7 petabytes of storage. More information on the Sun Constellation System is available at: www.sun.com.

Ranger offers more than six times the performance of the previous largest system for open science research. The boost in performance offered by Ranger relative to the previously largest open science machine is comparable to reducing the flight time from New York to London to just one hour.

Ranger and other petascale systems to follow will address many of society's most pervasive grand challenges including global climate change, water resource management, new energy sources, natural disasters, new materials and manufacturing processes, tissue and organ engineering, patient-specific medical therapies, and drug design. These issues cannot be addressed or overcome without modeling and simulation.

“Our world is facing great challenges and grappling with big questions across a broad spectrum,” said William Powers Jr., president of The University of Texas at Austin. “Advances in computer technology, like the new Ranger supercomputer at UT Austin, will help us manage and understand the vast streams of data flowing into our research. Ranger will attract the nation’s leading researchers and accelerate their work to produce faster, more probing analyses of the information they generate. There is no question that Ranger’s massive computing power will lead to some of the most significant discoveries of our time. We are pleased that UT Austin and the Texas Advanced Computing Center are playing a leadership role in this endeavor.”

The award for Ranger represents the largest NSF grant given to The University of Texas at Austin.

Omar Ghattas, professor of geological sciences and mechanical engineering and director of the Center for Computational Geosciences at ICES, said computational scientists have been making the case for petascale systems for more than a decade.

“NSF has responded with an aggressive campaign to provide U.S. computational scientists with the most powerful open science systems in the world,” Ghattas said. “The age of petascale computing — which began when Ranger went live on Feb. 4 — brings with it tremendous opportunities for addressing societal problems and boosting national competitiveness, as well as a tremendous responsibility to society to fulfill the promise of petascale computing. This is undoubtedly the most exciting time in history to be working as a computational scientist.”

Ghattas leads a team that will produce the highest resolution models of convection in the Earth’s mantle to date, enabling a better understanding of the evolution of tectonic deformation. Their work is emblematic of how larger HPC systems allow for more accurate simulations, finer-grained models, shorter time to results, better statistical analysis, higher-resolution visualization — in other words, bigger, better science.

The additional power, memory and storage of Ranger will enable new research in at least three ways: First, it will allow finer grids for resolution-starved problems, leading to more accurate solution of models. Second, it will permit additional physics to be incorporated into these models, leading to higher-fidelity simulations. Third, it will allow scientists to conduct parameter sweeps for very large-scale models, which are essential for data assimilation and uncertainty quantification, leading to better tools to support decision making.

Ninety percent of Ranger is dedicated to the TeraGrid (www.teragrid.org). Ten percent of Ranger’s time is allocated by TACC, with five percent going to research projects at Texas higher education institutions, and five percent going to help industrial partners develop more advanced computational practices.

Any researcher at a U.S. institution can submit a proposal to request an allocation of cycles on the system. The request must describe the research, justify the need for such a powerful system to achieve new scientific discoveries and demonstrate that the proposer's team has the expertise to use the resource effectively. To submit a proposal to request an allocation, please visit the TeraGrid website.

Source: R A N G E R
Posted on Reply
#23
ChillyMyst
the kitten reminds me of zek lol, he should use it as his avatar!!!!
Posted on Reply
Add your own comment
Dec 27th, 2024 10:44 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts