[Beowulf] Inside Tsubame - the Nvidia GPU supercomputer
Vincent Diepeveen
diep at xs4all.nl
Fri Dec 12 02:50:51 PST 2008
On Dec 12, 2008, at 8:56 AM, Eugen Leitl wrote:
>
> http://www.goodgearguide.com.au/article/270416/inside_tsubame_-
> _nvidia_gpu_supercomputer?fp=&fpid=&pf=1
>
> Inside Tsubame - the Nvidia GPU supercomputer
>
> Tokyo Tech University's Tsubame supercomputer attained 29th ranking
> in the
> new Top 500, thanks in part to hundreds of Nvidia Tesla graphics
> cards.
>
> Martyn Williams (IDG News Service) 10/12/2008 12:20:00
>
> When you enter the computer room on the second floor of Tokyo
> Institute of
> Technology's computer building, you're not immediately struck by
> the size of
> Japan's second-fastest supercomputer. You can't see the Tsubame
> computer for
> the industrial air conditioning units that are standing in your
> way, but this
> in itself is telling. With more than 30,000 processing cores
> buzzing away,
> the machine consumes a megawatt of power and needs to be kept cool.
>
1000000 watt / 77480 gflop = 12.9 watt per gflop.
If you run double precision codes on this box it is a big energy
waster IMHO.
(of course it's very well equipped for all kind of crypto codes using
that google library).
Vincent
> Tsubame was ranked 29th-fastest supercomputer in the world in the
> latest Top
> 500 ranking with a speed of 77.48T Flops (floating point operations
> per
> second) on the industry-standard Linpack benchmark.
>
> While its position is relatively good, that's not what makes it so
> special.
> The interesting thing about Tsubame is that it doesn't rely on the raw
> processing power of CPUs (central processing units) alone to get
> its work
> done. Tsubame includes hundreds of graphics processors of the same
> type used
> in consumer PCs, working alongside CPUs in a mixed environment that
> some say
> is a model for future supercomputers serving disciplines like material
> chemistry.
>
> Graphics processors (GPUs) are very good at quickly performing the
> same
> computation on large amounts of data, so they can make short work
> of some
> problems in areas such as molecular dynamics, physics simulations
> and image
> processing.
>
> "I think in the vast majority of the interesting problems in the
> future, the
> problems that affect humanity where the impact comes from
> nature ... requires
> the ability to manipulate and compute on a very large data set," said
> Jen-Hsun Huang, CEO of Nvidia, who spoke at the university this
> week. Tsubame
> uses 680 of Nvidia's Tesla graphics cards.
>
> Just how much of a difference do the GPUs make? Takayuki Aoki, a
> professor of
> material chemistry at the university, said that simulations that
> used to take
> three months now take 10 hours on Tsubame.
>
> Tsubame itself - once you move past the air-conditioners - is split
> across
> several rooms in two floors of the building and is largely made up of
> rack-mounted Sun x4600 systems. There are 655 of these in all, each
> of which
> has 16 AMD Opteron CPU cores inside it, and Clearspeed CSX600
> accelerator
> boards.
>
> The graphics chips are contained in 170 Nvidia Tesla S1070 rack-
> mount units
> that have been slotted in between the Sun systems. Each of the 1U
> Nvidia
> systems has four GPUs inside, each of which has 240 processing
> cores for a
> total of 960 cores per system.
>
> The Tesla systems were added to Tsubame over the course of about a
> week while
> the computer was operating.
>
> "People thought we were crazy," said Satoshi Matsuoka, director of
> the Global
> Scientific Information and Computing Center at the university.
> "This is a ¥1
> billion (US$11 million) supercomputer consuming a megawatt of
> power, but we
> proved technically that it was possible."
>
> The result is what university staff call version 1.2 of the Tsubame
> supercomputer.
>
> "I think we should have been able to achieve 85 [T Flops], but we
> ran out of
> time so it was 77 [T Flops]," said Matsuoka of the benchmarks
> performed on
> the system. At 85T Flops it would have risen a couple of places in
> the Top
> 500 and been ranked fastest in Japan.
>
> There's always next time: A new Top 500 list is due out in June
> 2009, and
> Tokyo Institute of Technology is also looking further ahead.
>
> "This is not the end of Tsubame, it's just the beginning of GPU
> acceleration
> becoming mainstream," said Matsuoka. "We believe that in the world
> there will
> be supercomputers registering several petaflops in the years to
> come, and we
> would like to follow suit."
>
> Tsubame 2.0, as he dubbed the next upgrade, should be here within
> the next
> two years and will boast a sustained performance of at least a
> petaflop (a
> petaflop is 1,000 teraflops), he said. The basic design for the
> machine is
> still not finalized but it will continue the heterogeneous
> computing base of
> mixing CPUs and GPUs, he said.
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
More information about the Beowulf
mailing list