Virginia Tech Supercomputer Up To 12.25 Teraflops 215
gonknet writes "According to CNET news and various other news outlets, the 1150-node Hokie supercomputer rebuilt with new 2.3 GHz Xserves now runs at 12.25 Teraflops. The computer, the fastest computer owned by an academic institution, should still be in the top 5 when the new rankings come out in November."
hrm (Score:5, Funny)
Re:hrm (Score:5, Interesting)
Re:hrm (Score:5, Interesting)
Re:hrm (Score:4, Interesting)
(Although I don't believe brain scanning quite hits the resolution mark required yet.)
Re:hrm (Score:5, Funny)
Re:hrm (Score:3, Informative)
If you read the article (I know, I know) you'll find that the peak performance of this Cray system is 144 teraflops with 30,000 processors.
Re:hrm (Score:3, Insightful)
So the intelligence level of this thing would prob. double in accordance to Moore's law, and in a year outclass it's master two fold. In about another year it will be four times as intelligent as any human being. And, of course, it doesn't stop there....
The implications that this would have on society would be very interesting. Would we believe everything it told us, or claimed
Re:hrm (Score:2, Informative)
http://singinst.org/index.html [singinst.org]
Re:hrm (Score:3, Funny)
Of course not. How are the philosophers going to get booked on talk shows?
Re:hrm (Score:3, Funny)
As a professional philosopher, let me ask what talk shows these are that you are watching. Also, what are the phone numbers of their producers?
Re:hrm (Score:4, Interesting)
Hearing it referenced in terms of AI helps, but is that the only purpose for a research facility to build one of these mammoths? Are there practical applications for the business world (other then the readily available (read commercial) clustered data warehousing)?
I'm not trolling, just curious.
Simulations (Score:5, Informative)
You can get a few hints by looking just at their names.
The number one "Earth Simulator Centre" [top500.org] is fairly self-explanatory, going to their website [jamstec.go.jp] show they create a variety of models for things such as weather, tectonic plate movement, etc.
The number 3 LANL supercomputer [com.com] "is a key part of DOE's plan to simulate nuclear weapons tests in the absence of actual explosions. The more powerful computers are designed to model explosions in three dimensions, a far more complex task than the two-dimensional models used in weapons design years ago." I imagine that most US government simulations would be doing something simmilar.
Re:hrm (Score:2, Informative)
Re:hrm (Score:5, Informative)
--Rob
Re:hrm (Score:2)
Where do we have this amazing capability ?
I mean not only have a rough model ignoring a lot of important influences on weather like water temperatures in the Oceans etc. on a very rough grid , like we have now, but a really accurate weather model.
A recent article I read about NEC's Earth simulator stated that even if this amazing machine was supposed to deliver beneath other things climate calculations wit
Re:hrm (Score:4, Insightful)
Think about it; how is throwing more and more hardware at it going to solve the problem? What we're lacking is the software itself needed to do this, and it's obviously not going to be an easy task to write. I see no reason why an AI as intelligent as a human couldn't be implemented on a slower system, unless "thinks as fast as a human" is among the requirements.
(disclaimer, I've never read the book, these are just my opinions)
Re:hrm (Score:2)
After that, the complicated bit (training the neural network) is much the same as it is with a baby - talk to it, show it simple things, put liquidised food in one end, keep the other end as clean as possible.
The only minor snag with current technology is the limits to how much it can learn and how long it takes to do so.
much the same as a baby (Score:3, Funny)
Researcher: "Go to your machine room! And no Command and Conquer until you do your homework!"
Joshua:"Oh yeah? Would you LIKE TO PLAY A GAME?"
Re:hrm (Score:2)
Theoretically Speaking (Score:2, Funny)
"Everything works in theory, but not pratice."
In theory, anyway.
Re:hrm (Score:5, Interesting)
Actually, it's not quite that simple. As someone whose research is in modeling the hippocampal region CA3 (about 2.5 million neurons in humans, 250k neurons in rats), I can tell you that the connectivity of the system is a very important variable. And there is still much we don't know about the connectivity of the human brain. Furthermore, there are hundreds of different types of neurons in the human brain. Why so many different types if only 2 or 3 would do? Seems evolution took an inefficient path - unless, as is probably the case, the differences in the neuron types are crucial for the human computer to work the way it does. Granted, some differences might be due to speed or energy efficiencies which are not absolutely critical for early stages, but I suspect that many differences have to do with the software (or wetware in this case) that makes us intelligent.
After we've solved that minor problem, I think teaching the system will be relatively trivial. I.e., if we understand the wetware enough to reconstruct it, we most likely understand how its inputs relate to our inputs, etc., and we could teach it much the same as we teach a human child. Of course, we might also figure out a better way to teach it, and in so doing we might even find a better way to teach human children. (Some of our research has recreated certain known best learning strategies, it is probably only a matter of time before simulators disover a better one!)
Re:hrm (Score:2)
Re:hrm (Score:2)
I doubt that it will be entirely successful, but will be happy if I'm proven wrong. Nevertheless, I'm certain that we (as a community) will learn much by studying its shortcomings. I'm really excited about the project!
I hasten to add that when they claim
Re:hrm (Score:2)
What if it only works for 95% of the people who aren't really exceptional mentally.
Or maybe it only works for narrow-minded people
Re:hrm (Score:2)
Re:hrm (Score:5, Interesting)
I'm a robotic software researcher, so this notion really affects me. IMO Software will lag well behind hardware, since it doesn't scale out nearly as well. Representation is of course a huge problem I won't even try to touch... But rest assured lots of people are working on all these things. Btw, It also doesn't help that CPU designs aren't even trying to make AI-style algorithms fast, but we can't blame manufacterers for that util there is demonstrable money to be made.
Re:hrm (Score:3, Interesting)
Re:hrm (Score:4, Interesting)
It's true that designing a CPU to *be* a neural or bayesian network is infeasible, but that doesn't mean we can't add instructions to accelerate their evaluation. The evaluation and update of a neural net, traditional or biologically modeled, is a rather simple algorithmic process, though people who have worked with such simulations (see Ben Hocking's post above, he was my quite capable AI TA) will tell you that they make rather obscene optimizations to make it run reasonably fast. I'm talking about things that might sound familiar to graphics people, like removing all multiplications from a program that's supposed to be doing them more than all other operations combined. It's a particularly good candidate for SIMD instructions. Most large neural nets are sparsely connected, so even if your net is substantially larger than your cache, you can beat that with prefetching. Threshold conditional addition is an example of something that can be done very quickly in hardware, and is much more of a pain to code and optimize in software.
If you prefer RISC to CISC, recall that even the original SPARC had special DSP instructions. Putting the sigmoid function and arctan on silicon is really not all that outrageous.
Re:hrm (Score:2)
There were AI CPUs (Score:2, Informative)
It failed in bankrupcy. My vague understanding was that the designing dedicated LISP processors was hard and slow and with little resources they could not keep up. Essentially the Symbolics computers ran LIPS pretty quickly given the MHZ but SUN and Intel kept moving up the MHZ faster than Symbolics could keep up. In the end there were not speed advantage to a dedicated LISP machine, just a
this is ground-breaking (Score:3, Funny)
I'm a robotic software researcher, so this notion really affects me.
This post deserves its own slashdot article all to itself. Not only has an AI-driven robot posted on slashdot, but apparently someone has designed the robot to research software. So it would make sense that the robot would be reading slashdot. I think the editors should set up an interview with this AI drone known as SnowZero.
Re:hrm (Score:2, Insightful)
Re:hrm (Score:4, Interesting)
Re:hrm (Score:2)
Worse... even if they're analog, they're probably noncomputable [edge.org].
--Rob
Re:hrm (Score:3, Interesting)
Elaborate... (Score:2)
So.... are they partly digital? Entirely analog? Quantum in nature?
Don't tease.
Re:hrm (Score:2, Funny)
Re:hrm (Score:3, Interesting)
Re:hrm (Score:4, Insightful)
We can build a machine that has human intelligence and run it on a 2ghz process. The only issue is that instead of answering a question in a second. Perhaps it will take 1 or 2 hours to deliver an intelligence reply. But it should be able to pass a turing test with time thrown at the window.
Go read what 3D researchers said about graphics in the 70's. I bet they believed a 10ghz was good enough for real life 3D graphics.
What is hindering us is not speed, but our approach to AI research.
Re:hrm (Score:2)
That's because it takes no time at all to answer a question, whether you're a man or a machine. To provide an accurate or informed answer, though, takes time. It
Re:hrm (Score:4, Funny)
Yeah. I've been waiting for years for those dumbasses to make a computer that can outperform my ability to perform 100 trillion double precision floating point operations a second flawlessly.
Re:hrm (Score:4, Insightful)
Re:hrm (Score:2)
Well, that's probably because it requires more than just computing hardware. (Disclaimer: I haven't read the aforementioned book). Right now, programming is nowhere near enough to simulate a human intelligence,
Speed at top (Score:4, Interesting)
Re:Speed at top (Score:5, Insightful)
That's an interesting way of looking at it, but I think so far most of the commentators have failed to pick up what makes this system so incredible. Srinidhi Varadarajan, the designer of the system:
Think about that for a second. The system isn't just in the top 5 (or at least top 10), but it's the cheapest by a factor of at least 2. What's even funnier from a tech standpoint is that the creator doesn't expect it to be beaten until another Apple system is built - which puts a very interesting spin on the old "Apple's more expensive".
Anyway as to in/out of the top 5, Varadarajan reckons there's another 10-20% in optimisations left in the tank...
Data taken from the recent Wired Article [wired.com] on the subject.
Re:Speed at top (Score:3, Informative)
The $5.8M number is how much the computers (and maybe racks) cost, not the whole system. AFAICT, that number appears leaves out US$2-3M worth of InfiniBand hardware that somebody (probably Apple) must've "donated" so it wouldn't show up as part of the purchase price. IB gear costs ~US$2k/node in bulk, on top of the cost of the node itself. It's highly unlikely someone else could build this exact con
don't forget... (Score:2, Interesting)
Not to poo poo their efforts, but the whole system was essentially a 'loss-leader' for future supercomputers projects using the G5's and Xserve....
Re:So he's saying that... (Score:4, Insightful)
what planet are you pricing yoru "similar" x86 hardware on? look, i know mac doesn't have a low end $200 pc. but their high end offerings are not only competitive, but cheap.
Density (Score:5, Interesting)
If that were feasible, you could be looking at toppling Earth Simulator at a fraction of the cost.
Re:Density (Score:3, Insightful)
Also, the Earth Simulator has been around for how many years? 2? 3? Quite frankly, it would be downright embarrassing if it couldn't be toppled at a fraction of its cost by now.
Re:Density (Score:3, Funny)
Of course, the Earth Sumulator wasn't built (just) to run linpack.
I think most super computers weren't built just to run benchmark tests.
Well, at least I hope.
Re:Density (Score:2, Funny)
Re:Density (Score:5, Informative)
First, as they try to increase the speed of the system, the bottlenecks start becoming more of a factor. Interconnects is one big obstacle. While the new System X may use the latest and greatest interconnects between the nodes, they still run at a fraction of the speed that the processors can run.
Also the computing problems that they are trying to solve may not scale either with more processors. For example, clusters like this can be used to predict and simulate weather. To do so, the target area (Europe for example) is divided into small parts called cells. Each node takes a cell and handles the computations of that cell.
In this case adding more processors does not necessarily mean that each cell is processed faster. Getting 4 processors to do one task may hurt performance as they may interfere with each other. More likely the cell is further subdivided into 4 smaller cells and the detail of the information is increased not the speed. So add 4x processors only increases data 4x but it doesn't mean that the data is solved any faster.
Re:Density (Score:4, Informative)
Many processes are indeed easy to divide to parts. Take for example ray-tracing, you can have one processor run each ray if you want, getting huge benefits compared to singleprocessor designs. But many tasks are such that the normal way of calculting them requires you to know the previous result. Trying to break up these tasks is one of the focuses in the reserearch around supercomputing.
"Dick factor" aside (Score:4, Interesting)
Re:"Dick factor" aside (Score:3, Informative)
Assembly - Completed!
System Stablization - In Progress
Benchmarking - In Progress
When up and going the system will probubly do some high end scientific calculations.
Re:"Dick factor" aside (Score:4, Informative)
Their site is out of date then: http://www.wired.com/news/mac/0,2125,65476,00.htm
If there's a Wired article, and a Cnet article, go with the Wired article every time. It's written by people who love tech.
Re:"Dick factor" aside (Score:4, Informative)
Would be interesting to know exactly what stuff do these machines do? Maybe they would even be able to share some code so that people can fiddle around with it optimizing
I don't know about the VT cluster specifically, but here's a couple of typical supercomputer applications that happen to be open source:
ABINIT [abinit.org], a DFT code.
CP2K [berlios.de], another DFT code, focused more on Car-Parinello MD.
Gromacs [gromacs.org], a molecular dynamics program.
(should be fun)
Well, if optimizing 200 000 line Fortran programs parallelized using MPI sounds like fun to you, jump right in!
Note: Above applies to abinit and cp2k only, I don't know anything about gromacs except that it's written in C, not Fortran (though inner loops are in Fortran for speed).
Oh, and then there's MM5 [ucar.edu], a weather prediction code which I think is also open source. I don't know anything about it, though.
Re:"Dick factor" aside (Score:2)
Yeah, talk about Car-Parinello. Great stuff, but I know past versions have sucked up >1GB per node for even small jobs. But I'd love to get my hands on some CP simulations with 400-500 CPUs at once.
Other open-source comp. chemistry packages include MPQC (Massively Parallel Quantum Chemistry): http://www.mpqc.org/ [mpqc.org]
-Geoff
What do they do with it? (Score:2)
So compare it to...... (Score:4, Interesting)
Compare it to this new Cray system [slashdot.org]. Bang for the buck would make the Apple system better.
Crays... (Score:5, Insightful)
Re:Crays... (Score:5, Interesting)
To be fair to the original poster, the Cray system he was referencing is a cluster system. Then again, its a cluster system with very impressive interconnects for which System X just isn't comparable (ie. The Cray system will scale far far better), not to mention the Cray software (UNICOS, CRMS, SFW), and the fact that the Cray system is an "out of the box" solution. So you are right, there is no comparison.
Jedidiah.
Re:So compare it to...... (Score:5, Insightful)
Yup, except the Cray comes with far superior interconnect technology, a better range of hardware and software reliability features built in, software designed (by people who do nothing but supercomputers) specifically for monitoring maintaining and administrating massively parallel systems, and most importantly it all works "out of the box". You buy a cabinet, you plug it in, it goes.
Why do these Apple fans, who justifiably claim that comparing a homebuilt PC to a "take it out of the box and plug it in" Apple system is silly, want to compare a build it yourself supercomputer to one that's just plug and go?
And yes, comparing MacOS X to UNICOS for supercomputers is like comparing Linux to OS X for desktops (in fact that's very flattering to OS X as a cluster OS).
Jedidiah.
Re:So compare it to...... (Score:3, Funny)
Bang for the buck would make the Apple system better.
Sure, but what would you rather say: "I just bought an Apple computer" or "I just bought a Cray computer"?
The list of Supercomputers (Score:5, Informative)
http://www.top500.org/
Obligatory: (Score:2, Funny)
Yes, but... (Score:2)
Peace
Old stuff... (Score:3, Insightful)
Re:Old stuff... (Score:5, Interesting)
and yet... (Score:4, Funny)
Re:and yet... (Score:2, Interesting)
Thank you VT (Score:5, Funny)
Not disease solving, not genetic mapping, not calculating weather patterns.
No, what they're going to do is remaster the Original Star Wars series, right from the laser disc versions!!!!
Imagine, a digitallly remastered bar scene where Han shoots first!!@$!@#!one!@
Re:Thank you VT (Score:2)
No, what they're going to do is remaster the Original Star Wars series, right from the laser disc versions!!!!
So, the question of how much it would cost to get an unadulterated version of Star Wars is finally answered: it would cost $5.8million.
Alright Slashdot, everyone chip in ten bucks!
What is a supercomputer ? (Score:4, Interesting)
The reason is this.. more and more of these 'supercomputer' entries appear to be many machines hooked up together, possibly doing a distributed calculation.
However, would projects such as SETI, GRID, and UD qualify with their many thousands of computers all hooked up and performing a distributed calculation ?
If not, then what about the WETA/Pixar/ILM/Digital Domain/Blur/You-name-it renderfarms ? Any one machine on those renderfarms could be put to use for only a single purpose: to render a movie sequence. Any one machine could be working on a single frame of that sequence. Does that count ?
I seem to think more and more that the answer is 'no', from my perspective. They mostly appear to me as rather simple computers (very often not even the top-of-the-line in their own class), with the only thing going for them that there are many of them.
The definition of supercomputer (thanks Google, and by linkage dictionary.reference.com ) is
And for mainframe
Doesn't the above imply that a supercomputer should really be just a single computer, and not a network or cluster of many computers ?
( The mention of 'terminals' does not mean they're nodes. Terminals are, after all, chiefly CPU-less devices intended for data entry and display only. They are not part of the mainframe's computing capabilities. )
If the above holds true, then what is *really* the world's top 3 of supercomputers ? I.e. which aren't 'simply' a cluster of nodes.
Any mistakes in the above write-up/though process ? Please do point them out
Re:What is a supercomputer ? (Score:2)
Anyway,
I think we can disqualify @HOME style projects, since the individual nodes are not under the control of the manager. Similarly, you can't submit some small batch job to a @HOME system and expect to have results within a short time. Uh, that wasn't a very good description but I hope you understand what I mean.. i.e. that to qualify as a supercomputer all the nodes should be dedicated to the supercomputing stuff, and be
Re:What is a supercomputer ? (Score:2, Interesting)
But if all of the networked/clustered computers are all working on the same task with information flowing between nodes dependant on other nodes processing , doesn't that make them all effectively one large computer?
A renderfarm is similar in many ways to a supercomputer, but I wouldn't think of it as one. Renderfarm nodes generally work on a specific task t
Re:What is a supercomputer ? (Score:2, Informative)
However, would projects such as SETI, GRID, and UD qualify with their many thousands of computers all hooked up and performing a distributed calculation ?
If not, then what about the WETA/Pixar/ILM/Digital Domain/Blur/You-name-it renderfarms ? Any one machine on those renderfarms could be put to use for only a single purpose: to render a movie sequence.
Actually, VT will be #8 this time around (Score:5, Interesting)
http://www.netlib.org/benchmark/performance.pdf [netlib.org] (see page 54)
There have been some new entries, including IBM's BlueGene/L, at 36Tflops, finally displacing Japan's Earth Simulator, and a couple other new entries in the top 5.
Here's just the top 16 as of 10/25/04:
http://das.doit.wisc.edu/misc/top500.jpg [wisc.edu]
No matter what anyone says, Virginia Tech pulled an absolute coup when they appeared on the list at the end of 2003: no one will likely EVER be able to be #3 on the Top 500 list for a mere US$5.2M...even if the original cluster didn't perform much, or any, "real" work, the publicity and recognition that came of it was absolutely more than worth it.
Also interesting is that there is also a non-Apple PowerPC 970 entry in the top 10, using IBM's JS20 blades...
What is the point? (Score:2)
Re:What is the point? (Score:4, Interesting)
Does that answer your question?
Maybe. (Score:2)
Re:Maybe. (Score:2)
It's probably safe to assume that the work in setting up the first cluster decreased the deployment of the second cluster so it wasn't just an exercise is vanity even if it didn't perform a lot of production calculations.
Not exactly (Score:2)
The one critical problem with the initial cluster was that the Power Mac G5 didn't have ECC memory, meaning that any long calculation would really have to be run twice - or at least until the result was the same - to essentially insure a soft error did not go unnoticed (and no, VT's special "error detecting" software didn't account for this).
The Xserve G5, however, does have ECC memory, making the current cluster just as capable as anything els
Hm.. with this much compute power.. (Score:5, Funny)
Great... (Score:2)
Rankings (Score:3, Funny)
Wow, ranked higher than the Virginia Tech football team this year.
Don't let them Virginians hear ya... (Score:2)
IronChefMorimoto
P.S. - Take my word on this as an ex-North Carolinian-- I called an Appalachian State University server farm rather "dairy" and nearly got my ass shot off.
Was Ty Pennington there with a megaphone? (Score:2)
"The fastest Mac supercomputer has gotten faster, thanks to an Xserve makeover." (http://news.com.com/Virginia+Tech+beefs+up+Mac+s u percomputer/2100-1016_3-5426091.html?tag=nefd.top [com.com])
Was that neurotic TLC-to-ABC crossover Ty Pennington (http://abc.go.com/primetime/xtremehome/bios/ty_pe nnington.html [go.com]) onsite to help with the installation upgrades?
Sorry -- this post was in honor of my wife, who tortures me with that damned show every Sunday night.
IronChefMorimoto
Renting Time? (Score:2)
Are there any supercomputer rental outfits out there?
I've heard IBM will truck in a box for you, but that's not really 'net savvy. There was a story about Weta leasing during downtime, but that's a side-line.
Re:Renting Time? (Score:2)
Re:2.3GHz? (Score:5, Interesting)
From the article: What I really want to know is what they do with the old machines. The articles speaks of the cluster being 'upgraded' - are the older G5s replaced, or do they just become part of the new cluster?
Still, I suppose there's one or two unwanted G5s - anyone want to send me a couple?
Re:2.3GHz? (Score:5, Informative)
Re:Article Comparison... (Score:3, Informative)
Re:Article Comparison... (Score:2)
I ask the question having freshly read this article [samag.com] which explains why Opterons are such a huge leap forward compared to Intel's designs. In a nutshell, the bottleneck on a server isn't generally the CPU but interconnects between components. An opteron has dedicated interconnects to other CPUs, its own
Re:Article Comparison... (Score:2)
My question would have to be: Teraflops - is it purely an aggregation of processor power, or does it take into account things like interconnects?
Take a look at top500.org. For every supercomputer you'll see two numbers, Rpeak and Rmax. Rpeak is a purely theoretical estimate, basically number of floating point operations per clock cycle times the clock frequency times the number of cpu:s in the system. Rmax is the result from running the linpack benchmark.
The linpack benchmark is run on the entire syste
Re:Article Comparison... (Score:2)
Thank you. Thank you. All of my friends laughed at me when I said I didn't need a motherboard-- I could save a bundle instead and just use a breadboard.
Re:Wow! (Score:2)
Re:Wow! (Score:3, Funny)
In Soviet Russia, Beowulf cluster jokes are sick of you.
Re:Everyone is getting their rigs ready (Score:2)
Good point...