IBM Demos Cray-Matching Linux Cluster 129
An anonymous reader sent us a link to an InfoWorld story
where you can read about IBM slapping together an
Open Source Supercomputer
capable of matching a Cray on PovRay benchmarks.
It's basically just a cluster of Xeon based Netfinitys.
Smooth.
IBM WAS (Score:1)
And I thought marketing was supposed to be Microsoft's -strong- point...
IBM move Caldera, Debian, SuSE out of the picture (Score:1)
reconsider that. what you say isn't logical. You're saying that the fact that IBM and others only invest in RedHat proves that the other distros don't have merit, but I could line up thousands of slashdotters who'd argue that debian or suse or whatever kicks redhat in the arse. I'm a redhat user, mostly because that's what I have used in the past, and it fits how I need the OS to install/function. I'd imagine that IBM and others are choosing RedHat to be their Linux prodigy child because it's a smart marketing move. From where I stand, redhat is the frontrunner in the corporate world, and companies will just run with that because of redhat's established name.
The other distros have qualities to them that are better for some people than redhat's distro... IBM picking redhat is purely a marketing move and says very little about the quality of other distros compared to rh.
What if? (Score:1)
What if IBM had this package in a box scheme along with
- UDB (AKA DB2)
- Visuage Age suite
- Lotus Notes (Dominos)
- e.commerce
- well the list is very long i guess !
I bet Microsoft have no future in corporate IT Departments!
Cheers
Real-time rendering? (Score:1)
Something like that could make a really cool video game. Of course, in ten years your Playstation will be able to do it.
IBM WAS (Score:1)
Interesting.... (Score:1)
"Hey, Bob! We got this nifty IBM cluster here. What'd you want to do with it?"
"Wait a minute. I'll run out and grab a Linux cd from the bookstore down the street."
Hahahahhahaha. That's fucking great!
hmmm (Score:1)
Not a very big T3E in that test... (Score:1)
Beowulf clusters are nice if you've got a parallel problem that only scales well out to a moderate number of processors (32-64 at most), or if price/performance matters more than raw performance. They get clobbered if the problem is bounded by I/O, communication latency, or per-CPU memory bandwidth.
Frankly, IBM has a lot more to worry about from Beowulf clusters than SGI does. Their supercomputer class machine, the SP, is just a cluster of rackmount RS/6000s with a very high speed internal network, and it has all the same problems as a Beowulf cluster relative to a more tightly coupled parallel system like a T3E or an Origin. Plus, AIX is eeeeeeevil; IRIX is much nice IMHO.
And before anybody asks, yes, I work with both traditional supercomputers (Cray T94, Cray T3E/600-136LC, SGI Origin 2000/24xR10-250, IBM SP-2/8) *and* a Beowulf cluster. We've been doing benchmarks to compare our Beowulf to our big machines; in some cases, the Beowulf wins, and in others, the big machines win. It really depends on the problem. We (i.e. my group at OSC) may be announcing some benchmark pages here in a few weeks.
--Troy
Slow down, people! (Score:1)
why not use SCSI? (Score:1)
Where's NUMA (Score:1)
The question is... (Score:1)
the inter-node communication/bandwidth needs?
In clustering and parallel computation, bandwidth
counts. My guess is that a different application
that requires much more communication between
nodes, the T3E would step on the Netfinitys. 100
megabit ethernet does it for low-communication
jobs, but what about those that require much more
intensive inter-node communication.
Yet another x86 Abomination (Score:1)
Oooh graphics cards are expensive aren't they?
You have not covered the performance aspect. Alphas systems have twice the FPU power of any intel system at the same price, new (per MHz is a little different, but cost-performance is more important than CPI. For clustering, that is very good. Remember that all new Alphas currently have 64 bit PCI slots, like such used for gigabit or four port duplex 100bTX cards, reducing memory system bottleneck and increasing raw comm throughtput for parallel cluster/node computing like this. Communication is the key for parallel.
JRDM
36 PII-400 Xeons match 48 Alpha-450's... --- OLD (Score:1)
If Microway makes a new cluster, its memory performance/bandwidth will probably multiply by 10 given the new chipset.
No Subject Given (Score:1)
The cluster may be able to outrun big blue... but if you install Linux on Big Blue, that should speed it up quite a bit i believe.
Then again, I seem to recall something about Linux being not quite THAT scalable, although I could be wrong..
Can Linux handle thousands of processors?
Big Blue / Power PC's (Score:1)
It's actually information on another supercomputer, but it compares the system to the IBM Blue Pacific.
Under the processors section, it says that the IBM system uses "5,856 Power PC 604 processors"
hmmm (Score:1)
Any comments on this? Obviously a dual Pentium II is pretty damn good at this too, being only 1/3 the speed of a $5.5 million supercomputer. Anyone have any idea why adding so many processors to the Linux cluster would improve results so little?
hmmm... I'll hazard a guess on this one... (Score:1)
In this case, 18 times the processors should give 18 times the speed -- unless the test really isn't processor bound. I'm not sure what it would be bound by, however... lousy implementation? I/O? Network?
If the test isn't really processor bound then the comparison to the cray is meaningless, because there's something wrong with the way the software is coded to work on parallel machines, I'd think.
I disagree that the $12k cost means it was a cluster of Pentium II's. I've bought a couple Pentium II systems in that range, its easy to get up there when you add a lot of RAM, lot of harddrive space, etc. Using name parts jacks the price up a lot. (ie, VA's selling systems with Intel boards rather than supermicro or some other lower-cost company...)
On a side note, I remember reading a year or two ago that someone was working on a networking layer that allowed IP and other protocols to be routed between cluster machines over a 40MB/sec SCSI bus. Anyone know if that ever got to completion? A four-fold jump in network speed would make quite a difference to I/O bound applications. (And SCSI cards are a lot cheaper than Gigabit ethernet or other real high-speed networking technologies...)
Poor SGI and others (Score:1)
SGI's NUMA architecture means data can be pumped *much* more quickly between nodes, 100-1000 times as fast. Network-based Linux clustering is useful only for calculations that are fairly self-contained and don't need a lot of data to process.
What I think would be more interesting, given SGI's leanings towards supporting Linux on their MIPS and Intel platforms, is if they eventually tweak the multiprocessing in the kernel to support NUMA style multiprocessing and I can throw Linux on an Origin server. Or maybe better yet a NUMA-architecture Intel machine (i'm not really up on floating point speed comparisions between newer MIPS and newer Intel chips). Since they've dropped real PC-compatibility on their new Intel machines, that sort of a shift is a lot less painful than the initial dropping of support for DOS/16 bit apps.
So SGI doesn't get hurt by Linux. Linux *can't* really compete with a Cray at any real-world tasks (not yet...). And SGI is in a *real* good spot to be the ones selling the Linux-compatible hardware that actually could. In which case, why would they care? Their profit may be lower on a $500k Cray-comparible NUMA linux system than on the Cray, but I'd bet they'd sell enough more of them to make up the difference.
Time will tell.
Second Place Results - NOT 1 machine (Score:1)
No, but 2.2 is an update to RH5.2 (Score:1)
--Lenny
not really... (Score:1)
Things like Crays are expensive mainly because they have very special, very fast hardware for this purpose. It may be extraneous hardware for something straight ahead like a render farm, but there are many cases where such massive bandwidth is very necessary. Thus, for most applications, replacing a Cray with a Beowulf cluster just isn't an acceptable solution.
Beowulf clustering has been proven to be a cost-effective non-real-time rendering system, however.
--Lenny
Could it be off the shelf? Close enough (Score:1)
Impressive failover (Score:1)
I'm also pleased with IBM's recent decision to release their Websphere Application Server on Linux - although the person in marketing who thought up that name should be demoted. The acronym is "IBM WAS." Both passive and past-tense. Sheesh!
hmmm / Check single benchmarks (Score:1)
Yet another x86 Abomination (Score:1)
People did this with BeOS when CodeWarrior/x86 was spitting out terrible machine code. Say what you must about M$, but their x86 codegen kicks serious butt.
For alphas, how about building the code on DEC Unix with the cool compilers?
IBM move Caldera, Debian, SuSE out of the picture (Score:1)
Get a clue.
.
Could it be off the shelf? Close enough (Score:1)
100bT and a switched hub, DEC Tulips bought on sale, donated hardware, etc... we paid approx $400 for our 8 node cluster.
i mean, who the heck wants to drop in a 2.2.x kernel rpm? c'mon! sources have been out for a while....
Where to find software? (Score:1)
Deep Blue (Score:1)
Impressive failover (Score:1)
A few problems here... (Score:1)
Second, the article says they used Xeons. I don't know about what prices IBM gets, but the cheapest I could find a Xeon was about $700. At this price, just the Xeons would cost half a million. The $150k price tag on this setup is just unbelievable, unless either (a) I really misunderstood how many processors they have, and/or (b) $150k was just what they had to buy in addition to what they already had lying around.
A few problems here... (Score:1)
A few problems here... (Score:1)
aint no 36 headed xeons arounds.
Partially correct (Score:1)
This shouldn't be true for much longer. Compaq released their math library for the Alpha last week (see here [digital.com] for details), and, acording to posts to comp.lang.fortan they will be releasing their Fortran compiler as well (as a commercial product, not for free). This should make Alphas much more appealing for cluster use.
-jason
Entry level supercomputing (Score:1)
case) bottleneck. Thus these computers need super fast, specialized networking connections.
Umm, I think you're confused. Vector computers, such as the older Cray machines, use special vector processors that can operate very efficiently on long vectors of data, applying the same operations (hence the name). Things get a little more confusing with later machines which are actually parallel-vector computers (i.e. they had multiple vector processors that worked in parallel).
It is generally accepted that parallel computers, whether they are "big iron" type machines, such as the T3E, Origin 2000 or SP2, or clusters of workstations and PCs, are the way to go for high-performance computing. Of course, some people would point to the latest vector machines from Japan to contradict this...
You are right that the true measure of performance is based on applications, and there are applications suited to each of these architectures. We have found that for our problems (large-scale reservoir modelling) clusters of commodity PCs perform quite well in comparision to an SP or T3E, even with 100 Mbps networking, but there certainly are other applications with more fine-grained communication requirements for which even a T3E or O2k is barely sufficient.
Yet another x86 Abomination (Score:1)
Check www.microway.com [microway.com] for an Alpha cluster priced at 2500$ per node and $4,500 for the master console. This means that for the $150000 used by IBM one could assemble a 50+ node alpha cluster instead of 17 PCs...
God, when will people ever learn that x86 just does not worth it...
Partially correct (Score:1)
Still, I would bet that you can actually get a very decent special deal if you purchase all the stuff together.
Overall it depends on the application. GCC is not that bad on Alpha integer. The performance loss is mostly in floating point and math libraries.
Anyway, I will bet for the Alpha for most of the cases
And what about the 1G network to support this (Score:1)
So getting a bunch of sloppy boxen is not an idea. There has to a compromise between box speed, box quantity and price of network equipment.
IBM SP2 (Score:1)
TA
IBM SP2 (Score:1)
No, I'm not impressed by the SP switch. And besides, it's a terrible beast to work with.
TA
IBM SP2 (Score:1)
It's interesting that you have measured 100MB on the latest equipment, the application should in theory be running on new hardware when it gets operational. It's very useful to have an idea of how the switch will perform, so thanks a lot for that info.
TA
Amusing that IBM Didn't Bench Against RS/6000 (Score:1)
InitZero
hmmm (Score:1)
A proud Debian user (Score:1)
Could it be off the shelf? (Score:1)
http://www.haveland.com/cgi-bin/getpovb.pl?sear
D
Amusing that IBM Didn't Bench Against RS/6000 (Score:1)
This demo was done at a Linux show. Linux on RS/6000 is still a work in progress, so it is not surprising that they aren't ready to show that. Doing a demo with AIX at that show would have been a political faux pas.
Could it be off the shelf? (Score:1)
bnf
Impressive failover (Score:1)
Yawn (Score:1)
Neat but not really. A cluster of PCs connected over fast ethernet is not as flexible as a Cray. On the other hand, a Cray is a waste of money for rendering.
hmmm... I'll hazard a guess on this one... (Score:1)
There has for some time been a rule of thumb that adding a second (or third, or 17th) processor to a problem doesn't get you double the performance, because there is overhead deciding "which processor is going to do what."
What this means is that at some point, adding parallel processors to a problem ceases to be cost effective answer.
See my other note on the main thread (this one got submitted first, so be a little patient!!) as to what I think is of greater long-term significance to the Linux world.
Leading us to the future. (Score:1)
This demonstration all about mind share -- something that Microsoft doesn't want Linux to achieve. Let me explain.
For many years, Cray build absolutely the highest performing mainframe number crunching computer in the world -- and every computer scientist knew it. We used to joke about having our own desktop Crays -- if someone would just lend us (in this case) $5.5 million dollars per workstation. So here's the point that IBM was really trying to get across to the corporate IS people out there -- that Linux is competitive with anything Microsoft can produce. Follow the steps:
Whether or not it could be done with NT-based machines misses the point.
Although I'm not always a fan of Big Blue, in this case we should all thank them for a great job in once again proving the power of Linux to the rest of the computing world.
Take that, Microsoft!!
Xenons vs Celerons (Score:1)
hmmm (Score:1)
Interesting.... (Score:1)
The question is... (Score:1)
I'm not sure I agree with that assessment... But I'm just a dumb systems researcher. What do I know about applications?
Fair where it matters... (Score:1)
Is Avalon faster then Cray? (Score:1)
Just in case anyone was in a coma all last year and doesn't know what Avalon is, here's [lanl.gov] the link.
I wonder if the Avalon folks ever tried anything as trivial as Ray Tracing.
Entry level supercomputing (Score:1)
There's nothing special about this news other than the fact that the individual nodes are running linux. Which basically makes this an SP2 minus the superfast network (and the dent in the wallet).
The measuring stick for all computer hardware issues is application. There is supercomputing (vectorcomputing) (like Cray, traditionally) and there is parallel computing (like any old cluster of workstations). The distinction is the type of operation. In vectorcomputing, each node computes a very small part of the big picture, making communication time a very big (or small in this case) bottleneck. Thus these computers need super fast, specialized networking connections. There are many problems/programs which may be parallelized and yet, still have a significant sequential segment, causing the bulk of the processor cycles to be spent on processing, as opposed to waiting for data communication.
The problems described by the later are becoming more and more popular. Vector computing, however, is primarily core scientific applications (physics, math, weather prediction, etc.) which have not seen dramatic computational advances in the last decade.
An SP2 is sort of in the middle of the spectrum since on top of having high powered nodes, it has a fast network. A COW running linux catches the bottom end of this spectrum, it's nodes are high powered by its network is slow. With ethernet, fast ethernet, or ATM it could never match the performance of Crays or Connection Machines. But then what do you expect for $2000 a node.
Also
trivial problem (Score:1)
(by the way, I bet I could probably piss further than you)
36 PII-400 Xeons match 48 Alpha-450's... (Score:1)
CJK
Where's NUMA (Score:1)
Sure, the architecture of the system may be old school, but that's not why people set up Beowulf clusters...they buy them for the untouchable price/performance for coarsely-grained problems. End of story.
Don't worry, though...Linux development won't stand still...we'll see changes in the future to allow for more flexible architectures.
CJK