SGI Demos 64-Proc Linux Box 253
foobar104 writes "Details are scarce, but SGI announced this morning that their prototype Itanium 2 system has demonstrated more than 120 GB/s to and from main memory on the STREAM TRIAD benchmark, which is the fourth best result in the world. For comparison, the Cray C90 sustains 105 GB/s, while an even larger Sun Fire 15K clocks a measly 55 GB/s. The interesting part? The system wasn't running IRIX, SGI's proprietary version of UNIX. It was running Linux. More information on STREAM TRIAD, including results from other systems, is available here. The system, incidentally, was an Origin 3800 straight out of manufacturing equipped with Itanium 2 processor modules. SGI will start selling the systems early next year."
What is this good for? (Score:3, Insightful)
Am I wrong about what this benchmark means? Or am I missing something basic?
Re:What is this good for? (Score:2)
Re:What is this good for? (Score:2)
Most of the super-computing problems are simulations, and I would have thought that being able to simulate more of the environemnt (therefore, more data to crunch) would be an advantage.
Simon
Re:What is this good for? (Score:2)
Re:What is this good for? (Score:3, Informative)
Most of these computations are pretty intensive in CPU and memory usage. Network speed and disk speed are less important (although you need lots of storage). I would like to try one of these babies, must be fast.
Re:What is this good for? (Score:2)
That's about the only use I can see for it. I could easily replace every workstation and server in our building with one of these.
I guess colo could be another use, but I'd have to question what you're hosting that needs 64 Itanium processors. More importantly, how well does it handle VM?
Re:What is this good for? (Score:2, Funny)
Wow, that's going to be expensive, and how will they fit those in their cubicles?
Imagine an openMosix cluster of these though.
Re:What is this good for? (Score:2)
Throw out that old Z/390 or AS/400 and replace it with centrally managed GUI terminals. Saves the company $$$ My understanding is the only thing stopping Intel from taking over the mainframe space was a stable OS and memeory bandwidth, looks like that's solved.
But given SGIs history this is probably destined for running simulations and large factoring jobs.
Re:What is this good for? (Score:2, Informative)
Re:What is this good for? (Score:5, Insightful)
Typical super-computing problems are weather prediction, air flow computations and nuclear reaction modelling. Physical models in other words.
Generally, you attack these kinds of problem by partitioning 3-d space into many small cells, and then running relatively simple calculations on every cell. The better the resolution, the better the model.
The thing about three dimensions is, storage space increases with resolution^3... For instance, I believe the weather guys are currently pushing 1kmx1kmx100m resolutions. That means about 3,2e11 cells. If each cell has 1 kB of state, the total memory usage would be about 320 TB.
Super computing problems eat memory like Takeru Kobayashi [cnn.com] eats hot dogs. In many (most?) cases the calculations are simple. Hence, bandwidth is King.
Re:What is this good for? (Score:3, Informative)
is a bit much for today's (affordable) supers.
We use a 22 km x 22 km horizontal grid for
predicting the weather 48 hours ahead over the
North Atlantic + Europe (406 x 324 cells).
We use 31 layers in the vertical (from ~30 meters
thick in the lowest level to ~2 km for the few in
the stratosphere.
This is for a so-called "limited area" model. A
global model such as the model of the European
Centre uses about half the resolution (40 km)
over the entire globe.
Toon Moene.
Re:What is this good for? (Score:4, Insightful)
With no disrespect intended, I think you might be missing something basic.
Any activity that involves moving data into and out of RAM will benefit from the ability to do it faster. That includes things such disparate things as database processing (if you're lucky, you can cache your indices in RAM), media encoding, hell, even compiling. Memory bandwidth is one of the few aspects of computer design that touches just about every application, with the exception of those that are small enough-- or sufficiently well optimized-- to fit into cache.
Re:What is this good for? (Score:5, Informative)
One of the areas this is meaningful is data warehousing. There are three major competitors in the very large data warehousing environment and one wanna be competitor:
One of the biggest drawbacks to Linux adoption in the commercial Enterprise space is its lack of SMP scalability. If the SGI platform works out we will start seeing Linux scaling into an arena that will allow for acceptance in the Enterprise.
Re:What is this good for? (Score:2)
I'd be very happy to find a storage solution that gave us transfer rates that would get us anywhere near utilizing the full CPU capacity even with entry level servers these days for non-computing intensive processes such as mail delivery, serving DNS queries or fault tolerant message queueing... (and preferrably one that doesn't cost ten times more than any potential savings from reducing the number of servers...)
Re:What is this good for? (Score:3, Insightful)
Chances are good that they will build very few full scale machines. Those that are built go towards data-warehousing, research (atmospheric, oceanic and space science, nuclear modeling, etc) and to the government. Factoring large primes is a use, for instance, as it's a problem that can be performed in parallel.
But they will have the ability to say that x, y, and z companies/ gov't agencies have our equipment, it can't be exported (so it must be good), and our lower end machines will suit your job until you need an upgrade - in other words we can be with you for the whole ride and promise application compatability.
-Adam
it's the other way 'round (Score:2, Informative)
In fact, most of the computationally-intensive problems require LOTS of mem-IO.
And there's one more thing: there's a huge difference between the 64-CPU SGI machine, and a Mosix cluster of 64 1-CPU nodes: the SGI has one single memory space contiguous on the same machine. That means you can actually use a very large matrix to process your data, instead of shoving bits of it over the network back and forth.
There are entire classes of problems that will be solved orders of magnitude faster on the SGI server than on a network-distributed Mosix cluster (or any other kind of cluster, Beowulf, etc.). That's the advantage of true SMP systems (all CPUs on the same hardware) as opposed to networked clusters.
stats? (Score:2)
Re:stats? (Score:4, Funny)
The testers tried, but it scrolled by too fast to see anything.
Re:stats? (Score:2)
Re:stats? (Score:2, Informative)
So what is faster than it in the TRIAD? (Score:5, Interesting)
Found the answer here [virginia.edu].
And if you were wondering about a Beowolf cluster of these, the top ten ranking excludes "cluster results".
Re:So what is faster than it in the TRIAD? (Score:3, Informative)
Re:So what is faster than it in the TRIAD? (Score:2)
Re:So what is faster than it in the TRIAD? (Score:3, Informative)
512 processor Origin 3000 quoated as 716 GB/sec.
I have no idea why they are using Itanics for this but its not because they are better processors.
Re:So what is faster than it in the TRIAD? (Score:2)
Windows? No way! (Score:2, Interesting)
Irix can scale up to 1024 CPUs and beyond. Solaris can scale up to 100. Here's Linux, now it's scaling close to 100. How much to you think Windows can scale? 10 CPUs? 20?
SGI's thing was always that it had machines running one single copy of the OS across hundreds (or thousands) of CPUs on the same machine (not in a cluster). You simply cannot do that with Windows, period.
They had some graphics workstations running Windows, but that was on the lowest end of things, and now those systems are not available anymore.
Re: (Score:2)
Re:Windows? No way! (Score:2)
Then SGI realized NT wasn't going to be for big machines, and let that bad dream fade away.
Dude, that's simply not true. SGI never even built a prototype Windows system that ran any version of NT before 4.0. By the time their Windows NT workstations made it out the door, Windows 2000 was very nearly a reality. So close that rather than adding support for certain features, SGI just punted. For example, dual monitor support was never offered on the NT systems until Windows 2000 came out, because NT 4.0 didn't support running two graphics pipes with different drivers.
SGI never, ever, spent any time or effort on Windows NT for anything other than workstations.
But likewise, finding x86 hardware with more processors is probably the largest reason that x86 Linux, Windows, whatever isn't running on bigger machines.
So, let me get this straight. They don't make big x86 boxes, and that's "probably the largest reason" why there are no OSs for big x86 boxes? Brilliant!
Re: (Score:2)
Re:Windows? No way! (Score:2)
If true, this is the first I've heard of this. Can you back this up with some sort of evidence?
Re: (Score:2)
Re:Windows? No way! (Score:2)
In any case, the large systems capable of running Windows didn't appear until after he left and by the time they did, the descision had been made to unload Windows and go with a useful OS (Linux). This is what triggered SGI to get involved heavily in Linux development. Irix was another OS that was considered for the Itanium platform, but there were a variety of reasons why that wasn't picked.
So, that's the short story on why SGI is presently making this system with Linux and why some people have mentioned Windows on large SGI's.
Re:So what is faster than it in the TRIAD? (Score:2, Informative)
That's a peak speed, not a STREAM speed. Some of these machines (like the NEC SX-6) have peak speeds that are *much* higher. STREAM is an attempt at showing how a system performs on a somewhat more realistic workload.
Re:So what is faster than it in the TRIAD? (Score:2)
impressive w/Linux (Score:5, Interesting)
Linux running at 120 GB/s with 64 processors is impressive for an OS that has been criticized as inefficient when running on more than 8.
I would be very interested to know what version of the kernel they are using.
Re:impressive w/Linux (Score:5, Interesting)
What I'm more curious about is what the licensing of all this will be like... are they just doing standard kernel patching, in which case the changes might get rolled back into the vanilla kernel? I'm a little worried that they might be doing it all via binary-only modules, which means that Linux proper gets none of the changes rolled back in...
Re:impressive w/Linux (Score:4, Interesting)
With the major animation companies going to Linux server farms to save cost and get better performance, maybe moving back away from x86 architecture to these large machines may be beneficial cost/productivity wise.
Re:impressive w/Linux (Score:2)
You would be foolish to pay for interprocessor memory bandwidth when clusters are just as fast for that task.
Re:impressive w/Linux (Score:2)
On the other hand, however, many of the modelling techniques used to generate/animate the scenes to be rendered are memory bandwidth intensive as they basically amount to physics simulations in themselves. (Think particle systems, water effects (fluid dynamics), motion of things like hair and fabric, etc.)
Re:impressive w/Linux (Score:5, Informative)
They are working hard to get a number of their changes into the offical kernel, I imagine this is one of them .
Re:impressive w/Linux (Score:2)
I tried really hard to find that info this morning before submitting, but to no avail. But the test was demonstrated at the Intel Developer's Conference, according to the press release, so maybe we could find somebody who knows somebody?
Re:impressive w/Linux (Score:2)
1) A K42 [ibm.com] -like exokernel with some parts of the linux kernel bolted on.
2) Something like Larry McVoys idea of OsLets [bitmover.com], i.e. many kernels running on the system collaborating to provide a single system image to the user.
3) The traditional way, i.e. implementing super-fine-grained locking in the linux kernel. This would of course make linux hard to maintain and slow on "normal" hardware, just like say, solaris.
SGI and Linux (Score:2)
This new system is news, but it's hardly groundbreaking news. Back in '99, SGI spun off MIPS [mips.com] and announced they would do commodity systems -- including supercomputers with commodity processors. At that they had a choice: port IRIX to the Itanium, or teach Linux to scale so they could use it on their supercomputers. It's been no secret that they chose the latter. Or why: it was less expensive, and catered to an established user community.
Note that Itanium/Linux systems are not meant to replace MIPS/Irix systems. Unless they've changed their strategy since I worked there, SGI plans to keep developing Irix systems for another 10 years, at least. Of course, that depends on maintaining loyalty to Irix solutions, and the buzz is that they're having trouble with that.
Historical comparison... (Score:3, Interesting)
They also mention the SV1, which is a "low-end" Cray. I'm curious how the new X1 (nee SV2) does on the STREAM suite.
It's good to see that their "scalable linux" work seems to be doing pretty well! I'm sure it was much easier for them to use the IA-64 port of Linux than to port IRIX...
Re:Historical comparison... (Score:3, Insightful)
The part that I really find interesting is that the top three in the list all outperform this by twice as much, the #1 spot being held by a machine that can do over 500GB/sec.
It's still over 12x faster then the quad Itaniums I used to work with, and probably much cheeper then the NEC machines and the Cray...
Re:Historical comparison... (Score:3, Informative)
If you read the STREAM TRIAD web site linked above, you'll see that SGI didn't compare itself to the C90 exactly; it just ran a benchmark and published the results. Also in that approximate rank are other machines from NEC and Cray and, further down, Sun.
But you're right. Cray was way ahead of their time when it came to things like memory bandwidth. I remember a friend (ex-Crayon) telling me once that access to main memory on the T-90 was faster than access to the on-chip cache on the Pentium III. That sounds implausible, though, so he might have been exaggerating.
I'm curious how the new X1 (nee SV2) does on the STREAM suite.
The last word I got is that X1 is still in the PCB design phase. It's only running as a simulator right now. So it'll be a while before you see those numbers.
(That info is several months old, so I may be wrong.)
Re:Historical comparison... (Score:2)
Re:Historical comparison... (Score:2)
Ordinarily I'm all for spoiling benchmarks by pointing out ways in which they're not applicable, but in this case, you're wrong. This test measures bandwidth into and out of main memory. That's it. It makes no difference whether the processors have vector registers and instructions or not. Noting matters except the factors that contribute to moving data between main memory and the CPUs.
Whoa. Now let's parallellize! (Score:2)
Re:Whoa. Now let's parallellize! (Score:2)
Re:Whoa. Now let's parallellize! (Score:2)
Two words: SOLID STATE (Score:2)
http://slashdot.org/article.pl?sid=99/06/27/13462
href="http://slashdot.org/article.pl?sid=01/02/09
One word (Score:2)
Stock Kernel? (Score:2)
Re:Stock Kernel? (Score:2, Informative)
Re:Stock Kernel? (Score:2, Informative)
Linux isn't really optimized for a lot of processors, but companies like SGI are working to change that, and contributing a lot to the community in the process.
SGI Origin 3800 (Score:2)
Re:SGI Origin 3800 (Score:2)
so he next question.... (Score:2, Funny)
Ow!... ow, ow, ow, OW! stop throwing rock at me!
Ok so it was a bad joke....
Re:so he next question.... (Score:3, Funny)
Well, Windows is notorious for demanding a lot from the hardware. You have to expect it to be a dog on a low-end machine like this one.
NT once ran on MIPS machines, as I recall. I don't have my NT4 disks handy, but I think that I recall that they included binaries for Alpha and Mips. Wouldn't it be nifty to be able to boot NT on that and see it run one cpu, straight into a bluescreen? After all, a computer without MS Windows is like a person without cancer.
Re:so he next question.... (Score:2)
I may be wrong...
Re:so he next question.... (Score:2)
Make the demo Open Source! (Score:2, Insightful)
What is to say that the demo's code isn't buggy and shoddy, holding the power Itanium processors back?
If we realize the vast potential that the Open Source developer community provides then we can tackle such complex tasks as this Itanium performance measurement.
Two things (Score:2, Interesting)
The second thought is: can it be partitioned? This is a rather big machine and goes against the trend I have witnessed to use many smaller machines to accomplish your goal. I'll have to ask some of the guys at Oracle if they've looked at Linux installs of this size, but as far as I know they only make x86 ports right now. So, I wonder what linux apps would someone run on a system this big? (I know. Insert obligatory Quake, Beowolf and porn server reference here.)
Disclaimer: I work for an SGI competitor. But I have personally installed Linux on every piece of harware I can get my hands on. Just to play usually, but still. They just pay my mortgage.
Re:Two things (Score:4, Interesting)
Since this machine is a standard Origin 3000 with McKinley processor modules, I'm going to assume the answer will be yes. You can partition an O3000 down to a single processor brick + base IO brick, so I imagine that SGI will implement the necessary software bits to make that happen on the SN1-IA systems. I know there are both user space bits (mkpart, partmgr) and kernel space bits (the TCP-over-NUMAlink driver).
I personally have only seen partitioning used on HA systems and lab systems. For a fully fault-tolerant N-processor system, you can buy one 2N-processor Origin and partition it down the middle. The two nodes can run in parallel, passing data back and forth over the NUMAlink via TCP/IP, until one goes down. Also, partitioning is great in a lab environment. It's nice to be able to carve up a big multiprocessor system and give each user a 4-processor (or multiple of 4) node.
I wonder what linux apps would someone run on a system this big?
Anything you'd run on an IRIX system of that size, I'd imagine. I believe-- not positive-- that MSC has already released Nastran for Itanium 2 Linux. (Nastran is a computer-aided engineering tool used extensively in the automotive industry, and other manufacturing industries. It's used for things like stress, heat transfer, and vibration analysis.)
And, as long as the Fortran compilers are worth a damn, you can run just about any other scientific, analytical, or technical software, I'd imagine.
What would you do with it? (Score:2)
Me, personally, I do lotsa 3D stuff and would love to see what it'd take to bring that machine to it's knees. However, I get the impression I'm but of a few 3D dudes here. So what would you non-3D dudes wanna do with it?
Re:What would you do with it? (Score:2, Informative)
Everyone on the team used SGIs (I used an Indigo 2, arguably the slowest box in the office) running IRIX. The Origin system sat two floors below us, with the 3D programmer only having the keyboard, mouse and monitor in his office. It made it difficult when we wanted to run a game of Quake, as everyone could easily sneak up on him.
Re:What would you do with it? (Score:2)
Re:What would you do with it? (Score:2)
heh, typo? or what he meant... (Score:2)
"For those applications that need to scale, SGI has just proven that Linux need not be synonymous with clutter."
cluster? or clutter? a good cluster is not cluttered
Released in the nick of time.... (Score:4, Funny)
to meet the system requirements for Doom III.
-R
great.... (Score:2)
I know that someone somewhere is going to use a box like this - but tell me for what real world application will you use it. (serious question - curious. I want to know the reall apps these are used for)
Well... (Score:2)
Models for plasma dynamics and astrophysics are also run on these heavy-duty machines. I'm sure others have had some experience running other things, but I know that the NCEP IBM-SP gets a workout at least 2 times a day running at least four different weather models that have average runtimes around an hour each.
-Jellisky
Re:great.... (Score:2)
In the health insurance industry, which I happen to work in, large SMP or MPP machines are used for data warehousing and fraud and abuse detection. Machines ranging from 16 to 64 CPU's (generally UltraSPARC or IBM Power). When you are dealing with claims records for 5 or 10 million beneficiaries over a 5 or 10 year time span you need a lot of processing power and disk space. The data warehouses are used for trend analysis, fraud investigation and the like. Anyone with a background in statistics knows just how much number crunching we are talking about.
STREAM and SGI past history (Score:4, Informative)
big national id databases (Score:2)
I have a dream... (Score:2)
Perhaps they should update the song [zdnet.co.uk]
Impressive memory crossbar (Score:5, Insightful)
That said, it's an impressive result. And it's done in an unusual way. SGI has a 1.6GB/s channel running through routers [sgi.com] connecting the processors and memory. A computer is made up of multiple rackmount "bricks" connected by cables and routers. The "router" is a 2U rackmount device.
Processors and memory reside in rackmount boxes with 4 CPUs and 8 GB (max) of local memory. These boxes interconnect through a single 1.6GB/s link per box, which, in a big system, goes through several layers of routers. So a memory access to another box is routed through what is essentially a fast LAN. All this is cached, of course.
It's not clear to what extent application programs have to be aware of this. Clearly, if you lay things out in memory badly, with lots of CPUs reading and writing the same memory from all over the memory net, the system will bottleneck. (Everybody reading the same stuff is OK; it's cached. But writes have to propagate back to the home location of the data.)
Since the whole monster crashes all at once, you don't want to build your web server farm this way. It's for applications that really need all that crunch power in one machine.
Re:Impressive memory crossbar (Score:5, Informative)
Speaking as somebody who's done his share of IRIX programming, I'd say "none at all."
In some cases, on Origin 2000 hardware with older versions of IRIX, you could see notable performance differences if you went out of your way to place memory in banks adjacent to the running processors. But the Origin 3000 architecture, with its significant reductions in memory latency, and newer versions of IRIX, with their improved page replication algorithms, have made manual memory placement almost obsolete. Almost.
SGI spent a lot of time and trouble trying to reduce the impact of accessing remote memory. The caching mechanisms and page replication stuff are really well thought-out.
Re:Impressive memory crossbar (Score:2)
arp who-has cpu53 tell cpu4
arp who-has ram1G-2G tell cpu3
Re:Impressive memory crossbar (Score:2)
Even the relatively simple uniprocessor x86 architecture offers OS implementors numerous ways to kill performance (shameless plug: a benchmark example [www.enyo.de]). I would be suprised if SGI achieved this result without some tweaking.
Statics, Benchmarks, and lies... (Score:5, Informative)
Unfortunately, memory transfers are not the world when it comes to large multiprocessor boxes. The overhead comes in when you're trying to synchronize a large number of threads/CPUs to do a large task. For example, an Oracle database.
Sun has proven that it scales up the tree very well with large numbers of processors. But from my understanding, Linux is more efficient with a low processor count, and less and less efficient with more processors.
I question its ability to do anything with a real workload. And I've even more suspicious because they use a benchmark I've never heard of (STREAM TRIAD) to push its superiority on a single-aspect synthetic benchmark.
Good. The machine looks like it has a decent memory bus, and memory modules with a good configuration and speed rating. Now, what can the machine actually do well that makes it a real winner?
Re:Statics, Benchmarks, and lies... (Score:5, Informative)
You know, before you piss in SGI's Cheerios, you might want to do a little reading. The Origin 3000 architecture, on which this prototype system was based, has no memory bus at all. It uses a fabric of switched multi-gigabyte-per-second interconnects to attach CPUs to RAM and to other CPU nodes.
CPU benchmarks (like SPEC) are synthetic and irrelevant, because they fit in cache. Virtually no real application fits in cache, and the sort of applications you run on a machine this big deal with data sets no the order of tens or even hundreds of gigabytes. Memory-to-CPU bandwidth is probably the only real indicator of the ability of the system to handle real-world workloads.
It's also the only thing-- other than the dimensions and the color of the plastics-- that differentiates SGI's big Itanium 2 server from everybody else's big Itanium 2 servers.
Re:Statics, Benchmarks, and lies... (Score:2)
You say that synthetic benchmarks are irrelevant. Then you go on to say that this particular synthetic benchmark is highly relevant. It can't be both. I'd like to see this run a TPC variant, which is closer to real-world than it is synthetic.
The Origin 3000 architecture, on which this prototype system was based, has no memory bus at all. It uses a fabric of switched multi-gigabyte-per-second interconnects to attach CPUs to RAM and to other CPU nodes.
What, do I have to explicitly call out the components and subcomponents? It is a memory bus, for the purpose of this discussion.
Re:Statics, Benchmarks, and lies... (Score:2)
No, I don't. I really can't emphasize this enough: read. I said, "SPEC is synthetic and irrelvant." Big difference.
I'd like to see this run a TPC variant, which is closer to real-world than it is synthetic.
The TPC benchmarks are measurements of database performance. Since SGI was trying to demonstrate the features and capabilities of their hardware, it would have been completely inappropriate for them to use a database benchmark. STREAM TRIAD is great because it measures only one thing: the rate at which data can be moved from memory to the CPU or vice versa. The TPCs measure aggregate systems, including hardware, storage, OS, database software, and so on. They may be relevant if you're looking for a fast database server system, but they're hardly useful for evaluating one hardware architecture over another.
What, do I have to explicitly call out the components and subcomponents? It is a memory bus, for the purpose of this discussion.
The whole point of this discussion is that the SGI system can outperform virtually everything else on STREAM TRIAD because it has no memory bus. Memory busses are bottlenecks, and pumping a lot of data through them is very hard. The SGI system eliminates the bottleneck and thus demonstrates amazing bandwidth. When you miss the whole point of the discussion, I'm going to call you on it.
Re:Statics, Benchmarks, and lies... (Score:2)
No. You said... "CPU benchmarks (like SPEC) are synthetic and irrelevant, because they fit in cache." You also said "Memory-to-CPU bandwidth is probably the only real indicator of the ability of the system to handle real-world workloads." I'd call that even more synthetic and irrelevant than SPEC.
The TPCs measure aggregate systems, including hardware, storage, OS, database software, and so on.
No argument there. But I was saying that it was MORE relevant than SPEC. And extremely more relevant than that STREAM TRIAD test they're pushing.
The whole point of this discussion is that the SGI system can outperform virtually everything else on STREAM TRIAD because it has no memory bus.
Really? I don't recall reading that in the story introduction or SGI's Press Release. Only the link to the STREAM TRIAD itself pointed out that it was talking about memory bandwidth. IIn fact, that is what my original message was trying to point out.
So they've got a machine that gets great ratings on this synthetic benchmark? Who cares. It doesn't mean much if you've bolted a kernel on top of it which isn't mature in a large CPU environment. (And other hardware issues, as you mentioned which the TPCs would bring into play.)
Re:Statics, Benchmarks, and lies... (Score:2)
I don't think I've missed the boat. Okay. Let's take rendering. On a pure economic level, they're going to be hard pressed to sell this configuration vs 64 single processor (perhaps even blade) servers.
On a technical level, let's see how well performance ramps up when you go from 32 to 33 processors. (Hint: you won't be getting a full CPU's worth of extra performance.) Actually, it can get even worse with lock contention and kernel issues so where you can LOSE performance by adding a CPU.
The point I was trying to make is they're touting superiority based on a single benchmark which measures memory bandwidth. Great. The company who produced the box picked a single benchmark which puts the best shine on the hardware/os combination.
Now, what does the box really crank?
C90 is 12 years old! (Score:2)
Only Thurston Howell III is eligible (Score:3, Funny)
Well, Hmph! The rest of us low-life customers wouldn't want it anyways!
Itanium on the rebound? (Score:2)
Intel must be pleased. If SGI could manage to sell one of these that would double the number of Itaniums that Intel has managed to flog.
Not an Origin 3800 (Score:2)
In any case, the poster made it sound like you can just plug Itanium 2's into an Origin 3000 and *bang* you've got a Linux system which is not correct.
Re:Well, that's nice, but what about... (Score:3, Insightful)
Zero. Origin servers don't have graphics cards. Which means, unfortunately, the Slashdot community is going to have to try to wrap its collective head around a more meaningful measurement of potential performance.
Re:Well, that's nice, but what about... (Score:2)
So how man LOC's/sec is that?
Re:Well, that's nice, but what about... (Score:2)
Hmm. (rubs chin). Well, there's that ASCII graphics version of Quake for the console- that should work!
graspee
Re:Well, that's nice, but what about... (Score:2)
NOOOOOOOOOoooooooooo...
Re:Well, that's nice, but what about... (Score:2)
Re:MIPS is to IA64 as Irix is to Linux? (Score:4, Informative)
SGI started working on porting IRIX to the IA-64 architecture back in (I think it was) 1995 or 1996. Not long after, they found that it would be easier and cheaper to get Linux to scale more efficiently and to port some key libraries and services from IRIX than it would be to port all of IRIX over to the new architecture.
It's all about time and money.
there's no point in doing that (Score:5, Interesting)
Networked clusters (Mosix, Beowulf) split the CPU bunch across the network, and the memory is split too. That means there's a huge latency when a CPU wants to access data that happens to be on a different node on the network: the network latency is many times larger than memory latency.
There are problems that simply cannot be solved on networked clusters, precisely because of network latency. While true supercomputers (all CPUs on the same machine) do not have this limitation.
Well, ok, so you can split the matrix across nodes in a Beowulf, but even if you have the same CPU power as the SGI supercomp, you're going to solve the problem several times slower (if not several orders of magnitude slower). Such is the importance of latency.
This is why there's no point in clusterising this kind of computers: you lose their biggest advantage: single OS copy, all memory on the same machine.
Re:Imagine. . . (Score:2)
Re:Imagine. . . (Score:2)