Is IBM's Power4 A Threat To Alpha, Sparc, IA-64? 103
HiyaPower writes: "There is an interesting discussion here about the IBM
Power4 chip. While it is most directly compared with the upcoming Alpha, it also has ramifications for the penetration of the IA-64 and/or Sledgehammer into the server market. Conclusion drawn is that the Alpha, etc., may be in for some very tough sledding. Now if only Apple could be persuaded to use these instead of what the article terms its "embedded controller chips...""
POWER4 runs PPC binaries? (Score:2)
Jobs has got to be pretty pissed at Motorola by now. Rumour has it he's shopping around for new chips. I bet AltiVec is the only thing holding him back. AltiVec is truly amazing for certain tasks...
Is there a technical reason why IBM is avoiding AltiVec? Could AltiVec somehow be responsible for the problems Motorola is having boosting PPC clock speed?
Oh, and can you imagine a Beowulf cluster of these? Sorry, just had to say it. It would be pretty mind-blowing, wouldn't it?
Re:(troll 2) (Score:1)
I can really identify with you, so much.
What about compiler support (Score:2)
Re:POWER4 runs PPC binaries? (Score:1)
Re:Drop Motorolla like a hot potato (Score:1)
another 15 year delay (Score:2)
it took MicroSoft until Windows 95,
I hope they are faster this time around!
i wish/not bloody likely (Score:1)
the likelyhood of these puppies ever landing in a colorful case with a fruit on it is like, nil.
a) does IBM even sell these for other companies use in their boxes? i think no.
b) apple would have to redesign like everything to put one of these in a workstation/server at which point their whole unified motherboard architecture plan goes out the window
c) they require a fan and you know how steve feels about fans
d) even if IBM would sell Apple the chips, can you imagine the price? you think a mac's expensive now?
and lastly, just for the obnoxious value, can you imagine an Appleseed cluster of these?! sweet!
Re:Not terribly far fetched. (Score:2)
An over clocked micro-sequencer could have the same effect without the problems of pipeline refilling at a branch. Overall, I suspect the over clocked sequencer approach might offer comparable performance with a much quicker design cycle - and or - lower design cost.
Re:Step in the right direction (Score:2)
Bzzt. Wrong answer.
Granted, S/390 is not the most popular hardware for ISP's, plenty use S/390. Here's an article [crn.com] about one.
Here's an article [cnet.com] where ebay discsses the possibility of moving to the S/390 platform.
This article [internetwk.com] discusses how some government agencies are web-enabling their mainframes.
I'll grant that traditionally IBM mainframes can be a bear from the usability perspective. However, things are changing quite quickly, especially with the advent of Linux on the S/390.
have a day,
-l
Re:Why Apple Could (Score:1)
Of course, there are plenty of other reasons this will never ever happen which I don't feel like typing yet again but it's not the contract.
Re:Processor design... (Score:1)
Re:Step in the right direction (Score:1)
Well, considering that IBM researcher John Cocke invented RISC... [ibm.com] (scroll down to 1980). Also do a keyword search on 'Ted Codd'. Guess where he worked?
IBM basically suffered from BigCorpBlah syndrome, which afflicted most big corps during the '50s thru the '90s.. So much cool shit got invented and totally ignored, and left to their inventors to splinter off and start dozens of revolutions..
One has to wonder if Micro$oft R&D is sitting on something interesting that is being smooshed because it doesn't fit into some marketdroid's PowerPoint slide... M$ has the size, hubris, and complacency of a BigCorpBlah victim..
Your Working Boy,
Re:Mac OS Rumors (Score:1)
IBM's uses for G4 ..Re:Apple IBM and Moto (Score:1)
1) IBM could use additional revenue.
2) IBM could use the reputation boost from the Apple community of breaking the long standing mhz freeze.
3) Never underestimate what IBM will find a use for. If (and when, if my predictions are correct)
they do produce a PPC7400 (or variant) be sure they'll have other uses for it.
A host is a host from coast to coast, but no one uses a host that's close
Re:Give me a break (Score:1)
And that is what a lot of people do : open their mail or plan their next move as photoshop chugs on. It isnt a fatal flaw for many people.
MacOS X Server has absolutely nothing to do with MacOS X. It is a totally different operating system. MacOS X Server is based on NeXT, not BSD like MacOS X.
Um , "absolutely nothing" seems a bit harsh. NeXT, mac os x , and mac os x server are all based on both mach and BSD. The point of my bringing up macosxs was because you declared that mac os x and every other promised os was vapor.
RELEASE means that it is a FINAL PRODUCT
oh i thought release meant it was something that had been released. mac OS X Server has been released as a final product
On the contrary, your post is the one that reads like something written by a 15 year old. If you want to get particular, the only word that you capitalized correctly is multitasking - something that your precious Macs can't even perform!
my macs can TOO capitalize! ;) (btw i don't proofread for punctuation either)
In retrospect questioning your maturity wasn't appropriate. But... to me both of your posts read as whiney and less than what i would expect from an adult. You reply by saying "oh yeah i'm rubber and you're glue ...!" and then you procede to take out your computer list and wave it around.
MacOS is shit.
how eloquent. and the worst is that apple will probably screw up macosx
Only on Slashdot can your post be moderated down for stating facts.
welcome to slashdot! everywhere else you'd get flamed... : )
OS support is the real question (Score:3)
Don't get me wrong, I am one of the few who actually like AIX. I think it's a mature, useful operating system with some really cool characteristics (fairly integrated hw support and debugging, excellent logical volume manager (better than veritas, imho)). nevertheless, it remains to be be seen whether IBM can actually bring Linux to their whole server platform (including these bad boys).
(There have been instruction set changes in the IBM processor line in the past, particularly between the POWER, POWER2 and POWER3 architectures, so I'm interested to see what the differences in this instruction set are...).
It's a server, guys (Score:1)
The answer is a multiprocessor (because Oracle is threaded) with a shared L2 cache (because the threads share most of the L2 footprint.) The advantage is quite large.
So, what IBM is building here is a server engine. Which is why there will be a 128-processor system, but there won't be Altivec.
Re:the dominant 64 bit processor (Score:1)
I have feeling we are going to be seeing a lot more from Apple, the Linux community and maybe even BE.
- tsi
Re:Jobs is no diplomat but i agree with his statem (Score:1)
To IBM's credit... (Score:4)
So just for that IBM's not a bit bad, and their NUMA-Q architecture looks REALLY neato. As for putting Alpha and Sparc out of business...Hey, you build a better mousetrap. Big Blue has always had great R&D and put out some of the best products out there. That doesn't mean Alpha and Sparc and such are going to plummet.
I say kudos to Big Blue.
Why Apple Can't (Score:1)
Motorolla - sounds like detroit house (Score:1)
Re:System design (Score:1)
The line with the 386 (and 486, and P5, and PPro) was that it was destined to remain a workstation chip for some time. This was Intel marketing bluster, yes, but it was moderately true--the initial versions of each of those chips was produced on an old process; once the chip moved to a new process, it became feasible for upper-end mainstream machines. In a two years, they were mainstream.
Right. But the difference here is, these chips are not intended for workstations. They're not intended for moderately sized servers. They're intended to replace mainframes, and to run high-high-end scientific code. In case you didn't read my other post [slashdot.org], these things are going to cost *at least* $10,000 *apiece* to MAKE. Just for the MPU. Moreover, they will not work up to their full potential without *massive* bandwidth, which still costs mucho $, last time I checked.
What servers really need is multiple CPUs and huge I/O bandwidth, not faster individual CPUs.
Oh wait, you didn't even read the article. The POWER4 *is* multiple MPU's--it's 8 cores on 1 die. This thing doesn't come in anything less than 8-way configurations. As for I/O bandwidth, would 84 GB/s be enough for you?
You still think this thing is a desktop chip???
On the other hand, Apple can't afford to change CPUs again.
Or maybe they can't afford not to. Contrary to all those MacOS Rumors, there are no definite plans for Apple to move even to the upcoming G4+ MPU's, which are essentially another incarnation of the tired G4, just with a stretched out pipeline which will get it to 800MHz at the cost of lower IPC. There aren't definite plans yet because the G4+, like the G4 and the G3 before it, is not a desktop MPU but rather an embedded/DSP chip which Motorola happens to sell to Apple to use in Macs. The design, ramping, pricing, roadmap, are all influenced primarily by Moto's embedded customers first, not by Apple. Unfortunately, despite the fact that they are utterly dependent on them, Apple has decided to treat first IBM and then Motorola rather poorly, and thus haven't gotten much in the way of support when they decided that they may, perhaps, want to increase the speed of their top MPU more than 50 MHz in a year. Whoops. BTW, did you catch Apple's earnings today? Ouch.
All of this is too bad--OS X looks like perhaps the best thing going as far as operating systems goes. There are always rumors that Apple's going to finally make their surprise move to x86. Their experience with PPC the last year or so, and the accompanying beating their bottom line has taken, might be the thing to finally push them over. I personally think they might still be able to carry over enough incompatabilities to stay the sole supplier of MacOS hardware--after all, the XBox uses x86, and it will be plenty incompatible with PCs. Migrating software will be a gigantic pain...but on the other hand, it's not like the Mac has too much in the way of software anyways. (OS X, and any Cocoa programs, will port very very quickly.) Who knows?
Re:Drop Motorolla like a hot potato (Score:1)
"AMD is 64 bits! Which one is more advanced? Signal 11!?"
"Do the math!"
You heard it here first.
great review, but some nits.... (Score:4)
Re:NO NEW INFORMATION (Score:1)
However, since you mention another source of information, would you be so kind as to post a link, perhaps?
---
pb Reply or e-mail; don't vaguely moderate [ncsu.edu].
Apple has nothing to do with this. (Score:1)
The PowerX line of chips don't run word, they don't process photoshop filters real well, and they need MASSIVE cooling. In fact, I don't see anything at all that would appeal to an actual Macintosh customer. What DO the chips do?
Well they crunch numbers, run molecular simulations, etc... I do know enough about the Power4 to give a decent speach on the target market and application uses, however all I wish to do is say upfront, that if your comment has anything to do with Apple and the Power4, you are waisting time, and most of all, showing your lack of knowledge when it comes to computing. Forgive the typos, I have to get back to hour 37 of work.
chow
bort
REal-world performnce gap? (Score:2)
NOTE: for what it's worth, Sun SPARCs give excellent MAQSIP performance, but even more miserable MM5 performance (as compared to processor-peak) than POWER3.
Re:Step in the right direction (Score:2)
"[...]no unix server can currently compete with even a middle of the road OS/390 machine for heavy server/transaction/database type workloads."
This would be true if you added "in a ultra-high-reliabilty environment". Fact is that not a single ISP uses S/390 systems for serving web content. If the IO of these machines would be so excellent, why don't they use them? What makes things even worse is the fact that serving web pages is similar to the IO load envisioned by Amdahl, namely relativly large chunks of data being tranfered. As a result even the terminals (3270) are based on transactions of this type. The user edits on the screen and commits changes every once in a while. This is very different to the character based aproach of Unix ttys.
"[...]a modern OS/390 the IO is handled by up to 1024 of these processers called 'channels'."
This is unclear to me. OS/390 is an operating system but you seem to make a statement about hardware. True is that 390-IO is based on a channel subsystem. All models from G3 through G7 (GA 2000) have 256 channels. Wrong is that a channel has something to do with a processor. A channel is an IO line with an interrupt on its own. Each of these channels may end in a channel controller. To the channel controller one can attach 256 subchannels. The subchannels end in a device. This makes a total of 64k devices.
A modern G6 has 16 processors (390 architecture). 14 for workload, 2 for the IO subsystem, 2 in case two of the others die.
"Nearly everything relating to transactions was done forst on an OS/390, databases in general, relational databases, messageing and queueing software, & etc. are all areas where the intial and continuing innovation took place on OS/390s"
True. One of the last "innovations" of DB2 on 390 was to optimize the data distribution on the harddisks depending on the speed of the movement of the HD heads. The last 390 harddisk physically built was a 4 Gig drive (I beleave it was called 3390 model 4) in the early 1980s. The disks had a diameter of almost 1 yard. Since then IBM simulated these 3390s through (SCSI-) disk arrays. The controlling software of these RAIDs introduce special waits to not disturb and crash DB2 that relies on specific timings.
The last disk logically defined was a 8 Gig drive. In case all 64k devices are 8 Gig drives this makes 512 TB storage. This will be a boarder very difficult to cross.
Is OS/390 innovative? All I know is that the OS/390 filesystem is non-hierarachical, i.e. does not know of directories! The filesystem is not block oriented. If you append data to the end of a file the file may overflow. This means that the user must create a new, bigger file and copy the old file into the new one and delete the old one.
Programs in OS/390 must be started using a special "scripting" language to supply parameters!
I could go on with this list. 390 were nice in the 70es. They still do a good job in some places that can afford them and need high reliability. All I'm saying is that the future belongs to a different kind of machine (definitly not 31bit like S/390) with a different kind of OS.
Re:To IBM's credit... (Score:1)
--
Re:Processor design... (Score:1)
First off, I'd like to say that that was indeed a great response. While I'm not sold on IBM's approach just yet, at least I'm more informed now than I was when I read the article.
The main technique I've seen exploited in DSP programming (I know someone who does this for a living) is software pipelining, which often involves like loop unrolling, except that you have to pay attention to the instructions to make sure you use all your instruction units.
Compilers these days can do loop unrolling, and past that I guess you'd just hope to be able to reorder the instructions somewhat, to get a decent instruction mix out of the code in the loop, and maximize that magical "IPC" number. However, yes, it's hard to get rid of all the dependencies, and run-time profiling (a la Transmeta) will probably get more popular as research into VLIW systems gains in popularity.
SMT sounds interesting; does it refer to "threads" in the software sense, or just separate processes? For the moment I'll view it with skepticism, just as I did when Sun built streams into the kernel. Any multiprocessing Unix system should be able to run separate processes on separate processors, and some of them can surely do the same with different threads, depending on the implementation. I guess IBM would just have a more compact and possibly more scalable solution in this case.
And yes, I realize these are supposed to be server chips for now, and knowing IBM, they might just stay that way. But this sort of technology usually filters down into the PC market quicker than you'd think, especially if it improves the price/performance ratio, as this might do, eventually.
---
pb Reply or e-mail; don't vaguely moderate [ncsu.edu].
There is no reason why Linux should not be forked (Score:1)
Linux could survive quite easily with at least 3 forks. It is arguable that it has two at the moment with the 8086 versions.
I see mebbe Embedded Linux, 'Official' (Linus) Linux, and Big Iron Linux as 3 forks that could happily live with each other, sharing code wherever possible or neccessary.
I mean, its not like the code is going away ... this is Free Software we are talking about here.
Step in the right direction (Score:2)
IBM politics was that certain techniques (like the ceramic multi-chip modules, copper process, etc.) were reserved for the S/390s. This was the reason why the RS/6000s with AIX were never really competitive (S/70 AIX server).
When SUN brought their E10000 server into shops IBM thought they'd never loose to SUN they obviously started changing their minds. The (RS/6000 based) 24 processor S/80 outperformes the 64 processor E10000.
This new POWER4 design makes clear that IBM favours modern Unix-based RISC servers over the old S/390 mainframes.
This is a good thing especially for IBM S/390 customer who start having problems finding talented people who want to work with dinosaur machines and OSes. (IBM also has major problems implementing new or even innovative software for the S/390s. Exception: Linux/390 ;-).
Better IBM offers them a safe way into the RISC/Unix world (we are looking at huge amounts of enterprise critical data sitting on all these S/390s) than when they try to migrate on their own.
Re:Processor design... (Score:2)
I have to wonder if an over clocked microprogram unit with a CISC instruction set couldn't be made competitive again. Ultimately that might be a simpler design - and thus potentially faster than a modern RISC chip.
One of the ultimate limits on processor performance is how fast you can get instructions into the processor. If you think of CISC as a compressed (Huffman encoded) version of RISC it is easy to see that CISC does have a theoretical advantage there.
I know this is heresy, and that the modern religion is RISC == GOOD, CISC == BAD; but it might at least be worth someone spending some time thinking about it. If nothing else, the reduced transistor count could do something about the spiraling power consumption problems in processors. Or you could integrate multiple processors on one die and do the SMP on chip with the much smaller cores this would make possible.
Re:the dominant 64 bit processor (Score:1)
Damn /. effect... (Score:1)
An Error Occurred While Attempting to Process the Request
Date/Time: 10/18/00 09:13:50
Template: D:\html\users\deankent\realworldtechcom\html\page
Remote Address: 12.20.217.130
Refering URL: http://slashdot.org/
Diagnostics
ODBC Error Code = S1000 (General error)
[Microsoft][ODBC Microsoft Access Driver] Could not update; currently locked by user 'admin' on
machine 'CFX1'.
SQL = "UPDATE adlog SET total = 12394 WHERE stamp = '2000/10/18' AND adid = 10 AND type
= 1.0"
Data Source = "ADS"
Re:REal-world performnce gap? (Score:1)
Also, article's conclusions are pretty badly flawed, esp. regarding the Alpha's future. The supposition that the POWER4 will eclipse the EV68 should be obvious but inconsequential, given how far the POWER4 is behind the EV68 in terms of getting to market, and the POWER4 has yet to convince anyone that it will be able to compete with EV7 in real world performance (of course, everyone's guesses could be wrong and EV7 could end up a big dog, but its unlikely).
Furthermore, considering the article is dated October 16th (2 days ago) I found the following statement interesting:
Considering that UltraSPARC-III is already shipping, this statement was confusing. But it doesn't really matter, UltraSPARC-III at 900MHz is already outclassed by Alpha EV67 at 667MHz [cite] [ideasinternational.com], no one even considers them a contender in the performance race.And what about HP? They gambled too heavily on IA-64 and cut back development of their PA-RISC, which was just recently getting interesting. Now that IA-64 is delayed and will probably be a dog (notice how Intel hasn't been leaking any spec numbers?), HP is realizing what a mistake they made and has revitalized their own development; it should be enough to keep them from completely being left behind, but just barely.
Re:Give me a break (Score:1)
Oh wait you obviously have done that. OS X is actually pretty damn nice looking........
Jeremy
Re:Drop Motorolla like a hot potato (Score:1)
baby if you're going to correct peoples' spelling, you really ought to double-check your own.
Steve Jobs is as much to blame as others. (Score:4)
Steve Jobs said in a visit to Motorola
"It will be great in two years when we arn't using your chips."
After this statement is when Motorola publicly started calling the PPC line 'embedded'
How often in YOUR relationships can you walk up to your relationship partners and tell them 'to hell with you, I'll be leaving in 2 years.' and NOT expect said partner to keep giving a damn about you.
Apple then made the problem WORSE by pubically calling altivec 'the future' and spent hours about how wonderful altivec is. Apple will have a hard time leaving AltiVec with all the statements about how wonderful altivec is.
Jobs ego put Apple in the place Apple is. Motorola only reacted to the actions Jobs took. It is not like Motorola NEEDS Apple, and took actions to protect Motorola's investment.
Jobs wants to be the 'saviour' of Apple, fine. Then Jobs must also take the mantle of the person who helps kill Apple also. Amazing how the history of Jobs repeats.
Re:Apple and Motorola (Score:1)
Yeah, it seems to work very well with MP3 ripping.. My 500MHz Cube can do 128kbps VBR normal-stereo at between 2.5x and 5x realtime..
Your Working Boy,
Re:Apple and Motorola (Score:1)
So can my 500mhz P3 - And I'm doing the mid/high quality encoding which goes up to ~224kbps. I don't know if that makes encoding faster or slower. MP3 encoding (ripping is a function of CD-Rom speed and I/O bandwidth, and in and of itself has nothing to do with mp3) does not parallelize well, at least as it's done today. So altivec really doesn't buy you anything there.
System design (Score:2)
That line first appeared with the 386. And it was wrong then, too.
What servers really need is multiple CPUs and huge I/O bandwidth, not faster individual CPUs. Loaded servers always have lots of threads running. On desktops, one thread typically is using most of the CPU time. Thus, the desktop is the place where the fastest CPUs are used. Servers are configured for max price/performance without sacrificing reliability, and tend to run a bit behind the fastest desktops.
On the other hand, Apple can't afford to change CPUs again. The last transition cost them a big fraction of their applications (for example, almost all the CAD vendors bailed out) and a big chunk of their user base. In retrospect, better 680x0 machines would have worked out better than going to PowerPC. The whole PowerPC thing was supposed to get IBM into MacOS, remember, and that was a total disaster. Now that everybody knows how to make CISC machines faster, there's no reason 68K machines couldn't be up there with x86 machines. And the architecture is much better.
Drop Motorolla like a hot potato (Score:1)
Re:M$ on something else than x86. I doubt it. (Score:1)
Re:Apple and Motorola (Score:2)
Re:POWER4 runs PPC binaries? (Score:2)
Now THERE'S a bad business decision.
This all just proves one thing. It's not PPC that's broken. It's Motorola, and AIM that are broken. AIM doesn't allow any REAL competition between Motorola and IBM. Too bad that the entire computing world is held hostage by the whims of evil Bill Walker of Motorola.
Re:Windows database server load (Score:1)
Now, I never said I hadn't written stupid code - And you don't have to be an experienced programmer to do that, either. Anyone can write bad code. Including me.
Processor design... (Score:1)
However, it *is* nice to have this depth of technical information to examine, and also it's good to know that they're still doing this.
I think the big advantage that VLIW instruction sets will have is strictly architectural, and I'm not sure how IBM's approach fits in yet, but it looks interesting. Throwing more chips at the problem is one approach, but remember that your competitors can do that too, *and* make the chips do more as well...
However, IBM will have to make sure people design their apps with more than one processor in mind, which will be a Good Idea for the future, since more people might have multiprocessor computers.
---
pb Reply or e-mail; don't vaguely moderate [ncsu.edu].
Re:Step in the right direction (Score:1)
I also used to have an IBM PC Server Model 500, which is a Pentium 90. It had something like 32mb of ram, and when I got it it actually had a MCA bus mainframe card in it, and a number of terminals hooked up to it, because it was being used to test a networking product a company I worked for developed back in the day for IBM.
I ended up yanking out the card, throwing it away, and putting NT 4 Server on the system, and using it for quite a while. It had a Mylex DAC960 in it, and 11 (!) 2.25gb UW disks in it. Eventually I pulled the ECC RAM, the disks, the 2.88mb floppy, DDS-2 DAT and 2x :) SCSI CDROM, saved that, and ditched the case, which was about the size of two full tower AT cases.
I would have kept it, but have you ever seen the price on a 100mbps NIC for MCA bus?
Re:System design (Score:1)
My understanding of RISC has always been lower CPI, higher clock at the cost of higher instruction counts. 'Deep' (and wide) pipelines had the advantage of both increasing the clock rate and lowering the CPI but at the cost of requiring more instructions to do similar actions. I don't really remember the RISC camp loosing on the clock rate war until the early '90s when the Pentium came out.
Re:Drop Motorolla like a hot potato (Score:1)
Why get faster when you can get MORE for less (that is as long as you don't run windows to suck all your CPU cycles trying to look pretty)
Re:Processor design... (Score:5)
Not sure how IBM's approach fits in yet?? Read the article.
Amongst other things, the POWER4 is *not* VLIW, it's straight-ahead modern RISC at its finest. With massively gigantic buffers, bandwidth and execution resources (8 functional units/core * 8 cores = wow), this chip'll do quite nicely on IPC/core, not to mention combined IPC for all 8. While presumably not quite as elegant, the design for the individual cores bears a lot in common with the archetype of perfect RISC cores, the Alpha 21264, and it has even more aggressive resources.
Essentially what this means is, assuming this design is as good as it appears, the only way the competition will be able to catch up (without going the way IBM has and deciding on a prohibitively expensive 8-in-one design and packaging) will be through the use of innovative design tricks. The upcoming P4 has a few of those, incidentally, but the big one--and the one the P4 *doesn't* have--appears to be SMT, Simultaneous MultiThreading. Alpha has an 8-way SMT core coming out in a bit, and it ought to compete well with IBM's much more expensive 8-way SMP design here. And AMD appears ready to do 2-way SMT (or something similar) with the Sledgehammer in about 15-18 months. And Sun is rumored to have SMT in the USV design due in several years. But the POWER4 looks to lead in the "big bad" category for quite some time to come.
(As for Intel's EPIC, the VLIW-like design strategy for their IA-64 chips, at the moment it's looking like a rather poor competitor to SMT. A quick explanation of why:
There are exactly two ways to make an MPU run faster: 1) increase the clock speed, or 2) increase the IPC (instructions per clock). Unfortunately, the best we've been able to do so far in the IPC realm is about 1.4 IPC on SPEC benchmarks (Alpha EV6x). IPC on a P3 runs about 40% lower. Now, these IPC numbers are despite the fact that the Alpha can theoretically retire 8 instructions/clock, and the P3 5 (5 internal ops, not 5 x86 ops). Furthermore, simulations show that as far as attacking the IPC problem by adding more functional units, we're nearing the point of diminishing returns.
The problem is, in order to run lots of instructions in parallel, you have to be able to safely extract parallelism from your code. And the problem with this is, you can't run instructions in parallel if they have dependencies, etc. And furthermore, nowadays all this parallelism has to be safely extracted in real-time by special hardware in the MPU itself; this makes your chip more complicated, and means you need to build a big buffer to hold instructions in flight so you can pick and choose which ones you want to run each clock.
So many many years ago, HP had the idea, which it later sold to Intel (and which wasn't really there idea at all but has indeed been used in DSP chips for years and years), of getting rid of all that complex insruction-level parallel-finding logic on the MPU and doing it all at compile-time instead. This is the basic idea behind EPIC, the philosophy of Intel's IA-64 line.
It sounds very nice, especially because in theory it means simpler chips (no complicated control logic), and simpler chips means faster chips. Heh heh heh. See it turns out that the amount of instruction-level parallelism which can be safely discovered at compile-time is way way less than the amount that could be found in the chip at run-time (which, as we recall, is too small already). Thus EPIC was modified to allow the compiler to just place "hints" in the code. Well, this means you still need all that complicated control logic back in place, because you still don't have deterministically scheduled instructions. But following the "hints" and other changes to the ISA ends up making everything *more* complicated, not less. This, in a nutshell, is why Itanium is 3 years late, way over budget, unable to meet its very modest clock speed goal of 800 MHz, and fitted with a laugh-enducing 96kb of on-die cache, lower even than the lowliest Celeron: all this added complexity means bigger, slower, more complicated chips that don't have the room for cache or the elegance for high (or even adequate) clock speeds. Plus we have very strong evidence that compiler technology is still not nearly good enough to make the kinds of insightful IPC-giving "hints" which are necessary to even make the damn fool scheme work. Thus the only benchmark Intel has "released" for the Itanium is that of an RSA-encryption--a routine simple enough to be hand-tuned in assembly. Meanwhile they have made the patently ridiculous claim that the SPEC benchmarks--directed precisely at the mid-cost server/workstation market which Itanium is aiming for--are "not relevant" to Itanium's market.
A completely opposite approach is SMT, which uses a relatively small number of core changes to allow not just instruction-level parallelism to be gleaned, but also thread-level parallelism. In other words, the chip will run several threads in parallel, confident in the fact that their instructions will not have dependencies on each other, and thus be able to use much more of its full execution capabilities. Early indications are that SMT can improve IPC by remarkable amounts, like on the order of 2x the performance on otherwise similar cores!
Unfortunately, it is too early to tell whether SMT will be as easy a design enhancement as is being claimed. Furthermore I've heard tell that SMT on IA-64 will be a lot more difficult than on a RISC MPU, so Intel could be missing out on a huge speed-up with this technique.)
However, IBM will have to make sure people design their apps with more than one processor in mind, which will be a Good Idea for the future, since more people might have multiprocessor computers.
These chips are not to be confused with PowerPC chips. They are server chips only, intended for seriously expensive machines.
Not thoroughly discussed (Score:2)
Re:Why Apple Can't (Score:1)
I'd love to see the source for that.
Even if they did have a contract, they could probably get out of it due to non-delivery of faster chips by Moto.
What an arse of a title (Score:1)
It's as silly as the debate when AtheOS came out - is this going to compete with Linux? Is it a threat?
No, it's an option. When it comes to computer's, options are good.
(honestly, it looks like a newspaper headline - that's a bad bad bad thing.)
Re:Give me a break (Score:1)
a lot of macs are single purpose : for fast photoshop. Multitasking not required.
So far, it is vapor
-bzzzzt-!! try again ! mac os x server shipped over a year ago (not to mention the release of macosxbeta)
on a side note i wonder how old you are ? 15? 16? your post certainly doesnt read like it was written by a grown up.
Mac OS Rumors (Score:2)
The rumor is now archived at:
http://macosrumors.com/?view=archive/8-00
If you dont feel like going to MOSR the link in the rumor to the Power4 info is:
http://www.austin.ibm.com/resource/features/1999/
Re:IBM != Motorola (Score:1)
Apple didn't supported the PREP or later CHRP efforts made by IBM, Motorola and other hardware companies around the PowerPC and instead prefered to stay on its little Mac-island with Apple-crap.
Imagine that: You might have been able to buy your PowerPC based PC (with working OpenFirmware!) from your local dealer at low-low prices and install your favourite OS as you do with IA32 PCs today...
Re:the dominant 64 bit processor (Score:1)
Re:great review, but some nits.... (Score:1)
There is (almost) nothing to fix. Linux kernel does indeed handle more than 24 processors. Check some recent post on linux-kernel (I can remember some post about a 128GB-equipped alpha).
There is no explicit limit to the # of CPUs Linux kernel can handle. It is just that it is not known or proven tha Linux scales well on this amout of CPUs. But the work on improving scalability is in progress (NUMA memory allocators, etc). Stay tuned :-)
-Yenya
--
Re:Processor design... (Score:1)
Re:great review, but some nits.... (Score:3)
Actually, yes there is. Linux currently uses a bitmask to specify certain CPU operations, so the number of CPUs is limited to the word length. In other words, Linux supports up to 32 CPUs on 32-bit platforms, and up to 64 CPUs on real machines (Sparc64, Alpha, Itanium (and MIPS64?)). Of course the fact that it supports that many CPUs doesn't mean that it scales linearly, but it looks like the 2.4 kernel will be good for at least 16 CPUs before performance starts dropping off. Various people (Dave Miller, Ralf Baechle and others) are working to remove the bitmask, and allow more CPUs than the word length. SGI in particular need Linxu to be able to support more than 64 CPUs for some of their machines.
Not terribly far fetched. (Score:2)
Now we've gone full circle and memory is the bottleneck again. CISC could provide a performance advantage again.
M$ on something else than x86. I doubt it. (Score:2)
Hey, this chipset is for some serious computing. Serious, serious. The range of boring, mundane software that would get a big boost from these fat, fat pipes into these fast, fast cores is limited.
Quake]|[ would absolutely drip with 3D VR gore. (I get ill just thinking about it. Gibs everywhere!)
But would you really need that kind of horsepower to run Word or an Excel spread sheet of even the maximal complexity that Excel can handle? I thought not. (Excel plays fast and loose with some math functions, Newton's approximations, etc. I just implemented algorithms which don't. Banks can't use Excel for real world amounts.)
Face it, M$ can't use it. Even GHz x86 chipsets are a waste for the desktop.
The server market is better served by Unix solutions that runs multi-user(NT is not), multi-threaded,) and across a range of big iron that's growing steadily bigger.
M$ support this? I hope the [expletive deleted] not!
Threat to IA-64? (Score:1)
Re:Apple and Motorola (Score:1)
Re:Step in the right direction (Score:1)
Whatever the hype from Sun and HP no unix server can currently compete with even a middle of the road OS/390 machine for heavy server/transaction/database type workloads.
The OS/390 was architedted by Gene Amhdal in the late 60s. The main problem was to get through the very large workloads required for NASA etc. using quite slow electronics. This was done by using the now well established tecniques of pipeline and parallel processing. What they also did (which is not common) is unload all the IO processing to seperate processors. On a modern OS/390 the IO is handled by up to 1024 of these processers called "channels".
You can ask a channel for thing like "read all the blocks in cylendars 12,14 and 24 into the this list of buffers" and then go do something else, when the buffers have been filled, an interupt will be posted.
As for innovative software. Nearly everything relating to transactions was done forst on an OS/390, databases in general, relational databases, messageing and queueing software, & etc. are all areas where the intial and continuing innovation took place on OS/390s (with an honorable mention for VAX VMS). For all the hype about UN*X very little innivation has taken place on these machines, (great exception being web servers).
In fact the most hyped recent advances, in the areas of clustering and high availability, have been around on OS/390 (in wierd and wonderful ways) and VMS (perfectly formed and complete in the first release of VMS) for well over 10 years.
Re:Not terribly far fetched. (Score:1)
An ISA designed with this in mind might be a win (say first few bits determine instruction length for example), but you still end up with decode of an instruction depending on finding out where it begins which depends on the previous instruction type which depends on the one before it which depends on... This makes decoding a lot of instructions at the same time hard, which means having a fat pipe is harder.
Like anything, it's a tradeoff.
Apple IBM and Moto (Score:1)
IBM is the primary source for all Apple G3 processors. Moto is the source for G4, solely because IBM up until now, has opted to not produce a chip with AltiVec.
I have said for over a year now that if IBM fabbed the G4 for Moto, that the high speed yields would come up, and that if IBM produced the G4 that the speed rating would increase.
IBM just dropped 5 billion on new fabrication plants. IF IBM wanted to *own* the OEM contract for all of Apple's processors, they'd only have to produce their flavor of a PPC7400.
I predict they will within a years time, and at speeds comparable to Power4
A host is a host from coast to coast, but no one uses a host that's close
Re:Processor design... (Score:1)
BeOS should have no problems, and Linux should do better now with glibc...
I think a model like this would be better served with processes rather than threads; in all of these systems, will there be unified access to memory? I know the POWER4 will have it, since this is just a beast of a CPU grafted onto a traditional computer, but I can see problems in any NUMA system, where the memory for one thread might be closer to a separate processor. I guess they'll have to take that into consideration as well, for systems like that...
---
pb Reply or e-mail; don't vaguely moderate [ncsu.edu].
Re:Not terribly far fetched. (Score:1)
What we need is a new architecture that supports direct 1 clock access to RAM. With that, we don't need complex caching algorithms and can actually do interesting things with the instruction sets like the self-optimizing microcode that Zilog was working on for the Z-80000 or actually having instruction sets optimized for programming rather than ease of cache design.
Re:Give me a break (Score:1)
-bzzzzt-!! try again ! mac os x server shipped over a year ago (not to mention the release of macosxbeta)
MacOS X Server has absolutely nothing to do with MacOS X. It is a totally different operating system. MacOS X Server is based on NeXT, not BSD like MacOS X. MacOS X Public Beta is NOT A RELEASE. RELEASE means that it is a FINAL PRODUCT.
on a side note i wonder how old you are ? 15? 16? your post certainly doesnt read like it was written by a grown up.
On the contrary, your post is the one that reads like something written by a 15 year old. If you want to get particular, the only word that you capitalized correctly is multitasking - something that your precious Macs can't even perform!
FYI, I am 29 years old. My first computer was a TRS-80 Model 1. I got a Color Computer when I was in the 6th grade. Since then, I've personally moved through an Amiga 1000, an Amiga 2000, a Quadra 840av, a 7200/75, a Power Computing PowerCenter Pro and various Intel boxes. I've used Macs almost every day for the last seven years, and I know the operating system inside and out. Believe me when I say that MacOS is shit.
I used to like Macs because at one time, they actually performed better than PCs. Sure, the G4's AltiVec crap is wicked fast, but the OS does not take advantage of it. Advantages are found only in applications written specifically for the use of it. The only thing that it is good for is so that Apple can take benchmarks performed on AltiVec (which have no relevance to 99.9% of the computer's real performance) and use the results in a misleading ad campaign about G4 Macs being "supercomputers." The only thing that I have found that actually uses AltiVec is the Distributed RC5 client. It is fast as hell, but to what gain? When I'm not cracking RC5 keys, my 18 month old PIII/500 Linux box kicks the shit out of the G4/400 that I am writing this on 99% of the time - and for FAR less money.
So, let me say it again: Give me a break! :P
Only on Slashdot can your post be moderated down for stating facts.
You Mac zealots just can't stand the fact that I'm right! I know, I used to be one of you. There is help available to you all! [linuxppc.com]
(Score: -1 Troll)
Re:"If only Apple could be persuaded to use these. (Score:1)
Re:Step in the right direction (Score:1)
I have one word for you: ROMP. Well, I suppose it's an acronym.
I used to have several IBM RT-PC model 135 machines; 16mb of ram, very classy. The model 135 was the most advanced RT developed, and it wasn't very exciting by the time I got them - of course, when I got them they were very far behind the curve. Even so, that machine supposedly cranked 5.6 MIPS (Meaningless Indicator of Processor Speed) and had an 80ns cycle time, so it's a bit faster than a 33mhz 486 as far as MIPS ratings go, yes? You could run AOS 4.3 (IE, IBM's port of BSD-4.3-lite) on them, normal BSD 4.3-lite, and then someone ported BSD 4.4-lite to them somewhat later, by which time the only RT-PC systems were legacy, and mostly running AIX 2.2.
Apparently, someone is trying to organize a port [daemonz.org] of OpenBSD to ROMP. There's an article on the ROMP architecture (which included a full MMU) in IBM Systems Journal [ncu.edu.tw], Volume 26, Number 4, 1987. Unfortunately, you have to pay IBM $30 for back journals, and the text isn't available via the web. There's also a RT PC FAQ [jmas.co.jp]. It has the following interesting tidbit:
Ahh, gotta love 'dat internet. There's also a RT Hardware FAQ [jmas.co.jp] which has the following to say:
That's fairly entertaining.
Re:What an arse of a title (Score:1)
Re:Drop Motorolla like a hot potato (Score:1)
Anyway, with Intel and AMD coming, "64-bit" is going to be a key marketing requirement for desktop systems in the next few years. (If it sells game consoles, it will sell computers.) Apple, however, has to hold off for practical reasons until they get their base moved to 32-bit OSX.
Re:Apple IBM and Moto (Score:1)
Of course, they don't really have a use for it personally, inside of IBM, because they'd have to develop a new RS6k architecture to support it. And do you really think that someone with the savvy (recently reacquired savvy, but savvy nonetheless) that IBM contains wants to get involved with someone as comfortable screwing their industry associates as Apple? I kind of doubt it. IBM has no problem providing the G3 chips because they use them themselves, so if Apple suddenly said "Today we decided to stop buying these chips from you and get them all from motorola when motorola can actually get their shit together and produce them" then IBM would just say "That's cool, fuck you too and good luck buying anything from us at a reasonable price ever again."
Meanwhile, it's curious that someone with the silicon ability of motorola can't meet Apple's demand for the less-used G3 chips at this time. I tend to wonder if that's because they've geared up for G4, or if it's because they don't want to depend too much on Apple.
Re:M$ on something else than x86. I doubt it. (Score:1)
They did a PPC version too, then dropped out of that particular rat race when they discovered that people didn't get behind PPC like IBM/Moto had hoped, and there was a distinct lack of both CHRP and PREP platforms out there. There were two standards, and even if they were put together there wouldn't be enough hardware to bother supporting NT on.
Re:Windows database server load (Score:1)
The error you received
indicates that someone was doing maintenance on the table in which the data is stored, probably updating a typo or something similar in the story.
The only crime committed here is by a stupid programmer who doesn't know how to redirect you to a less-lame error page than the default.
Re:Windows database server load (Score:1)
The error you received
indicates that someone was doing maintenance on the table in which the data is stored, probably updating a typo or something similar in the story.
The only crime committed here is by a stupid programmer who doesn't know how to redirect you to a less-lame error page than the default.
Where is AMD's Hammer? (Score:3)
Hammer does not have a track record in the marketplace, but neither does Itanium, and it's odd to ignore an architecture that in all likelihood will sell in much greater volume than several of the chips profiled here. Even if AMD's 64-bit implementation turns out less than ideal, it will probably outsell the Power, Alpha and Sparc offerings by virtue of the vastly larger market it targets.
A simulator for a Hammer chip has been released. A comparison, or at least an acknowledgement, would have made the article more valuable.
Apple and Motorola (Score:2)
That said, I don't what Motorola's plans for a G5 are, if any. It may turn out that Apple has no choice but to go with IBM's chips after a while.
GCC supports AltiVec (Score:1)
Re:Drop Motorolla like a hot potato (Score:3)
PPC chips are optimized for cost, POWER chips are optimized for performance, screw the cost.
Re:the dominant 64 bit processor (Score:2)
Lee Reynolds
Embedded Alphas (Score:2)
Imagine, embedded Alphas!
As this is a major deal, Compaq will have an output for years to come and the Alphas seem far from dead, or even threated.
Re:Not terribly far fetched. (Score:2)
Apple, IBM, PReP & CHRP (Score:1)
Isn't it amazing what entrenched bureaucracies can do to each other (with the users as innocent bystanders)
Re:the dominant 64 bit processor (Score:1)
Re:Processor design... (Score:2)
Re:"If only Apple could be persuaded to use these. (Score:1)
Another approach is to just port Darwin to RS/6000's.
(OT)Only 'Official' Linux could be called Linux® (Score:1)
I see mebbe Embedded Linux, 'Official' (Linus) Linux, and Big Iron Linux as 3 forks that could happily live with each other
But marketing would have to think of a name for the embedded kernel and the big iron kernel; as Linux® is taken by 'Official' (Linus) Linux.
Linux is a registered trademark of Linus Torvalds.Sparc and Alpha at Risc, Not AAPL... (Score:2)
On The Other Hand, Apple's direction has lately been to using convection cooling, not the water cooling that would be necessary for the number of watts the Power4 dissipates.
Re:Step in the right direction (Score:1)
Hehe, that's a smidge before my time on IBM RISC hardware.. I started with a 7012-340 workstation at about the time the 390 was released.. What a wonderful frankenbeast that was (ripped a sabine adapter out of one system and used it until AIX v4, various 8mm, QIC, HDDs chained off, 16mbps token ring..)..
Though I did work with a dude who was on a team that developed a mainframe-on-a-card for PCs. Not terminal emulators (like Wang cards), but an actual System 370 system that fit into an AT slot used for software development.. PC/370 IIRC..
Your Working Boy,
Re:Processor design... (Score:1)
The problems with compilers and EPIC is that current popular languages like C, C++ and Java, are not designed for parallel machines. There are extensions that add explicit parallel keywords to these languages but they're all non-standard. Anyway, explicitly stating what statements can run in parallel increases the application programmer's burden. And extracting parallelism from a imperative language puts a helluva burden on the compiler writer.
Now languages like Lisp or Scheme should lend themselves well to the EPIC architecture. It's easier to extract implied parallelism from a functional language because more of the scheduling decisions are left up to the compiler. A Scheme like lowlevel language should produce code that screams on EPIC.
I think HP/Intel are in unfamiliar territory -- both on architecture and languages. Once they gain that experience, later generation EPICs should be phenomenal.
Jobs is no diplomat but i agree with his statement (Score:1)
Re:System design (Score:1)
I agree though that apple may have screwed up going to the PPC but any apple customer should have been aware that apple has a track record of dumping there existing base of users as soon as a great new technology comes along.
Its sort of interesting to note though that Transmeta is actually swinging the 'CISC' is better than 'RISC' argument around again. Even though every one claims that the P6 is a 'RISC' arch its really not true in the real sense of the word. Original CISC machines had 'translation' units that translated the ISA to an internal micro-OP execution engine. This is actually what is happening on nearly all processors now. Most of the 'RISC' processors acquired a number of 'CISC'y like instructions in the mid '90s which are turning out to be incredibly difficult to completely implement with hardwired SI and still maintain high clock rates. So instead they break these instructions up into more basic operations and feed them though the pipelines. Originally 'RISC' was designed to do LESS work per cycle and but have a higher clock rate and therefore get more work done.
Re:Apple and Motorola (Score:1)
How many servers need altivec? none im sure a 8 cpu power4 box would out do any altivec desktop.
gee... what an easy solution, now send me the $10,000 for it.
"If only Apple could be persuaded to use these..." (Score:5)
To quote Paul's response [aceshardware.com]:
Maybe another way of looking at it is perhaps the price of four POWER4 known good die and the ceramic substrate and metal carrier totals $3000 (although I suspect that a tested and 100% functional ceramic substrate itself might approach or exceed $3000 in cost).
The real question is the cost of a fully assembled and tested, 100% functional, POWER4 8-way module? After all what are the chances one of these can be reworked if even just one of the 20,000+ solder ball joints was bad?
So for one of these 8-way on a chip jobs (unsure if they'll be offering 4-way configurations too or if those were just a prototype) it's looking like upwards of $10,000 just for IBM to fab, package, and test the darn things. Add in a system capable of feeding it the tremendous bandwidth it requires to run up to its full potential--8 GB/s to DRAM and a phenomenal 84 (!) GB/s I/O--and...ok, so I know Hemos was just joking when he made that comment about Apple, but you get the idea. These are MPUs you use to fold proteins and run gigantic dynamic-content websites, not surf the web and edit the home video of your kid's elementary school graduation.
On a related note, man these things oughtta show Intel a thing or two about how to marry clever instruction scheduling to brute-force functional units--forget about Itanium; it's gonna take a several-way McKinley system to even take a swing at this these. And it oughtta show Sun a thing or two about the dangers of resting on the laurels of your marketing success when designing new chips. And, as Paul notes in the article, it really oughtta make Alpha engineers worry that for the first time, having the most elegant design may not guarantee the best performance. Compaq has an 8-way SMT Alpha core on the way as well (EV8); too bad the Alpha group's customary position in the world--stepped on and neglected by their corporate masters--means they haven't got the money or manpower to bring it to market until well after POWER4.