Design Philosophy of the IBM PowerPC 970 232
D.J. Hodge writes "Ars Technica has a very detailed article on the PowerPC 970 up that places the CPU in relation to other desktop CPU offerings, including the G4 and the P4. I think this gets at what IBM is doing: 'If the P4 takes a narrow and deep approach to performance and the G4e takes a wide and shallow approach, the 970's approach could be characterized as wide and deep. In other words, the 970 wants to have it both ways: an extremely wide execution core and a 16-stage (integer) pipeline that, while not as deep as the P4's, is nonetheless built for speed.'"
What is: 2H03? (Score:2, Interesting)
Re:What is: 2H03? (Score:2, Informative)
Re:What is: 2H03? (Score:5, Funny)
03 = 2003
In the case of Blizzard that means Fall 2005.
Re:What is: 2H03? (Score:3, Funny)
Re:What is: 2H03? (Score:1)
Re:What is: 2H03? (Score:3, Informative)
Re:What is: 2H03? (Score:2, Funny)
It's got nothing on the Pentium 4 (Score:5, Funny)
tcd004
IBM Chip (Score:2, Troll)
Re:IBM Chip (Score:1)
Re:IBM Chip (Score:2)
It's very likely that Apple intend to use it in future high end boxes. (They would be rather stupid not to.) Unless Motorola can one-up IBM by then.
I don't think it has all that much to do with Apples future though. Apple boxes have not really been about high performance lately, and people buy them anyways. Mostly becuase they like the OS/apps and such I assume.
And the specs are pretty much solid now I assume. Creating hardware is
Question (Score:5, Interesting)
Why wouldn't Apple go with the Power4 over the PPC970? And I already know that nothing official has been announced by Apple and that this is all probably going to be a lot of sturm und drang signifying nothing, but that's what keeps us Mac guys going I guess.
Re:Question (Score:3, Informative)
While the Power4 core has two processor cores and massive caches for MP implementations, the PowerPC 970 has only one processor core, an SIMD unit and a 512K on-die L2 cache. The cache includes error correction. The PowerPC 970, as described today, has no connectors for an L3 cache.
Re:Question (Score:5, Informative)
This implies that the Power4 is not 64 bit -- which is of course wrong.
I would say that the PowerPC 970 trades the second core and fancier interconnects of the Power4 for lower power, cost, and the SIMD unit.
Re:Question (Score:2)
Re:Question (Score:2, Informative)
Re:Question (Score:2)
As long as you don't compile with -arch=pwr4 (etc), xlc (IBM's C compiler) will compile the code to run on any Power or PowerPC.
Re:Question (Score:2)
Re:Question IT IS ONLY 40 BITS not 64. (Score:5, Interesting)
Your desire to use address pins (or is it max pinned space per process?) to measure size puts you in a distinct minority. That doesn't make you wrong. But neither does it help make you right in this particular jungle.
Systems whose physical addressing match their claimed "bitness" are probably in the minority.
Some systems provide more physical addressing than register width (later PDP-11s, 8086, S/390), some less (68000, classic CDCs, early POWER). The 970 falls into the less category. Nothing unusual there.
Apple, like EVERY OTHER OS KNOWN, will steal a bit or two
Some bits come from physical addresses, some from virtual addresses. These should be addressed [pun slipped in, sorry] separately. AIX, btw, steals less than one bit. Linux can also be configured to steal less than one bit. (Assertions I can get away with no loss of credibility, since AC's have none to start with.) Were you frightened by a VAX in your formative years?
Why do fanboys mod stuff like this down?
Because we can't figure out why someone who needs 512GB, or 1TB, or more (which is it?) cares that a Linux process is limited to 1GB and not 2GB or 4GB.
Re:Question IT IS ONLY 40 BITS not 64. (Score:4, Informative)
And the 40 bit address bus is most likely a pin packaging limitation. They did not see a need to bring those extra 24 address lines out to the chip package. Internally, it is 64 bit. Much like the venerable MC68000 was 24 bit externally, but 32 bit internally.
But seriously, in the life-span of THIS processor implementation - do you seriously see ANY desktop manufacturer even thinking about putting that much RAM in their CPUs?? Heck 1GB of RAM is not 'standard' yet. Extrapolating w/ Moore's Law, we'll be approaching 40bits in 8 years. Apple will undoubtedly have another chip before then!
If you truly need THAT much physical storage today, you'll need to shell out for a SERIOUSLY large server. IBM's high-end p690 currently maxes out at 256GB. The virtual address space is undoubtedly much higher.
Tom
Using Big RAM in the Real World (Score:2)
Almost two decades ago, I was supporting a network modelling application that needed 12MB, and our computer was a VAX with 4MB of RAM, but fortunately 4.1BSD and System 5.2p were able to give it enough virtual memory to scrape along. Our typical runs took about a week, until a couple years later when the price of RAM dropped to the point that we could afford to upgrade to 16MB, at which time it dropped to an hour per run.
One thing we found out was that as you approach the limits of a machine's capacity, all the details of the architecture that you were able to ignore on smaller problems become visible, like how the TLBs work and what the memory page sizes are; things became somewhat clunky at 6MB and much more clunky at 22MB (or maybe 24MB), and some of that may have been the OS rather than the hardware. The extra 12MB RAM cost us approximately one person-year's salary, but unfortunately corporate accounting rules made it much harder to buy capital equipment than to make our study take a year longer.
At this point, I'm supposed to include the obligatory old-geezer rant about walking to the mainframe, five miles in the snow, uphill both ways, carrying punch-cards and hand-winding magtape, but that was back when I was an undergrad, and we didn't actually go to the mainframe, just the card-reader/printer/keypunch room which was half a mile away (still snow and hills), and the reason I handwound the magtape was because the professor's only copy had a cracked reel, though we did use hand-cranked papertape tools to do it, and I really _did_ wind papertapes by hand :-)
I haven't dealt with these problems lately - Moore's Law has long passed the limits of any problems I solve in practice, laptop diskdrive sizes have gotten two jumps ahead of Microsoft bloatware (though fitting backups onto CD-Rs feels a lot like fitting them onto floppies used to), and the only practical application I've got that could think about pushing the terabyte disk boundary is Tivo, if I'm willing to devote that much resources to something that makes me watch more television... The reason it took so long to get the 120GB drive on my 233MHz Pentium wasn't cost, it was BIOS upgrades
Re:No connectors for L3 cache? (Score:2)
Re:Question (Score:2, Informative)
Re:Power4 vs PowerPC 970 (Score:5, Informative)
The first generations (601, 603/604 and the ?aborted? 620) of the PowerPC line were scaled-back versions of the Power and Power2 architectures respectively [the original Power architecture was mounted on a 3x5 daughter card with 4-5 separate chips [I'll have to go looking for my tech papers] making-up the core
Additionally, IBM has tended to work-out new capabilities -- such as the move to 64-bit and dual cores -- on the larger scale Power architecture, before attempting to stuff it into the smaller PowerPC pacakge [besides, IBM has to keep something to distinguish its pricier iron from the OEMs.
Natty
Re:Power4 vs PowerPC 970 (Score:2)
The first generations (601, 603/604 and the ?aborted? 620) of the PowerPC line were scaled-back versions of the Power and Power2 architectures respectively [the original Power architecture was mounted on a 3x5 daughter card with 4-5 separate chips [I'll have to go looking for my tech papers] making-up the core ... because of this the migration of everything into one die for the PowerPC was amazing.
The PowerPC 601 was not a scaled back version of the Power series. To say this would imply that they took the design and modified it. In fact, they took the Power instruction set, modified that and then designed the processor to support it for the target markets.
The 64 bit PowerPC 620 was not "aborted" per se (like the PowerPC 615 was), rather IBM decided that its role was filled by the higher clocked 604 series and the then soon-to-come IBM Rochester, MN designed 64 bit PowerPC 630 (aka Power 3).
To verify my claim that the PowerPC 620 was not aborted, Motorola got suckered into manufacturing [kegel.com] them for Bull [club-internet.fr].
Re:Question (Score:4, Informative)
Plus, the Power4 is really designed as a server/Big Iron chip - it's really 2 CPUs on 1 die - and that's just not what an iMac needs.
Re:Question (Score:2)
Re:Question (Score:5, Informative)
It is HUGE [com.com].
The picture at the top right shows the Power4 multichip module as used in the p690. Yes, it is the 5" square thing in the guys hand.
There are better pictures of the MCM itself, but I couldn't find the close-up showing just the MCM in someone's hand.
The large size (along with everything it entails: it uses 125W power, and supposedly costs about $3500 to manufacture) is one indication that IBM designed the Power4 for its big-iron. Nevermind that IBM does offer the Power4 (sans MCM) in some of their smaller servers.
The PowerPC970 is the equivalent processor tweaked for the desktop/low-end servers.
Re:Question (Score:4, Informative)
The power4 costs 4-5k per cpu. Obviously too expensive for desktop systems because of high end server features and very large caches in the chip that will offer no performance benefit to desktop apps. Only heavily threaded multitasking apps running in parrallel will see the performance improvments by a power4. A web server running servlets and databases are the examples I refer to as heavily threaded multitasking applications. Adobe photoshop will show little performance difference and may even run slower on a power4 vs a powerpc 970 due to the lack of simd instructions.
IBM did good with this processor and its leaps and bounds ahead of the g4. The main limitation of the g4 is the lack of ddr memory support. In ddr macs the chipset has to slow down memory access to the cpu to 133mhz speeds and it creates a very serious bottleneck. This alone is bottlneckintgthe processor down to half its potential in +1 ghz processors. Expect a %200-300 performance increase with these new processors.
Re:Question (Score:2)
They have a slightly divergent instruction set, but considering neither actually perform the instructions they "support" that's not anything worth noting.
An overview of pipelining (Score:3, Informative)
Basically, all modern processors are pipelined. This means that they execute various instructions at the same time. Whereas doing a load of wash, waiting for it to finish, putting it into the dryer, waiting to finish, and then folding would take 30 minutes * 3 steps * 3 loads = 4.5 hours, one could PIPELINE such a process, thus removing sequentialism and doing the first load, then while that's drying put the second load into the washer, and so on
This is all a processor really does. It does a FETCH, an INSTRUCTION DECODE, then an EXECUTION, then perhaps a MEMORY READ/WRITE, and then a WRITE BACK, perhaps. So this 16-stage pipeline can have 16 different instructions executed all at the same time, but just in different points of its execution. The example in CAPS above is a 5-stage pipeline that's similar to those in MIPS processors.
Hope this was helpful!
Re:An overview of pipelining (Score:3, Funny)
Re:An overview of pipelining (Score:2)
It wasn't. You only explained "deep", not "wide". Since both words occured in the sentence you were trying to explain, you should try to do better than only explaining the one you remembered from college...
Re:An overview of pipelining (Score:1)
For a decent intro to pipelining, Google for design notes for the MC6502, the first pipelined processor.
$$$/performance (Score:4, Interesting)
I recall IBM's PPC boards going for over a grand, which is (to me) far too much. Especially when it was a 'G3' chip.
Even if the new chip is faster, will I be able to buy 2 pentium 4's (5?) for the price of it?
What? (Score:2)
-jason m
Re:$$$/performance (Score:1, Insightful)
Tom
Die size as an indicator of cost (Score:2, Interesting)
Every time I readone of these articles... (Score:4, Insightful)
Kudos to the Ars team yet again for going deeper into CPU designs than 99% of the populace need to go :)
Re:Every time I readone of these articles... (Score:2)
lol... did anyone else read an 'e' too many here ? Surely smells of goatse that way
Re:Every time I readone of these articles... (Score:2)
by Hennesy-Patterson.
A really good book, not too hard to read..
Comparison without AMD? (Score:4, Interesting)
Re:Comparison without AMD? (Score:1)
Re:Comparison without AMD? (Score:4, Informative)
Decoded cache in the P4. (Score:3, Insightful)
The P4 has its flaws, but IMO cacheing decoded instructions isn't one of them. It shortens the pipeline, and paves the way for a true trace cache (cache of decoded basic blocks indexed by entry point; very handy for renaming and scheduling).
Whoa (Score:5, Interesting)
Before all my fellow Mac users start A) thinking about going to Linux B) drooling C) wondering about Darwin or D) some combination of the above, let me remind you that Darwin scales very well. You can now return to your previous state of awe.
PS - How much you want to bet good ol Steve is already having wet dreams about doing the traditional Photoshop test at a Macworld with 4-way SMP?
PPC, not just for Apple any more (Score:5, Insightful)
This chip could be the start of something big in the Linux space as well. Think about it, we are now at a point where a few companies other than Intel are now poised to take the center stage in the next gen workstation, most notably AMD, Apple, and now IBM themselves.
While Linux has run on PPC chips for a long time, it is difficult to come upon a G4 chip without paying the "Apple Tax" for the hardware. If IBM steps up to the plate with this chip, which can then run OS X, Mach, Linux, *BSD, (insert other OS'es here), and can be purchased directly or in a package from IBM, we may see a good set of Windows challengers for the desktop and server room. Obviously OS X will still only run on Apple derivatives.
These chips will be big, I guarantee it, and not just for Apple. It will be interesting to see if Microsoft ports Win XP to these chips.
Re:PPC, not just for Apple any more (Score:3, Informative)
Re:PPC, not just for Apple any more (Score:2)
Don't guarantee it 'till you see the prices for the chips. The G4 is pretty expensive.
Re:PPC, not just for Apple any more (Score:2, Flamebait)
Hey, I'll be delighted if it really happens, but that sounds like the usual "BIG NEWS in two weeks" type of stuff from my Amiga days. I can't believe it anymore, and if I tried, I would go insane. Mock my unrealistic fanboy idealism once, shame on you. Mock it twice, shame on me.
Re:PPC, not just for Apple any more (Score:3, Insightful)
Bear in mind that when IBM says "desktop workstation" they mean a $20k+ machine. Consumer desktop machines these aren't.
built for speed (Score:2, Informative)
A deep pipeline has as much to do with speed as the number of characters in the processors name. Deep pipelines allow for higher MHz. That's all there's about it. Granted, for two processors of the same architacture, the one with higher MHz is faster. But you can't claim a CPU to be built for speed solely based on it's pipeline depth.
UUUUUUUUUH I hate explaining this shit, but my god
Re:built for speed (Score:3, Funny)
Yeah, sorry.
Re:built for speed (Score:5, Informative)
If you're interested, take a look at the following documents (you might wanna check the urls for spaces):
http://systems.cs.colorado.edu/ISCA2002/FinalPa
http://systems.cs.colorado.edu/ISCA2002/FinalPa
http://systems.cs.colorado.edu/ISCA2002/FinalPa
very deep pipelines (Score:3, Informative)
If you're interested, take a look at the following documents (you might wanna check the urls for spaces):
What makes me leery of taking these results at face value is that the performance peak is very broad (i.e. incremental benefit is low beyond a certain point), while the first paper, at least, seemed to gloss over a few important concerns (keeping clock skew and jitter very low when distributing to that many more stages, increasing overhead from the bypassing network, etc).
Still an interesting set of papers, but I'm not (yet) convinced.
Re:very deep pipelines (Score:3, Interesting)
(which becomes a more and more significant overall factor as k and f go up...)
i think i saw at least one cpu where latch delay per stage was equivalent to stage-execute time.
the biggest reason im skeptical of deep pipelines is that they suck unless the instruction mix is hand picked. in nature, 30% of instructions are branches. compilers can only do so much to lengthen the basic block, and predictors can only be right "most" of the time.
Holy B-Jesus! (Score:4, Funny)
Understanding 64-Bit Processing (Score:4, Informative)
Re:Understanding 64-Bit Processing (Score:1)
No one will ever need more than that.
Re:Understanding 64-Bit Processing (Score:2)
of silicon til the end of time, right?
Re:Understanding 64-Bit Processing (Score:2)
Re:Understanding 64-Bit Processing (Score:4, Funny)
wide / transistors (Score:5, Insightful)
it operates very much like itanium, w.r.t to group bundling / dispatch of IOPs. Very much like itaniums 3-bundle EPIC codes, but itanium requires the compiler to best pack the templates, whereas the 970 builds each bundle based on dependencies. funny how they both punt with nops.
i wouldn't be surprised if it has similar int performance as itanium, but better fp. i would expect integer to be better on p4 by the time the 970 hits the streets in systems. intel keeps increasing the number of int execution ports.
although, i wouldn't call it wide-and-deep -- it has distinct vector and FP units, but they are rarely used simulataneously, so i suspect they will be idle most of the time. there's only 2 integer units compared to p4's 4.5. so if most of the time the execution engines are idle, it isn't really 'wide'.
goes to show how bulky x86 decode is -- 970 has fewer transistors, but the same cache size, more branch resources, and more Vector/FP hardware than p4!!
now watch the bias on
Re:wide / transistors (Score:5, Interesting)
And VLIW (even EPIC) requires the compiler to adapt the code to the hardware. The 970 uses hardware to adapt the software to match it's needs. A next generation version could have 10 units, and the software wouldn't be optimized for 3 instruction bundles (like EPIC), etc....
Also, while there are more int units in the P4, my understanding is they are not all full units. And in the 970 case, you were ignoring that the vector unit is a better int unit for most things than the Int unit (and it has 2 more ALU's, that are actually up to 16 separate ALU's each).
The point is that if I'm doing int calculations or moves, then the AltiVec unit is much better than the Int units in the P4. So the only thing the 2 int units in the 970 should be used for is address calculations, branches and a few scalars that don't fit the vectors.
Re:wide / transistors (Score:2)
Ah, I see. I was associating "vector" with packed floating point, and completely forgot about vector integer.
But doesn't that depend on the compiler? Would it dispatch any integer op to the vector engine, or is there a way for the 970 to know if an instruction is an address calculation via a special opcode? Otherwise you again have the problem of compiler dependency like Itanium.
Re:wide / transistors (Score:5, Informative)
This is a common misconception, probably stemming from some early coverage of the Power4 by Keith Dieffendorff(sp?) where he for some unknown reason called the Power4's instruction bundles "VLIW-like". The problem is that the bundles are strictly in program order. There is no dependency checking or code scheduling that goes on in the building of these bundles. They're built along completely different rules than bundles in a VLIW machine.
All the out-of-order stuff happens in the back end, in the scheduling queues, just like on any other non-VLIW processor.
Wide, Deep and Built for Speed (Score:3, Funny)
Wide and Deep - Hail Mary pass perhaps? (Score:2, Interesting)
if.... (Score:1)
G5? (Score:2)
Re:G5? (Score:3, Informative)
http://e-www.motorola.com/webapp/sps/site/prod_
Re:G5? (Score:2)
Finally, my girlfriend can be happy (Score:3, Funny)
For those /.ers who will not read the article (Score:5, Insightful)
For those not planning to read the article, I wanted to mention the following so you do not get the wrong impression. The speed that the article refers to (of a long integer pipeline, like a 16-stage or like the Pentium IV's 20-stage) is clockspeed, not necessarily actual performance. The P4's super long pipeline, for example, allows it to run at higher clock speeds, but less work gets done in the same number of clock cycles. This is the "braniac" vs "speed demon" philosophy (with a high clock speed but low instructions-per-clock representing "speed demon") and neither is necessarily better than the other (though one is obviously better for the marketing dept.)
Just don't assume that "built for speed" always means "built to be fast" -- a confusing but important distinction.
Speculate... (Score:3, Interesting)
Re:They should make it work three ways (Score:5, Insightful)
Intel and AMD have the x86 market pretty well locked down.
More importantly, why would ANYBODY want to implement the x86 ISA (Instruction Set Architecture or smtn like that). It's the most horrid instruction set in use today.
Some instruction sets can't really be mapped to others easily, and optimizing for good performance with PPC would probably not have good x86 performance anyways.
In Pentiums and Athlons, the instruction set isn't really emulated. It's translated to a smaller instruction set (uops, iops, pick whatever term you like and run with it). However, these smaller sets are still made pretty much specifically to cover the overlying ISA (x86 in this case).
Re:They should make it work three ways (Score:3, Interesting)
Because that is where most of the desktop CPU money is going, some of the high end, and frighteningly enough a fair bit of embedded CPU money too.
In short if you can navigate the patent mine field, the brutal competition mine field, and deal with the instruction set making things a royal bitch doing an x86 CPU is a total no-brainer.
Other then needing a whole new decoding front end, and being forced to use a trace cache because decoding multiple instructions in x86 land is very hard... the instruction thing isn't a big deal. Handling the odd-ball 80 bit FP format is. So is emulating all of the trap stuff and the other little odd bits close to the instructions set (like the MMU).
A big pain. But with much of the effort not being where folks think it is!
Re:They should make it work three ways (Score:2)
Re:They should make it work three ways (Score:3, Informative)
The x86 world seemed to move faster than the design for this and it fell away. It made more sense to concentrate on PPC stuff rather than try to do PPC and changing x86 stuff. Also, if it ran x86, why should anyone bother to write for PPC?
The difference is Pentiums and Athlons are intended to be x86 family upgrades, while the PPC is not. The PPC 970 is meant as an upgrade to earlier PPCs. One could as well ask why AMD doesn't make an Athlon that can run PPC code.
Re:They should make it work three ways (Score:2)
Puto
Re:unfotunatly Apple is going with Intel instead.. (Score:5, Insightful)
I think Apple will stick with a company that it knows, IBM, since they have been working together for years. It doesn't seem that Apple will just jump ship to the x86 platform. This would also mean redoing the Mac OS X code and optimization (not like they will have to do some anyway, but they will have to do more). It is highly unlike that Apple will go with a heat producing, energy wasting x86 Intel chip.
They wouldn't have to redo anything... (Score:2, Insightful)
Are they going to jump ship to x86? Not likely if they can help it... but they're keeping the option open. Kind of like how Dr. Evil KNOWS his plans will never fail, but he always has that Big Boy rocket hidden in the back--just in case.
And who ya calling energy wasting? My Palomino keeps my room nice and toasty on those lonely nights and makes great julienne fries!
Re:They wouldn't have to redo anything... (Score:1)
Re:They wouldn't have to redo anything... (Score:2, Interesting)
Apple selects a winner! (Score:3, Funny)
Re:unfotunatly Apple is going with Intel instead.. (Score:5, Insightful)
Re:unfotunatly Apple is going with Intel instead.. (Score:2)
Re:unfotunatly Apple is going with Intel instead.. (Score:2)
I said "competitive", but I meant in performance terms. Price competitiveness is a whole different issue, though as another poster hinted, price/performance might not be as bad as one would think, esp. if IBM isn't plagued by the same fab issues Motorola has come across with the G4 (and history has shown that they probably won't be).
Re:unfotunatly Apple is going with Intel instead.. (Score:2)
Mac OS X binaries are capable of being FAT. They can contain both PPC and Intel instructions. OS X binaries are actually folders which can contain multiple architecture implementations within them. This also applies to the "libraries". Ok, the Framework bundles.
Re:unfotunatly Apple is going with Intel instead.. (Score:2)
Re:unfotunatly Apple is going with Intel instead.. (Score:2)
Go count up the PPC registers.. then the registers in the x86 architecture...
With JIT "acceptable" speed would be possible, but I wouldn't expect more than 601-66 performance from a high clocked P4 running interpretive PPC emulation..
Re:unfotunatly Apple is going with Intel instead.. (Score:4, Insightful)
Re:unfotunatly Apple is going with Intel instead.. (Score:2)
I stared at him and just said "Yes...yes it will"
--Joey
Re:unfotunatly Apple is going with Intel instead.. (Score:2, Insightful)
P.S. My 333Mhz P2 runs great at 400Mhz, but largely because Intel underclocked the identical core to run at 333Mhz. Intel plays up the importance of clock speed, so they do dumb things like underclocking, and multiplyer locking.
Re:I am tired of 64 bit lies ! (40 bit only, folks (Score:3, Interesting)
I dunno 'bout Macs (I don't know the M68k's "bitness"), but Intel introduced the 386 (their first 32-bit CPU) in 1986. And I certainly don't think the M68k was a RISC processor.
at current prices and projected prices, 512 gigabytes or RAM will barely cost more than a couple of the fastest processors of this type.
Really? I would LOVE to be able to buy 512 gigabytes of RAM for the cost of a couple of fast desktop processors. Don't forget that the PowerPC 970 is meant to be a desktop processor.
Blah blah blah... (Score:2, Insightful)
And now that you mention it, I do remember reading that the M68k was 32-bit, but it only had a 24-bit address bus, which meant the max. amount of RAM it could physically have was 16 megabytes. Again, I'm not trying to bash anybody, I'm trying to point out that your "64-bit CPUs aren't really 64-bit because they only have a 48-bit address bus" argument is flawed.
Why you went off on the whole Apple vs. Wintel thing is beyond me, but if you want to play ball, OK. For the record, the PC wasn't meant to compete with the Lisa or the Mac, and both of those computers were introduced after the PC. The original IBM PC was a competitor to the Apple II, but more oriented towards business use rather than home use. If you remember, the Apple II also used a cassette tape drive (just like the original PC), but, like the PC's successors, the PC-XT and the PC-AT (all modern PCs are descended from the PC-AT), later had the ability to use floppy drives and hard drives.
The 1982 lisa had windows, scolling, dialogs, fonts, buttons, WYSIWYG text editing with graphics, etc.
Which were all "borrowed" from Xerox PARC. The fact that Apple later whined and bitched about Microsoft "borrowing" those ideas from the Lisa and/or Mac (when Apple themselves had stolen those concepts from somebody else in the first place) is too amusing for words. I can't stomach Bill Gates, but I have just as hard of a time putting up with Steve Jobs ("You stole Windows! It's not fair! We stole it first!").
And as for Mac OS always being 8 years ahead of Windows, well, I'm no lover of Windows, but Windows had preemptive multitasking years before Mac OS (Windows got it in Win95, Mac OS didn't have it until OS X).
the Apple II had 75% of us market
Although I don't have any hard facts, I have a hard time with this. It wasn't like IBM and Apple were the only players in the personal computer market. There was Commodore with their highly successful Commodore 64 computer (not to mention the Commodore PET, VIC-20, and Commodore 128), Sinclair, the TRS-80 (from Tandy and RadioShack IIRC), and a whole host of others.
Re:Blah blah blah... (Score:2, Insightful)
Windows NT/2000XP had real pre-emptive multitasking before the half bit crappy win95 implementation. 1993? Think NT 3.1 and Nt 3.51. Those OS's, much maligned then are now the foundation for MS future OS's...
Re:Blah blah blah... (Score:2)
Re:All this talk... (Score:5, Informative)
Now, openMP is good for programming extremely high-performance shared-memory applications, like scientific computation applications and stuff like that. It really sounds like overkill for a desktop environment where it's probably easier to program a multithreaded application with standard IPC mechanisms where communication is required. And really high-performance applications could also be programmed using MPI and a message passing communication scheme, which is far more widely used (compare the # of people who know about openmp versus those who know about mpi), probably wouldn't be much less efficient, and would quite likely scale much better than a shared memory implementation.
Re:All this talk... (Score:2, Interesting)
0. When I say I want openMP added to gcc, that sort of implies that I want the compiler directives added, and the library routines added to the standard libraries that are shipped with the compiler. I realize gcc isn't going to affect any environment variables.
1. I write scientific code. That's how I know openmp. I think it's great for that.
2. It's not really overkill, because it's quite easy to program for (in a portable way!) Much easier than fork(), and IMHO easier than pthreads.
3. Admittedly, I've never needed much in the way of IPC for the codes I've written for smp (scientific codes on smp machines don't need much of it) but you could probably use a pipe, a socket, or whatever else. I do mostly want it for the scientific uses, though.
4. MPI is good. But openMP is like 1000x easier to write for, and on good hardware is usually better ... for my problems.
5. If IBM, or apple even, makes affordable, good smp boxes with this processor, openmp would be quite useful.
6. The same features it's good at in science ought to make it perfect for other processor intensive tasks. Anything that needs a for (do) loop can be scaled quite well. Anything that has chunks that don't need communication can be scaled well as well. I imagine video/sound encoding might be easily parallelized using this...much easier than pthreads, and for this mpi would be overkill, don't you think?
7. All I'm saying is that openMP is out there, is supported on most commercial compilers, and it's noticably missing from gcc. *I* would use it. I suspect many more people would use it if it was available .... especially if 4 processor boxes become available on the commodity market as people seem to be expecting (dreaming) in this topic.