Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
IBM

Design Philosophy of the IBM PowerPC 970 232

D.J. Hodge writes "Ars Technica has a very detailed article on the PowerPC 970 up that places the CPU in relation to other desktop CPU offerings, including the G4 and the P4. I think this gets at what IBM is doing: 'If the P4 takes a narrow and deep approach to performance and the G4e takes a wide and shallow approach, the 970's approach could be characterized as wide and deep. In other words, the 970 wants to have it both ways: an extremely wide execution core and a 16-stage (integer) pipeline that, while not as deep as the P4's, is nonetheless built for speed.'"
This discussion has been archived. No new comments can be posted.

Design Philosophy of the IBM PowerPC 970

Comments Filter:
  • What is: 2H03? (Score:2, Interesting)

    by Malc ( 1751 ) on Monday October 28, 2002 @05:32PM (#4551087)
    When will the "projected 2H03 release date" be? I'm not familiar with this term.
  • Question (Score:5, Interesting)

    by wcbrown ( 184278 ) on Monday October 28, 2002 @05:34PM (#4551105) Homepage
    What's the difference between the Power4 and the PowerPC 970? As a Mac guy, I've been following all of the rumors and announcements with interest but I keep seeing the PPC970 referred to as a scaled-back version of the Power4.

    Why wouldn't Apple go with the Power4 over the PPC970? And I already know that nothing official has been announced by Apple and that this is all probably going to be a lot of sturm und drang signifying nothing, but that's what keeps us Mac guys going I guess.
  • $$$/performance (Score:4, Interesting)

    by teamhasnoi ( 554944 ) <teamhasnoi AT yahoo DOT com> on Monday October 28, 2002 @05:36PM (#4551127) Journal
    That's what I want to see, and missing in the article and links. Anyone have an idea?

    I recall IBM's PPC boards going for over a grand, which is (to me) far too much. Especially when it was a 'G3' chip.

    Even if the new chip is faster, will I be able to buy 2 pentium 4's (5?) for the price of it?

  • by rsborg ( 111459 ) on Monday October 28, 2002 @05:41PM (#4551179) Homepage
    I think that since this is a 64 bit chip, why not compare it with other 64 bit consumer desktop chips (ie, AMD Clawhammer)? A lot of Intel's questionable moves (12K micro-ops instruction cache?) for the P4 were obviously not copied by AMD, and x86-64 seems to be the 64 bit desktop chip of the future.
  • Whoa (Score:5, Interesting)

    by MalleusEBHC ( 597600 ) on Monday October 28, 2002 @05:42PM (#4551190)
    The PowerPC 970 has other potential customers as well, though, not the least of which is IBM itself who, with its large investments in Linux, would love to see a high-performance, 970-based 4-way or 8-way SMP Linux desktop workstation halt the steady flow of former 64-bit *NIX workstation users who began switching to Wintel hardware in the late 90's.

    Before all my fellow Mac users start A) thinking about going to Linux B) drooling C) wondering about Darwin or D) some combination of the above, let me remind you that Darwin scales very well. You can now return to your previous state of awe.

    PS - How much you want to bet good ol Steve is already having wet dreams about doing the traditional Photoshop test at a Macworld with 4-way SMP?
  • by twoslice ( 457793 ) on Monday October 28, 2002 @06:18PM (#4551448)
    In football (the oblong ball not the round one) going wide and deep is sometimes referred to as the hail mary pass. Just like a hail mary pass, If IBM succeeds it will look brilliant, otherwise it will look like a futile attempt to score a winning TD in the war of the processors.
  • by IIci ( 611404 ) on Monday October 28, 2002 @06:54PM (#4551749)
    Looking purely at die size, one can expect manufacturing costs of the p4 and ppc970 as being roughly equal: PowerPC 970 1.8 GHz, 0.13um, 121 mm2, 52 million transitors Pentium 4 2.8 GHz, 0.13um, 131 mm2, 55 million transistors As long as IBM is not using the exotic materials of the power4, then the main advantage Intel has for pricing is that their R&D can be spread over many more chips. What the R&D costs of the ppc 970 are is interesting, especially since IBM is trying to position themselves as a maker of custom chips leveraging their ppc ISA and the experience gained through their big bucks power series.
  • Actually not that much of a joke... when I moved into my new house I noticed that the area under my desk stays noticably warmer than the rest of the room and my feet were kept downright toasty if I propped them up on top of my Athlon XP 1800 tower... I think the GeForce 4 and the 400w power supply with all the fans piping heat outta the back add to the effect also. :)
  • by KewlPC ( 245768 ) on Monday October 28, 2002 @07:01PM (#4551797) Homepage Journal
    *Sigh* The size of a CPU's address/data bus does not reflect a processor's "bitness". 64-bit means that it has a 64-bit word size (as opposed to the 32-bit word size on x86 processors), 64-bit registers, etc. Most 64-bit CPUs don't actually have a 64-bit address bus. Like you said, the Alpha's is 48-bit. This is usually done to keep the pin count down to a sane level (if you need all the physical RAM that a 64-bit address bus would provide, you need something bigger than a desktop CPU). You can expect that as 64-bit chips become more common on the desktop to see somebody introduce a 64-bit CPU that has a 64-bit address bus just so that they can say, "Hey, look, we have a 64-bit address bus, the other guys only have a 48-bit one!" and (like you are trying to do) will insinuate that this means the competition's CPUs aren't "true" 64-bit (even though they are).

    I dunno 'bout Macs (I don't know the M68k's "bitness"), but Intel introduced the 386 (their first 32-bit CPU) in 1986. And I certainly don't think the M68k was a RISC processor.

    at current prices and projected prices, 512 gigabytes or RAM will barely cost more than a couple of the fastest processors of this type.

    Really? I would LOVE to be able to buy 512 gigabytes of RAM for the cost of a couple of fast desktop processors. Don't forget that the PowerPC 970 is meant to be a desktop processor.
  • by dke ( 608042 ) on Monday October 28, 2002 @07:05PM (#4551825) Homepage
    This is a little like EPIC; except that bundling 5 instructions (IOP's), is bigger than 3.

    And VLIW (even EPIC) requires the compiler to adapt the code to the hardware. The 970 uses hardware to adapt the software to match it's needs. A next generation version could have 10 units, and the software wouldn't be optimized for 3 instruction bundles (like EPIC), etc....

    Also, while there are more int units in the P4, my understanding is they are not all full units. And in the 970 case, you were ignoring that the vector unit is a better int unit for most things than the Int unit (and it has 2 more ALU's, that are actually up to 16 separate ALU's each).

    The point is that if I'm doing int calculations or moves, then the AltiVec unit is much better than the Int units in the P4. So the only thing the 2 int units in the 970 should be used for is address calculations, branches and a few scalars that don't fit the vectors.
  • by stripes ( 3681 ) on Monday October 28, 2002 @07:10PM (#4551855) Homepage Journal
    More importantly, why would ANYBODY want to implement the x86 ISA (Instruction Set Architecture or smtn like that). It's the most horrid instruction set in use today.

    Because that is where most of the desktop CPU money is going, some of the high end, and frighteningly enough a fair bit of embedded CPU money too.

    In short if you can navigate the patent mine field, the brutal competition mine field, and deal with the instruction set making things a royal bitch doing an x86 CPU is a total no-brainer.

    In Pentiums and Athlons, the instruction set isn't really emulated. It's translated to a smaller instruction set (uops, iops, pick whatever term you like and run with it). However, these smaller sets are still made pretty much specifically to cover the overlying ISA (x86 in this case).

    Other then needing a whole new decoding front end, and being forced to use a trace cache because decoding multiple instructions in x86 land is very hard... the instruction thing isn't a big deal. Handling the odd-ball 80 bit FP format is. So is emulating all of the trap stuff and the other little odd bits close to the instructions set (like the MMU).

    A big pain. But with much of the effort not being where folks think it is!

  • Speculate... (Score:3, Interesting)

    by skinlayers ( 621258 ) on Monday October 28, 2002 @08:36PM (#4552400)
    What I really want to know is how much this chip is going to cost. If its cheap for Apple to put 2 or 4 of these in a machine, then how much will it matter that an expensive P4 (P5) out performs it? Hmmm.... The current Wind-Tunnel G4s raised a few eyebrows when it first came out do to the new case design. These things were designed to disapate heat! A HUGE (7 lbs) heat sink w/ matching fan, a small case fan, 2 fans on the power supply, and a ton of ventalation in the back. WAY more cooling that those 2 little G4s require. I think Apple is trying to avoid the fiasco it had with the Sawtooth (1st gen) G4s where they just slapped a G4 onto a G3 mobo. This time around, I believe they're releasing a new mobo first and then put a new proc in it down the road. I've also read stuff in forums suggesting that the power supply for the Wind-Tunnel had way more juice than the system currently demands. Can anyone out there do the math on this? We know how much power the PPC 970 eats. Can we figure out how much heat the Wind-Tunnel case is designed to disapate? What about how much power the power supply is putting? With these numbers, can we figure out how many PPC 970 the Wind-Tunnel case could power and cool? I've been suffering with a 266MHz G3 iMac, and I refuse to upgrade until Apple comes out with a system that really is worth that premium they charge, and a G4 is not it.
  • by bmajik ( 96670 ) <matt@mattevans.org> on Monday October 28, 2002 @08:41PM (#4552432) Homepage Journal
    asynchronus logic might mitigate clock propagation, and have added benefit of avoiding latch delay
    (which becomes a more and more significant overall factor as k and f go up...)

    i think i saw at least one cpu where latch delay per stage was equivalent to stage-execute time.

    the biggest reason im skeptical of deep pipelines is that they suck unless the instruction mix is hand picked. in nature, 30% of instructions are branches. compilers can only do so much to lengthen the basic block, and predictors can only be right "most" of the time.

  • by Anonymous Coward on Monday October 28, 2002 @09:07PM (#4552549)
    IT IS ONLY 40 BITS not 64

    Your desire to use address pins (or is it max pinned space per process?) to measure size puts you in a distinct minority. That doesn't make you wrong. But neither does it help make you right in this particular jungle.

    Systems whose physical addressing match their claimed "bitness" are probably in the minority.
    Some systems provide more physical addressing than register width (later PDP-11s, 8086, S/390), some less (68000, classic CDCs, early POWER). The 970 falls into the less category. Nothing unusual there.

    Apple, like EVERY OTHER OS KNOWN, will steal a bit or two

    Some bits come from physical addresses, some from virtual addresses. These should be addressed [pun slipped in, sorry] separately. AIX, btw, steals less than one bit. Linux can also be configured to steal less than one bit. (Assertions I can get away with no loss of credibility, since AC's have none to start with.) Were you frightened by a VAX in your formative years?

    Why do fanboys mod stuff like this down?

    Because we can't figure out why someone who needs 512GB, or 1TB, or more (which is it?) cares that a Linux process is limited to 1GB and not 2GB or 4GB.
  • Re:All this talk... (Score:2, Interesting)

    by gnuadam ( 612852 ) on Monday October 28, 2002 @09:44PM (#4552748) Journal

    0. When I say I want openMP added to gcc, that sort of implies that I want the compiler directives added, and the library routines added to the standard libraries that are shipped with the compiler. I realize gcc isn't going to affect any environment variables.

    1. I write scientific code. That's how I know openmp. I think it's great for that.

    2. It's not really overkill, because it's quite easy to program for (in a portable way!) Much easier than fork(), and IMHO easier than pthreads.

    3. Admittedly, I've never needed much in the way of IPC for the codes I've written for smp (scientific codes on smp machines don't need much of it) but you could probably use a pipe, a socket, or whatever else. I do mostly want it for the scientific uses, though.

    4. MPI is good. But openMP is like 1000x easier to write for, and on good hardware is usually better ... for my problems.

    5. If IBM, or apple even, makes affordable, good smp boxes with this processor, openmp would be quite useful.

    6. The same features it's good at in science ought to make it perfect for other processor intensive tasks. Anything that needs a for (do) loop can be scaled quite well. Anything that has chunks that don't need communication can be scaled well as well. I imagine video/sound encoding might be easily parallelized using this...much easier than pthreads, and for this mpi would be overkill, don't you think?

    7. All I'm saying is that openMP is out there, is supported on most commercial compilers, and it's noticably missing from gcc. *I* would use it. I suspect many more people would use it if it was available .... especially if 4 processor boxes become available on the commodity market as people seem to be expecting (dreaming) in this topic.

  • by Anonymous Coward on Monday October 28, 2002 @10:49PM (#4553018)
    probably closer to $400 for the top of the cutting edge at the begining, with the 1.2g versions being around $270 (lower misfab rates). Prices will go down as this is IBM's competition to "comodity x86" hardware they are forced to use on their blades currently, I wouldn't doubt but within 10 years IBM and AMD will be fighting a price war over the desktop.

    Intel seems happy to sit out this round, I'm wondering if they are going to be there next round though.

Today is a good day for information-gathering. Read someone else's mail file.

Working...