Inside the Itanium 135
vanguard writes: "Extreme Tech has a detailed overview of the Itanium. It's fairly long but it's worth your time if such things interest you."
Understanding is always the understanding of a smaller problem in relation to a bigger problem. -- P.D. Ouspensky
Better link (Score:3, Informative)
Re:Better link (Score:3, Interesting)
Of course, for some perspective on the nature of processor speculation, I point you to nearly any issue in Byte's print archive [byte.com].
Re:Better link (Score:2)
Woof! Someone help that guy get his GRE...
Re:Better link (Score:1)
No $64,000 question (Score:2)
I believe that you are mistaken. There are many people still living who remember it and you can buy video tape copies [cduniverse.com] of the show. This seams a little much for "an urban myth"
-- MarkusQ
Re:No $64,000 question (Score:1)
Partly right? (Score:1)
I fail to see how you conclude that I am only "partly right" when the link you posted mentions the existence of the very TV show that you initially claimed was "an urban legend". If you recall, nobody said that there wasn't a $64 question on the radio first; but I do claim that the existance of the radio show doesn't magically make the TV "an urban legend".
-- MarkusQ
328 registers!!! (Score:2)
Re:328 registers!!! (Score:1)
On the other hand, for a context switch (switching between threads or processes), yes, all registers will have to be saved.
Re:328 registers!!! (Score:2, Interesting)
It also had 2 16-bit index registers, a 16-bit stack pointer and a 16-bit program counter. Wich, of course, shouldn't be used for calculations.
So you could count the registers on your hand. Ye good ol' times.
Re:328 registers!!! (Score:2)
Kids these days...
Re:328 registers!!! (Score:1)
My 8 hamsters-on-wheels-pushing-buttons seem to be VERY interested in this "register" technology.
I wonder why....
PS- Yes, that was a Simpson's quote.
Re:328 registers!!! (Score:1)
Anyway, I'm glad to see other people out there remember the 'ole 6502. Just to clear things up a bit, the 6502 had 5 registers. You forgot the Processor status (P) register which held the test flags. On the other hand though there was no way to access this register directly, just through test instructions.
Re:328 registers!!! (Score:1)
Re:sounds awfully familiar.. (Score:1)
But usually is anyway.
Re:328 registers!!! (Score:2)
Some programmers do not divide my code in little functions to save some register swapping, but to make it easier to read, manage, and of course to code.
Although I can see how not having to fight for your registers is a big help at the assembler level, I would be very, very afraid of someone who codes a function/method/procedure with 300 variables "to take advantage of the registers", be it C, Assembly or whatever. Fear does not always imply respect.
I would think all those registers would be much more useful in context switches and/or pipelining issues.
Say, allocate, 30 registers to each context and keep 10 contexts running on the processor without much penalty. Prepare your variables for your next few operations for each conditional outcome on the registers. Stuff like that would be more useful (and less visible) in most cases.
Re:328 registers!!! (Score:1)
Say, allocate, 30 registers to each context and keep 10 contexts running on the processor without much penalty. Prepare your variables for your next few operations for each conditional outcome on the registers. Stuff like that would be more useful (and less visible) in most cases.
That said, the new AMIGA TAO OS has Virtual Registers (not that I've seen diddly squat in the way of programs for it).
Re:328 registers!!! (Score:2)
Re:328 registers!!! (Score:1)
Try before you buy (Score:3, Funny)
It was worth reading a long article to come across this nugget of wisdom. I think i'll embroider it on a sampler and hang it in my cube.
Re:Why this strange name ? (Score:2)
Uhh, no, try titanium, a material known for its phenomenal strength. Lop off the first letter, and you have a word that calls to mind the Internet, Intel, and titanium. Pretty slick name, if you ask me.
The sad thing is that titanium has gotten to the gimmick level lately. They use it in everything from sunglasses to golf balls to notebook cases, and it usually isn't even the right material for the job. It's just used for the "coolness" factor....
Re:Why this strange name ? (Score:2, Interesting)
The PowerPC G4 chip is a hot beast (nothing like the P3 or P4, but hot enough) that the titanium is used in part as a heat sink. Apple's application of the metal seems to get lots of oohs and ahhs as a result of this blend.
Re:Why this strange name ? (Score:2)
Now if it had been a rhodium or palladium Visa card, I might've thought of it differently...
Dell Launches 64-Bit Workstation (Score:2, Informative)
The price? $7,999 at the time.
Re:Dell Launches 64-Bit Workstation (Score:1)
Next time somebody complains about the Mac OS and their PowerPC RISC processors, I'm going to beat them over the head with this Dell box.
Re:Dell Launches 64-Bit Workstation (Score:1)
(Not really, but I like your idea and want to promote it.)
Also, so much for that whole MHz/GHz race.
Re:Dell Launches 64-Bit Workstation (Score:1)
http://www.theinquirer.net/15120102.htm
Re:Dell Launches 64-Bit Workstation (Score:1)
Re:Dell Launches 64-Bit Workstation (Score:1)
Why all the hubbub? (Score:2, Insightful)
What is so special about the Itanium other than the fact that it's from Intel? We've had 64-bit processors for years now. Moreover, it's not like everyone is going to ditch the IA32 architecture overnight and go to Itanium. It seems to me that anyone who needs/wants 64-bit computing already has it.
Re:Why all the hubbub? (Score:2)
Re:Why all the hubbub? (Score:1)
Re:Why all the hubbub? (Score:1)
Re:Why all the hubbub? (Score:2)
This is actually an interesting perception issue. Itanium has hardware support for x86 and software support for PA-RISC. The original article attributes this to a lesser priority being given to PA-RISC. While that may be true, it may also be due to the PA-RISC customers having less of a "real PA-RISC" hardware vs. software hangup than x86 customers.
Re:Why all the hubbub? (Score:1)
As far as I am aware, the IA-64 is the first instruction set to use explicit parallelism. This removes much work from the CPU when determining which instructions can be executed in parallel. I believe the IA-64 is also extensible--it is extremely easy to add more pipelines or make other significant changes to the architecture without a lot of redesign.
From the article:
An assembly program would call [4 instruction types] nuts. You'd think that Itanium's designers would have been satisfied with 241 different opcodes, but no? An assembly program would call it nuts. You'd think that Itanium's designers would have been satisfied with 241 different opcodes, but no?
The Intel engineer I spoke to about IA-64 said that it would be virtually impossible to write good assembly for IA-64 because humans don't think about explicit parallelism in their heads very well. Itanium/IA-64 relies heavily on good compilers.
Re:Why all the hubbub? (Score:2)
Itanium -- what's the point? (Score:1, Flamebait)
Is there ANYONE doing ANYTHING useful with them yet?
As far as I can tell, current Itanium stuff is a mere curiosity.
Intel is making an enormous gamble with IA-64, it is a huge investment and the whole thing may blow up in their face.
AMD is taking the 'safe road' -- bolting 64bit onto an existing design, taking advantage of the huge momentum of the x86 architecture.
Intel on the other hand is going to have to spend a huge amount of effort getting the Itanic moving.
Will be interesting to see how things develop. Will AMD's 64bit products derail Intel's efforts? Can they crack the 64bit server market fast enough and make IA-64 irrelevant? Stay tuned!!
Re: Bold and Italics, what's the point? (Score:2)
Well right there you have already invalidated the rest of your comment. Intel is not going to have to do anything to get their IA64 architecture moving.
As announced, HP's new PA-RISC chipset will support both McKinley and their RISC processor. Sounds crazy but HP is doing it.
Furthermore nobody and I mean nobody is developing anything for AMD's 64bit systems. On the other hand I have seen significant interest in McKinely. True enough, Merced is just considered a development platform and is pretty much dead in the water. I work for a test and measurement company and I deal directly with our front side bus solutions. You are wrong. AMD is facing the same problems in the IA64 arena that they are in the IA32 arena. Intel's chips say Intel and theirs say AMD. So everybody is ignoring them, for now.
This is not to say that AMD can't come in beat them. I am just saying that Intel's solutions are already in development in many of the high-end server labs. I haven't seen a single group working on AMD's 64bit proc, nor have I been asked if we plan on supporting it.
Re: Bold and Italics, what's the point? (Score:2)
> developing anything for AMD's 64bit systems.
It doesn't matter. AMD will sell plenty of Hammer chips in their usual markets, who will just use it as a faster 32bit x86 chip. (Just like the 386 was used as a faster 16bit chip when it came out.)
BUT one day someone will wake up and realize that Hammer makes for fast, cheap, cool and backwards compatible servers. And then Intel will release their secret x86-64-compatible CPU, and IA64 will be cancelled.
Not about 64 bits... (Score:4, Insightful)
First of all, with HP being a co-developer of the entire architecture, they are a big backer of the Itanium. So is Compaq, who sold their entire Alpha technology to Intel to focus on implementing the Itanium in all their high-end products (makes you think, was this all decided because they already knew they were going to merge with HP? Probably...) Dell is still sticking with 100% Intel, so the Itanium will be their bet for capturing more of the high-end segment. Even SGI is selling Itanium workstations (although, with the recent announcement of the MIPS-only, IRIX-only Fuel workstation, they might abandon the Itanium as well.)
So what's holding it back? I think that although there's now Linux available for it as well as a prerelease version of Windows Server along with some other systems (like HP UX) we still need to see more applications. Databases alone just aren't enough - and with the high prices of Itanium machines (the cheapest dual-Itanium 733 is around $22K at Dell, everyone else is probably more expensive) developers are not really happy about buying a couple of those machines and start hacking. So I think that because we don't see the Itanium much, developers are not investing in writing the software and business are not investing in buying the hardware.
Maybe Intel should start giving out IA-64 machines to opensource hackers and watch it fly? Where can I submit my address info?
Oh and about the subject of this post, the fact that the Itanium is 64 bits is not really all that important - the fact that a processor is 64 instead of 32 bits doesn't say anything about how fast it is. If you think it does, you can buy my R4400 Indigo2 for $10K
or my R10K Indigo2 for $20K... (Score:2)
Or my R10000 Indigo2 for $20K... at least it runs "IRIX64" the 64-bit kernel and the 64-bit ABIs.
elwood 6# uname -aR
IRIX64 elwood 6.5 6.5.15m 01091821 IP28
AFAIK, the only SGIs that use the R4400 in a 64-bit manner are the Challenge L, Challenge XL, and (original) Onyx. R4400 in desktop machines is limited to 32-bit support for memory contraint issues. 64-bit on the desktop from SGI requires an R8000/R10000 based Indigo2, R10K/R12K/R14K Octane/Octane2, or R14K Fuel. All other desktop configurations are limited to O32 and N32.
Re:Not about 64 bits... (Score:1)
Huge installed 32 bit codebase.
$1200 price tag for little speed gains.
Unproven platform.
I think thoses are good enough reasons. Also, many hacks are in place to allow 32 bit systems to do a lot of stuff 64 bit ones can do, for example, creating files > 2GB.
Re:Not about 64 bits... (Score:1)
Re:Not about 64 bits... (Score:2)
Unproven platform.
Whats worse, the next release will require a new motherboard, so there's no hope of getting used to it now, then a simple upgrade once the bugs are out.
It *IS* about 64 bits! (Score:1, Interesting)
In the EDA industry (Electronic Design Automation i.e. tools for making computer chips) we routinely hit the 4GB memory limit. 99% of EDA tools run on Solaris but EDA companies are slowly recompiling their apps to be 64-bit clean on Solaris.
Meanwhile Linux is picking up steam in the EDA world but the 4GB limit is holding it back. We're forcing into complex partitioning of our chips to break it into small enough chunks to fit under 4GB.
We need cheap (non-SPARC) 64-bit chips, say like oh AMD Hammer?
Can't design 64-bit chips with 32-bit tools? (Score:1)
Re:Not about 64 bits... (Score:1)
The big changes in PC hardware are successful only when both Intel and Microsoft marketing convince others (both resellers and users) that having the new tech is the Big Thing--even if neither group really understands the tech and why they should buy into it. I still get questions about USB, which Microsoft and Intel began to support better after the debut of the first iMacs and their USB-only support.
Another factor may be cost. Are Itanium chips much more expensive than P4s or AMDs? If so, adoption will be further slowed. This is not a game of better tech (IMHO, other companies typically win that game over classic Intel motherboard architecture), but of commodity pricing. The best tech is NEVER achieved by the lowest bidder.
Re:Not about 64 bits... (Score:1)
Re:Not about 64 bits... (Score:2)
Re:Not about 64 bits... (Score:1)
With this datastructure, for instance obtaining the free or occupied squares is easily done by oring them together. In today's 32bit CPUs this requires a few instructions, in the Itanium (or any 64bit CPU) this is only one.
Play a little chess on the net at http://mobilsjakk.no. This service (my employer's service in fact) would definitely benefit from being run on a server with dual itaniums
Re:Not about 64 bits... (Score:3, Informative)
For real-world numerical applications, using state-of-the-art Intel compilers, the Pentium 4 is faster than the Itanium. Of course, people still use the Itanium because of its substantially larger address space, which is a very, very significant issue. And we can expect the IA-64 architecture to catch up as compilers improve, as it is common with such architectures.
So what's holding it back? (Score:3, Informative)
cache hits (Score:3, Interesting)
they took one look at the people trying to do predictive memory loads and decided not too. this was a LONG time ago and now people have solved the problem so that most of the time you can get things from cache
IA64 fails to get things from cache too well (one of the reasons why they stuck such a large one on) so suffers from the latencey problems more than most
simple
regards
john 'try runnning spec marks on it' jones
HP is actually cheaper [Re:Not about 64 bits...] (Score:1, Informative)
$7995 single / $14995 dual
Check the prices yourself:
http://www.hp.com/workstations/products/itanium
link [hp.com]
beginning of the end of hand crafted assembly (Score:1, Informative)
Anyway, To all those who think it's performance is rather low, It seems more like a proof of concept chip rather than something intended for mass production.
What about power / heat ? (Score:3, Informative)
Power/heat this high gives system designers problems, plus it can't be easy getting ~100 Amps to&from a chip.
Otherwise, AFAIK, Linux has working ia64 so code size can be compared. I'd expect 4x x86.
Re:What about power / heat ? (Score:3, Informative)
As a complete processor lamer (Score:1)
I don't feel that any of the current reports are aimed at people such as myself, and don't feel that I'm getting the real deal in terms that I understand.
I understand what a register is, the advantages of 64bit, 128bit etc., even what a pipeline is.
What I would like to see is a bullet pointed list of advantages put in executive summary style, dumb-down, type!
Also, I read a few years ago about Elberus, who have some pretty neat claims here:http://www.elbrus.ru/mcst_e/proect_e/e2k_arch _e.htm [elbrus.ru]
You may be interested in their claims.
Re:As a complete processor lamer (Score:1)
Re:As a complete processor lamer (Score:2)
Re:As a complete processor lamer (Score:1)
What's the point? (Score:3, Insightful)
Except for the Itanium coming from Intel, what's the point? This is a prototype for a new architecture (IA64), prototype proven to be seriously lacking in speed, stability etc. I got to see a dual Itanium prototype from HP a few months ago, and all the comments I got about it were that it essentially sucked.
Really, if you need 64 bit, why not just go and get yourself some UltraSparc, Alpha etc.? I have gotten myself a used Ultra 30, will soon get an used AlphaServer, and I sure don't need to go buy an expensive, unstable processor that's not even got decent compiler support yet.
And if it has to be IA64, at the very least wait for McKinley - HP's engineers are supposed to be doing a much better job of IA64 than Intel did. Or even wait for the version after McKinley, which is supposed to profit from good ol' Alpha.
Re:What's the point? (Score:2)
The point of this article is to introduce you to the IA64 architecture, which McKinley and Madison are going to be based on. The point of developing for Itanium/Merced is to learn how IA64 is different from IA32 so that when you do a McKinley platform you are ready and not fighting with as many "what the hell is it doing now."
I've seen many protoytpes as well from HP and others and the reponse is not "it sucks." The response is "wow, we've got a lot of work ahead of us."
wrong direction? (Score:2, Insightful)
Software is by far more costly and complex than processors these days, and we just don't need extra complications in the form of processors that shift even more complexity into software.
I can't pretend to know what a "good" 64bit architecture should look like. But for the time being, something like Alpha or AMD Hammer seems like a better choice to me. And even Intel seems to be reconsidering and keeping a 64bit version of the Pentium as a backup strategy.
Re:wrong direction? (Score:1)
The main bottleneck of modern microprocessors is, in fact, the extra space and heat produced by the complex logic you defend.
The RISC processors was born because the CISC complexity was impairing performance as hell. CISC was good when the processing bottleneck was the instruction fetching : every slice of clock saved by reading hard to decode but compact bytecode worth the pain.
But now, the bottleneck was shifted to inner spheres. Nowadays a CISC processor waste more time translating the bytecode and executing the microcode that anything else. In this panorama Itanium made the right thing : lets get ripped out of complex bytecode. The really bad drawback (and yes, you are right on this) is the huge increase in complexity in coding in a cripped machine language level.
But think : how many compilers was written in the computer history, and how many applications was written with these compilers? This ratio will prove that it's worth the trade. Of course compilers will be more complex and hard to code, but once that damn thing was done, it was done.
Thet's the way MIPS, SPARC and Alpha was done, and they did very well in the past.
Re:wrong direction? (Score:2, Interesting)
Complex compiling issues are NOT a result of CISC or RISC in this case. In fact, RISC is far easier to write an efficient compiler for than CISC. The instructions offered by RISC more closely mimic the kinds of basic operations compilers manipulate in the very back end of compilation. Register sets are usually general and very orthogonal, compared to CISC (Intel in particular) where you have very few registers and they all have special meaning depending on context.
The complexity in building compiling tools with respect to the itanium is all about VLIW, parallelization and scheduling. These are incredibly complex topics with many subtle features that make optimization and analysis very difficult.
Also, think again if you think compilers are written once for an achitecture and then set in stone -- 'once the damn thing is done' it definately will not be done. It will probably be buggy and poor at doing these new complicated tasks compiler writers have never had to do before. It will likely take a few iterations before the compiler tools start to show off the architecture. The question is which will come first, the latter or industry's frustration with poor performance out of expensive silicon.
Re:wrong direction? (Score:2)
The problem doesn't magically go away by shifting it into software. A static compiler cannot do instruction scheduling and parallelization correctly--you need runtime instrumentation and JITting. The end result is something that likely performs more poorly than if you had let the processor do this.
Nowadays a CISC processor waste more time translating the bytecode and executing the microcode that anything else
CISC vs. RISC has nothing to do with it.
But think : how many compilers was written in the computer history, and how many applications was written with these compilers?
Not nearly enough compilers have been written.
This ratio will prove that it's worth the trade. Of course compilers will be more complex and hard to code, but once that damn thing was done, it was done.
Yes, if you are happy muddling through with C and C++. Itanium will further cement the dominance of languages that we already know to be absolutely lousy from a software engineering point of view, because almost nobody will be able to make the investment to write a competitive compiler for any other language.
compiler technology crucial (Score:2, Interesting)
Re:compiler technology crucial (Score:1)
compilers harder to write (Score:2)
Altogether, that can't be good for the industry in the long run. We need more, not less, support for new software architectures and languages. Instruction scheduling and parallelization are things that a processor can handle much easier than a static compiler because the processor can efficiently keep runtime statistics on what a program actually does.
Potentially a better approach to me appears to be hyperthreading, which redefines the problem. No, individual threads won't get very high performance, but code generation is pretty easy, and (unlike VLIW) the programmer can take explicit control of parallelism at a higher level through threads. To me, that seems like an overall better approach.
Can anyone understand this? (Score:3)
I can't even conceive of having to write assembly code for these monsters. Anyone happen to browse through the instruction set reference? All 900+ pages of it? It's all cryptic as hell. I could sooner build a rocket bound for Pluto than write a simple recursive factorial program in IA-64 assembly.
I sure hope someone can figure it out. I doubt I'll be doing any assembly optimizations in the future.
Re: yeah (Score:2)
I've had to do worse, our logic analyzers have to decode the bus traffic to figure out what was going on. As you can imagine, the fact that 1 OP code could mean 4 different thigns depending on when it shows up (or what happened before it) poses a significant challenege. You might say "big deal." Sure, big deal, if you are developing a system. We have to be able to provide tools that are bug free BEFORE the first silicon hits market. That way when a guy in a lab encounters what looks like a bug isn't fighting our mistake.
I fear one day we are going to get to a point where we can't provide a solution for hardware engineers to probe their system with.
Some weirdness, but good (Score:2)
Overview? How about Autopsy... (Score:4, Informative)
Years late the "Itanium" finally ships (although no one buys it) as Intel says, "But wait for McKinley! Then it will really work!"
The McKinley is the product of the "rethought" Merced project. McKinley is shipping later this year - with a completely different socket system so even the arch surrounding the "Itanium" is dead in the water.
Let's compare this to the REAL competition:
IBM Power4 1.3GHz - shipping for a while now:
SPECint2000 = 814 SPECint_base2000 = 790
SPECfp2000 = 1169 SPECfp_base2000 = 1098
Sun UltraSparc III Cu 1.05GHz:
SPECint2000 = 610 SPECint_base2000 = 537
SPECfp2000 = 827 SPECfp_base2000 = 701
Even the best Itanium 800Mhz reported int numbers are:
SPECint2000 = 365 SPECint_base2000 = 358
(Same box) SPECfp2000 = 610 SPECfp_base2000 = 526
Even if the McKinley (which doesn't ship for 6 months or so) produces double the Itanium numbers it'll still lag the currently shipping Power4 chips.
Remember the hype and FUD surrounding the launch of the "Itanium" chip that eventually hasn't even caused a ripple in the marketplace? Intel has sunk billions into this EPIC project and refuses to let it go even though it's years late so far hasn't produced the clear advantage over the RISC arch it was supposed to make obsolete. In many cases the "consumer" chips continue to make better results than the "server" chip series - and with AMD knocking on Intel's door throttling back production/performance of the consumer ship is not an option.
Will the McKinley better than the Itanium? Certainly.
Will it be compelling? Wihtout Intel behind it - probably not. (Alpha was the clear performance winner for so long but couldn't get any traction.)
Is VLIW^H^H^H^H EPIC the future of computing? "Answer unclear... ask again later."
=tkk
Re:Overview? How about Autopsy... (Score:1)
As for your SPEC figures, you could have at least made your post worthwhile by not fudging the numbers to make the Itanium worse than it actually is (although it still isn't very good).
Source: http://www.aceshardware.com/read_news.jsp?id=3000
800MHz Itanium
SPECint2000 base: 403 (Your number was 358)
SPECfp2000 base: 701 (Your number was 526)
The McKinley should be much faster. Hell, it's got an 8 stage pipeline instead of 10 stage, an additional 2 integer units (6 total), MUCH more efficient cache, on-die L3 cache, more L2 cache, MUCH faster system bus, etc.
Don't write off an entire architecture because you didn't like how the experimental implementation came out. Itanium was just to get a product out there for IA-64 early adopters to start getting code working on.
Re:Overview? How about Autopsy... (Score:2)
I didn't. These are the numbers from official submissions to the spec organization [spec.org]. (If they can't bother to submit results then they don't count.) I took the machine from with the HIGHEST Int performance - as I said in my post. The FP is the result for the same machine - as I said in my post.
There is a >700 SpecFP machine claimed by Dell but there is no corresponding SpecInt submission. I think Intel claimed nearly 800 Spec2000FP for the Itanium but no one else has been able to re-create those results. That's why non-submitted results don't count.
Don't write off an entire architecture because you didn't like how the experimental implementation came out.
But it wasn't supposed to be a proof of concept chip. It was supposed to be the future of computing.
An exec actually admitted this in 2000.
Which is at least 3 years after they knew it. Intel instead spread FUD around while refusing to talk performance numbers.
From Intel Press Release:
SANTA CLARA, Calif., Oct. 4, 1999 - Intel Corporation today announced it has selected Itanium(TM) as the new brand name for the first product in its IA-64 family of processors, formerly code-named Merced. The Itanium brand extends Intel's reach into the highest level of computing enabling powerful servers and high-performance workstations which will address the increasing demands that the Internet economy places on e-Businesses. "The Intel Itanium processor represents a new level of processor capability that will be the driving force for the Internet economy,"
Ummmm... okay. I see, by "highest level of computing" and "new level of processor capability" they meant "proof-of-concept place-holder chip". It's all clear to me now...
The full text is here on Intel's site [intel.com] since you seem to think I make this stuff up.
Now McKinley is supposed to be the next, big thing.
From an article about McKinley [com.com] previously on /.
"Applications will be about one and a half to two times faster than what you get on a (current) Itanium," said John Crawford, an Intel fellow in the enterprise platforms group.
The additional bus and processor speed and 3 megs of on-chip on speed cache should deliver nearly a 50% boost all by itself. If the "new" features of the McKinley don't add much more beyond that then where are they going?
Forgive me if I appear skeptical...
=tkk
Re:PPC Really clean Platform (Score:1)
"PPC is the cleanest Risc architecture"
For whatever definition of "clean", the closest were probably research chips like the original MIPS or predecessor of SPARC. PPC has quite a bit of extra baggage, not a fatal flaw by any means but hardly merits your description.
"IBM's design expertise and manufacturing prowress guarantees it will stomp on the competition."
LOL. If that were any guarantee of anything, history would be very different (for one thing, Digital would probably still be a company, and Intel would've vanished years ago).
"VLIW in and of itself has not come into the stage where it is a good design concept it is not mature enough."
And it may never. VLIW seemed like a pretty good idea 10 or so years ago, when processor speeds were still under 100MHz. But things have changed, and there is a whole new set of problems VLIW isn't prepared to solve.
"...Intel is retarded for not just buying the Alpha processor cleaning up the instruction set..."
Alpha was probably the cleanest commercial RISC processor. It was certainly minimalistic, giving the compiler freedom to optimize code.
Intel on the other hand is not exactly known for clean instruction sets. So what exactly would you propose they do to Alpha?
(*sigh* why do people keep posting bs to
What you really need to know about IA-64 (Score:1, Interesting)
A bit more specific, look at the bottom of the article where it mentions the "use merced for development, mcinley will actually sell". And the fact that even Dell no longer sells Itanium.
Sanity check:
system specint/specfp cost
Hp server rx4610 342/701 $23k
AMD XP2000+ epox 8HK+ 734/642 $1k
Keep in mind that the Itanics are supposed to be for the server market, so the specint figures will more likely track actual performance. Intel has been claiming that McKinley will be a vast improvement (actuall claims seem to have been steadily downgraded from "dominate the market" to "actually sell a few" to "won't make management look like complete idiots"...). Present claims of McKinley performance are 1.5 to 2.0 times Itanic performance, i.e. unlikely to keep up with Athlon, let alone the hammer.
Why does it suck?
While Intel does know how to design processors, the architectures are annother story. Aside from the 8086 kludge, intel has produced such "successors" as the 432 and the 860. The 432 was even slower compared to the competition as the itanic, and the 860 was even harder to write a compiler for (and impossible to write an interupt handler, let alone an OS).
EPIC is supposed to be VLIW with enough "extras" to allow the compiler to write code that won't require out-of-order execution. It is also supposed to allow intel to create several generations of compatible chips (something hard to do with pure VLIW). Somewhere along they way they forgot that the point of VLIW is to make execution simple. The lessons of RISC and the Cray machines is that the more simple and clean an architecture is, the faster it can go. Check out how long it takes to explain the architecture, then examine Alpha and ARM. Granted, ease of explaining does not always translate into ease of design, but it ussually does, and certainly did in this case.
What now?
It looks like the architecture of the future is x86-64. Hammer should appear this year (maybe not for sale, but at least samples). Intel is claimed to have a project called Yamhill that adds AMD compatiblity to the next generation X86. Right now, any support at intel for x86-64 appears to be a CLM (Career Limiting Move and no, it is still a CLM to support it now even after yamhill is enabled). After McKinley goes down in flames (there doesn't seem to be a chance of anything else), it will be interesting to see how long it takes intel to produce anything AMD compatible. Assuming that the next generation of X86 was started after the Pentium 4 finished, that would place it about 4 years from now. By the time the politics get straightened out, that is likely the earlies option (rushing that job will just make it happen later).
Why x86-64?
Anyone familiar with the x86 architecture ussually runs away in horror at the thought of actually using/designing one. Having said that, the 386 architecture fixes almost all of the problems with the 8086/80286. The problems left are:
variable length instructions - this problem has basically disappeard, all modern high power CPUs have bigger problems than this.
insufficent registers - hammers doubles the number of registers (note that itanics 200+ registers become a hindrence if not fully used).
addressing size (32bits) - obviously hammer fixes this.
In summary, intel had a chance to create a usable architecture (anything noticably better than X86 probably would have worked), and would simply owned the market. It is possible that some PHBs thought that anti-trust laws might actually be inforced and then created an architecture too complex to clone. If so, they certainly did so, admitedly making one too complex to build themselves.
Scott
it's always been about the compilers (Score:4, Interesting)
Hardware is getting more complex; it takes more sophistication to deal with it. Binding a (general purpose) processor to a language in order to make language implementation easier is exactly the wrong way to support a wider variety of languages. Making the most of a processor's capabilities is what compilers are for. That's what compiler writers get paid for.
That's not to say I'm in love with the Itanium. At first glance I found it a baroque rehash of old ideas. But time--and compiler writer's--will tell.
Re: (Score:3, Insightful)
RISC? (Score:1)
If they want to call it RISC, then don't make the instructions so large, don't include hundreds of possible instructions, and make the clock cycle time shorter.
Where have the basics of RISC gone?
A visionary's gutfeel regarding 64-bit widespread (Score:3, Interesting)
Given that it was coming out from the mouth of the CEO of a company that :
- can afford the move quickly and nicely (PowerPC architecture is clean compare to IA-64 + x86 and is 32-bit backward compatible).
- had successfully shifted the kernel to a clean replacement (less kludges) allowing the transition in a blink of an eye (ok, maybe 6 months)
- has a park of installed machines in places like labs (see gentech), and design studio.
- runs applications that would benefit the most are all in the Apple camp (A/V and number crunching apps like photoshop, maya and final cut)
- develops a big chunck of the major apps for its platform leading the way in term of design and adoption of new tech.
it would seem that we have about 8 more years of 32-bit glory or galore in front of us, before the current cpu architectures get displaced and eventually die.
Which 64-bit architecture will succeed is not clear today. Knowing that MS doesn't rush their OS out of the door to support the IA-64, it seems to be a little premature to tell.
PPA, the girl next door.
Re:A visionary's gutfeel regarding 64-bit widespre (Score:1)
That's not true. There will be a 64-bit version with 32-bit backwards compatibility and a 32-bit version (just like with the PowerPC 620 vs 601). That is what it says on the roadmap [motorola.com]. Since programmers are still transitioning to OS X, I wouldn't expect the switch soon, at least not until the high-end users start complaining about a lack of RAM. Don't forget, 64-bit will hardly make a difference in speed for most uses, especially since the G4 already has a 128-bit processing unit (Altivec). Any application that would benefit greatly from 64-bit instructions already supports Altivec. On the other hand, 64-bit instructions/data means moving twice as much data, even for 32-bit instructions. Given that modern CPU's are for a good part limited by bandwith issues, this may make a 64-bit CPU slower than the 32-bit version.
The larger registers and caches will increase the size of the die and thus the cost of the chip. So I'm not looking forward to it yet, my next computer will probably not (need to) move beyond 2GB of RAM, so I can wait.
Gad, what a turkey (Score:3, Interesting)
Then came the Inanium. VILW, code bloat, ugly architecture, requires near-omniscience from the compiler, very tough to program in assembler, a power hog, and with mediocre performance. If anybody else had launched this, it would have died before first shipment. As it is, it's dying anyway. Dell dropped their Itanium workstation recently. The Itanium may end up as a niche product, like the forgotten i860, i960, and iapx432 processors.
I'm hearing rumors of a new 64-bit machine from Intel that's basically an improved x86, like the AMD Sledgehammer. That may be what actually gets used.
Re:Gad, what a turkey (Score:1, Interesting)
Then came the Inanium. VILW, code bloat, ugly architecture, requires near-omniscience from the compiler, very tough to program in assembler, a power hog, and with mediocre performance.
eh?
even though the IA-64 arch does seem to have some weird stuff into it, i wouldn't call it UGLY especially when comparing it to the IA-32 "architecture" (or rather lack of it.) and who programs in assembler nowadays? (excluding the MMX/SSE stuff which however is a direct consequence of the crap fpu on IA-32)
mediocre performance might the first implementation be, i can agree with that. but is the IA-64 a power hog? itanium might be, but if you look at the article really carefully, you'll learn that the itanium CPU core only contains approx 25 million transistors. this is MUCH less than a P4! being a x86 chip always required carrying that extra baggage to decode those mysteriously coded x86 instructions. i'm not saying that the IA-64 instruction decoding is simple, but at least it's worth the effort (whereas the legacy x86 baggage is not).
and no, i don't like itanium or intel very much. i have an alpha 21066 at home :-)
Re:Gad, what a turkey (Score:2, Interesting)
even though the IA-64 arch does seem to have some weird stuff into it, i wouldn't call it UGLY especially when comparing it to the IA-32 "architecture" (or rather lack of it.) and who programs in assembler nowadays? (excluding the MMX/SSE stuff which however is a direct consequence of the crap fpu on IA-32)
IA-64 still has backwards compatibility with ia-32, which has realmode, v86, and protected mode. that makes the ia-64 a mess to start with. compilers still generate assembler in some cases and some people have to use asm for low level things in the kernel and doing things that you can't do in C, like calling software interrupts, which, by the way, requires that you enter either v86 or real mode which isn't as simple as changing the PE bit. you have to setup the stack, and memory segments again and real mode can only physically access 2^20/1024/1024=1 megabyte of memory at once. maybe if intel would stop building on their old crap the whole thing would get a little simplier.
but i guess it would be boring if everything was as simple and stable as a calculator.
Re:Gad, what a turkey (Score:2)
Granted, this needs improvement, but x86 is hardly a prize either. The architecture of the Pentium Pro and up screams kludge. Huge amounts of circuitry that do nothing but break up CISC instructions into something resembling RISC so that they can be executed at reasonable speed. x86 didn't beat CISC because it avoided code bloat, (Does anyone eve notice code bloat anymore? I mean most of the world is running Word processor that consumes 32+ megs of RAM, who are we kidding?) it won for the same reason Windows wins, which is because it runs all your stuff. Installed user base. It was compatible with what had come before, and no one wanted to buy new programs, so we've stuck with it. It's like an AMC Gremlin with V12 engine welded onto the roof because you didn't want to have to move your stuff out of the trunk when you move your stuff. And to steal somebody eles's analogy, that V12 has uses side-injection to remain compatible with your old Model-T. I mean, segmented memory architecture? Please.
So what if you can't write assembly to make your code faster. Aside from the "Real Programmers don't eat quiche" mystique, this is a problem why? You probably can't beat a good C compiler on a P3 or P4 either. With very few very specialized exceptions, the compiler is smarter than you are. Granted, it's slow now, but this is the first generation of the chip. With some architectural improvements(please, for the love of god do something different with the cache!) this could be a pretty decent chip, and one that will still run your old apps.
Re:Gad, what a turkey (Score:2)
Your instruction cache does.
Try out Linux Itanium on IA-32 systems (Score:2, Informative)
If you want to know more technical details about Linux on the Itanium Architecture, David and Stephane just released a book "IA-64 Linux Kernel: Design and Implementation". David was signing copies at HP's booth at LinuxWorld NY.
Intel's next chip will be... (Score:2)
Unobtainium
-
Re:Intel's next chip will be... (Score:1)
Itanium? who cares (Score:1)
It's different, not better (Score:2)
This is good for Intel, but not for anybody else. Go back and look at Intel CPU prices from just before AMD processors caught up.
And that's the real reason for the Inanium.
I Say, enough of registers (Score:1)
Pipeline depth (Score:1)