Intel Itanium 2 Benchmarks 186
Pablo writes "Over at VR-Zone we saw some
interesting benchmarks of the upcoming Intel Itanium 2 processor codenamed
McKinley that is on schedule to be launched during second half of this year.
With a faster 3MB on-die L3 cache, 6 instructions/cycle and 6.4GB/s of
bandwidth, it is poised to perform at 1.5-2x of the current Itanium processor.
There is an overview of how the Intel Itanium 2 at 1Ghz clock frequency will
perform against the current Itanium 800Mhz and Sun's Ultra Sparc III RISC
processor."
No benchmarks (Score:4, Informative)
Mod Me down Please
Re:No benchmarks (Score:4, Insightful)
Nothing to see here folks please move along.
Well other than the slow death of competing high end architectures.
Lets see here we have:
SPARC
Itanium
Alpha
PA-RISC
MIPS
Power4
Cray
What eles is out there? I haven't really been in the market for a high end system in a while, but it feels like the market is shrinking and soon Itanium will be "the choice," unless legacy support is a concern .
Re:No benchmarks (Score:2, Interesting)
Note that Sun has been cheating on Spec, that's the only way they can make recent Sparcs look competitive. It's a shame that it will force every other vendor to cheat in the same way.
MIPS is dominating... (Score:1)
Re:MIPS is dominating... (Score:1)
Even though intel controls the desktop market pretty tightly they do not control the high end and low end. IMO this is because inteligent people can take their time to make decisions either when designing an embedded device or spending tens-of-thousand$ or million$.
Re:No benchmarks (Score:2)
Hmm, what makes the Power4 not suitable for your HPC needs. IBM seems to think it's more than fine since it's used in ASCI White (and anything else they sell for that purpose).
Re:No benchmarks (Score:1)
Re:No benchmarks (Score:1)
"to power4 has had to do with reliability at the hardware level. IIRC they have multiple identical cpus that execute the same instruction stream and check the results, or is that only in the BIG mainframes?"
Yes, that only applies to the really big iron, i.e. S/390 mainframes. The other boxen use redundant caches a la Sun's SPARC processors. Keep in mind that running N redundant processors is extremely expensive, not only putting the extra CPUs, but the 'magic' hardware required to compare every instruction executed by them.Re:No benchmarks (Score:1)
MIPS is dead on the workstation/server scene, SGI went the Itanium way...MIPS is today almost only for embedded devices.
SUN is building SMT into a variant of the next US generation, used for Quad machines and bigger.
But remember the strength of a SUN SPARC machine is not the processors, which hasn't been cutting edge for many years, but the big picture (Overall performance) - the machines are extremely well build.
But lately I've heard rumours about Compaq reconsidering the death of the Alpha - because of the possible future Itanium2 flop, even internal forces of the HP part of Compaq is reconsidering the death plans of PA-RISC.
Re:No benchmarks (Score:1)
Has SGI really gone Itanium? They have waffled on a LOT of things for years now. They kind of went wintel then backed out, then kind of went x86 linux then backed out [sgi.com]. Are they planning any new ia-64 products? The 750 [sgi.com] is a legacy product and the Pro64 [sgi.com] compiler seems to be gone.
Re:No benchmarks (Score:1, Interesting)
I'd love to know where people hear these kinds of things! I'd like to find the source and plug it good.
MIPS is SGI's primary platform for their worstation and server product lines. They will shortly be releasing Itanium based servers running Linux but they have stated again and again that MIPS/IRIX and ITANTIUM/LINUX are seperate product lines. Some of SGI's troubles stem from the fact that Intel is 2+ years late brining Itanium to market, they bet the farm on somone elses vaporware instead of their own (H1 & H2).
Re:No benchmarks (Score:1)
The sales of MIPS processors for embedded devices is quite good....AMD Alchemy is also MIPS based (Current is MIPS 32, the next generation will be MIPS 64).
Re:No benchmarks (Score:2)
Power4
Huh??? What are you smoking?
BRAG MODE ON
And I have an account on that baby!!! *Drool* Wonder how many fps quake would get?
BRAG MODE OFF
No seriously, they naturally have a strict policy on what you are allowed to run on it. You have to fill out forms requesting cpu hours with project descriptions etc. etc. Anyway, my plan is to run ab initio calculations on it. Hopefully that is. They're having some serious problems, related to MPI, I think... Which has led to the fact that everyone is submitting the big jobs they planned to run on it to the old T3E, which is rapidly getting overloaded...:(
Re:No benchmarks (Score:2)
It is a great box but only if you use half of the CPU's. not enough cache to go around on the dual cpu boards so IBM has to disable half of them and sell them as 16 ways. Kinda sad.
And their switch architech for connecting them together is slow and has a relative high latency. But for what they are charging...
It is a SUN killer not a HPTC box. They have been telling their customers for the last two years that this box is for TPC not HPTC. HPTC low margin. TPC high margin. They want the HPTC customers to migrate towards linux.
Thier switch is not even really ready for prime time on these boxes.
Here [ibm.com] you can see some benchmarks comparing the 32-way p690 with the 16-way HPC (ie. same machine but with only one cpu per power4 chip). The HPC achieves about 80% of the p690 performance. While this is impressive and suggests that the 2-cpu power4 suffers from L2 cache bandwidth I still think this is a semi-low-end configuration. When you want many cpu:s the internode bandwidth and latency is probably going to be more important. I.e. if you want say 512 cpu:s then 16 32-way p690:s are probably going to be faster than 32 16-way HPC:s, depending on your application of course. Not to mention reduced space, electricity and cooling requirements.
And also IIRC their switch so far supports only 16 nodes, although IBM claims it will support 128 nodes by the end of the year. And I wouldn't call it slow either, speed is 1GB/s (that's gigaBYTES, not gigaBITS), which is about the same as the T3E IIRC.
Finally, I'd say it's very much a HPC box in addition to a SUN killer. Take a look at the top500 list. There's lots of IBM SP:s there at the top isn't it? The p690 is essentially a SP with the 375 MHz power3:s replaced by 1+ GHz power4:s. Same goes for the switch. It's an upgraded version of the SP switch.
Of course all these IBM machines are seriously smoked by this new Japanese earth simulator machine, which is supposed to be 5 times faster than the current top500 #1, the ASCI white (an IBM SP with 8192 cpu:s).
Re:No benchmarks (Score:1)
Re:No benchmarks (Score:2)
Re:No benchmarks (Score:2, Informative)
Re:No benchmarks (Score:2)
Re:No benchmarks (Score:4, Insightful)
They sorted the benchmark results in ascending order and the connected the data points of completely different and independet benchmarks by a line!
What shall the line tell you? The faker the benchmark the better the results? Or
"This is a line graph that doesn't make sense at all. But look: It shows an increase, increase is good, so Itanion 2 is good!"
Re:No benchmarks (Score:4, Informative)
And, absolutely none of the benchmarks are substantiated with real data!!!
Only a fool would accept any of this presentation as fact. An even bigger fool would use this presentation in a decision whether to buy Sun or Intel.
Re: (Score:2, Insightful)
Re:On schedule??? (Score:1, Interesting)
Oh, good luck to AMD...
Yamhill? (Score:3, Insightful)
Well, this can only mean good for Linux...
X86-64 (Score:3, Insightful)
There is always Linux64.
Not to mention the fact that many a beowolf supercomputer would like to be designed on a Itanium 2. There is one at NCSA from IBM with 800 some IA64 chips. They're just waiting for the Itanium2.
Re:X86-64 (Score:1)
If anyone has first hand experience with 64bit applications on windows, please share your experience. I'm not trolling, just honestly curious about real world deployments of 64bit apps on windows.
You are right. (Score:2)
However, under Red Hat Linux 64 there are tons.
And note the release date of Itanium 2. Right along side Windows XP. There are supposedly 64 bit versions of SQL waiting to release with
Though honestly most people that ask me about 64 bit computing are Unix (Solaris, AIX, others) wanting to migrate to a less expensive hardware plaform running Linux to replace some lower end Sparc or Power3 boxes.
Though working for IBM I tend to work hard at the Solaris conversion than the AIX ones
Re:You are right. (Score:1)
Re:You are right. (Score:2)
But SQL is the main focus to start the new major cash cow to compete with Oracle.
Keep in mind, I work for a hardware vendor though. They don't tell us everything.
64 bit windows, close but not quite (Score:1)
There's not much to comment on. It worked, slowly. There weren't many apps. Tools were very expensive.
I switched it to linux way, way back in the 2.0.xx era. Since then it has acquired my personal uptime record of 350+ days while working 24/7 as home-office smtp/http/squid/ftp/firewall/etc. box. It's as solid as a rock and completely invulnerable to intel-style buffer overflow exploits and other arbitrary(x86) code execution attacks. I get a hit every other day from some skiddie that doesn't know what an alpha is or is not.
Re:X86-64 (Score:2)
Re:Yamhill? (Score:1, Insightful)
Re:Yamhill? (Score:2)
I don't know what you mean by a 'real OS' however having worked at the extreeme high end I can tell you that there are plenty of real fast machines with O/S that are in most respects (except performance) pure junk. And by high end I mean that some of the machines I worked on 10 years ago still outperform top end desktops.
The fact is that high performance machines are usually bought for fairly narrow purposes and as a result tend not to need a lot of an operating system. Back in the early 1990s an awfull lot of computationally intensive work was still being run on IBM mainframes running MVS and JCL which in many respects is not a whole lot better than MSDOS but you would never get the people who ran those piles of junk to admit it.
If you want high performance from a multiprocessor machine the architecture of the O/S does matter a lot, but may not determine the outcome. consider this analogy, when it comes to writing an optimizing compiler modern languages such as Eifel or Java give the compiler writer a heck of a lot more help than a language like Fortran. However until very recently many of the top benchmarks for compiled code were for Fortran compilers simply because brute effort could be used to compensate for poor architecture.
When it comes to the 'architectural features' required to make a multiprocessor machine work fast the cards are all with Microsoft. WNT was designed to work well on multiprocessor platforms and the design team were mainly DEC ex-VMS people who had a lot of experience in that area.
UNIX was originally architected for a UNIprocessor and there are a lot of design decisions that you just would not take if making it easy to run fast on multiprocessors was your objective. On the other hand this matters less than it might because there has been a lot of work since on compensating. The cost being that to make an SMP system work well under 'UNIX' often means having to use features that are proprietary.
The claim that UNIX has a better architecture than WNT is essentially as unprovable as the claim that vi is better than emacs. There are certainly still people in the computing world whose only experience of editing programs is with vi in line mode but will nevertheless post many gigabytes worth of posts to USEnet arguing the point. Having actually worked on O/S design, having used 20 odd O/S and having done system level programming on 4 (including UNIX) I can tell you that UNIX certainly did not get where it is on the strength of design merit alone. Many of the internals of UNIX are as confused and obfuscated as the syntax of the csh.
To port to a new 64-bit arch. with modern compilers and libs, it's not much more difficult than a recompile.
If you have nice 64 bit clean code that may be the case. The problem is that most people don't start from good code and even if they have tried to keep the code base clean they may not have succeeded. But remember that WNT has already been ported to a 64 bit architecture (Alpha) and although WNT is no longer supported on Alpha you have to believe that the compiler rules are still in place to detect code that is not 64 bit clean.
The issue that is probably more important for Microsoft is how .NET performs on itanium. With .NET they have the ability in theory to take applications and compile them for any architecture at installation time. I don't think there can be much doubt that this architecture is designed to support Itanium.
Re:Yamhill? (Score:2)
Slowlaris, oh yes, I had Sun sell me that when it was still a real heap of junk. Took the sysadmin a day each to configure four machines. He never did manage to get the multimedia (sound, video camera) features to work, nor had anyone else in the building. You might have heard of it, 545 Technology Square Cambs, Mass.
If they can't get an O/S to function two years after release it does not qualify in my book as superb...
Solaris kicks its ass clean out of the arena. Solaris has an extremely sophisticated internal threading model and scales nearly perfectly linearly to over 100 processors.
As you will note in my original post I differentiated the architectural contributions of Ken Thomson and co from later developers who just had the misfortune to start from the constraints of their design. It is possible to get a UNIX shell to run on a machine whose internal architecture bears little resemblance to UNIX.
The question is what the application developer is required to do to take advantage of the result. Does their application work in scalable fashion across all 100 processors if written for 'UNIX'? Or is it the case (of course it is) that to scale the application must either be 100 independent applications running on each processor or if it is genuinely one application does it have to be rewritten for the specific UNIX variant?
In the old days some processors would have a mode switch. So the VAX could emulate a PDP-11 but while it was doing so it wasn't doing the types of thing you would expect of a VAX, like virtual memory or such.
I have written code for machines with several thousand processors. Absent some development in hardware I may be unaware of nobody has a machine that can load more than about 16 processors that does not have some pretty significant limitations. There is a good reason for this, any given generation of silicon can support a certain processor technology and a certain interconnection technology. Unless you go to some form aof non shared memory system performance tends to top out at 8 processors. Yes I can believe that Sun showed off a machine that can calculate the mandelbrot set on 100 processors with perfect linearity, I doubt that it responds in that manner to all applications. Even exotic MIMD machines don't do that when they are purpose built for the application.
What iron has NT ever run on with more that 16/32 processors?
If Intel are right and they can provide 10 times the performance per processor over Sun then Windows need only run on 16 processors to kick Sun's ass.
However since the benchmark in question compared a 2 processor Intel machine with an 8 processor Sun machine the nature of the game should be obvious. Intel clearly chose a problem for which Sun's platform does not respond with near perfect linearity and the benchmark is obviously skewed in that respect. Even so, if Intel have the higher raw processor performance they have the engineers to build machines with more processors. Intel has at various times developed extreemly fast research prototypes (Bill Dally's hypercube being an example. Intel spends 3.2 Bn on processor research a year, Sun spends only 2 Bn on all its research.
Asking the number of processors NT runs on when Intel is only starting its 64 bit processor line is kind of beside the point. The real issue is where we will be in 3 years time...
Re:Yamhill? (Score:2)
MIT LCS and AI Labs, if they can't get Solaris to work properly without spending a week doing it the problem is the o/s vendor.
We thought this somewhat strange since at the time SunOS had a realy good reputation on campus. Then we discovered that what people were using was a custom distribution cut by MIT...
As for hypercubes, the ones I saw were Transputer machines. Never wrote any code for them though. They only let me loose on the VAX cluster with 10 lines of FORTRAN *puke*.
You can't build a hypercube from a device with 4 links, its a geometric fact that you need 2 links per dimension. I did use Transputers, including some of the 1000+ node machines. Perhaps if you had some experience of the machines you are ignorantly blathering about your posts would not get modded down.
Your entire post confirms the argument I was originally making, that the term 'UNIX' no longer describes the system internals, it merely describes the application interface. However the application interface is still a major constraint on application design.
If you study the internal architecture of NT you will find that, despite M$'s continual assertions that UNIX is dead, outmoded etc. etc., they have come up with a kernel that, with every new release, is looking more and more like some of the big commercial UNIX kernels of 5-10 years ago.
Clearly from the tone of your post you don't like Windows, I have great difficulty therefore in believing that you have spent such a great deal of time examining the WNT internals that you can come up with such an assessment.
It would not be so surprising if WNT shared some of the concepts in Mach given that Rashid has been working for Microsoft for some time. However the architectural principles have from start to finish been based on VMS.
Re:Yamhill? (Score:2)
So you think your second hand account of some unspecified lecturer's opinions on O/S internals gives you the authority to pontificate on the issue?
I have first hand experience of VMS, NT and Unix internals. I suspect that either the lecturer was as opinionated as yourself, or more likely you didn't spend very much time listening to him either.
As for being a troll, if as you claim you are at Karma 50 and post at zero it sounds to me if a member of the slashcrew has come to an uncomplimentary opinion of you.
Re:Yamhill? (Score:2)
NT does not have the UNIX thread model, or rather it has pthreads but the system apps all use a different model. This is not suprising if you know about thread models. pthreads is not generally considered the acme of good design. It is based on Tony Hoare's monitor design which is almost thirty years old now and Hoare introduced a much better design only a few years after monitors. However MULTICS was designed in the intervening years and when SMP Unix systems started to emerge in the late 80s the work was largely driven by the MULTICS work.
Opinionated is my middle name. However, I have to take his word for it. The guy is very experience in such matters and has written a lot of kernel-level networking code.
I don't know who your lecturer was but I'll bet he wasn't Tony Hoare, my Oxford Tutor who along with Per Brich Hansen and Dijkstra invented most of the concepts on which all modern O/S concurrency models are based. If Microsoft thought there was a problem with their threading model I guess they would call him up since he now works for them.
Re:Yamhill? (Score:1, Informative)
The market has decided already. It voted for "cheaper", and even HP can't spend the money to keep those other processors competitive without pricing its machines out of the market.
I don't give more than 5 years to Sparc
Re:Yamhill? (Score:2)
x86-64 on the other hand, is a good solution for today's demand.
So who will have an upper hand 5 years from now? probably intel, but who knows, Ghz sells.
Re:Yamhill? (Score:1)
If HP has been consitent on one front, its that PA-RISC is being phased out and IA-64 phased in. The question is how fast, and how long will the be able to milk the PA-RISC for all the support. (But they're not even betting on that too much it seems, since the IA-64 HP-UX comes with PA-RISC binary emulation -- its really more of runtime translation than emulation -- you get about 80% orignal speed, pretty nifty :) )
Yikes! (Score:2, Interesting)
Not to mention 130 watts of power consumption. And you thought Athlons were hard to cool!
Re:Yikes! (Score:2)
Re:Yikes! (Score:1)
Re:Yikes! (Score:1)
Its not pretty. The 4-CPU Itanium 1 systems are composed almost entirely of fans on the front to create a sort-of wind tunnel for the damn CPUs. :)
When you turn the machine on, there's so much interference on fan spinup that all the monitors around you degauss
run benchmarks in cache == FAST (Score:4, Interesting)
real systems are about BANDWIDTH
memory bandwidth/latency is the reason AMD killed the P4 in benchmarks
lets see INTEL go up aganst a SUN on a large oracle DB then I will take notice
really this is where SUN make their money
regards
john jones
Re:run benchmarks in cache == FAST (Score:3, Insightful)
Actually, Intel systems do pretty well [tpc.org], indeed, better than Sun running Oracle with a 3000G test database. And they do a good job [tpc.org] on transaction throughput too.
Re:run benchmarks in cache == FAST (Score:1)
Exactly, the single most important mark (Score:1)
Always.
Everytime.
Everywhere.
Why has a vendor with balls(willing to post FLOPS/$ or whatever transactions/$) never been born? Somebody must have the highest performing system, who are they? Why don't they show themselves?
Re:run benchmarks in cache == FAST (Score:1)
Re:run benchmarks in cache == FAST (Score:1)
Re:run benchmarks in cache == FAST (Score:1)
Re:run benchmarks in cache == FAST (Score:2)
Actually Hammer's long mode is a significant cleanup of the x86 instruction set. A number of rarely-used instructions and architectural features (e.g., segments) are removed, new general purpose registers added, various other features regularized (e.g., you can address the low 8 bits of every single register), and you can even ignore the x87 nightmare and use SSE2 as a clean floating point architecture
I just like to say "AMD's Hammer will crush Itanium"
Re:run benchmarks in cache == FAST (Score:2)
There's definitely some confusion here. P4 (with Rambus) has much better bandwith than Athlon. On any memory bandwidth benchmark (e.g. Sisfot Sandra) the slowest P4 is faster than the fastest Athlon. The 533 MHz bus version of P4 does 4.26 GB/s, while the Athlon only does up to 2.4 GB/s or 2.7 GB/s. Latency-wise they're pretty much equivalent. They're the same using DDR, and 400 MHz Rambus is slower than DDR, and 533 MHz Rambus is faster than Rambus. Of course, for most applications, latency is more important than bandwidth which is why they're closer in overall performance than in memory bandwidth performance.
Re:run benchmarks in cache == FAST (Score:2)
Companies buy Sun because if for some awful reason a processor in an E10k dies they don't have to shut the machine down. The machine can be opened, still running, and the processors hot swapped.
The fact is, Sun machines are really really dependable, and thats what companies pay so much more for.
Re:run benchmarks in cache == FAST (Score:1)
can you say ecache data parity error? (Score:2)
Sun did change the error message in later versions of Solaris 8.
I know a large company (you would know the name) who's production R3 system was off line for more than a day, mid week, because of this (crashed, and that trigered the problem that kept them down, thanks for the sucky support, EMC!).
Yet more Intel vapor-hardware (Score:3, Insightful)
It seems that in comparison to finding ways to rev up the clock speed, PC-based innovation in processors has stagnated -- at least as far as those innovations that actually reach the market.
Perhaps I'm just picky.
Re:Yet more Intel vapor-hardware (Score:1)
Sure, you could introduce compatible stuff like HyperThreading, but I believe this is scheduled for the generation after Itanium2 (or the one after that). Nobody should really expect the Itanium2 to be much more than an incremental improvement over the Itanium. The Itanium has afterall always been regaded as sort of a developer's beta version of the architecture.
Re:Yet more Intel vapor-hardware (Score:1)
Re:Yet more Intel vapor-hardware (Score:2)
Intel sold two or three thousand Itaniums, total. Pretty small "market".
I'm ashamed of you guys (Score:1)
seriously, on one hand it sure sounds like marketing hype but on the other hand what I want to know is how well it will benchmark on Doom III!
Re:I'm ashamed of you guys (Score:1, Offtopic)
What about I/O? (Score:1)
Until they get working implementations of a new I/O bus and a memory architecture that gets RAM bandwidth and latency up to a point where it can keep up with the CPU, this will continue to be nothing more than trivia.
Re:What about I/O? (Score:2, Informative)
In 4-way configuration, each CPU gets only 1.6GB/s shared I/O and memory bandwidth.
UltraSparc III has 2.4GB/s memory bandwidth + some I/O bandwidth for each CPU.
Sledgehammer will have 5.4GB/s memory bandwidth for each CPU
Re:What about I/O? (Score:2)
Properly written applications that take advantage of the cache (think video encoders that apply multiple filters on content already in the cache, for one example) are going to scream on this architecture.
Cool, but... (Score:3, Funny)
There's another variation on this story..... (Score:2, Informative)
Compilers (Score:2, Redundant)
When I can run my C++ through an Itanium compiler and have it come out good, then I'll believe it. Benchmarks? Right.
Re:Compilers (Score:2)
Solid information content: 0
Repetition of things heard elsewhere: 10
Re:Compilers (Score:3, Interesting)
What I really don't get, though, is why no one is focusing on using JITs with these. It strikes me that this is the ideal platform for a JIT, where it can recode parts of the program on-the-fly based on where the bottlenecks are and so forth. I mean, wasn't this the whole point of a just-in-time compiler in the first place? IBM's Java runtime can rival C++ in speed if it is allowed to run for a reasonable length of time, allowing the code to become truly optimized, and since Intel is targeting this thing in a server environment where applications will run for a similarly long length of time I fail to see why they aren't going that route. This has the additional benefit that as our understanding of how to optimize for VLIW improves, the programs do not need to be recompiled, but instead can immediately get the benefit. (I am fully aware that Itanium is supposed to do some of this type of optimization itself, but current specs are utter crap, to put it lightly.)
Interestingly, Sun's MAJC architecture does exactly that, expecting that a JVM or similar virtual machine will run on top. I have no clue what happened to that chip, but it struck me that it had much better potential to kick ass than Intel's Itanium despite having similar designs precisely because it was designed for a JIT to be on top.
Re:Compilers (Score:3, Insightful)
Although a JIT would be able to discover and exploit behavior patterns that didn't show up until runtime (and therefore not exploitable by a static compiler), it's not a panacea. Lots of programs are unpredictable even down to the level of individual loop iterations. Such programs really need small branch penalties and hardware support for instruction reordering
you can profile and recompile the kernel/modules (Score:1)
Re:Compilers (Score:1)
McKinley is a fast SOB (Score:1, Interesting)
I have had the chance to work with a McKinley box for a few months now, and it is with no doubt the fastest chip in the West, especially for some applications like public key crypto algorithms.
Indeed, McKinley running at 1 GHz can do a 1024-bit private key operation in 0.2 milliseconds - something well beyond any other existing processor. For high-volume secure electronic transactions, McKinley rules.
Re:McKinley is a fast SOB (Score:1)
especially for some applications like public key crypto algorithms.
Oh, here we go again with the crypto.You fail to mention that crypto is the ONLY application where Itanium (1) was any good at -- it has a lot of shift units. Is McKinley going to be the same story? Great crypto performance, really crappy integer / everything else performance?
McKinley had better be better than that, or its going to get the same lukewarm reception as the Itanium.
And they do seem to be having the same difficulties of pushing the clockspeed to decent levels as they did with Itanium. It was only by the C-rev of the original Itanium (the last pre-production beta chip realease) that they achieved 800Mhz, while the original target, I believe, was over 1Ghz.
Its sad, since we all thought McKinley, being desgined more by HP and less by Intel, would have good performance besides SSL.
Re:McKinley is a fast SOB (Score:2)
> McKinley rules.
For the price of a McKinley you could buy a Pentium 4 and a pile of someone's crypto ASICs, and blow the McKinley away.
better parallelism (Score:3, Interesting)
Intel's claimed specint2000 and specfp2000 are both about 1.75x the 800MHz itanium. And this with only a 25% clock speedup to 1GHz.
They claim specint2000 is 1.3x Sun Ultrasparc3 1050MHz, and specfp is 2x.
Unfortunately, there is no indication of what the frequency headroom/scalability might be. The main point of the pentium4 architecture is to scale to 4+GHz. Can we assume anything similar for the itanium?
Re:better parallelism (Score:1)
Performance improvements are conservative... (Score:2)
I don't follow this too closely any more, but I would presume they'll get to 2+ GHz, maybe 3 GHz, but probably not 4 GHz. A 4x jump is a lot to ask for without some additional redesign, especially if you are talking 4-way SMP running at those rates.
Given that they're claiming a 2x boost in SPECint2000 and SPECfp2000 from Itanium, on the same
--LP
Good For Linux (Score:1)
Most scientific heavy-duty work, such as EDA (chip design, etc.) is done on Sun and HP stations running their brand of Unix. Now, if an Intel processor based station starts to perform better than a comparable Sun station, at a much lower price, PLUS you run Linux instead of SunOS or HP-UX, a solution that costs a fraction of a price, and you get the same or better performance- well, you now have a VERY good reason for companies to start using Linux based workstations.
So cheer Intel and AMD on- because it's good for Linux! :-)
If you can profile and recompile the kernel/module (Score:1)
Itanium/Itanium-2 , infact i think most of the development money went on producing good compiler algorithms for optimisation.
This kind of processor would be great for running OS software because you could compile you kernel/modules/XFree/KDE/postgress etc... with a majic +profile for profiling info, and recompile it with better performance later on. this would mean that diffrerent prople would have the kernel/db software optimised for exactly there needs not for a general peropus implementation.
SPECint???? (Score:3, Informative)
I like the part where they said that Itanium 2 has 2x the SPECint performance of the original Itanium, since they never published it!! The SPECint performance for Itanium was so bad that only published SPECfp data!
It's just the same thing that happened when IBM published the SPECint/fp for POWER 4 processors. They only publish the data using 1 processor on the p690, so they run the hole SPEC benchmark suite un the 128MB SRAM cache memory, avoiding using regular DRAM. The easy way to see this is that they never published any SPECrate number, so to avoid showing that they don't scale as all processors start competing for the cache.
Sun USIII 1050MHz is almost 54% faster that USII 750MHz, as anyone can check going to the SPEC page (Sun Blade 1000 Model 1750 against Sun Blade 2050), with a 40% clock speed-up (this 14% increment is due to the compiler). This is exaclty the same processor at a faster clock, while Itanium 2 has more cache and a different architecture that Itanium, so a 1.5x to 2x speedup is less than spectacular, I will said.
For transaction processing, thay don't give any clue to show where they get the info from. While they expect to get the best OLTP number for 4-way systems, I don't think they will be able to surpass the AlphaServer ES45 MoDel 68/1000, which is by far the best 4-way system ever. What's worst, WLIW is know for been a poor performer for OLTP, and a great performer for floating-pont (that's why the only publisehd SPECfp!!). They never published any OLTP benchmark for Itanium (nor SAP, Peoplesoft, ORACLE, or even the raped PTC-C), so you can have an idea of how poor it is...
As of today, Fijutso PrimePower with 128 SPARC processors is the faster OLTP server ever (both SAP and TPC-C numbers!), with IBM p690 a close second for TPC-C and Sun SF15K a close second for SAP SD2-tier. Intel never showed in this kind of performance numbers, and Itanium certainly won't (unless while they keep running Windows).
Re:SPECint???? (Score:1, Informative)
Go to spec.org, you'll find all kinds of SpecInt results for Itanium (aka Merced). Those for Itanium 2 (aka McKinley) will appear at launch.
Sun USIII 1050Mhz Spec benchmark is the result of cheating on one test where the compiler has been taught to recognize the benchmark and apply a conversion.
Re:Wrong-O Jim... There are 4 published results (Score:1)
By the way, the SPECint for the Itanium are as low as I expected to be (less than half of the USIII or even the fastest Pentium 4!), that's why the will suck in OLTP. OLTP is marely int, and Itanium is only competitive (not the fastest) on fp, that's why the talk about DSS.
By the way, the tunning that Sun uses for the compiler, is only for SPECfp, it does not work on SPECint, so USIII is a real number, but of course you knew that, right?
Itanium 2? (Score:4, Funny)
NY Times Credulous (Score:1)
The article also mentions that Jack Dongarra - keeper of the Linpack-based 500 fastest computer systems - now shows an Itanium-based cluster at the top of the heap.
Unfortunately for programs that don't run out of the cache, there are three dimensions to computer system performance: processor, memory, and io.
Intel marketing has successfully skewed the common perception to the detriment of a more balanced system viewpoint.
Bad codename for a next-gen processor (Score:3, Funny)
Pfffft... One Gigaherz! (Score:2)
No, seriously, it seems that initial releases of the Hammer will have 2X the clock frequency of the McKinleys. I hope Intel includes an "Opteron rating" into the names of the various models just to help us keep things straight!
DDR, no Rambus! (Score:2)
Benchmark? What Benchmark? (Score:1, Insightful)
I'm getting interesting error messages (Score:1, Funny)
Logo.... (Score:2)
Why Is This Chip Necessary? (Score:1)
What is the current advantage to buying this chip instead of an older chip that is 5-10% of the cost of these ultra-fast state-of-the-art CPUs?
In October 2000 I bought several thousand dollars of Intel stock. That very night the stock price lost 40% of its value; going from $62 per share to the low $40s. It has never recovered and is currently in the upper $20s per share. What is this chip going to do to restore the value of Intel's stock?
I'm serious. Please give me some of your Slashdot insight as to why anyone would want to buy this thing? Will the sales of this thing ever generate the funds needed to recoup the R&D investment (never mind generate enough excitement to actually boost the depressed stock price)?
Re:Why Is This Chip Necessary? (Score:2)
Or, maybe I should just invest heavily in Microsoft.
News? Fluff! (Score:1, Redundant)
On second thought, don't even bother to call me then either. I can currently buy a Sun Enterprise 420R [sun.com] right now. What was the point of the story again?
--Mid
Address the real reason I use Sun over Intel... (Score:1)
Faster is not always better. As a system admin with both NT and Unix systems, my goal is availability and managability first, savings second. Let's face it, I could rebuild a Sun Solaris box remotely with a Palm, a VT-100 emulator, and a cell modem from just about anywhere in the country if I had to.
Why is that important? I can only speak for my company, but being able to do the Sun maintenance from the comfort of our homes/desks is very important to me and my staff. We have equipment in remote locations (over 100 miles away from the office), and not having to drive over there to rebuild a server or install patches saves $100 in expenses, plus takes only 20 minutes instead of all day.
This is not a M$ bashing bit either. If we were using Linux on Intel we would have the same issues. What I need Intel to do is very simple, restore a serial console to the platform. Let me have access to the BIOS from the command line and during startup. Let me power on/power off equipment from the console port too.
Yes...I can run a cardpunch machine too!!!
Re:AMD (Score:4, Funny)
Re:Odd. (Score:1, Insightful)
800,750,900,1050 yet the Itanium 2 stays the same speed.
Re:Open Source like AMD? (Score:1)
I don't want to have to compile my cpu too.
---
Re:It better be quicker (Score:3, Insightful)
Bullshit. You don't know this, and Intel's history with new processor development suggests otherwise. Maybe this means that they finally have the heat issue under control, and can *finally* reach the clockrates they want to advertise. Maybe it means clock distribution is reducing chip performance to the point that heat isn't an issue. Maybe it means that you work for Intel marketing, and think you are repeating something you heard an Intel engineer saying at a party.
-Paul Komarek
Re:A fast CPU , oh wahey , the adrenaline buzz ... (Score:2)
webber carbs vs bosch jetronic.
>>>>>>>>>>>
Here's your profile:
You use AOL.
You're VCR blinks 12:00
All you're Windows ME software still has the default settings.
You don't do your own oil changes.
You call a handyman whenever something leaks or breaks or burns.
You've never read the instruction manual to anything you own.
In total, you're a complete marshmellow who doesn't like to get his hands dirty.
Some people like tweaking, whether on cars or computers. More power to them!