It's not that supporting the old things slows things down, it's that it doesn't speed things up. It actually does cause some problems, because various things in the X11 protocol use 8-bit fields of which a significant space is used by legacy stuff that no one uses anymore, but that's largely worked around in newer extensions.
If you're in a world where most applications are sending commands like 'draw line from x,y to x1,y1' then X11 network transparency is really fast. At the protocol layer, anyway - if you use xlib then performance will suck unless network latency is very low because it adds a synchronous API on top of an asynchronous protocol (XCB fixes this). Modern applications don't do that, they typically render pixmaps and just have the X server composite them. X11 can still do a reasonable job here, with XDAMAGE, XFIXES, and XRENDER, allowing you to keep most of a pixmap (a Picture, in fact) on the server, update image data in selected parts, and do all of the compositing in the server. The problem is that none of the X11 toolkits actually do this very well. Wayland doesn't solve this at all - it simply says 'well, grab an OpenGL context and send drawing commands'. That works okay - the OpenGL protocol allows you to copy textures to the server (and the GPU) and composite them very fast. The problem is that this approach also works fine in X11, and with X11 you get network transparency when you do it (which works reasonably).
The main criticism I'd have of X11 is that it puts too much state on the server. There is no way, at the X protocol layer (or even in the low-level X libraries) of saying 'disconnect this window from this display, reconnect it here', or 'oh, my X server has crashed, recreate my state on this newly restarted version'. The latter worked fine in BeOS almost 20 years ago and works fine in Windows today. The former worked on NeWS 30 years ago. Both are use cases that I'd love to see addressed for modern devices. The Wayland solution to this is 'write a web app'.
So under your plan, now all the evil geniuses get to rule the world.
It's hard to say whether this is better or worse than the current plan to allow evil idiots to run the world...
If the average IQ is 100 (and it is, by definition), that means for everyone with a 160 IQ, there has to be someone with a 40 IQ, or two people with 70 IQ, or four with 80...
There is an incredible number of stupid, uneducated idiots in this world, right around you.
IQ curve is a normalized bell curve. Equal on both sides, reaching into infinity on both sides.
BUT... There is neither infinite IQ nor 0 intelligence. Neither of those would be a living human being.
So right there, the curve itself is a broken representation. If taken in such a simplistic "or two people with 70 IQ, or four with 80" way.
Back in reality, those numbers actually mean something.
Anything in the 71 - 84 range is considered "Borderline Intellectual Functioning".
These are people with difficulties learning to read, write, do math or solve complex problems.
People who don't get "When is a door not a door? When it's ajar." jokes.
70 and below is Mental Retardation.
At 50 - 70 range - reading, writing and basic math is an accomplishment, while communicating is a difficulty.
Do you REALLY see many people like that around you? Cause those are only about 2% of population.
And nobody is including their opinions in pols as they are incapable of understanding such complex questions or formulating meaningful answers.
Meanwhile, that curve represents ALL HUMANS. Including kids and babies. And senile old people.
So, a lot of those low IQ numbers are actually AGAIN people unable to understand or answer such questions.
At the same time, that right part of the curve are actual people too. 100+ IQ, and going up to 160 and more...
Major difference being that THOSE people really ARE intellectually functional.
Some of them MAY lack education or they may have prejudices and biases preventing them in reaching accurate or logical conclusions - but IQ is there.
Present and accountable.
And then there is a part where those IQ numbers actually have a +/- error built in due to the nature of the test.
And when the test favors those with higher IQ, who can breeze through the test faster, scoring more points, making less errors... guess which group gets penalized the most from pondering about the solution a bit longer?
Hint: It ain't the IQ 85 and below crowd. They hit their ceiling early on. Never get to the point where seconds mean additional IQ points.
Again, curve is a broken representation.
In reality, it is a lot flatter in the middle and steeper on the left side.
Cause while those standard deviations are rather arbitrary (representation of a measuring tool - not the measured value) - there IS a real cut off line below which it is obvious that people have problems with intellectual functioning.
Your view is distorted by the fact that you are probably standing a bit low (indicating higher IQ) on the right side of the curve, looking up-curve at all those people below you and going "OMG! There are SO MANY of them."
So you don't see that in actuality, most of those people are actually on your side of the curve. Closer to you, than to those below IQ 85.
Education on the other hand... that's a different matter.
And so are biases and prejudices and simply faulty information and reasoning.
No one is immune to that. Just remember Linus Pauling, his double Nobels and his ideas about vitamin C.
Or any person still believing in the dude in the sky, working in mysterious ways while murdering babies in Africa.
Those people can't be all below average. There are simply too many of them for that. And the curve is broken.
No. RAID isn't better handled at other layers. If you don't know about the filesystem semantics then you need NVRAM or journalling at the block level to avoid the RAID-5 write hole. RAID-Z doesn't have this problem. If you're recovering a failed block-level RAID, then you need to copy all of the data, including unused space. With ZFS RAID (all levels), you only copy the used data. There are numerous other advantages to rearranging the layers, including being a lot more flexible in the provisioning.
It's also a mistake to think of ZFS as a layer. ZFS has three layers: the lowest handles physical disks and presents a linear address space, the middle presents a transactional object store, and the top presents something that looks like a filesystem (or a block device, which is useful for things like VM disk images).
Once a subject goes over the mock horizon, it gets pretty hard to distinguish entire discussion threads from line noise.
You have mail.