Even for going small distances like to Mars space radiation is a big problem. The fastest probes that we send out (that don't have to carry a complete ecosystem for us to live) could need more than 25k years just to get to the closest star system, at more than 100 times less distance than that planet. Probably no human will ever reach another solar system, so visiting there is badly out of the question.
Whats left? Contacting with a possible civilization there? Our planet has been with this size and in this orbit for more than 4000 millon years, and had a capable to send signals to other systems (maybe in very short range) for just 0,000000025% of that time, and who knows for how much time we will be around or trying to communicate. Was a civilization willing to communicate be around there 500 years in the past sending signals to us so we could get now a hint that someone is there?
You should try to become replaceable. Make most your task become automatic or trivial, that systems try to heal themselves when known problems arise. That anyone else can understand how exactly the systems work based on your documentation, or see that a problem is about to happen based on your monitoring.
That will make your work easier, be able to take appropiate vacations, and be irreplaceable when (not if) things change.
prawns fed on the new diet grow 40% faster and are healthier and more robust.
Look similar to the the claims for the rations given to cows, while their meat is not the healthier one. When the ultimate metric is rate of production instead of quality (specially if have health consequences) a lot of consumers will be harmed.
Why that wide range? It is taking into account if we take active measures to diminish it or try to not make it worse, or keep running as if nothing is happening? Or just the uncertain of predicting a so complex system with so much unknowns as is the global climate system?
In any case, with so uncertain final impact, maybe food and water shortages will be just the tip of the iceberg. Rising the average world temperature so much (at least, for close to the worst case) should have a lot of very visible effects in all the ecosystems.
... but rationalizing it. Sometimes you just need to run more or less isolated single apps, not for a full blown OS. In a lot of usage scenarios is far more efficient, (both in disk/memory/cpu usage and app density) and probably more flexible. In others full OS virtualization or running on dedicated hardware may be the best option.
It also brings a virtualization-like approach for apps in the cloud. You can have cointainerized apps in aws, google apps and many others, something like having a vm inside a vm.
Is not the only solution of its kind. Google is heavily using containers in Omega (you can try their container stack with lmctfy), you can use openvz, lxc, or solaris zones or bsd jails. But the way that docker mixes containers (not just lxc by 0.9) with union fs, making them portable and to have inheritance, is a touch of genius.
The missing pieces are being added by different projects. CoreOS as a dedicated OS for containers (that coupled with etcd and fleet could become a big player in a near future), OpenStack/OpenShift bringing manageability, and maybe someone could bring to the table what Omega does with google containers.
Maybe those credentials were posted on github by devels and then scraped from there. Or from google, there is a bunch of id_rsa that pop up with trivial searchs.
Anyway, 25.000 linux/unix servers looks like a very low number, considering the 500.000.000 servers running apache or nginx, even with multiple domain hosted in a lot of them.
And if people start buying from that brand over rivals (or having country legislation forbidding not open enough and/or so backdoored hardware) it may move others to do the same.
Also, if a "hidden" functionality is exposed in major brands using that executable code to perform malware-like activities that brands should be punished in security aware circles. That won't reach the majority of people, but will be an start.