What, too soon?
A lot of those early mathematicians were a bit on the crazy side, having come to that realization and not having any of the framework for coping with the idea.
Not being able to ack important message packets seems like a design flaw.
Even though we have a LOT more hardware now than we did back in the day, you still can't BFI your way through a lot of the big data applications that companies are starting to try to get into. In the past, the company would just throw more hardware at a poorly designed application and that would "solve" the problem. I once saw a team throw 48 gigabytes of RAM at a leaky Java program, and schedule weekly restarts for the goddamn thing. But it's a lot easier to hit hard walls with big data, to the point where you absolutely can't throw more hardware at the problem.