Short answer, No. TCP doesn't back off until packets are lost. uTP looks for latency increases which happen before packet loss (and therefore, before TCP congestion control kicks in) and throttles itself preemptively. Put another way, TCP treats all senders as having an equal right to bandwidth. uTP doesn't want to assert an equal right to bandwidth, it wants to send and receive in the unused portion of the available connection.
As they say [citation needed].
According to their 07-08 annual report the NPG had an annual revenue of 16.6M pounds, liquid assets of 7.5M pounds, and total assets of 69M pounds.
Relevant bits start around page 40.
"Hacked" is rather misleading.
The NPG made high res images viewable online as a series of tiles, i.e. you might deliver the full image as twenty smaller images each showing a portion of the total.
His "hack" was to download each of these tiles individually and reassemble them into a coherent whole.
Nothing was done to gain unauthorize access or compromise their systems in any way.
No, no, no.
Berne requires that the US protect foreign copyright holders if and only if equivalent works published in the US by US citizens would be protected.
If a work is intrinsically ineligible for copyright in the US then the US does not and will not honor any foreign laws that say otherwise.
The flip-side is that, on some issues, it is quite common to refuse to accept any explanation that _isn't_ grounded in science - a religions explanation isn't scientific, ergo it's wrong. I (largely) agree with what you seem to be saying, but the follow-on reasoning from that of many is decidedly suspect IMO.
They think they're not aiming at me, but I'm still on their radar. It's absurd that I should have to buy an expensive 'professional' model to achieve something the cheap model would be quite capable of doing were it not given an artificial restriction - assuming this is even offered as an option. Buying laptops with XP Pro rather than Home got difficult enough and I don't want to end up having to jump through fifteen hoops to achieve something that should be perfectly normal.
Linux returns of netbooks have been far higher than Windows, from what I've seen. People have got used to running lots at once and, if they're suddenly told they can't, I'd expect Windows returns to suddenly climb as well.
(Assuming of course that the netbook market doesn't go the way of the PDA market. I _love_ having proper portable computers again - I was an old-style Psion user and having that sort of thing at my fingertips is just brilliant. Kinda worried that we might have an overhyped market that promptly dies because it can't sustain the hype though...)
That's about it, really (though I've written quite a lot of code in a freezing office on the weekend - huge motivation boosts productivity for obvious reasons).
What kills productivity? Colleagues interrupting my train of thought, either by requesting my input or simply by doing something that inherently distracts. Sharing an office with sales staff can be a killer, simply because they're so often on the phone or running round assembling information. Music can help with the happy place but isn't even always on at home (and I love music, have far too many albums at my fingertips
I've been in offices where you shivered all morning, or where every last movement caused sweat to drip off you - neither was very productive. I can type just as easily on a laptop (heck, I've written a fair bit of code on a 9" netbook) but accept I'm unusual in that way
What can be the biggest killer though? Motivation. You tell me you code as efficiently when presented with a task which will achieve almost nothing of benefit if it ever goes live and involves large-scale maintenance on a poorly-built legacy codebase. We do our best work when there's a reward of pride, and when we know that our best work is still only polishing a turd, it's far harder to summon the energy.
Yes, but this surely presupposes that there are no observable phenomena with unobservable causes?
I confess I'm not remotely an expert in the field, but my interested observer perception is of a bootstrap problem in pretty much all scenarios. If we believe all matter was formed in a Big Bang of which we detect traces that match current hypothesese, what was the cause of the Big Bang event? I love and value science, but am deeply uncomfortable with the quasi-religious assertions from some that if it can't be measured then it isn't real. We've learnt to measure a great many things which we previously couldn't.
In all honesty though, if you wish to oppose intelligent design, let it be taught. The underground, opposed, rebel argument that They don't want you to hear will always have power - if you honestly believe it's rubbish, teach its tenets and then teach why you believe they don't match the data. If you're a good teacher with good information and the students are intellectually up to the argument, they'll likely agree and the rest weren't likely to have their minds changed either way in any case.
I've got a netbook, which gets used heavily as an ultraportable machine. As long as you're sensible, it's fine. It's far from unusual for it to be running:
* Visual Studio
* OpenOffice showing some documentation or notes
* Web browser
* DB program of some description, usually SQLite Admin.
Why, why, why? Anyway, as has been pointed out, plenty of apps seem to have already found ways round this. Annoy your customers in their day-to-day use and they'll find ways to stop the annoyance - if that means you're creating a group motivated to hack your security, that's just a terrible idea.
Stay out of your users' way and let them work the way they want to. If I'm daft enough to want to try to host a commercial website or want to do serious software development on a netbook, that's my problem.
The faster I go, the behinder I get. -- Lewis Carroll