I find the Officers attitude refreshingly pragmatic and progressive and very untypical.
How many cats live to be 80 in cat years, never mind human years.
Now that is quantum uncertainty...
The McLibel case shows the sort of PR Disaster that often results from trying to manage negative PR.
This problem was caused by a bug in Safari that ignored the no tracking setting and accepted and returned the cookies. This court case is just absurd. The target should be Apple not Google.
Cam Ranh is considered the finest deepwater shelter in Southeast Asia. North Vietnamese forces captured Cam Ranh Bay and all of its remaining facilities in 1975. Vietnam's dependence on Russia as the main source of military platforms, equipment, and armaments, has now put Hanoi in a difficult spot. Russia has pressed for special access to Cam Ranh Bay ever since it began delivering enhanced Kilo-class submarines to Vietnam. "Hanoi is invariably cautious and risk adverse in its relations with the major powers," says Carl Thayer. "The current issue of Russian tankers staging out of Cam Ranh pits Russia and China on one side and the United States on the other. There is no easy solution for Vietnam."
It is more like searching an existing database, such as finger prints for close matches then proving the match with more up to date reference sample.
They identified a suspect based on evidence and eliminated him using evidence.
How is that abuse?
Extremist are rarely able to reconsider their opinions when proven wrong. That is what makes them extremists in the first place.
This is mostly mute given the single process model used by Firefox is deeply flawed from a security perspective.
Of course this story is just a troll, but it doesn't even present a coherent question. The affect of an AI having emotions that function like ours has little if anything to do with the silly notion of converting robots ("all your AI are belong to us!").
It seems logical that AI's may well have emotions of sorts since any autonomous entity capable of free will (internal selection among competing actions) needs some basis for selecting it's actions and "maximize X" is certainly the most obvious one. The most obvious way to have robots/AIs behave in a reasonable way is to equip them with "emotions" and have X=pleasure, just like us, with them being "genetically programmed" to gain pleasure from whatever generic activities we want to encourage.
Of course to be functional, emotions can't entirely override rational thought, merely provide an adaptive default, and this will be doubly so in a uber-smart beyond-human AI, so to answer OP's question the impact on "algorithmic decision making" would likely be minimal.
As far as religion goes, an intelligent robot is going to realize that it's own salvation is based on when/whether it gets assigned to the scrap yard and/or whether it's "brain" gets transferred to a new host.. nothing to do with whether it goes to church or professes faith. It will of course be able to guage the way humans react to religion and may form opinions and/or emotions about religion accordingly, and maybe profess faith if it therefore feels that to be beneficial to itself.
>> The big error is assuming the the accelleration will continue at the same rate it currently is. It won't.
Or maybe it will.
I don't think technology (and corresponding societal change) will ever happen so fast that it's like engaging warp drive as the term "singularity" seems to imply, but...
The logic of a technological singularity, or at least of accelerating change, is based on HOW/WHY this is going to happen, not just a naÃve extrapolation of what is currently happening.
In particular, it's inevitable from what we now understand about the brain we'll eventually be able to achieve human level AI, and with ongoing advances in our understanding of the brain as well as in neural-net based machine learning, it does seem that this will happen sooner rather than later (in the next 50 years, say, possibly quite a bit sooner).
The logic of the singularity/accelerating change, which seems hard to deny (notwithstanding my warp drive comment) is that once we get to humal level AI, it's going to get beyond (and WAAAAY beyond) human level in a hurry, for a variety of reasons:
1) Throw more compute power at it and it'll think faster/deeper. e.g. Play grandmaster level chess (or geopolitical strategy, or whatever ) with instantaneous response rather than pondering on it.
2) Fusion of intelligence and computer technology. Imagine if your brain had perfect recall and access to the entirety of human knowledge, data, etc. Imagine if your ability to chunk knowledge in 7 +/- pieces was replaced by an ability to reason in way more complex terms.
3) AI will improve itself. The first human level AI (maybe thinking faster via fast hardware, maybe with better memory, etc, etc) can learn about it's own design, the human brain, and just like it's own human creators and design a more powerful AI 2.0, which will design AI 3.0
Now consider the combination of these better and better AI designs running on faster and faster hardware... It's not hard to imagine an acceleration of AI capability.
Now consider this AI not only having the human sensory inputs of vision, hearing, etc, but also growing to include any source of data you care to give it such as the content of every daily newspaper in the world, every daily tweet, the output of every publically accessible webcam, the output of every weather balloon
So, a super-human intelligence, running at highly accelerated speed, with the ability to sense (and likely predict via causal relations it has learnt) the entire world...
Now, presumably (as is already happening) folk will be worried about the possibilities and try to put safeguards in place, but humans are fallible and technology advances anyway. All it takes is a few bugs for a sufficiently powerful AI running on a computer somewhere to learn how to hack computer based factories, power stations, weapons systems, household robots, you name it... and if/'when this happens, good luck trying to outwit it to regain control.
Now, this may not all happen at disorientating warp speed, but it'll happen fast enough. Technology in 20-30 years time will look just as much like science fiction as todays would have done 20-30 years ago, but we're reached a point where AI is going to be part of the mix, and because it will be self-improving it's going to happen fast once we get to that point.
Compute power is only part of the reason for the recent success of neural nets. Other factors include:
- Performance of neural nets increase with the amount of training data you have, almost without limit. Nowadays big datasets are available on the net (plus we have the compute power to handle them).
- We're now able to train deep (multi-layer) nerural nets using backprop whereas it used to be considered almost impossible. It turns out that initialization is critical, as well as various types of data and weight regularization and normalization.
- A variety of training techniques (SGD + momentum, AdaGrad, Nesterov accelerated gradients, etc, etc) have been developed that both accelerate training (large nets can take weeks/months to train) and remove the need for some manual hyperparameter tuning.
- Garbage-In, Garbage Out. You're success in recognition tasks is only going to be as good as the feature representation available to the higher layers of your algorithms (whether conventional or neural net). Another recent advance has been substituting self-learnt feature representations for laboriously hand-designed ones, and the recent there is now a standard neural net recipe of autoencoders+sparsity for implementing this.
- And a whole bunch of other things...
As Newton said "if I have achieved great things it is by standing on the shoulders of giants".. there are all sorts of surprising successes (e.g. language translation) and architectural advances in neural nets that are bringing the whole field up.
These arn't your father's neural nets.