Comment Re:I don't think he is talking about satellite lev (Score 1) 147
Have they even considered how much radiation hardening they need - even though they will still operate within the Van Allen belt, the increased radiation is a serious concern.
Have they even considered how much radiation hardening they need - even though they will still operate within the Van Allen belt, the increased radiation is a serious concern.
As a lawyer would say "not suitable for the purpose sold".
For a story about how Oxide avoided them, see "Holistic boot", at
https://rfd.shared.oxide.compu...
["Really quite horrid" is British for "<expletive deleted/> piece of <expletive deleted/> junk"]
Yup, same as the feedback loops in "cold readings"
Charlie Stross(@cstross@wandering.shop) wrote, in Mastadon:
The LLMentalist effect: Large Language Models replicate the mechanisms used by (fake) psychics to gull their victims: https://softwarecrisis.dev/let...
The title of the paper is "The LLMentalist Effect: how chat-based Large Language Models replicate the mechanisms of a psychic’s con"
Interests, like "I want a trailer hitch for my Subaru".
All that other stuff? Wanted for someone else.
It's just piggybacking the blame onto advertisers, because people don't like them anyway
I used to work in advertising, and I saw Google as the personification of "moral hazard" (which see). Other things? Way nicer.
If you scan a thousand British faces and compare them to a thousand criminals, you will do 1,000,000 comparisons. (that's the birthday paradox part).
If your error rate is 0.8%, you'll get roughly 8,000 false positives and negatives.
That's bad enough if they are all false positives: people get arrested, then released.
It's way worse if they are all false negatives: 8,000 criminals get ignored by the police dragnet.
That was Britain: false positives are life-threatening in countries where the police carry guns.
0.8% is a good error rate. 34% wrong is typical in matching black women. See
https://www.aclu-mn.org/en/news/biased-technology-automated-discrimination-facial-recognition#:~:text=Studies%20show%20that%20facial%20recognition%20technology%20is%20biased.,published%20by%20MIT%20Media%20Lab.
At a certain company long long ago, managers had a mainframe-based planning app that looked like a sort of spreadsheet.
The company did a study to see how much it improved the manager's team's productivity...
Oops! Use of the tools was correlated with declining productivity.
[A computer is] like an Old Testament god, with a lot of rules and no mercy. -- Joseph Campbell