Comment Re: Not helpful (Score 1) 28
This is the sound of the other shoe dropping.
This is the sound of the other shoe dropping.
We need to stop pretending like it's perfectly OK to film strangers in public. Legal? Sure. Should you be doing it? 9 times out of 10, no.
It's long past time we had a real debate about the law, too. Just because something has been the law for a long time, that doesn't necessarily mean it should remain the law as times change. Clearly there is a difference between the implications of casually observing someone as you pass them in a public street, when you probably forget them again a moment later, and the implications of recording someone with a device that will upload the footage to a system run by a global corporation where it can be permanently stored, shared with other parties, analysed including through image and voice recognition that can potentially identify anyone in the footage, where they were, what they were doing, who they were doing it with, and maybe what they were saying and what they had with them, and then combined with other data sources using any or all of those criteria as search keys in order to build a database at the scale of the entire global population over their entire lifetimes to be used by parties unknown for purposes unknown, all without the consent or maybe even the knowledge of the observed people who might be affected as a result.
I don't claim to know a good answer to the question of what we should allow. Privacy is a serious and deep moral issue with far-reaching implications and it needs more than some random guy on Slashdot posting a comment to explore it properly. But I don't think the answer is to say anything goes anywhere in public either just because it's what the law currently says (laws should evolve to follow moral standards, not the other way around) or because someone likes being able to do that to other people and claims their freedoms would be infringed if they couldn't record whatever they wanted and then do whatever they wanted with the footage. With freedom comes responsibility, including the responsibility to respect the rights and freedoms of others, which some might feel should include more of a right to privacy than the law in some places currently protects.
That all said, people who think it's cool to film other human beings in clear distress or possibly even at the end of their lives just for kicks deserve to spend a long time in a special circle of hell. Losing a friend or family member who was, for example, killed in a car crash is bad enough. Having to relive their final moments over and over because people keep "helpfully" posting the footage they recorded as they drove past is worse. If you're not going to help, just be on your way and let those who are trying to protect a victim or treat a patient get on with it.
The question was literally, "What happens when all car manufacturers are displaying ads?"
The escape of medical information is truly well under way already, independent of AI.
In the UK, most medical information will be classified as sensitive personal data, which means it has significant extra protections under our regular data protection law, in addition to the medical ethics implications of breaching patient confidentiality. Letting it escape is a big deal and potentially a serious threat to the business/career of any medical professional who does it. Fortunately the days of people sending that kind of data around over insecure email are finally giving way to more appropriate methods of communication as the technology improves. It's usually governments seeing pound signs and/or businesses who aren't providing direct care to the patients that are pushing for wider distribution (and also those organisations who act as if impossible claims about sanitising the data effectively before releasing it are true).
I'm serious. I don't fucking pay for ads. Ever.
Good for you! Unfortunately, for a lot of people, having no car isn't really an option, so the answer to what happens next with your strategy is really that all of those people get an inferior product because there's no effective competition or regulation in the market to prevent that, while people like you don't get any product at all.
What should happen is that governments recognise a failure of the market to maintain adequate standards for customers and introduce regulation to enforce minimum acceptable standards accordingly. Whether that actually happens obviously depends on whether your government is more interested in looking out for the people or the businesses.
Take sons or grandsons of these lobbyists, enlist them into army, arm them with newest and costliest equipment from their fathers' companies and send to Syria or some other place with bad climate and actual actions. I think parents would in one moment change their mind about right to repair.
So I'm all for evidence-based medicine as a starting point, but when you realize it isn't behaving normally, you should adjust accordingly.
The thing about adopting evidence-based policy is that you also need to review and if necessary change policy when more evidence becomes available. The kind of situation you're describing would surely qualify.
Hedge funds and private equity firms demurred on investing in NTP, reportedly claiming that the foundation didn't have a credible plan for improving engagement.
How on earth can you "exhaustively" deidentify millions of chat logs that could contain literally any personal details, and presumably all without OpenAI's own employees also sifting through personal information in exactly the way they're claiming would be bad if others did it?
...This is my embarrassed face.
I had previously assumed you were speaking of allocating $1M across all projects used by Google. In fact, you were speaking of giving $1M to each such project.
One would wonder what sorts of strings would be attached to such largesse. Still, that would indeed be game-changing and amazing.
Google could create a new corporate policy to provide a minimum of $1M/year to any open source project it uses.
That would be real innovation.
While acknowledging your noble intentions, no, it wouldn't be innovation. It would be cheaping out.
In the San Francisco bay area, $1.0E+06/year gets you maybe five skilled engineers. Set against the quantity of Open Source projects used by such organizations -- FFmpeg, GStreamer, OpenSSL, ssh, rsync, gcc, gdb, coreutils, nanopb, Samba, Lua, Python, Perl, Git, Vim/Neovim, Yocto, ImageMagick, Blender, the Pipewire framework, the Linux kernel, the Debian packaging system, etc. etc. etc. etc. etc... -- five engineers is miserly.
Google appears to have understaken the expense of spinning up an ocean-boiling slop machine to automagically generate plausible bug reports, and then casually fire off an email to the maintainers.
Note that Google has not undertaken the expense of assigning an engineer to also write a fix.
That they are not doing that is a conscious, management-approved choice.
...Y'know how Google relishes in closing bug reports with "WONTFIX - Working as designed?" I think FFmpeg should close slop reports from Google with, "WONTFIX - Unfunded."
"Hey, everyone! Don't pay any attention to those Japanese translators who'd been volunteering their time and expertise for the last 20 years that we just insensitively and comprehensively shit on... Look! New mascot logo! Giz cash..."
(Narrator: New revenues did not materialize.)
Aren't you glad you're not getting all the government you pay for now?