Comment Re:Ironically, this Slashdot summary title is a li (Score 1) 104
Which it was.
Which it was.
There was a time when the people who complained about soldered RAM (and I was one of those people) were a significant enough proportion of the community that manufacturers would pay attention. This was the age when gaming PCs were constructed from high end pieces from the wild-assed cases to the heavy duty PSUs to overclocked CPUs and next gen GPUs.
But overall, that segment of the consumer market has dwindled. Most folks just want to charge their new machine up, connect it to their WiFi network and get going. On the corporate end of things, save for pretty niche areas like engineering and R&D, a cube you can plug a keyboard, mouse and camera into and will last through a few upgrade cycles before it's sold back to a refurb outfit is all that is needed. Nobody in IT departments is pulling RAM chips anymore, particularly at RAM prices right now! Even the folks writing operating systems are starting to get it, and have rediscovered the glory of native apps that don't required bloated Javascript engines just to select a few radio buttons.
Yes, Windows 11 is really that bad. It's cluttered, slow, inconsistent. I've seen it on pretty high end hardware, and it's a dog. And that's before we even talk about how they tried to insert Copilot into everything. It's a shitty version of Windows and even Redmond acknowledges it. It was the impending EOL of Windows 10 that lead me to buy an M1 MacBook Pro, and I've never looked back. If I want to run Linux, I've got servers set up to do that kind of heavy lifting, but I have absolutely no need for whatever it is MS is trying to sell me these days.
Since pertinent information was withheld (that it didn't know), then by your own post you acknowledge it was a lie of omission.
The stupidity of people these days is truly beyond belief. And, yes, get the f off my lawn.
We learned back in the 80s that trying to get a neural net to emphasise what you want is actually very difficult. What it will tend to emphasise are the assumptions that underly the test data, and that's usually a completely different sort of fiction.
But was that figure provided by AI?
Even if not, we all know that 793% of all statistics are invented.
If something is inaccurately presented as being the truth, then it is a lie of omission because it is dishonest about the fact that the information isn't actually known.
Gemini is exceptionally bad, as LLMs go. I really have no idea why it is so dreadful, even compared to other LLMs. It isn't context window. and it doesn't seem to be training material either.
Cyber Implications have been noted. Mondas security is to be Cyber Vibed until we have Cyber Security capable of defeating The Doctor.
When I test the different AI systems, Google's AI system loses track of complex problems incredibly quickly. It's great on simple stuff, but for complex stuff, it's useless.
Unfortunately.... advice, overviews, etc, are very very complex problems indeed, which means that you're hitting the weakspot of their system.
I've designed a few machines - some rather more insane than others - in meticulous detail using AI. What I have not done, so far, is get an engineer to review the designs to see if any of them can be turned into something that would be usable. My suspicion is that a few might be made workable, but that has to be verified.
Having said that, producing the design probably took a significant amount of compute power and a significant amount of water. If I'd fermented that same quantity of water and provided wine to an engineering team that cost the same as the computing resources consumed, I'd probably have better designs.But, that too, is unverified. As before, it's perfectly verifiable, it just hasn't been so far.
If an engineer looks at the design and dies laughing, then I'm probably liable for funeral costs but at least there would be absolutely no question as to how good AI is at challenging engineering concepts. On the other hand, if they pause and say that there's actually a neat idea in a few of the concepts, then it becomes a question of how much of that was ideas I put in and how much is stuff the AI actually put together. Again, though, we'd have a metric.
That, to me, is the crux. It's all fine and well arguing over whether AI is any good or not (and, tbh, I would say that my feeling is that you're absolutely right), but this should be definitively measured and quantified, not assumed. There may be far better benchmarks than the designs I have - I'm good but I'm not one of the greats, so the odds of someone coming up with better measures seems high. But we're not seeing those, we're just seeing toy tests by journalists and that's not a good measure of real-world usability.
If no such benchmark values actually appear, then I think it's fair to argue that it's because nobody believes any AI out there is going to do well at them.
(I can tell you now, Gemini won't. Gemini is next to useless -- but on the Other Side.)
This means you shoud NOT, under any circumstance, run Claude at 88mph. Unless you really want to.
I was hoping at the bottom of the article it would say that Professor Utonium accidentally added Chemical X.
The fines should be proportional to actual damage caused (ie: 100% coverage of any interest on loans, any extra spending the person needed to do in consequence, loss of compound interest, damage to credit rating along with any additional spending this resulted in, and any medical costs that can reasonably be attributed to stress/anxiety). It would be difficult to get an exact figure per person, but a rough estimate of probable actual damage would be sufficient. Add that to the total direct loss - not the money that went through any individual involved, and THEN double that total. This becomes the minimum, not the maximum. You then allow the jury to factor in emotional costs on top of that.
In such cases as this, the statutary upper limit on fines should not apply. SCOTUS has repeatedly ruled that laws and the Constitution can have reasonable exceptions and this would seem to qualify.
If a person has died in the meantime, where the death certificate indicates a cause of death that is medically associated with anxiety or depression, each person invovled should also be charged with manslaughter per such case.
I suggest:
First offence: Have to watch CSPAN for 5 hours a day, for a week, without sleeping through it - evidence to be provided in court
Second offence: Have to sing Miley Cyrus songs and Baby Shark on TikTok - sober
Third offence: License to practice and all memberships of country clubs and golf courses revoked
"I have not the slightest confidence in 'spiritual manifestations.'" -- Robert G. Ingersoll