Forgot your password?
typodupeerror

Comment Re:All radiologists do is analyze digital images (Score 1) 89

Which is peanuts compared to the recurring $500-600k cost of a radiologist's annual salary.

It's not peanuts at all, which is precisely why there is also competition in the medical industry to reduce radiologists, and push more work onto larger centralized corporately hosted radiology mills.

I don't get the sense that these models are ready to replace radiologists yet

They're not even close.
LLMs outperform doctors in diagnostics. ImageNets do not outperform them clearly*.

but the closer we get the more tempting it'll be for individuals like the CEO quoted in this article. The ROI is massive.

Radiologists today are being "replaced" (in that more work is piled onto less radiologists now augmented by AI)
That is still replacement.

* there are some cases where ImageNets do outperform radiologists, and there are also cases where they do significantly poorly.
VLMs also have a very noted problem in the way that they're trained that leads them to producing reports that are quite bad when evaluated by other radiologists (implying the trainers aren't using radiologists to train the output)

Comment Re:All radiologists do is analyze digital images (Score 1) 89

Physicians are also only as good as their training data.

It's inherently different.
The ImageNet is trained to recognize images, and pair it with contextual information given in its scanning.
I.e., it is- at best- distilled from the knowledge of rad techs.

On the other hand, it's pretty easy (i.e. cheap) to train an image classifier on orders of magnitude more cases than any pathologist or radiologist could ever see in their lifetime.

Indeed. Many ImageNets have been trained on more dogs than I will ever see in my lifetime, and will still- a few percent of the time- call them a duck.

I find it funny that you start off talking about how doctors don't stay up on medicine (which is very true), and then point out that these aren't LLMs.
LLMs used in diagnostics regularly outperform doctors for precisely the reason you mentioned.

However, image classifiers match doctors at best, and are sometimes much worse in the case of false positives, most likely due to their inability to consider broader context than the image they're looking at.
This doesn't mean they don't have a place, but it's not real clear what it is yet.
Studies are mixed. Sometimes doctors augmented with AI do worse (suspected anti-AI bias), sometimes they do considerably better. Sometimes doctors alone outperform AI (critical miss rate on low sensitivity scans), sometimes worse- critical miss rate on high-sensitive scans.

Doctors are not liable for missing something on a scan. They're liable for negligence in interpreting a scan.

Comment Re:All radiologists do is analyze digital images (Score 1) 89

The types of neural networks used in these classifiers are not LLMs.

Correct.

The expense and difficulty of collecting an image set is going to far outweigh the compute time used to train them.

Like any network, how good it is is a mix of how much data you throw at it, and how many parameters it is.
Large ImageNets cost millions to train.

Comment Re:Clean room? (Score 5, Interesting) 111

Even if you use an AI to extract an extremely condensed specification out of the source code, it's hardly clean room if the LLM was pre-trained on the source code any way.

I once worked at a place that had a clean room process to create code compatible with a proprietary product. Anybody who had ever seen the original code or even loaded the original binary into a debugger was not allowed to write any code at all for the cloned product. The clone writers generally worked only off of the specifications and user documentation.

There were a handful of people who were allowed to debug the original to resolve a few questions about low-level compatibility. The only way they were allowed to communicate with the software writers was through written questions and answers that left a clear paper trail, and the answers had to be as terse as possible (usually just yes or no). Everyone knew that these memos were highly likely to be used as evidence in legal proceedings.

I highly doubt that any AI tech bros have ever been this rigorous, and I'd bet that most of these AIs have been trained on the exact same source code that they are cloning.

Comment Re:hmmm (Score 1) 64

Lol, now you're accusing me of not writing my own posts here? Hilarious!

Not an accusation in the slightest. Was giving you an out for disclaiming authorship of that post ;)
It is impressively ignorant.

Show evidence. Like I did.

You showed evidence of popularity of a thing within a tiny non-representative sample in a discussion about feature parity. Put scientifically- you showed nothing.
As mentioned, no amount of Claude Code's codebase is going to make you give OpenCode more github stars. Idiots are led by a different metric than code.

Comment Re:All radiologists do is analyze digital images (Score 2) 89

Problem with "AI and digital images" is that classifier is only as good as its training data.
The cost to train goes up dramatically the more training data you give it. This means there is a financial incentive to not train on edge cases, which means your AI *will not catch them.*

Like all things, it's not the tech I fear. It's the executives in charge of monetizing it.

Comment Re:Here it comes (Score 1) 69

They don't, and it isn't.
There is no "Kessler event". You're imagining some shit like the movie Gravity, but that's not Kessler Syndrome. It's just good graphics.

Kessler Syndrome is merely the state of increasing orbital debris from ongoing impacts being greater than that reduced by atmospheric drag.
By some measures, we're already well within a Kessler Syndrome. Station Keeping sats, or sats that can otherwise dodge debris are just fine.

To quote an AC quoting a wise man,

As a wise man once said, "Space is big. You just won't believe how vastly, hugely, mind-bogglingly big it is. I mean, you may think it's a long way down the road to the chemist's, but that's just peanuts to space."

Some kind of hypothetical future situation- a "Super Kessler" where even mobile sats cannot possibly dodge the amount of debris up there is just bad science fiction.

Comment Re:Sloppary (Score 1) 64

lol- the data points, which are like pointing out the sky is blue except for 1 of them, are not what was referred to, you illiterate buffoon.
It are the fluff opinions- the sound bites- that are devoid of fact.
Read again:

The rest, and ChatGPT's opinion on them is just padding your word count:
"This is the brain of the operation." "leveraging its dead code elimination for feature flags and its faster startup times." "there's a command system as rich as any IDE."

And of course, they're also simply not fact-based in the slightest.

This is why elementary school is important.

Comment Re:hmmm (Score 1) 64

Christ. Only in 2026 would I have to do this.

1) We weren't talking about a popularity contest. Though I do imagine Claude Code definitely had the most mentions by respondents to the, uhhh /me glances at notes, "pragmaticengineer.com survey". OpenClaw has more stars on github than software with tens of millions more in installed base than it.
2) After this comment, I'm not sure you should ever reply to anything technical ever again.

If there's anything that can be learned here, it's that people like you love having opinions. Even when shit like "My sense of it is that this code is not the LLM itself, it is the infrastructure and interface layer between the user and the LLM." has escaped your fingers on a public forum. Did an LLM write that? Be honest.

Claude and OpenCode have near feature-parity. I don't think anything in Claude Code's codebase is going to help it win your github stars. Maybe more flashy lights and buzzwords might, though.

Comment Re:hmmm (Score 1) 64

Context engineering for modern agents. The distinction is what you can imagine. In context engineering, you're dynamically altering the entire context window instead of just prompts added to them.

As for "of value", perhaps the system and agent prompts. That's a pretty fast-moving target. But since Claude has locked down Opus access to just API keys for external tools, those are really just for trying to glean ideas from.

Comment Re:TypeScript? (Score 1) 64

Incorrect.

There is nothing in it that should demand high performance, I agree.
But Claude Code is never waiting on your input. It's waiting on its frame timer to expire so it can render the next frame of its interface. Because decisions... were made... my people of questionable wisdom.

It's a real time React renderer that runs at 60fps.

Slashdot Top Deals

Your code should be more efficient!

Working...