
Hippocratic Is Building a Large Language Model For Healthcare 36
An anonymous reader quotes a report from TechCrunch: AI, specifically generative AI, has the potential to transform healthcare. At least, that's the sales pitch from Hippocratic AI, which emerged from stealth today with a whopping $50 million in seed financing behind it and a valuation in the "triple-digit millions." The tranche, co-led by General Catalyst and Andreessen Horowitz, is a big vote of confidence in Hippocratic's technology, a text-generating model tuned specifically for healthcare applications.
Hippocratic -- hatched out of General Catalyst -- was founded by a group of physicians, hospital administrators, Medicare professionals and AI researchers from organizations including Johns Hopkins, Stanford, Google and Nvidia. After co-founder and CEO Munjal Shah sold his previous company, Like.com, a shopping comparison site, to Google in 2010, he spent the better part of the next decade building Hippocratic. "Hippocratic has created the first safety-focused large language model (LLM) designed specifically for healthcare," Shah told TechCrunch in an email interview. "The company mission is to develop the safest artificial health general intelligence in order to dramatically improve healthcare accessibility and health outcomes."
Shah emphasized that Hippocratic isn't focused on diagnosing. Rather, he says, the tech -- which is consumer-facing -- is aimed at use cases like explaining benefits and billing, providing dietary advice and medication reminders, answering pre-op questions, onboarding patients and delivering "negative" test results that indicate nothing's wrong. [...] Shah claims that Hippocratic's AI outperforms leading language models including GPT-4 and Claude on more than 100 healthcare certifications, including the NCLEX-RN for nursing, the American Board of Urology exam and the registered dietitian exam. Hippocratic aims to have its LLM detect tone and communicate empathy better than its rivals -- in part by "building in" good bedside manner, says Shah. They have designed a benchmark to evaluate their model's humanistic qualities, and it scored higher than other models, including GPT-4.
As for whether or not it can replace a healthcare worker, Hippocratic argues that their models, trained under medical professionals' supervision, possess high capabilities.
"We're only releasing each role -- dietician, billing agent, genetic counselor, etc. -- once the people who actually do that role today in real life agree the model is ready," Shah said. "In the pandemic, labor costs went up 30% for most health systems, but revenue didn't. Hence, most health systems in the country are financially struggling. Language models can help them reduce costs by filling their current large level of vacancies in a more cost-effective way."
Hippocratic -- hatched out of General Catalyst -- was founded by a group of physicians, hospital administrators, Medicare professionals and AI researchers from organizations including Johns Hopkins, Stanford, Google and Nvidia. After co-founder and CEO Munjal Shah sold his previous company, Like.com, a shopping comparison site, to Google in 2010, he spent the better part of the next decade building Hippocratic. "Hippocratic has created the first safety-focused large language model (LLM) designed specifically for healthcare," Shah told TechCrunch in an email interview. "The company mission is to develop the safest artificial health general intelligence in order to dramatically improve healthcare accessibility and health outcomes."
Shah emphasized that Hippocratic isn't focused on diagnosing. Rather, he says, the tech -- which is consumer-facing -- is aimed at use cases like explaining benefits and billing, providing dietary advice and medication reminders, answering pre-op questions, onboarding patients and delivering "negative" test results that indicate nothing's wrong. [...] Shah claims that Hippocratic's AI outperforms leading language models including GPT-4 and Claude on more than 100 healthcare certifications, including the NCLEX-RN for nursing, the American Board of Urology exam and the registered dietitian exam. Hippocratic aims to have its LLM detect tone and communicate empathy better than its rivals -- in part by "building in" good bedside manner, says Shah. They have designed a benchmark to evaluate their model's humanistic qualities, and it scored higher than other models, including GPT-4.
As for whether or not it can replace a healthcare worker, Hippocratic argues that their models, trained under medical professionals' supervision, possess high capabilities.
"We're only releasing each role -- dietician, billing agent, genetic counselor, etc. -- once the people who actually do that role today in real life agree the model is ready," Shah said. "In the pandemic, labor costs went up 30% for most health systems, but revenue didn't. Hence, most health systems in the country are financially struggling. Language models can help them reduce costs by filling their current large level of vacancies in a more cost-effective way."
Good luck with that (Score:3)
No matter how empathetic the computer appears, people are not going to like getting bad news from a machine.
In America they won't have a choice (Score:4, Insightful)
Re: (Score:3)
Re: (Score:2)
Then you are fucked. I have a right to see an actual MD whenever I think it is necessary.
Re: (Score:2)
you got leprosy
Re: (Score:2)
No matter how empathetic the computer appears, people are not going to like getting bad news from a machine.
Well - no, I'm sure they don't. I don't like getting bad news from anyone for that matter. 8^)
Re: (Score:2)
No matter how empathetic the computer appears, people are not going to like getting bad news from a machine.
Doctor: I have good news and bad news
Patient: Let's get the bad news first
Doctor: you have operable brain cancer
Patient: What's the good news?
Doctor: I'm not going to have to take out a loan for the Ferrari I'm buying next week.
Re: (Score:1)
Doctor: I have good news and bad news
Patient: Uh-oh. Let's hear the good news first.
Doctor: You're going to be dead in 24 hours.
Patient: My God! That's the GOOD news? What's the bad news?
Doctor: I should have told you this yesterday.
Re: (Score:2)
I thought it goes like this:
Doctor: I have good news and bad news
Patient: Let's get the bad news first
Doctor: you have Alzheimer.
Patient: Oh my gosh, and what is the good news?
Doctor: you will have forgotten this conversation when you ar back home.
Re: (Score:2)
The article specifically spells out that the AI is intended only to be delivering routine negative test results that indicated no problem.
I used to write guidelines for how hospital systems communicated about certain conditions. A common case would be Pap smear results. A women who's been sent for a pap can be anxious about the results, right? Relying on human communications about negative tests can involve more time -- we identified one hospital where the mean was around 2 extra days! there was variation u
Re: (Score:2)
An "AI" generating more immediate communications about those negative test results seems like pointless overkill; why not:
Dear ,
We are happy to report that your pap smear results all came back negative. A negative result means that there is nothing to worry about and everything is fine.
If you have any questions or concerns please contact our office.
You should schedule your next pap smear in .
Thanks for coming in,
For virtually all negative pap smear results that's sufficient. /SMS text/ro
And then mail merge
AI not needed here, Bad Idea (Score:2)
The article specifically spells out that the AI is intended only to be delivering routine negative test results that indicated no problem.
)
You don't need an "AI", much less a GPT chatbot, to inform a patient of a negative test result. Or a set of results that taken together give a prognosis.
It is easy to do that computation and select an appropriate communication of pre-written material that explains it all. This would also include links and FAQs. All of it is 100% reliable.
I suppose you could glorify that with an (e.g. rule-based) system and call it "AI". But none of that has nothing to do with anything GPT.
That would all be relatively minor
Re: (Score:3)
If you read the education curriculum for medical doctors and more specifically general practitioners, you'll find that the education is rapidly dated and due to the nature of universities, the amount of retained information is extremely low to begin with. There are medical professors at universities who manage to properly accumulate and maintain their medical educations, but general practitioners who typically
Re: (Score:2)
My healthcare company uses AI assist for its chat support agents (system suggests a response array, agent approves type model). It is absolutely atrocious, useless, and makes the agents worse than they already are since they are not US based. The âoerules for AIâ discussion yesterday pointed loud and clear at a regulatory proposal for companies using this crap: full disclosure that it is being used and an easy way to opt out on the customer side.
So you are now getting AI with an Indian accent eh?
ChatGPT: we hallucinate the em-PHA-sis on the wrong SYLL-able.
Regulation (Score:2)
My healthcare company uses AI assist for its chat support agents [...] It is absolutely atrocious, useless, and makes the agents worse than they already are
[...] pointed loud and clear at a regulatory proposal for companies using this crap: full disclosure [...]
"Regulation" again.
The problem is that it is impossible to characterize "AI" in such a way that it could be identified as subject to any regulation.
I'm not sure what to do about the problem, but all the current "Regulation" ideas are either useless "We're the Government and Don't Fear, We're Doing Something About It" plattitudes, or the early companies trying to write the laws so that they remain dominant.
The kind of regulation that would be good is something clarifying vast liability dangers for companies
Built by Doctors to do jobs Doctors don't do. (Score:2)
"Shah emphasized that Hippocratic isn't focused on diagnosing. Rather, he says, the tech -- which is consumer-facing -- is aimed at use cases like explaining benefits and billing, providing dietary advice and medication reminders, answering pre-op questions, onboarding patients and delivering "negative" test results that indicate nothing's wrong."
Benefits and Billing, handled by Revenue Cycle teams, at the healthcare institution. The other features dietary advice, medication reminders, pre-op and on boardin
Re: (Score:2)
Re: (Score:2)
You need to be more explicit as to what you mean.
As I read it you're expecting the system to know what the appropriate answer is without testing.
Perhaps you actually meant something more like "compare the success rates of different doctors on cases that are equally difficult", which is a more reasonable desire, but also requires information that NOBODY has access to. There's no easy way to compare whether two cases are equally difficult. You might be able to come up with a reasonable statistic, but it wou
HAL 9000 of Healthcare (Score:2)
Re: (Score:3)
What was that line from "Passengers"?
No treatment can meaningfully extend your life, these pills ( they rattle out of the dispenser) may ease your transition.
Re: (Score:2)
2001, a bit misquoting HAL.
More harm than good (Score:3)
What's already been happening for many years is that people who do not have medical degrees are second-guessing doctors' diagnoses and courses of treatment. Oh, and to add insult to injury, those people usually get paid a lot more than the doctors do. As an added bonus, those people are immune from any sort of malpractice lawsuit. Here's what's going to happen: the same people without a medical degree are going to use this AI system as justification for their second-guessing while doing even less work and getting paid even more. Don't believe me? See the recent slashdot story where a university professor asked AI to determine if students used AI in their papers which dutifully said yes and the students got screwed because the professor (and the overpaid healthcare bureaucrat) is the gatekeeper.
Re: (Score:3)
Very good point.
In case anyone is wondering, I believe parent is referring to insurance companies and Medicare where some bureaucrat decides if your doc's recommendation is going to be paid for.
However, this process is already 99% done by computer where the CPT (procedure) code is matched against the ICD (diagnosis) code. It's just table lookup. It's only in the rare cases when an appeal is made that a human even looks at it. In those cases, they usually just reject it, handling like 200 appeals a day.
Thank you for participating in civilization (Score:4, Insightful)
Your Consumer Score is sub-optimal.
Your Credit Score is sub-optimal.
Your Social Score is sub-optimal.
Your Health Score is sub-optimal
Probable outcomes using similar scores show that you will provide no further increase in productivity.
Please report to the nearest self-recycling unit. We appreciate our customers! Thanks for your business!
Re: (Score:2)
Please report to the nearest self-recycling unit. We appreciate our customers! Thanks for your business!
And please settle your bill and fill out the customer satisfaction feedback form before beginning your final treatment.
Explaining things (Score:2)
Almost all of these require human level emotional intelligence, an attribute completely lacking in LLM's. Tell a 60 year old man why he can't have red wine anymore, easy. Do it in a way that keeps him happy with the system and also stops him from drinking red wine, very difficult. Go through a bullet poin
Re: (Score:2)
Re: (Score:2)
It's basically an FAQ in dialogue format. If a question has already been asked enough times, it can provide a good answer, just like an FAQ.
It can (and often will) also provide a wrong answer, unlike an FAQ.
haha -- bedside manner (Score:2)
That's a good one. The AI will have good bedside manner. I can tell whether a doc gives a shit about me from his bedside manner, including whether he is pretending to "care" about me. I already know the AI doesn't give a shit about me, so no matter how "caring" it is, I still know it couldn't give a shit. They intend to replace information with flattery. Maybe that works for most idiots.
robots with a gleam in their eye (Score:2)
I for one look forward to receiving healthy enemas from a robotic AI enema nurse. I mean what could possibly go wrong?
Re: (Score:1)