A friend sent this article to me today.
It feels like a good moment to remind us: the LLM outperformed physicians on STRUCTURED cases.
Someday the AIs may perform better than us in the unstructured, messy, real-world patient encounter...but that day is not today.🩺🛟
https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395
It feels like a good moment to remind us: the LLM outperformed physicians on STRUCTURED cases.
Someday the AIs may perform better than us in the unstructured, messy, real-world patient encounter...but that day is not today.🩺🛟
https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395
Comments
1. It mixed multiple specialties in a way that wasn’t helpful
2. It had a small sample size, which included trainees
3. They provided one vignette as an example and the differential diagnosis wasn’t an EM ddx
https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825399
@jama.com @sumantranji.bsky.social
All patients are not good communicators.
They remember things later and call back.
The power distance between MD and patient may mean that patients hold back essential info.
And a thousand other things.
LLM = not ready for prime time.
Thanks @meganranney.bsky.social for your kind words about my editorial
AIs are poor at recognizing data gaps and more prone to assume based on past data.
Naturally elevates its tendency to fish out the associations underlying uncommon disease patterns.
http://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/
LLMs have promise
but they have not proven themselves yet, at least not for triage
Better data science and mixed methods will do better.
Where can one get best current data on the safety of AI use for diagnostics with break up of type of information it used on?