TUESDAY, Oct. 8, 2024 (HealthDay News) -- AI isn’t ready to run a hospital’s emergency room just yet, a new study concludes. ChatGPT likely would ask for unnecessary x-rays and antibiotics for some patients, and admit others who don’t really need hospital treatment, researchers reported Oct.

8 in the journal Nature Communications . “This is a valuable message to clinicians not to blindly trust these models,” said lead researcher Chris Williams , a postdoctoral scholar with the University of California, San Francisco. “ChatGPT can answer medical exam questions and help draft clinical notes, but it’s not currently designed for situations that call for multiple considerations, like the situations in an emergency department,” Williams added in a UCSF news release.

For the new study, researchers challenged the ChatGPT AI model to provide the sort of recommendations an ER doctor would make after initially examining a patient. The team ran data from 1,000 prior ER visits past the AI, drawn from an archive of more than 251,000 visits. The AI had to answer “yes” or “no” as to whether each patient should be admitted, sent for X-rays or prescribed antibiotics.

Overall, ChatGPT tended to recommend more services than were actually needed, results showed. The ChatGPT-4 model was 8% less accurate than human doctors, and ChatGPT-3.5 was 24% less accurate.

This tendency to overprescribe might be explained by the fact that the AI models are trained on the internet, Will.