“Primary care doctors now spend, on average, two hours a day filling out patients’ electronic health records. From 2009 to 2018, EHR length has increased by 60 percent. And yet most of this work is simply documentation, not problem-solving or reasoning. Why take doctors away from their patients or hire other teams of humans, like medical scribes, to transcribe and enter all this data, when AI can now do it instead?

[..] they listen in on the conversation between a patient and physician, create a transcript, then organize the information from the transcript into the standard format doctors use, generating a note within minutes or even seconds of the clinical conversation. This is, for many medical professionals, promising, exciting, and nerve-wracking. AI is powerful, but it is also known for being biased, for making untrue statements with great confidence, and for its poor quality of training data; to say nothing of all the issues around models potentially not being up to date with the newest medical literature and guidelines. [..]

As flawed as AI notes may be, it’s entirely possible that human-generated medical notes could be more flawed. One study conducted at the VA found that when compared with secret recordings of medical encounters, 90 percent of doctor-written medical notes contained at least one error. Another study found that 96 percent of speech recognition–generated notes contained errors; 42 percent of final signed notes still had errors. An emergency room–based study found that physical exams are often “documented” even if they didn’t occur; only slightly above 50 percent of documented physical exams, by body system, were verified by an observer. And many issues raised by patients during visits don’t make it into the medical record.

Compared with this status quo, AI scribes don’t look so dangerous. Right now, there is already inaccurate information in medical records that is used to make clinical decisions, evaluate risk profiles, and develop new predictive algorithms. AI may reinforce biases, but those biases likely come from humans; if the human-authored texts on which the tools are based contain biased views of race or gender (which they often do), the resulting tools will as well.

None of this addresses the privacy risk of AI tools, which is serious. Recordings of your doctor’s appointments may be stored on a server hosted by a third party and at risk for breach. But your medical data is already at high risk for breach. Since 2021, there have been more than 700 health data breaches each year. The 703 in 2024 alone affected more than half of the U.S. population (over 181 million people). And if you wanted to ask your hospital not to use your data for AI or for research, you likely can’t. Sharing your data is a prerequisite for getting care at many, if not most, hospitals. The primary U.S. law governing health data, HIPAA, may be long overdue for an update, but AI tools for documentation aren’t really going to make what is already a bad situation any worse.

AI may be just as (un)reliable as our current methods of medical documentation, but it is also much, much faster. This speed can bring about many benefits. It may allow doctors more time with their patients. It may decrease doctors’ after-hours workload. It may help reduce rates of burnout among medical professionals, which the pandemic pushed to all-time highs. These are all good things. If AI can help doctors get back to the point of doctoring—focusing on their patients rather than documentation—and the notes it produces are no worse than what we already have, then that would be a step forward.”

Full article, SH Fischer and SL Gebauer, RAND, 2025.4.4