People keep telling me to watch The Pitt, and now that the show is tackling generative AI in healthcare, I must. As Mashable’s tech editor, I field a lot of questions about generative AI, and I also spend a lot of time talking to people who are extremely enthusiastic about artificial intelligence, and people who are extremely hostile to AI.
So, what does the latest episode of The Pitt Season 2 get right about AI in medicine, and what does it get wrong?
How AI factors into The Pitt episode “8: 00 AM.”

Credit: Warrick Page / HBO Max
AI wasn’t the entire focus of Season 2, episode two, “8: 00 AM.” We also get to see a nun with gonorrhea in her eye, an unhoused man with a colony of maggots inside his moldy arm cast, and a clear view of an erection, which, of course, has lasted more than four hours.
However, one of the plotlines focuses on newcomer Dr. Baran Al-Hashimi’s (Sepideh Moafi) attempt to modernize the ER and integrate AI into patient care. She introduces an AI app that automatically listens to patient visits and summarizes the pertinent details in their charts.
When an excited student says, “Oh my god, do you know how much time this will save?”, Dr. Al-Hashimi has an answer: With the new AI app, the ER doctors will spend 80 percent less time charting. Later in the episode, the good doctor tells Dr. Robby (Noah Wyle) that it will also allow physicians to spend 20 percent more time with patients.
So far, so good, but the app immediately makes a mistake. It documents the wrong medication in the patient’s chart, substituting a similar-sounding medication instead.
This doesn’t diminish Dr. Al-Hashimi’s enthusiasm at all. “Generative AI is 98 percent accurate at present,” she says. “You must always carefully proofread and correct minor errors. It’s excellent but not perfect.”
AI transcription is really good (and it still makes mistakes).
Dr. Al-Hashimi states that generative AI is 98 percent accurate. But is this really true?
Fact-checking Dr. Al-Hashimi’s claim is actually tricky, because it’s not entirely clear what she means. If she’s referring purely to AI transcription, she’s closer to the truth. That’s a task that generative AI excels at.
Last year, a group of researchers conducted a systematic review of 29 studies that measured AI transcription accuracy specifically in healthcare settings. (You can check out the entire review, published in the BMC Medical Informatics and Decision Making journal.)
Mashable Light Speed
Some of those studies did find accuracy rates of 98 percent or higher, and some found significant time savings for doctors. However, those studies involved controlled, quiet environments. In multi-speaker environments with a lot of crosstalk and medical jargon, like you’d find in a crowded emergency room, accuracy rates were much lower, sometimes as low as 50 percent.
Still, rapid advancements in large-language models are improving AI transcriptions all the time. So, we could be generous and say Dr. Al-Hashimi’s claim is close to the truth, for the latest LLM models, in certain settings.
Generative AI is definitely not 98 percent accurate.
Let’s take the latest version of ChatGPT as an example. When the GPT-5.2 model was released a few months ago, OpenAI published documentation on the model’s tendency to hallucinate and provide false information.
According to OpenAI, its GPT-5.2 Thinking model has an average hallucination ra
