Why AI in healthcare needs a safety net
Ambient voice technology is transforming NHS documentation. Instead of spending hours typing up notes after patient consultations, clinicians can now speak naturally and have AI convert their words into structured medical documents. It’s a game-changer for reducing administrative burden, but there’s a catch that few people talk about.
A recent BMJ survey revealed that one in five GPs are already using generative AI tools like ChatGPT in their clinical practice, despite a lack of formal guidance or clear work policies. Of those using AI, 29% employ it to generate documentation after patient appointments and 28% use it to suggest potential diagnoses.
“When we started exploring AI for clinical documentation in 2022, we discovered something fascinating,” explains Dr Andrew Whiteley, Managing Director of Lexacom. “Large language models are incredibly good at being helpful, sometimes too helpful. They’ve been trained to recognise patterns and provide useful summaries, which is brilliant for most applications. But in medicine, there’s a fundamental difference between recording what a patient says and interpreting what it means.“
The helpful AI problem
Think about how you might describe a drink to someone, if you said, “juniper based spirit with carbonated quinine water“, most people would helpfully summarise that as “a gin and tonic“. Large language models do the same thing, they’ve learned this is the ‘helpful’ response, recognising patterns and generating the most probable response.
In everyday contexts, this is fantastic. In medical documentation, it becomes problematic. When a patient describes symptoms like chest discomfort and shortness of breath, the AI might helpfully write “patient presents with cardiac symptoms” in the notes. It’s made a reasonable connection, but it’s crossed an important line, only qualified clinicians should interpret symptoms and suggest diagnoses.
This isn’t the AI making things up or ‘hallucinating’ in the traditional sense. It’s doing what it was designed to do, being helpful by recognising patterns. The challenge is that medical records need to be precise, not helpful interpretations.
But there’s another, equally serious problem: what AI leaves out. “A model might brilliantly summarise a 30-minute consultation into a concise letter, but if it’s dropped the patient’s penicillin allergy or their history of adverse reactions, that ‘helpful’ summary becomes dangerous”, notes Dr Whiteley. “The AI thinks it’s being helpful by condensing information, but in medicine, what’s left unsaid can be as dangerous as what’s wrongly added.”
