When you leave hospital, the discharge summary written about your stay acts as the bridge between the hospital and your GP. These often involve requesting your GP to perform some follow-up action, e.g. “check bloods in 1 week,” or “review medication”.

But these aren’t always followed.

In my new HDR UK poster, I explored if the language used by hospital clinicians affects this. The result? While politeness doesn’t make a difference, language precision does, and that an AI model can predict which follow-up requests are likely to be missed.

The Problem

Hospital discharge summaries often include requests for primary care, e.g. “check bloods in 1 week,” “review medication,” and so on. But studies show that these actions are inconsistently followed.

Sometimes the request seems unnecessary, sometimes the patient doesn’t attend, and sometimes the workload in primary care just makes it hard to keep up.

We wanted to know:

  • Does the way a request is written affect whether it’s completed?
  • Can AI help identify which requests are at risk of being missed?

Some Data

We gathered thousands of discharge summaries from Salford Royal Hospital, and used a langauge model to extract requests from the text:

Diagram showing how requests can be extracted from free-text. A discharge summary is turned into structured data extracting both the request name and the timeframe requested for completion

We can then compare the requests to primary care records to see if and when any blood tests were actually done:

Adding actual timeframe

We label any requests where the blood test is done within 7 days of the requested timeframe as “concordant”:

Adding concordance

Predicting Concordance

Now we have a dataset of requests and concordance, we can train a model.

Diagram showing how a discharge summary can be passed into the model and outputs "concordant"

The model predicts concordance directly from request text with 83% accuracy.

These metrics indicate that on internal validation, the model utilising linguistic features was both accurate and robust.

Investigating the patterns the model learned, we found that:

  • Precise, clinical language (words like “renal,” “potassium,” “kinase”) predicted higher follow-through.
  • Conditional or vague terms (“if possible,” “consider checking”) predicted non-concordance.
  • Surprisingly, words of gratitude (“thanks,” “thank you”) often appeared in requests less likely to be completed, possibly because they softened uncertain or tentative instructions.

Why it matters

The fact that we can accuratly predict concordance suggests that how we write discharge summaries affects how care is delivered.

More importantly, an AI model can flag which requests are unlikely to be actioned, allowing clinicians to write better discharge summaries.

Imagine an EPR system that quietly warns:

“This request is phrased in a way that historically leads to low follow-through.”