A Framework for AI Integration in GP Training

Ambient Voice Technologies (AVTs)

Summary
The rapid adoption of AI scribes in the form of Ambient Voice Technologies (AVTs) and Large Language Models (LLMs) for synthesising and summarising clinical notes presents a dual opportunity for GP trainees who will be our future GP workforce: reducing administrative burnout and a risk of eroding “clinical “muscle memory.”

This paper proposes a tiered roadmap to ensure GP trainees develop core consultation and reasoning skills before augmenting them with AI.


Objective
To establish a tiered competency model for the use of LLMs and AVT scribes.



Educational Rationale
The primary risk of early-stage AI adoption in GP training is “Cognitive Offloading.”

If a trainee uses an AI scribe before they have mastered the art of history-taking and clinical synthesis and summarisation, they fail to build the foundational levels of Bloom’s Taxonomy.

By aligning AI access with the trainee’s progression through the taxonomy, we ensure that technology augments rather than replaces clinical intelligence.


Bloom’s Taxonomy
This is a hierarchical framework used by educators to classify learning objectives into levels of complexity and specificity.

Developed in 1956 (and revised in 2001), it rests on the idea that you cannot effectively perform higher-level cognitive tasks (like designing an AI workflow) until you have mastered the foundational ones (like understanding the clinical guidelines).

In the context of GP Training, Bloom’s taxonomy explains why an ST1 shouldn’t jump straight into AI scribing which would lead to fragility in expertise.

Here is an example hierarchy:

LevelActionApplication to GP Training
1. RememberRecall facts and basic conceptsMemorising the Red Flags for eg a headache
2. UnderstandExplain ideas or conceptsExplaining why a certain medication is contraindicated
3. ApplyUse information in new situationsConducting a consultation and writing a manual note
4. AnalyseDraw connections among ideasComparing an AI-generated note against a manual one to find errors
5. EvaluateJustify a stand or decisionCritiquing the safety of an AI tool’s summary of a complex patient
6. CreateProduce new or original workDeveloping new AI protocols for a PCN or Practice


The ST-Phased Roadmap for AVT use within live clinical consultations

Training YearTaxonomy LevelAI Access StatusPrimary Objective
ST1Remember & UnderstandProhibitedBuild “Clinical Muscle Memory” and manual note-taking skills
ST2Apply & AnalyseControlledDevelop “Human-in-the-Loop” scepticism and prompt engineering
ST3Evaluate & CreateIntegratedMaster efficiency and ethical oversight for post-CCT life


Phase I: ST1 – The “Analogue” Foundation
Policy
No use of AI scribes or LLMs for live clinical documentation. Focus on developing their “internal scribe”.
Rationale
To ensure the trainee can identify red flags and confidently complete a concise clinical entry into the electronic medical record independently. Will require ability to listen, summarise and record a patient history whilst maintaining rapport.
Key Activity
Education on mandatory modules on AI Literacy (eg how LLMs work). Trainees learn the mechanics of “hallucinations” and the governance of data privacy (GDPR/DPIA/DCB) without using them on live patients.

Phase 2: ST2 – Analytical Integration
Policy
Supervised use by clinical trainer during tutorials and Consultation Observation Tools (COTs).  
Rationale
Trainees begin to distinguish between their own clinical findings and AI-generated summaries, once clinical reasoning is established and confirmed by supervisor.
Key Activity
The “Parallel Note Audit.” Trainees write a note manually, then generate one via AI, and critique the differences. This builds the ability to spot omissions in automated text.
“Prompt Engineering” Learning to use scribes for onward tasks (eg letter summarising).
“Informing patients” Developing proficiency in informing patients of the use of scribe within a consultation.

Phase 3: ST3 – Evaluative Mastery
Policy
Full use of practice-approved, NHS-accredited AI scribes for routine clinics. Preparing for post CCT.
Rationale
At this stage, the trainee has the mastery to appraise the safety of AI-generated plans and design efficient workflows for high-volume General Practice.
Key Activity
Mastery of informing patients. Learning how to explain the “Ambient Listener” to patients while maintaining the confidence of the doctor-patient relationship.
Demonstrate ability to check scribe output comprehensively.
Understanding the liability of generating notes and entering in to the electronic patient record.


Conclusion
In conclusion, the integration of AI into GP training must be deliberately informed, staged, and grounded in educational theory to avoid undermining the development of core clinical competencies.

By aligning access to AI tools such as AVTs and LLMs with progression through Bloom’s Taxonomy, this framework ensures that trainees first build essential skills in history-taking, clinical reasoning, and documentation before augmenting their practice with technology.

The proposed ST-phased roadmap balances the benefits of AI efficiency, reduced administrative burden, and scalability — with the need to preserve clinical “muscle memory” and professional judgement.

Ultimately, this approach positions future GPs not as passive users of AI, but as critical, informed practitioners capable of safely integrating these tools into high-quality patient care.

Dr Osman Bhatti MBBS FRCGP
GP Trainer & Programme Director
Digital Health Lead
Clinical Consultant – Accurx

TAGS

One response

  1. This is a thoughtful and much-needed contribution — particularly the framing around cognitive offloading and the use of Bloom’s taxonomy, which is exactly the right educational lens for this problem.

    The “parallel note audit” in ST2 is especially strong. That is where real clinical judgement is forged — not by avoiding AI, but by learning how and when to challenge it.

    Where I think this becomes even more interesting is when we zoom out from training into the system we are actually preparing trainees for.

    We are no longer training doctors for a purely human documentation environment. We are training them for an AI-augmented system where:

    documentation is medico-legal, not just cognitive
    time pressure is extreme
    and administrative burden is already distorting clinical work

    So there are really two competing risks here:

    Cognitive fragility from premature AI reliance
    System inefficiency and burnout from not adopting AI fast enough

    The NHS is already feeling the second quite acutely.

    A couple of areas that might strengthen this further if it were to inform national guidance:

    Moving from stage-by-year → stage-by-competence
    Not all ST1s are equal, and progression based on demonstrated capability (note quality, clinical reasoning, ability to detect AI error) may be more robust than training year alone
    Bringing governance earlier into the model
    Patient consent, auditability, and error detection probably need to start at first exposure, not just at ST3
    Adding objective metrics
    If this is to scale, we’ll need to show:
    does AI improve note quality?
    does it reduce error or introduce new ones?
    what happens to consultation quality and time?

    The core principle you’ve outlined — competence before convenience, augmentation not replacement — feels exactly right.

    The next step, I suspect, is translating this from an educational framework into an operational one that works at NHS scale.

    Curious to see how this evolves — it’s a space that’s moving faster than most training structures are designed to handle.

Leave a Reply to Dr Nigel Beejay Cancel reply

Your email address will not be published. Required fields are marked *