Teaching Hospital–Based Medical Education Quality Framework

Learn–Assess–Improve–Govern

What it is: A lightweight but rigorous structure that helps teaching hospitals and universities build a measurable medical education system—without over-engineering.

Where it fits: Designed for teaching hospitals and health sciences universities — including Medicine, Dentistry, and Dental Public Health — where clinical training, community placements, prevention programs, and population health initiatives require structured supervision, measurable assessment, continuous improvement, and accountable governance.

Why this framework exists

Common problems this addresses:

  • “We teach well” but outcomes and progression are unclear

  • Assessment varies by assessor and department

  • Feedback exists but does not drive improvement actions

  • Governance is ad-hoc; decisions are not data-driven

The four pillars

These four pillars connect learning to reliable assessment, continuous improvement, and leadership governance.

LEARN
  • Curriculum & rotation design (outcomes, competencies, logbook)

  • Clinical supervision & teaching (bedside teaching, feedback moments)

  • Learning resources (cases, simulation, guidelines, e-learning)

ASSESS
  • Competency assessment system (WBAs and OSCE as applicable)

  • Assessment quality (blueprinting, standard setting, inter-rater reliability)

  • Learner progression (entrustment decisions, remediation triggers)

IMPROVE
  • CQI cycle (audit → CAPA → re-audit)

  • Faculty development (train-the-trainer, assessor calibration)

  • Learning environment & safety (patient safety + learner safety)

GOVERN
  • Medical Education Governance (committee, roles, decision rights)

  • Quality & risk management (risk register, incident learning, escalation)

  • Transparency & reporting (dashboard, quarterly review, annual plan)

KPI starter set (example list)

Starter KPIs (customizable)

  • % rotations with defined outcomes + logbook requirements (target ≥ 90%)

  • Supervision ratio compliance (e.g., % shifts meeting supervision standard)

  • Faculty teaching engagement (% faculty meeting minimum sessions)

  • WBA completion rate per trainee per month (minimum threshold defined locally)

  • Assessor calibration coverage (% assessors calibrated/6 months)

  • Assessment reliability signal (e.g., moderation concordance %)

  • CAPA closure on time (% actions closed by due date)

  • Re-audit pass rate (% issues improved after 60–90 days)

  • Remediation success rate (% learners improved after support plan)

  • Quarterly Medical Education Quality Review held (Y/N + attendance rate)

  • Learner experience score (pulse survey, response rate + score)

  • Patient safety learning integration (# cases/M&M linked to learning objectives)

Optional Extension KPIs: Community / Dental Public Health (add-on)

  • % community/outreach sites meeting minimum training criteria (target ≥ 90%)

  • Prevention pathway adherence rate (target defined locally)

  • Follow-up/referral completion rate after screening (target defined locally)

  • Education delivery rate (% eligible encounters documented) (target ≥ 85% or local)

  • Underserved reach indicator (target defined locally)

  • Outreach IPC & safety compliance (target ≥ 95% or local)

Adoption pathway (90 days)

90-day adoption (high-level)

  • Weeks 1–2: baseline scan + gap map

  • Weeks 3–4: KPI definitions + dashboard skeleton

  • Weeks 5–8: pilot in 1–2 departments

  • Weeks 9–12: governance cadence + scale plan