LLMs Face Rigorous Testing in Plastic Surgery Education
- •Researchers benchmarked 14 LLMs using 7,000 evaluations of plastic surgery training exams.
- •Proprietary models like Claude Opus 4.5 and GPT-5.2 Pro outperformed open-source competitors.
- •Study highlights that clinical reliability—not just raw accuracy—is critical for medical education AI.
The integration of artificial intelligence into specialized medical training is accelerating, but a new study warns that academic performance doesn't always equal clinical reliability. Researchers recently conducted a comprehensive benchmark of 14 large language models (LLMs) against the Plastic Surgery In-Service Training Examination (PSITE). While headline accuracy numbers provide a quick metric, this study dug deeper into consistency, measuring how stable these models are across multiple, independent runs.
The findings reveal a clear tiering in the landscape. High-end proprietary models, such as Claude Opus 4.5 and GPT-5.2 Pro, dominated the field with accuracy scores topping 90%. However, the research team emphasized that accuracy is only half the battle. They introduced metrics like the Coefficient of Variation to capture 'stochastic instability'—the erratic behavior where a model might give different answers to the same question over time.
For students looking toward the future of medical education, this research provides a vital reality check. While LLMs are clearly capable of absorbing vast amounts of specialized medical knowledge, their tendency to vary in reliability means they cannot yet be treated as foolproof tutors. As we integrate these tools into professional training environments, developers and educators must prioritize stability and consistency alongside raw performance scores to ensure these systems are actually safe for clinical use.