Research Paper ML Hub

Proceedings of the 24th Workshop on Biomedical Language Processing / 2025

LLMs as Medical Safety Judges: Evaluating Alignment with Human Annotation in Patient-Facing QA

Yella Diekmann, Chase Fensore, Rodrigo M. Carrillo-Larco, Eduard Castejon Rosales, Sakshi Shiromani, Rima Pai, Meghan Shah, Joyce C. Ho

AI Safety

The increasing deployment of LLMs in patient-facing medical QA raises concerns about the reliability and safety of their responses. Traditional evaluation methods rely on expert human annotation, which is costly, time-consuming, and difficult to scale. This study explores the feasibility of using LLMs as automated judges for medical QA evaluation. We benchmark LLMs against human annotators across eight qualitative safety metrics and introduce adversarial question augmentation to assess LLMs’ robustness in evaluating medical responses. Our findings reveal that while LLMs achieve high accuracy in objective metrics such as scientific consensus and grammaticality, they struggle with more subjective categories like empathy and extent of harm. This work contributes to the ongoing discussion on automating safety assessments in medical AI and in-forms the development of more reliable evaluation methodologies.

4 citations1 influential

Full paper

Read the original paper

A direct open-access PDF is not available in the database yet. Use the source page or learning resources below to open the complete paper from the publisher or index.