Ran Score: a LLM-based evaluation metric promises to rethink radiology-report scoring
What the paper proposes
Researchers have posted a new preprint, arXiv:2603.22935, introducing "Ran Score" — a large language model (LLM)-based framework for evaluating chest X‑ray report generation. The authors argue that current automated metrics struggle with clinically important language (negation, uncertainty) and with low-prevalence findings that matter for patient care. Their approach combines clinician-guided label definitions with LLMs for multi-label finding extraction from free-text radiology reports, aiming to produce an evaluation score that better reflects clinical correctness than surface-level text matches.
Why this could matter
Why does evaluation matter? Generative models for radiology reports are increasingly tested with metrics like BLEU, ROUGE or similarity scores that often miss clinical nuance. Ran Score aims to quantify whether a generated report correctly captures the presence, absence or uncertainty of specific findings — including rare abnormalities — rather than just mirroring phrasing. It has been reported that the authors find improved alignment with clinician judgments compared with existing automated scorers, though this is drawn from a preprint and not yet peer reviewed.
Caveats and next steps
The work is currently a preprint on arXiv, meaning methods and claims have not passed formal peer review. Reportedly, Ran Score’s performance will depend on the quality of clinician-provided labels and the behaviour of underlying LLMs, which can vary across models and deployment contexts. How will regulators and hospitals view LLM-assisted evaluation in safety-critical imaging workflows? That remains an open question, especially as medical AI faces increasing scrutiny in the US, Europe and China. Readers can review the full technical description at https://arxiv.org/abs/2603.22935.
