Comparative analysis of generative LLMs for labeling entities in clinical notes

Abstract This paper evaluates and compares different fine-tuned variations of generative large language models (LLM) in the zero-shot named entity recognition (NER) task for the clinical domain. As part of the 8th Biomedical Linked Annotation Hackathon, we examined Llama 2 and Mistral models, includ...

Full description

Saved in:
Bibliographic Details
Main Authors: Rodrigo del Moral-González, Helena Gómez-Adorno, Orlando Ramos-Flores
Format: Article
Language:English
Published: BioMed Central 2025-02-01
Series:Genomics & Informatics
Subjects:
Online Access:https://doi.org/10.1186/s44342-024-00036-x
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract This paper evaluates and compares different fine-tuned variations of generative large language models (LLM) in the zero-shot named entity recognition (NER) task for the clinical domain. As part of the 8th Biomedical Linked Annotation Hackathon, we examined Llama 2 and Mistral models, including base versions and those that have been fine-tuned for code, chat, and instruction-following tasks. We assess both the number of correctly identified entities and the models’ ability to retrieve entities in structured formats. We used a publicly available set of clinical cases labeled with mentions of diseases, symptoms, and medical procedures for the evaluation. Results show that instruction fine-tuned models perform better than chat fine-tuned and base models in recognizing entities. It is also shown that models perform better when simple output structures are requested.
ISSN:2234-0742