Comparative analysis of generative LLMs for labeling entities in clinical notes
Abstract This paper evaluates and compares different fine-tuned variations of generative large language models (LLM) in the zero-shot named entity recognition (NER) task for the clinical domain. As part of the 8th Biomedical Linked Annotation Hackathon, we examined Llama 2 and Mistral models, includ...
Saved in:
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
BioMed Central
2025-02-01
|
Series: | Genomics & Informatics |
Subjects: | |
Online Access: | https://doi.org/10.1186/s44342-024-00036-x |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Abstract This paper evaluates and compares different fine-tuned variations of generative large language models (LLM) in the zero-shot named entity recognition (NER) task for the clinical domain. As part of the 8th Biomedical Linked Annotation Hackathon, we examined Llama 2 and Mistral models, including base versions and those that have been fine-tuned for code, chat, and instruction-following tasks. We assess both the number of correctly identified entities and the models’ ability to retrieve entities in structured formats. We used a publicly available set of clinical cases labeled with mentions of diseases, symptoms, and medical procedures for the evaluation. Results show that instruction fine-tuned models perform better than chat fine-tuned and base models in recognizing entities. It is also shown that models perform better when simple output structures are requested. |
---|---|
ISSN: | 2234-0742 |