Ontology-based prompt tuning for news article summarization

Ontology-based prompt tuning and abstractive text summarization techniques represent an advanced approach to enhancing the quality and contextual relevance of news article summaries. Despite the progress in natural language processing (NLP) and machine learning, existing methods often rely on extrac...

Full description

Saved in:
Bibliographic Details
Main Authors: A. R. S. Silva, Y. H. P. P. Priyadarshana
Format: Article
Language:English
Published: Frontiers Media S.A. 2025-02-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2025.1520144/full
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1823859167441453056
author A. R. S. Silva
Y. H. P. P. Priyadarshana
author_facet A. R. S. Silva
Y. H. P. P. Priyadarshana
author_sort A. R. S. Silva
collection DOAJ
description Ontology-based prompt tuning and abstractive text summarization techniques represent an advanced approach to enhancing the quality and contextual relevance of news article summaries. Despite the progress in natural language processing (NLP) and machine learning, existing methods often rely on extractive summarization, which lacks the ability to generate coherent and contextually rich summaries. Moreover, these approaches rarely integrate domain-specific knowledge, resulting in generic and sometimes inaccurate summaries. In this study, we propose a novel framework, which combines ontology-based prompt tuning with abstractive text summarization to address these limitations. By leveraging ontological knowledge, our model fine-tunes the summarization process, ensuring that the generated summaries are not only accurate but also contextually relevant to the domain. This integration allows for a more nuanced understanding of the text, enabling the generation of summaries that better capture the essence of the news articles. Our evaluation results demonstrate significant improvements over state-of-the-art methods such as BART, BERT, and GPT-3.5. The results show that the proposed architecture achieved a 5.1% higher ROUGE-1 score and a 9.8% improvement in ROUGE-L compared to baseline models. Additionally, our model showed significance in F1, precision, and recall metrics, with major improvements of 6.7, 3.9, and 4.8%, respectively. These results underscore the effectiveness of integrating ontological insights into the prompt tuning process, offering a robust solution for generating high-quality, domain-specific news summaries.
format Article
id doaj-art-3ee5acd20ab5496b89ed0216ee2174aa
institution Kabale University
issn 2624-8212
language English
publishDate 2025-02-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Artificial Intelligence
spelling doaj-art-3ee5acd20ab5496b89ed0216ee2174aa2025-02-11T07:00:21ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122025-02-01810.3389/frai.2025.15201441520144Ontology-based prompt tuning for news article summarizationA. R. S. SilvaY. H. P. P. PriyadarshanaOntology-based prompt tuning and abstractive text summarization techniques represent an advanced approach to enhancing the quality and contextual relevance of news article summaries. Despite the progress in natural language processing (NLP) and machine learning, existing methods often rely on extractive summarization, which lacks the ability to generate coherent and contextually rich summaries. Moreover, these approaches rarely integrate domain-specific knowledge, resulting in generic and sometimes inaccurate summaries. In this study, we propose a novel framework, which combines ontology-based prompt tuning with abstractive text summarization to address these limitations. By leveraging ontological knowledge, our model fine-tunes the summarization process, ensuring that the generated summaries are not only accurate but also contextually relevant to the domain. This integration allows for a more nuanced understanding of the text, enabling the generation of summaries that better capture the essence of the news articles. Our evaluation results demonstrate significant improvements over state-of-the-art methods such as BART, BERT, and GPT-3.5. The results show that the proposed architecture achieved a 5.1% higher ROUGE-1 score and a 9.8% improvement in ROUGE-L compared to baseline models. Additionally, our model showed significance in F1, precision, and recall metrics, with major improvements of 6.7, 3.9, and 4.8%, respectively. These results underscore the effectiveness of integrating ontological insights into the prompt tuning process, offering a robust solution for generating high-quality, domain-specific news summaries.https://www.frontiersin.org/articles/10.3389/frai.2025.1520144/fullknowledge representationnatural language processing (NLP)ontologyprompt tuningtext summarization
spellingShingle A. R. S. Silva
Y. H. P. P. Priyadarshana
Ontology-based prompt tuning for news article summarization
Frontiers in Artificial Intelligence
knowledge representation
natural language processing (NLP)
ontology
prompt tuning
text summarization
title Ontology-based prompt tuning for news article summarization
title_full Ontology-based prompt tuning for news article summarization
title_fullStr Ontology-based prompt tuning for news article summarization
title_full_unstemmed Ontology-based prompt tuning for news article summarization
title_short Ontology-based prompt tuning for news article summarization
title_sort ontology based prompt tuning for news article summarization
topic knowledge representation
natural language processing (NLP)
ontology
prompt tuning
text summarization
url https://www.frontiersin.org/articles/10.3389/frai.2025.1520144/full
work_keys_str_mv AT arssilva ontologybasedprompttuningfornewsarticlesummarization
AT yhpppriyadarshana ontologybasedprompttuningfornewsarticlesummarization