AI language model rivals expert ethicist in perceived moral expertise
Abstract People view AI as possessing expertise across various fields, but the perceived quality of AI-generated moral expertise remains uncertain. Recent work suggests that large language models (LLMs) perform well on tasks designed to assess moral alignment, reflecting moral judgments with relativ...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Nature Portfolio
2025-02-01
|
Series: | Scientific Reports |
Online Access: | https://doi.org/10.1038/s41598-025-86510-0 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1823862474676371456 |
---|---|
author | Danica Dillion Debanjan Mondal Niket Tandon Kurt Gray |
author_facet | Danica Dillion Debanjan Mondal Niket Tandon Kurt Gray |
author_sort | Danica Dillion |
collection | DOAJ |
description | Abstract People view AI as possessing expertise across various fields, but the perceived quality of AI-generated moral expertise remains uncertain. Recent work suggests that large language models (LLMs) perform well on tasks designed to assess moral alignment, reflecting moral judgments with relatively high accuracy. As LLMs are increasingly employed in decision-making roles, there is a growing expectation for them to offer not just aligned judgments but also demonstrate sound moral reasoning. Here, we advance work on the Moral Turing Test and find that Americans rate ethical advice from GPT-4o as slightly more moral, trustworthy, thoughtful, and correct than that of the popular New York Times advice column, The Ethicist. Participants perceived GPT models as surpassing both a representative sample of Americans and a renowned ethicist in delivering moral justifications and advice, suggesting that people may increasingly view LLM outputs as viable sources of moral expertise. This work suggests that people might see LLMs as valuable complements to human expertise in moral guidance and decision-making. It also underscores the importance of carefully programming ethical guidelines in LLMs, considering their potential to influence users’ moral reasoning. |
format | Article |
id | doaj-art-4d246ee29a324489b8d47d35d5f02442 |
institution | Kabale University |
issn | 2045-2322 |
language | English |
publishDate | 2025-02-01 |
publisher | Nature Portfolio |
record_format | Article |
series | Scientific Reports |
spelling | doaj-art-4d246ee29a324489b8d47d35d5f024422025-02-09T12:30:52ZengNature PortfolioScientific Reports2045-23222025-02-0115111510.1038/s41598-025-86510-0AI language model rivals expert ethicist in perceived moral expertiseDanica Dillion0Debanjan Mondal1Niket Tandon2Kurt Gray3Department of Psychology and Neuroscience, University of North Carolina at Chapel HillAllen Institute for Artificial IntelligenceAllen Institute for Artificial IntelligenceDepartment of Psychology and Neuroscience, University of North Carolina at Chapel HillAbstract People view AI as possessing expertise across various fields, but the perceived quality of AI-generated moral expertise remains uncertain. Recent work suggests that large language models (LLMs) perform well on tasks designed to assess moral alignment, reflecting moral judgments with relatively high accuracy. As LLMs are increasingly employed in decision-making roles, there is a growing expectation for them to offer not just aligned judgments but also demonstrate sound moral reasoning. Here, we advance work on the Moral Turing Test and find that Americans rate ethical advice from GPT-4o as slightly more moral, trustworthy, thoughtful, and correct than that of the popular New York Times advice column, The Ethicist. Participants perceived GPT models as surpassing both a representative sample of Americans and a renowned ethicist in delivering moral justifications and advice, suggesting that people may increasingly view LLM outputs as viable sources of moral expertise. This work suggests that people might see LLMs as valuable complements to human expertise in moral guidance and decision-making. It also underscores the importance of carefully programming ethical guidelines in LLMs, considering their potential to influence users’ moral reasoning.https://doi.org/10.1038/s41598-025-86510-0 |
spellingShingle | Danica Dillion Debanjan Mondal Niket Tandon Kurt Gray AI language model rivals expert ethicist in perceived moral expertise Scientific Reports |
title | AI language model rivals expert ethicist in perceived moral expertise |
title_full | AI language model rivals expert ethicist in perceived moral expertise |
title_fullStr | AI language model rivals expert ethicist in perceived moral expertise |
title_full_unstemmed | AI language model rivals expert ethicist in perceived moral expertise |
title_short | AI language model rivals expert ethicist in perceived moral expertise |
title_sort | ai language model rivals expert ethicist in perceived moral expertise |
url | https://doi.org/10.1038/s41598-025-86510-0 |
work_keys_str_mv | AT danicadillion ailanguagemodelrivalsexpertethicistinperceivedmoralexpertise AT debanjanmondal ailanguagemodelrivalsexpertethicistinperceivedmoralexpertise AT nikettandon ailanguagemodelrivalsexpertethicistinperceivedmoralexpertise AT kurtgray ailanguagemodelrivalsexpertethicistinperceivedmoralexpertise |