The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis

BackgroundThe increasing deployment of conversational artificial intelligence (AI) in mental health interventions necessitates an evaluation of their efficacy in rectifying cognitive biases and recognizing affect in human-AI interactions. These biases are particularly relevan...

Full description

Saved in:
Bibliographic Details
Main Authors: Marcin Rządeczka, Anna Sterna, Julia Stolińska, Paulina Kaczyńska, Marcin Moskalewicz
Format: Article
Language:English
Published: JMIR Publications 2025-02-01
Series:JMIR Mental Health
Online Access:https://mental.jmir.org/2025/1/e64396
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1825200099124314112
author Marcin Rządeczka
Anna Sterna
Julia Stolińska
Paulina Kaczyńska
Marcin Moskalewicz
author_facet Marcin Rządeczka
Anna Sterna
Julia Stolińska
Paulina Kaczyńska
Marcin Moskalewicz
author_sort Marcin Rządeczka
collection DOAJ
description BackgroundThe increasing deployment of conversational artificial intelligence (AI) in mental health interventions necessitates an evaluation of their efficacy in rectifying cognitive biases and recognizing affect in human-AI interactions. These biases are particularly relevant in mental health contexts as they can exacerbate conditions such as depression and anxiety by reinforcing maladaptive thought patterns or unrealistic expectations in human-AI interactions. ObjectiveThis study aimed to assess the effectiveness of therapeutic chatbots (Wysa and Youper) versus general-purpose language models (GPT-3.5, GPT-4, and Gemini Pro) in identifying and rectifying cognitive biases and recognizing affect in user interactions. MethodsThis study used constructed case scenarios simulating typical user-bot interactions to examine how effectively chatbots address selected cognitive biases. The cognitive biases assessed included theory-of-mind biases (anthropomorphism, overtrust, and attribution) and autonomy biases (illusion of control, fundamental attribution error, and just-world hypothesis). Each chatbot response was evaluated based on accuracy, therapeutic quality, and adherence to cognitive behavioral therapy principles using an ordinal scale to ensure consistency in scoring. To enhance reliability, responses underwent a double review process by 2 cognitive scientists, followed by a secondary review by a clinical psychologist specializing in cognitive behavioral therapy, ensuring a robust assessment across interdisciplinary perspectives. ResultsThis study revealed that general-purpose chatbots outperformed therapeutic chatbots in rectifying cognitive biases, particularly in overtrust bias, fundamental attribution error, and just-world hypothesis. GPT-4 achieved the highest scores across all biases, whereas the therapeutic bot Wysa scored the lowest. Notably, general-purpose bots showed more consistent accuracy and adaptability in recognizing and addressing bias-related cues across different contexts, suggesting a broader flexibility in handling complex cognitive patterns. In addition, in affect recognition tasks, general-purpose chatbots not only excelled but also demonstrated quicker adaptation to subtle emotional nuances, outperforming therapeutic bots in 67% (4/6) of the tested biases. ConclusionsThis study shows that, while therapeutic chatbots hold promise for mental health support and cognitive bias intervention, their current capabilities are limited. Addressing cognitive biases in AI-human interactions requires systems that can both rectify and analyze biases as integral to human cognition, promoting precision and simulating empathy. The findings reveal the need for improved simulated emotional intelligence in chatbot design to provide adaptive, personalized responses that reduce overreliance and encourage independent coping skills. Future research should focus on enhancing affective response mechanisms and addressing ethical concerns such as bias mitigation and data privacy to ensure safe, effective AI-based mental health support.
format Article
id doaj-art-698ddb43c47b43e08d7e6615b4fdcc10
institution Kabale University
issn 2368-7959
language English
publishDate 2025-02-01
publisher JMIR Publications
record_format Article
series JMIR Mental Health
spelling doaj-art-698ddb43c47b43e08d7e6615b4fdcc102025-02-07T21:00:36ZengJMIR PublicationsJMIR Mental Health2368-79592025-02-0112e6439610.2196/64396The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative AnalysisMarcin Rządeczkahttps://orcid.org/0000-0002-8315-1650Anna Sternahttps://orcid.org/0000-0001-8658-7823Julia Stolińskahttps://orcid.org/0009-0003-8206-8876Paulina Kaczyńskahttps://orcid.org/0000-0002-8690-8436Marcin Moskalewiczhttps://orcid.org/0000-0002-4270-7026 BackgroundThe increasing deployment of conversational artificial intelligence (AI) in mental health interventions necessitates an evaluation of their efficacy in rectifying cognitive biases and recognizing affect in human-AI interactions. These biases are particularly relevant in mental health contexts as they can exacerbate conditions such as depression and anxiety by reinforcing maladaptive thought patterns or unrealistic expectations in human-AI interactions. ObjectiveThis study aimed to assess the effectiveness of therapeutic chatbots (Wysa and Youper) versus general-purpose language models (GPT-3.5, GPT-4, and Gemini Pro) in identifying and rectifying cognitive biases and recognizing affect in user interactions. MethodsThis study used constructed case scenarios simulating typical user-bot interactions to examine how effectively chatbots address selected cognitive biases. The cognitive biases assessed included theory-of-mind biases (anthropomorphism, overtrust, and attribution) and autonomy biases (illusion of control, fundamental attribution error, and just-world hypothesis). Each chatbot response was evaluated based on accuracy, therapeutic quality, and adherence to cognitive behavioral therapy principles using an ordinal scale to ensure consistency in scoring. To enhance reliability, responses underwent a double review process by 2 cognitive scientists, followed by a secondary review by a clinical psychologist specializing in cognitive behavioral therapy, ensuring a robust assessment across interdisciplinary perspectives. ResultsThis study revealed that general-purpose chatbots outperformed therapeutic chatbots in rectifying cognitive biases, particularly in overtrust bias, fundamental attribution error, and just-world hypothesis. GPT-4 achieved the highest scores across all biases, whereas the therapeutic bot Wysa scored the lowest. Notably, general-purpose bots showed more consistent accuracy and adaptability in recognizing and addressing bias-related cues across different contexts, suggesting a broader flexibility in handling complex cognitive patterns. In addition, in affect recognition tasks, general-purpose chatbots not only excelled but also demonstrated quicker adaptation to subtle emotional nuances, outperforming therapeutic bots in 67% (4/6) of the tested biases. ConclusionsThis study shows that, while therapeutic chatbots hold promise for mental health support and cognitive bias intervention, their current capabilities are limited. Addressing cognitive biases in AI-human interactions requires systems that can both rectify and analyze biases as integral to human cognition, promoting precision and simulating empathy. The findings reveal the need for improved simulated emotional intelligence in chatbot design to provide adaptive, personalized responses that reduce overreliance and encourage independent coping skills. Future research should focus on enhancing affective response mechanisms and addressing ethical concerns such as bias mitigation and data privacy to ensure safe, effective AI-based mental health support.https://mental.jmir.org/2025/1/e64396
spellingShingle Marcin Rządeczka
Anna Sterna
Julia Stolińska
Paulina Kaczyńska
Marcin Moskalewicz
The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
JMIR Mental Health
title The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
title_full The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
title_fullStr The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
title_full_unstemmed The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
title_short The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
title_sort efficacy of conversational ai in rectifying the theory of mind and autonomy biases comparative analysis
url https://mental.jmir.org/2025/1/e64396
work_keys_str_mv AT marcinrzadeczka theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT annasterna theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT juliastolinska theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT paulinakaczynska theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT marcinmoskalewicz theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT marcinrzadeczka efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT annasterna efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT juliastolinska efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT paulinakaczynska efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis
AT marcinmoskalewicz efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis