Language writ large: LLMs, ChatGPT, meaning, and understanding

Apart from what (little) OpenAI may be concealing from us, we all know (roughly) how Large Language Models (LLMs) such as ChatGPT work (their vast text databases, statistics, vector representations, and huge number of parameters, next-word training, etc.). However, none of us can say (hand on heart)...

Full description

Saved in:
Bibliographic Details
Main Author: Stevan Harnad
Format: Article
Language:English
Published: Frontiers Media S.A. 2025-02-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2024.1490698/full
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1823856626816253952
author Stevan Harnad
author_facet Stevan Harnad
author_sort Stevan Harnad
collection DOAJ
description Apart from what (little) OpenAI may be concealing from us, we all know (roughly) how Large Language Models (LLMs) such as ChatGPT work (their vast text databases, statistics, vector representations, and huge number of parameters, next-word training, etc.). However, none of us can say (hand on heart) that we are not surprised by what ChatGPT has proved to be able to do with these resources. This has even driven some of us to conclude that ChatGPT actually understands. It is not true that it understands. But it is also not true that we understand how it can do what it can do. I will suggest some hunches about benign “biases”—convergent constraints that emerge at the LLM scale that may be helping ChatGPT do so much better than we would have expected. These biases are inherent in the nature of language itself, at the LLM scale, and they are closely linked to what it is that ChatGPT lacks, which is direct sensorimotor grounding to connect its words to their referents and its propositions to their meanings. These convergent biases are related to (1) the parasitism of indirect verbal grounding on direct sensorimotor grounding, (2) the circularity of verbal definition, (3) the “mirroring” of language production and comprehension, (4) iconicity in propositions at LLM scale, (5) computational counterparts of human “categorical perception” in category learning by neural nets, and perhaps also (6) a conjecture by Chomsky about the laws of thought. The exposition will be in the form of a dialogue with ChatGPT-4.
format Article
id doaj-art-e2d839d54f0f4ddd97f2fb6e97cdcc93
institution Kabale University
issn 2624-8212
language English
publishDate 2025-02-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Artificial Intelligence
spelling doaj-art-e2d839d54f0f4ddd97f2fb6e97cdcc932025-02-12T07:25:58ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122025-02-01710.3389/frai.2024.14906981490698Language writ large: LLMs, ChatGPT, meaning, and understandingStevan HarnadApart from what (little) OpenAI may be concealing from us, we all know (roughly) how Large Language Models (LLMs) such as ChatGPT work (their vast text databases, statistics, vector representations, and huge number of parameters, next-word training, etc.). However, none of us can say (hand on heart) that we are not surprised by what ChatGPT has proved to be able to do with these resources. This has even driven some of us to conclude that ChatGPT actually understands. It is not true that it understands. But it is also not true that we understand how it can do what it can do. I will suggest some hunches about benign “biases”—convergent constraints that emerge at the LLM scale that may be helping ChatGPT do so much better than we would have expected. These biases are inherent in the nature of language itself, at the LLM scale, and they are closely linked to what it is that ChatGPT lacks, which is direct sensorimotor grounding to connect its words to their referents and its propositions to their meanings. These convergent biases are related to (1) the parasitism of indirect verbal grounding on direct sensorimotor grounding, (2) the circularity of verbal definition, (3) the “mirroring” of language production and comprehension, (4) iconicity in propositions at LLM scale, (5) computational counterparts of human “categorical perception” in category learning by neural nets, and perhaps also (6) a conjecture by Chomsky about the laws of thought. The exposition will be in the form of a dialogue with ChatGPT-4.https://www.frontiersin.org/articles/10.3389/frai.2024.1490698/fullsymbol groundingcategorical perceptioncategory learningfeature abstractionmeaning and understandingChatGPT and LLMs
spellingShingle Stevan Harnad
Language writ large: LLMs, ChatGPT, meaning, and understanding
Frontiers in Artificial Intelligence
symbol grounding
categorical perception
category learning
feature abstraction
meaning and understanding
ChatGPT and LLMs
title Language writ large: LLMs, ChatGPT, meaning, and understanding
title_full Language writ large: LLMs, ChatGPT, meaning, and understanding
title_fullStr Language writ large: LLMs, ChatGPT, meaning, and understanding
title_full_unstemmed Language writ large: LLMs, ChatGPT, meaning, and understanding
title_short Language writ large: LLMs, ChatGPT, meaning, and understanding
title_sort language writ large llms chatgpt meaning and understanding
topic symbol grounding
categorical perception
category learning
feature abstraction
meaning and understanding
ChatGPT and LLMs
url https://www.frontiersin.org/articles/10.3389/frai.2024.1490698/full
work_keys_str_mv AT stevanharnad languagewritlargellmschatgptmeaningandunderstanding