Meaning, Form and the Limits of Natural Language Processing
This article engages the anthropological assumptions underlying the apprehensions and promises associated with language in artificial intelligence (AI). First, we present the contours of two rivalling paradigms for assessing artificial language generation: a holistic-enactivist theory of language an...
Authors: | ; ; |
---|---|
Format: | Electronic Article |
Language: | German |
Check availability: | HBZ Gateway |
Journals Online & Print: | |
Fernleihe: | Fernleihe für die Fachinformationsdienste |
Published: |
Mohr Siebeck
2023
|
In: |
Philosophy, theology and the sciences
Year: 2023, Volume: 10, Issue: 1, Pages: 42-72 |
RelBib Classification: | NBE Anthropology NCJ Ethics of science VA Philosophy |
Further subjects: | B
Common-sense
B Enactivism B Language B Artificial general intelligence (AGI) B Ai B Understanding B Embodied cognition B Information B philosophy of language B Large Language Models (LLMs) B Meaning B Autonomous machine intelligence |
Online Access: |
Volltext (kostenfrei) |
Summary: | This article engages the anthropological assumptions underlying the apprehensions and promises associated with language in artificial intelligence (AI). First, we present the contours of two rivalling paradigms for assessing artificial language generation: a holistic-enactivist theory of language and an informational theory of language. We then introduce two language generation models - one presently in use and one more speculative: Firstly, the transformer architecture as used in current large language models, such as the GPT-series, and secondly, a model for "autonomous machine intelligence" recently proposed by Yann LeCun, which involves not only language but a sensory-motor interaction with the world. We then assess the language capacity of these models from the perspectives of the two rivalling language paradigms. Taking a holistic-enactivist stance, we then argue that there is currently no reason to assume a human-comparable language capacity in LLMs and, further, that LeCun's proposed model does not represent a significant step toward artificially generating human language because it still lacks essential features that underlie the linguistic capacity of humans. Finally, we suggest that proponents of these rivalling interpretations of LLMs should enter into a constructive dialogue and that this dialogue should continuously involve further empirical, conceptual, and theoretical research. |
---|---|
ISSN: | 2197-2834 |
Contains: | Enthalten in: Philosophy, theology and the sciences
|
Persistent identifiers: | DOI: 10.1628/ptsc-2023-0005 |