Esanu, Andreea (2024) Scrutinizing the foundations: could large Language Models be solipsistic? [Preprint]
|
Text
Manuscript_LLMs_v4_preprint.pdf Download (1MB) | Preview |
Abstract
In artificial intelligence literature, “delusions” are characterized as the generation of unfaithful output from reliable source content. There is an extensive literature on computer-generated delusions, ranging from visual hallucinations, like the production of nonsensical images in Computer Vision, to nonsensical text generated by (natural) language models, but this literature is predominantly taxonomic. In a recent research paper, however, a group of scientists from DeepMind successfully presented a formal treatment of an entire class of delusions in generative AI models (i.e., models based on a transformer architecture, both with and without RLHF – reinforcement learning with human feedback, such as BERT, GPT-3 or the more recent GPT-3.5), referred to as auto-suggestive delusions. Auto-suggestive delusions are not mere unfaithful output, but are self-induced by the transformer models themselves. Typically, these delusions have been subsumed under the concept of exposure bias, but exposure bias alone does not elucidate their nature. In order to address their nature, I will introduce a formal framework that clarifies the probabilistic delusions capable of explaining exposure bias in a broad manner. This will serve as the foundation for exploring auto-suggestive delusions in language models. Next, an examination of self- or auto-suggestive delusions will be undertaken, by drawing an analogy with the rule-following problematic from the philosophy of mind and language. Finally, I will argue that this comprehensive approach leads to the suggestion that transformers, large language models in particular, may develop in a manner that touches upon solipsism and the emergence of a private language, in a weak sense.
Export/Citation: | EndNote | BibTeX | Dublin Core | ASCII/Text Citation (Chicago) | HTML Citation | OpenURL |
Social Networking: |
Item Type: | Preprint | ||||||
---|---|---|---|---|---|---|---|
Creators: |
|
||||||
Additional Information: | Accepted for publication in Synthese | ||||||
Subjects: | General Issues > Data General Issues > Causation Specific Sciences > Probability/Statistics General Issues > Technology |
||||||
Depositing User: | Dr Andreea Esanu | ||||||
Date Deposited: | 01 Apr 2024 05:32 | ||||||
Last Modified: | 01 Apr 2024 05:32 | ||||||
Item ID: | 23249 | ||||||
Subjects: | General Issues > Data General Issues > Causation Specific Sciences > Probability/Statistics General Issues > Technology |
||||||
Date: | 29 March 2024 | ||||||
URI: | https://philsci-archive.pitt.edu/id/eprint/23249 |
Monthly Views for the past 3 years
Monthly Downloads for the past 3 years
Plum Analytics
Actions (login required)
View Item |