PhilSci Archive

Simulated Selfhood in LLMs: A Behavioral Analysis of Introspective Coherence

de Lima Prestes, José Augusto (2025) Simulated Selfhood in LLMs: A Behavioral Analysis of Introspective Coherence. [Preprint]

This is the latest version of this item.

[img] Text
Simulated_Selfhood_in_LLMs_Preprint_v2.pdf

Download (356kB)

Abstract

Large Language Models (LLMs) increasingly generate outputs that resemble introspection, including self-reference, epistemic modulation, and claims about their internal states. This study investigates whether such behaviors reflect consistent, underlying patterns or are merely surface-level generative artifacts.We evaluated five open-weight, stateless LLMs using a structured battery of 21 introspective prompts, each repeated ten times to yield 1,050 completions. These outputs were analyzed across four behavioral dimensions: surface-level similarity (token overlap via SequenceMatcher), semantic coherence (Sentence-BERT embeddings), inferential consistency (Natural Language Inference with a RoBERTa-large model), and diachronic continuity (stability across prompt repetitions). Although some models exhibited thematic stability, particularly on prompts concerning identity and consciousness, no model sustained a consistent self-representation over time. High contradiction rates emerged from a tension between mechanistic disclaimers and anthropomorphic phrasing. Following recent behavioral frameworks, we heuristically adopt the term pseudo-consciousness to describe structured yet non-experiential self-referential output in LLMs. This usage reflects a functionalist stance that avoids ontological commitments, focusing instead on behavioral regularities interpretable through Dennett’s intentional stance. The study contributes a reproducible framework for evaluating simulated introspection in LLMs and offers a graded taxonomy for classifying such reflexive output. Our findings carry significant implications for LLM interpretability, alignment, and user perception, highlighting the need for caution when attributing mental states to stateless generative systems based on linguistic fluency alone.


Export/Citation: EndNote | BibTeX | Dublin Core | ASCII/Text Citation (Chicago) | HTML Citation | OpenURL
Social Networking:
Share |

Item Type: Preprint
Creators:
CreatorsEmailORCID
de Lima Prestes, José Augustocontato@joseprestes.com0000-0001-8686-5360
Keywords: Large Language Models Introspective Simulation Pseudo-consciousness Self-reference Behavioral Evaluation Epistemic Modulation AI Alignment Philosophy of Cognitive Science Philosophy of Artificial Intelligence Explainability
Subjects: Specific Sciences > Computation/Information > Classical
Specific Sciences > Artificial Intelligence > AI and Ethics
Specific Sciences > Artificial Intelligence > Classical AI
Specific Sciences > Cognitive Science > Computation
Specific Sciences > Cognitive Science > Concepts and Representations
Specific Sciences > Cognitive Science > Consciousness
General Issues > Explanation
General Issues > Models and Idealization
Depositing User: José Augusto de Lima Prestes
Date Deposited: 21 Sep 2025 11:13
Last Modified: 21 Sep 2025 11:13
Item ID: 26706
Subjects: Specific Sciences > Computation/Information > Classical
Specific Sciences > Artificial Intelligence > AI and Ethics
Specific Sciences > Artificial Intelligence > Classical AI
Specific Sciences > Cognitive Science > Computation
Specific Sciences > Cognitive Science > Concepts and Representations
Specific Sciences > Cognitive Science > Consciousness
General Issues > Explanation
General Issues > Models and Idealization
Date: 26 July 2025
URI: https://philsci-archive.pitt.edu/id/eprint/26706

Available Versions of this Item

Monthly Views for the past 3 years

Monthly Downloads for the past 3 years

Plum Analytics

Actions (login required)

View Item View Item