PhilSci Archive

Nothing New Under the Sun – Large Language Models and Scientific Method

Pietsch, Wolfgang (2026) Nothing New Under the Sun – Large Language Models and Scientific Method. [Preprint]

[img] Text
pietsch-llms_and_scientific_method.pdf

Download (847kB)

Abstract

The fundamental principle of large language models (LLMs) is next-word prediction based on a modeling architecture that is—in comparison to the achievements of LLMs—strikingly simple consisting of just a vast number of vector and matrix operations. For predictions, large language models rely on induction inferring generalized relationships from an enormous number of particulars without any hypothetical assumptions concerning the modeled subject matter itself, i.e. language. Regarding the specific type of induction, I argue that LLMs employ a difference-making logic (or variational induction). I show that central aspects of such variational induction are realized by LLMs. In particular, the training of large language models requires variation in evidence drawing on as much text data from as wide a range of contexts as possible. If sufficient data is available, difference and indifference makers among the input token sequence are identified during training. The resulting large language models essentially consist of aggregations of vast numbers of probabilistic laws, where each law relates Boolean combinations of ordered input tokens with output probability vectors. Linguistic meaning can be extracted, because variational induction allows for distinguishing between spurious and necessary relationships. The hierarchical layer structure of LLMs can be interpreted as a continuous probabilistic generalization of the deterministic binary Boolean logic of conventional variational induction. Specific features of LLMs that go beyond a simple neural network architecture such as token embeddings or self-attention are discussed to determine their role in the context of variational induction. Token embeddings transform the input sequence into a compressed representation capturing semantic relationships so that difference and indifference makers can be found more efficiently. Self-attention allows for analysing long-range difference making dependencies in token sequences, which are prevalent in natural language. In the history of science, the difference making logic employed by large language models has been widely used. Maybe most importantly, it underlies the experimental method, where causal relations are derived by systematically varying individual circumstances to determine their influence on a phenomenon.


Export/Citation: EndNote | BibTeX | Dublin Core | ASCII/Text Citation (Chicago) | HTML Citation | OpenURL
Social Networking:
Share |

Item Type: Preprint
Creators:
CreatorsEmailORCID
Pietsch, Wolfgangpietsch@cvl-a.tum.de
Keywords: machine learning; neural network; large language model; LLM; transformer; embedding; induction; causation; difference making; variational induction
Subjects: General Issues > Data
General Issues > Causation
Specific Sciences > Complex Systems
Specific Sciences > Computer Science
Specific Sciences > Artificial Intelligence
General Issues > Computer Simulation
General Issues > Confirmation/Induction
Specific Sciences > Engineering
General Issues > Experimentation
Specific Sciences > Artificial Intelligence > Machine Learning
General Issues > Models and Idealization
Specific Sciences > Probability/Statistics
Specific Sciences > Physics > Statistical Mechanics/Thermodynamics
General Issues > Structure of Theories
General Issues > Technology
Depositing User: Wolfgang Pietsch
Date Deposited: 17 Feb 2026 12:53
Last Modified: 17 Feb 2026 12:53
Item ID: 28259
Subjects: General Issues > Data
General Issues > Causation
Specific Sciences > Complex Systems
Specific Sciences > Computer Science
Specific Sciences > Artificial Intelligence
General Issues > Computer Simulation
General Issues > Confirmation/Induction
Specific Sciences > Engineering
General Issues > Experimentation
Specific Sciences > Artificial Intelligence > Machine Learning
General Issues > Models and Idealization
Specific Sciences > Probability/Statistics
Specific Sciences > Physics > Statistical Mechanics/Thermodynamics
General Issues > Structure of Theories
General Issues > Technology
Date: 16 February 2026
URI: https://philsci-archive.pitt.edu/id/eprint/28259

Monthly Views for the past 3 years

Monthly Downloads for the past 3 years

Plum Analytics

Actions (login required)

View Item View Item