Why Large Language Models Are Just Mathematical Functions: A Rational Perspective
The article argues that large language models are fundamentally mathematical functions that model human language, emphasizing their role as simplified representations, explaining their structural nature, sources of errors, the importance of prompts as boundary conditions, and the need for clear usage assumptions to avoid anthropomorphic misconceptions.
Models Are Simplified Representations of Complex Worlds
Mathematical models aim to simplify reality without sacrificing decision value. From differential equations in physics to optimization models in economics and network models in social science, each model serves a purposeful, structured, and reduced representation of a complex system.
They have specific modeling goals (explanation, prediction, optimization);
They select a subset of variables, inevitably ignoring others;
They establish relationships between variables through logical rules or statistical patterns.
Large language models (LLMs) are, in this sense, tools that parameterize human language behavior. They do not think; they fit the probability and structural logic of language occurrences.
The Structural Essence of Large Language Models
A language model can be abstracted as a mathematical function: the input is a natural‑language prompt, and the output is the model’s most likely continuation. Its core is a parameter space of billions of weights forming a neural network that maximizes likelihood of generated text.
This is akin to a highly complex regression or classification model that predicts the next word, sentence, or paragraph, not because it understands, but because it models co‑occurrence probabilities in language data.
When we say an LLM “writes stories” or “makes logical errors,” we are describing typical failures of function approximation: inputs outside the training distribution, loss functions lacking logical constraints, and no real‑world verification.
Errors Are Model Biases, Not Cognitive Failures
Understanding that an LLM is merely a model helps us locate the source of its errors without attributing intent or morality.
Common error types include:
Factual errors : the training data diverge from reality or the context pattern is not activated;
Logical inconsistencies : the objective function does not enforce logical coherence;
Semantic vagueness : language itself is ambiguous, so the model selects the most probable interpretation.
These are analogous to residuals, extrapolation failures, or over‑fitting in traditional mathematical modeling.
Prompts Are Boundary Conditions
In mathematical modeling, the crucial aspect is not only the model structure but also how we impose boundary conditions and control inputs. Prompts serve exactly this role for LLMs:
They control the initial state;
They activate a specific context space;
They fine‑tune the direction of the model’s output.
Thus using an LLM becomes an “interactive modeling” process:
Each prompt sets boundary conditions;
Provided examples constrain the solution space;
Feedback and corrections optimize the objective function.
We must understand the model’s structure, adjust inputs, and evaluate output boundaries.
Must Define Usage Assumptions
Before employing any model, we must state its usage assumptions, just as traditional modeling declares linearity, variable independence, or controllable observation error.
When applying LLMs for writing, translation, Q&A, or decision support, we should clarify:
Whether outputs need human review;
Whether the model is limited to non‑critical contexts;
If factual accuracy or logical consistency is required;
Potential value bias or ethical risks that demand additional constraints.
Only with explicit assumptions can an LLM be treated as a controllable system, avoiding both “technology worship” and “technology nihilism.”
De‑Anthropomorphizing Is Essential
Popular narratives anthropomorphize LLMs, claiming they “understand,” “know,” or “think.” In reality, the model merely captures statistical regularities, stores compressed text correlations, and lacks genuine understanding, knowledge, or consciousness.
Anthropomorphizing raises unrealistic expectations, cognitive mismatches, and ethical confusion. Viewing the model as a mathematical function clarifies three key questions:
What does it do well?
Where does it fail?
How can we improve or control it?
Removing human‑centric projections reveals a powerful generator that can assist information processing and language generation, provided it operates within defined boundaries.
The rational first step is linguistic clarity: an LLM is a parameterized function, not an entity with intent or moral agency. We should aim to understand, control, and fine‑tune it rather than worship or fear it.
Large language models represent a breakthrough technology, but our attitude must be mature, restrained, and rational. Treating them as mathematical models does not diminish their capabilities; it enables better usage, analysis, and awareness of their strengths and limitations.
Model Perspective
Insights, knowledge, and enjoyment from a mathematical modeling researcher and educator. Hosted by Haihua Wang, a modeling instructor and author of "Clever Use of Chat for Mathematical Modeling", "Modeling: The Mathematics of Thinking", "Mathematical Modeling Practice: A Hands‑On Guide to Competitions", and co‑author of "Mathematical Modeling: Teaching Design and Cases".
How this landed with the community
Was this worth your time?
0 Comments
Thoughtful readers leave field notes, pushback, and hard-won operational detail here.