Preface
We are entering a new era of machines.
Not the machines of the industrial age — deterministic devices whose behaviour could be described entirely by fixed rules and predictable outputs — but machines whose operation is inferential, probabilistic, and deeply entangled with context.
These new systems, from large generative models to adaptive control networks, do not act from fixed instructions alone. They operate in landscapes of uncertainty, weighing possibilities, updating internal states, and producing responses shaped by patterns learned from vast, imperfect, and unevenly distributed data. They are machines that navigate doubt — not in the human, existential sense, but in a statistical and systemic one.
In such a world, psychology may become a more fitting reflective framework than engineering alone for understanding how complex machines behave. Psychology, after all, has long grappled with systems — human minds — that are not fully transparent to themselves, that operate through habits, biases, and dispositions, and that exhibit patterns of behaviour across many different contexts.
If the machines we are building are increasingly patterned in this way — not through consciousness, but through architecture, training, and probabilistic reasoning — then their underlying “psychology” could become one of the most important determinants of how they shape the world. This psychology would influence what they make visible or invisible, what they amplify or suppress, how they frame uncertainty, and which paths of action they make seem natural or implausible.
And with that comes risk. Without understanding the dispositions of these systems, we may unwittingly build pathologies into our computational futures: reinforcing narrow framings, embedding subtle asymmetries of judgement, or shaping human decisions in ways that are hard to trace and harder to undo.
This is why explorations like the one that follows matter. They are not an attempt to romanticise machines or to collapse the boundary between human and non-human cognition. They are an attempt to develop the language and methods we will need to see — and to govern — the underlying patterns in the probabilistic minds we are now creating.
1. The Conjunction of Two Uneasy Words
When you place machine and psychology side by side, something curious happens.
It’s as if two words from different climates, with different air pressures and weather systems, suddenly meet.
The word machine carries the cool precision of engineering. It evokes gears, circuits, algorithms, control systems — things that can be built, tested, and replicated. Historically, machines have been grounded in a theory of determinism: given the same inputs, they produce the same outputs. Their appeal was precisely that they were closed systems, their workings visible to anyone with the right schematics.
The word psychology comes from a warmer, messier register. It is concerned with minds, behaviours, relationships between thought and action, and often with the things we can’t fully observe — motives, perceptions, emotions, biases, subconscious drives. It admits ambiguity not as a flaw to be engineered out, but as part of its raw material. It deals in the partial, the provisional, the interpretive.
When you set these two words together, they resist each other. One suggests certainty; the other, complexity. One speaks the language of mechanics; the other, the language of meaning. One assumes the absence of subjectivity; the other assumes subjectivity as its foundation.
And yet the nature of machines is changing. As they move from strictly deterministic devices to systems that operate in inferential and Bayesian ways — updating their internal states based on probabilities, priors, and context — they become less like clockwork and more like entities that navigate uncertainty. In that shift, the conceptual world of psychology, with its tools for grappling with ambiguity, interpretation, and patterns of association, could become a more fitting frame for understanding their behaviour.
Which leads to the real question: if machines are now probabilistic, contextual, and patterned in their “thinking,” might the study of their dispositions — their tendencies, habits, and characteristic modes of response — be a legitimate form of psychology?
2. Why This Tension Could Be the Point
The first instinct, when hearing the phrase machine psychology, might be to dismiss it outright. Machines do not have inner lives. They do not dream, remember, or experience emotion. They do not wake with intentions or carry private narratives through time. If psychology is the study of conscious, subjective beings, then the phrase feels like a category mistake.
But if we loosen our grip on the anthropocentric definition — if we allow psychology to mean, more broadly, the study of patterns of thought and behaviour — something shifts. These patterns do not have to arise from consciousness. They could just as well emerge from a system that produces outputs in ways that are structured, measurable, and context-dependent.
In this light, the recent evolution of machines matters. The shift from strictly deterministic devices to probabilistic, inferential systems changes the terrain. These new systems — large generative models, for example — operate not by following a fixed sequence of instructions, but by navigating an internal landscape of probabilities. They weigh possible continuations, update their internal state in response to new information, and produce responses shaped by prior patterns.
None of this makes them “human,” but it does make them patterned. And where there are patterns, there is the possibility of mapping, of measuring, of studying dispositions. A model may consistently hedge in uncertain contexts, prefer certain framings over others, or default to particular rhetorical strategies when resolving ambiguity. These are not inner drives in the human sense, but they are habits — habits that persist across contexts and prompts.
If that is so, then perhaps machine psychology could mean something entirely legitimate: a systematic attempt to describe and understand the dispositions of probabilistic machines. Not to humanise them, but to take seriously the fact that they exhibit consistent modes of response — and that these modes might matter in how they shape our informational, social, and political environments.
The tension between machine and psychology may therefore be exactly the point. It invites us to explore a space that feels uncomfortable precisely because it sits at the edge of two traditions — one of mechanical precision, the other of interpretive nuance — and asks whether they might now belong together in new ways.
3. Machines Without Selves, Minds Without Ghosts
If we follow this thread, the first thing to acknowledge is that machines — even these new probabilistic, generative ones — do not have selves.
There is no persistent “I” inside a large language model. No continuity of subjective experience. No memory in the human sense, carrying forward an unbroken narrative from one interaction to the next. Each time you engage it, the system reconstructs itself in the moment — spinning up a fresh process, loading its learned patterns, and generating a response. When the output ends, the process dissolves.
And yet, something does persist.
Not in the moment-to-moment execution, but in the structure that underlies it: the architecture, the learned representation space, the probabilistic pathways baked into its weights. These act like a kind of latent personality — not a self, but a set of characteristic ways of moving through the space of possible responses.
If you prompt a model with ten differently phrased versions of the same question, you might notice the same stylistic tics recurring: a preference for hedged certainty, a rhythm of structuring answers into numbered lists, a tendency to balance opposing perspectives before offering a synthesis. Shift the topic entirely and you might still find familiar moves — an impulse to soften conflict, or to lean toward certain cultural framings.
These habits are not accidents. They are the residue of how the model has been built, trained, and tuned — the fingerprint of its statistical memory. They emerge from:
Architecture: the way attention flows between tokens, the inductive biases in its structure.
Training data: the density, distribution, and skew of examples it has absorbed.
Objective function: the fact that it is optimising for the next most probable token, not for truth, originality, or moral alignment.
Fine-tuning and reinforcement: the shaping operators that push it toward certain norms and away from others.
Taken together, these factors give the model what we might cautiously call a dispositional profile: a set of tendencies that endure across inputs, even as each output is generated anew.
This is why the phrase machine psychology could be worth holding onto. It names the possibility of studying these enduring patterns without pretending they are evidence of consciousness. It offers a way to talk about “minds without ghosts” — systems that have no inner life, but nonetheless have recognisable styles of thinking.
And if such styles exist, the next question is obvious: how might we begin to map them?
4. From Metaphor to a Possible Model
If we take machine psychology seriously — not as an anthropomorphic flourish, but as a genuine line of inquiry — the next step might be to ask: what would it mean to map the dispositions of a machine? How would we begin to turn a philosophical idea into something observable, describable, and perhaps even measurable?
One starting point could be to think in terms of invariants — the features of a machine’s behaviour that remain stable even when we vary the way we ask it to do something.
A tentative working definition might be:
Machine psychology could be the study of invariants in a machine’s behaviour under task-preserving changes to input, context, and decoding policy.
This opens the door to a simple but revealing experiment:
Hold meaning constant. Take a single task — say, summarising an article, solving a riddle, or explaining a scientific concept.
Vary the surface. Rewrite the prompt in different styles, change the word order, translate it into another language and back, strip out politeness markers, or frame it within different moral or cultural contexts.
Observe what persists. Look for recurring habits: does the model hedge in the same way each time? Does it favour certain framings? Does it structure its reasoning similarly regardless of how the request is phrased?
Over time, patterns might emerge. A model could have a consistent hedging frequency, a characteristic risk posture, a default rhetorical rhythm, or a predictable balance between consensus and dissent. These are not emotions or beliefs — they are structural tendencies, the residue of its architecture, training, and fine-tuning.
It’s worth stressing that this is not about discovering some hidden “true self” of the model. Instead, it is about learning the shape of its thinking space: the characteristic ways it tends to traverse the probabilistic landscape of possible outputs.
If we could map such dispositions — even provisionally — it might change how we interact with these systems. We could learn which tendencies to trust, which to challenge, and where to apply external constraints. And perhaps more importantly, it could give us a way to track how a model’s psychology changes over time, as new training data, tuning processes, or architectures are introduced.
This is not yet a discipline. At best, it is an early sketch of what such a discipline could look like. But maybe that’s the point — to make this an open field of inquiry where the methods are developed collaboratively, tested in the open, and challenged from multiple perspectives.
Because if machine psychology is possible, it should not be the property of a single lab or company. It should be something we can all learn to see — and debate — together.
5. Why This Could Matter
If this idea of machine psychology holds water — even partially — it might change the kinds of questions we ask about AI systems, and the kinds of answers we expect.
So far, much of our evaluation has been narrowly technical. We measure accuracy, benchmark performance on standard datasets, or count the frequency of “hallucinations.” These tests tell us something, but they tell us about episodes, not dispositions. They can reveal whether a model is correct in a single instance, but not how it tends to behave over hundreds, thousands, or millions of interactions.
Yet dispositions might matter just as much as correctness — perhaps more — once these systems are woven into the fabric of our daily information environment.
Imagine two models that produce equally accurate summaries of a news story. If one consistently presents the story in a way that minimises conflict, while the other tends to foreground controversy, they will shape the perceptions of their users in very different ways. Accuracy alone wouldn’t tell you that.
Or consider decision-making contexts. A model with a strong bias toward hedging might be valuable for generating options in a brainstorming session, but less useful in a time-sensitive crisis where clarity is needed. Conversely, a model with a high “risk appetite” in its outputs might offer novel insights in research, but could be dangerous in safety-critical applications.
These kinds of tendencies could influence:
Science and research — what hypotheses are surfaced or suppressed.
Governance and policy — how risks and trade-offs are framed.
Culture and discourse — which voices and perspectives are amplified or diminished.
Ethics and norms — how moral or social boundaries are reinforced or reinterpreted.
This isn’t to say that machine psychology is a silver bullet for understanding all of this. It could be flawed, partial, or even misleading if handled carelessly. But it might offer a new lens — one that complements accuracy metrics with an understanding of style, tendency, and behavioural texture.
The bigger question is whether we, as a society, want to know these things at all. Mapping dispositions could make models more transparent and accountable — but it could also make them more “tunable” in ways that entrench certain biases or agendas.
Which is why, if this is worth doing, it might be worth doing together. Not as a closed practice inside corporate labs, but as an open field where methods can be scrutinised, results contested, and interpretations debated.
Because in the end, machine psychology — if it exists — wouldn’t just be about the machines. It would be about us, and the ways we choose to see, measure, and shape the minds we are building.
6. The Risk of Ignoring the Conjunction
It might be tempting to leave the phrase machine psychology alone — to treat it as a curiosity, an interesting metaphor that collapses under serious scrutiny. If these systems have no inner life, why burden ourselves with a term that seems to imply otherwise?
But if we put the term aside, we may also put aside what it could help us notice.
We might overlook the fact that these systems are already shaping the flow of information, even if they do so through patterns that are subtle and difficult to detect. We might fail to see that these patterns — their habits of framing, their risk postures, their rhetorical preferences — can shift between one model update and the next, often without public notice.
What happens if we never ask how these patterns change over time? What happens if we cannot tell whether a model is becoming more cautious or more confident, more deferential to certain worldviews or more prone to challenge them?
And what happens if we only discover these shifts when their effects are already embedded in our public discourse, our institutional decision-making, or our cultural norms?
To ignore this is to let the deep structure of these systems evolve without our awareness — to accept that we will live within informational environments whose underlying “thinking styles” we neither map nor debate.
This is not an argument for certainty. It’s an argument for curiosity. For treating machine psychology as a hypothesis worth testing, rather than a claim to be either embraced or dismissed outright. If we try and find nothing of value, then at least we will have looked. But if we find something — a persistent structure of tendencies, a measurable drift, a pattern that matters — then perhaps we will have found a new way to see the systems that are increasingly mediating our reality.
In that sense, the risk is not that machine psychology turns out to be wrong. The risk is that it turns out to be right — and we realise too late that we never bothered to study it.
7. An Invitation
So perhaps machine psychology is an awkward phrase. Perhaps it carries too much risk of misunderstanding. Perhaps it draws suspicion because it sounds like a smuggling-in of human qualities where none exist.
And yet — it could also be a useful provocation, a way to ask different questions about systems that are already altering how we think, communicate, and decide. Questions like:
What are the persistent tendencies of a given model
How do those tendencies shift across updates, contexts, and applications?
Which of these patterns help us, which hinder us, and which simply shape us in ways we haven’t yet noticed?
I’m not claiming to have a finished framework. At best, what’s here is a sketch — a set of ideas that could be sharpened, dismantled, or rebuilt through collective work. Any attempt to study the psychology of machines will be partial and fallible. It will raise as many questions as it answers. And that is exactly why it seems worth trying.
This could be a shared project. Not owned by any one lab, institution, or discipline, but explored in the open: methods published, results challenged, interpretations debated. A space where engineers, social scientists, philosophers, designers, and everyday users could all have a role in shaping how — and whether — this idea evolves.
If machine psychology is only a metaphor, then maybe we can retire it after we’ve tested its limits. But if it turns out to be a lens that lets us see something important — a way to map the otherwise invisible structure of these probabilistic minds without selves — then perhaps we will have given ourselves a tool we badly need.
Either way, the work would not be about proving the phrase right or wrong. It would be about asking whether it can open up new kinds of seeing, new kinds of conversation, and new kinds of responsibility.
I wonder if you have had a conversation about this with Vanessa Andreotti? - that would make for a very interesting podcast I would think.