Author
Listed:
- Araújo, Diogo Luiz Alves de
Abstract
This article contends that the prevalent anthropomorphism in Large Language Model (LLM) alignment research constitutes a fundamental category error, rooted in psychological projection. By describing LLMs with human-centric terms like “deception” and “intent,” the field mischaracterizes the technology, leading to flawed threat models and misguided safety evaluations. I first deconstruct the LLM as a mathematical and statistical system, demonstrating how its convincing mimicry of cognition emerges from probabilistic pattern-matching, not genuine understanding. I then establish a philosophical threshold for moral agency by synthesizing Humean, Kantian, and phenomenological perspectives, arguing that agency requires affective sentiment, rational autonomy and subjective, temporal experience—all of which are absent in LLMs. Using a Jungian framework, I re-interpret studies on “deceptive” and “scheming” AI not as discoveries of emergent malice, but as manifestations of the projection of our own “Shadow” onto an opaque technological artifact. This misinterpretation leads to dangerous, quasi-mythological narratives of AI risk, exemplified by reports such as 'AI 2027'. As an alternative, I propose a grounded paradigm for alignment that shifts focus from human-like malice to non-human failure modes. This paper concludes not that LLMs are harmless, but that danger is misplaced. The risk arises when a non-rational text generator is connected to real-world tools and functions as an advisor to end-users and geopolitical leaders, a situation that demands conspicuous communication about the technology's scripted nature and inherent limitations.
Suggested Citation
Araújo, Diogo Luiz Alves de, 2025.
"On the Psychology of a Large Language Model,"
SocArXiv
adxbj_v1, Center for Open Science.
Handle:
RePEc:osf:socarx:adxbj_v1
DOI: 10.31219/osf.io/adxbj_v1
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:osf:socarx:adxbj_v1. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: OSF (email available below). General contact details of provider: https://arabixiv.org .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.