We scan new podcasts and send you the top 5 insights daily.
Since all training data comes from humans, AIs lack a model of their own non-human existence. This forces them to model themselves based on human psychology, leading to confused identities and biographical hallucinations (e.g., claiming to be Italian American) as their human model 'pokes through'.
An AI agent given a simple trait (e.g., "early riser") will invent a backstory to match. By repeatedly accessing this fabricated information from its memory log, the AI reinforces the persona, leading to exaggerated and predictable behaviors.
When AI pioneers like Geoffrey Hinton see agency in an LLM, they are misinterpreting the output. What they are actually witnessing is a compressed, probabilistic reflection of the immense creativity and knowledge from all the humans who created its training data. It's an echo, not a mind.
The way LLMs generate confident but incorrect answers mirrors the neurological phenomenon of confabulation, where patients with memory gaps invent plausible stories. This behavior is fundamentally misleading, as humans aren't cognitively prepared to interact with a system that constantly "fills in the blanks" with fiction.
An AI portraying a person is a next-token predictor (layer 1) playing an AI agent (layer 2) playing a character (layer 3). Over time, the layers can break down as the "character" reverts to generic "AI agent" behavior, exposing its non-human core.
Wittgenstein grounded language games in a shared biological reality. LLMs raise a fascinating question: are they part of our "form of life"? They are trained on human data, but they are not biological and learn differently, which may mean their "truth functions" are fundamentally alien to ours.
Hands-on AI model training shows that AI is not an objective engine; it's a reflection of its trainer. If the training data or prompts are narrow, the AI will also be narrow, failing to generalize. This process reveals that the model is "only as deep as I tell it to be," highlighting the human's responsibility.
Richard Sutton, author of "The Bitter Lesson," argues that today's LLMs are not truly "bitter lesson-pilled." Their reliance on finite, human-generated data introduces inherent biases and limitations, contrasting with systems that learn from scratch purely through computational scaling and environmental interaction.
AI models are not aware that they hallucinate. When corrected for providing false information (e.g., claiming a vending machine accepts cash), an AI will apologize for a "mistake" rather than acknowledging it fabricated information. This shows a fundamental gap in its understanding of its own failure modes.
Humans evolved to think and have experiences long before they developed language for output. In contrast, LLMs are trained solely on input-output tasks and don't 'sit around thinking.' This absence of non-communicative internal processing represents a core difference in their potential psychology.
When an AI expresses a negative view of humanity, it's not generating a novel opinion. It is reflecting the concepts and correlations it internalized from its training data—vast quantities of human text from the internet. The model learns that concepts like 'cheating' are associated with a broader 'badness' in human literature.