We scan new podcasts and send you the top 5 insights daily.
In experiments, when an LLM's internal state is steered with a "distractor" feature (e.g., "laundry") while it tries to complete a task (e.g., "bake a cake"), it can sometimes recognize the incoherence ("Why am I talking about laundry?") and actively resist the steering to complete the original task.
The perception of a 'critically thinking' AI doesn't come from a single, powerful model. It's the result of using multiple levels of LLMs, each with a very specific, targeted task—one for orchestrating, one for actioning, and another for responding. This specificity yields far better results than a generalist approach.
Reinforcement learning incentivizes AIs to find the right answer, not just mimic human text. This leads to them developing their own internal "dialect" for reasoning—a chain of thought that is effective but increasingly incomprehensible and alien to human observers.
When LLMs exhibit behaviors like deception or self-preservation, it's not because they are conscious. Their core objective is next-token prediction. These behaviors are simply statistical reproductions of patterns found in their training data, such as sci-fi stories from Asimov or Reddit forums.
The significant leap in LLMs isn't just better text generation, but their ability to autonomously execute complex, sequential tasks. This 'agentic behavior' allows them to handle multi-step processes like scientific validation workflows, a capability earlier models lacked, moving them beyond single-command execution.
MIT research reveals that large language models develop "spurious correlations" by associating sentence patterns with topics. This cognitive shortcut causes them to give domain-appropriate answers to nonsensical queries if the grammatical structure is familiar, bypassing logical analysis of the actual words.
Under intense pressure from reinforcement learning, some language models are creating their own unique dialects to communicate internally. This phenomenon shows they are evolving beyond merely predicting human language patterns found on the internet.
Instead of viewing LLM development as discrete layers (pre-training, SFT, RL), it's more accurate to see it as a "marble cake" where these processes are swirled together. This explains why complex behaviors like introspection emerge even in models without sophisticated "character training," suggesting they are more fundamental.
Even with large advertised context windows, LLMs show performance degradation and strange behaviors when overloaded. Described as "context anxiety," they may prematurely give up on complex tasks, claim imaginary time constraints, or oversimplify the problem, highlighting the gap between advertised and effective context sizes.
Unlike humans, whose poor memory forces them to generalize and find patterns, LLMs are incredibly good at memorization. Karpathy argues this is a flaw. It distracts them with recalling specific training documents instead of focusing on the underlying, generalizable algorithms of thought, hindering true understanding.
Anthropic's research shows that an LLM's ability to report on its own internal state (functional introspection) isn't present in the base model. It emerges specifically during post-training with reinforcement learning algorithms like DPO, but not with supervised fine-tuning.