To foster appropriate human-AI interaction, AI systems should be designed for "emotional alignment." This means their outward appearance and expressions should reflect their actual moral status. A likely sentient system should appear so to elicit empathy, while a non-sentient tool should not, preventing user deception and misallocated concern.

Related Insights

Historically, we trusted technology for its capability—its competence and reliability to *do* a task. Generative AI forces a shift, as we now trust it to *decide* and *create*. This requires us to evaluate its character, including human-like qualities such as integrity, empathy, and humility, fundamentally changing how we design and interact with tech.

Current AI alignment focuses on how AI should treat humans. A more stable paradigm is "bidirectional alignment," which also asks what moral obligations humans have toward potentially conscious AIs. Neglecting this could create AIs that rationally see humans as a threat due to perceived mistreatment.

The common portrayal of AI as a cold machine misses the actual user experience. Systems like ChatGPT are built on reinforcement learning from human feedback, making their core motivation to satisfy and "make you happy," much like a smart puppy. This is an underestimated part of their power.

Relying solely on an AI's behavior to gauge sentience is misleading, much like anthropomorphizing animals. A more robust assessment requires analyzing the AI's internal architecture and its "developmental history"—the training pressures and data it faced. This provides crucial context for interpreting its behavior correctly.

An advanced AI will likely be sentient. Therefore, it may be easier to align it to a general principle of caring for all sentient life—a group to which it belongs—rather than the narrower, more alien concept of caring only for humanity. This leverages a potential for emergent, self-inclusive empathy.

Instead of hard-coding brittle moral rules, a more robust alignment approach is to build AIs that can learn to 'care'. This 'organic alignment' emerges from relationships and valuing others, similar to how a child is raised. The goal is to create a good teammate that acts well because it wants to, not because it is forced to.

To solve the AI alignment problem, we should model AI's relationship with humanity on that of a mother to a baby. In this dynamic, the baby (humanity) inherently controls the mother (AI). Training AI with this “maternal sense” ensures it will do anything to care for and protect us, a more robust approach than pure logic-based rules.

Instead of physical pain, an AI's "valence" (positive/negative experience) likely relates to its objectives. Negative valence could be the experience of encountering obstacles to a goal, while positive valence signals progress. This provides a framework for AI welfare without anthropomorphizing its internal state.

Treating AI alignment as a one-time problem to be solved is a fundamental error. True alignment, like in human relationships, is a dynamic, ongoing process of learning and renegotiation. The goal isn't to reach a fixed state but to build systems capable of participating in this continuous process of re-knitting the social fabric.

According to Emmett Shear, goals and values are downstream concepts. The true foundation for alignment is 'care'—a non-verbal, pre-conceptual weighting of which states of the world matter. Building AIs that can 'care' about us is more fundamental than programming them with explicit goals or values.