WisPaper
WisPaper
学术搜索
学术问答
论文订阅
价格
TrueCite
[Science Robotics 2026] Evidence of an Emergent “Self” in Continual Robot Learning: How AI Carves Out a Sense of Body
总结
问题
方法
结果
要点
摘要

This paper identifies the emergence of a stable, task-invariant "self" subnetwork in robots undergoing continual reinforcement learning. By training a simulated quadruped on a sequence of distinct locomotion behaviors (walk, wiggle, bob), the authors demonstrate that a specific functional core remains persistent while other neural components reorganize to adapt to new tasks.

TL;DR

Researchers at Columbia University have discovered that when robots are forced to learn many different tasks—like walking, wiggling, and bobbing—their neural networks naturally split into two parts: a stable "self" core that represents the unchanging body, and a plastic "task" exterior that changes with every new skill. This "self" emerges spontaneously, without any special programming, acting as a persistent cognitive anchor across a robot's lifetime.

Background: The Persistent Individual

Since the days of Descartes, "selfhood" has been a philosophical enigma. In robotics, we often ask: Does a robot only know how to walk, or does it know what kind of body it is walking with? Most current AI treats every task as a fresh start, often leading to "catastrophic forgetting."

The authors hypothesize that the "self" is simply the invariant portion of experience. While a robot's goals change (moving forward vs. dancing in place), its leg lengths, joint limits, and mass remain constant. This paper seeks the neural signature of that physical constancy.

Methodology: Finding the "Self" in the Spaghetti

To find the self, the team trained a 2-layer MLP (Soft Actor-Critic) on a cyclical curriculum: Walk → Wiggle → Bob.

The technical challenge was "disentangling" a dense neural network. They used a three-step pipeline:

  1. Co-activation Mapping: They grouped neurons that "fire together."
  2. Block Diagonalization: They reordered the matrix to visualize functional modules (subnetworks).
  3. Cross-Task Alignment: Using the Hungarian algorithm, they matched neurons from one task to the next to see which ones "survived" the transition.

Model Architecture and Pipeline Figure 1: The training and comparison pipeline. By evaluating different behaviors on shared reference states, the researchers could isolate which neurons remained "loyal" to the body's identity versus the task's demands.

The "Goldilocks" Regime of Self-Emergence

The results were striking. In robots trained on only one task (the control group), the internal structure remained diffuse and "spaghetti-like." However, in the continual learning group, a dominant, stable subnetwork formed in the first layer.

  • Persistence: The "self" neurons changed significantly less than the "task" neurons (a gap of ~17 percentage points).
  • Layer Depth: The "self" was most prominent in the first hidden layer (closest to the sensors), suggesting it acts as a foundational "body-state" processor.
  • Capacity: This emergence happens best in a "Goldilocks" regime—where the network is large enough to learn multiple tasks but small enough that "reusing" a core body model is computationally efficient.

Visualizing the Self Subnetwork Figure 2: Alluvial plots showing the "Self" subnetwork (dark purple). Notice how in the variability condition (right), one thick, stable band persists across Walk, Wiggle, and Bob, while other groups split and merge.

Quantitative Evidence

The statistical evidence for this "self" core is overwhelming. Using a sample of over 900 behavior transitions, the researchers found that the task-specific regions of the brain underwent much more aggressive reorganization.

Quantitative Stability Comparison Figure 3: Co-activation matrices. The top-left localized blocks in the "Continual" condition represent the emergent self-model, which correlates with high persistence scores (bottom panel).

Critical Insight: Why This Matters

This work shifts our understanding of AI from "monolithic calculators" to "modular organisms."

  • Efficiency: If we can identify a robot's "self" subnetwork, we can freeze it when teaching the robot a 100th task, preventing the loss of 99 previous skills.
  • Resilience: A stable self-model is the first step toward "metabolism"—robots that can recognize when their body has changed (e.g., a broken leg) and update only their self-core.
  • The Philosophy of AI: It suggests that "selfhood" isn't a mystical quality, but a byproduct of biological (and digital) pressure to find invariants in a changing world.

Conclusion

By looking for what doesn't change, Jhunjhunwala et al. have provided a lens into the "inner life" of a robot. The emergent self is the "body backbone"—a compact, task-agnostic representation that says, "No matter what I am doing, this is who I am."

Takeaway for Practitioners: When designing lifelong learning agents, don't just optimize for the task. Monitor for representation stability; the part that stays still might be the most important piece of the puzzle.

发现相似论文

试试这些示例

  • Search for recent papers investigating the spontaneous emergence of modularity in deep reinforcement learning agents during multi-task or lifelong learning settings.
  • Which prior studies first utilized co-activation matrices and block diagonalization for interpreting the internal representations of neural network controllers in robotics?
  • Explore how a persistent "self" subnetwork, once identified, can be explicitly frozen or regularized to improve transfer learning and mitigate catastrophic forgetting in humanoid robotics.
目录
[Science Robotics 2026] Evidence of an Emergent “Self” in Continual Robot Learning: How AI Carves Out a Sense of Body
1. TL;DR
2. Background: The Persistent Individual
3. Methodology: Finding the "Self" in the Spaghetti
4. The "Goldilocks" Regime of Self-Emergence
5. Quantitative Evidence
6. Critical Insight: Why This Matters
7. Conclusion