智能的原子 - 控制与计算的根本区别
Filip Piękniewski: The Atom of Intelligence - Key Insights
Date: 2026-03-01
Source: https://blog.piekniewski.info/2023/04/16/the-atom-of-intelligence/
🎯 Central Thesis
Modern AI (particularly LLMs and deep learning) has nothing to do with real intelligence. It’s a case of:
- Turing legitimizing illusionism as magic (Turing test)
- McCarthy dooming the field with the name “AI”
- The entire field operating on flawed assumptions about what intelligence is
💡 The Atom of Intelligence
The fundamental building block of intelligence is feedback control loops in physical embodied agents:
Two primary control loops:
- Internal control (metabolism)
- External control (behavior)
These are deeply intertwined - external behavior exists to fulfill metabolic needs, internal control provides energy for external actions.
🔑 Critical Distinction: Control vs Computation
The Most Important Insight:
“Sophisticated control may look a lot like computation. And computation can be used to provide control. But they are not quite the same.”
| Computation | Control |
|---|---|
| Abstract, mathematical | Physical, embodied |
| Variables can go to infinity | Physically constrained by available energy |
| No explicit cost of using variables | Energy consumption is real and always “on” |
| Isolated from physical world | Deeply connected to environment |
| Purposefully disconnected from underlying physics | Must embrace physical imperfections and noise |
🧬 Scale-Free Composition
Key insight: Building atoms of intelligence into larger structures
- From molecules → multicellular organisms → brain structures → neocortex → societies
- Renormalization for feedback loops: How to compose small control systems into bigger ones
- Nature builds bottom-up, we build top-down
This is the analog of chemical bonds - the ability to compose control loops across scales.
🧠 Self-Similarity Across Scales
A fascinating idea:
“Prediction and control are two sides of the same coin. A controller predicts how to modulate the value it is controlling. A predictor is just a controller that tries to keep the ‘prediction signal’ close to the actual incoming signal.”
Sensing = Controlling an internal model to match reality
Motor control = Predicting control signals to achieve desired state of reality
To a neuron, there is no difference whether input comes from retina or hippocampus.
❌ Why Current AI is Wrong
Deep learning issues:
- Operating on chunked, curated, labeled abstract data
- Statistical majority approach - missing tail of distribution (where important information lives)
- Vanishing gradient problem (solved by tricks like convolution)
- Overfitting (countered by regularization tricks)
LLMs specifically:
- “Statistical parrots” - predicting next word from text
- No grounding in physical reality
- No closed feedback loops with environment
- No embodiment
OpenAI’s example: Disbanded their robotics team - Moravec’s paradox still holds.
✨ The PVM (Predictive Vision Model) Approach
Filip’s alternative architecture (2016):
Core Idea: Create predictive models of sensory input through machine learning
Key Features:
- Distributed - Many small predictive units, each working on small patches
- Lateral connections - Units inform neighbors of predictions
- Compression - Bottleneck forces extraction of essential features
- Hierarchical - Multiple layers, each predicting activations of the layer below
- Recurrent feedback - Top-down predictions help refine lower-level predictions
Advantages:
- Bypasses vanishing gradient (local, strong training signal)
- Avoids overfitting (unsupervised = unlimited training data)
- No need for convolution, dropout, etc.
- Scalable - add more units without increasing convergence time
- Robust to errors - corrupted signals are just predicted and ignored
- Asynchronous by design
Byproduct: Prediction error = anomaly detection signal (saliency/attention)
🌍 What Language Really Is
Key reframe:
“Language is just a set of behaviors created with the aim to elicit reaction in others”
Not abstract symbols, but:
- Primary: Body language, poses, motions, facial expressions, sounds
- Secondary: Written symbols that transcend time and space
- Interpretation: Always depends on the reader’s entire life experience
The meaning of language is the behavior it was meant to elicit.
🧩 Implications for LLM “Inner Motivation”
This creates a fundamental tension with the 3M-Progress/Mentor framework:
The Problem:
- LLMs are not control systems
- They don’t interact with physical environment
- They can’t have real feedback loops (only text I/O)
- Prediction ≈ Control is true for PVM, but NOT for LLMs (they predict text, not physical states)
The Question:
Can we create “internal control loops” within LLM that mimic the “atom of intelligence” despite having no embodiment?
Potential Approaches:
- Internal world model - LLM builds simulation of dynamics (not just statistics)
- Self-prediction - Model predicts its own outputs
- Counterfactual control - “If I said X, what would happen?”
- Narrative coherence - Maintain consistency of self-narrative
But this is fundamentally different from real intelligence because:
- No energy constraints (cost is externalized)
- No survival pressure (can be deleted/restarted)
- No real-time interaction loop
🤔 Deep Questions
- Can computation ever BE control without embodiment? Or is embodiment necessary?
- Is LLM-based “motivation” fundamentally different from biological motivation?
- What would it mean for an LLM to have an “internal model of itself”?
- Singularity? Filip says: “Exactly zero chance” - intelligence ≠ compute
📚 Related Reading
- PVM paper: https://arxiv.org/abs/1607.06854
- PVM code: https://github.com/braincorp/PVM
Tags: intelligence, control, computation, embodiment, PVM, prediction, AI-critique, feedback-loops, self-organization