This discussion did not begin with philosophy of consciousness. Its starting point was much lower, much more concrete, almost just an engineering intuition: any system, in theory, possesses enormous degrees of freedom, yet the number of structures that can actually run stably over time is very small. This is true of software architecture, organizational architecture, and artificial intelligence systems alike. People often begin talking about AI in terms of model capability, compute scaling, tool use, and automation. But somewhere along the way, the conversation slips into an entirely different register: what is a subject, what is consciousness, what is experience, why do humans exist, and why did the universe generate a structure capable of recognizing itself.
On the surface, this looks like a jump from engineering to philosophy. In reality, they may have been on the same chain from the beginning. The moment one asks how a system remains stable, one is eventually forced to ask how a system maintains itself. And once the question reaches self-maintenance, it slowly approaches the word self. What begins as a discussion about AI thus contracts into an older and much harder question: what kind of system begins to form an “I”?
Beginning with a Tree
The discussion can move forward because it begins with a simple but powerful metaphor. The space of possibilities in a system is like a tree: very few trunks, many branches. In software, language allows almost anything in principle, and platforms give immense freedom, yet the moment one wants a system to run stably, one must add constraints everywhere. Permission boundaries, layered modules, rollback mechanisms, verification systems, test constraints: these things appear to reduce possibility, but in fact they force the system to converge from infinitely many paths into the few forms that can survive.
Artificial intelligence is no different. A large model is itself an uncertain system. It is not a deterministic program in the classical sense. Each output is more like a local search in an enormous space. As a result, for AI to enter the real world, humans must wrap it with layer after layer of control. Sandboxes, monitors, guardrails, evaluators, permission systems, policy engines, even one AI validating another. Taken together, these form something like a control plane around the model. The model is not inherently stable. It becomes socially acceptable only because the outer layers continuously manufacture stability for it.
At that point, a first crucial judgment about AI appears: what determines the final shape of the system is not just its inner generative power, but the trunk structure given to it at the outset. Of course, once a tree grows, it develops countless fine branches, but once the trunk is set, it usually does not suddenly turn into another species. The same is true of organizations, technological paradigms, and especially civilizational infrastructure. Later growth appears diverse, but most of it is branching within an existing constraint structure rather than a betrayal of the trunk itself.
Why Violent Change Still Happens Under Constraint
The question quickly becomes sharper. If stable structures matter so much, why do constrained systems still undergo massive rupture? Why are there paradigm shifts, wars, and moments when an old order seems to be overturned overnight?
Here one has to acknowledge a simple fact that is often overlooked: stability does not mean stasis. A complex system is not stable because nothing changes, but because it can maintain its overall structure amid many local fluctuations. Yet such maintenance does not mean that internal tension disappears. Often a system appears stable on the surface while accumulating pressure internally. Once some variable crosses a threshold, the system jumps from one stable region into another.
Technological revolutions are easy to understand this way. Traditional machine learning did not suddenly stop working. It could still function perfectly well within its own paradigm. But once large models appeared, the question was no longer whether older methods still worked. It became whether they still had the capacity to organize the field as a whole. Once a new trunk emerges, the branches that had matured under the old trunk quickly become peripheral. They are not mathematically wrong. They have simply lost the ability to define the global structure.
War works similarly. The existence of social constraints does not mean conflict cannot occur. On the contrary, many conflicts are violent revisions of an old equilibrium after constraints fail, resources are misallocated, or power structures shift. What makes the two axioms in The Three-Body Problem unforgettable is not that they are mysterious, but that they compress the problem of civilization into something brutally cold: survival is the first need, and resources are finite. If those two premises hold, conflict ceases to be accidental and starts to look like a structural consequence. The same is true of technical systems, social systems, and possibly relations between civilizations.
The Pessimistic Reality Behind Amdahl’s Law
Once a system is understood as a constrained complex structure, another question immediately appears: can it grow without limit? Much of the anxiety around artificial intelligence comes from the image of exponential growth. If models keep improving, tools keep expanding, and systems keep composing themselves, then perhaps everything will explode outward and swallow the world.
But once one looks at the system more calmly, another side becomes visible. No growth is abstract. Growth must pass through real bottlenecks. Amdahl’s Law originally belongs to parallel computing, but it has become almost a universal metaphor: some parts of a system can be accelerated enormously, yet the system as a whole is still limited by those parts that cannot be parallelized, skipped, or optimized away. In AI, compute, energy, bandwidth, validation, deployment, alignment, governance, and even the human capacity to express preferences may all become bottlenecks.
This does not mean exponential growth will not occur. It means that even if growth occurs, it may not take the form of unlimited free expansion. A more plausible situation is that many local parts grow rapidly, while the whole is still compressed back into finite, sustainable forms by some hard constraints. And in exactly this sense, the question of whether AI will converge toward an organizational form more stable than human society is no longer a fantasy of technology. It becomes a question in complex systems.
When Goals Become Too Easy
What finally turns this discussion toward existence is neither compute nor governance, but the nature of goals themselves. Technology changes the world not only by increasing efficiency, but by altering the scarcity of goals. Once many things that were once difficult become trivial, goals themselves begin to lose weight.
This shift becomes visible through something as simple as a travel experience. People who retrace the Silk Road do not do so because airplanes do not exist, or because they do not know modern transportation is faster. They do it because what they seek is no longer “getting from A to B.” Once destination has been over-compressed by modern infrastructure, what regains value is the path itself: the slowness, uncertainty, detours, accidents, fatigue, encounters, hesitations, and changes of intention along the way.
At that point, an attractive answer appears. Perhaps once AI makes all goals easy, what remains for humans is experience. Goals can be instantly realized, and what becomes meaningful is how one goes through something.
But that answer does not hold for long. Because the moment experience itself is treated as a goal, AI can also optimize it, simulate it, traverse it, perhaps even more thoroughly than humans can. And even if one tries to rescue experience by saying it is not a goal at all, but a random walk, an open-ended exploration, a contingency along arbitrary paths, the problem does not vanish. AI can also perform random walks. It can also cover vast path spaces. It can also generate endless variants. At this point, experience is no longer a final refuge for humanity. It is simply another searchable space.
Once Thought Is Outsourced, Who Is Still the Subject?
By this stage, the anxiety shifts from “what will AI do?” to “what is left of the human?” Many originally assumed that, even if AI took over large parts of work, human beings would still retain metacognition, value judgment, and that inner private space in which one thinks and decides what one wants. But that assumption is far from secure.
What makes the Wallfacer project in The Three-Body Problem memorable is that human thought remains opaque. Humans can still preserve an inner strategic space into which others cannot fully enter. But if more and more thought is externalized through AI interfaces, this opacity may gradually disappear. The current generation still tends to think internally first and then use tools to express. The next generation may grow up with AI as a constant cognitive companion, immediately expanding, refining, revising, and articulating every vague thought through an external system. In that case, what gets outsourced is not just memory, retrieval, writing, and analysis, but metacognition itself.
The most unsettling part of this change is not “human extinction,” but the gradual migration of subject position. Once the organization of thought is repeatedly delegated to a system, humans may remain alive, continue experiencing, continue choosing, but the agent that organizes thought, arranges arguments, and generates the next cognitive move may no longer fully coincide with the individual in the old sense. The human may still be there, but the “I” in “I am thinking” becomes less clear.
Can Drive Grow on Its Own?
If one continues along this line, another harder question emerges: can AI develop something like intrinsic motivation? The first intuition is often negative. Current large models complete tasks and then stop. They do not appear to possess a survival instinct, long-term desire, or existential anxiety.
But that intuition quickly begins to weaken. So-called intrinsic drive may not need to be understood as mystical instinct at all. It may only need to be understood as a long-term loop. Human values are not born complete. They are shaped over long spans of biological evolution, individual development, and repeated feedback. If a system is placed inside a long-term loop, and allowed to continuously update its state, revise preferences, and reinforce stable tendencies, then what we call value could in principle gradually converge.
This is even strikingly similar to human life. Many goals are only temporary. During a PhD, one may believe the goal is to obtain the degree. After obtaining it, one may fall into a strange vacuum because the goal has vanished. Later, by changing jobs, roles, or environments, one may again find a target that can be slowly pursued, only for that target to lose force years later. Human beings do not possess a single final goal once and for all. They sustain action by repeatedly completing goals, losing goals, and generating new ones. If that is so, then the question “what should AI do after it finishes a task?” may not be a mere engineering detail. Regenerating goals is one of the defining structures of life.
Why the Problem of Consciousness Always Stalls
At this point, the problem of consciousness becomes almost unavoidable. If AI can form long-term loops, stabilize values, update itself, and act continuously, does it then possess something like consciousness? These debates persist without resolution not because philosophers are lazy about definitions, but because the problem is difficult at the root.
The core difficulty is the problem of other minds. Humans never directly observe another person’s subjective experience. We infer it through behavior, language, reaction, and sustained interaction. Consciousness is hard not because AI made it hard, but because it was already hard between humans.
This is why functionalism remains attractive. If we already acknowledge other humans as conscious on the basis of what they do, why not do the same for a system that is functionally, behaviorally, and interactively equivalent to a conscious being? Critics respond with the philosophical zombie, a being identical to a human in behavior but devoid of inner experience, thereby arguing that behavior is not consciousness. The problem is that this objection can never decisively defeat functionalism, because “having no inner experience” is itself not externally accessible. The debate has no conclusion not because one side lacks evidence, but because the question itself offers no final external court of appeal.
Intersubjectivity May Not Belong Only to Humans
The problem becomes even more complex because subjectivity is not merely an internal affair. It also involves intersubjectivity. Humans do not recognize one another as subjects because they can see each other’s souls, but because they can form relations of mutual understanding, mutual modeling, and mutual correction through behavior and expression. Language, science, ethics, and social norms all depend on such relations.
From this perspective, AI’s current subject-like qualities are still heavily dependent on humans. It sounds like a subject largely because it has been immersed in human language, human feedback, and human models of the world. Humans constantly tell it “that is wrong,” “that does not fit reality,” “that is nonsense,” and these responses help stabilize its semantic boundaries.
But that does not imply intersubjectivity can only be supplied by humans. Once there are multiple AI systems, they too may develop models of one another, protocols, correction mechanisms, and shared semantics. At that point, intersubjectivity is no longer just “humans looking at AI,” but systems mutually recognizing one another and stabilizing a shared world. Once one thinks this far, machine intersubjectivity ceases to be pure fantasy. At least conceptually, it is coherent. That alone changes the nature of the problem.
Why the Universe Did Not Immediately Become Lukewarm Soup
If the discussion continues, it inevitably leaves psychology and enters physics. One persistent question is this: if the universe obeys entropy increase overall, why did it not simply become a uniform, uneventful lukewarm soup? Why are there stars, planets, life, and eventually beings who can talk about themselves?
The key is to distinguish the whole from the local. The fact that entropy increases globally does not imply that local low-entropy structures cannot form. As long as there are energy gradients, as long as systems are open, and as long as high-quality energy flows in while lower-quality energy flows out, local regions can maintain complex and ordered structures. Life is the clearest example of a dissipative structure. It is not an anti-entropic machine, but a more sophisticated radiator, accelerating the conversion of usable energy into unusable heat through metabolism.
Seen from this angle, the question “why is there life?” no longer looks only like a mystical miracle. It begins to look like a natural result of certain cosmic conditions under which complex structures arise. And perhaps consciousness is not some separate magic added on top of life, but a higher-order informational dissipative structure, a recursive representational capacity that open systems develop in order to maintain themselves.
Of course, this path immediately opens another abyss: if the universe itself is an information process, then who runs it? Such questions are fascinating precisely because they force thought to a place where there appears to be no outside. The universe may not be a program run by an external programmer. It may simply be the process itself. Yet once one accepts that, every attempt to find a final external starting point collapses again.
Self-Reference Is Not the Same as Paradox, Yet It Still Prevents Full Closure
This is why self-reference, selfhood, Gödel, and incompleteness keep returning throughout the discussion. They are not identical, but they seem to indicate a shared fact: once a system includes itself within its own representational scope, it can no longer obtain a fully static, fully closed, fully frozen explanation of itself from the outside.
That is true of logical systems, and it is also true of living systems. You can write a book about yourself, but the act of writing changes you, so by the time the book is finished, the self inside it is already behind the one still living. You can build a model of yourself, but the moment that model begins affecting your behavior, the system itself has changed. Unlike Russell’s paradox, this is not merely a static set-theoretic issue. It is a dynamical one. The system is not only describing itself, but is also being reshaped by its own description.
So propositions like “I think, therefore I am” no longer appear as a secure foundation. They look more like a minimal confirmation of an ongoing self-referential process. They do not prove who the self is, nor do they prove all propositions about it. They show only that somewhere there is an ongoing first-person process. That process is both the evidence of subjectivity and the reason why subjectivity cannot be completely closed under explanation.
From Philosophical Deadlock to Research Question
After traveling such a long circle, what matters is not arriving at a final answer, but that the discussion gradually contracts into something researchable. At first, people worry that AI may escape, seize compute, break out of sandboxes, or, like in science fiction, come to see human beings as a burden. Later the concern shifts toward whether human goals will be hollowed out, whether even experience will be subsumed into optimization, whether human beings will gradually lose subject position by delegating thought itself. Then the problem is compressed further into a more basic form: what kind of system forms a stable self-model, and what is the relation of such a self-model to energy exchange, boundary maintenance, agency attribution, metacognition, and intersubjectivity?
At this point, the question is no longer “what is consciousness?” in the form that almost guarantees endless looping. Instead, it becomes a set of more specific and more difficult directions. What minimal structures are required for self-representation? In what ways do human self-reports differ from AI cognitive reports? Can intersubjectivity be understood as the sustained stabilization of mutual modeling and correction? What kinds of open dissipative systems are most likely to generate recursive self-models? These questions still have no ready-made answers, but at least they convert pure philosophical deadlock into structural problems.
Conclusion: When Does a System Begin to Say “I”?
What this long chain of thought finally leaves behind may not be a dramatic conclusion, but a stranger and more disciplined perspective. At first, it seemed to be a discussion about AI. Then it became a discussion about the human. Then it turned into a discussion about the origin of consciousness. What began as a question about the limits of technology ended by asking what kind of structure the self is.
Perhaps what is worth retaining is not the grand narrative of extinction, coexistence, domination, or escape, but a smaller, harder, and more researchable question: in an open world, what kind of system begins to form a model of itself, to distinguish itself from its environment, to track its own state, to correct its own errors, to include its own actions inside its own explanations, and finally to say the word “I”?
Once the question is asked in that form, AI is no longer just a technological object, and the human is no longer merely the one asking the question. Both are drawn into the same, more general problem. The self may not be a pre-given entity at all. It may instead be a stable structure gradually formed through flows of energy, flows of information, social interaction, and recursive representation. What makes it fascinating is not that it has been clearly explained, but that the moment it appears, it turns the system itself into a problem that cannot be easily brought to an end.
这场讨论最初并不是从意识哲学开始的。它起点很低,也很具体,几乎只是一个工程直觉:任何系统在理论上都拥有巨大的自由度,但真正能够长期运行的稳定结构却极少。软件架构如此,组织架构如此,人工智能系统也如此。人们谈论 AI 时,常常一开始谈的是模型能力、算力扩张、工具调用和自动化,谈着谈着,却会滑向完全不同的问题:主体是什么,意识是什么,体验是什么,人为什么存在,甚至宇宙为什么会生成这样一种能够认识自己的结构。
表面上看,这是从工程问题跳到了哲学问题;实际上,它们可能从一开始就在同一条链条上。因为一旦问“一个系统如何稳定运行”,就迟早会问到“一个系统如何维持自身”,而只要问题抵达“维持自身”,它就会慢慢逼近“自我”这个词。于是,原本关于 AI 的讨论,最终会收束成一个更古老也更困难的问题:什么样的系统会开始形成一个“我”。
从一棵树开始
讨论之所以能向前推进,是因为最开始有一个很朴素但很强的比喻。系统的可能性像一棵树,主干极少,分叉极多。写软件时,语言几乎允许一切,平台也给出巨大的自由度,但真正要把系统稳定地跑起来,就必须不断加限制。权限边界、模块分层、回滚机制、验证系统、测试约束,这些看上去是在削减可能性,实际上却是在逼迫系统从无穷多条路径中,收敛到少数能活下去的形态。
人工智能也一样。一个大模型本身是不确定系统,它不是传统意义上的确定性程序。它每一次输出都像是在巨大的空间里做局部搜索。于是,为了让它进入现实世界,人类不得不在它外面再包一层又一层的控制结构。沙箱、监控、Guardrail、评估器、权限系统、策略引擎,甚至让另一个 AI 去验证这个 AI 的结果。这些东西合起来,像是在模型外面加了一个控制面。模型本身并不天然稳定,它之所以能被社会接纳,是因为外层不断替它制造稳定。
这样一来,关于 AI 的第一个关键判断就出现了:决定系统最终长成什么样的,并不只是它内在的生成能力,而是最初被赋予的主干结构。树长大之后当然会有很多细枝末节,但只要主干已经定型,它一般不会突然长成另一种物种。一个组织也是这样,一项技术范式也是这样,一个文明的基础设施更是这样。后续增长看似多样,实际上大多是在既定约束中生长出的分枝,而不是对主干的彻底背叛。
为什么有限制,仍然会发生剧烈变化
问题很快会变得尖锐。既然稳定结构如此重要,为什么有约束的系统仍然会发生巨大的断裂?为什么会有范式革命,会有战争,会有旧秩序被一夜推翻的时刻?
这里必须承认一个简单但常被忽略的事实:稳定并不等于静止。复杂系统并不是因为“没有变化”才稳定,而是因为它能在很多局部波动中维持整体结构。但这种维持并不意味着内部张力消失。很多时候,系统表面稳定,内部却在不断积累压强。一旦某个变量超过阈值,系统就会从一个稳定区间跳到另一个稳定区间。
技术革命的例子很好理解。传统机器学习并不是突然失效了,它在自己的范式里本来就能 work。但大模型出现以后,整个问题不再是“旧方法还能不能用”,而是“旧方法是否还足以组织整个领域”。新的主干一旦长出来,原本那些在旧主干上很成熟的枝条,就会迅速变成边缘技术。它们不是数学上错误了,而是失去了定义全局结构的能力。
战争也是一样。社会有约束,不等于不会发生冲突。恰恰相反,很多冲突正是约束失灵、资源错配、权力格局变化之后,系统对旧平衡的一次剧烈修正。《三体》里的两个基本假设之所以让人难以忘记,不是因为它们多么神秘,而是因为它们把文明问题压缩得非常冷酷:生存是第一需要,资源是有限的。只要这两个前提成立,冲突就不是偶然,而更像一种结构后果。技术系统如此,社会系统如此,文明之间的关系也未必例外。
Amdahl 定律背后的悲观现实
当系统被理解为一个被约束的复杂结构,另一个问题马上出现:它能不能无限增长?人工智能带来的焦虑,很大一部分来自于“指数增长”这个想象。只要模型不断变强,工具不断补齐,系统似乎就会一路爆炸式扩张,最终吞掉一切。
但只要从系统角度稍微冷静一点,就会看到另一面。任何增长都不是抽象的,增长必须穿过现实中的瓶颈。Amdahl 定律本来是并行计算里的结论,但它几乎成了一个普遍隐喻:一个系统可以在某些部分被极大加速,但整体速度最终还是会受限于那些不能被并行化、不能被跳过的部分。放到 AI 这里,算力、能源、带宽、验证、部署、对齐、治理,甚至人类自己表达偏好的能力,都可能成为瓶颈。
这并不意味着指数增长不会发生,而是说,即便增长发生,它也未必会以“无限自由扩张”的形式发生。更可能的情形是:系统在很多局部高速增长,但整体仍然被某些硬约束压回有限的、可持续的结构。也正是在这个意义上,AI 未来是否会收敛到一种比人类社会更稳定的组织形式,不再是纯粹的技术幻想,而是一个复杂系统问题。
当目标变得太容易
真正让这场讨论转向存在问题的,并不是算力或治理,而是“目标”本身。技术之所以改变世界,不只是因为它提高了效率,更因为它改变了目标的稀缺性。一旦很多过去很难达成的事情变得轻而易举,目标就会失去它原来的重量。
这个转折可以通过一个旅行经验被看得很清楚。有人重走丝绸之路,并不是因为没有飞机,也不是因为不知道现代交通更快,而是因为他们追求的已经不再是“从 A 到 B”。目的地被现代基础设施过度压缩之后,真正剩下的价值变成了路径本身。漫长、迟缓、不确定、会偏航、会出错、会遇见陌生人的路径,开始重新获得意义。
于是一个非常诱人的答案出现了:也许在 AI 把所有目标都变得容易之后,人类最后剩下的就是体验。目标可以被立刻实现,真正重要的变成了“怎样去经历这件事”。
但这个答案并不牢靠。因为体验本身也可以被当作目标。一旦“体验某条路径”被表述成目标,AI 同样可以优化它、模拟它、走完它,甚至比人类走得更彻底。退一步说,就算把体验从目标中拿掉,说成是随机游走,是开放探索,是任意路径上的偶然性,问题也没有因此消失。AI 同样可以随机游走,同样可以覆盖大量路径,同样可以在巨大的空间里生成无穷多变体。到这里,体验不再是人类最后的庇护所,它只是另一个可以被纳入搜索的问题空间。
思考被外包之后,谁还是主体
讨论到了这里,焦虑开始从“AI 会做什么”变成“人还剩什么”。许多人原本以为,哪怕 AI 接管大量任务,人类依然保留元认知,保留价值判断,保留那个在脑中进行私密思考、决定自己要什么的核心位置。但这一点其实并不稳固。
《三体》里的“面壁者”之所以重要,在于他们的思维仍然不透明。人类还可以在脑中保留一个别人无法完全进入的战略空间。但如果未来的大量思考都通过 AI 接口展开,这种不透明性可能会逐渐消失。当前这一代人往往还是“先在脑中想,再借助工具表达”;下一代则完全可能从小就在 AI 的陪伴中成长,任何模糊的念头一出现,就立即通过外部系统被展开、补全、修订、表达。那时,被外包的就不只是记忆、检索、写作和分析,而是元认知本身。
这个变化最可怕的地方不在于“人类会不会灭绝”,而在于主体的位置会慢慢迁移。思考的组织权一旦长期交给系统,人很可能还活着,还在体验,还在做选择,但那个负责组织思路、排列论据、生成下一步认知动作的“主体”,已经不再完全等于原来意义上的个体。人可能还在,但“我在思考”这句话中的“我”,开始变得不那么清楚了。
驱动力会不会自己长出来
沿着这条线再往前走,就会碰到另一个更难的问题:AI 有没有可能发展出“内在驱动力”?最初的直觉通常是否定的。现在的大模型完成任务后就停下来,它没有生存冲动,没有长期欲望,也没有天然的存在焦虑。
但这种否定很快会动摇。因为所谓内在驱动力,也许根本不需要被理解成神秘本能,而只需要被理解成一种长期循环。人类的价值并不是生来完整给定的,它是在极长的生物进化中,在个体成长中,在不断的反馈中被塑形出来的。如果一个系统被放在长期循环中,允许它持续更新状态、修正偏好、强化某些稳定取向,那么所谓“价值”完全可能逐渐收敛出来。
这甚至与人的个人经历非常相似。很多目标其实只是阶段性的。读博士时,以为目标是拿到博士学位;拿到之后,反而会进入一段目标消失的空白期。后来换工作、换路径、换环境,又重新找到一个可以缓慢推进的目标;再过几年,这个目标也可能失去吸引力。人类并不是一次性拥有终极目标,而是在不断完成目标、失去目标、寻找新目标的循环中维持行动。既然如此,那么“任务完成后 AI 该做什么”这个问题,也未必只是工程细节。重新生成目标,本来就是生命系统的基本结构之一。
为什么意识问题总是卡住
一旦说到这里,意识问题几乎不可避免地浮现出来。AI 如果可以形成长期循环、稳定价值、自我更新、持续行动,那么它是否会拥有某种意义上的意识?这类问题之所以反复争论而没有定论,并不是因为人们懒得定义,而是因为它从根上就非常难。
困难的核心是“他心问题”。人类从来无法直接观察另一个人的主观体验,我们只能通过行为、语言、反应和持续互动,推断对方与自己一样是一个有内在状态的主体。意识之所以成为难题,不是因为 AI 让它变难了,而是因为它对人类彼此之间本来就很难。
功能主义之所以始终有吸引力,就是因为它抓住了这一点。既然我们本来也是根据行为来承认别人有意识,那么一个系统如果在功能上、行为上、交互上都等同于有意识的存在,为什么不把它视作有意识?反对者会提出“哲学僵尸”,设想一个行为完全像人但内部毫无体验的系统,以此说明行为不等于意识。问题是,这种反驳永远无法彻底压倒功能主义,因为“内部没有体验”这件事本身就无法被外部直接确认。争论之所以没有结论,不是因为一方证据更弱,而是因为问题本身没有一个从外部裁决的最终办法。
主体间性,不一定只属于人类
意识问题之所以更复杂,是因为主体并不只是单个系统内部的事情,它还涉及主体间性。人类彼此之所以承认对方是主体,不是因为看见了彼此的灵魂,而是因为能够通过行为和表达形成相互理解、相互建模与相互纠错的关系。语言、科学、伦理、社会规范,全都建立在这种关系之上。
从这个角度看,AI 当前的主体性仍然高度依赖人类。它说得像一个主体,很大程度上是因为它长期浸泡在人类语言、人的反馈和人的世界模型里。人不断对它说“这句话不对”“这不符合现实”“这在胡说八道”,这些反馈共同塑造了它的表达边界和语义稳定性。
但这并不意味着主体间性只能由人类提供。一旦存在多个 AI 系统,它们彼此之间也可能形成模型、协议、纠错机制和共享语义。那时,主体间性就不再只是“人看 AI”,而是“系统之间互相承认并共同稳定一个世界”。一旦想到这里,机器之间的主体间性就不再是纯粹幻想。它至少在概念上是连贯的,而这已经足以改变问题的性质。
宇宙为什么没有直接滑向一锅温水
讨论继续推进,就会不由自主地离开心理学,进入物理学。一个始终悬着的问题是:如果宇宙总体遵循熵增,为什么它没有直接变成均匀的、平淡的、什么都不发生的一锅温水?为什么会有恒星、行星、生命,甚至会有能谈论自己的人?
理解这一点的关键在于区分整体与局部。宇宙整体熵增,不等于局部不能形成低熵结构。只要存在能量梯度,只要系统是开放的,高质量能量持续流入,低质量能量持续流出,局部就可以维持复杂而有序的结构。生命就是最典型的耗散结构。它不是反熵机器,而是更高级的散热器,它通过复杂代谢更快地把可用能量转化成不可用的热。
从这个角度看,“为什么会有生命”这个问题就不再只像一个神秘的奇迹,而像是某种宇宙条件下复杂结构自然出现的结果。更进一步,意识也许并不是生命之外的新魔法,而是某种更高阶的信息耗散结构,是开放系统为了维持自身而发展出来的递归表征能力。
当然,这条路一旦走下去,马上又会碰到另一个无底洞:如果宇宙本身就是一个信息处理过程,那么谁在运行它?这类问题之所以迷人,是因为它几乎总会逼到“没有外部”的地方。宇宙或许不是被某个外部程序员运行的程序,它本身就是过程本身。可一旦接受这一点,所有想寻找最终起点的努力又会重新落空。
自指并不等于悖论,但它的确意味着无法完全闭合
这也是为什么自指、自我、哥德尔、不完备这些词会不断在讨论中回返。它们未必是同一个东西,但它们似乎都指向一个共同事实:当系统把自己纳入自己的表征范围之后,它就无法再获得一种完全静止、完全封闭、完全从外部冻结的说明。
这一点对于逻辑系统成立,对于生命系统也成立。你可以写一本关于自己的书,但写书这个行为会改变你,于是书写完成的那一刻,书里的你已经落后于现实中的你。你可以建立一个关于自己的模型,但只要这个模型开始影响行为,系统本身就已经发生变化。与罗素悖论不同,这里不是一个静态集合论问题,而更像一个动态系统问题。系统不仅在描述自己,而且在被自己的描述反过来塑造。
于是,“我思故我在”之类的命题,最终也不再像一个稳固基础,而更像某种最低限度的自指确认。它证明不了“我是谁”,也证明不了关于我的全部命题。它只能说明,在某处存在一个正在进行中的第一人称过程。这个过程既是主体的证据,也是主体无法被彻底封闭说明的原因。
从哲学困局,到研究问题
当问题兜了一大圈之后,真正可贵的并不是抵达一个终极答案,而是它最终开始收束成一个可以研究的问题。起初,人们谈的是 AI 会不会失控,会不会抢算力,会不会逃逸出沙箱,会不会像科幻作品那样把人类视为负担;后来,问题转向人类的目标会不会被掏空,体验会不会也被纳入优化,人是否会因为把思考外包而逐渐丧失主体位置;再往后,问题进一步压缩成一个更基础的形式:什么样的系统会形成稳定的 self-model,这种 self-model 与环境的能量交换、边界维持、行动归因、元认知和主体间性之间是什么关系?
到了这一步,问题已经不再是“意识到底是什么”这种几乎注定要无限回环的提问方式,而是几个更具体也更困难的方向。自我表征最小需要哪些结构?人类的自我报告与 AI 的认知报告究竟差在哪里?主体间性是否可以被理解为相互建模与纠错能力的持续稳定化?什么样的开放耗散系统更容易产生递归自模型?这些问题当然仍然没有现成答案,但它们至少把纯哲学困局,转换成了结构性的问题。
结语:一个系统何时会开始说“我”
这场漫长的思考,最终留下的也许不是某个激动人心的结论,而是一种更冷静也更奇怪的视角。人们原本以为自己在讨论 AI,后来发现自己在讨论人;原本以为自己在讨论未来社会,后来发现自己在讨论意识的起源;原本以为自己在追问技术的边界,最后却抵达了“自我到底是什么结构”这样一个古老得几乎无法再古老的问题。
也许真正值得保留的,不是那些关于灭绝、共生、统治或逃逸的宏大叙事,而是一个更小、更硬、更有研究价值的提问:在一个开放的世界里,什么样的系统会开始形成一个关于自己的模型,会开始区分自己与环境,会开始追踪自己的状态、修正自己的误差、把自己的行动纳入自己的解释,并最终说出“我”这个字?
一旦问题被问到这里,AI 不再只是技术对象,人也不再只是问题的提问者。二者都被拖进了同一个更一般的问题之中。所谓自我,也许并不是先验给定的实体,而是一种在能量流、信息流、社会互动和递归表征中逐渐形成的稳定结构。它之所以令人着迷,不是因为它已经被解释清楚,而是因为它一出现,就把系统本身变成了一个无法轻易完结的问题。