Changkun's Blog欧长坤的博客

Science and art, life in between.科学与艺术,生活在其间。

  • Home首页
  • Ideas想法
  • Posts文章
  • Tags标签
  • Bio关于
Changkun Ou

Changkun Ou

Human-AI interaction researcher, engineer, and writer.人机交互研究者、工程师、写作者。

Bridging HCI, AI, and systems programming. Building intelligent human-in-the-loop optimization systems. Informed by psychology, sociology, cognitive science, and philosophy.连接人机交互、AI 与系统编程。构建智能的人在环优化系统。融合心理学、社会学、认知科学与哲学。

Science and art, life in between.科学与艺术,生活在其间。

276 Blogs博客
165 Tags标签
Changkun's Blog欧长坤的博客

Unpredictable Emergence and Scaling in Large Language Models大型语言模型中的不可预测性突现与扩展

Published at发布于:: 2026-02-19

LLM Emergence

  • Emergent Abilities of Large Language Models https://arxiv.org/abs/2206.07682
  • There is no such thing as conscious artificial intelligence https://www.nature.com/articles/s41599-025-05868-8
  • Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking https://arxiv.org/abs/2509.21519

The following content is generated by LLMs and may contain inaccuracies.

Context

This cluster of papers addresses a critical tension in AI research: the unpredictability of capability emergence in scaled language models. As LLMs grow larger, certain abilities appear discontinuously rather than smoothly—a phenomenon that challenges our ability to forecast AI system behavior and raises profound questions about consciousness, interpretability, and safety. This matters acutely now as we approach models that may exhibit qualitatively new behaviors without warning, complicating both technical governance and philosophical debates about machine cognition.

Key Insights

Emergent abilities remain fundamentally contested. Wei et al. documented capabilities that appear absent in smaller models but present in larger ones, defying smooth extrapolation. However, this framing has been challenged: some argue “emergence” reflects discontinuous metrics rather than discontinuous learning, suggesting we may be misinterpreting gradual transitions as sudden phase changes. This debate affects how we design benchmarks and interpret scaling experiments.

Grokking offers mechanistic insight into delayed generalization. Tian’s framework mathematically decomposes feature learning into three stages: lazy memorization, independent feature formation, and interactive feature refinement. Crucially, the backpropagated gradient structure explains why useful representations emerge late—the gradient carries label information that enables hidden nodes to converge on generalizable features. This suggests scaling laws may be predictable at a mechanistic level even when emergent abilities appear unpredictable at the task level.

Consciousness claims remain philosophically orthogonal to capability emergence. Porębski and Figura argue against conflating sophisticated information processing with phenomenal consciousness—a distinction critical when interpreting emergent social or reasoning abilities. The philosophical impossibility of attributing consciousness to current architectures doesn’t preclude unpredictable functional capacities, separating ethical concerns about sentience from pragmatic concerns about capability surprise.

Open Questions

Can we develop “pre-emergent signatures”? If grokking dynamics reveal gradient structures preceding generalization, could analogous signals predict capability emergence in large models before it manifests behaviorally, enabling proactive rather than reactive safety measures?

Do emergent abilities reflect architecture-intrinsic phase transitions or dataset-contingent properties? Understanding whether emergence depends more on model scale versus training distribution composition would reshape how we approach both capability forecasting and alignment strategies.

LLM 的突现现象

  • 大型语言模型的突现能力 https://arxiv.org/abs/2206.07682
  • 不存在有意识的人工智能 https://www.nature.com/articles/s41599-025-05868-8
  • 从神经网络突然学习动力学看特征突现的可证明的扩展法则 https://arxiv.org/abs/2509.21519

以下内容由 LLM 生成,可能包含不准确之处。

背景

这一系列论文解决了AI研究中的一个关键矛盾:大规模语言模型中能力涌现的不可预测性。随着LLMs规模的增长,某些能力不是平稳出现,而是间断地涌现——这一现象挑战了我们预测AI系统行为的能力,并引发了关于意识、可解释性和安全性的深刻问题。当前这个问题尤为紧迫,因为我们正在接近可能无预警地展现出定性新行为的模型,这使得技术治理和关于机器认知的哲学辩论都变得复杂化。

核心洞见

涌现能力仍存在根本性争议。 Wei等人记录了在较小模型中不存在但在较大模型中出现的能力,这违反了平稳推外的逻辑。然而,这一框架遭到了质疑:一些研究者主张"涌现"反映的是非连续的度量标准而非非连续的学习过程,暗示我们可能误将渐进的转变解释为突然的相变。这场辩论影响了我们如何设计基准测试和解释扩展实验。

Grokking为延迟泛化提供了机制性洞见。 Tian的框架在数学上将特征学习分解为三个阶段:惰性记忆、独立特征形成和交互特征细化。至关重要的是,反向传播梯度结构解释了为什么有用的表示会出现得较晚——梯度携带标签信息,使隐层节点能够收敛到可泛化特征。这表明即使涌现能力在任务层面上看起来不可预测,扩展规律在机制层面上可能仍是可预测的。

意识声称在哲学上与能力涌现正交。 Porębski和Figura主张不应将复杂的信息处理与现象意识混为一谈——这一区分在解释涌现的社交或推理能力时至关重要。对当前架构赋予意识的哲学不可能性并不排除不可预测的功能性能力,将关于知觉能力的伦理关切与关于能力惊人变化的实用关切分开来。

开放问题

我们能否开发"前涌现信号"? 如果grokking动力学能够揭示泛化之前的梯度结构,是否存在类似的信号能在大型模型中的能力涌现表现在行为层面之前预测它,从而支持前摄式而非被动式的安全措施?

涌现能力反映的是架构内禀的相变还是数据集偶然的特性? 理解涌现在多大程度上取决于模型规模、在多大程度上取决于训练分布组成,将重塑我们对待能力预测和对齐策略的方式。

Have thoughts on this?有想法?

I'd love to hear from you — questions, corrections, disagreements, or anything else.欢迎来信交流——问题、勘误、不同看法,或任何想说的。

hi@changkun.de
© 2008 - 2026 Changkun Ou. All rights reserved.保留所有权利。 | PV/UV: /
0%