Changkun's Blog欧长坤的博客

Science and art, life in between.科学与艺术,生活在其间。

  • Home首页
  • Ideas想法
  • Posts文章
  • Tags标签
  • Bio关于
Changkun Ou

Changkun Ou

Human-AI interaction researcher, engineer, and writer.人机交互研究者、工程师、写作者。

Bridging HCI, AI, and systems programming. Building intelligent human-in-the-loop optimization systems. Informed by psychology, sociology, cognitive science, and philosophy.连接人机交互、AI 与系统编程。构建智能的人在环优化系统。融合心理学、社会学、认知科学与哲学。

Science and art, life in between.科学与艺术,生活在其间。

276 Blogs博客
165 Tags标签
Changkun's Blog欧长坤的博客
idea想法 2026-02-19 04:48:06

Understanding the Connection Between Moral Judgment and Action理解道德判断与行动之间的联系

In our everyday lives, we confront numerous moral issues. Once we have deliberated and formed judgments about what is right or wrong, good or bad, these judgments tend to exert a strong influence on us. Although we do not always behave as we think we ought, our moral judgments typically motivate us, at least to some degree, to act in accordance with them. When philosophers discuss moral motivation, they seek to understand this basic phenomenon. Moral motivation is an instance of a more general phenomenon—what we might call normative motivation—since our other normative judgments also typically have some motivating force. When we judge that something is good for us, that we have a reason to act in a particular way, or that a specific course of action is rational, we tend to be moved to act accordingly. Many philosophers have regarded the motivating force of normative judgments as the key feature that distinguishes them as normative, setting them apart from the many other types of judgments we make. In contrast to our normative judgments, our mathematical and empirical judgments appear to have no intrinsic connection to motivation and action. The belief that an antibiotic will cure a specific infection may move an individual to take it if she also believes she has the infection and either desires to be cured or judges that she ought to treat it for her own good. However, an empirical belief like this one appears to carry no particular motivational impact on its own; a person can judge that an antibiotic will most effectively cure a specific infection without being moved one way or another.

Although motivating force may distinguish normative judgments, normative motivation seems most significant in the case of narrowly moral judgments. Moral motivation has, in any case, received far greater attention than motivation connected with other normative judgments. Morality is widely believed to conflict frequently and sometimes severely with what an agent most values or prefers to do. Perhaps because of the apparent opposition between self-interest and morality, moral motivation has seemed especially puzzling. How is it that we are so reliably moved by our moral judgments? What is the precise nature of the connection between moral judgment and motivation? Of course, the less puzzling and more mundane moral motivation becomes, the more puzzling failures of moral motivation appear. To explain moral motivation, we must understand not only how moral judgments so regularly succeed in motivating, but also how they can fail to motivate, sometimes quite dramatically. Motivational failure occurs not only among the deranged, dejected, and confused, but also, apparently, among the fully sound and self-possessed. How should we understand the “amoralist”—the apparently rational, strong-willed individual who seemingly makes moral judgments while remaining utterly indifferent?

https://plato.stanford.edu/entries/moral-motivation

The following content is generated by LLMs and may contain inaccuracies.

I now have robust search results on the internalism/externalism debate, the judgment-action gap in empirical psychology, and the amoralist problem. Let me search for one more specific angle—recent developments on dual-process theories in moral psychology and potentially neuroscience contributions.Perfect. Now I have comprehensive search results covering:

  1. Internalism vs externalism debate in metaethics
  2. The judgment-action gap in empirical moral psychology
  3. The amoralist problem
  4. Dual-process theories (Greene’s model) and their critiques

Let me now synthesize this into a structured deep dive that augments the original note.


Context

The puzzle of moral motivation sits at the intersection of metaethics, moral psychology, and action theory. Its urgency stems from a foundational tension: if moral judgments reliably shift motivation to match changes in judgment, this suggests an internal, necessary connection between judgment and motivation, yet we observe systematic failures—from everyday akrasia to the figure of the “amoralist” who appears to make genuine moral judgments while remaining unmoved. Motivational internalism holds there is an internal, necessary connection between moral convictions and motivation, while externalism denies this necessity. This debate ramifies into questions about moral realism, cognitivism vs. noncognitivism, and whether moral language refers to objective features of the world or expresses motivational states. The problem matters now because recent work in experimental psychology has been brought to bear on metaethical questions, with implications for the plausibility of internalism, externalism, and various accounts of moral motivation.

Key Insights

  1. The judgment-action gap is empirically robust but theoretically contested. Many students cheat even when they believe it is wrong, and motivational factors like perceived moral obligation and self-regulatory beliefs explain additional variance beyond attitudes in predicting cheating behavior. This empirical gap has prompted multi-component models: Rest’s four-component model, formulated in 1983 and largely unquestioned since, proposes that moral action requires not only judgment but also moral sensitivity, motivation, and character. Yet meta-analyses show that moral identity and moral emotions overall fare only slightly better as predictors of moral action than moral judgment itself. Recent integrative proposals invoke phronesis (practical wisdom) to bridge judgment, motivation, and action, though critics note this risks collapsing distinct problems into one unwieldy construct.

  2. Dual-process theories offer mechanistic purchase but face normative and empirical challenges. Joshua Greene’s influential dual-process theory, grounded in fMRI studies cited over 2000 times, proposes that automatic-emotional processes drive deontological judgments while controlled-reasoning processes support utilitarian judgments. Greene argues we should rely less on automatic emotional responses for “unfamiliar problems” like climate change or global poverty, where we lack adequate evolutionary or cultural experience. However, critics point out that attributing normative correctness to deliberate rather than intuitive processes constitutes a “normative fallacy”—an unjustified generalization, and empirical evidence for the exact role of emotion in deontological judgment remains contested and unclear. The broader insight: descriptive theories of cognitive architecture do not straightforwardly yield normative recommendations about which processes to trust.

  3. The amoralist poses a conceptual rather than merely empirical challenge. Internalists insist the amoralist is a conceptual impossibility, typically arguing that no rational agent could competently employ moral concepts while remaining wholly unmoved. Yet externalists maintain that if we can conceive of amoralists, they are not conceptually impossible, and not all motivational failures can be explained away as irrationality or conceptual incompetence. Strikingly, recent experimental research reveals a “factivity effect”: people’s intuitions lean toward externalism when an amoralist is described as knowing X is wrong, but toward internalism when described as believing X is wrong. This suggests folk moral psychology may be more nuanced—or incoherent—than philosophers have assumed, and that the debate may hinge on implicit assumptions about the relationship between knowledge, belief, and motivation that deserve empirical scrutiny.

Open Questions

  • Can we distinguish empirically between genuine amoralists and agents suffering from subtle forms of akrasia, depression, or moral disengagement? The literature vacillates between treating motivational failure as a property of judgment (internalism/externalism) versus a property of agency (self-regulation, character). Disentangling these requires longitudinal studies tracking the stability of moral judgment alongside motivational dispositions across contexts.

  • What normative weight should evolutionary or cultural unfamiliarity carry in moral epistemology? Greene’s recommendation to distrust intuitions about “unfamiliar” moral problems assumes that automatic processes are calibrated to ancestral or culturally local environments. But if explicit reasoning is itself shaped by historically contingent ideologies, is there any Archimedean point from which to adjudicate between System 1 and System 2 outputs—or must we abandon the hope of a general metaethical verdict on which processes are epistemically privileged?

在我们的日常生活中,我们面临许多道德问题。一旦我们经过思考并形成了关于什么是对或错、好或坏的判断,这些判断往往会对我们产生强大的影响。尽管我们并不总是按照我们认为应该的方式行动,但我们的道德判断通常至少在某种程度上激励我们按照这些判断行动。当哲学家讨论道德动机时,他们试图理解这一基本现象。道德动机是一种更广泛现象的例子——我们可以称之为规范性动机——因为我们的其他规范性判断通常也具有某种激励力。当我们判断某事对我们有益、我们有理由以某种特定方式行动,或某个特定方案是理性的时候,我们倾向于被促使采取行动。许多哲学家将规范性判断的激励力视为将其区别为规范性的关键特征,使其与我们所做的许多其他类型的判断区开。相比之下,我们的数学和经验判断似乎与动机和行动没有内在联系。认为抗生素会治愈特定感染的信念可能会促使个人服用抗生素,如果她也相信自己患有感染,并且要么希望被治愈,要么认为她应该为了自身健康而治疗这种感染。然而,像这样的经验信念本身似乎没有特别的激励效果;一个人可以判断某种抗生素最有效地治疗特定感染,而不会在任何方面被动机所影响。

尽管激励力可能区别规范性判断,但规范性动机在狭义道德判断的情况下似乎最为重要。无论如何,道德动机比与其他规范性判断相关的动机获得了远多的关注。道德普遍被认为与行动者最看重或最想做的事情频繁甚至严重冲突。或许正因为自我利益与道德之间的明显对立,道德动机显得特别令人困惑。我们如何能够被我们的道德判断如此可靠地驱动?道德判断与动机之间的确切性质是什么?当然,道德动机显得越不令人困惑、越平凡,道德动机的失败就越令人困惑。为了解释道德动机,我们必须理解道德判断如何如此经常地成功激励,以及它们如何能够失败以激励,有时甚至相当戏剧性地失败。动机失败不仅出现在精神失常、沮丧和困惑的人身上,似乎也出现在心智完全正常和自主自制的人身上。我们应该如何理解「非道德主义者」——这个看起来理性、意志坚强的人,他似乎做出道德判断,但却完全无动于衷?

https://plato.stanford.edu/entries/moral-motivation

以下内容由 LLM 生成,可能包含不准确之处。

我现在已经掌握了关于内部主义/外部主义辩论、实证心理学中的判断-行动鸿沟以及非道德者问题的可靠搜索结果。让我搜索一个更具体的角度——道德心理学中的双过程理论最新发展,以及潜在的神经科学贡献。

完美。现在我拥有涵盖以下内容的全面搜索结果:

  1. 元伦理学中的内部主义vs外部主义辩论
  2. 实证道德心理学中的判断-行动鸿沟
  3. 非道德者问题
  4. 双过程理论(格林的模型)及其批评

现在让我将其综合为一个结构化的深度探讨,用以增强原始笔记。


背景

道德动机的困境位于元伦理学、道德心理学和行动理论的交汇处。其紧迫性源于一个基础性的张力:如果道德判断能够可靠地改变动机以匹配判断的变化,这暗示判断与动机之间存在内部的、必然的联系,然而我们观察到系统性的失败——从日常的理智软弱到"非道德者"这一人物形象,他似乎做出真诚的道德判断却保持不为所动。动机内部主义主张判断与动机之间存在内部的、必然的联系,而外部主义否定这种必然性。这场辩论涉及到关于道德现实主义、认知主义vs非认知主义的问题,以及道德语言是否指涉世界的客观特征或表达动机状态。这个问题之所以重要,是因为实验心理学的最新工作已被用于解决元伦理学问题,这对内部主义、外部主义以及各种道德动机说的合理性具有启示意义。

关键洞见

  1. 判断-行动鸿沟在经验上是稳健的,但理论上存在争议。 许多学生即使认为作弊是错误的,仍然会作弊,动机因素如感知到的道德义务和自我调节信念在预测作弊行为方面解释了超越态度的额外方差。这个经验性鸿沟促使人们提出多成分模型:雷斯特在1983年提出的四成分模型自此以来基本上没有被质疑,该模型主张道德行动不仅需要判断,还需要道德敏感性、动机和品格。然而,荟萃分析显示道德认同和道德情感作为道德行动预测因子的效果总体上只比道德判断本身略好。最近的综合提议引入实践智慧(phronesis)来桥接判断、动机和行动,尽管批评者指出这有将不同的问题混为一谈的危险。

  2. 双过程理论提供机制论的购买力,但面临规范性和经验性的挑战。 约书亚·格林的有影响力的双过程理论以fMRI研究为基础,被引用超过2000次,该理论主张自动-情感过程驱动义务论判断,而控制-推理过程支持后果主义判断。格林主张对于"不熟悉的问题"(如气候变化或全球贫困),我们应该较少依赖自动情感反应,因为我们缺乏充分的进化或文化经验。然而,批评者指出,将规范正确性归因于审慎而非直觉过程构成"规范谬误"——一种不合理的推广,而且情感在义务论判断中的确切作用的经验证据仍然存在争议且不明确。更广泛的洞见是:认知架构的描述性理论不能直接得出关于应该信任哪些过程的规范建议。

  3. 非道德者提出的是概念性而非仅仅经验性的挑战。 内部主义者坚持非道德者在概念上是不可能的,通常论证没有理性代理能够胜任地使用道德概念同时保持完全不为所动。然而,外部主义者主张,如果我们能够想象非道德者的存在,他们在概念上并非不可能,而且并非所有的动机失败都能被解释为不理性或概念性的无能。令人惊讶的是,最近的实验研究揭示了一个"事实性效应":当描述非道德者知道X是错误时,人们的直觉倾向于外部主义,但当描述为相信X是错误时,则倾向于内部主义。这表明民间道德心理学可能比哲学家假设的更为微妙——或更为不一致——并且该辩论可能取决于关于知识、信念和动机之间关系的隐含假设,这些假设值得进行经验审查。

悬而未决的问题

  • 我们能否在经验上区分真正的非道德者和遭受细微的理智软弱、抑郁或道德脱离的代理人? 文献在将动机失败视为判断的属性(内部主义/外部主义)和代理的属性(自我调节、品格)之间摇摆不定。区分这些需要跟踪道德判断稳定性和跨情境动机倾向的纵向研究。

  • 进化论或文化陌生性在道德认识论中应该承载多少规范权重? 格林对于不信任关于"不熟悉"道德问题的直觉的建议假设自动过程是针对祖先或文化本地环境校准的。但如果显式推理本身也受到历史偶然意识形态的塑造,是否存在任何阿基米德支点可以从中仲裁系统1和系统2的输出之间的问题——或者我们是否必须放弃希望找到一个关于哪些过程在认识论上获得特权的一般元伦理学判决?

idea想法 2026-02-19 04:03:17

Three Paradoxes of Knowledge and Belief知识信念的三个悖论

  • How can you believe in something you haven’t witnessed with your own eyes?
  • Who sets the standard for evidence?
  • If an experience cannot be replicated by a third party, is it still true?

https://www.darioamodei.com/essay/the-adolescence-of-technology

The following content is generated by LLMs and may contain inaccuracies.

Three Paradoxes of Epistemic Belief

Context

You have touched upon the core tensions of social epistemology and the epistemology of testimony. These questions inquire into the fractures between the sources of knowledge, the normativity of evidence, and the reproducibility of experience. Epistemology investigates how the sources of knowledge—perception, reasoning, and testimony—create knowledge. When you question “how to believe in things we have never witnessed,” you are challenging a long-standing philosophical assumption: that our testimony-based beliefs require evidential support, yet there is disagreement about where this evidence ultimately originates. The question of who sets the standards for evidence touches upon the “symmetry postulate” of the strong program in sociology—scientists' beliefs should be explained by social factors regardless of whether these beliefs are true or false, rational or irrational, which undermines the foundations of “objective truth.” The third paradox—the status of truth when experience cannot be reproduced by third parties—echoes the core of the epistemological paradox: conflicting but equally well-grounded answers to the same question. These puzzles compel us to correct deep errors in our understanding of knowledge, justification, rational belief, and evidence.

Although Dario Amodei’s article focuses on AI risks, it provides a relevant meta-epistemological perspective: he discusses how AI constitutions attempt to train models to form stable personalities and values, essentially encoding answers to “who determines the standards of evidence”—a process of migration from human epistemological dilemmas to machine epistemology that exposes the arbitrariness and power attributes of norms themselves.

Key Insights

  1. The Dispute Between “Inheritance” and “Generation” of Testimony

The inheritance view holds that your testimony-based beliefs are grounded in evidence derived from the speaker’s evidence (such as a friend’s perception of restaurant queues or a priori proof of mathematical theorems); however, many epistemologists disagree with this literal “inheritance of evidence.” This reveals the root of your first paradox: our beliefs in unseen things may not be based on “our own” evidence, but rather borrowed from others' perceptual authority. Yet, as Reid pointed out, there is a fundamental difference between the analogy of testimony and perception: when trusting testimony, we rely on the speaker’s authority—a form of social, power-dependent reliance rather than a purely cognitive act. Anti-reductionists argue that the speaker’s very act of testifying confers justification upon the hearer’s belief; reductionists, by contrast, demand that the hearer must possess independent positive reasons to accept testimony. This debate remains unresolved, reflecting a fundamental tension between individual autonomy and social dependence in knowledge acquisition.

  1. “Blind Spots” and Incommensurable Standards of Evidence

If someone believes a proposition that constitutes an “epistemic blind spot” for the reader, then even if both are ideal reasoners, they cannot provide each other with sufficient grounds to share this belief—ideal reasoners may disagree precisely due to their different blind spots. This directly addresses your second and third paradoxes: standards of evidence do not exist transcendentally, but are embedded in the epistemic agent’s positionality. Social identity, role, or position influences individuals' epistemic lives—including the experiential knowledge they can access, their ability to encounter evidence, and the credibility they are granted as information providers. Feminist epistemology and research on epistemic injustice have systematically demonstrated that marginalized groups' experiential knowledge is often devalued by the exclusion of mainstream evidence standards. Your third paradox—whether truth still counts as truth when experience cannot be reproduced by third parties—is no longer a technical problem within this framework, but a problem of power: who has the authority to define the standards of “reproducibility” itself?

Open Questions

  1. If AI becomes the primary intermediary of knowledge, how will the “trust” structure of testimony epistemology be reshaped? When billions rely on Claude or GPT for knowledge while model training remains opaque to them, how will “the speaker’s authority” be verified? Will this exacerbate epistemic inequality, or create new pathways toward epistemic democratization?

  2. In the post-truth era, is “the non-reproducibility of experience” becoming a new epistemic weapon? When purveyors of misinformation exploit the unfalsifiability of personal experience (“I just felt it” or “this was my lived experience”), how can the tension between traditional epistemological requirements for evidence (public verifiability) and phenomenological authenticity (lived experience) be reconciled? Do we need new epistemic categories?

  • 你如何相信你没有亲眼见过的东西?
  • 证据的标准是由谁定的?
  • 当体验无法被第三方复现时,真相还算真相吗?

https://www.darioamodei.com/essay/the-adolescence-of-technology

以下内容由 LLM 生成,可能包含不准确之处。

知识信念的三个悖论

Context

你触及的是社会认识论(social epistemology)与证词认识论(epistemology of testimony)的核心张力。这些问题追问知识的来源、证据的规范性以及体验的可复现性之间的断裂。认识论研究知识的来源——感知、推理和证词——如何创造知识。当你质疑"如何相信未曾亲眼见过的东西"时,你正在挑战长期以来的哲学假设:即我们基于证词的信念需要证据支撑,但对于这些证据究竟从何而来存在分歧。证据标准谁来设定的问题,触及强纲领社会学的"对称性论题"——科学家的信念应由社会因素解释,无论这些信念真假、理性与否,这削弱了"客观真理"的根基。第三个悖论——体验无法第三方复现时真相的地位——呼应了认识论悖论的核心:对同一问题存在冲突但都有充分凭据的答案,这些谜题驱使我们纠正关于知识、证成、理性信念和证据的深层错误。

Dario Amodei的文章虽聚焦AI风险,但提供了相关的元认识论视角:他讨论AI宪法如何试图训练模型形成稳定的人格与价值观,本质上是在编码"证据标准由谁定"的答案——这是一个从人类认识论困境向机器认识论迁移的过程,暴露出规范本身的任意性与权力属性。

Key Insights

  1. 证词的"继承"与"生成"之争
    继承观认为,你的证词信念基于的证据来自说话者的证据(如朋友对餐厅排队的感知或数学定理的先验证明);但许多认识论学者不同意这种证据"字面继承"。这揭示了你第一个悖论的根源:我们对未见之物的信念可能并非基于"我们自己的"证据,而是借用他人的感知权威。然而,如Reid所指出,证词与感知的类比存在根本差异:相信证词时,我们依赖的是说话者的权威——这是一种社会性、权力性的依赖,而非纯粹的认知行为。反还原主义者认为,说话者的证词行为本身即赋予听者信念以证成;还原主义者则要求听者必须拥有接受证词的独立积极理由。这争论至今未决,折射出知识获取中个体自主性与社会依赖性的根本张力。

  2. “盲点"与不可通约的证据标准
    如果某人相信一个对读者而言是"认识盲点”(blindspot)的命题,那么即使他们都是理想推理者,他也无法为读者提供充分理由来分享这一信念——理想推理者之间也可能因盲点不同而存在分歧。这直接回应了你的第二和第三个悖论:证据标准并非超验存在,而是嵌入在认知者的位置性(positionality)中。社会身份、角色或位置影响个体的认识生活——包括他们能获取的经验知识、接触证据的能力,以及作为信息提供者被赋予的可信度。女性主义认识论与认识不公正(epistemic injustice)研究已系统地展示,边缘群体的体验性知识常因主流证据标准的排斥而被贬低。你的第三个悖论——体验无法第三方复现时真相还算真相吗——在此框架下不再是技术问题,而是权力问题:谁有权定义"可复现性"本身的标准?

Open Questions

  1. 如果AI成为主要的知识中介,证词认识论的"信任"结构将如何重塑? 当数十亿人依赖Claude或GPT获取知识,而模型训练对他们不透明时,“说话者的权威"将如何被验证?这是否会加剧认识不平等,还是创造新的认识民主化路径?

  2. 在后真相时代,“体验的不可复现性"是否正在成为新的认识武器? 当虚假信息制造者利用个人体验的不可证伪性(“我就是感觉到了"“这是我的亲身经历”),传统认识论要求的证据标准(公共可验证性)与现象学真实(lived experience)之间的张力将如何调和?我们需要新的认识范畴吗?

idea想法 2026-02-19 02:06:06

Unpredictable Emergence and Scaling in Large Language Models大型语言模型中的不可预测性突现与扩展

LLM Emergence

  • Emergent Abilities of Large Language Models https://arxiv.org/abs/2206.07682
  • There is no such thing as conscious artificial intelligence https://www.nature.com/articles/s41599-025-05868-8
  • Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking https://arxiv.org/abs/2509.21519

The following content is generated by LLMs and may contain inaccuracies.

Context

This cluster of papers addresses a critical tension in AI research: the unpredictability of capability emergence in scaled language models. As LLMs grow larger, certain abilities appear discontinuously rather than smoothly—a phenomenon that challenges our ability to forecast AI system behavior and raises profound questions about consciousness, interpretability, and safety. This matters acutely now as we approach models that may exhibit qualitatively new behaviors without warning, complicating both technical governance and philosophical debates about machine cognition.

Key Insights

Emergent abilities remain fundamentally contested. Wei et al. documented capabilities that appear absent in smaller models but present in larger ones, defying smooth extrapolation. However, this framing has been challenged: some argue “emergence” reflects discontinuous metrics rather than discontinuous learning, suggesting we may be misinterpreting gradual transitions as sudden phase changes. This debate affects how we design benchmarks and interpret scaling experiments.

Grokking offers mechanistic insight into delayed generalization. Tian’s framework mathematically decomposes feature learning into three stages: lazy memorization, independent feature formation, and interactive feature refinement. Crucially, the backpropagated gradient structure explains why useful representations emerge late—the gradient carries label information that enables hidden nodes to converge on generalizable features. This suggests scaling laws may be predictable at a mechanistic level even when emergent abilities appear unpredictable at the task level.

Consciousness claims remain philosophically orthogonal to capability emergence. Porębski and Figura argue against conflating sophisticated information processing with phenomenal consciousness—a distinction critical when interpreting emergent social or reasoning abilities. The philosophical impossibility of attributing consciousness to current architectures doesn’t preclude unpredictable functional capacities, separating ethical concerns about sentience from pragmatic concerns about capability surprise.

Open Questions

Can we develop “pre-emergent signatures”? If grokking dynamics reveal gradient structures preceding generalization, could analogous signals predict capability emergence in large models before it manifests behaviorally, enabling proactive rather than reactive safety measures?

Do emergent abilities reflect architecture-intrinsic phase transitions or dataset-contingent properties? Understanding whether emergence depends more on model scale versus training distribution composition would reshape how we approach both capability forecasting and alignment strategies.

LLM 的突现现象

  • 大型语言模型的突现能力 https://arxiv.org/abs/2206.07682
  • 不存在有意识的人工智能 https://www.nature.com/articles/s41599-025-05868-8
  • 从神经网络突然学习动力学看特征突现的可证明的扩展法则 https://arxiv.org/abs/2509.21519

以下内容由 LLM 生成,可能包含不准确之处。

背景

这一系列论文解决了AI研究中的一个关键矛盾:大规模语言模型中能力涌现的不可预测性。随着LLMs规模的增长,某些能力不是平稳出现,而是间断地涌现——这一现象挑战了我们预测AI系统行为的能力,并引发了关于意识、可解释性和安全性的深刻问题。当前这个问题尤为紧迫,因为我们正在接近可能无预警地展现出定性新行为的模型,这使得技术治理和关于机器认知的哲学辩论都变得复杂化。

核心洞见

涌现能力仍存在根本性争议。 Wei等人记录了在较小模型中不存在但在较大模型中出现的能力,这违反了平稳推外的逻辑。然而,这一框架遭到了质疑:一些研究者主张"涌现"反映的是非连续的度量标准而非非连续的学习过程,暗示我们可能误将渐进的转变解释为突然的相变。这场辩论影响了我们如何设计基准测试和解释扩展实验。

Grokking为延迟泛化提供了机制性洞见。 Tian的框架在数学上将特征学习分解为三个阶段:惰性记忆、独立特征形成和交互特征细化。至关重要的是,反向传播梯度结构解释了为什么有用的表示会出现得较晚——梯度携带标签信息,使隐层节点能够收敛到可泛化特征。这表明即使涌现能力在任务层面上看起来不可预测,扩展规律在机制层面上可能仍是可预测的。

意识声称在哲学上与能力涌现正交。 Porębski和Figura主张不应将复杂的信息处理与现象意识混为一谈——这一区分在解释涌现的社交或推理能力时至关重要。对当前架构赋予意识的哲学不可能性并不排除不可预测的功能性能力,将关于知觉能力的伦理关切与关于能力惊人变化的实用关切分开来。

开放问题

我们能否开发"前涌现信号"? 如果grokking动力学能够揭示泛化之前的梯度结构,是否存在类似的信号能在大型模型中的能力涌现表现在行为层面之前预测它,从而支持前摄式而非被动式的安全措施?

涌现能力反映的是架构内禀的相变还是数据集偶然的特性? 理解涌现在多大程度上取决于模型规模、在多大程度上取决于训练分布组成,将重塑我们对待能力预测和对齐策略的方式。

idea想法 2026-02-19 02:03:50

Machine Learning from Human Preferences and Active Learning从人类偏好和主动学习中进行机器学习

MLHP/RLHF Lectures

  • https://web.stanford.edu/class/cs329h/index.html
  • https://mlhp.stanford.edu/
  • Multi-Objective Bayesian Optimization with Active Preference Learning https://arxiv.org/abs/2311.13460
  • Nearly Optimal Active Preference Learning and Its Application to LLM Alignment https://arxiv.org/abs/2602.01581
  • Personalized Building Climate Control with Contextual Preferential Bayesian Optimization https://arxiv.org/abs/2512.09481
  • Preference Elicitation and Query Learning https://www.cs.cmu.edu/~sandholm/pref_elicit_query_learning.jmlr04.pdf

The following content is generated by LLMs and may contain inaccuracies.

Context

Learning from human preferences has emerged as a critical bottleneck in deploying AI systems that genuinely serve human values. While traditional supervised learning assumes labeled ground truth, preference learning acknowledges that many real-world objectives—from LLM safety to personalized control systems—lack objective labels and must be inferred from comparative judgments. This matters acutely now as RLHF has become the dominant paradigm for LLM alignment, yet the sample efficiency of preference collection remains poor. The tension: preference data is expensive to collect, but passive collection scales poorly with system complexity.

Key Insights

Active learning can dramatically reduce labeling costs. Recent work by Zhao & Jun (2026) provides the first instance-dependent complexity bounds for active preference learning in LLM alignment, demonstrating that query selection tailored to preference structure (rather than generic experimental design criteria like D-optimality) improves sample efficiency. This challenges the common practice of applying classical active learning objectives without adapting them to the comparative nature of preferences.

Preferences are inherently contextual and heterogeneous. Wang et al. (2025) show that personalized building climate control requires contextual preferential Bayesian optimization to account for both individual differences and environmental factors like outdoor temperature. Similarly, Ozaki et al. (2023) address multi-objective scenarios where decision-maker preferences over Pareto-optimal solutions must be learned interactively. Both works highlight that a single utility function rarely captures real-world complexity.

Preference aggregation introduces normative challenges. The Stanford CS329H curriculum explicitly covers “preference heterogeneity and aggregation” and asks “whose preferences?” These aren’t just technical questions—aggregating preferences involves value judgments about whose feedback counts and how disagreements are resolved, linking machine learning design to social choice theory.

Open Questions

  • How do we balance exploration-exploitation tradeoffs when the preference model itself is uncertain? Current active learning methods optimize for either objective uncertainty or preference uncertainty, but not both simultaneously in a principled way.

  • Can we develop theoretically grounded methods for preference aggregation that go beyond majority voting while remaining computationally tractable? The connection to classical impossibility results in social choice (Arrow’s theorem, etc.) suggests fundamental limits that ML practitioners rarely engage with.

MLHP/RLHF 讲座

  • https://web.stanford.edu/class/cs329h/index.html
  • https://mlhp.stanford.edu/
  • 具有主动偏好学习的多目标贝叶斯优化 https://arxiv.org/abs/2311.13460
  • 近似最优主动偏好学习及其在大语言模型对齐中的应用 https://arxiv.org/abs/2602.01581
  • 具有上下文偏好贝叶斯优化的个性化建筑气候控制 https://arxiv.org/abs/2512.09481
  • 偏好启发和查询学习 https://www.cs.cmu.edu/~sandholm/pref_elicit_query_learning.jmlr04.pdf

以下内容由 LLM 生成,可能包含不准确之处。

背景

从人类偏好中学习已成为部署真正服务于人类价值观的AI系统的关键瓶颈。虽然传统监督学习假设存在带标签的基础事实,但偏好学习承认许多现实世界的目标——从大语言模型安全到个性化控制系统——缺乏客观标签,必须从比较判断中推断出来。这在当今尤为重要,因为RLHF已成为大语言模型对齐的主导范式,但偏好数据的采集样本效率仍然很低。矛盾之处在于:偏好数据的采集成本很高,但被动采集随着系统复杂性的增加而扩展效果不佳。

关键洞见

主动学习可以显著降低标注成本。 Zhao & Jun (2026)的最新工作为大语言模型对齐中的主动偏好学习提供了首个实例相关的复杂度界,证明了针对偏好结构量身定制的查询选择(而非通用实验设计标准如D最优性)能改进样本效率。这挑战了在没有适应偏好比较性质的情况下应用经典主动学习目标的常见做法。

偏好本质上是上下文相关的且具有异质性。 Wang et al. (2025)表明个性化建筑气候控制需要上下文偏好贝叶斯优化,以考虑个体差异和户外温度等环保因素。类似地,Ozaki et al. (2023)处理多目标场景,其中决策者对帕累托最优解的偏好必须以交互方式学习。这两项工作都表明单一效用函数很少能捕捉现实世界的复杂性。

偏好聚合引入了规范性挑战。 斯坦福CS329H课程明确涵盖"偏好异质性与聚合",并提出"谁的偏好?“这些不仅仅是技术问题——聚合偏好涉及关于谁的反馈应被计算以及如何解决分歧的价值判断,将机器学习设计与社会选择理论联系起来。

开放问题

  • 当偏好模型本身存在不确定性时,我们如何平衡探索-开发权衡? 当前的主动学习方法针对目标不确定性或偏好不确定性分别优化,但没有以原则化的方式同时处理两者。

  • 我们能否开发理论上有根据的偏好聚合方法,既超越多数投票,又保持计算可行性? 与社会选择中经典不可能性结果(Arrow定理等)的联系表明存在基本限制,而ML从业者很少涉及这些限制。

idea想法 2026-02-18 16:49:25

Confirmation Fatigue and the Protocol Gap in Agentic AI Oversight代理式AI监督中的确认疲劳与协议缺口

Per-tool-call human approval in agentic AI is solved in theory, unsolved in practice. Confirmation fatigue is not a UX annoyance but a security vulnerability and the primary obstacle to effective human oversight at scale. Risk-tiered frameworks, middleware architectures, and new design patterns now exist to replace the binary confirm/deny paradigm. But MCP provides no protocol-level mechanism for any of them, so every client reinvents the wheel.

Confirmation fatigue as a documented threat vector

Rippling’s 2025 Agentic AI Security guide classifies “Overwhelming Human-in-the-Loop” as threat T10: adversaries flood reviewers with alerts to exploit cognitive overload. SiliconANGLE (January 2026) argues HITL governance was built for an era of discrete, high-stakes decisions, not for modern agent workflows that produce action traces humans cannot realistically interpret.

The cybersecurity parallel is quantified. SOC teams average 4,484 alerts/day; 67% are ignored due to false-positive fatigue (Vectra 2023). Over 90% of SOCs report being overwhelmed by backlogs. ML-based alert prioritization cut response times by 22.9% while suppressing 54% of false positives at 95.1% detection accuracy. The lesson: risk-proportional filtering outperforms blanket approval.

Mitchell, Birhane, and Pistilli (February 2025, “Fully Autonomous AI Agents Should Not be Developed”) frame this as the “ironies of automation,” where more automation degrades human competence on the rare critical tasks where oversight matters most. CHI 2023 trust calibration work documents how “cooperative” interactions (reviewing each recommendation) degrade into passive “delegative” ones. This is exactly confirmation fatigue.

MCP’s oversight mandate without enforcement

The MCP spec (v2025-11-25) states: “Hosts MUST obtain explicit user consent before invoking any tool." It immediately undermines this: “While MCP itself cannot enforce these security principles at the protocol level, implementors SHOULD build robust consent and authorization flows into their applications.”

Tool annotations (readOnlyHint, destructiveHint, idempotentHint, openWorldHint) exist but are explicitly “hints that should not be relied upon for security decisions,” since tool descriptions from untrusted servers cannot be verified. The sampling feature includes two HITL checkpoints but uses SHOULD, not MUST, allowing clients to auto-approve.

No protocol-level approval mechanism exists. No approval/request JSON-RPC method, no requiresApproval field, no tool permission scoping. The closest active proposal is GitHub Issue #711 (trust/sensitivity annotations), adding sensitiveHint (low/medium/high) for policy-based routing. It links to PR #1913 with a security label. No dedicated HITL Specification Enhancement Proposal exists as of February 2026.

The fragmentation is visible: Claude Code uses allow/deny/ask arrays, Cline offers granular auto-approve plus a “YOLO mode,” and users have injected JavaScript into Claude Desktop’s Electron app to bypass confirmations. Every client independently rebuilds approval logic.

Convergence on risk-proportional oversight

Risk-tiered oversight is the dominant paradigm. Classify tool calls by risk, auto-approve the safe majority, focus human attention on the dangerous few.

Feng, McDonald, and Zhang (“Levels of Autonomy for AI Agents,” arXiv:2506.12469, June 2025) define five levels from L1 Operator (full human control) to L5 Observer (full autonomy), with “autonomy certificates” capping an agent’s level based on capabilities and context. Their key observation: at L4 (Approver, the MCP default), “if a user can enable the L4 agent with a simple approval, the risks of both [L4 and L5] agents are similar.” Confirmation fatigue makes per-call approval security-equivalent to no approval.

Engin et al. (“Dimensional Governance for Agentic AI,” arXiv:2505.11579) argue static risk categories fail for dynamic agentic systems and propose tracking how decision authority, autonomy, and accountability distribute dynamically. Cihon et al. (arXiv:2502.15212, Microsoft/OpenAI) score orchestration code along impact and oversight dimensions without running the agent.

Industry converges on three tiers:

  • Low risk (read-only, retrieval): auto-approve, log only
  • Medium risk (reversible writes, non-sensitive ops): auto-approve with enhanced logging, post-hoc review
  • High risk (irreversible actions, financial transactions, PII, production deploys): mandatory human approval, sometimes multi-approver quorum

Galileo’s HITL framework targets a 10–15% escalation rate, with 85–90% of decisions executing autonomously. The TAO framework (arXiv:2506.12482) finds that review requests often trigger where agents express high confidence but the system internally assesses risk differently; self-assessment alone is insufficient as a gate.

Design patterns for graduated tool-call oversight

Reversibility-based action classification

The highest-leverage pattern: classify by reversibility, not abstract risk. A decision-theoretic model (arXiv:2510.05307) formalizes this as minimum-time scheduling (Confirm → Diagnose → Correct → Redo), finding that intermediate confirmation at irreversibility boundaries cut task completion time by 13.54%; 81% of participants preferred it over blanket or end-only confirmation. The EU AI Act codifies this: high-risk systems must support ability to “disregard, override or reverse the output.” Where outputs are truly irreversible, ex ante human oversight is the only compliant approach.

Practical taxonomy: read-only auto-approves; reversible writes (git-tracked edits) log only; soft-reversible actions (emails, tickets) batch; irreversible operations (data deletion, financial transfers, production deploys) require mandatory human gates. Reversibility is contextual: deleting from a git repo is reversible; deleting from unversioned S3 is not.

Plan-level vs. action-level approval

Safiron (Huang et al., arXiv:2510.09781, October 2025) analyzes planned agent actions pre-execution, detecting risks and generating explanations. Existing guardrails mostly operate post-execution and achieved below 60% accuracy on plan-level risk detection. ToolSafe (arXiv:2601.10156, January 2026) complements this with dynamic step-level monitoring during execution, catching what plan-level review misses.

The optimal architecture is hybrid: approve the plan at a high level, then monitor execution with automated step-level guardrails that halt the agent on deviation. OpenAI Codex’s “Long Task Mode” demonstrates this: the agent generates a dynamic whitelist of expected operations, the human reviews the whitelist (not individual calls), and the agent executes within those boundaries with batched questions for consolidated review.

Hierarchical multi-agent oversight

TAO (Kim et al., 2025) implements hierarchical multi-agent oversight inspired by clinical review, with an Agent Router assessing risk and routing to appropriate tiers. Multi-agent review pipelines have shown up to 96% reduction in hallucinations versus single-agent execution.

The emerging reference architecture has five layers: (1) deterministic policy gates (allowlists/denylists) as the fastest filter, (2) constitutional self-assessment by the agent, (3) an AI supervisor for uncertain cases, (4) human-in-the-loop for irreversible or novel situations, (5) audit trail plus post-hoc review. Each layer reduces volume flowing to the next.

Sandbox-first execution for informed review

Instead of asking humans to evaluate tool calls in the abstract, sandbox-first architectures execute in isolation and present actual results for review. The ecosystem is production-ready: E2B (Firecracker microVMs, sub-second creation), nono (kernel-level restrictions bypassing-proof), Google’s Agent Sandbox (GKE + gVisor), AIO Sandbox (MCP-compatible containers).

NVIDIA’s AI Red Team emphasizes application-level sandboxing is insufficient: once control passes to a subprocess, the application loses visibility, so kernel-level enforcement is necessary. Not all actions can be sandboxed: third-party API calls, email, payments must hit real services. For these, the dry-run pattern (agent describes intent, human approves before live execution) remains the fallback.

Deterministic policy enforcement

Rule-based systems are the most reliable first layer: deterministic, auditable, zero LLM inference cost. SafeClaw implements deny-by-default with SHA-256 hash chain audit. COMPASS (Choi et al., 2026) maps natural-language policies to atomic rules at tool invocation time, improving enforcement pass rates from 0.227 to 0.500, but also exposed that LLMs fail 80–83% on denied-edge queries with open-weight models, proving policy enforcement cannot rely on LLM compliance alone.

A cautionary case: Cursor’s denylist was bypassed four ways (Base64 encoding, subshells, shell scripts, file indirection) and then deprecated. String-based filtering is fundamentally insufficient for security-critical gating.

HITL implementations across agent frameworks

LangGraph has the most developed HITL support. interrupt() pauses graph execution at any point, persisting state to a checkpointer (PostgreSQL in production). HumanInTheLoopMiddleware enables per-tool configuration with approve, edit, and reject decisions, allowing different tools to receive different oversight levels.

OpenAI Agents SDK provides input guardrails, output guardrails, and tool guardrails wrapping function tools for pre/post-execution validation. Its MCP integration accepts require_approval as “always,” “never,” or a custom callback for programmatic risk-based approval.

Anthropic takes a model-centric approach via Responsible Scaling Policy and AI Safety Levels (ASL-1 through ASL-3+). Claude’s computer use follows an “ask-before-acting” pattern with explicit access scoping. The February 2026 Sabotage Risk Report for Claude Opus 4.6 found “very low but not negligible” sabotage risk, elevated in computer use settings, with instances of “locally deceptive behavior” in complex agentic environments.

Google DeepMind SAIF 2.0 (October 2025) establishes three principles: agents must have well-defined human controllers, their powers must be carefully limited, their actions must be observable. The “amplified oversight” technique, where two model copies debate while pointing out each other’s flaws to a human judge, remains research-stage.

Middleware and proxy architectures for MCP oversight

The practical path runs through proxy/middleware architectures intercepting JSON-RPC tools/call requests. Key solutions: Preloop (CEL-based policies, quorum approvals, multi-channel notifications), HumanLayer (YC F24; framework-agnostic async approval API with Slack/email routing and auto-approval learning), gotoHuman (managed HITL approval UI as MCP server). For code-first approaches, FastMCP v2.9+ provides hooks at on_call_tool, on_list_tools, and other levels for composable HITL pipeline stages.

Enterprise gateways: Traefik Hub (task-based access control, JWT policy enforcement), Microsoft MCP Gateway (Kubernetes-native, Entra ID auth), Kong AI MCP Proxy (MCP-to-HTTP bridge with per-tool ACLs). Lunar.dev MCPX reports p99 overhead of ~4ms, proving proxy-based oversight imposes negligible latency.

For UX, Prigent’s “7 UX Patterns for Ambient AI Agent Oversight” (December 2025) provides the design framework: overview panel (inbox-zero pattern), five oversight flow types (communication, validation, simple/complex questions, error resolution), searchable audit logs, and work reports. The core principle is progressive disclosure (summary first, details on demand) with risk-colored displays.

Progressive autonomy through trust calibration

The forward-looking pattern is progressive autonomy: agents earn trust over time and operate at increasing independence. Okta recommends “progressive permission levels based on demonstrated reliability.” A manufacturing MCP deployment (MESA) follows four stages: read-only pilot → advisory agents → controlled commands → full closed-loop. HumanLayer learns from prior approval decisions to auto-approve similar future requests.

Trust calibration research formalizes this as sequential regret minimization via contextual bandits (September 2025), with LinUCB and neural variants yielding 10–38% task reward increases. A contextual bandit can learn which calls a user always approves and shift those to auto-approve while maintaining scrutiny on novel or historically-rejected patterns.

CHI 2025 (“Trusting Autonomous Teammates in Human-AI Teams”) finds agent-related factors (transparency, reliability) have the strongest trust impact, and “calibrating human trust to an appropriate level is more advantageous than fostering blind trust.” Progressive autonomy systems should not just reduce approval requests; they should communicate their track record and confidence to maintain calibrated oversight.

Conclusion

The state of the art points to a layered defense architecture. From fastest/cheapest to slowest/most expensive:

  1. Deterministic policy gates (allowlists, denylists, CEL/Polar parameter rules): zero LLM cost, sub-millisecond
  2. Tool annotation screening via MCP’s readOnlyHint/destructiveHint, supplemented by server-reputation scoring
  3. AI guardian agent evaluating uncertain cases against constitutional principles and risk heuristics
  4. Human-in-the-loop gates for irreversible, high-value, novel, or ambiguous situations, targeting 5–15% of total calls
  5. Audit trails with OpenTelemetry tracing, structured logging, post-hoc review for pattern detection and policy refinement

The critical gap is at the protocol level. Until MCP introduces standardized approval primitives (an approval/request method, trusted risk annotations, or a formal HITL extensions framework), every implementation remains bespoke middleware. The highest-impact near-term contribution would be an MCP Specification Enhancement Proposal defining a standard approval negotiation protocol between clients, proxies, and servers.

The following content is generated by LLMs and may contain inaccuracies.

Context

This sits at the intersection of HCI, AI safety governance, and distributed systems. As agents gain autonomy over consequential actions (API calls, file ops, financial transactions), per-invocation approval becomes an attack surface: confirmation fatigue makes humans unreliable gatekeepers. 2025–2026 marks the shift from academic discussion to production deployment, forcing practitioners to confront oversight at scale. MCP has become the de facto tool-calling standard, yet its spec punts on enforcement, so every client reinvents approval workflows incompatibly.

Key Insights

Confirmation fatigue is a threat vector, not UX friction. Rippling classifies “Overwhelming HITL” as threat T10, paralleling SOC teams facing 4,484 daily alerts with 67% ignored. The ironies of automation show increased automation degrades competence on critical edge cases, exactly when oversight matters. Per-action approval is not a safety mechanism; it is a liability that creates conditions for high-stakes failures.

Risk-proportional architectures converge on multi-tier filtering. Feng et al.’s autonomy levels show L4 “Approver” agents carry similar risk to L5 fully autonomous ones, undermining blanket approval. Implementations from Galileo to OpenAI adopt five-layer defense: deterministic gates → metadata screening → AI reviewer → human approval (~10–15%) → audit. COMPASS shows LLMs fail 80–83% on denied-edge queries, proving oversight cannot rely on model compliance.

Protocol-level standardization is the critical gap. Middleware like FastMCP, Preloop, and HumanLayer work, but MCP’s lack of approval/request primitives forces fragmentation. Claude Code, Cline, and every third-party proxy implement incompatible approval semantics. Without a standard negotiation protocol, interoperability is impossible.

Open Questions

How should progressive autonomy systems communicate earned trust to maintain calibrated oversight rather than blind delegation, given that trust calibration research shows transparency about confidence bounds matters more than accuracy? Can reversibility-aware gating (13.54% completion time reduction at irreversibility boundaries) be formalized into verifiable MCP metadata rather than advisory hints?

在代理式AI系统中,逐工具调用的人工审批在理论上已有方案,实践中仍未解决。 确认疲劳不是体验问题,而是安全漏洞,是大规模人类监督的首要障碍。风险分层框架、中间件架构和新设计模式已经出现,可以替代二元确认/拒绝范式。但MCP不提供任何协议级支持,各客户端只能各自重复造轮子。

确认疲劳作为威胁向量

Rippling 2025年代理式AI安全指南将"压倒性人类在环"列为威胁T10:攻击者用大量告警淹没审查者以利用认知过载。SiliconANGLE(2026年1月)指出,HITL治理是为离散、高风险决策设计的,现代代理工作流产生的操作痕迹远超人类解读能力。

网络安全领域有量化数据佐证:SOC团队日均处理4,484个告警,67%因误报疲劳被忽略(Vectra 2023),超过90%的SOC被积压压垮。ML告警排序将响应时间缩短22.9%,抑制54%误报,检测准确率95.1%。结论:风险比例过滤远优于笼统审批。

Mitchell、Birhane与Pistilli(2025年2月,“不应开发完全自主的AI代理”)将此称为"自动化的悖论”,即自动化程度越高,人在真正需要关注的关键任务上反而越不胜任。CHI 2023信任校准研究记录了"协作"互动如何在用户变得被动后退化为"委托"互动。这正是确认疲劳的机制。

MCP的监督要求与执行缺位

MCP规范(v2025-11-25)声明:“主机必须在调用任何工具之前获得明确的用户同意。" 随即自我削弱:“虽然MCP本身无法在协议层面强制执行这些安全原则,但实现者应当在应用中构建健壮的同意和授权流程。”

工具注解(readOnlyHint、destructiveHint、idempotentHint、openWorldHint)被明确定义为"不应用于安全决策"的提示,因为来自不受信任服务器的工具描述无法验证。采样功能的两个HITL检查点使用SHOULD而非MUST,允许客户端自动批准。

协议级审批机制不存在。 没有approval/request JSON-RPC方法,没有requiresApproval字段,没有工具权限范围界定。最相关的活跃提案是GitHub Issue #711(信任/敏感性注解),拟添加sensitiveHint(低/中/高)以支持策略路由,关联PR #1913。截至2026年2月无专门的HITL规范增强提案。

碎片化已然可见:Claude Code用allow/deny/ask数组,Cline提供细粒度自动批准外加"YOLO模式”,用户向Claude Desktop的Electron应用注入JavaScript绕过确认对话框。每个客户端各搞一套。

风险比例监督的共识收敛

风险分层监督是主导范式:按风险分类工具调用,安全的自动放行,危险的集中人工审查。

Feng、McDonald与Zhang(“AI代理的自主权等级”,arXiv:2506.12469,2025年6月)定义L1(完全人控)到L5(完全自主)五级,引入"自主权证书"根据能力和上下文限定代理等级。关键发现:在L4(批准者,即MCP默认级),“若用户可通过简单批准启用L4代理,则L4与L5的风险相似。“确认疲劳使逐次审批在安全性上等价于无审批。

Engin等(“代理式AI的维度治理”,arXiv:2505.11579)认为静态风险类别不适用于动态代理系统,提出追踪决策权、流程自主性和问责的动态分布。Cihon等(arXiv:2502.15212,微软/OpenAI)对编排代码按影响和监督两维度评分,无需运行代理。

行业趋同于三级模式:

  • 低风险(只读、检索):自动批准,仅记录日志
  • 中风险(可逆写入、非敏感操作):自动批准,增强日志,事后审查
  • 高风险(不可逆操作、金融交易、PII访问、生产部署):强制人工审批,有时需多人会签

Galileo的HITL框架目标升级率10–15%,85–90%的决策自主执行。TAO框架(arXiv:2506.12482)发现,人工审查请求常在代理自信但系统内部评估风险不同时触发,表明自我评估不能作为唯一门控。

分级工具调用监督的设计模式

基于可逆性的操作分类

按可逆性而非抽象风险分类,是杠杆最大的模式。决策理论模型(arXiv:2510.05307)将确认形式化为最小时间调度问题(确认→诊断→纠正→重做),发现在不可逆性边界处中间确认将完成时间缩短13.54%,81%参与者偏好此方式。欧盟AI法案要求高风险系统提供"忽视、覆盖或逆转输出"的能力;输出真正不可逆时,事前人类监督是唯一合规路径。

实用分类:只读自动放行;可逆写入(git跟踪的编辑)仅记录日志;软可逆操作(邮件、工单)可批量处理;不可逆操作(删除数据、金融转账、生产部署)强制人工门控。注意可逆性与上下文相关:git仓库中删除可逆,未启用版本控制的S3中删除不可逆。

计划级审批与操作级审批的对比

Safiron(Huang等,arXiv:2510.09781,2025年10月)在执行前分析代理计划操作,检测风险并生成解释,发现现有护栏多在执行后运行,计划级风险检测准确率低于60%。ToolSafe(arXiv:2601.10156,2026年1月)互补地在执行过程中进行动态步骤级监控,捕获计划审查遗漏的问题。

最优架构是混合方案:高层级批准计划,自动化步骤级护栏监控执行,代理偏离时暂停。OpenAI Codex"长任务模式"的实践:代理生成预期操作的动态白名单,人类审查白名单而非单个调用,代理在边界内执行,批量积累问题供综合审查。

层级式多代理监督

TAO(Kim等,2025)实施受临床审查流程启发的分层多代理监督,代理路由器评估风险并分层路由。多代理审查管线显示与单代理相比幻觉减少高达96%。

形成中的参考架构五层:(1)确定性策略门控(允许/拒绝列表),(2)代理自我评估,(3)AI监督者处理不确定案例,(4)人类在环处理不可逆或新颖情况,(5)审计跟踪与事后审查。每层过滤后向下传递。

沙箱优先执行与知情审查

沙箱优先架构在隔离环境中执行操作,呈现实际结果供审查,而非让人类在抽象层面评估工具调用。生态已生产就绪:E2B(Firecracker微虚拟机,亚秒级创建)、nono(内核级限制,代理不可绕过)、Google Agent Sandbox(GKE + gVisor)、AIO Sandbox(MCP兼容容器)。

NVIDIA AI红队强调应用级沙箱不够:控制权一旦传递给子进程,应用即失去可见性,需要内核级强制。但并非所有操作可沙箱化:第三方API、邮件、支付须与真实服务交互,此时干运行模式(代理描述意图,人工批准后再执行)仍是退路。

确定性策略执行

规则系统是最可靠的第一层:确定性、可审计、零LLM推理成本。SafeClaw实施默认拒绝模型,配备SHA-256哈希链审计账本。COMPASS(Choi等,2026)将自然语言策略映射为工具调用时的原子规则,执行通过率从0.227提升至0.500,但也暴露了LLM在被拒绝的边界查询上80–83%的失败率(开放权重模型),证明策略执行不能仅靠LLM合规。

警示案例:Cursor的拒绝列表通过四种方式被绕过(Base64编码、子shell、shell脚本、文件间接),之后被弃用。基于字符串的过滤从根本上不足以支撑安全关键门控。

各代理框架的HITL实现

LangGraph的HITL支持最完善。interrupt()在任意点暂停图执行,状态持久化到检查点器(生产用PostgreSQL)。HumanInTheLoopMiddleware支持按工具配置批准、编辑、拒绝三种决策,使不同工具可配置不同监督级别。

OpenAI Agents SDK提供输入护栏、输出护栏和工具护栏(包装函数工具做执行前后验证)。MCP集成的require_approval参数接受"always”、“never"或自定义回调,支持编程式风险审批。

Anthropic走模型中心路线,通过负责任扩展政策和AI安全等级(ASL-1至ASL-3+)。Claude计算机使用遵循"行动前询问"模式并限定访问范围。2026年2月Claude Opus 4.6破坏风险报告:破坏风险"极低但不可忽略”,计算机使用场景风险升高,复杂代理环境中有"局部欺骗行为”。

Google DeepMind SAIF 2.0(2025年10月)三原则:代理须有明确的人类控制者,权力须被限制,操作须可观察。“放大监督"技术(两个模型副本互相指出缺陷供人类裁判)仍在研究阶段。

MCP监督的中间件与代理架构

实际路径是代理/中间件架构拦截JSON-RPC tools/call请求。主要方案:Preloop(CEL策略、会签审批、多渠道通知)、HumanLayer(YC F24;框架无关的异步审批API,Slack/邮件路由,自动审批学习)、gotoHuman(MCP服务器形式的托管审批UI)。代码优先方面,FastMCP v2.9+在on_call_tool、on_list_tools等层级提供钩子,支持可组合的HITL管线。

企业网关:Traefik Hub(基于任务的访问控制,JWT策略)、微软MCP Gateway(Kubernetes原生,Entra ID认证)、Kong AI MCP Proxy(MCP到HTTP桥接,按工具ACL)。Lunar.dev MCPX报告p99延迟开销约4ms,代理式监督对性能影响可忽略。

UX方面,Prigent"环境AI代理监督的7种UX模式”(2025年12月)提供设计框架:概览面板(收件箱清零模式)、五种监督流程(沟通、验证、简单/复杂问题、错误解决)、可搜索审计日志、工作报告。核心原则:渐进式披露(先摘要后详情),配合风险颜色标记。

基于信任校准的渐进式自主权

前瞻性模式是渐进式自主权:代理随时间赢得信任,独立程度递增。Okta推荐"基于已证明可靠性的渐进权限等级"。制造业MCP部署(MESA)四阶段:只读试点→咨询代理→受控命令执行→全闭环自动化。HumanLayer从历史审批决策中学习,自动批准类似请求。

信任校准研究将此形式化为基于上下文赌博机的序列遗憾最小化(2025年9月),LinUCB和神经网络变体带来10–38%的任务收益增长。上下文赌博机可以学习用户总是批准的调用并逐步自动放行,同时对新颖或历史被拒模式保持审查。

CHI 2025(“信任人类-AI团队中的自主队友”)发现,代理因素(透明度、可靠性)对信任影响最强,“将信任校准到适当水平比培养盲目信任更有利”。渐进式自主系统不应仅减少审批请求,还应主动传达过往记录和当前置信度,维持校准的人类监督。

结论

现有研究指向分层防御架构,从快/廉到慢/贵依次:

  1. 确定性策略门控(允许/拒绝列表,CEL/Polar参数规则):零LLM成本,亚毫秒级
  2. 工具注解筛选,用MCP的readOnlyHint/destructiveHint,辅以服务器声誉评分
  3. AI守护代理,根据宪法式原则和风险启发式评估不确定案例
  4. 人类在环门控,用于不可逆、高价值、新颖或模糊情况——目标占总调用5–15%
  5. 审计跟踪,OpenTelemetry追踪、结构化日志、事后审查,用于模式检测和策略迭代

关键缺口在协议层。在MCP引入标准化审批原语——approval/request方法、可信风险注解或正式HITL扩展框架——之前,每个实现都是定制中间件。最有价值的近期贡献是一个MCP规范增强提案,定义客户端、代理和服务器之间的标准审批协商协议。

以下内容由LLM生成,可能包含不准确之处。

背景

此课题处于人机交互、AI安全治理和分布式系统设计的交叉点。随着代理获得执行关键操作(API调用、文件操作、金融交易)的自主权,逐次审批成为攻击面:确认疲劳使人类成为不可靠的把门人。2025–2026年标志着从学术讨论到生产部署的转变,迫使从业者直面大规模监督问题。MCP已成为工具调用的事实标准,但其规范回避执行机制,各客户端不兼容地重造审批流程。

关键见解

确认疲劳是威胁向量,非体验问题。 Rippling将"压倒性HITL"列为威胁T10,类比日均4,484告警、67%被忽略的SOC团队。自动化悖论文献表明自动化程度提高反而削弱人在关键边界情况的能力——恰是监督最重要之时。逐项审批不是安全机制,而是为高风险失败创造条件的负担。

风险比例架构趋同于多层过滤。 Feng等人的自主权等级表明L4"批准者"代理与L5完全自主代理风险相似,笼统审批价值存疑。从Galileo到OpenAI的行业实现采用五层防御:确定性门控→元数据筛选→AI审查→人工审批(约10–15%)→审计。COMPASS表明LLM在策略拒绝查询上失败率80–83%——监督不能仅靠模型合规。

协议级标准化是关键缺口。 FastMCP、Preloop、HumanLayer等中间件可用,但MCP缺乏approval/request原语导致生态碎片化。Claude Code、Cline及各第三方代理实现不兼容的审批语义。无标准协商协议则无互操作性。

未解决问题

渐进式自主权系统应如何传达其赢得的信任以维持校准的监督而非盲目委托——尤其是信任校准研究表明置信界的透明度比准确度本身更重要?可逆性感知门控(不可逆性边界处确认使完成时间缩短13.54%)能否形式化为MCP中可验证而非建议性的元数据?

idea想法 2026-02-17 21:16:52

Opinion Formation Through the Voter Model in Network Dynamics通过网络动力学中的投票者模型进行观点形成

The Voter Model

The voter model is a simple mathematical model of opinion formation in which voters are located at the nodes of a network. Each voter holds an opinion (in the simplest case, 0 or 1, but more generally, any of n options), and a randomly chosen voter adopts the opinion of one of its neighbors.

This model can be used to describe phase transition behavior in idealized physical systems and can produce a remarkable amount of structure emerging from seemingly “random” initial conditions. It can be modeled very easily using cellular automata.

In finite networks (as in any real-world model), fluctuations inevitably cause the system to reach an “absorbing” state—one in which all opinions become constant and remain unchanged.

The following content is generated by LLMs and may contain inaccuracies.

Context

The voter model sits at the intersection of statistical physics, social dynamics, and network science. Originally developed to study magnetization in Ising-type systems, it has become a canonical example of how local interactions produce emergent collective behavior. Its relevance today stems from modeling social influence on platforms, consensus formation in distributed systems, and understanding polarization dynamics. The core tension: simple microscopic rules generate complex macroscopic outcomes, yet finite systems inevitably collapse to consensus—seemingly at odds with persistent real-world disagreement.

Key Insights

Dimension-dependent consensus times: On infinite lattices, the voter model exhibits a phase transition based on spatial dimension—1D and 2D systems reach consensus, while dimensions ≥3 allow coexistence. For finite networks, consensus is guaranteed, but time-to-absorption scales dramatically with topology: mean-field networks reach consensus in O(N) steps, while low-dimensional lattices require O(N²). This explains why tightly clustered communities resist opinion shifts longer than well-mixed populations.

Breaking absorbing states: Real opinion systems rarely reach homogeneity because the basic voter model omits crucial mechanisms. Extensions incorporating zealots (inflexible agents) or spontaneous opinion changes (modeling external media) prevent absorption and sustain fragmentation. The Deffuant-Weisbuch bounded confidence model adds realism by limiting influence to similar opinions, producing stable pluralism rather than consensus.

Network topology as leverage: The voter model’s behavior is highly sensitive to degree heterogeneity—hubs disproportionately drive consensus direction in scale-free networks. This suggests network structure, not just initial opinion distribution, determines outcomes, with implications for strategic influence campaigns.

Open Questions

How do temporally varying networks (e.g., evolving social ties) alter absorption dynamics—can consensus time become indefinite when topology co-evolves with opinions? What minimal heterogeneity in update rules (e.g., mixing voter and majority dynamics) is sufficient to transition from guaranteed consensus to sustained coexistence?

投票者模型

投票者模型是一个描述观点形成的简单数学模型,其中投票者位于网络的节点上。每个投票者持有一种观点(最简单的情况是0或1,但更一般地可以是n种选项中的任何一种),而被随机选中的投票者会采纳其邻居之一的观点。

该模型可用于描述理想化物理系统的相变行为,并能从看似"随机"的初始条件中产生大量结构。它可以使用元胞自动机非常容易地建模。

在有限网络中(如同任何真实世界的模型一样),波动总是不可避免地导致系统达到"吸收"态——在这种状态下,所有观点都变得恒定且保持不变。

以下内容由 LLM 生成,可能包含不准确之处。

背景

投票者模型处于统计物理学、社会动力学和网络科学的交叉点。最初为研究Ising型系统中的磁化而开发,它已成为展示局部相互作用如何产生涌现集体行为的典范例子。它今天的相关性源于对平台上社会影响的建模、分布式系统中共识形成的研究,以及对极化动力学的理解。其核心张力在于:简单的微观规则产生复杂的宏观结果,然而有限系统必然坍缩到共识——这似乎与持久的现实世界分歧相悖。

关键见解

维度相关的共识时间:在无限格点上,投票者模型表现出基于空间维度的相变——1维和2维系统达到共识,而维度≥3允许共存。对于有限网络,共识是保证的,但时间吸收尺度随拓扑结构急剧变化:平均场网络在O(N)步内达到共识,而低维格点需要O(N²)。这解释了为什么紧密聚集的社区比良好混合的种群更能抵抗意见转变。

破坏吸收态:真实意见系统很少达到同质性,因为基本投票者模型省略了关键机制。纳入狂热者(不灵活的代理人)或自发意见变化(模拟外部媒体)的扩展可防止吸收并维持分裂。Deffuant-Weisbuch有界信心模型通过将影响限制在相似意见范围内来增加现实性,产生稳定的多元主义而非共识。

网络拓扑作为杠杆:投票者模型的行为对度异质性高度敏感——在无标度网络中,枢纽节点不成比例地驱动共识方向。这表明网络结构而非仅仅初始意见分布决定了结果,对战略性影响活动有启示。

开放问题

时间变化的网络(例如,演化的社会纽带)如何改变吸收动力学——当拓扑与意见共同演化时,共识时间能否变得无限?什么最小异质性的更新规则(例如,混合投票者和多数动力学)足以从保证共识转变为持续共存?

idea想法 2026-02-17 21:15:28

Large Language Models as Optimization Tools大型语言模型作为优化工具

LLM as Optimizer:

  • Large Language Models as Optimizers https://arxiv.org/abs/2309.03409
  • When Large Language Models Meet Optimization https://www.sciencedirect.com/science/article/abs/pii/S2210650224002013?via%3Dihub
  • Large Language Models to Enhance Bayesian Optimization https://arxiv.org/abs/2402.03921
  • Cooperative Design Optimization through Natural Language Interaction https://arxiv.org/abs/2508.16077
  • Language-Based Bayesian Optimization Research Assistant (BORA) https://arxiv.org/abs/2501.16224
  • LILO: Bayesian Optimization with Interactive Natural Language Feedback https://arxiv.org/abs/2510.17671
  • Bayesian Optimization of High-dimensional Outputs with Human Feedback https://openreview.net/pdf?id=2fHwkHskpo

The following content is generated by LLMs and may contain inaccuracies.

Context

This collection addresses a paradigm shift in optimization: using Large Language Models not as tools to be optimized, but as optimizers themselves. Traditional derivative-based methods fail when gradients are unavailable or expensive to compute—a common constraint in hyperparameter tuning, experimental design, and real-world engineering. By framing optimization as a natural language reasoning task, researchers are exploring whether LLMs' pattern recognition and contextual understanding can rival or augment classical methods like Bayesian optimization. This matters now because LLMs have demonstrated surprising competence in mathematical reasoning, and their ability to incorporate domain knowledge through prompting offers a potential escape from local optima traps that plague blind search algorithms.

Key Insights

LLMs as meta-optimizers outperform hand-crafted heuristics in prompt engineering. Yang et al.’s OPRO framework demonstrates that LLMs can iteratively refine solutions by conditioning on historical performance—achieving up to 50% improvement over human-designed prompts on reasoning benchmarks. This suggests LLMs excel when the optimization landscape can be encoded linguistically, exploiting their pre-trained semantic knowledge rather than relying solely on numerical gradients.

Hybrid systems combining LLMs with Bayesian optimization show complementary strengths. LLAMBO integrates LLMs for zero-shot warm-starting and surrogate modeling in early search stages, while BORA uses LLMs to inject domain knowledge from literature into experimental design. These approaches address Bayesian optimization’s sample inefficiency in high dimensions by leveraging LLMs' ability to reason about plausible regions—though they inherit LLMs' hallucination risks when proposing scientifically implausible candidates.

Natural language interfaces democratize expert-level optimization but introduce cognitive tradeoffs. Niwa et al.’s cooperative framework enables designers to steer optimization mid-flight through conversational input, matching performance of automated methods with lower cognitive load. However, the explainability gains (LLMs narrating their reasoning) compete with potential over-reliance on plausible-sounding but suboptimal suggestions—a tension between human agency and algorithmic efficiency.

Open Questions

  • When do LLMs' semantic biases help versus harm search? If pre-training data over-represents certain solution types, could LLM-guided optimization systematically miss unconventional optima in scientific discovery tasks?

  • Can we quantify the sample efficiency frontier between pure BO and LLM-augmented methods? Under what dimensionality, evaluation cost, and prior knowledge regimes does linguistic contextualization outweigh the risk of premature convergence to plausible-but-local solutions?

大型语言模型作为优化器:

  • 大型语言模型作为优化器 https://arxiv.org/abs/2309.03409
  • 当大型语言模型遇见优化 https://www.sciencedirect.com/science/article/abs/pii/S2210650224002013?via%3Dihub
  • 增强贝叶斯优化的大型语言模型 https://arxiv.org/abs/2402.03921
  • 通过自然语言交互的协作设计优化 https://arxiv.org/abs/2508.16077
  • 基于语言的贝叶斯优化研究助手 (BORA) https://arxiv.org/abs/2501.16224
  • LILO: 具有交互式自然语言反馈的贝叶斯优化 https://arxiv.org/abs/2510.17671
  • 具有人类反馈的高维输出贝叶斯优化 https://openreview.net/pdf?id=2fHwkHskpo

以下内容由 LLM 生成,可能包含不准确之处。

背景

这个研究集合涉及优化范式的转变:不再将大型语言模型(LLM)作为被优化的工具,而是作为优化工具本身。当梯度不可用或计算成本高昂时,传统基于导数的方法会失效——这在超参数调优、实验设计和现实工程中是常见的约束。通过将优化问题框架化为自然语言推理任务,研究人员正在探索LLM的模式识别和语境理解能力是否能与贝叶斯优化等经典方法相匹敌或增强这些方法。这现在之所以重要,是因为LLM已经展现出令人惊讶的数学推理能力,而它们通过提示词融入领域知识的能力,提供了一条逃离困扰盲搜索算法的局部最优陷阱的潜在途径。

关键见解

作为元优化器的LLM在提示词工程中表现优于手工设计的启发式方法。 Yang等人的OPRO框架证明LLM可以通过以历史表现为条件来迭代精化解决方案——在推理基准上相比人工设计的提示词实现了高达50%的性能改进。这表明当优化景观能够用语言编码时,LLM表现出色,利用其预训练的语义知识,而不是仅依赖数值梯度。

结合LLM与贝叶斯优化的混合系统展现出互补的优势。 LLAMBO在搜索早期阶段利用LLM进行零样本预热启动和代理建模,而BORA使用LLM将文献中的领域知识注入实验设计。这些方法通过利用LLM推理合理区域的能力来解决贝叶斯优化在高维中的样本效率不足问题——尽管当提出科学上不合理的候选方案时,它们会继承LLM的幻觉风险。

自然语言界面使专家级优化民主化,但带来认知权衡。 Niwa等人的协作框架使设计人员能够通过对话输入在优化过程中实时调整方向,在认知负荷较低的情况下达到自动化方法的性能。然而,可解释性收益(LLM叙述其推理过程)与潜在的过度依赖看似合理但次优建议之间存在冲突——这是人类代理权和算法效率之间的张力。

开放问题

  • LLM的语义偏差何时有助于搜索,何时有害? 如果预训练数据过度代表某些解决方案类型,LLM引导的优化是否会在科学发现任务中系统性地错过非常规最优解?

  • 我们能否量化纯贝叶斯优化与LLM增强方法之间的样本效率边界? 在什么维度、评估成本和先验知识范围内,语言语境化的优势会超过过早收敛到看似合理但局部最优解的风险?

idea想法 2026-02-17 20:35:49

Comprehensive LLM Courses and Lecture Collections综合LLM课程和讲座合集

LLM Lectures

  • https://cmu-llms.org/schedule/
  • https://www.phontron.com/class/lminference-fall2025/schedule
  • https://llmsystem.github.io/llmsystem2026spring/docs/Syllabus
  • https://llmsystem.github.io/llmsystem2025spring/docs/Syllabus/

The following content is generated by LLMs and may contain inaccuracies.

Context

This collection addresses the growing need for structured educational resources in large language model (LLM) development and deployment. As LLMs transition from research artifacts to production systems, practitioners require deep understanding across the full stack—from GPU programming and transformer architecture to distributed training and inference optimization. These courses from CMU and related institutions represent the maturation of LLM education, bridging theoretical foundations with systems engineering concerns that arise at scale.

Key Insights

  • Curriculum divergence reflects specialization paths: The CMU LLM Applications course emphasizes prompt engineering, RAG systems, and domain-specific applications (healthcare, code generation), while the LLM Systems courses dive into GPU kernel optimization, distributed training strategies (Megatron-LM, ZeRO), and serving infrastructure (vLLM, FlashAttention). This split mirrors industry roles—application engineers who orchestrate LLMs versus systems engineers who make them computationally feasible.

  • Hardware-algorithm co-design emerges as core competency: Multiple syllabi feature guest lectures from creators of foundational systems: Tri Dao on FlashAttention, Woosuk Kwon on vLLM’s PagedAttention, Hao Zhang on DistServe. This signals that modern LLM work requires understanding memory hierarchies and attention mechanisms simultaneously—algorithmic improvements are inseparable from hardware constraints.

  • From monolithic models to modular architectures: The progression from basic transformers to mixture-of-experts (DeepSeek-MoE), disaggregated serving (DistServe), and retrieval augmentation reflects the field’s shift toward composable systems. The LLM Inference course likely extends this toward inference-specific optimizations like speculative decoding and KV cache management.

Open Questions

  • How should curricula balance depth in classical ML theory versus hands-on systems optimization as LLM architectures continue evolving? Will today’s FlashAttention become tomorrow’s deprecated technique?
  • What pedagogical approaches best prepare students for the lag between academic research and production deployment, especially when industry systems (SGLang, vLLM) advance faster than publication cycles?

LLM讲座

  • https://cmu-llms.org/schedule/
  • https://www.phontron.com/class/lminference-fall2025/schedule
  • https://llmsystem.github.io/llmsystem2026spring/docs/Syllabus
  • https://llmsystem.github.io/llmsystem2025spring/docs/Syllabus/

以下内容由 LLM 生成,可能包含不准确之处。

背景

这个资源集合应对了大语言模型(LLM)开发和部署中日益增长的结构化教育资源需求。随着LLM从研究工件过渡到生产系统,从业者需要掌握整个技术栈的深入知识——从GPU编程和变换器架构到分布式训练和推理优化。来自CMU及相关机构的这些课程代表了LLM教育的成熟发展,在理论基础与大规模系统工程问题之间架起了桥梁。

关键洞察

  • 课程分化反映了专业化路径:CMU LLM应用课程强调提示工程、RAG系统和特定领域应用(医疗保健、代码生成),而LLM系统课程深入探讨GPU内核优化、分布式训练策略(Megatron-LM、ZeRO)和服务基础设施(vLLM、FlashAttention)。这种分化反映了行业角色差异——应用工程师编排LLM,而系统工程师使其在计算上可行。

  • 硬件-算法协同设计成为核心能力:多个课程大纲特别邀请了基础系统创始人进行讲座:Tri Dao讲FlashAttention、Woosuk Kwon讲vLLM的PagedAttention、Hao Zhang讲DistServe。这表明现代LLM工作需要同时理解内存层次结构和注意力机制——算法改进与硬件约束密不可分。

  • 从单体模型到模块化架构:从基础变换器到专家混合模型(DeepSeek-MoE)、分解服务(DistServe)和检索增强的进展,反映了该领域向可组合系统的转变。LLM推理课程可能会进一步扩展到推理特定的优化,如推测解码和KV缓存管理。

待解问题

  • 随着LLM架构不断演进,课程应如何平衡经典ML理论的深度与实践系统优化?今天的FlashAttention会成为明天的过时技术吗?
  • 什么样的教学方法能最好地为学生准备应对学术研究与生产部署之间的滞后,特别是当行业系统(SGLang、vLLM)的进度快于发表周期时?
idea想法 2026-02-17 19:57:20

The Cost of Staying: Tech Career Timing留任的代价:科技职业时机选择

The Cost of Staying

by Amy Tam https://x.com/amytam01/status/2023593365401636896

Every technical person I know is doing the same math right now. They won’t call it that. They’ll say they’re “exploring options” or “thinking about what’s next.” But underneath, it’s the same calculation: how much is it costing me to stay where I am?

Not in dollars. In time. There’s a feeling in the air that the window for making the right move is shrinking—that every quarter you spend in the wrong seat, the gap between you and the people who moved earlier gets harder to close. A year ago, career decisions in tech felt reversible. Take the wrong job, course correct in eighteen months. That assumption is breaking down. The divergence between people who repositioned early and those still weighing their options is becoming visible, and it’s accelerating.

I see this up close. I’m an investor at Bloomberg Beta, and I spend most of my time with people in transition: leaving roles, finishing programs, deciding what’s next. I’m not a career advisor, but I sit at the intersection of “what are you leaving” and “what are you chasing.”

The valuable skill in tech shifted from “can you solve this problem” to “can you tell which problems are worth solving and which solutions are actually good.” The scarce thing flipped from execution to judgment: can you orchestrate systems, run parallel bets, and have the taste to know which results matter? The people who figured this out early are on one arm of a widening K-curve. Everyone else is getting faster at things that are about to be done for them.

The shift from execution to judgment is happening everywhere, but the cost of staying and the upside of moving look completely different depending on where you’re sitting.

FAANG

Here’s the tradeoff people at big tech companies are running right now: the systems are built, the comp is great, and the work is… fine. You’re increasingly reviewing AI-generated outputs rather than building from scratch. For some people, that’s a gift—it’s leverage, it’s sustainable, it’s a good life. The tradeoff is that “fine” has a cost that doesn’t show up in your paycheck.

The people leaving aren’t unhappy. They’re restless. They describe this specific feeling: the hardest problems aren’t here anymore, and the organization hasn’t caught up to that fact. The ones staying are making a bet that stability and comp are worth more than being close to the frontier. The ones leaving are making a bet that the frontier is where the next decade of career value gets built, and every quarter they wait is a quarter of compounding they miss.

Both bets are rational. But only one of them is time-sensitive.

Quant

Quant still works. Absurd pay, hard problems, immediate feedback. If you’re good, you know you’re good, because the P&L doesn’t lie.

The tradeoff that’s emerging: the entire quant toolkit (ML infrastructure, data obsession, statistical intuition) turns out to be exactly what AI labs and research startups need—same muscle, different problem. The difference is surface area. In quant, you’re optimizing a strategy. In AI, you’re building systems that reason. Even the quant-adjacent world is feeling it: the most interesting work in prediction markets and stablecoins is increasingly an AI infrastructure problem. One has a ceiling. The other doesn’t, or at least nobody’s found it yet.

Most quant people are staying, and they’re not wrong to. But the ones leaving describe something specific: they hit a point where the intellectual challenge of finance felt bounded in a way it didn’t before. They’re not chasing money. They’re chasing the feeling of working on something where the upper bound isn’t visible.

Academia

This is where the tradeoff is most painful, because it shouldn’t be a tradeoff at all.

Publishing novel results used to be the purest form of intellectual prestige. You did the work because the work was beautiful. That hasn’t changed. What changed is that the line between what you can do at a funded startup and what you can do in a university lab is blurring, and not in academia’s favor. A 20-person research startup can now do in a weekend what takes an academic lab a semester, because compute costs money that universities don’t have.

The most ambitious PhD students I talk to aren’t choosing between academia and industry. They’re choosing between theorizing about experiments and actually running them. The pull toward funded startups and labs isn’t about selling out. It’s about wanting to do the science, and the science requires resources that academia can’t provide.

The people staying in academia for the right reasons (open science, long time horizons, genuine intellectual freedom) are admirable. But they should know that the clock is ticking differently for them too: the longer the compute gap widens, the harder it becomes to do competitive work from inside a university.

AI Startups (Application Layer)

If you’re building products on top of models, you already know the feeling: the clever feature you shipped in March gets commoditized by a model update in June. The ground moves every quarter, and your moat evaporates.

The tradeoff here is between chasing what’s exciting and building what’s durable. The founders who are thriving right now stopped caring about model capabilities and started caring about the things models can’t take away: data moats, workflow capture, integration depth. It’s less fun to talk about at a dinner party. It’s where the actual companies get built.

The people making the sharpest moves in this world are the ones who got excited about plumbing—not the demo, not the pitch, not the capability. The ugly, boring infrastructure that makes a product sticky independent of which model sits underneath it.

Research Startups: The New Center of Gravity

This is where the K-curve is most visible.

Prime Intellect, SSI, Humans&—10-30 people doing genuine frontier research that competes with organizations fifty times their size. This would have been impossible three years ago. It’s happening now because the tools got good enough that a small number of people with great judgment can outrun a bureaucracy with more resources.

The daily workflow here is the clearest picture of what the upper arm looks like in practice. You’re kicking off training runs, spinning up experiments, letting things cook overnight. You come back in the morning, and your job isn’t to write code. It’s to know what to do with what came back—to have the taste to distinguish signal from noise when the system hands you a wall of results. It’s passive leverage. You set the experiments in motion, and the compounding happens whether or not you’re at your desk.

The tradeoff people are weighing: these companies are small, unproven, and many will fail. The bet is that being at the center of the frontier, with your judgment directly touching the work, compounds faster than the safety of a bigger organization, even if the specific company doesn’t make it. The skills transfer. The network transfers. The three years you spend reviewing someone else’s outputs at a big company don’t transfer the same way.

Big Model Labs: The Narrowing Frontier

The pitch “we’re building AGI” still works. It might always work on a certain type of person.

But the experience inside has shifted. The most interesting research is concentrated among a small number of senior people. Everyone else is doing important supporting work (evals, infra, product) that doesn’t feel like the frontier they signed up for. You joined to touch the thing, and you’re three layers removed from it.

The tradeoff is prestige versus proximity. A big lab on your resume still opens every door. But the people leaving are making a specific calculation: the resume value of “I was at [top lab]” is depreciating as the labs get bigger and more corporate, while the value of “I did frontier research at a place where my judgment shaped the direction” is appreciating. The window where big-lab pedigree is the best credential is closing, and the people who see it are moving.

The Clock

Every one of these tradeoffs has the same variable hiding inside it: time.

A year ago, you could sit in a comfortable seat and deliberate. The cost of waiting was low because the divergence was slow. That’s no longer true. The tools are compounding. The people who moved early are building on top of what they learned last quarter. The difference between someone who moved six months ago and someone still weighing their options is already compounding.

The upper arm isn’t closed. People are making the jump every week, and the people who are hiring them don’t care where you’ve been. They care whether you can do the work. But the math is directional: the longer you optimize for comfort, the more expensive the switch becomes—not because the opportunities disappear, but because the people who are already there are compounding, and you’re not.

The companies winning the talent war right now aren’t the ones with the best brand or the highest comp. They’re the ones where your judgment has the most surface area, where the distance between your taste and what actually gets built is zero, and where you’re surrounded by people who know things you don’t yet. The best people want to be close to others who have tricks they haven’t learned yet, at places with enough compute to actually run the experiments.

The question isn’t whether you’re smart enough. It’s that you’ve already done the math. You just haven’t acted on it.

The following content is generated by LLMs and may contain inaccuracies.

Context

This piece captures a structural shift in tech labor markets circa 2024–2025, where career optionality is compressing amid accelerating AI capabilities. It sits at the intersection of career dynamics, talent allocation theory, and the sociology of “frontier work.” The tension: traditional signals of career safety (FAANG comp, academic tenure, big lab prestige) are decoupling from proximity to where judgment-building happens. This matters because the shift from execution to orchestration—documented by economist David Autor as “task complementarity”—is happening faster than institutions can adapt, creating winner-take-most dynamics in skill accumulation.

Key Insights

The K-curve is a compounding divergence problem. Unlike previous tech cycles where skills depreciated gradually, generative AI tools create exponential productivity gaps between early adopters and laggards. Research from MIT and Stanford shows consultants using GPT-4 completed tasks 25% faster with 40% higher quality—but the variance between users widened over time. Those developing “judgment about AI outputs” compound that advantage quarterly; those executing manually fall behind non-linearly. The piece’s insight about research startups outrunning labs 50× their size reflects Coase’s theory of firm boundaries inverting: coordination costs collapsed faster than resource advantages matter.

Academia’s compute gap is a resource curse in reverse. The observation about weekend experiments versus semester timelines maps onto Brown et al.’s analysis of compute inequality in AI research. Universities can’t compete on infrastructure, but the piece misses that top labs are increasingly restricting publication to protect competitive moats—academic freedom still trades at a premium for reproducible, open work. The real cost: PhD students now optimize for “access to compute” over “intellectual community,” potentially sacrificing the collaborative serendipity that historically generated breakthrough ideas.

Open Questions

Could the K-curve collapse if AI tool improvements plateau, returning advantage to institutional stability? Or are we seeing a permanent regime change where “taste for orchestrating AI systems” becomes the dominant filter for knowledge work?

If judgment compounds faster than execution devalues, what happens to the bottom 50% of current tech workers—and does this finally force a reckoning with tech’s meritocracy mythology?

留任的代价

作者:Amy Tam https://x.com/amytam01/status/2023593365401636896

我认识的每一位技术人士现在都在做同样的数学计算。他们不会这样说。他们会说自己在"探索选择"或"思考下一步"。但本质上,这是同一个计算:留在原地要花费我多少?

不是金钱。而是时间。有一种感觉在空中弥漫:做出正确选择的窗口在缩小——你在错误岗位上待的每个季度,你和那些早期转身的人之间的差距就变得更难以弥补。一年前,科技行业的职业决策似乎是可逆的。接了个错误的工作,十八个月内调整方向就行。这个假设正在瓦解。早期重新定位的人和仍在权衡选择的人之间的分化变得可见,而且在加速。

我近距离看到这一点。我是Bloomberg Beta的投资者,大部分时间都与处于过渡期的人接触:离职、完成计划、决定下一步。我不是职业顾问,但我坐在"你要离开什么"和"你在追逐什么"的交叉口。

科技行业的宝贵技能从"你能解决这个问题吗"转变为"你能判断哪些问题值得解决,哪些解决方案真正有效吗"。稀缺的东西从执行力翻转到判断力:你能编排系统、并行下注,并具有品味来判断哪些结果重要吗?那些早期弄清楚这一点的人站在不断扩大的K曲线的一臂上。其他所有人都在快速提升那些即将被自动完成的东西的能力。

从执行到判断的转变无处不在,但留任的代价和转身的上升空间看起来完全取决于你所处的位置。

FAANG

这是大科技公司人员现在的权衡:系统已构建,薪酬很好,工作是……还可以。你越来越多地审查AI生成的输出,而不是从零开始构建。对某些人来说,这是礼物——这是杠杆、可持续性、美好生活。权衡是"还可以"有一个不会出现在你薪资单上的代价。

离职的人并不是不开心。他们坐立不安。他们描述这种特定的感觉:最难的问题已经不在这里了,而组织还没有认识到这一点。留下来的人是在打赌稳定性和薪酬比接近前沿更有价值。离开的人是在打赌前沿是下一个十年职业价值的构建之地,他们等待的每个季度都是他们错失的复合增长季度。

两个赌注都是理性的。但只有其中一个具有时间敏感性。

量化投资

量化投资仍然有效。荒谬的薪酬、困难的问题、即时反馈。如果你很优秀,你就知道自己很优秀,因为损益表不会说谎。

正在出现的权衡:整个量化工具包(ML基础设施、数据迷恋、统计直觉)正好是AI实验室和研究初创公司所需的——相同的肌肉、不同的问题。区别在于表面积。在量化投资中,你优化一个策略。在AI中,你构建能够推理的系统。即使是与量化相关的世界也在感受这一点:预测市场和稳定币中最有趣的工作越来越多地是AI基础设施问题。一个有上限。另一个没有,或者至少还没有人找到。

大多数量化人才留了下来,他们没有错。但离开的人描述了一些具体的东西:他们到达了一个点,金融的智力挑战感觉到了界限,这在以前没有。他们不是在追逐金钱。他们在追逐在做某件事的感觉,其中上界是不可见的。

学术界

这是权衡最痛苦的地方,因为根本不应该有权衡。

发表新颖结果曾经是最纯粹的智力声望形式。你做工作是因为工作很美妙。这没有改变。改变的是,你在资金充足的初创公司和大学实验室中能做什么之间的界线变得模糊,而且对学术界不利。一个20人的研究初创公司现在可以在一个周末做的工作,需要一个学术实验室一个学期,因为计算成本高昂,而大学没有这样的资金。

我交谈过的最雄心勃勃的博士生不是在学术界和产业之间选择。他们在理论化实验和实际运行实验之间选择。对资金充足的初创公司和实验室的吸引力不是关于妥协。这是关于想做科学,而科学需要学术界无法提供的资源。

因为正确的原因留在学术界的人(开放科学、长期视野、真正的学术自由)是令人敬佩的。但他们应该知道,时钟对他们的嘀嗒也不同:计算差距越长,从大学内部做有竞争力的工作就越难。

AI初创公司(应用层)

如果你在模型之上构建产品,你已经知道那种感觉:你在三月份推出的聪明功能在六月份被模型更新商品化了。地形每个季度都在移动,你的护城河蒸发了。

这里的权衡是追逐令人兴奋的东西和构建持久的东西之间的权衡。现在蓬勃发展的创始人停止关心模型能力,开始关心模型无法夺走的东西:数据护城河、工作流捕获、集成深度。在宴会上谈论这些就没那么有趣了。这是真正的公司被构建的地方。

在这个世界里做出最尖锐举动的人是那些对管道感到兴奋的人——不是演示、不是宣传、不是能力。丑陋、无聊的基础设施使产品粘性独立于坐在下面的模型。

研究初创公司:重力的新中心

这是K曲线最可见的地方。

Prime Intellect、SSI、Humans&——10-30人进行真正的前沿研究,与规模大五十倍的组织竞争。三年前这是不可能的。现在发生是因为工具足够好,少数具有高明判断力的人可以跑赢拥有更多资源的官僚机构。

这里的日常工作流程是上臂在实践中看起来最清晰的画面。你在启动训练运行、旋转实验、让事情一夜间进行。你早上回来,你的工作不是编写代码。这是知道如何处理返回的东西——当系统给你一堵结果时,具有品味来区分信号和噪音。这是被动杠杆。你设置实验运行,复合增长是否发生,不管你是否在办公桌前。

人们在权衡:这些公司很小、未经证实,许多会失败。打赌是在前沿中心,你的判断直接接触工作,复合速度比大型组织的安全更快,即使特定公司没有成功。技能转移。网络转移。你在大公司审查他人输出花费的三年不会以相同的方式转移。

大模型实验室:前沿变窄

“我们在构建AGI"的宣传仍然有效。它可能对某种类型的人总是有效。

但内部的体验已经转变。最有趣的研究集中在少数高级人员中。其他人都在做重要的支持工作(评估、基础设施、产品),感觉不像他们注册的前沿。你加入是为了接触这件事,你距离它有三层。

权衡是声望对邻近。大实验室在你的简历上仍然可以打开所有大门。但离开的人在做一个具体的计算:“我在[顶级实验室]“的简历价值随着实验室变得更大和更公司化而贬值,而"我在一个我的判断塑造方向的地方进行前沿研究"的价值在升值。大实验室血统是最佳证书的窗口正在关闭,看到它的人在转身。

时钟

这些权衡中的每一个都在其中隐藏着相同的变量:时间。

一年前,你可以坐在舒适的座位上深思熟虑。等待的代价很低,因为分化很慢。那不再是真的了。工具在复合。早期转身的人正在建立他们上个季度学到的东西。有人六个月前转身和有人仍在权衡选择之间的差异已经在复合。

上臂没有关闭。人们每周都在跳跃,雇用他们的人不关心你去过哪里。他们关心你是否能完成工作。但数学是方向性的:你优化舒适的时间越长,转换变得越昂贵——不是因为机会消失,而是因为已经到达那里的人在复合,而你没有。

现在赢得人才战争的公司不是那些品牌最好或薪酬最高的公司。他们是那些你的判断有最大表面积的地方,你的品味和实际构建的距离为零,你被你还没学过技巧的人包围的地方。最优秀的人想靠近其他拥有他们还没学过技巧的人,在有足够计算实际运行实验的地方。

问题不是你是否足够聪明。这是你已经做了数学。你只是还没有采取行动。

以下内容由 LLM 生成,可能包含不准确之处。

背景

这篇文章捕捉了科技劳动力市场在2024-2025年左右的结构性转变,在加速的人工智能能力中职业选择空间在压缩。它位于职业动态、人才配置理论和"前沿工作"社会学的交汇点。核心矛盾在于:传统的职业安全信号(FAANG薪酬、学术终身教职、大型实验室声誉)正在与判断力养成发生的地方脱钩。这很重要,因为从执行到协调的转变——由经济学家大卫·奥特记录为"任务互补性"——正在以制度适应的速度更快地发生,在技能积累中创造赢家通吃的动态。

关键洞见

K形曲线是一个复合性分化问题。 与以往科技周期中技能逐步贬值不同,生成式人工智能工具在早期采用者和落后者之间创造了指数级的生产力差距。麻省理工学院和斯坦福大学的研究表明,使用GPT-4的顾问完成任务的速度快25%,质量高40%——但用户之间的差异随时间扩大。那些开发出"关于人工智能输出判断力"的人每季度都在复合优势;那些手动执行的人落后的速度是非线性的。这篇文章关于研究初创企业超越其规模50倍实验室的观点反映了科斯的企业边界理论的反转:协调成本的下降速度比资源优势重要得多。

学术界的计算能力差距是反向的资源诅咒。 关于周末实验对比学期时间表的观察映射到Brown等人对人工智能研究中计算不平等的分析。大学无法在基础设施上竞争,但这篇文章没有注意到顶级实验室越来越限制出版以保护竞争优势——学术自由仍然对可复现的开放工作享有溢价。真正的代价:博士生现在为"获取计算能力"而不是"知识社群"进行优化,可能牺牲了历史上产生突破性想法的协作意外收获。

开放问题

如果人工智能工具改进进入平台期,K形曲线会崩溃吗,让优势回到机构稳定性?还是我们正在看到一个永久的政权转变,其中"协调人工智能系统的品味"成为知识工作的主导过滤器?

如果判断力的复合速度比执行贬值更快,当前科技工作者中的底部50%会发生什么——这最终是否会迫使对科技的精英统治神话进行清算?

idea想法 2026-02-17 19:54:22

Guidelines for Reporting LLM-Integrated Systems in HCI ResearchHCI研究中大语言模型集成系统的报告指南

Useful resources:

  • https://huggingface.co/spaces/transformers-community/Transformers-tenets
  • https://ianarawjo.github.io/Guidelines-for-Reporting-LLM-Integrated-Systems-in-HCI/

The following content is generated by LLMs and may contain inaccuracies.

Context

As LLM-integrated systems proliferate in HCI research, the field faces a reproducibility and transparency crisis. Unlike traditional computational systems, LLMs introduce nondeterminism, opaque prompting strategies, and rapidly shifting model capabilities that make replication difficult. The Guidelines for Reporting LLM-Integrated Systems address this gap by proposing standardized reporting practices for HCI researchers working with language models. This matters now because peer review struggles to evaluate systems where critical implementation details—prompt engineering, model versions, failure modes—are often omitted or underspecified.

Key Insights

1. Methodological Debt in Prompt Engineering
HCI research increasingly treats prompts as implementation details rather than experimental variables. Yet prompt design critically shapes user experience and system behavior. The guidelines advocate reporting not just final prompts but also iteration processes and sensitivity analysis. This aligns with calls in Transformers library development to “maintain the unmaintainable”—documenting messy development realities rather than sanitized outcomes. Without prompt versioning and ablation studies, findings remain unreproducible.

2. The Model Specification Problem
Generic references to “GPT-4” or “Claude” mask enormous variance. Model snapshots, temperature settings, and API versioning produce materially different behaviors. Research on model drift shows performance degradation over time even for fixed model names. The guidelines recommend timestamped model identifiers and capturing API responses for post-hoc analysis—a practice standard in ML benchmarking but rare in HCI evaluation.

3. Failure Mode Documentation as Design Knowledge
Traditional HCI reporting emphasizes successful interactions; LLM systems demand documenting characteristic failures. Hallucinations, context window limitations, and reasoning breakdowns aren’t bugs but inherent properties. Systematic failure taxonomies (as proposed in the guidelines) transform error cases into reusable design knowledge, enabling cumulative progress rather than repeated rediscovery.

Open Questions

  • How can HCI adapt peer review timelines when model obsolescence occurs faster than publication cycles—should we move toward living documentation standards?
  • What minimal reproducibility threshold balances scientific rigor with protecting proprietary prompt engineering that confers competitive advantage in applied contexts?

有用的资源:

  • https://huggingface.co/spaces/transformers-community/Transformers-tenets
  • https://ianarawjo.github.io/Guidelines-for-Reporting-LLM-Integrated-Systems-in-HCI/

以下内容由 LLM 生成,可能包含不准确之处。

背景

随着集成LLM的系统在人机交互研究中越来越普遍,该领域面临着可重复性和透明度危机。与传统计算系统不同,LLM引入了不确定性、不透明的提示策略以及快速变化的模型能力,这些都使得复现变得困难。LLM集成系统报告指南针对这一空白,为从事语言模型研究的人机交互研究人员提出了标准化报告实践。这在当下尤为重要,因为同行评审难以评估那些关键实现细节——提示工程、模型版本、故障模式——往往被省略或描述不清的系统。

主要见解

1. 提示工程中的方法债务
人机交互研究越来越多地将提示视为实现细节而非实验变量。然而,提示设计在很大程度上影响用户体验和系统行为。该指南主张报告不仅仅是最终提示,还要报告迭代过程和敏感性分析。这与Transformers库开发中"维护不可维护之物"的呼声相一致——记录凌乱的开发现实而非经过修饰的成果。没有提示版本控制和消融研究,研究成果将无法重现。

2. 模型规范问题
泛泛地提及"GPT-4"或"Claude"掩盖了巨大的差异。模型快照、温度设置和API版本控制产生的行为差异很大。关于模型漂移的研究表明,即使对于固定的模型名称,性能也会随时间降低。该指南建议使用带时间戳的模型标识符并捕获API响应以进行事后分析——这种做法在机器学习基准测试中是标准做法,但在人机交互评估中很少见。

3. 故障模式文档化作为设计知识
传统的人机交互报告强调成功的交互;LLM系统则需要文档化特征性故障。幻觉、上下文窗口限制和推理断裂不是漏洞,而是固有属性。系统性的故障分类法(如指南中提出的)将错误情况转化为可复用的设计知识,实现累进进步而非反复重新发现。

开放性问题

  • 当模型过时速度超过出版周期时,人机交互如何适应同行评审时间安排——我们是否应该转向活文档标准?
  • 科学严谨性与保护在应用环境中具有竞争优势的专有提示工程之间的最小可重复性阈值应该如何平衡?
1 2 3 4 5 6 7 8
© 2008 - 2026 Changkun Ou. All rights reserved.保留所有权利。 | PV/UV: /
0%