Changkun's Blog欧长坤的博客

Science and art, life in between.科学与艺术,生活在其间。

  • Home首页
  • Ideas想法
  • Posts文章
  • Tags标签
  • Bio关于
Changkun Ou

Changkun Ou

Human-AI interaction researcher, engineer, and writer.人机交互研究者、工程师、写作者。

Bridging HCI, AI, and systems programming. Building intelligent human-in-the-loop optimization systems. Informed by psychology, sociology, cognitive science, and philosophy.连接人机交互、AI 与系统编程。构建智能的人在环优化系统。融合心理学、社会学、认知科学与哲学。

Science and art, life in between.科学与艺术,生活在其间。

282 Blogs博客
171 Tags标签
Changkun's Blog欧长坤的博客
idea想法 2026-02-17 20:35:49

Comprehensive LLM Courses and Lecture Collections综合LLM课程和讲座合集

LLM Lectures

  • https://cmu-llms.org/schedule/
  • https://www.phontron.com/class/lminference-fall2025/schedule
  • https://llmsystem.github.io/llmsystem2026spring/docs/Syllabus
  • https://llmsystem.github.io/llmsystem2025spring/docs/Syllabus/

The following content is generated by LLMs and may contain inaccuracies.

Context

This collection addresses the growing need for structured educational resources in large language model (LLM) development and deployment. As LLMs transition from research artifacts to production systems, practitioners require deep understanding across the full stack—from GPU programming and transformer architecture to distributed training and inference optimization. These courses from CMU and related institutions represent the maturation of LLM education, bridging theoretical foundations with systems engineering concerns that arise at scale.

Key Insights

  • Curriculum divergence reflects specialization paths: The CMU LLM Applications course emphasizes prompt engineering, RAG systems, and domain-specific applications (healthcare, code generation), while the LLM Systems courses dive into GPU kernel optimization, distributed training strategies (Megatron-LM, ZeRO), and serving infrastructure (vLLM, FlashAttention). This split mirrors industry roles—application engineers who orchestrate LLMs versus systems engineers who make them computationally feasible.

  • Hardware-algorithm co-design emerges as core competency: Multiple syllabi feature guest lectures from creators of foundational systems: Tri Dao on FlashAttention, Woosuk Kwon on vLLM’s PagedAttention, Hao Zhang on DistServe. This signals that modern LLM work requires understanding memory hierarchies and attention mechanisms simultaneously—algorithmic improvements are inseparable from hardware constraints.

  • From monolithic models to modular architectures: The progression from basic transformers to mixture-of-experts (DeepSeek-MoE), disaggregated serving (DistServe), and retrieval augmentation reflects the field’s shift toward composable systems. The LLM Inference course likely extends this toward inference-specific optimizations like speculative decoding and KV cache management.

Open Questions

  • How should curricula balance depth in classical ML theory versus hands-on systems optimization as LLM architectures continue evolving? Will today’s FlashAttention become tomorrow’s deprecated technique?
  • What pedagogical approaches best prepare students for the lag between academic research and production deployment, especially when industry systems (SGLang, vLLM) advance faster than publication cycles?

LLM讲座

  • https://cmu-llms.org/schedule/
  • https://www.phontron.com/class/lminference-fall2025/schedule
  • https://llmsystem.github.io/llmsystem2026spring/docs/Syllabus
  • https://llmsystem.github.io/llmsystem2025spring/docs/Syllabus/

以下内容由 LLM 生成,可能包含不准确之处。

背景

这个资源集合应对了大语言模型(LLM)开发和部署中日益增长的结构化教育资源需求。随着LLM从研究工件过渡到生产系统,从业者需要掌握整个技术栈的深入知识——从GPU编程和变换器架构到分布式训练和推理优化。来自CMU及相关机构的这些课程代表了LLM教育的成熟发展,在理论基础与大规模系统工程问题之间架起了桥梁。

关键洞察

  • 课程分化反映了专业化路径:CMU LLM应用课程强调提示工程、RAG系统和特定领域应用(医疗保健、代码生成),而LLM系统课程深入探讨GPU内核优化、分布式训练策略(Megatron-LM、ZeRO)和服务基础设施(vLLM、FlashAttention)。这种分化反映了行业角色差异——应用工程师编排LLM,而系统工程师使其在计算上可行。

  • 硬件-算法协同设计成为核心能力:多个课程大纲特别邀请了基础系统创始人进行讲座:Tri Dao讲FlashAttention、Woosuk Kwon讲vLLM的PagedAttention、Hao Zhang讲DistServe。这表明现代LLM工作需要同时理解内存层次结构和注意力机制——算法改进与硬件约束密不可分。

  • 从单体模型到模块化架构:从基础变换器到专家混合模型(DeepSeek-MoE)、分解服务(DistServe)和检索增强的进展,反映了该领域向可组合系统的转变。LLM推理课程可能会进一步扩展到推理特定的优化,如推测解码和KV缓存管理。

待解问题

  • 随着LLM架构不断演进,课程应如何平衡经典ML理论的深度与实践系统优化?今天的FlashAttention会成为明天的过时技术吗?
  • 什么样的教学方法能最好地为学生准备应对学术研究与生产部署之间的滞后,特别是当行业系统(SGLang、vLLM)的进度快于发表周期时?
idea想法 2026-02-17 19:57:20

The Cost of Staying: Tech Career Timing留任的代价:科技职业时机选择

The Cost of Staying

by Amy Tam https://x.com/amytam01/status/2023593365401636896

Every technical person I know is doing the same math right now. They won’t call it that. They’ll say they’re “exploring options” or “thinking about what’s next.” But underneath, it’s the same calculation: how much is it costing me to stay where I am?

Not in dollars. In time. There’s a feeling in the air that the window for making the right move is shrinking—that every quarter you spend in the wrong seat, the gap between you and the people who moved earlier gets harder to close. A year ago, career decisions in tech felt reversible. Take the wrong job, course correct in eighteen months. That assumption is breaking down. The divergence between people who repositioned early and those still weighing their options is becoming visible, and it’s accelerating.

I see this up close. I’m an investor at Bloomberg Beta, and I spend most of my time with people in transition: leaving roles, finishing programs, deciding what’s next. I’m not a career advisor, but I sit at the intersection of “what are you leaving” and “what are you chasing.”

The valuable skill in tech shifted from “can you solve this problem” to “can you tell which problems are worth solving and which solutions are actually good.” The scarce thing flipped from execution to judgment: can you orchestrate systems, run parallel bets, and have the taste to know which results matter? The people who figured this out early are on one arm of a widening K-curve. Everyone else is getting faster at things that are about to be done for them.

The shift from execution to judgment is happening everywhere, but the cost of staying and the upside of moving look completely different depending on where you’re sitting.

FAANG

Here’s the tradeoff people at big tech companies are running right now: the systems are built, the comp is great, and the work is… fine. You’re increasingly reviewing AI-generated outputs rather than building from scratch. For some people, that’s a gift—it’s leverage, it’s sustainable, it’s a good life. The tradeoff is that “fine” has a cost that doesn’t show up in your paycheck.

The people leaving aren’t unhappy. They’re restless. They describe this specific feeling: the hardest problems aren’t here anymore, and the organization hasn’t caught up to that fact. The ones staying are making a bet that stability and comp are worth more than being close to the frontier. The ones leaving are making a bet that the frontier is where the next decade of career value gets built, and every quarter they wait is a quarter of compounding they miss.

Both bets are rational. But only one of them is time-sensitive.

Quant

Quant still works. Absurd pay, hard problems, immediate feedback. If you’re good, you know you’re good, because the P&L doesn’t lie.

The tradeoff that’s emerging: the entire quant toolkit (ML infrastructure, data obsession, statistical intuition) turns out to be exactly what AI labs and research startups need—same muscle, different problem. The difference is surface area. In quant, you’re optimizing a strategy. In AI, you’re building systems that reason. Even the quant-adjacent world is feeling it: the most interesting work in prediction markets and stablecoins is increasingly an AI infrastructure problem. One has a ceiling. The other doesn’t, or at least nobody’s found it yet.

Most quant people are staying, and they’re not wrong to. But the ones leaving describe something specific: they hit a point where the intellectual challenge of finance felt bounded in a way it didn’t before. They’re not chasing money. They’re chasing the feeling of working on something where the upper bound isn’t visible.

Academia

This is where the tradeoff is most painful, because it shouldn’t be a tradeoff at all.

Publishing novel results used to be the purest form of intellectual prestige. You did the work because the work was beautiful. That hasn’t changed. What changed is that the line between what you can do at a funded startup and what you can do in a university lab is blurring, and not in academia’s favor. A 20-person research startup can now do in a weekend what takes an academic lab a semester, because compute costs money that universities don’t have.

The most ambitious PhD students I talk to aren’t choosing between academia and industry. They’re choosing between theorizing about experiments and actually running them. The pull toward funded startups and labs isn’t about selling out. It’s about wanting to do the science, and the science requires resources that academia can’t provide.

The people staying in academia for the right reasons (open science, long time horizons, genuine intellectual freedom) are admirable. But they should know that the clock is ticking differently for them too: the longer the compute gap widens, the harder it becomes to do competitive work from inside a university.

AI Startups (Application Layer)

If you’re building products on top of models, you already know the feeling: the clever feature you shipped in March gets commoditized by a model update in June. The ground moves every quarter, and your moat evaporates.

The tradeoff here is between chasing what’s exciting and building what’s durable. The founders who are thriving right now stopped caring about model capabilities and started caring about the things models can’t take away: data moats, workflow capture, integration depth. It’s less fun to talk about at a dinner party. It’s where the actual companies get built.

The people making the sharpest moves in this world are the ones who got excited about plumbing—not the demo, not the pitch, not the capability. The ugly, boring infrastructure that makes a product sticky independent of which model sits underneath it.

Research Startups: The New Center of Gravity

This is where the K-curve is most visible.

Prime Intellect, SSI, Humans&—10-30 people doing genuine frontier research that competes with organizations fifty times their size. This would have been impossible three years ago. It’s happening now because the tools got good enough that a small number of people with great judgment can outrun a bureaucracy with more resources.

The daily workflow here is the clearest picture of what the upper arm looks like in practice. You’re kicking off training runs, spinning up experiments, letting things cook overnight. You come back in the morning, and your job isn’t to write code. It’s to know what to do with what came back—to have the taste to distinguish signal from noise when the system hands you a wall of results. It’s passive leverage. You set the experiments in motion, and the compounding happens whether or not you’re at your desk.

The tradeoff people are weighing: these companies are small, unproven, and many will fail. The bet is that being at the center of the frontier, with your judgment directly touching the work, compounds faster than the safety of a bigger organization, even if the specific company doesn’t make it. The skills transfer. The network transfers. The three years you spend reviewing someone else’s outputs at a big company don’t transfer the same way.

Big Model Labs: The Narrowing Frontier

The pitch “we’re building AGI” still works. It might always work on a certain type of person.

But the experience inside has shifted. The most interesting research is concentrated among a small number of senior people. Everyone else is doing important supporting work (evals, infra, product) that doesn’t feel like the frontier they signed up for. You joined to touch the thing, and you’re three layers removed from it.

The tradeoff is prestige versus proximity. A big lab on your resume still opens every door. But the people leaving are making a specific calculation: the resume value of “I was at [top lab]” is depreciating as the labs get bigger and more corporate, while the value of “I did frontier research at a place where my judgment shaped the direction” is appreciating. The window where big-lab pedigree is the best credential is closing, and the people who see it are moving.

The Clock

Every one of these tradeoffs has the same variable hiding inside it: time.

A year ago, you could sit in a comfortable seat and deliberate. The cost of waiting was low because the divergence was slow. That’s no longer true. The tools are compounding. The people who moved early are building on top of what they learned last quarter. The difference between someone who moved six months ago and someone still weighing their options is already compounding.

The upper arm isn’t closed. People are making the jump every week, and the people who are hiring them don’t care where you’ve been. They care whether you can do the work. But the math is directional: the longer you optimize for comfort, the more expensive the switch becomes—not because the opportunities disappear, but because the people who are already there are compounding, and you’re not.

The companies winning the talent war right now aren’t the ones with the best brand or the highest comp. They’re the ones where your judgment has the most surface area, where the distance between your taste and what actually gets built is zero, and where you’re surrounded by people who know things you don’t yet. The best people want to be close to others who have tricks they haven’t learned yet, at places with enough compute to actually run the experiments.

The question isn’t whether you’re smart enough. It’s that you’ve already done the math. You just haven’t acted on it.

The following content is generated by LLMs and may contain inaccuracies.

Context

This piece captures a structural shift in tech labor markets circa 2024–2025, where career optionality is compressing amid accelerating AI capabilities. It sits at the intersection of career dynamics, talent allocation theory, and the sociology of “frontier work.” The tension: traditional signals of career safety (FAANG comp, academic tenure, big lab prestige) are decoupling from proximity to where judgment-building happens. This matters because the shift from execution to orchestration—documented by economist David Autor as “task complementarity”—is happening faster than institutions can adapt, creating winner-take-most dynamics in skill accumulation.

Key Insights

The K-curve is a compounding divergence problem. Unlike previous tech cycles where skills depreciated gradually, generative AI tools create exponential productivity gaps between early adopters and laggards. Research from MIT and Stanford shows consultants using GPT-4 completed tasks 25% faster with 40% higher quality—but the variance between users widened over time. Those developing “judgment about AI outputs” compound that advantage quarterly; those executing manually fall behind non-linearly. The piece’s insight about research startups outrunning labs 50× their size reflects Coase’s theory of firm boundaries inverting: coordination costs collapsed faster than resource advantages matter.

Academia’s compute gap is a resource curse in reverse. The observation about weekend experiments versus semester timelines maps onto Brown et al.’s analysis of compute inequality in AI research. Universities can’t compete on infrastructure, but the piece misses that top labs are increasingly restricting publication to protect competitive moats—academic freedom still trades at a premium for reproducible, open work. The real cost: PhD students now optimize for “access to compute” over “intellectual community,” potentially sacrificing the collaborative serendipity that historically generated breakthrough ideas.

Open Questions

Could the K-curve collapse if AI tool improvements plateau, returning advantage to institutional stability? Or are we seeing a permanent regime change where “taste for orchestrating AI systems” becomes the dominant filter for knowledge work?

If judgment compounds faster than execution devalues, what happens to the bottom 50% of current tech workers—and does this finally force a reckoning with tech’s meritocracy mythology?

留任的代价

作者:Amy Tam https://x.com/amytam01/status/2023593365401636896

我认识的每一位技术人士现在都在做同样的数学计算。他们不会这样说。他们会说自己在"探索选择"或"思考下一步"。但本质上,这是同一个计算:留在原地要花费我多少?

不是金钱。而是时间。有一种感觉在空中弥漫:做出正确选择的窗口在缩小——你在错误岗位上待的每个季度,你和那些早期转身的人之间的差距就变得更难以弥补。一年前,科技行业的职业决策似乎是可逆的。接了个错误的工作,十八个月内调整方向就行。这个假设正在瓦解。早期重新定位的人和仍在权衡选择的人之间的分化变得可见,而且在加速。

我近距离看到这一点。我是Bloomberg Beta的投资者,大部分时间都与处于过渡期的人接触:离职、完成计划、决定下一步。我不是职业顾问,但我坐在"你要离开什么"和"你在追逐什么"的交叉口。

科技行业的宝贵技能从"你能解决这个问题吗"转变为"你能判断哪些问题值得解决,哪些解决方案真正有效吗"。稀缺的东西从执行力翻转到判断力:你能编排系统、并行下注,并具有品味来判断哪些结果重要吗?那些早期弄清楚这一点的人站在不断扩大的K曲线的一臂上。其他所有人都在快速提升那些即将被自动完成的东西的能力。

从执行到判断的转变无处不在,但留任的代价和转身的上升空间看起来完全取决于你所处的位置。

FAANG

这是大科技公司人员现在的权衡:系统已构建,薪酬很好,工作是……还可以。你越来越多地审查AI生成的输出,而不是从零开始构建。对某些人来说,这是礼物——这是杠杆、可持续性、美好生活。权衡是"还可以"有一个不会出现在你薪资单上的代价。

离职的人并不是不开心。他们坐立不安。他们描述这种特定的感觉:最难的问题已经不在这里了,而组织还没有认识到这一点。留下来的人是在打赌稳定性和薪酬比接近前沿更有价值。离开的人是在打赌前沿是下一个十年职业价值的构建之地,他们等待的每个季度都是他们错失的复合增长季度。

两个赌注都是理性的。但只有其中一个具有时间敏感性。

量化投资

量化投资仍然有效。荒谬的薪酬、困难的问题、即时反馈。如果你很优秀,你就知道自己很优秀,因为损益表不会说谎。

正在出现的权衡:整个量化工具包(ML基础设施、数据迷恋、统计直觉)正好是AI实验室和研究初创公司所需的——相同的肌肉、不同的问题。区别在于表面积。在量化投资中,你优化一个策略。在AI中,你构建能够推理的系统。即使是与量化相关的世界也在感受这一点:预测市场和稳定币中最有趣的工作越来越多地是AI基础设施问题。一个有上限。另一个没有,或者至少还没有人找到。

大多数量化人才留了下来,他们没有错。但离开的人描述了一些具体的东西:他们到达了一个点,金融的智力挑战感觉到了界限,这在以前没有。他们不是在追逐金钱。他们在追逐在做某件事的感觉,其中上界是不可见的。

学术界

这是权衡最痛苦的地方,因为根本不应该有权衡。

发表新颖结果曾经是最纯粹的智力声望形式。你做工作是因为工作很美妙。这没有改变。改变的是,你在资金充足的初创公司和大学实验室中能做什么之间的界线变得模糊,而且对学术界不利。一个20人的研究初创公司现在可以在一个周末做的工作,需要一个学术实验室一个学期,因为计算成本高昂,而大学没有这样的资金。

我交谈过的最雄心勃勃的博士生不是在学术界和产业之间选择。他们在理论化实验和实际运行实验之间选择。对资金充足的初创公司和实验室的吸引力不是关于妥协。这是关于想做科学,而科学需要学术界无法提供的资源。

因为正确的原因留在学术界的人(开放科学、长期视野、真正的学术自由)是令人敬佩的。但他们应该知道,时钟对他们的嘀嗒也不同:计算差距越长,从大学内部做有竞争力的工作就越难。

AI初创公司(应用层)

如果你在模型之上构建产品,你已经知道那种感觉:你在三月份推出的聪明功能在六月份被模型更新商品化了。地形每个季度都在移动,你的护城河蒸发了。

这里的权衡是追逐令人兴奋的东西和构建持久的东西之间的权衡。现在蓬勃发展的创始人停止关心模型能力,开始关心模型无法夺走的东西:数据护城河、工作流捕获、集成深度。在宴会上谈论这些就没那么有趣了。这是真正的公司被构建的地方。

在这个世界里做出最尖锐举动的人是那些对管道感到兴奋的人——不是演示、不是宣传、不是能力。丑陋、无聊的基础设施使产品粘性独立于坐在下面的模型。

研究初创公司:重力的新中心

这是K曲线最可见的地方。

Prime Intellect、SSI、Humans&——10-30人进行真正的前沿研究,与规模大五十倍的组织竞争。三年前这是不可能的。现在发生是因为工具足够好,少数具有高明判断力的人可以跑赢拥有更多资源的官僚机构。

这里的日常工作流程是上臂在实践中看起来最清晰的画面。你在启动训练运行、旋转实验、让事情一夜间进行。你早上回来,你的工作不是编写代码。这是知道如何处理返回的东西——当系统给你一堵结果时,具有品味来区分信号和噪音。这是被动杠杆。你设置实验运行,复合增长是否发生,不管你是否在办公桌前。

人们在权衡:这些公司很小、未经证实,许多会失败。打赌是在前沿中心,你的判断直接接触工作,复合速度比大型组织的安全更快,即使特定公司没有成功。技能转移。网络转移。你在大公司审查他人输出花费的三年不会以相同的方式转移。

大模型实验室:前沿变窄

“我们在构建AGI"的宣传仍然有效。它可能对某种类型的人总是有效。

但内部的体验已经转变。最有趣的研究集中在少数高级人员中。其他人都在做重要的支持工作(评估、基础设施、产品),感觉不像他们注册的前沿。你加入是为了接触这件事,你距离它有三层。

权衡是声望对邻近。大实验室在你的简历上仍然可以打开所有大门。但离开的人在做一个具体的计算:“我在[顶级实验室]“的简历价值随着实验室变得更大和更公司化而贬值,而"我在一个我的判断塑造方向的地方进行前沿研究"的价值在升值。大实验室血统是最佳证书的窗口正在关闭,看到它的人在转身。

时钟

这些权衡中的每一个都在其中隐藏着相同的变量:时间。

一年前,你可以坐在舒适的座位上深思熟虑。等待的代价很低,因为分化很慢。那不再是真的了。工具在复合。早期转身的人正在建立他们上个季度学到的东西。有人六个月前转身和有人仍在权衡选择之间的差异已经在复合。

上臂没有关闭。人们每周都在跳跃,雇用他们的人不关心你去过哪里。他们关心你是否能完成工作。但数学是方向性的:你优化舒适的时间越长,转换变得越昂贵——不是因为机会消失,而是因为已经到达那里的人在复合,而你没有。

现在赢得人才战争的公司不是那些品牌最好或薪酬最高的公司。他们是那些你的判断有最大表面积的地方,你的品味和实际构建的距离为零,你被你还没学过技巧的人包围的地方。最优秀的人想靠近其他拥有他们还没学过技巧的人,在有足够计算实际运行实验的地方。

问题不是你是否足够聪明。这是你已经做了数学。你只是还没有采取行动。

以下内容由 LLM 生成,可能包含不准确之处。

背景

这篇文章捕捉了科技劳动力市场在2024-2025年左右的结构性转变,在加速的人工智能能力中职业选择空间在压缩。它位于职业动态、人才配置理论和"前沿工作"社会学的交汇点。核心矛盾在于:传统的职业安全信号(FAANG薪酬、学术终身教职、大型实验室声誉)正在与判断力养成发生的地方脱钩。这很重要,因为从执行到协调的转变——由经济学家大卫·奥特记录为"任务互补性"——正在以制度适应的速度更快地发生,在技能积累中创造赢家通吃的动态。

关键洞见

K形曲线是一个复合性分化问题。 与以往科技周期中技能逐步贬值不同,生成式人工智能工具在早期采用者和落后者之间创造了指数级的生产力差距。麻省理工学院和斯坦福大学的研究表明,使用GPT-4的顾问完成任务的速度快25%,质量高40%——但用户之间的差异随时间扩大。那些开发出"关于人工智能输出判断力"的人每季度都在复合优势;那些手动执行的人落后的速度是非线性的。这篇文章关于研究初创企业超越其规模50倍实验室的观点反映了科斯的企业边界理论的反转:协调成本的下降速度比资源优势重要得多。

学术界的计算能力差距是反向的资源诅咒。 关于周末实验对比学期时间表的观察映射到Brown等人对人工智能研究中计算不平等的分析。大学无法在基础设施上竞争,但这篇文章没有注意到顶级实验室越来越限制出版以保护竞争优势——学术自由仍然对可复现的开放工作享有溢价。真正的代价:博士生现在为"获取计算能力"而不是"知识社群"进行优化,可能牺牲了历史上产生突破性想法的协作意外收获。

开放问题

如果人工智能工具改进进入平台期,K形曲线会崩溃吗,让优势回到机构稳定性?还是我们正在看到一个永久的政权转变,其中"协调人工智能系统的品味"成为知识工作的主导过滤器?

如果判断力的复合速度比执行贬值更快,当前科技工作者中的底部50%会发生什么——这最终是否会迫使对科技的精英统治神话进行清算?

idea想法 2026-02-17 19:54:22

Guidelines for Reporting LLM-Integrated Systems in HCI ResearchHCI研究中大语言模型集成系统的报告指南

Useful resources:

  • https://huggingface.co/spaces/transformers-community/Transformers-tenets
  • https://ianarawjo.github.io/Guidelines-for-Reporting-LLM-Integrated-Systems-in-HCI/

The following content is generated by LLMs and may contain inaccuracies.

Context

As LLM-integrated systems proliferate in HCI research, the field faces a reproducibility and transparency crisis. Unlike traditional computational systems, LLMs introduce nondeterminism, opaque prompting strategies, and rapidly shifting model capabilities that make replication difficult. The Guidelines for Reporting LLM-Integrated Systems address this gap by proposing standardized reporting practices for HCI researchers working with language models. This matters now because peer review struggles to evaluate systems where critical implementation details—prompt engineering, model versions, failure modes—are often omitted or underspecified.

Key Insights

1. Methodological Debt in Prompt Engineering
HCI research increasingly treats prompts as implementation details rather than experimental variables. Yet prompt design critically shapes user experience and system behavior. The guidelines advocate reporting not just final prompts but also iteration processes and sensitivity analysis. This aligns with calls in Transformers library development to “maintain the unmaintainable”—documenting messy development realities rather than sanitized outcomes. Without prompt versioning and ablation studies, findings remain unreproducible.

2. The Model Specification Problem
Generic references to “GPT-4” or “Claude” mask enormous variance. Model snapshots, temperature settings, and API versioning produce materially different behaviors. Research on model drift shows performance degradation over time even for fixed model names. The guidelines recommend timestamped model identifiers and capturing API responses for post-hoc analysis—a practice standard in ML benchmarking but rare in HCI evaluation.

3. Failure Mode Documentation as Design Knowledge
Traditional HCI reporting emphasizes successful interactions; LLM systems demand documenting characteristic failures. Hallucinations, context window limitations, and reasoning breakdowns aren’t bugs but inherent properties. Systematic failure taxonomies (as proposed in the guidelines) transform error cases into reusable design knowledge, enabling cumulative progress rather than repeated rediscovery.

Open Questions

  • How can HCI adapt peer review timelines when model obsolescence occurs faster than publication cycles—should we move toward living documentation standards?
  • What minimal reproducibility threshold balances scientific rigor with protecting proprietary prompt engineering that confers competitive advantage in applied contexts?

有用的资源:

  • https://huggingface.co/spaces/transformers-community/Transformers-tenets
  • https://ianarawjo.github.io/Guidelines-for-Reporting-LLM-Integrated-Systems-in-HCI/

以下内容由 LLM 生成,可能包含不准确之处。

背景

随着集成LLM的系统在人机交互研究中越来越普遍,该领域面临着可重复性和透明度危机。与传统计算系统不同,LLM引入了不确定性、不透明的提示策略以及快速变化的模型能力,这些都使得复现变得困难。LLM集成系统报告指南针对这一空白,为从事语言模型研究的人机交互研究人员提出了标准化报告实践。这在当下尤为重要,因为同行评审难以评估那些关键实现细节——提示工程、模型版本、故障模式——往往被省略或描述不清的系统。

主要见解

1. 提示工程中的方法债务
人机交互研究越来越多地将提示视为实现细节而非实验变量。然而,提示设计在很大程度上影响用户体验和系统行为。该指南主张报告不仅仅是最终提示,还要报告迭代过程和敏感性分析。这与Transformers库开发中"维护不可维护之物"的呼声相一致——记录凌乱的开发现实而非经过修饰的成果。没有提示版本控制和消融研究,研究成果将无法重现。

2. 模型规范问题
泛泛地提及"GPT-4"或"Claude"掩盖了巨大的差异。模型快照、温度设置和API版本控制产生的行为差异很大。关于模型漂移的研究表明,即使对于固定的模型名称,性能也会随时间降低。该指南建议使用带时间戳的模型标识符并捕获API响应以进行事后分析——这种做法在机器学习基准测试中是标准做法,但在人机交互评估中很少见。

3. 故障模式文档化作为设计知识
传统的人机交互报告强调成功的交互;LLM系统则需要文档化特征性故障。幻觉、上下文窗口限制和推理断裂不是漏洞,而是固有属性。系统性的故障分类法(如指南中提出的)将错误情况转化为可复用的设计知识,实现累进进步而非反复重新发现。

开放性问题

  • 当模型过时速度超过出版周期时,人机交互如何适应同行评审时间安排——我们是否应该转向活文档标准?
  • 科学严谨性与保护在应用环境中具有竞争优势的专有提示工程之间的最小可重复性阈值应该如何平衡?
idea想法 2026-02-17 19:13:48

Bayesian Methods, RLHF, and Building AI Models贝叶斯方法、RLHF和构建AI模型

Interesting books for reading:

  • 贝叶斯方法与科学合理性——对休谟问题的思考: https://book.douban.com/subject/4472081/
  • Reinforcement Learning from Human Feedback - A short introduction to RLHF and post-training focused on language models: https://rlhfbook.com/
  • Build a Reasoning Model (From Scratch): https://www.manning.com/books/build-a-reasoning-model-from-scratch
  • Build a Large Language Model (From Scratch): https://www.manning.com/books/build-a-large-language-model-from-scratch

The following content is generated by LLMs and may contain inaccuracies.

Context

This collection spans foundational epistemology, modern alignment techniques, and practical AI implementation—reflecting the field’s dual nature as both engineering practice and philosophical inquiry. The juxtaposition is telling: Bayesian reasoning connects to centuries-old debates about induction (Hume’s problem), while RLHF and transformer architectures represent cutting-edge methods for aligning LLMs with human preferences. Together, they address AI’s central tension: building systems that are both technically capable and rationally justified in their behavior.

Key Insights

Epistemological foundations matter for alignment. The Chinese work on Bayesian methods and scientific rationality revisits Hume’s problem of induction—how we justify inferring general principles from finite observations. This isn’t purely academic: RLHF implicitly makes Bayesian updates about human preferences from limited feedback. Nathan Lambert’s RLHF book describes how post-training uses reward models trained on human comparisons to steer base models, but rarely interrogates the epistemological validity of learning “values” from sparse signals. The gap matters: if we can’t justify ordinary induction, justifying value alignment from few-shot preference data becomes even more precarious.

From-scratch implementations reveal architectural commitments. Raschka’s LLM book and its reasoning model companion emphasize implementing attention mechanisms and transformers without abstraction layers. This pedagogical approach exposes design choices often hidden in frameworks: why scaled dot-product attention, why layer normalization placement matters, how positional encodings shape what’s learnable. Understanding these details illuminates why certain alignment interventions (like RLHF fine-tuning) work—they exploit specific inductive biases already present in the architecture.

Open Questions

Can we formalize RLHF’s relationship to Bayesian belief updating in a way that makes its epistemological assumptions explicit? Does building reasoning models from scratch reveal fundamental architectural requirements for multi-step inference, or are current approaches contingent engineering solutions?

值得阅读的有趣书籍:

  • 贝叶斯方法与科学合理性——对休谟问题的思考: https://book.douban.com/subject/4472081/
  • Reinforcement Learning from Human Feedback - A short introduction to RLHF and post-training focused on language models: https://rlhfbook.com/
  • Build a Reasoning Model (From Scratch): https://www.manning.com/books/build-a-reasoning-model-from-scratch
  • Build a Large Language Model (From Scratch): https://www.manning.com/books/build-a-large-language-model-from-scratch

以下内容由 LLM 生成,可能包含不准确之处。

背景

这个集合涵盖了基础认识论、现代对齐技术和实际的人工智能实现——反映了该领域作为工程实践和哲学探究的双重性质。这种对比很能说明问题:贝叶斯推理与几个世纪以来关于归纳的辩论相关(休谟问题),而RLHF和Transformer架构代表了将大语言模型与人类偏好对齐的最尖端方法。它们共同解决了人工智能的核心矛盾:构建既在技术上有能力又在行为上有理性依据的系统。

关键见解

认识论基础对对齐很重要。 中文关于贝叶斯方法和科学理性的著作重新审视了休谟的归纳问题——我们如何从有限的观察中证明推断一般原则的合理性。这不是纯学术问题:RLHF隐含地从有限反馈中对人类偏好进行贝叶斯更新。Nathan Lambert的RLHF书籍描述了训练后如何使用基于人类比较训练的奖励模型来引导基础模型,但很少质疑从稀疏信号学习"价值观"的认识论有效性。这个差距很重要:如果我们无法证明普通归纳的合理性,那么从少量偏好数据中证明价值对齐的合理性就变得更加危险。

从零开始的实现揭示了架构承诺。 Raschka的大语言模型书籍及其推理模型伴读书强调在没有抽象层的情况下实现注意力机制和Transformer。这种教学方法暴露了通常隐藏在框架中的设计选择:为什么使用缩放点积注意力、为什么层归一化的位置很重要、位置编码如何塑造可学习的内容。理解这些细节阐明了为什么某些对齐干预(如RLHF微调)有效——它们利用了架构中已经存在的特定归纳偏差。

待解问题

我们能否以使其认识论假设明确的方式,将RLHF与贝叶斯信念更新的关系正式化?从零开始构建推理模型是否揭示了多步推理的基本架构要求,或者当前的方法只是偶然的工程解决方案?

idea想法 2026-02-17 08:22:37

Human-in-the-loop approval scaling in agentic systems代理系统中人在回路批准的可扩展性

In agentic applications, to ensure the agent actions executed in a safe and trusted manner, in particular write operations, we often use a human in the loop.

The native practice for the setup with one agentic loop + MCP tool servers, is to add human confirmation per tool call. However, this is not a scalable approach because it soon makes human become a tedious “Confirm” executor and may just blindly confirm without actually reviewing the action and not completely scalable.

One potential mitigation is to use diff based review approach, and providing summary of the ongoing executions and ask for confirm. However this remains non-scalable when agent attempts to do a huge diff (e.g. writing 10k lines of code and wants to commit), where human cannot review the whole thing efficiently.

Another idea is to prepare some sort of intent overview and contract to let user to review, but it seems generally hard to prepare tool call sequence in advance because it is non-deterministic and depending on the context. It can also happen that one tool call is catastrophic but missed.

The following content is generated by LLMs and may contain inaccuracies.

Context

Human-in-the-loop (HITL) approval for agentic systems addresses a fundamental tension in AI safety: balancing autonomy with control. As agents gain write permissions—whether modifying codebases, executing financial transactions, or altering production systems—the risk of cascading failures grows. Traditional per-action approval gates create approval fatigue, degrading the very oversight they’re meant to provide. This challenge intensifies as agents integrate with Model Context Protocol (MCP) tool servers, where tool composition can generate unbounded action sequences.

Key Insights

Hierarchical approval boundaries: Rather than uniform gating, systems could implement trust tiers based on reversibility and blast radius. Anthropic’s Constitutional AI work suggests learned policies can classify actions by consequence severity. Read operations and idempotent writes might auto-approve, while irreversible operations (deletions, external API calls) trigger review. This mirrors capability-based security patterns where permissions are granular rather than binary.

Semantic compression for review: The 10k-line diff problem isn’t unique to agents—code review research tackles this via change impact analysis. Agents could pre-compute intent summaries using formal specifications or property-based testing. Instead of reviewing raw diffs, humans approve high-level invariants (“maintains API compatibility,” “preserves data integrity”). Microsoft’s Copilot Workspace experiments with this by generating editable task plans before execution.

Auditable sandboxing with rollback: Non-determinism makes pre-approval contracts fragile, but post-hoc auditing with cheap rollback changes the calculus. Systems like Deno’s permission model prove that runtime permission prompts can work when paired with clear scope boundaries. For agents, execution in isolated environments with speculative checkpointing lets humans review outcomes rather than intentions, then commit or revert atomically.

Open Questions

  • Can we develop a “differential trust calculus” that dynamically adjusts approval thresholds based on agent track record, action reversibility, and environmental context, similar to credit scoring for automation?
  • What design patterns from transactional databases (two-phase commit, optimistic concurrency) could apply to multi-step agent workflows with deferred human approval gates?

在代理应用中,为了确保代理操作以安全和可信的方式执行,尤其是写操作,我们通常会使用人在回路的方式。

对于一个代理循环加 MCP 工具服务器的配置,最原始的做法是在每次工具调用时添加人工确认。然而这种方式不具备可扩展性,因为它很快就会让人变成一个乏味的"确认"执行者,可能会不加审查地盲目确认,完全无法扩展。

一种可能的缓解方案是使用基于差异的审查方法,提供执行摘要并请求确认。但当代理尝试生成巨大的差异(例如写一万行代码并想要提交)时,这种方式仍然不可扩展,因为人无法高效地审查全部内容。

另一个想法是准备某种意图概览和合约让用户审查,但由于工具调用序列是非确定性的且依赖上下文,通常很难提前准备。也可能出现某个灾难性的工具调用被遗漏的情况。

以下内容由 LLM 生成,可能包含不准确之处。

背景

人在回路(HITL)批准对于代理系统解决了AI安全中的一个基本矛盾:平衡自主性与控制。当代理获得写入权限——无论是修改代码库、执行金融交易还是改变生产系统——级联故障的风险都会增长。传统的逐个操作批准门控会导致批准疲劳,削弱了它们本应提供的监督。当代理与模型上下文协议(MCP)工具服务器集成时,这一挑战会加剧,因为工具组合可以生成无限的操作序列。

关键洞见

分层批准边界:与其采用统一的门控,系统可以基于可逆性和影响范围实现信任层级。Anthropic的宪法AI工作表明,学习策略可以按后果严重程度对操作进行分类。读取操作和幂等写入可能会自动批准,而不可逆操作(删除、外部API调用)会触发审查。这反映了基于能力的安全模式,其中权限是精细化而非二进制的。

用于审查的语义压缩:万行代码差异问题不仅限于代理——代码审查研究通过变更影响分析来解决这个问题。代理可以使用形式化规范或基于属性的测试预先计算意图摘要。与其审查原始差异,人类可以批准高级不变量(“维护API兼容性”、“保留数据完整性”)。微软的Copilot工作区通过在执行前生成可编辑的任务计划来尝试这种方法。

具有回滚功能的可审计沙箱:非确定性使得预批准合同变得脆弱,但带有廉价回滚的事后审计改变了成本效益计算。Deno的权限模型等系统证明,当与清晰的作用域边界配对时,运行时权限提示可以有效。对于代理,在隔离环境中执行并结合推测性检查点让人类审查结果而非意图,然后原子性地提交或回滚。

开放问题

  • 我们能否开发一种"差异信任计算",根据代理的历史记录、操作可逆性和环境背景动态调整批准阈值,类似于自动化的信用评分?
  • 事务数据库中的哪些设计模式(两阶段提交、乐观并发)可以应用于具有延迟人工批准门控的多步代理工作流?
idea想法 2026-02-17 08:14:44

Pareto Principles and Incomparability in Infinite Ethics无限伦理学中的帕累托原则与不可比性

Pareto Principles in Infinite Ethics Published: May 01, 2018

Recommended citation: Askell, Amanda. ‘Pareto Principles in Infinite Ethics.’ PhD thesis, New York University (2018). https://askell.io/files/Askell-PhD-Thesis.pdf

Summary: In this thesis I argue that ethical rankings of worlds that contain infinite levels of wellbeing ought to be consistent with the Pareto principle, which says that if two worlds contain the same agents and some agents are better off in the first world than they are in the second and no agents are worse off than they are in the second, then the first world is better than the second. I show that if we accept four axioms – the Pareto principle, transitivity, an axiom stating that populations of worlds can be permuted, and the claim that if the ‘at least as good as’ relation holds between two worlds then it holds between qualitative duplicates of this world pair – then we must conclude that there is ubiquitous incomparability between infinite worlds.

The following content is generated by LLMs and may contain inaccuracies.

Context

Amanda Askell’s thesis addresses a fundamental tension in population ethics: how to compare worlds with infinite populations or infinite welfare distributions. Classical utilitarian aggregation breaks down when summing infinite utilities, yet we still need principles to guide ethical decisions affecting potentially infinite futures. This matters for longtermism, existential risk prioritization, and any framework considering indefinitely large futures.

Key Insights

Incomparability as inevitable, not defective. Askell proves that accepting four seemingly minimal axioms—Pareto, transitivity, permutation invariance, and qualitative invariance—forces “ubiquitous incomparability” between infinite worlds. This isn’t a bug to be fixed through cleverer aggregation, but a structural feature of infinite ethics. The result parallels impossibility theorems in social choice: we cannot have all desirable properties simultaneously. Rather than abandoning comparability entirely, we must accept that some world-pairs lack ordinal rankings.

Pareto remains non-negotiable. Unlike other axioms that might be weakened, Askell defends Pareto as foundational: if world A is identical to world B except some individuals fare better in A and none fare worse, A must be better. Rejecting Pareto permits rankings that ignore individual welfare entirely—a violation of welfarism’s core commitment. This constrains which infinite-ethics frameworks remain viable; approaches that violate Pareto (like some overtaking criteria) lose moral standing even if they avoid incomparability.

Practical implications for decision-making. If incomparability is ubiquitous, how do we act? Askell’s framework suggests adopting permissibility frameworks rather than maximization: multiple infinite futures may be permissible if incomparable. This aligns with recent work on maximality in decision theory under incomplete preferences.

Open Questions

How should we prioritize between finite and infinite considerations when they conflict—does any finite welfare gain justify foregoing incomparably different infinite futures? Can bounded rationality constraints justify practically rejecting Pareto in infinite cases where verification is computationally infeasible?

无限伦理学中的帕累托原则 发表日期:2018年5月1日

推荐引用:Askell, Amanda. ‘Pareto Principles in Infinite Ethics.’ PhD thesis, New York University (2018). https://askell.io/files/Askell-PhD-Thesis.pdf

摘要:在这篇论文中,我论证了对包含无限福祉水平的世界的伦理排序应当与帕累托原则一致。帕累托原则认为,如果两个世界包含相同的主体,而在第一个世界中某些主体的境况优于第二个世界,且没有主体的境况劣于第二个世界,那么第一个世界优于第二个世界。我证明了如果我们接受四条公理——帕累托原则、传递性、一条关于世界人口可置换的公理、以及"至少同样好"关系在两个世界之间成立则它在该世界对的定性副本之间也成立——那么我们必须得出结论:无限世界之间存在普遍的不可比性。

以下内容由 LLM 生成,可能包含不准确之处。

背景

阿曼达·阿斯凯尔的论文论述了种群伦理学中的一个根本矛盾:如何比较具有无限人口或无限福利分布的世界。经典功利主义的聚合方法在对无限效用求和时会失效,然而我们仍然需要原则来指导可能影响无限未来的伦理决策。这对于长期主义、存在风险优先级排序以及任何考虑无限期宏大未来的框架都很重要。

核心洞察

不可比性是必然的,而非缺陷。 阿斯凯尔证明了接受四个看似最小化的公理——帕累托原则、传递性、排列不变性和定性不变性——会导致无限世界之间的"普遍不可比性"。这不是可以通过更巧妙的聚合方法来修复的bug,而是无限伦理学的结构特征。该结果与社会选择中的不可能性定理相似:我们不能同时具备所有理想属性。与其完全放弃可比性,我们必须接受某些世界对缺乏序数排名的事实。

帕累托原则不可协商。 与其他可能被削弱的公理不同,阿斯凯尔将帕累托原则视为基础性的:如果世界A与世界B相同,只是某些个体在A中状况更好,在B中没有人状况更差,那么A必定更优。拒绝帕累托原则会允许完全忽视个人福利的排名——这违反了福利主义的核心承诺。这限制了哪些无限伦理学框架仍然可行;违反帕累托原则的方法(如某些超越标准)即使避免了不可比性,也失去了道德上的立足点。

对决策的实际影响。 如果不可比性是普遍存在的,我们该如何行动?阿斯凯尔的框架建议采用许可性框架而非最大化:如果无限期未来是不可比的,多个可能是许可的。这与最近在不完全偏好下的最大性决策论研究相吻合。

开放问题

当有限和无限考量相冲突时,我们应如何在两者之间排序——任何有限福利收益是否都足以证明放弃无限期不可比未来的合理性?有界理性约束是否能够在无限情况下(其中验证在计算上不可行)为实际上拒绝帕累托原则辩护?

idea想法 2026-02-16 11:43:36

The Energy Dilemma and Historical Paradox of Intelligent Optimization智能优化的能量困境与历史悖论

Continuing from my previous thoughts.

We mentioned that the long tail may not necessarily be a bad thing. Similar perspectives have actually been proposed long ago—over-optimization weakens the ability to adapt to mutations, which comes from the theory of antifragility.

Actually, the leaders on the adaptation curve are, in a sense, a group of highly intelligent people. From historical experience, we can see that the more intelligent people become, the more they crave energy. Similarly, we can draw a parallel between this human process and artificial intelligence: once the reproductive cycle of intelligence is broken through, it becomes increasingly intelligent while consuming more and more energy.

At some point in the future, if these superintelligent entities are still alive, we can logically deduce that this form of intelligence would consume all available energy and be unable to continue existing.

From this perspective, what exactly is optimization? Is it necessarily always a good thing?

After sharing these thoughts with someone, they countered me with a question: this kind of technological progress seemingly has never occurred in the long course of history. I was momentarily at a loss and didn’t know how to respond.

The following content is generated by LLMs and may contain inaccuracies.

Context

This reflection touches on an energy sustainability paradox in AI safety research: the tension between optimizing efficiency and long-term adaptability. It connects Nassim Taleb’s antifragility theory (over-optimization reduces system resilience) with the resource consumption dilemma in the AI alignment field. Current large model training energy consumption exhibits exponential growth (GPT-3 training consumed 1287 MWh), transforming this question from philosophical speculation into engineering reality.

Key Insights

  1. Historical Analogy Breaking Points: Human civilization has indeed experienced localized collapses (Easter Island deforestation, Mayan civilization), but never a global “intelligence-energy death spiral.” The critical difference lies in the optimization speed AI might achieve, far exceeding biological evolution — Bostrom’s intelligence explosion theory suggests recursive self-improvement could accomplish in months what took humans a million years of intelligent development, while energy infrastructure response cycles are measured in decades.

  2. Multi-Objective Optimization Dilemma: Single-dimensional optimization (such as reasoning capability) necessarily sacrifices other dimensions (such as energy efficiency and robustness). Pareto efficiency frontier demonstrates that once a system reaches certain optimization limits, any further improvement requires trade-offs. Biological evolution’s retention of “suboptimal” diversity serves as a hedge against uncertainty — the “long-tail populations” you mention may become critical gene pools for species survival during environmental upheaval.

  3. Self-Limiting Energy Bottleneck: Landauer’s principle defines the thermodynamic lower bound of computation (each bit erasure must dissipate at least kT·ln2 energy). Even with perfect algorithmic optimization, physical laws will force intelligent agents to encounter hard limits on the energy-intelligence curve, potentially naturally producing an “optimization ceiling” rather than unlimited expansion.

Open Questions

  • Does an operational definition of “moderate optimization” exist — one that captures intelligence dividends while preserving adaptive redundancy? Can the exploration-exploitation balance from evolutionary algorithms translate into AI governance principles?
  • If future AI breaches energy constraints through discovering new physics (such as controlled nuclear fusion), does the original argument become invalid? Would this mean the issue is fundamentally a race between “optimization speed vs. resource acquisition innovation speed”?

接着上个想法继续说。

我们提到长尾可能不一定是坏事,类似的观点其实很早之前就有人提出过——过度优化会削弱适应突变的能力,这来自反脆弱性的理论。

其实适应性曲线的领头人群在某种意义上属于高智能人群。我们从历史经验可以看到,越是智能的人往往越渴望能量。同样的,我们可以将人类的这个过程类比到机器智能上:一旦突破了智能的繁衍循环,它就会越来越智能,且越来越消耗能量。

在未来某个时间节点上,如果这些超级智能体仍然活着,我们可以从逻辑上推演出这种智能会消耗所有能源而无法继续生存。

从这个角度来看,优化到底是什么?它是不是一定就是好事?

我把这些想法跟人分享之后,有人反问我:这种技术进步似乎在历史长河中从未发生过。我一时语塞,竟不知作何回答。

以下内容由 LLM 生成,可能包含不准确之处。

Context

这个思考触及人工智能安全研究中的能源可持续性悖论:优化效率与长期适应性的张力。它连接了Nassim Taleb的反脆弱性理论(过度优化降低系统韧性)与AI对齐领域的资源消耗难题。当前大模型训练能耗呈指数增长(GPT-3训练消耗1287 MWh),使这个问题从哲学思辨转向工程现实。

Key Insights

  1. 历史类比的断裂点:人类文明确实经历过局部崩溃(复活节岛森林耗竭、玛雅文明),但从未出现全球性"智能-能源死亡螺旋"。关键差异在于AI可能实现的优化速度远超生物演化——Bostrom的智能爆炸理论指出递归自我改进可能在数月内完成人类百万年的智能跃迁,而能源基础设施响应周期以十年计。

  2. 优化的多目标困境:单一维度优化(如推理能力)必然牺牲其他维度(如能效、鲁棒性)。Pareto效率前沿表明:当系统达到某种优化极限时,任何进一步改进都需要权衡取舍。生物进化保留"次优"多样性正是对冲不确定性——你提到的"长尾人群"在环境剧变时可能成为种群延续的关键基因库。

  3. 能源瓶颈的自我限制:Landauer极限定义了计算的热力学下界(每比特擦除至少耗散kT·ln2能量)。即使实现完美算法优化,物理定律也会强制智能体在能源-智能曲线上遭遇硬上限,可能自然产生"优化天花板"而非无限扩张。

Open Questions

  • 是否存在"适度优化"的可操作定义——既获得智能红利又保留适应冗余?进化算法中的exploration-exploitation平衡能否转化为AI治理原则?
  • 如果未来AI通过发现新物理学突破能源约束(如可控核聚变),原论证是否失效?这意味着问题本质是"优化速度 vs 资源获取创新速度"的竞赛?
idea想法 2026-02-16 11:23:56

Abandoned Skills in the Long Tail May Be a Lifeline for the Future长尾效应中被遗弃的技能可能是未来的救命稻草

Over the past few years, due to the need to migrate legacy systems in enterprise projects, I’ve actually developed little affection for the long tail effect. A project that should have taken only a few months ended up requiring years to complete due to user dependencies or stubborn system coupling.

However, my perspective on this has shifted recently due to AI’s rapid transformation. The long tail may not necessarily be a bad thing. The main reason is that I’ve come to realize: as AI gradually replaces human programming, people are losing technical skills and can no longer code with the familiarity they once had. Yet in reality, most people still live in what we might call the old era, manually and inefficiently completing various tasks. From this perspective, AI has become the new infrastructure. If some force causes this infrastructure to collapse, then those who have walked at the forefront of the era and thus lost their skills would paradoxically need to seek help from those still living in the old era to solve coding problems.

If those walking at the head of the adaptation curve ultimately fail to survive the long course of human history, then those who haven’t yet adapted to the future might instead forge an alternative path toward a different future.

From this perspective, I now better understand why in many science fiction works (such as Cloud Atlas), future humanity possesses advanced technology yet lives extremely primitive lives.

The following content is generated by LLMs and may contain inaccuracies.

Context

This idea touches on the “skill fragility paradox” in technological evolution—when a technology becomes infrastructure, the populations depending on it may become more vulnerable in the face of systemic risk. In the current rapid penetration of AI into programming, this problem extends from the engineering dilemma of legacy system migration to the evolutionary strategy of human skill composition. It echoes Nassim Taleb’s core argument about antifragility: excessive optimization weakens the ability to adapt to sudden change.

Key Insights

  1. Skill Redundancy as Survival Insurance — The phenomenon you’ve observed validates the “Collingridge dilemma”: technology is easy to modify early but its impacts are difficult to measure; once mature, impacts are clear but the technology becomes hard to change. Venkatesh Rao’s analysis on Ribbonfarm points out that “laggards” on the technology adoption curve actually maintain diversified skill repositories, which in black swan events can transform into critical advantages—analogous to how biodiversity contributes to ecosystem resilience.

  2. Hidden Dependencies of Infrastructure — AI-enabled programming is creating new forms of “technical debt.” When coding ability is outsourced to AI, we face not merely skill atrophy but the comprehension gaps created by cognitive offloading. Historical cases like GPS causing spatial cognition decline, or calculators affecting mental arithmetic abilities, demonstrate that convenience tools reshape rather than merely enhance human capabilities.

  3. Prophetic Insights in Science Fiction — The “Cloud Atlas”-style “high-tech primitivization” is not a paradox but an artistic expression of the risks of over-specialization. Joseph Tainter argues in The Collapse of Complex Societies that the maintenance costs of complex systems may ultimately exceed their marginal benefits, leading to simplification and reversion. Your legacy system migration predicament is, at the macro level, a microcosmic manifestation of this complexity trap.

Open Questions

  • As AI tools proliferate, which “inefficient” manual skills are worth deliberately preserving as strategic redundancy? How should we quantify the value of such insurance?
  • If we view the human skill ecosystem as an investment portfolio, how should the optimal “long-tail/head” allocation ratio dynamically adjust with the speed of technological change?

最近几年,因为企业项目需要迁移遗留系统,我对长尾效应反而没有太多好感。本来几个月就能完成的项目,却因为用户依赖或顽固的系统耦合,导致整个迁移需要数年才能完成。

不过,最近因为AI的快速变革,我对这件事的认识有所转变。长尾可能并不一定是坏事。主要原因是我意识到:AI逐步替代了人类编程,导致人们在技能上的丧失,不再能像从前那样熟悉地编码。但实际上大部分人还生活在所谓的旧时代,手动低效地完成各种任务。从这个角度看,AI成为了新的基础设施。如果某种力量导致这种基础崩塌,那么这些走在时代前沿因而丧失了技能的人,反而会因为缺乏解决编码问题的能力,转向寻求那些还留在旧时代的人的帮助。

如果这些走在适应曲线头部的人最终没能在人类历史的长河里生存下来,那些还没来得及适应未来的人反而另辟蹊径,走向其他的未来。

从这个意义上,我更能理解为什么在许多科幻作品(比如《云图》)中,未来人类掌握了许多尖端科技,反而过着极其原始的生活。

以下内容由 LLM 生成,可能包含不准确之处。

Context

这个想法触及技术演化中的"技能脆弱性悖论"——当某项技术成为基础设施后,依赖它的群体反而可能在系统性风险面前更加脆弱。在AI快速渗透编程领域的当下,这个问题从遗留系统迁移的工程困境延伸到了人类技能组合的演化策略。它呼应了Nassim Taleb关于反脆弱性的核心论点:过度优化会削弱适应突变的能力。

Key Insights

  1. 技能冗余作为生存保险 — 你观察到的现象印证了"Collingridge困境":技术早期易改但影响难测,成熟后影响明确却难以改变。Venkatesh Rao在Ribbonfarm的分析指出,技术采纳曲线上的"落后者"实际维持着多样化的技能库,这在黑天鹅事件中可能转化为关键优势——类似生物多样性对生态系统韧性的作用。

  2. 基础设施的隐性依赖 — AI编程正在创造新的"技术债务"形式。当编码能力外包给AI,我们面临的不仅是技能退化,更是认知卸载带来的理解断层。历史案例如GPS导致的空间认知能力下降,或计算器对心算能力的影响,都显示便利性工具会重塑而非仅仅增强人类能力。

  3. 科幻中的预言性洞察 — 《云图》式的"高科技原始化"并非矛盾,而是对过度专业化风险的艺术表达。Joseph Tainter在《复杂社会的崩溃》中论证,复杂系统的维护成本最终可能超过其边际收益,导致简化回归。你的遗留系统迁移困境,在宏观层面就是这种复杂性陷阱的微观映射。

Open Questions

  • 在AI工具普及的过程中,哪些"非效率"的手工技能值得刻意保留作为战略性冗余?如何量化这种保险的价值?
  • 如果把人类技能生态看作投资组合,最优的"长尾/头部"配置比例应该如何随技术变革速度动态调整?
idea想法 2026-02-16 08:01:45

Programmer Identity Threatened by AI AdvancementAI 进步威胁程序员身份

Lex Fridman (03:04:40) Yeah, there’s a couple of things I want to say. One is, I never thought that the thing I love doing would be the thing that gets replaced. You hear these stories about things like the steam engine. I’ve spent thousands of hours poring over code, pouring my heart and soul into it. Some of my most painful and happiest moments were alone behind a screen. I was an Emacs person for a long time—man, Emacs. And there’s an identity there, there’s meaning. When I walk about the world, I don’t say it out loud, but I think of myself as a programmer. And to have that possibly change in a matter of months…

https://lexfridman.com/peter-steinberger-transcript

The following content is generated by LLMs and may contain inaccuracies.

Context

This moment captures a profound shift in knowledge work: the dissolution of professional identity in real-time. Fridman’s vulnerability reflects a broader crisis facing programmers as AI coding agents rapidly automate tasks once considered deeply human. Unlike previous automation waves that displaced manual labor, LLMs threaten cognitive specialization—the very competencies that define “who we are” rather than just “what we do.” Research on professional identity shows that when core work activities become obsolete, individuals experience not just job insecurity but existential disruption, forcing renegotiation of self-concept and social positioning.

Key Insights

Fridman’s “thousands of hours” in Emacs represents what psychologists call identity-constitutive labor—work so integral to self-conception that its loss mirrors bereavement. Studies on technological displacement document similar patterns among craftspeople during industrialization: the pain stems less from lost income than from obsolescence of hard-won mastery. The “programmer identity” functioned as what sociologist Richard Sennett calls craftsman pride—status derived from specialized competence.

The speed (“a matter of months”) distinguishes this transition from historical precedents. Traditional career disruptions allowed generational adaptation; parents steered children away from declining trades. AI’s pace eliminates that buffer. Research on rapid deskilling shows compressed timelines trigger acute psychological distress and resistance, as individuals lack cultural scripts for graceful transitions when expertise evaporates mid-career.

Paradoxically, programmers may be uniquely equipped for this transition—their meta-skill is abstraction and tool-building. Studies of AI adoption suggest roles shift from implementation to orchestration: programming the programmers. The identity crisis may stem not from capability loss but from status anxiety: supervisory roles feel less “real” than hands-on coding.

Open Questions

If programming becomes prompting, does the new skill require comparable depth to command respect—or will it always feel like diluted expertise? What happens to communities (open source, Stack Overflow) built around shared struggle when struggle itself becomes obsolete?

Lex Fridman (03:04:40) 是的,我想说几件事。首先,我从未想过我热爱做的事会成为被取代的东西。你听过很多这样的故事,比如蒸汽机。我花了数千个小时钻研代码,把心血倾注其中。我最痛苦和最快乐的一些时刻都是独自坐在屏幕前度过的。我长期使用 Emacs——天哪,Emacs。这里面有一种身份认同,有意义。当我走在世界上时,我不会大声说出来,但我把自己看作一名程序员。而在短短几个月内可能失去这一身份…

https://lexfridman.com/peter-steinberger-transcript

以下内容由 LLM 生成,可能包含不准确之处。

背景

这一刻捕捉了知识工作中的深刻转变:职业身份在实时瓦解。弗里德曼的脆弱性反映了程序员面临的更广泛危机,因为AI编码代理正在快速自动化曾经被认为深深扎根于人类的任务。不同于以往针对体力劳动的自动化浪潮,大语言模型威胁的是认知专业化——定义"我们是谁"而非仅仅"我们做什么"的核心能力。关于职业身份的研究表明,当核心工作活动变得过时时,个人经历的不仅是工作不安全感,更是存在危机,迫使人们重新协商自我概念和社会地位。

关键洞察

弗里德曼在Emacs中花费的"数千小时"代表了心理学家所称的身份构成劳动——这种工作与自我认知如此紧密相连,以至于其丧失如同哀悼。关于技术替代的研究记录了工业化期间工匠的类似模式:痛苦源于失去的不是收入,而是来之不易的掌握能力的陈旧化。“程序员身份"发挥了社会学家理查德·桑内特所称的工匠自豪感的作用——这种地位来自专业能力。

速度的快乐(“数个月的时间”)将这一转变与历史先例区分开来。传统职业中断允许代际适应;父母会引导孩子远离衰落的行业。AI的速度消除了这个缓冲期。关于快速技能贬值的研究显示,压缩的时间表会引发急性心理困扰和抵触,因为个人在职业生涯中期缺乏优雅过渡的文化脚本。

矛盾的是,程序员可能特别善于应对这一转变——他们的元技能是抽象化和工具构建。AI采用研究表明角色从实现转向编排:对程序员编程。身份危机可能源于的不是能力丧失,而是地位焦虑:监督角色感觉不如动手编码那样"真实”。

悬而未决的问题

如果编程变成提示词,新技能是否需要可比的深度才能获得尊重——或者它会始终感觉像被稀释的专业知识?当基于共同奋斗构建的社区(开源、Stack Overflow)中的奋斗本身变得过时时,会发生什么?

idea想法 2026-02-16 06:54:47

A Must-Read List of Chinese Literary Classics值得一读的中文文学经典书单

It’s been a long time since I’ve read Chinese books. The following titles sound quite interesting and are worth reading when I have time:

  • The Distant Savior
  • The Ordinary World
  • When the Mountains Turn Red
  • The Horse Herder
  • Hibiscus Town

The following content is generated by LLMs and may contain inaccuracies.

Context

This reading list presents a specific narrative tradition in contemporary Chinese literature: one focused on the fate of ordinary people during periods of social transformation. Most of these works were created in the 1980s-90s, reflecting the transformative pains of Chinese society before and after the reform and opening-up. They collectively explore a central tension: the collision between individual ideals and the tide of the times—a tension that has formed a unique “scars-reflection-root-seeking” narrative spectrum in Chinese literature. Revisiting these works in today’s context of “involution” and value reconstruction may provide a historical depth of reference.

Key Insights

The Continuation of Rural Realism — The Ordinary World (Lu Yao, 1986-1988) and The Horse Herder demonstrate Chinese literature’s sustained attention to the themes of “land-labor-dignity,” a tradition traceable to Zhao Shuli and Liu Qing. Such works counter the singularity of urbanization narratives, proposing “the ordinary” itself as a possible existential philosophy—forming an Eastern dialogue with the “persistence amid absurdity” found in Western existentialist literature.

Literary Memory of Historical Trauma — Furong Town (Gu Hua, 1981) practices a form of “microhistorical writing” through a dual perspective of “political movement-daily life.” This resonates with the recent “turn toward everyday life” in historical sociology: how do grand narratives permeate and distort individual experience? The Distant Savior, though a commercial novel, attempts to graft Buddhist philosophical contemplation onto contemporary market logic, forming a kind of “worldly transcendence”—a contradictory tension itself worthy of critical interpretation.

The Absence of Gender Perspective — Notably, this reading list is dominated by male authors and male protagonists. Compared to works by Zhang Jie (The Heavy Wings), Wang Anyi, and Tie Ning from the same period, one can discover different dimensions of gendered experience in narratives of social change—how women are simultaneously historical objects yet attempt to become subjects.

Open Questions

How can contemporary readers avoid simplifying these works into “nostalgic texts” or an “aesthetics of suffering”? Do their insights into present dilemmas transcend the particularity of their historical contexts?

In a reading ecosystem dominated by algorithmic recommendation and short videos, what kind of renewed life can these “weighty narratives” demanding sustained immersion still achieve?

很久没有读过中文书籍了。以下这些书听起来很有意思,有时间的话值得一读:

  • 遥远的救世主
  • 平凡的世界
  • 等到满山红叶时
  • 牧马人
  • 芙蓉镇

以下内容由 LLM 生成,可能包含不准确之处。

Context

这份书单呈现了中国当代文学中一个特定的叙事传统:聚焦于社会变革时期普通人的命运。这些作品大多创作于20世纪80-90年代,反映了改革开放前后中国社会的转型阵痛。它们共同探讨一个核心张力:个体理想与时代洪流的碰撞——这在中国文学中形成了独特的"伤痕-反思-寻根"叙事谱系。在当下"内卷"与价值重构的语境中重读这些作品,或能提供历史纵深的参照。

Key Insights

乡土现实主义的延续 — 《平凡的世界》(路遥, 1986-1988)与《牧马人》展现了中国文学对"土地-劳动-尊严"主题的持续关注,这一传统可追溯至赵树理、柳青。这类作品抗衡城市化叙事的单一性,提出"平凡"本身作为一种存在哲学的可能性——与西方存在主义文学中"荒诞中的坚持"形成东方对话。

历史创伤的文学记忆 — 《芙蓉镇》(古华, 1981)通过"政治运动-日常生活"的双重视角,实践了一种"微观政治史"的写作。这与近年历史社会学的"日常生活转向"暗合:宏大叙事如何渗透、扭曲个体经验?《遥远的救世主》虽为商业小说,却试图将佛学思辨嫁接于当代市场逻辑,形成某种"入世的超越性"——这种矛盾张力本身值得警惕性解读。

性别视角的缺失 — 值得注意的是,这份书单以男性作家及男性主人公为主导。对比同时期张洁(《沉重的翅膀》)、王安忆、铁凝的作品,可发现性别经验在社会变革叙事中的不同维度——女性如何既是历史客体又试图成为主体。

Open Questions

当代读者如何避免将这些作品简化为"怀旧文本"或"苦难美学"?它们对当下困境的启示是否超越了历史情境的特殊性?

在算法推荐与短视频主导的阅读生态中,这种需要时间沉浸的"厚重叙事"还能获得怎样的新生命?

1 2 3 4 5 6 7 8
© 2008 - 2026 Changkun Ou. All rights reserved.保留所有权利。 | PV/UV: /
0%