Changkun's Blog欧长坤的博客

Science and art, life in between.科学与艺术,生活在其间。

  • Home首页
  • Ideas想法
  • Posts文章
  • Tags标签
  • Bio关于
Changkun Ou

Changkun Ou

Human-AI interaction researcher, engineer, and writer.人机交互研究者、工程师、写作者。

Bridging HCI, AI, and systems programming. Building intelligent human-in-the-loop optimization systems. Informed by psychology, sociology, cognitive science, and philosophy.连接人机交互、AI 与系统编程。构建智能的人在环优化系统。融合心理学、社会学、认知科学与哲学。

Science and art, life in between.科学与艺术,生活在其间。

276 Blogs博客
165 Tags标签
Changkun's Blog欧长坤的博客

Bayesian Methods, RLHF, and Building AI Models贝叶斯方法、RLHF和构建AI模型

Published at发布于:: 2026-02-17

Interesting books for reading:

  • 贝叶斯方法与科学合理性——对休谟问题的思考: https://book.douban.com/subject/4472081/
  • Reinforcement Learning from Human Feedback - A short introduction to RLHF and post-training focused on language models: https://rlhfbook.com/
  • Build a Reasoning Model (From Scratch): https://www.manning.com/books/build-a-reasoning-model-from-scratch
  • Build a Large Language Model (From Scratch): https://www.manning.com/books/build-a-large-language-model-from-scratch

The following content is generated by LLMs and may contain inaccuracies.

Context

This collection spans foundational epistemology, modern alignment techniques, and practical AI implementation—reflecting the field’s dual nature as both engineering practice and philosophical inquiry. The juxtaposition is telling: Bayesian reasoning connects to centuries-old debates about induction (Hume’s problem), while RLHF and transformer architectures represent cutting-edge methods for aligning LLMs with human preferences. Together, they address AI’s central tension: building systems that are both technically capable and rationally justified in their behavior.

Key Insights

Epistemological foundations matter for alignment. The Chinese work on Bayesian methods and scientific rationality revisits Hume’s problem of induction—how we justify inferring general principles from finite observations. This isn’t purely academic: RLHF implicitly makes Bayesian updates about human preferences from limited feedback. Nathan Lambert’s RLHF book describes how post-training uses reward models trained on human comparisons to steer base models, but rarely interrogates the epistemological validity of learning “values” from sparse signals. The gap matters: if we can’t justify ordinary induction, justifying value alignment from few-shot preference data becomes even more precarious.

From-scratch implementations reveal architectural commitments. Raschka’s LLM book and its reasoning model companion emphasize implementing attention mechanisms and transformers without abstraction layers. This pedagogical approach exposes design choices often hidden in frameworks: why scaled dot-product attention, why layer normalization placement matters, how positional encodings shape what’s learnable. Understanding these details illuminates why certain alignment interventions (like RLHF fine-tuning) work—they exploit specific inductive biases already present in the architecture.

Open Questions

Can we formalize RLHF’s relationship to Bayesian belief updating in a way that makes its epistemological assumptions explicit? Does building reasoning models from scratch reveal fundamental architectural requirements for multi-step inference, or are current approaches contingent engineering solutions?

值得阅读的有趣书籍:

  • 贝叶斯方法与科学合理性——对休谟问题的思考: https://book.douban.com/subject/4472081/
  • Reinforcement Learning from Human Feedback - A short introduction to RLHF and post-training focused on language models: https://rlhfbook.com/
  • Build a Reasoning Model (From Scratch): https://www.manning.com/books/build-a-reasoning-model-from-scratch
  • Build a Large Language Model (From Scratch): https://www.manning.com/books/build-a-large-language-model-from-scratch

以下内容由 LLM 生成,可能包含不准确之处。

背景

这个集合涵盖了基础认识论、现代对齐技术和实际的人工智能实现——反映了该领域作为工程实践和哲学探究的双重性质。这种对比很能说明问题:贝叶斯推理与几个世纪以来关于归纳的辩论相关(休谟问题),而RLHF和Transformer架构代表了将大语言模型与人类偏好对齐的最尖端方法。它们共同解决了人工智能的核心矛盾:构建既在技术上有能力又在行为上有理性依据的系统。

关键见解

认识论基础对对齐很重要。 中文关于贝叶斯方法和科学理性的著作重新审视了休谟的归纳问题——我们如何从有限的观察中证明推断一般原则的合理性。这不是纯学术问题:RLHF隐含地从有限反馈中对人类偏好进行贝叶斯更新。Nathan Lambert的RLHF书籍描述了训练后如何使用基于人类比较训练的奖励模型来引导基础模型,但很少质疑从稀疏信号学习"价值观"的认识论有效性。这个差距很重要:如果我们无法证明普通归纳的合理性,那么从少量偏好数据中证明价值对齐的合理性就变得更加危险。

从零开始的实现揭示了架构承诺。 Raschka的大语言模型书籍及其推理模型伴读书强调在没有抽象层的情况下实现注意力机制和Transformer。这种教学方法暴露了通常隐藏在框架中的设计选择:为什么使用缩放点积注意力、为什么层归一化的位置很重要、位置编码如何塑造可学习的内容。理解这些细节阐明了为什么某些对齐干预(如RLHF微调)有效——它们利用了架构中已经存在的特定归纳偏差。

待解问题

我们能否以使其认识论假设明确的方式,将RLHF与贝叶斯信念更新的关系正式化?从零开始构建推理模型是否揭示了多步推理的基本架构要求,或者当前的方法只是偶然的工程解决方案?

Have thoughts on this?有想法?

I'd love to hear from you — questions, corrections, disagreements, or anything else.欢迎来信交流——问题、勘误、不同看法,或任何想说的。

hi@changkun.de
© 2008 - 2026 Changkun Ou. All rights reserved.保留所有权利。 | PV/UV: /
0%