能量基模型:专为LLM无法处理的正确性而生Energy-Based Models: Built for the Correctness LLMs Can't Deliver
关键要点:能量基模型(EBM)无 token 且非自回归,内置验证和自对齐机制,终于让 AI 在代码生成和芯片设计等关键任务中变得可靠。
Logical Intelligence 创始人兼 CEO Eve 领导一家基础 AI 公司,先用 LLM 原型验证,再长期投入 EBM 来填补确定性空白。LLM 把一切塞进语言和顺序猜测,制造昂贵的黑盒幻觉;而 EBM 用物理启发的能量最小化一次映射所有可能状态,提供鸟瞰视角避免走错路。
Eve 解释道:“EBM 天生是非自回归的。没有 token 序列,这就是它与众不同的根本原因。”这种架构允许实时检查和控制训练,通过潜变量处理稀疏数据并捕捉现实世界规则,还能与外部验证器完美配合——正好解决 LLM 无法进入飞机、汽车和生产工程领域的问题。
Logical Intelligence 创始人兼 CEO Eve 领导一家基础 AI 公司,先用 LLM 原型验证,再长期投入 EBM 来填补确定性空白。LLM 把一切塞进语言和顺序猜测,制造昂贵的黑盒幻觉;而 EBM 用物理启发的能量最小化一次映射所有可能状态,提供鸟瞰视角避免走错路。
Eve 解释道:“EBM 天生是非自回归的。没有 token 序列,这就是它与众不同的根本原因。”这种架构允许实时检查和控制训练,通过潜变量处理稀疏数据并捕捉现实世界规则,还能与外部验证器完美配合——正好解决 LLM 无法进入飞机、汽车和生产工程领域的问题。
The Takeaway: Energy-based models (EBMs) are token-free and non-autoregressive, delivering built-in verification and self-alignment that finally make AI safe for mission-critical systems like code generation and chip design.
Eve, founder and CEO of Logical Intelligence, runs a foundational AI company that prototypes on LLMs but is betting long-term on EBMs to close the determinism gap. LLMs force everything through language and sequential guessing, creating expensive black-box hallucinations; EBMs use physics-inspired energy minimization to map all possible states at once, giving a bird's-eye view that avoids wrong turns.
"EBMs are naturally non autoregressive. There are no sequences of tokens, and that's what makes it fundamentally different," Eve explains. The architecture lets you inspect and control training in real time, works beautifully with sparse data via latent variables that capture real-world rules, and pairs perfectly with external verifiers—solving the exact problems that keep LLMs out of planes, cars, and production engineering today.
查看原文 →
Eve, founder and CEO of Logical Intelligence, runs a foundational AI company that prototypes on LLMs but is betting long-term on EBMs to close the determinism gap. LLMs force everything through language and sequential guessing, creating expensive black-box hallucinations; EBMs use physics-inspired energy minimization to map all possible states at once, giving a bird's-eye view that avoids wrong turns.
"EBMs are naturally non autoregressive. There are no sequences of tokens, and that's what makes it fundamentally different," Eve explains. The architecture lets you inspect and control training in real time, works beautifully with sparse data via latent variables that capture real-world rules, and pairs perfectly with external verifiers—solving the exact problems that keep LLMs out of planes, cars, and production engineering today.