Aligned Pod


Pod6/S1: LLMs are Bayesian in Expectation, Not in Realization, Leon Chlon, Hassana Labs (2025).

Pod5/S1: On hallucinations & how Guardrails AI Keeps LLMs Honest via open source collab.

Pod3/S1: AI Alignment: Preventing Existential Risk from Superintelligent Machines

Pod2/S1: Navigating AI alignment risk and the future of human control.

This Podcast is produced using Google/DM’s NotebookLM ❤️