Epiplexity, Reasoning & The "Alien" Behavior of LLMs: my conversation with @Pavel_Izmailov, AI researcher at @AnthropicAI and professor at @nyuniversity. 00:00 - Intro 00:53 - Alien survival instincts: is continual learning about to create a major alignment problem? Riffing on the post by @iruletheworldmo 03:33 - Did AI learn deception from sci-fi literature? 05:55 - Defining Alignment, Superalignment & OpenAI teams 08:12 - Pavel’s journey: From Russian math to OpenAI Superalignment 10:46 - Culture check: OpenAI vs. Anthropic vs. Academia 11:54 - Why move to NYU? Academia and the need for exploratory research 13:09 - Does reasoning make AI alignment harder or easier? 14:22 - Sandbagging: When models pretend to be dumb 16:19 - Scalable Oversight: Using AI to supervise AI 18:04 - Weak-to-Strong Generalization 22:43 - Mechanistic Interpretability: Inside the black box 25:08 - The reasoning explosion 27:07 - Are Transformers enough or do we need a new paradigm?...