PodcastIntel
Sign in Get Started Free
Dwarkesh Podcast
Dwarkesh Podcast

Paul Christiano — Preventing an AI takeover

Oct 31, 2023 · 3h 7m
AI Summary
  • Paul Christiano has modest AGI timelines: 40% by 2040, 15% by 2030; addresses whether RLHF invention was regrettable and whether alignment is necessarily dual-use
  • Post-AGI world design remains unsettled: question of whether keeping superintelligent entities enslaved is ethical; pushing labs toward responsible scaling policies to prevent AI coups
  • New proof systems could solve alignment by explaining model behavior mathematically; preventing bioweapon development and AI takeover requires coordination on safety research

More from Dwarkesh Podcast

View all episodes →

Get AI Summaries for Every New Episode

Subscribe to Dwarkesh Podcast and get AI summaries, guest tracking, and email digests delivered automatically.

Sign Up Free →