Join Ads Marketplace to earn through podcast sponsorships.
Manage your ads with dynamic ad insertion capability.
Monetize with Apple Podcasts Subscriptions via Podbean.
Earn rewards and recurring income from Fan Club membership.
Get the answers and support you need.
Resources and guides to launch, grow, and monetize podcast.
Stay updated with the latest podcasting tips and trends.
Check out our newest and recently released features!
Podcast interviews, best practices, and helpful tips.
The step-by-step guide to start your own podcast.
Create the best live podcast and engage your audience.
Tips on making the decision to monetize your podcast.
The best ways to get more eyes and ears on your podcast.
Everything you need to know about podcast advertising.
The ultimate guide to recording a podcast on your phone.
Steps to set up and use group recording in the Podbean app.
Join Ads Marketplace to earn through podcast sponsorships.
Manage your ads with dynamic ad insertion capability.
Monetize with Apple Podcasts Subscriptions via Podbean.
Earn rewards and recurring income from Fan Club membership.
Get the answers and support you need.
Resources and guides to launch, grow, and monetize podcast.
Stay updated with the latest podcasting tips and trends.
Check out our newest and recently released features!
Podcast interviews, best practices, and helpful tips.
The step-by-step guide to start your own podcast.
Create the best live podcast and engage your audience.
Tips on making the decision to monetize your podcast.
The best ways to get more eyes and ears on your podcast.
Everything you need to know about podcast advertising.
The ultimate guide to recording a podcast on your phone.
Steps to set up and use group recording in the Podbean app.
Curtis Huebner on Doom, AI Timelines and Alignment at EleutherAI
Curtis, also known on the internet as AI_WAIFU, is the head of Alignment at EleutherAI. In this episode we discuss the massive orders of H100s from different actors, why he thinks AGI is 4-5 years away, why he thinks we're 90% "toast", his comment on Eliezer Yudkwosky's Death with Dignity, and what kind of Alignment projects is currently going on at EleutherAI, especially a project with Markov chains and the Alignment test project that he is currently leading.
Youtube: https://www.youtube.com/watch?v=9s3XctQOgew
Transcript: https://theinsideview.ai/curtis Death with Dignity: https://www.lesswrong.com/posts/j9Q8bRmwCgXRYAgcJ/miri-announces-new-death-with-dignity-strategy Alignment Minetest: https://www.eleuther.ai/projects/alignment-minetest Alignment Minetest update: https://blog.eleuther.ai/minetester-intro/
OUTLINE
(00:00) Highlights / Intro
(00:50) The Fuck That Noise Comment On Death With Dignity
(10:28) Th Probability of Doom Is 90%
(12:44) Best Counterarguments For His High P(doom)
(14:41) Compute And Model Size Required For A Dangerous Model
(17:59) Details For Curtis' Model Of Compute Required
(21:23) Why This Estimate Of Compute Required Might Be Wrong, Ajeya Cotra's Transformative AI report
(29:00) Curtis' Median For AGI Is Around 2028, Used To Be 2027
(30:50) How Curtis Approaches Life With Short Timelines And High P(Doom)
(35:27) Takeoff Speeds—The Software view vs. The Hardware View
(39:57) Nvidia's 400k H100 rolling down the assembly line, AIs soon to be unleashed on their own source code
(41:04) Could We Get A Fast Takeoff By Fuly Automating AI Research With More Compute
(46:00) The Entire World (Tech Companies, Governments, Militaries) Is Noticing New AI Capabilities That They Don't Have
(47:57) Open-source vs. Close source policies. Mundane vs. Apocalyptic considerations.
(53:25) Curtis' background, from teaching himself deep learning to EleutherAI
(55:51) Alignment Project At EleutherAI: Markov Chain and Language Models
(01:02:15) Research Philosophy at EleutherAI: Pursuing Useful Projects, Multingual, Discord, Logistics
(01:07:38) Alignment MineTest: Links To Alignmnet, Embedded Agency, Wireheading
(01:15:30) Next steps for Alignment Mine Test: focusing on model-based RL
(01:17:07) Training On Human Data & Using an Updated Gym Environment With Human APIs
(01:19:20) Model Used, Not Observing Symmetry
(01:21:58) Another goal of Alignment Mine Test: Study Corrigibility
(01:28:26) People ordering H100s Are Aware Of Other People Making These Orders, Race Dynamics, Last Message
Create your
podcast in
minutes
It is Free