Selection Theorems: A Program For Understanding Agents by johnswentworth
My take on Vanessa Kosoy's take on AGI safety by Steve Byrnes
How special are human brains among animal brains? by Alex Zhu
How uniform is the neocortex? by Alex Zhu
How "honest" is GPT-3?Q by Abram Demski
Bottle Caps Aren't Optimisers by DanielFilan
Deceptive Alignment by Evan Hubinger, Chris van Merwijk, Vladimir Mikulik, Joar Skalse, Scott Garrabrant
Problems in AI Alignment that philosophers could potentially contribute to by Wei Dai
Public Static: What is Abstraction? by johnswentworth
Updating the Lottery Ticket Hypothesis by johnswentworth
Siren worlds and the perils of over-optimised search by Stuart Armstrong
Arguments about fast takeoff by Paul Christiano
Alignment Newsletter #13: 07/02/18 by Rohin Shah
Risks from Learned Optimization: Conclusion and Related Work by Evan Hubinger, Chris van Merwijk, Vladimir Mikulik, Joar Skalse, Scott Garrabrant
2-D Robustness by Vladimir Mikulik
Learning Normativity: A Research Agenda by Abram Demski
To what extent is GPT-3 capable of reasoning? by Alex Turner
Environmental Structure Can Cause Instrumental Convergence by Alex Turner
[Book Review] "The Alignment Problem" by Brian Christian,Lsusr
Comment on decision theory by Rob Bensinger
Create your
podcast in
minutes
It is Free
Navigating Life After 40
Teaching Learning Leading K-12
Regenerative Skills
The Jordan B. Peterson Podcast
The Mel Robbins Podcast