AI 2027 predicts that superhuman AIs will not be aligned to the values and goals intended by their human developers. This supplement justifies that assumption by discussing the possibilities for what goals the AIs end up with.| ai-2027.com
AI’s most outspoken leader found direction in a personal tragedy.| www.bigtechnology.com
Ryan Greenblatt — lead author on the explosive paper “Alignment faking in large language models” and chief scientist at Redwood Research — thinks there’s a 25% chance that within four years, AI will be able to do everything needed to run an AI company, from writing code to designing experiments to making strategic and business decisions.| 80,000 Hours
Regulation cannot be written in blood alone.| hollyelmore.substack.com
Scaling reinforcement learning, tracing circuits, and the path to fully autonomous agents| www.dwarkesh.com
Misaligned hive minds, Xi and Trump waking up, and automated Ilyas accelerating AI progress| www.dwarkesh.com
Can AI Do Revolutionary Science? And If It Can, What Does That Mean?| O’Reilly Media
[Crossposted on lesswrong, see here for prior posts] The following statements seem to be both important for AI safety and are not widely agreed upon. These are my opinions, not those of my employer…| Windows On Theory
Why AI is a risk for the future of our existence, and why we need to pause development.| PauseAI