Appearance | dynalist.io
12 week online course, covering a range of policy levers for steering AI development. By taking this course, you’ll learn about the risks arising from future AI systems, and proposed governance interventions to address them. You’ll consider interactions between AI and biosecurity, cybersecurity and defence capabilities, and the disempowerment of human decision-makers. We’ll also provide an overview of open technical questions such as the control and alignment problems – which posit th...| BlueDot Impact
The availability of great public resources has wide-ranging benefits: both for the people publishing the content, and for others who want to use it. We think people often underestimate these benefits, but minor inconveniences like hiding information behind an email request turns a lot of people away. This is likely even more true for people […]| BlueDot Impact
AI could bring significant rewards to its creators. However, the average person seems to have wildly inaccurate intuitions about the scale of these rewards. By exploring some conservative estimates of the potential rewards AI companies could expect to see from the automation of human labour, this article tries to convey a grounded sense of ‘woah, this could […]| BlueDot Impact
How might AI-enabled oligarchies arise?| adamjones.me
Reinforcement learning from human feedback (RLHF) has emerged as a powerful technique for steering large language models (LLMs) toward desired behaviours. However, relying on simple human feedback doesn’t work for tasks that are too complex for humans to accurately judge at the scale needed to train AI models. Scalable oversight techniques attempt to address this […]| BlueDot Impact
If you thought we might be able to cure cancer in 2200, then I think you ought to expect there’s a good chance we can do it within years of the advent of AI systems that can do the research work humans can do.| Planned Obsolescence
PASTA: Process for Automating Scientific and Technological Advancement.| Cold Takes
Why would we program AI that wants to harm us? Because we might not know how to do otherwise.| Cold Takes