Scaling RL, sparse rewards, continual learning, and the progress wall when pretraining really stops.| www.interconnects.ai
Continual learning is a huge bottleneck| www.dwarkesh.com