MOAR AXRP
This time with Erik Jenner, on a paper he's presenting at NeurIPS tomorrow - check it out if you're there!
Lots of people in the AI safety space worry about models being able to make deliberate, multi-step plans. But can we already see this in existing neural nets? In this episode, I talk with Erik Jenner about his work looking at internal look-ahead within chess-playing neural networks.
Ben Weinstein-Raun likes this.