In contrast, large reasoning models like OpenAI o3 show a much weaker correlation, in line with other work indicating that LRMs struggle at ToM reasoning.
In contrast, large reasoning models like OpenAI o3 show a much weaker correlation, in line with other work indicating that LRMs struggle at ToM reasoning.
1. A robot helps a human reach one of 4 gems
2. An astronaut collects resources on the way to their spaceship
3. A hungry student looks for their favorite foodcart behind a building
How do we achieve this?
1. A robot helps a human reach one of 4 gems
2. An astronaut collects resources on the way to their spaceship
3. A hungry student looks for their favorite foodcart behind a building
How do we achieve this?
In a new paper led by @lanceying.bsky.social, we introduce a cognitive model that achieves this by synthesizing rational agent models on-the-fly -- presented at #EMNLP2025!
In a new paper led by @lanceying.bsky.social, we introduce a cognitive model that achieves this by synthesizing rational agent models on-the-fly -- presented at #EMNLP2025!
arxiv.org/abs/2411.09514
arxiv.org/abs/2411.09514
The trick is v neat: Turn an infinite limit into a telescoping sum, then arrange to have all terms after some step t cancel out.
The trick is v neat: Turn an infinite limit into a telescoping sum, then arrange to have all terms after some step t cancel out.