yijieisabelliu.bsky.social
@yijieisabelliu.bsky.social
🙌 This work was done with my incredible advisors @tomssilver.bsky.social and @ben-eysenbach.bsky.social and amazing collaborator Bowen Li.
November 5, 2025 at 3:18 PM
In Cleanup Table environment:
(8/8)
November 5, 2025 at 3:18 PM
In Cluttered Drawer environment:
(7/8)
November 5, 2025 at 3:18 PM
We demonstrate our results on PyBullet simulated environments. In Obstacle Tower environment:
(6/8)
November 5, 2025 at 3:18 PM
3️⃣ Self-Balancing. SLAP automatically navigates this spectrum between planning and learning. In the extremes, if shortcuts are too difficult to learn, SLAP reduces to the original planners; if the tasks are easy, SLAP reduces to RL—the plan collapses into a single shortcut.
(5/8)
November 5, 2025 at 3:18 PM
2️⃣ Hierarchical. SLAP leverages the hierarchical structure of planners to decompose the unstructured, long-horizon problem into smaller, tractable shortcuts for RL to learn.
(4/8)
November 5, 2025 at 3:18 PM
Our approach is:
1️⃣ Automatic. SLAP does not require any intermediate inputs or additional assumptions. The way we find, learn, and deploy shortcuts is completely automatic.
(3/8)
November 5, 2025 at 3:18 PM
We introduce Shortcut Learning for Abstract Planning (SLAP), a new method that uses reinforcement learning (RL) to discover shortcuts in the planning graphs induced by task and motion planning (TAMP) skill libraries. It is a plug-and-play module that can be trained on top of existing planners.
(2/8)
November 5, 2025 at 3:18 PM