4/5
4/5
No manual labeling is required. The narrations from egocentric datasets also provide free-form text supervision! (Eg. "pour milk into bowl")
3/5
No manual labeling is required. The narrations from egocentric datasets also provide free-form text supervision! (Eg. "pour milk into bowl")
3/5
Most affordance detection methods just segment object parts & do not predict actionable regions for robots!
Our solution?
Use egocentric bimanual human videos to extract precise affordance regions considering object relationships, context, & hand coordination!
2/5
Most affordance detection methods just segment object parts & do not predict actionable regions for robots!
Our solution?
Use egocentric bimanual human videos to extract precise affordance regions considering object relationships, context, & hand coordination!
2/5
Stop labeling affordances or distilling them from VLMs.
Extract affordances from bimanual human videos instead!
Excited to share 2HandedAfforder: Learning Precise Actionable Bimanual Affordances from Human Videos, accepted at #ICCV2025! 🎉
🧵1/5
Stop labeling affordances or distilling them from VLMs.
Extract affordances from bimanual human videos instead!
Excited to share 2HandedAfforder: Learning Precise Actionable Bimanual Affordances from Human Videos, accepted at #ICCV2025! 🎉
🧵1/5
Congratulations to the winners of the Best Paper Awards: EgoDex & DexWild!
The full recording is available at: youtu.be/64yLApbBZ7I
Some highlights:
Congratulations to the winners of the Best Paper Awards: EgoDex & DexWild!
The full recording is available at: youtu.be/64yLApbBZ7I
Some highlights:
We’re bringing together researchers exploring how egocentric perception can drive next-gen robot learning!
🔗 Full info: egoact.github.io/rss2025
@roboticsscisys.bsky.social
We’re bringing together researchers exploring how egocentric perception can drive next-gen robot learning!
🔗 Full info: egoact.github.io/rss2025
@roboticsscisys.bsky.social