http://dimadamen.github.io
HD-EPIC: A Highly-Detailed Egocentric Video Dataset
hd-epic.github.io
arxiv.org/abs/2502.04144
New collected videos
263 annotations/min: recipe, nutrition, actions, sounds, 3D object movement &fixture associations, masks.
26K VQA benchmark to challenge current VLMs
1/N
Also, we now provide access to the full HowTo100M dataset!
Download our datasets, or HowTo100M at github.com/ekazakos/grove
Also, we now provide access to the full HowTo100M dataset!
Download our datasets, or HowTo100M at github.com/ekazakos/grove
and internationally is open for applications - DL 29 Jan 2026.
uob-mavi.github.io/Summer@MaVi....
Would you like to work with any of the Faculty working in Machine Learning and Computer Vision #mavi as part of our summer of research at Bristol program?
uob-mavi.github.io/Summer@MaVi....
and internationally is open for applications - DL 29 Jan 2026.
uob-mavi.github.io/Summer@MaVi....
@cvprconference.bsky.social #CVPR2026
egovis.github.io/cvpr26/
CFP and challenge deadlines after the NY
Great lineup of 7 keynote speakers...
See you in Denver!
@cvprconference.bsky.social #CVPR2026
egovis.github.io/cvpr26/
CFP and challenge deadlines after the NY
Great lineup of 7 keynote speakers...
See you in Denver!
EgoVis 2024/2025 Distinguished Paper Awards.
Published a paper contributing to Ego Vision in 2024/25?
Innovative &advancing Ego Vision?
Worthy of a prize?
DL for nominations 20 Feb 2026
Awards announced @cvprconference.bsky.social #CVPR2026
egovis.github.io/awards/2024_...
EgoVis 2024/2025 Distinguished Paper Awards.
Published a paper contributing to Ego Vision in 2024/25?
Innovative &advancing Ego Vision?
Worthy of a prize?
DL for nominations 20 Feb 2026
Awards announced @cvprconference.bsky.social #CVPR2026
egovis.github.io/awards/2024_...
The N-Body Problem: Parallel Execution from Single-Person Egocentric Video
Input: Single-person egocentric video 👤
Out: imagine how these tasks can be performed faster by N > 1 people, correctly e.g. N=2 👥
📎 arxiv.org/abs/2512.11393
👀 zhifanzhu.github.io/ego-nbody/
1/4
The N-Body Problem: Parallel Execution from Single-Person Egocentric Video
Input: Single-person egocentric video 👤
Out: imagine how these tasks can be performed faster by N > 1 people, correctly e.g. N=2 👥
📎 arxiv.org/abs/2512.11393
👀 zhifanzhu.github.io/ego-nbody/
1/4
Would you like to work with any of the Faculty working in Machine Learning and Computer Vision #mavi as part of our summer of research at Bristol program?
uob-mavi.github.io/Summer@MaVi....
Would you like to work with any of the Faculty working in Machine Learning and Computer Vision #mavi as part of our summer of research at Bristol program?
uob-mavi.github.io/Summer@MaVi....
"Leveraging Multimodal Data for Egocentric Video Understanding" w no corrections
📜 in ICASSP23 CVPR24 CVPR25 3DV25 TPAMI25
jacobchalk.github.io
🙏examiners @hildekuehne.bsky.social @andrewowens.bsky.social &Wei-Hong Li
"Leveraging Multimodal Data for Egocentric Video Understanding" w no corrections
📜 in ICASSP23 CVPR24 CVPR25 3DV25 TPAMI25
jacobchalk.github.io
🙏examiners @hildekuehne.bsky.social @andrewowens.bsky.social &Wei-Hong Li
World models for evaluating autonomous driving, GAIA3 released! End-to-end driving model &loads of insights!
Thanks for visiting &spending the day talking to researchers.
World models for evaluating autonomous driving, GAIA3 released! End-to-end driving model &loads of insights!
Thanks for visiting &spending the day talking to researchers.
Zihui Xue, Kristen Grauman @dimadamen.bsky.social Andrew Zisserman, Tengda Han
tl;dr: in title. I love such "blind baseline" papers.
arxiv.org/abs/2511.21681
Zihui Xue, Kristen Grauman @dimadamen.bsky.social Andrew Zisserman, Tengda Han
tl;dr: in title. I love such "blind baseline" papers.
arxiv.org/abs/2511.21681
pics w @simonmcs.bsky.social and Sadaf Alam
pics w @simonmcs.bsky.social and Sadaf Alam
Abstract: Jan 23, 2026 AoE
Paper: Jan 28, 2026 AoE
Location: Seoul, South Korea 🇰🇷
icml.cc/Conferences/...
ICML'26 (abs): 71 days.
ICML'26 (paper): 76 days.
ECCV'26: 112 days.
But how is this skill learned, and can we model its progression?
We present CleverBirds, accepted #NeurIPS2025, a large-scale benchmark for visual knowledge tracing.
📄 arxiv.org/abs/2511.08512
1/5
Please help us share this post among students you know with an interest in Machine Learning and Biodiversity! 🤖🪲🌱
Please help us share this post among students you know with an interest in Machine Learning and Biodiversity! 🤖🪲🌱
By Gianluca Monaci, @weinzaepfelp.bsky.social and myself.
@naverlabseurope.bsky.social
arxiv.org/abs/2507.01667
🧵1/5
By Gianluca Monaci, @weinzaepfelp.bsky.social and myself.
@naverlabseurope.bsky.social
Form: support.conferences.computer.org/cvpr/help-desk
Form: support.conferences.computer.org/cvpr/help-desk
PointSt3R: Point Tracking through 3D Grounded Correspondence
arxiv.org/abs/2510.26443
Can point tracking be re-formulated as pairwise frame correspondence solely?
We fine-tuning MASt3R with dynamic correspondences and a visibility loss and achieve competitive point tracking results
1/3
PointSt3R: Point Tracking through 3D Grounded Correspondence
arxiv.org/abs/2510.26443
Can point tracking be re-formulated as pairwise frame correspondence solely?
We fine-tuning MASt3R with dynamic correspondences and a visibility loss and achieve competitive point tracking results
1/3
R. Guerrier, @adamharley.bsky.social, @dimadamen.bsky.social
Bristol/Meta
rhodriguerrier.github.io/PointSt3R/
R. Guerrier, @adamharley.bsky.social, @dimadamen.bsky.social
Bristol/Meta
rhodriguerrier.github.io/PointSt3R/
Chandan Yeshwanth and Yueh-Cheng Liu have added pano captures for 956 ScanNet++ scenes, fully aligned with the 3D meshes, DSLR, and iPhone data - multiple panos per scene
Check it out:
Docs kaldir.vc.in.tum.de/scannetpp/do...
Code github.com/scannetpp/sc...
Chandan Yeshwanth and Yueh-Cheng Liu have added pano captures for 956 ScanNet++ scenes, fully aligned with the 3D meshes, DSLR, and iPhone data - multiple panos per scene
Check it out:
Docs kaldir.vc.in.tum.de/scannetpp/do...
Code github.com/scannetpp/sc...
@bristoluni.bsky.social to give a #MaVi for a seminar: From Pixels to 3D Motion
We enjoyed your visit! Thanks for staying through for all 1-1s with the researchers.
@bristoluni.bsky.social to give a #MaVi for a seminar: From Pixels to 3D Motion
We enjoyed your visit! Thanks for staying through for all 1-1s with the researchers.
Read the full article: ellis.eu/news/ellis-s...
Read the full article: ellis.eu/news/ellis-s...
Kinaema: A recurrent sequence model for memory and pose in motion
arxiv.org/abs/2510.20261
By @mbsariyildiz.bsky.social, @weinzaepfelp.bsky.social, G. Bono, G. Monaci and myself
@naverlabseurope.bsky.social
1/9
Kinaema: A recurrent sequence model for memory and pose in motion
arxiv.org/abs/2510.20261
By @mbsariyildiz.bsky.social, @weinzaepfelp.bsky.social, G. Bono, G. Monaci and myself
@naverlabseurope.bsky.social
1/9