Hirokatsu Kataoka | 片岡裕雄
banner
hirokatukataoka.bsky.social
Hirokatsu Kataoka | 片岡裕雄
@hirokatukataoka.bsky.social
Chief Scientist @ AIST | Academic Visitor @ Oxford VGG | PI @ cvpaper.challenge | 3D ResNet (Top 0.5% in 5-yr CVPR) | FDSL (ACCV20 Award/BMVC23 Award Finalist)
We’ve released the ICCV 2025 Report!
hirokatsukataoka.net/temp/presen/...

Compiled during ICCV in collaboration with LIMIT.Lab, cvpaper.challenge, and Visual Geometry Group (VGG), this report offers meta insights into the trends and tendencies observed at this year’s conference.

#ICCV2025
October 31, 2025 at 5:46 PM
[Workshop Paper; 2/5; 19 Oct 16:40 - 18:00] Shinichi Mae, Ryousuke Yamada, Hirokatsu Kataoka, Industrial Synthetic Segment Pre-training, ICCV 2025 Workshop on LIMIT Workshop (Invited Poster). arxiv.org/abs/2505.13099
October 16, 2025 at 2:16 AM
[Workshop Paper; 1/5; 19 Oct 11:25 - 12:15] Misora Sugiyama, Hirokatsu Kataoka, Simple Visual Artifact Detection in Sora-Generated Videos, ICCV 2025 Workshop on Workshop on Human-Interactive Generation and Editing, 2025. arxiv.org/abs/2504.21334 / higen-2025.github.io
October 16, 2025 at 2:16 AM
[Main Conference Paper; 2/2; 22 Oct 10:45 - 12:45; Poster #451] Risa Shinoda, Nakamasa Inoue, Iro Laina, Christian Rupprecht, Hirokatsu Kataoka, AnimalClue: Recognizing Animals by their Traces, ICCV 2025 (Highlight). dahlian00.github.io/AnimalCluePa...
October 16, 2025 at 2:16 AM
[Main Conference Paper; 1/2; 21 Oct 15:00 - 17:00; Poster #246] Risa Shinoda, Nakamasa Inoue, Hirokatsu Kataoka, Masaki Onishi, Yoshitaka Ushiku, AgroBench: Vision-Language Model Benchmark in Agriculture, ICCV 2025. dahlian00.github.io/AgroBenchPage/
October 16, 2025 at 2:15 AM
We organized the "Cambridge Computer Vision Workshop" at the University of Cambridge together with Elliott Wu, Yoshihiro Fukuhara, and LIMIT.Lab! It was a fantastic workshop featuring presentations, networking, and discussions.
cambridgecv-workshop-2025sep.limitlab.xyz
October 2, 2025 at 12:15 PM
Finally, the accepted papers at #ICCV2025 / @iccv.bsky.social LIMIT Workshop has been publicly released!
--
- OpenReview: openreview.net/group?id=the...
- Website: iccv2025-limit-workshop.limitlab.xyz
October 2, 2025 at 12:06 PM
At ICCV 2025, I am organizing two workshops: the LIMIT Workshop and the FOUND Workshop.

◆ LIMIT Workshop (19 Oct, PM): iccv2025-limit-workshop.limitlab.xyz
◆ FOUND Workshop (19 Oct, AM): iccv2025-found-workshop.limitlab.xyz

We warmly invite you to attend at these workshops in ICCV 2025 Hawaii!
September 17, 2025 at 3:42 PM
I’m thrilled to announce my invited talk at BMVC 2025 Smart Cameras for Smarter Autonomous Vehicles and Robots!

supercamerai.github.io
September 2, 2025 at 2:35 PM
Our AnimalClue has been accepted to #ICCV2025 as a highlight🎉🎉🎉 We also released an official press release from AIST!! This is the collaboration between AIST x Oxford VGG.

Project page: dahlian00.github.io/AnimalCluePa...
Dataset: huggingface.co/risashinoda
Press: www.aist.go.jp/aist_j/press...
August 3, 2025 at 9:52 PM
Our AgroBench has been accepted to #ICCV2025 🎉🎉🎉 We released project page, paper, code, and dataset!!

Project page: dahlian00.github.io/AgroBenchPage/
Paper: arxiv.org/abs/2507.20519
Code: huggingface.co/datasets/ris...
Dataset: github.com/dahlian00/Ag...
August 3, 2025 at 9:49 PM
June 17, 2025 at 11:10 AM
For the research community, we’ve named it “http://LIMIT.Community.” If you’re interested, please feel free to contact us. Students are also welcome.
June 6, 2025 at 10:04 AM
LIMIT.Lab brings together computer vision researchers from Japan, UK, Germany, and Netherlands! Below are our current partner institutions:

🇯🇵 AIST, Science Tokyo, TUS
🇬🇧 Oxford VGG, Cambridge
🇩🇪 UTN FunAI Lab
🇳🇱 UvA
# Fields & partner institutions are continually expanding
June 6, 2025 at 10:03 AM
[LIMIT.Lab Launched]
limitlab.xyz

We’ve established "LIMIT.Lab" a collaboration hub for building multimodal AI models under limited resources, covering images, videos, 3D, and text, when any resource (e.g., compute, data, or labels) is constrained.
June 6, 2025 at 10:03 AM
“Industrial Synthetic Segment Pre-training” on arXiv!

Formula-driven supervised learning (FDSL) has surpassed the vision foundation model "SAM" on industrial data. It delivers strong transfer performance to industry while minimizing IP-related concerns.

arxiv.org/abs/2505.13099
May 21, 2025 at 10:28 AM
Very excited to announce that our Formula-Driven Supervised Learning (FDSL) series now includes audio modality 🎉🎉🎉
--
Formula-Supervised Sound Event Detection: Pre-Training Without Real Data, ICASSP 2025.
- Paper: arxiv.org/abs/2504.04428
- Project: yutoshibata07.github.io/Formula-SED/
April 11, 2025 at 1:53 PM
Our paper has been published ( lnkd.in/gyuPEWSA ) and issued an AIST press release (JPN: lnkd.in/gyifZauS ) on applying FDSL pre-training for microfossil recognition, showcasing an example of image recognition technology in AI for Science! 🎉🎉🎉

# The shared images are coming from our paper
March 20, 2025 at 2:52 PM
[ Reached 5,000 Citations! 🎉🎉🎉 ]

My research has reached 5,000 citations on Google Scholar! This wouldn’t have been possible without the support of my co-authors, colleagues, mentors, and the entire research community.

Looking forward to the next phase of collaboration! 🙌
February 7, 2025 at 3:33 PM
[New pre-training / augmentation dataset] MoireDB – a formula-generated interference-fringe image dataset for synthetic pre-training and data augmentation 🎉🎉🎉

Paper: arxiv.org/abs/2502.01490
February 4, 2025 at 3:56 PM
[ Research Paper Award🏅]
Our paper, "Efficient Load Interference Detection with Limited Labeled Data," has won the SICE International Young Authors Award (SIYA) 2025🎉🎉🎉 This work is the family of FDSL pre-training and its real-world application in advanced logistics using forklists.
January 25, 2025 at 4:55 AM
Our ICCV 2023 paper "SegRCDB" has reached 10 citations on Google Scholar! 🎉

We have proposed 'Synthetic Pre-training for Segmentation', establishing a strong baseline.

- Project: dahlian00.github.io/SegRCDBPage/
- GitHub: github.com/dahlian00/Se...
- Paper: openaccess.thecvf.com/content/ICCV...
January 23, 2025 at 3:58 PM
Excited to announce that our paper, "Text-guided Synthetic Geometric Augmentation for Zero-shot 3D Understanding ( arxiv.org/abs/2501.09278 )", is now available on arXiv pre-print 🎉🎉🎉

We propose a framework for generation-enhanced 3D dataset expansion, targeting zero-shot 3D understanding.
January 20, 2025 at 2:55 PM
January 14, 2025 at 4:37 PM
January 14, 2025 at 4:36 PM