Apratim Bhattacharyya
apratimbh.bsky.social
Apratim Bhattacharyya
@apratimbh.bsky.social
ML Researcher, Qualcomm AI Research | Postdoc, University of Tübingen | PhD, Max Planck Institute for Informatics | AI Assistants, Multi-modal LLMs, Autonomous Driving
Pinned
Join us at the CVPR 2025 Workshop on Vision-based Assistants in the Real-world (VAR) and tackle one of AI's biggest challenges: building systems that can comprehend and reason about dynamic, real-world scenes.

Workshop Page: varworkshop.github.io
📣📣📣Our team at Qualcomm AI Research is hiring Research Interns for Summer 2026 in Toronto to work on multi-modal LLMs and embodied AI.

👉Apply here:
1) Embodied AI:
qualcomm.wd12.myworkdayjobs.com/External/job...

2) Multi-modal LLMs:
FY26 Intern - Deep Learning Research Internship - Embodied AI - Canada (4 months)
Company: Qualcomm Canada ULC Job Area: Interns Group, Interns Group > Interim Engineering Intern - SW Qualcomm Overview: Qualcomm is a company of inventors that unlocked 5G ushering in an age of ra...
qualcomm.wd12.myworkdayjobs.com
October 16, 2025 at 5:05 AM
🚨Submit by 1st May @cvprconference.bsky.social: extended abstracts on streaming vision-language models, real-time activity understanding, grounding, ego-centric video understanding, language and robot learning. Contributions are encouraged to include a demo!

👉Details: varworkshop.github.io/calls/
April 28, 2025 at 9:45 PM
🚨🚨🚨 We are now accepting submissions!
Call for Participation @cvprconference.bsky.social: Multi-Modal LLMs - prepare to engage in a dynamic, face-to-face conversation with a real human user!

Details: varworkshop.github.io/challenges/

🚨🚨🚨The winning teams will receive a prize and a contributed talk.

P.S. GPT-4o does not do too well.
March 26, 2025 at 4:52 AM
Reposted by Apratim Bhattacharyya
Call for Participation @cvprconference.bsky.social: Multi-Modal LLMs - prepare to engage in a dynamic, face-to-face conversation with a real human user!

Details: varworkshop.github.io/challenges/

🚨🚨🚨The winning teams will receive a prize and a contributed talk.

P.S. GPT-4o does not do too well.
March 26, 2025 at 4:38 AM
Reposted by Apratim Bhattacharyya
Call for Participation: We're excited to announce a challenge focused on developing AI assistants that can guide users through workout sessions with intelligent feedback!

🚨The winning teams will receive a prize along with a contributed talk. 🚨

Website: varworkshop.github.io/challenges/
March 10, 2025 at 6:47 PM
🚨Submission are now open!
Call for Papers and Demos @cvprconference.bsky.social: on topics such as streaming vision-language models, real-time activity understanding, grounding, ego-centric video understanding, language and robot learning. Contributions are encouraged to include a demo!

Link: varworkshop.github.io/calls/
March 3, 2025 at 8:47 PM
Reposted by Apratim Bhattacharyya
Call for Papers and Demos @cvprconference.bsky.social: on topics such as streaming vision-language models, real-time activity understanding, grounding, ego-centric video understanding, language and robot learning. Contributions are encouraged to include a demo!

Link: varworkshop.github.io/calls/
March 3, 2025 at 8:30 PM
Join us at the @cvprconference.bsky.social Workshop on Vision-based Assistants in the Real-world (VAR) and tackle one of AI's biggest challenges: building systems that can comprehend and reason about dynamic, real-world scenes.

Workshop Page: varworkshop.github.io
February 28, 2025 at 7:56 PM
Reposted by Apratim Bhattacharyya
By popular demand, we are extending #CVPR2025 coverage to Bluesky. Stay tuned!
February 27, 2025 at 9:07 PM
Accepted to CVPR2025 🥳🥳

#CVPR2025
🚨Check out our new work on distilling reasoning skills from LLMs into efficient driving policies, to deal with critical "long-tail" scenarios.

arXiv: arxiv.org/abs/2501.09757
February 26, 2025 at 10:56 PM
🚨We present in "Enhancing Hallucination Detection through Noise Injection" [https://arxiv.org/pdf/2502.03799], an efficient approach to detect hallucinations in LLMs, within a Bayesian framework.

TL; DR - We use noise injection to capture both epistemic and aleatoric uncertainty!
February 11, 2025 at 10:01 PM
Join us at the CVPR 2025 Workshop on Vision-based Assistants in the Real-world (VAR) and tackle one of AI's biggest challenges: building systems that can comprehend and reason about dynamic, real-world scenes.

Workshop Page: varworkshop.github.io
January 21, 2025 at 6:43 PM
🚨Check out our new work on distilling reasoning skills from LLMs into efficient driving policies, to deal with critical "long-tail" scenarios.

arXiv: arxiv.org/abs/2501.09757
January 17, 2025 at 8:26 PM
🚨 The code for our NeurIPS 2024 (D&B track) paper: ClevrSkills: Compositional Language And Visual Understanding in Robotics (arxiv.org/abs/2411.09052), is now available.

GitHub Repo: github.com/Qualcomm-AI-...
Dataset Page: www.qualcomm.com/developer/so...
ClevrSkills AI Dataset
www.qualcomm.com
January 10, 2025 at 7:55 PM