Koustuv Sinha
banner
koustuvsinha.com
Koustuv Sinha
@koustuvsinha.com
🔬Research Scientist, Meta AI (FAIR).

🎓PhD from McGill University + Mila

🙇‍♂️I study Multimodal LLMs, Vision-Language Alignment, LLM Interpretability & I’m passionate about ML Reproducibility (@reproml.org)

🌎https://koustuvsinha.com/
Pinned
🚨 We are pleased to announce the first, in-person event for the Machine Learning Reproducibility Challenge, MLRC 2025! Save your dates: August 21st, 2025 at Princeton!
Reposted by Koustuv Sinha
Our team is hiring a postdoc in (mechanistic) interpretability! The ideal candidate will have research experience in interpretability for text and/or image generation models and be excited about open science!

Please consider applying or sharing with colleagues: metacareers.com/jobs/2223953961352324
careers.com
July 15, 2025 at 8:11 PM
Reposted by Koustuv Sinha
Excited to share the results of my recent internship!

We ask 🤔
What subtle shortcuts are VideoLLMs taking on spatio-temporal questions?

And how can we instead curate shortcut-robust examples at a large-scale?

We release: MVPBench

Details 👇🔬
June 13, 2025 at 2:47 PM
Reposted by Koustuv Sinha
The HuggingFace/Nanotron team just shipped an entire pretraining textbook in interactive format. huggingface.co/spaces/nanot...

It’s not just a great pedagogic support, but many unprecedented data and experiments presented for the first time in a systematic way.
February 19, 2025 at 7:13 PM
Reposted by Koustuv Sinha
Excited to have two papers at #NAACL2025!
The first reveals how human over-reliance can be exacerbated by LLM friendliness. The second presents a novel computational method for concept tracing. Check them out!

arxiv.org/pdf/2407.07950

arxiv.org/pdf/2502.05704
February 19, 2025 at 9:58 PM
Reposted by Koustuv Sinha
👋 Hello world! We’re thrilled to announce the v0.4 release of fairseq2 — an open-source library from FAIR powering many projects at Meta. pip install fairseq2 and explore our trainer API, instruction & preference finetuning (up to 70B), and native vLLM integration.
February 12, 2025 at 12:31 PM
Reposted by Koustuv Sinha
I am shocked by the death of Felix Hill. He was one of the brightest minds of my generation.

His last blog post on the stress of working in AI is very poignant. Apart from the emptiness of working mostly to make billionaires even richer, there's the intellectual emptiness of 'scale is all you need'
The blog post of the late Felix Hill is powerful. Stress for AI researchers today is real.

I did not know Felix Hill and I am sorry for those who did.
This story is perhaps a reminder for students, postdocs, founders and researchers to take care of their well being.

medium.com/@felixhill/2...
200bn Weights of Responsibility
The Stress of Working in Modern AI
medium.com
January 14, 2025 at 12:41 PM
We posted our paper on arxiv recently, sharing this here too: arxiv.org/abs/2412.141... - work led by our amazing intern Peter Tong. Key findings:

- LLMs can be trained to generate visual embeddings!!
- VQA data appears to help a lot in generation!
- Better understanding = better generation!
December 26, 2024 at 8:01 PM
🚨 We are pleased to announce the first, in-person event for the Machine Learning Reproducibility Challenge, MLRC 2025! Save your dates: August 21st, 2025 at Princeton!
December 13, 2024 at 7:06 PM
Reposted by Koustuv Sinha
Our paper PRISM alignment won a best paper award at #neurips2024!

All credits to @hannahrosekirk.bsky.social A.Whitefield, P.Röttger, A.M.Bean, K.Margatina, R.Mosquera-Gomez, J.Ciro, @maxbartolo.bsky.social H.He, B.Vidgen, S.Hale

Catch Hannah tomorrow at neurips.cc/virtual/2024/poster/97804
blog.neurips
December 11, 2024 at 4:20 PM
Checkout the MLRC 2023 posters at #NeurIPS 2024 this week: reproml.org/proceedings/ - do drop by to these posters and say hi!
Online Proceedings | MLRC
Machine Learning Reproducibility Challenge
reproml.org
December 10, 2024 at 4:15 PM
Reposted by Koustuv Sinha
The return of the Autoregressive Image Model: AIMv2 now going multimodal.
Excellent work by @alaaelnouby.bsky.social & team with code and checkpoints already up:

arxiv.org/abs/2411.14402
November 22, 2024 at 9:44 AM
Reposted by Koustuv Sinha
For those who missed this post on the-network-that-is-not-to-be-named, I made public my "secrets" for writing a good CVPR paper (or any scientific paper). I've compiled these tips of many years. It's long but hopefully it helps people write better papers. perceiving-systems.blog/en/post/writ...
Writing a good scientific paper
perceiving-systems.blog
November 20, 2024 at 10:18 AM
Reposted by Koustuv Sinha
How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this:

Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢

🧵⬇️
November 20, 2024 at 4:35 PM
When I first read this paper, I instinctively scoffed at the idea. But the more I look at empirical results, the more I’m convinced this paper highlights something fundamentally amazing. Lots of exciting research on this direction will come very soon!

arxiv.org/abs/2405.07987
November 20, 2024 at 12:29 AM
Reposted by Koustuv Sinha
November 19, 2024 at 3:48 AM
Reposted by Koustuv Sinha
Doing good science is 90% finding a science buddy to constantly talk to about the project.
November 9, 2024 at 10:53 PM
Reposted by Koustuv Sinha
New here? Interested in AI/ML? Check out these great starter packs!

AI: go.bsky.app/SipA7it
RL: go.bsky.app/3WPHcHg
Women in AI: go.bsky.app/LaGDpqg
NLP: go.bsky.app/SngwGeS
AI and news: go.bsky.app/5sFqVNS

You can also search all starter packs here: blueskydirectory.com/starter-pack...
November 9, 2024 at 9:13 AM