Rabiraj Banerjee
rabirajb.bsky.social
Rabiraj Banerjee
@rabirajb.bsky.social
I focus on data learnability, model representation and uncertainty for subjective NLP tasks
PhDing on Interpretable NLP + CSS @gesis.org Prev: Masters Student + Researcher at @ubuffalo.bsky.social and Sr. Data Scientist at Coursera
Reposted by Rabiraj Banerjee
I’m excited to share our Findings of EMNLP paper w/ @cocoscilab.bsky.social , @rtommccoy.bsky.social, and @rdhawkins.bsky.social !

Language models, unlike humans, require large amounts of data, which suggests the need for an inductive bias.
But what kind of inductive biases do we need?
November 7, 2025 at 9:17 AM
Reposted by Rabiraj Banerjee
In the year since LRMs ("reasoning models") hit the scene, we have been trying to understand, analyze and demystify them.. Here are our efforts to date--conveniently all in one place..👇

www.linkedin.com/posts/subbar...
In the year since LRMs ("reasoning models") hit the scene, we have been trying to understand, analyze and demystify them.. Here are our efforts to date--conveniently all in one… | Subbarao K...
In the year since LRMs ("reasoning models") hit the scene, we have been trying to understand, analyze and demystify them.. Here are our efforts to date--conveniently all in one place.. (𝗙𝗶𝗿𝘀𝘁..) 𝗘𝘃𝗮𝗹...
www.linkedin.com
September 14, 2025 at 10:00 PM
Reposted by Rabiraj Banerjee
A Survey of Reinforcement Learning for Large Reasoning Models

Five sections:

- Foundational Components
- Foundational Problems
- Training Resources
- Applications
- Future Directions
September 11, 2025 at 11:42 PM
Reposted by Rabiraj Banerjee
When reading AI reasoning text (aka CoT), we (humans) form a narrative about the underlying computation process, which we take as a transparent explanation of model behavior. But what if our narratives are wrong? We measure that and find it usually is.

Now on arXiv: arxiv.org/abs/2508.16599
Humans Perceive Wrong Narratives from AI Reasoning Texts
A new generation of AI models generates step-by-step reasoning text before producing an answer. This text appears to offer a human-readable window into their computation process, and is increasingly r...
arxiv.org
August 27, 2025 at 9:30 PM
Reposted by Rabiraj Banerjee
Great interview with @stevenstrogatz.com with a lot of discussion of research advising. Parts reminded me of @eegilbert.org and @informor.bsky.social's (excellent) guides to PhD mentorship, with a big focus on ideation.
Eric's: docs.google.com/document/d/1...
Mor's: s.tech.cornell.edu/phd-syllabus/
August 24, 2025 at 2:26 PM
Reposted by Rabiraj Banerjee
We try to avoid self-promoting too much, but we (with @sjgreenwood.bsky.social) built a personalized feed with posts about papers from your network. Many people say it's the closest they can get to old academic twitter, and I hope you enjoy it and share with others too!

bsky.app/profile/pape...
**Please repost** If you're enjoying Paper Skygest -- our personalized feed of academic content on Bluesky -- we'd appreciate you reposting this! We’ve found that the most effective way for us to reach new users and communities is through users sharing it with their network
August 20, 2025 at 12:46 PM
Reposted by Rabiraj Banerjee
August 19, 2025 at 5:02 AM
Reposted by Rabiraj Banerjee
🤖 But wait! There's more! You can check out @shiraamitchell.bsky.social 's most recent update on the details of Calibration, posted yesterday! statmodeling.stat.columbia.edu/2025/08/12/s...
August 13, 2025 at 6:38 PM
Reposted by Rabiraj Banerjee
#acl2025 anyone get a good quote of phil resnik's last comment?

context: (some?all?) panelists & him agree the field needs more deep, careful research on smaller models to do better science. everyone is frustrated with impossibility of large-scale pretraining experiments
July 28, 2025 at 3:24 PM
@kennyjoseph.bsky.social , Kenny check this thread out
What are your favorite recent papers on using LMs for annotation (especially in a loop with human annotators), synthetic data for task-specific prediction, active learning, and similar?

Looking for practical methods for settings where human annotations are costly.

A few examples in thread ↴
July 24, 2025 at 11:38 AM
Reposted by Rabiraj Banerjee
What are your favorite recent papers on using LMs for annotation (especially in a loop with human annotators), synthetic data for task-specific prediction, active learning, and similar?

Looking for practical methods for settings where human annotations are costly.

A few examples in thread ↴
July 23, 2025 at 8:10 AM
Reposted by Rabiraj Banerjee
UB's new Department of AI and Society is hiring faculty across ranks (Assistant, Associate, Full Professor). We’re looking for transdisciplinary scholars interested in building AI by society, for society. Start dates begin Fall 2025.

More info: www.ubjobs.buffalo.edu/postings/57734
Assistant, Associate or Full Professor, AI & Society
The Department of AI and Society (AIS) at the University at Buffalo (UB) invites candidates to apply for multiple positions as Assistant Professor, Associate Professor, or Full Professor. The new AIS ...
www.ubjobs.buffalo.edu
July 17, 2025 at 4:11 PM
Reposted by Rabiraj Banerjee
Reposted by Rabiraj Banerjee
Check out our take on Chain-of-Thought.
I really like this paper as a survey on the current literature on what CoT is, but more importantly on what it's not.
It also serves as a cautionary tale to the (apparently quite common) misuse of CoT as an interpretable method.
Excited to share our paper: "Chain-of-Thought Is Not Explainability"! We unpack a critical misconception in AI: models explaining their steps (CoT) aren't necessarily revealing their true reasoning. Spoiler: the transparency can be an illusion. (1/9) 🧵
July 1, 2025 at 5:45 PM
Reposted by Rabiraj Banerjee
Hi everyone. I'm excited to announce that I will be organizing a 2nd Summit on Responsible Computing, AI, and Society rcais.github.io October 27-29, 2025.

We will explore the future of computing for health, sustainability, human-centered AI, and policy.

Please consider submitting a 1-page abstract
July 1, 2025 at 4:41 PM
So ICWSM concluded today and it was a blast, was a great honor to attend @icwsm.bsky.social at Copenhagen and present my work with @kennyjoseph.bsky.social and other colleagues. The paper link is here :
ojs.aaai.org/index.php/IC...,
View of Measuring Dimensions of Self-Presentation in Twitter Bios and their Links to Misinformation Sharing
ojs.aaai.org
June 26, 2025 at 9:39 PM
Reposted by Rabiraj Banerjee
This is great! The idea is somewhat obvious (good!), and I'm sure many have toyed with the connection to learning-to-rank. However, no work had developed it. This should be relevant for constructing valid PIs from just preferential feedback. openreview.net/pdf?id=ENJd3...
June 26, 2025 at 11:17 AM
Reposted by Rabiraj Banerjee
🚨 New preprint! 🚨
Phase transitions! We love to see them during LM training. Syntactic attention structure, induction heads, grokking; they seem to suggest the model has learned a discrete, interpretable concept. Unfortunately, they’re pretty rare—or are they?
June 24, 2025 at 6:29 PM
Reposted by Rabiraj Banerjee
Generative language systems are everywhere, and many of them stereotype, demean, or erase particular social groups.
June 16, 2025 at 9:49 PM
Reposted by Rabiraj Banerjee
Alright, people, let's be honest: GenAI systems are everywhere, and figuring out whether they're any good is a total mess. Should we use them? Where? How? Do they need a total overhaul?

(1/6)
June 15, 2025 at 12:20 AM
Reposted by Rabiraj Banerjee
🧵 1/ Las redes están llenas de odio. ¿Puede la inteligencia artificial ayudarnos a detectarlo…

⚖️ sin discriminar,
🚫 sin reforzar estereotipos,
🔁 y sin aprender a odiar?

Esa es la gran pregunta de mi tesis.

👇 Te lo cuento en este #HiloTesis @crueuniversidades.bsky.social @filarramendi.bsky.social
June 10, 2025 at 9:20 AM
Reposted by Rabiraj Banerjee
As we go through a lot of excitement about RL recently with lots of cool work/results, here is a reminder that RL with a reverse KL-regularizer to the base model cannot learn any new skills that were not already present in the base model. It can only amplify the weak skills.
👇
May 27, 2025 at 5:40 PM
Reposted by Rabiraj Banerjee
My path into AI
The sort of small wins that accumulate into a real career in AI.
When I started grad school AI prof's didn't have space for me in their group and when I ended I had no papers at NeurIPS/ICLR/ICML, yet the process can still work.
www.interconnects.ai/p/my-path-in...
My path into AI
How I got here. Building a career brick by brick over 8 years.
www.interconnects.ai
May 14, 2025 at 2:29 PM
Reposted by Rabiraj Banerjee
I posted this on LinkedIn too and it has over 600 reactions there, with the caveat that I don't know how many are from bots.
I've updated my LaTeX CV template for PhD students. It now uses the article document class, which makes the CV better for learning LaTeX and fixes several small formatting irregularities. shomir.net/wilhom_rosin...
Shomir Wilson - Example CV for PhD students
An example CV for PhD students, with LaTeX code.
shomir.net
May 12, 2025 at 8:03 PM
Reposted by Rabiraj Banerjee
Out of all the mountains of work on "debiasing" word2vec, BERT, and LLMs, what anecdotes or evidence do we have that debiasing techniques have *actually been used in practice*, in industry or research?

(not referring to "cultural alignment" techniques)

(bonus if used and *to good effect*)
May 5, 2025 at 9:17 PM