Yonatan Belinkov ✈️ COLM2025
boknilev.bsky.social
Yonatan Belinkov ✈️ COLM2025
@boknilev.bsky.social
Assistant professor of computer science at Technion; visiting scholar at @KempnerInst 2025-2026
https://belinkov.com/
Reach out if you'd like to discuss anything related to model interpretability and controllability, robustness, multi-agent communication, biological LMs, etc.
Also happy to talk about PhD and Post-doc opportunities!
October 7, 2025 at 1:41 PM
In the #LM4Science workshop, Friday 111:30am-1pm, we have a poster on generating rich text descriptions from protein sequences, work by Edo Dotan, who couldn't travel.
Preprint: www.biorxiv.org/content/10.1...
Protein2Text: Providing Rich Descriptions for Protein Sequences
Understanding the functionality of proteins has been a focal point of biological research due to their critical roles in various biological processes. Unraveling protein functions is essential for adv...
www.biorxiv.org
October 7, 2025 at 1:41 PM
In #Interplay25 workshop, Friday ~11:30, I'll present on measuring *parametric* CoT faithfulness on behalf of
@mtutek.bsky.social , who couldn't travel:
bsky.app/profile/mtut...

Later that day we'll have a poster on predicting success of model editing by Yanay Soker, who also couldn't travel
October 7, 2025 at 1:41 PM
In the LLM explainability in reasoning and planning workshop, Friday 9am, I'll talk about scaling up model interpretability (xllm-reasoning-planning-workshop.github.io)
XLLM-Reason-Plan
Website for the Workshop on the Application of LLM Explainability to Reasoning and Planning at COLM 2025
xllm-reasoning-planning-workshop.github.io
October 7, 2025 at 1:41 PM
@zorikgekhman.bsky.social
will present on Wednesday a poster on hidden factual knowledge in LMs
bsky.app/profile/zori...
🚨 It's often claimed that LLMs know more facts than they show in their outputs, but what does this actually mean, and how can we measure this “hidden knowledge”?

In our new paper, we clearly define this concept and design controlled experiments to test it.
1/🧵
October 7, 2025 at 1:41 PM
@itay-itzhak.bsky.social
presenting today, morning, a spotlight talk and poster on the origin of cognitive biases in LLMs
bsky.app/profile/itay...
🚨New paper alert🚨

🧠
Instruction-tuned LLMs show amplified cognitive biases — but are these new behaviors, or pretraining ghosts resurfacing?

Excited to share our new paper, accepted to CoLM 2025🎉!
See thread below 👇
#BiasInAI #LLMs #MachineLearning #NLProc
October 7, 2025 at 1:41 PM
I think a scalable open source implementation would have many uses! Let’s say I can’t run all pretraining data because of cost. And I run a subset and get influential examples. What would that mean concerning what I’m missing?
April 25, 2025 at 11:59 AM
Looks great! What would it take to run this on another model and dataset?
April 25, 2025 at 12:31 AM
This has been a huge team effort with many talented contributors. Very thankful for everyone’s contributions!

See the list here:
bsky.app/profile/amuu...
This was a huge collaboration with many great folks! If you get a chance, be sure to talk to Atticus Geiger, @sarah-nlp.bsky.social, @danaarad.bsky.social, Iván Arcuschin, @adambelfki.bsky.social, @yiksiu.bsky.social, Jaden Fiotto-Kaufmann, @talhaklay.bsky.social, @michaelwhanna.bsky.social, ...
April 24, 2025 at 2:20 AM