Tomás Vergara Browne
tomvergara.bsky.social
Tomás Vergara Browne
@tomvergara.bsky.social
Interp & analysis in NLP

Mostly 🇦🇷, slightly 🇨🇱
Reposted by Tomás Vergara Browne
Our new paper in #PNAS (bit.ly/4fcWfma) presents a surprising finding—when words change meaning, older speakers rapidly adopt the new usage; inter-generational differences are often minor.

w/ Michelle Yang, ‪@sivareddyg.bsky.social‬ , @msonderegger.bsky.social‬ and @dallascard.bsky.social‬👇(1/12)
July 29, 2025 at 12:06 PM
Reposted by Tomás Vergara Browne
Started a new podcast with @tomvergara.bsky.social !

Behind the Research of AI:
We look behind the scenes, beyond the polished papers 🧐🧪

If this sounds fun, check out our first "official" episode with the awesome Gauthier Gidel
from @mila-quebec.bsky.social :

open.spotify.com/episode/7oTc...
02 | Gauthier Gidel: Bridging Theory and Deep Learning, Vibes at Mila, and the Effects of AI on Art
Behind the Research of AI · Episode
open.spotify.com
June 25, 2025 at 3:54 PM
Reposted by Tomás Vergara Browne
Overall I loved the paper, got lots of inspiration from it and would love to be part of a similar project in the future: for example an empirical investigation of many AI papers to answer "To what extent is AI is a science?"
April 15, 2025 at 9:56 PM
Reposted by Tomás Vergara Browne
Models like DeepSeek-R1 🐋 mark a fundamental shift in how LLMs approach complex problems. In our preprint on R1 Thoughtology, we study R1’s reasoning chains across a variety of tasks; investigating its capabilities, limitations, and behaviour.
🔗: mcgill-nlp.github.io/thoughtology/
April 1, 2025 at 8:07 PM
Reposted by Tomás Vergara Browne
Instruction-following retrievers can efficiently and accurately search for harmful and sensitive information on the internet! 🌐💣

Retrievers need to be aligned too! 🚨🚨🚨

Work done with the wonderful Nick and @sivareddyg.bsky.social

🔗 mcgill-nlp.github.io/malicious-ir/
Thread: 🧵👇
Exploiting Instruction-Following Retrievers for Malicious Information Retrieval
Parishad BehnamGhader, Nicholas Meade, Siva Reddy
mcgill-nlp.github.io
March 12, 2025 at 4:15 PM
Reposted by Tomás Vergara Browne
Agents like OpenAI Operator can solve complex computer tasks, but what happens when users use them to cause harm, e.g. spread misinformation?

To find out, we introduce SafeArena (safearena.github.io), a benchmark to assess the capabilities of web agents to complete harmful web tasks. A thread 👇
March 10, 2025 at 5:45 PM
Reposted by Tomás Vergara Browne
After a fun and long #EMNLP2024 I'm now travelling AGAIN to Uppsala 🇸🇪, to speak at the Transdisciplinary Queer Futures of AI Conference! Any Sweden/Uppsala recs?
November 20, 2024 at 1:27 PM