Stratis Tsirtsis
stratiss.bsky.social
Stratis Tsirtsis
@stratiss.bsky.social
Postdoc @ Hasso Plattner Institute working on machine learning. Previously @ Max Planck Institute, Meta, Stanford, NTUA.
💻 https://stsirtsis.github.io/
Pinned
todo:
* thesis defense ✅

Grateful to the committee and reviewers Marius Kloft, @arkrause.bsky.social, @rupakmajumdar.bsky.social, and @tobigerstenberg.bsky.social for their time and support. No words are enough to thank my advisor @autreche.bsky.social for everything I’ve learned from him so far 🙏
What if AI agents aren't here to replace us, but to facilitate our decisions? In a study with 1600 participants, we show that a human with action choices narrowed by an AI makes better sequential decisions than an AI or a human alone.
📜 arxiv.org/abs/2510.16097
October 22, 2025 at 11:45 AM
Reposted by Stratis Tsirtsis
The Causality in Cognition Lab at Stanford University is recruiting PhD students this cycle!

We are a supportive team who happened to wear bluesky appropriate colors for the lab photo (this wasn't planned). 💙

Lab info: cicl.stanford.edu
Application details: psychology.stanford.edu/admissions/p...
October 17, 2025 at 5:43 PM
Reposted by Stratis Tsirtsis
We (w/ Moritz Hardt, Olawale Salaudeen and
@joavanschoren.bsky.social) are organizing the Workshop on the Science of Benchmarking & Evaluating AI @euripsconf.bsky.social 2025 in Copenhagen!

📢 Call for Posters: rb.gy/kyid4f
📅 Deadline: Oct 10, 2025 (AoE)
🔗 More info: rebrand.ly/bg931sf
September 22, 2025 at 1:45 PM
Reposted by Stratis Tsirtsis
So excited and honored to receive an ERC Starting Grant for the project BrainAlign!! BrainAlign will bring LLMs closer to human understanding by directly aligning them with the human brain.

Stay tuned for our findings, and multiple postdoc and PhD openings in the coming years!
September 4, 2025 at 4:21 PM
todo:
* thesis defense ✅

Grateful to the committee and reviewers Marius Kloft, @arkrause.bsky.social, @rupakmajumdar.bsky.social, and @tobigerstenberg.bsky.social for their time and support. No words are enough to thank my advisor @autreche.bsky.social for everything I’ve learned from him so far 🙏
August 20, 2025 at 5:22 PM
Last week I had the pleasure of presenting a 2.5-hour tutorial on "Counterfactuals in Minds and Machines" at UAI 2025 in Rio 🇧🇷, prepared together with @autreche.bsky.social and @tobigerstenberg.bsky.social. We've made all materials and references available here: learning.mpi-sws.org/counterfactu...
August 1, 2025 at 11:40 AM
In Athens 🇬🇷 for the Greeks in AI symposium. Super excited to present our work on "Counterfactual Token Generation in LLMs" (bit.ly/4nMibs2) and see all the amazing work Greek people all over the world are doing on AI! If you are in Athens, let's meet! Next, heading to👇
July 18, 2025 at 12:38 PM
Reposted by Stratis Tsirtsis
did you check our amazing list of tutorials in Rio?
spanning

- hyperparameter optimization
- counterfactual reasoning
- bayesian nonparametrics for causality
- causal inference with deep generative models
- modern variational inference

👉 www.auai.org/uai2025/tuto...
Uncertainty in Artificial Intelligence
www.auai.org
June 4, 2025 at 9:25 AM
The LLM API you use returns (and charges you for) 5 tokens. Did the LLM actually generate 5 tokens? Or is the provider overcharging you? 🤔 In arxiv.org/abs/2505.21627, led by Ander Artola Velasco, we argue (game-theoretically) for a change from pay-per-token to pay-per-character.
May 30, 2025 at 11:24 AM
Presenting this today at 17:00 in Hall 4 #6
In Singapore for #ICLR2025! I'll be presenting our work on a causal methodology for evaluating LLMs (arxiv.org/abs/2502.01754) at the "Building Trust in LLMs" workshop on Monday. If you are working on causality, game theory and/or LLMs, let's grab a ☕️ during the conference!
April 28, 2025 at 4:57 AM
In Singapore for #ICLR2025! I'll be presenting our work on a causal methodology for evaluating LLMs (arxiv.org/abs/2502.01754) at the "Building Trust in LLMs" workshop on Monday. If you are working on causality, game theory and/or LLMs, let's grab a ☕️ during the conference!
April 25, 2025 at 1:53 PM
Reposted by Stratis Tsirtsis
LLMs rely on randomization to respond to a prompt: they may respond differently to the same prompt if asked multiple times. In “Evaluation of LLMs via Coupled Token Generation” (arxiv.org/abs/2502.01754), we argue that the eval of LLMs should control for this randomization 1/
Evaluation of Large Language Models via Coupled Token Generation
State of the art large language models rely on randomization to respond to a prompt. As an immediate consequence, a model may respond differently to the same prompt if asked multiple times. In this wo...
arxiv.org
February 5, 2025 at 8:33 AM
Let's talk causality and LLMs! Come find us at the posters in East Hall C. 11:30-12:00 & 14:30-15:00. #neurips2024
What would an LLM have said, counterfactually? Here is a short video illustrating our method for counterfactual token generation. We will present this work at the CaLM workshop at #neurips2024. See you in Vancouver!
📜 arxiv.org/abs/2409.17027
💻 made with manim in python
December 14, 2024 at 3:23 PM
What would an LLM have said, counterfactually? Here is a short video illustrating our method for counterfactual token generation. We will present this work at the CaLM workshop at #neurips2024. See you in Vancouver!
📜 arxiv.org/abs/2409.17027
💻 made with manim in python
November 27, 2024 at 5:24 PM
Hey there 🦋
Let's start with an intro. I'm a final-year PhD student at the Max Planck Institute for Software Systems, working on machine learning, decision making and social aspects of AI. Currently on the academic job market, looking for tenure-track positions👇
💻 stsirtsis.github.io
November 20, 2024 at 12:55 PM