Ethan Gotlieb Wilcox
banner
wegotlieb.bsky.social
Ethan Gotlieb Wilcox
@wegotlieb.bsky.social
Assistant Professor of Computational Linguistics @ Georgetown; formerly postdoc @ ETH Zurich; PhD @ Harvard Linguistics, affiliated with MIT Brain & Cog Sci. Language, Computers, Cognition.
I did not! Yikes! Another reason to include "pickle" and/or pickle-related emoji in any lab communication!
October 23, 2025 at 1:04 AM
Georgetown Linguistics has a dedicated Computational Linguistics PhD track, and a lively CL community on campus (gucl.georgetown.edu), including my faculty colleagues @complingy.bsky.social and Amir Zeldes.
GUCL: Computation and Language @ Georgetown
gucl.georgetown.edu
October 21, 2025 at 9:52 PM
PICoL stands for “Psycholinguistics, Information, and Computational Linguistics,” and I encourage applications from anyone whose research interests connect with these topics!
October 21, 2025 at 9:52 PM
We see this project as in line with some other recent papers seeking to cast typological variation in information-theoretic terms, with shout-outs to Michaela Socolof, @postylem.bsky.social @futrell.bsky.social (aclanthology.org/2022.coling-...) and Julius Steuer (aclanthology.org/2023.sigtyp-...)
Measuring Morphological Fusion Using Partial Information Decomposition
Michaela Socolof, Jacob Louis Hoover, Richard Futrell, Alessandro Sordoni, Timothy J. O’Donnell. Proceedings of the 29th International Conference on Computational Linguistics. 2022.
aclanthology.org
May 13, 2025 at 1:21 PM
⭐ ⭐This paper also makes several technical contributions to the mixed-pair mutual information estimation pipeline of Wolf et al., (aclanthology.org/2023.emnlp-m...). Shout out to @cuiding.bsky.social for all her hard work on this aspect of the paper! ⭐⭐
Quantifying the redundancy between prosody and text
Lukas Wolf, Tiago Pimentel, Evelina Fedorenko, Ryan Cotterell, Alex Warstadt, Ethan Wilcox, Tamar Regev. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023.
aclanthology.org
May 13, 2025 at 1:21 PM
✅In line with our prediction, we find that mutual information is higher in tonal languages than in non-tonal languages. BUT, the way one represents context is important. When full sentential context is taken into account (mBERT and mGPT), the distinction collapses.
May 13, 2025 at 1:21 PM
🌏🌍We test this prediction by estimating mutual information in an audio dataset of 10 different languages across 6 language families. 🌏🌍
May 13, 2025 at 1:21 PM
We propose a way to do so using …📡information theory.📡 In tonal languages, pitch reduces uncertainty about lexical identity, therefore, the mutual information between pitch and words should be higher.
May 13, 2025 at 1:21 PM
🌐But there are intermediate languages, which have lexically contrastive tone, but only sporadically, making some linguists doubt the tonal/non-tonal dichotomy. So, how can we measure how “tonal” a language is? 🧐🧐
May 13, 2025 at 1:21 PM
🌏 Different languages use pitch in different ways. 🌏 “Tonal” languages, like Cantonese, use it to make lexical distinctions. 📖 While others, like English, use it for other functions, like marking whether or not a sentence is a question. ❓
May 13, 2025 at 1:21 PM
I’ll also use this as a way to plug human-scale language modeling in the wild: This year’s BabyLM eval pipeline was just released last week at github.com/babylm/evalu.... For more info on BabyLM head to babylm.github.io
GitHub - babylm/evaluation-pipeline-2025
Contribute to babylm/evaluation-pipeline-2025 development by creating an account on GitHub.
github.com
May 12, 2025 at 3:48 PM
Couldn’t be happier to have co-authored this will a stellar team, including: Michael Hu, @amuuueller.bsky.social, @alexwarstadt.bsky.social, @lchoshen.bsky.social, Chengxu Zhuang, @adinawilliams.bsky.social, Ryan Cotterell, @tallinzen.bsky.social
May 12, 2025 at 3:48 PM
This version includes 😱New analyses 😱new arguments 😱 and a whole new “Looking Forward” section! If you’re interested in what a team of (psycho) computational linguists thinks the future will hold, check out our brand new Section 8!
May 12, 2025 at 3:48 PM
⚖️📣This paper was a big departure from my typical cognitive science fare, and so much fun to write! 📣⚖️ Thank you to @bwal.bsky.social and especially to @kevintobia.bsky.social for their legal expertise on this project!
February 19, 2025 at 2:25 PM
On the positive side, we suggest that LLMs can serve a role as “dialectic” partners 🗣️❔🗣️ helping judges and clerks strengthen their arguments, as long as judicial sovereignty is maintained 👩‍⚖️👑👩‍⚖️
February 19, 2025 at 2:25 PM
⚖️ We also show, through demonstration, that it’s very easy to engineer prompts that steer models toward one’s desired interpretation of a word or phrase. 📖Prompting is the new “dictionary shopping” 😬 📖 😬
February 19, 2025 at 2:25 PM
🏛️We identify five “myths” about LLMs which, when dispelled, reveal their limitations as legal tools for textual interpretation. To take one example, during instruction tuning, LLMs are trained on highly structured, non-natural inputs.
February 19, 2025 at 2:25 PM