Leonie Weissweiler
banner
weissweiler.bsky.social
Leonie Weissweiler
@weissweiler.bsky.social
Postdoc at Uppsala University Computational Linguistics with Joakim Nivre
PhD from LMU Munich, prev. UT Austin, Princeton, @ltiatcmu.bsky.social, Cambridge
computational linguistics, construction grammar, morphosyntax
leonieweissweiler.github.io
Oops, here are the three caused-motion examples that were meant to go in the first post:
November 19, 2025 at 6:20 PM
👥Joint work with Abdullatif Köksal and Hinrich Schütze

📰Check out the paper: nejlt.ep.liu.se/article/view...

💻The full dataset and code are available on GitHub: github.com/LeonieWeissw...

🧵7/7
View of Hybrid Human-LLM Corpus Construction and LLM Evaluation for the Caused-Motion Construction
nejlt.ep.liu.se
November 19, 2025 at 1:57 PM
We find that a range of models indeed struggle with this, but Gemma 27B solves it almost perfectly!

In grey are cases where the model struggles to answer both questions, in red the cases where it would have needed to reply on the caused-motion semantics.

🧵6/7
November 19, 2025 at 1:57 PM
Once we have manually annotated the final dataset in this way, we return to the original question and test if LLMs find the third question below easier than the second, which would indicate difficulties in making use of the semantics of the caused-motion construction.

🧵5/7
November 19, 2025 at 1:57 PM
Designing the right prompt is tricky and depends on annotation cost.

For example, giving examples and asking for a json with the sentences and labels is always a good idea, but using o1 over 4o-mini is only worth it if human annotation costs more than .5$ per sentence!

🧵4/7
November 19, 2025 at 1:57 PM
But this leaves many FPs, and filtering them by hand would be expensive.

To reduce this cost, we use few-shot prompt-based filtering, which greatly reduces the number of FPs that our human annotator will have to sift through, and therefore the annotation cost.

🧵3/7
November 19, 2025 at 1:57 PM
They are all instances of the so-called caused-motion construction, and collecting enough instances for testing was a challenge, given its rarity!

To construct our dataset, we first create a dependency filter based on the syntactic side of the construction.

🧵2/7
November 19, 2025 at 1:57 PM
Reposted by Leonie Weissweiler
There's many directions where this could go, multilingual, low-resource language, interpretability, depending on your profile, and the internship may lead to a PhD, provided we get funding!
November 6, 2025 at 9:07 AM
Reposted by Leonie Weissweiler
As we found in aclanthology.org/2025.coling-... that BPE-based LLMs (i.e. pretty much all LLMs) did not handle prefixations well
Unlike “Likely”, “Unlike” is Unlikely: BPE-based Segmentation hurts Morphological Derivations in LLMs
Paul Lerner, François Yvon. Proceedings of the 31st International Conference on Computational Linguistics. 2025.
aclanthology.org
November 6, 2025 at 9:06 AM
Reposted by Leonie Weissweiler
Basically the idea is to extend www.pnas.org/doi/10.1073/... to see how well LLMs model competition between affixes, not only suffixes (e.g. -ity vs. -ness) but also prefixes (e.g. un- vs. non-)
Derivational morphology reveals analogical generalization in large language models | PNAS
What mechanisms underlie linguistic generalization in large language models (LLMs)? This question has attracted considerable attention, with most s...
www.pnas.org
November 6, 2025 at 9:04 AM
This is great! Do you happen to also have a list of nice examples of lab websites? 🙂
November 4, 2025 at 12:15 PM
@kanishka.bsky.social congratulations on making the list!!
November 4, 2025 at 12:10 PM
Reposted by Leonie Weissweiler
Session 5: "BabyLM’s First Constructions." In a companion paper to the first one, we show that much of this constructional knowledge is even present in "babyLMs" with more developmentally plausible amounts of training: arxiv.org/abs/2506.02147. w/ @weissweiler.bsky.social
BabyLM's First Constructions: Causal probing provides a signal of learning
Construction grammar posits that language learners acquire constructions (form-meaning pairings) from the statistics of their environment. Recent work supports this hypothesis by showing sensitivity…
arxiv.org
November 2, 2025 at 9:57 PM
Reposted by Leonie Weissweiler
Session 4: "Constructions are Revealed in Word Distributions." We show that rich knowledge of syntactic constructions in masked language models (MLMs) is revealed by patterns of contextual constraint. arxiv.org/abs/2503.06048. w/ @weissweiler.bsky.social + @kmahowald.bsky.social
Constructions are Revealed in Word Distributions
Construction grammar posits that constructions, or form-meaning pairings, are acquired through experience with language (the distributional learning hypothesis). But how much information about…
arxiv.org
November 2, 2025 at 9:57 PM
I'll be dreaming of them once Swedish winter starts...
September 15, 2025 at 3:38 PM