PhD from LMU Munich, prev. UT Austin, Princeton, @ltiatcmu.bsky.social, Cambridge
computational linguistics, construction grammar, morphosyntax
leonieweissweiler.github.io
📰Check out the paper: nejlt.ep.liu.se/article/view...
💻The full dataset and code are available on GitHub: github.com/LeonieWeissw...
🧵7/7
📰Check out the paper: nejlt.ep.liu.se/article/view...
💻The full dataset and code are available on GitHub: github.com/LeonieWeissw...
🧵7/7
In grey are cases where the model struggles to answer both questions, in red the cases where it would have needed to reply on the caused-motion semantics.
🧵6/7
In grey are cases where the model struggles to answer both questions, in red the cases where it would have needed to reply on the caused-motion semantics.
🧵6/7
🧵5/7
🧵5/7
For example, giving examples and asking for a json with the sentences and labels is always a good idea, but using o1 over 4o-mini is only worth it if human annotation costs more than .5$ per sentence!
🧵4/7
For example, giving examples and asking for a json with the sentences and labels is always a good idea, but using o1 over 4o-mini is only worth it if human annotation costs more than .5$ per sentence!
🧵4/7
To reduce this cost, we use few-shot prompt-based filtering, which greatly reduces the number of FPs that our human annotator will have to sift through, and therefore the annotation cost.
🧵3/7
To reduce this cost, we use few-shot prompt-based filtering, which greatly reduces the number of FPs that our human annotator will have to sift through, and therefore the annotation cost.
🧵3/7
To construct our dataset, we first create a dependency filter based on the syntactic side of the construction.
🧵2/7
To construct our dataset, we first create a dependency filter based on the syntactic side of the construction.
🧵2/7