wordpredictabilityvisualized.vercel.app
wordpredictabilityvisualized.vercel.app
gpt2-small-dutch (huggingface.co/GroNLP/gpt2-...) or gpt2-medium-dutch-embeddings (huggingface.co/GroNLP/gpt2-...) are great options.
gpt2-small-dutch (huggingface.co/GroNLP/gpt2-...) or gpt2-medium-dutch-embeddings (huggingface.co/GroNLP/gpt2-...) are great options.
For very unpredictable words, a decrease in predictability has a much larger slowing-down effect on reading times than the same decrease for highly predictable words.
For very unpredictable words, a decrease in predictability has a much larger slowing-down effect on reading times than the same decrease for highly predictable words.
1. Smaller Dutch models often predict reading times better (= inverse scaling trend) ~ in line with evidence of English models.
But, with more context (in a book reading corpus), larger models catch up.
1. Smaller Dutch models often predict reading times better (= inverse scaling trend) ~ in line with evidence of English models.
But, with more context (in a book reading corpus), larger models catch up.
But, most evidence so far is limited to English.
How well do Dutch open-source language models fit reading times using their word predictability estimates?
But, most evidence so far is limited to English.
How well do Dutch open-source language models fit reading times using their word predictability estimates?
Check out our full dataset and code here:
osf.io/wr4qf/
Check out our full dataset and code here:
osf.io/wr4qf/
A challenge to the notion of an universal linear effect of surprisal.
A challenge to the notion of an universal linear effect of surprisal.
However, this effect varied depending on the corpus used.
However, this effect varied depending on the corpus used.
Does surprisal still show linear link with reading times when estimated with a Dutch-specific language model as opposed to a multilingual model?
Does surprisal still show linear link with reading times when estimated with a Dutch-specific language model as opposed to a multilingual model?
Do these Dutch-specific LLMs show a similar inverse scaling trend as English models?
That is, do the smaller transformer models' surprisal estimates account better for reading times than those of the very large models?
Do these Dutch-specific LLMs show a similar inverse scaling trend as English models?
That is, do the smaller transformer models' surprisal estimates account better for reading times than those of the very large models?
What is the best computational method for estimating word predictability in Dutch?
We compare 14 Dutch large language models (LLMs), a multilingual model (mGPT) and an N-gram model in their ability of explaining reading times.
What is the best computational method for estimating word predictability in Dutch?
We compare 14 Dutch large language models (LLMs), a multilingual model (mGPT) and an N-gram model in their ability of explaining reading times.
We adressed this and asked three questions:
We adressed this and asked three questions: