Using fine-tuned Qwen2.5-VL and TransOCR, we estimated the MI between images and word identity.
MI systematically drops: Full > Upper > Lower — perfectly mirroring human reading patterns! 🤯
Using fine-tuned Qwen2.5-VL and TransOCR, we estimated the MI between images and word identity.
MI systematically drops: Full > Upper > Lower — perfectly mirroring human reading patterns! 🤯
Higher mutual information → fewer samples → faster reading.
Higher mutual information → fewer samples → faster reading.
To test the theory, we created a reading experiment using the MoTR (Mouse-Tracking-for-Reading) paradigm 🖱️📖
We ran the study in both English and Chinese.
To test the theory, we created a reading experiment using the MoTR (Mouse-Tracking-for-Reading) paradigm 🖱️📖
We ran the study in both English and Chinese.
Our new #EMNLP2025 paper with @wegotlieb.bsky.social, Lena Jäger -- “Modeling Bottom-up Information Quality during Language Processing”, bridges psycholinguistics and multimodal LLMs.
🧠💡👇
arxiv.org/pdf/2509.17047
Our new #EMNLP2025 paper with @wegotlieb.bsky.social, Lena Jäger -- “Modeling Bottom-up Information Quality during Language Processing”, bridges psycholinguistics and multimodal LLMs.
🧠💡👇
arxiv.org/pdf/2509.17047