Working on multimodal fact-checking
https://jtonglet.github.io
Meet @igurevych.bsky.social, our postdocs, and our PhD students, and be sure to check out their posters 🎉
More information about our ACL and TACL papers below ⬇️
Meet @igurevych.bsky.social, our postdocs, and our PhD students, and be sure to check out their posters 🎉
More information about our ACL and TACL papers below ⬇️
We identify a key vulnerability of multimodal LLMs: they are easily deceived by misleading charts🫣
This vulnerability opens many doors for future work in automated chart understanding, be sure to check it out!
We show the extent of this vulnerability and how to fix it by modifying the misleading chart image or replacing it by another modality - 🧵(1/5)
📄Paper: arxiv.org/abs/2502.20503
💻Project: ukplab.github.io/arxiv2025-mi...
We identify a key vulnerability of multimodal LLMs: they are easily deceived by misleading charts🫣
This vulnerability opens many doors for future work in automated chart understanding, be sure to check it out!
We find that, like humans, multimodal LLMs are vulnerable to misleading charts, which severely impairs their chart QA abilities.
Check out our preprint, data, and code - more information ⬇️
We show the extent of this vulnerability and how to fix it by modifying the misleading chart image or replacing it by another modality - 🧵(1/5)
📄Paper: arxiv.org/abs/2502.20503
💻Project: ukplab.github.io/arxiv2025-mi...
We find that, like humans, multimodal LLMs are vulnerable to misleading charts, which severely impairs their chart QA abilities.
Check out our preprint, data, and code - more information ⬇️
»COVE: COntext and VEracity prediction for out-of-context images« by Jonathan Tonglet (@tongletj.bsky.social), Gabriel Thiem (@gthiem.bsky.social) & Iryna Gurevych (@igurevych.bsky.social)
»COVE: COntext and VEracity prediction for out-of-context images« by Jonathan Tonglet (@tongletj.bsky.social), Gabriel Thiem (@gthiem.bsky.social) & Iryna Gurevych (@igurevych.bsky.social)
go.bsky.app/tRMpng
go.bsky.app/tRMpng