Martin Jaggi
mjaggi.bsky.social
Martin Jaggi
@mjaggi.bsky.social
Prof at EPFL
AI • Climbing
June 20, 2025 at 8:08 PM
Our method results in >6x faster pre-training of LLMs across many languages. The approach is applicable to both high and low-resource languages, and also improves upon its English-speaking DCLM cousin, which inspired this work.
April 23, 2025 at 5:06 AM
Using the 'right' data can hugely speed up LLM training, but how to find the best training data in the vast sea of a whole web crawl?

We propose a simple classifier-based selection, enabling multilingual LLMs 🧵
April 23, 2025 at 5:06 AM
The Swiss AI Initiative has launched open calls for disruptive ideas - Democratizing large-scale AI for the benefit of society.

Send your idea by end of March 🏃‍♂️‍➡️ , and run on one of the largest public AI clusters globally. Everyone is eligible to apply!

swiss-ai.org
March 4, 2025 at 11:13 PM
new open weights, 24B model, with comparable performance to Llama 3.3 70B 😮. congrats mistral team!
mistral.ai/news/mistral...
January 30, 2025 at 7:01 PM