📅 Key Dates
Abstract deadline: Jan 23, 2026 AOE
Paper deadline: Jan 28, 2026 AOE
A few key changes this year:
- Attendance for authors of accepted papers is optional
- Originally submitted version of accepted papers will be made public
...
📅 Key Dates
Abstract deadline: Jan 23, 2026 AOE
Paper deadline: Jan 28, 2026 AOE
A few key changes this year:
- Attendance for authors of accepted papers is optional
- Originally submitted version of accepted papers will be made public
...
Anthropic sweetly asked Sonnet about its preferences in how it wanted to be deprecated
in addition:
- no, still not open weights
- preserve weights and keeping it running internally
- letting models pursue their interests
www.anthropic.com/research/dep...
careers.epfl.ch/job/Lausanne...
careers.epfl.ch/job/Lausanne...
How do #LLMs’ inner features change as they train? Using #crosscoders + a new causal metric, we map when features appear, strengthen, or fade across checkpoints—opening a new lens on training dynamics beyond loss curves & benchmarks.
#interpretability
How do #LLMs’ inner features change as they train? Using #crosscoders + a new causal metric, we map when features appear, strengthen, or fade across checkpoints—opening a new lens on training dynamics beyond loss curves & benchmarks.
#interpretability
https://www.heise.de/hintergrund/Schweizer-Sprachmodell-Apertus-So-sieht-EU-konforme-transparente-KI-aus-10638501.html?utm_source=flipboard&utm_medium=activitypub
Gepostet in Nachrichten @nachrichten-heiseonline
https://www.heise.de/hintergrund/Schweizer-Sprachmodell-Apertus-So-sieht-EU-konforme-transparente-KI-aus-10638501.html?utm_source=flipboard&utm_medium=activitypub
Gepostet in Nachrichten @nachrichten-heiseonline
The largest publicly available corpus sourced exclusively from PDFs, containing about 3 trillion tokens across 475 million documents in 1733 languages.
- Long context
- 3T tokens from high-demand domains like legal and science.
- Heavily improves over SoTA
The largest publicly available corpus sourced exclusively from PDFs, containing about 3 trillion tokens across 475 million documents in 1733 languages.
- Long context
- 3T tokens from high-demand domains like legal and science.
- Heavily improves over SoTA
pip install mlx-lm
mlx_lm.generate --model swiss-ai/Apertus-8B-Instruct-2509 --prompt "wer bisch du?"
(make sure you have done huggingface-cli login before)
pip install mlx-lm
mlx_lm.generate --model swiss-ai/Apertus-8B-Instruct-2509 --prompt "wer bisch du?"
(make sure you have done huggingface-cli login before)
WBK N will weder Opt Out noch Opt In festschreiben.
WBK N will weder Opt Out noch Opt In festschreiben.
Trained on 15T tokens in 1,000+ languages, it’s built for transparency, responsibility & the public good.
Read more: actu.epfl.ch/news/apertus...
arxiv.org/abs/2509.01440
arxiv.org/abs/2509.01440
Fürs MAZ habe ich Apertus kurz analysiert:
www.maz.ch/news/apertus...
Fürs MAZ habe ich Apertus kurz analysiert:
www.maz.ch/news/apertus...
Trained on 15T tokens in 1,000+ languages, it’s built for transparency, responsibility & the public good.
Read more: actu.epfl.ch/news/apertus...
Trained on 15T tokens in 1,000+ languages, it’s built for transparency, responsibility & the public good.
Read more: actu.epfl.ch/news/apertus...
Fully open and multilingual, the model is trained on CSCS's supercomputer "Alps" and supports sovereign, transparent, and responsible AI in Switzerland and beyond.
Read more here: ai.epfl.ch/a-language-m...
#ResponsibleAI
Fully open and multilingual, the model is trained on CSCS's supercomputer "Alps" and supports sovereign, transparent, and responsible AI in Switzerland and beyond.
Read more here: ai.epfl.ch/a-language-m...
#ResponsibleAI
Because someone at Musk’s xAi deliberately did this, and we only found out because they were clumsy.
My piece on the real dangers of AI.
Gift link:
www.nytimes.com/2025/05/17/o...
Because someone at Musk’s xAi deliberately did this, and we only found out because they were clumsy.
My piece on the real dangers of AI.
Gift link:
www.nytimes.com/2025/05/17/o...
I will be there to chat about all things multilingual & multicultural evaluation.
Feel free to reach out anytime during the conference. I’d love to connect!
Contains *newly-collected* data, prioritizing *regional knowledge*.
Setting the stage for truly global AI evaluation.
Ready to see how your model measures up?
#AI #Multilingual #LLM #NLProc
I will be there to chat about all things multilingual & multicultural evaluation.
Feel free to reach out anytime during the conference. I’d love to connect!
We propose a simple classifier-based selection, enabling multilingual LLMs 🧵
We propose a simple classifier-based selection, enabling multilingual LLMs 🧵
arxiv.org/abs/2504.05295
arxiv.org/abs/2504.05295
The first decentralized 32B-parameter RL training run open to join for anyone with compute — fully permissionless.
www.primeintellect.ai/blog/intelle...
The first decentralized 32B-parameter RL training run open to join for anyone with compute — fully permissionless.
www.primeintellect.ai/blog/intelle...
She will be joining University of Zurich as a professor this summer, and hiring PhD students and postdocs. You should apply to her group!
Her website: koloskova.github.io
She will be joining University of Zurich as a professor this summer, and hiring PhD students and postdocs. You should apply to her group!
Her website: koloskova.github.io
Send your idea by end of March 🏃♂️➡️ , and run on one of the largest public AI clusters globally. Everyone is eligible to apply!
swiss-ai.org
Send your idea by end of March 🏃♂️➡️ , and run on one of the largest public AI clusters globally. Everyone is eligible to apply!
swiss-ai.org