Antoine Bosselut
abosselut.bsky.social
Antoine Bosselut
@abosselut.bsky.social
Helping machines make sense of the world. Asst Prof @icepfl.bsky.social; Before: @stanfordnlp.bsky.social @uwnlp.bsky.social AI2 #NLProc #AI

Website: https://atcbosselut.github.io/
I don't see why the answer would be no but since you specifically say "October", what if we submitted to ARR in July and want to do early submission to ACL 2026 ?
September 29, 2025 at 8:03 PM
Special thanks to everyone that participated in this journey!
September 3, 2025 at 9:26 AM
(5) Transparency: We're fully open, pairing our weights with a full suite of reproduction artifacts.

Check out our artifacts and technical report here: huggingface.co/swiss-ai
swiss-ai (Swiss AI Initiative)
Org profile for Swiss AI Initiative on Hugging Face, the AI community building the future.
huggingface.co
September 3, 2025 at 9:26 AM
(4) Multilinguality: We pretrain the model on 15T tokens from 1811 languages, and post-train with 3.8 M examples from 149 languages
September 3, 2025 at 9:26 AM
(3) Memorization Prevention: Adopting the Goldfish objective, we suppress verbatim recall and reduce risks of memorization
September 3, 2025 at 9:26 AM
(2) Data Compliance: we pretrained exclusively on openly available data, retroactively respecting robots.txt exclusions and filtering for copyrighted, non-permissive, toxic, and personally identifiable content
September 3, 2025 at 9:26 AM
What makes Apertus special?
(1) Scale: Apertus-70B is the first fully open model to be trained at 70B parameter scale on 15T tokens, requiring us to scale out training to 4096 GPUs at
@cscsch.bsky.social
September 3, 2025 at 9:26 AM
Thank you for your incredible work!
September 2, 2025 at 6:23 PM