Antoine Bosselut
@abosselut.bsky.social
Helping machines make sense of the world. Asst Prof @icepfl.bsky.social; Before: @stanfordnlp.bsky.social @uwnlp.bsky.social AI2 #NLProc #AI
Website: https://atcbosselut.github.io/
Website: https://atcbosselut.github.io/
EPFL AI Center Postdocs: www.epfl.ch/research/fun...
NLP Lab Postdoc: docs.google.com/document/d/1...
NLP Lab Postdoc: docs.google.com/document/d/1...
EPFL AI Center and Swiss AI Initiative Postdoctoral Fellowships
The 2nd call is now open with a deadline for submissions of 3 November (17.00 CET)!Applications are encouraged from researchers at the postdoctoral level with a keen interest in collaborative, interdi...
www.epfl.ch
October 14, 2025 at 5:56 PM
EPFL AI Center Postdocs: www.epfl.ch/research/fun...
NLP Lab Postdoc: docs.google.com/document/d/1...
NLP Lab Postdoc: docs.google.com/document/d/1...
I don't see why the answer would be no but since you specifically say "October", what if we submitted to ARR in July and want to do early submission to ACL 2026 ?
September 29, 2025 at 8:03 PM
I don't see why the answer would be no but since you specifically say "October", what if we submitted to ARR in July and want to do early submission to ACL 2026 ?
Special thanks to everyone that participated in this journey!
September 3, 2025 at 9:26 AM
Special thanks to everyone that participated in this journey!
(5) Transparency: We're fully open, pairing our weights with a full suite of reproduction artifacts.
Check out our artifacts and technical report here: huggingface.co/swiss-ai
Check out our artifacts and technical report here: huggingface.co/swiss-ai
swiss-ai (Swiss AI Initiative)
Org profile for Swiss AI Initiative on Hugging Face, the AI community building the future.
huggingface.co
September 3, 2025 at 9:26 AM
(5) Transparency: We're fully open, pairing our weights with a full suite of reproduction artifacts.
Check out our artifacts and technical report here: huggingface.co/swiss-ai
Check out our artifacts and technical report here: huggingface.co/swiss-ai
(4) Multilinguality: We pretrain the model on 15T tokens from 1811 languages, and post-train with 3.8 M examples from 149 languages
September 3, 2025 at 9:26 AM
(4) Multilinguality: We pretrain the model on 15T tokens from 1811 languages, and post-train with 3.8 M examples from 149 languages
(3) Memorization Prevention: Adopting the Goldfish objective, we suppress verbatim recall and reduce risks of memorization
September 3, 2025 at 9:26 AM
(3) Memorization Prevention: Adopting the Goldfish objective, we suppress verbatim recall and reduce risks of memorization
(2) Data Compliance: we pretrained exclusively on openly available data, retroactively respecting robots.txt exclusions and filtering for copyrighted, non-permissive, toxic, and personally identifiable content
September 3, 2025 at 9:26 AM
(2) Data Compliance: we pretrained exclusively on openly available data, retroactively respecting robots.txt exclusions and filtering for copyrighted, non-permissive, toxic, and personally identifiable content
What makes Apertus special?
(1) Scale: Apertus-70B is the first fully open model to be trained at 70B parameter scale on 15T tokens, requiring us to scale out training to 4096 GPUs at
@cscsch.bsky.social
(1) Scale: Apertus-70B is the first fully open model to be trained at 70B parameter scale on 15T tokens, requiring us to scale out training to 4096 GPUs at
@cscsch.bsky.social
September 3, 2025 at 9:26 AM
What makes Apertus special?
(1) Scale: Apertus-70B is the first fully open model to be trained at 70B parameter scale on 15T tokens, requiring us to scale out training to 4096 GPUs at
@cscsch.bsky.social
(1) Scale: Apertus-70B is the first fully open model to be trained at 70B parameter scale on 15T tokens, requiring us to scale out training to 4096 GPUs at
@cscsch.bsky.social
Thank you for your incredible work!
September 2, 2025 at 6:23 PM
Thank you for your incredible work!