Machine Learning Engineer at 🤗 Hugging Face
Originally developed at the UKP Lab at @tuda.bsky.social, Sentence Transformers has become one of the world’s most widely used open-source libraries for semantic embeddings in natural language processing.
(1/🧵)
Originally developed at the UKP Lab at @tuda.bsky.social, Sentence Transformers has become one of the world’s most widely used open-source libraries for semantic embeddings in natural language processing.
(1/🧵)
This formalizes the existing maintenance structure, as I've personally led the project for the past two years on behalf of Hugging Face. I'm super excited about the transfer!
Details in 🧵
This formalizes the existing maintenance structure, as I've personally led the project for the past two years on behalf of Hugging Face. I'm super excited about the transfer!
Details in 🧵
Their blogpost covers all changes, including easier evaluation, multimodal support, rerankers, new interfaces, documentation, dataset statistics, a migration guide, etc.
🧵
Their blogpost covers all changes, including easier evaluation, multimodal support, rerankers, new interfaces, documentation, dataset statistics, a migration guide, etc.
🧵
It's a new multilingual text embedding retrieval benchmark with private (!) datasets, to ensure that we measure true generalization and avoid (accidental) overfitting.
Details in our blogpost below 🧵
It's a new multilingual text embedding retrieval benchmark with private (!) datasets, to ensure that we measure true generalization and avoid (accidental) overfitting.
Details in our blogpost below 🧵
It's a small patch release that makes the project more explicit with incorrect arguments and introduces some fixes for multi-GPU processing, evaluators, and hard negatives mining.
Details in 🧵
It's a small patch release that makes the project more explicit with incorrect arguments and introduces some fixes for multi-GPU processing, evaluators, and hard negatives mining.
Details in 🧵
One of the most requested models I've seen, @jhuclsp.bsky.social has trained state-of-the-art massively multilingual encoders using the ModernBERT architecture: mmBERT.
Stronger than an existing models at their sizes, while also much faster!
Details in 🧵
One of the most requested models I've seen, @jhuclsp.bsky.social has trained state-of-the-art massively multilingual encoders using the ModernBERT architecture: mmBERT.
Stronger than an existing models at their sizes, while also much faster!
Details in 🧵
Details in 🧵:
Details in 🧵:
Details in 🧵:
Details in 🧵:
See 🧵for the deets:
See 🧵for the deets:
See more in huggingface.co/openai
See more in huggingface.co/openai
🧵
🧵
Details in 🧵:
Details in 🧵:
Details in 🧵
Details in 🧵
Details in 🧵
Details in 🧵
Details in 🧵
Details in 🧵
The new patch introduces compatibility with the latest transformers and sentence-transformers versions.
The new patch introduces compatibility with the latest transformers and sentence-transformers versions.
Details in 🧵
Details in 🧵
These models were introduced in the Rank-DistiLLM paper, and distilled from LLMs like RankZephyr and RankGPT4.
Details in 🧵
These models were introduced in the Rank-DistiLLM paper, and distilled from LLMs like RankZephyr and RankGPT4.
Details in 🧵
I finetuned ModernBERT on my dataset, TriviaQA, just on my normal PC in a few hours. It outperforms all open models on the market on my eval set.
Finetuning is so worth it.
Blog in 🧵
I finetuned ModernBERT on my dataset, TriviaQA, just on my normal PC in a few hours. It outperforms all open models on the market on my eval set.
Finetuning is so worth it.
Blog in 🧵
I also prove that finetuning on your domain helps much more than you might think.
Details in 🧵
I also prove that finetuning on your domain helps much more than you might think.
Details in 🧵
Their new mxbai-rerank-...-v2 line is quite novel, using GRPO (akin to DeepSeek-R1), contrastive learning, and ranking objectives.
Details in 🧵
Their new mxbai-rerank-...-v2 line is quite novel, using GRPO (akin to DeepSeek-R1), contrastive learning, and ranking objectives.
Details in 🧵
Tried it again via hf.co/papers, same prompt, and the paper I wanted is literally the first result. 🩷🤗🩷
Tried it again via hf.co/papers, same prompt, and the paper I wanted is literally the first result. 🩷🤗🩷
It's a state-of-the-art multilingual encoder for 15 European languages, designed to be finetuned for retrieval, classification, etc.
Details in 🧵
It's a state-of-the-art multilingual encoder for 15 European languages, designed to be finetuned for retrieval, classification, etc.
Details in 🧵
Vibe checks still in the works.
Details & Links in 🧵
Vibe checks still in the works.
Details & Links in 🧵