@aubmindlab Alumni
Interested in AI, NLP, Video Games
wissamantoun.com
ModernBERT exhibits instabilities in downstream fine-tuning tasks.
While DeBERTaV3 offers more stable training dynamics.
ModernBERT exhibits instabilities in downstream fine-tuning tasks.
While DeBERTaV3 offers more stable training dynamics.
High-quality pretraining data accelerates convergence but offers minimal gains in final performance.
We suggest that current benchmarks may be saturated, limiting their ability to distinguish model improvements.
High-quality pretraining data accelerates convergence but offers minimal gains in final performance.
We suggest that current benchmarks may be saturated, limiting their ability to distinguish model improvements.
When trained on identical data, DeBERTaV3 outperforms ModernBERT in benchmark tasks.
ModernBERT's strength is faster training and inference, but it doesn't surpass DeBERTaV3 in accuracy on NLU tasks.
When trained on identical data, DeBERTaV3 outperforms ModernBERT in benchmark tasks.
ModernBERT's strength is faster training and inference, but it doesn't surpass DeBERTaV3 in accuracy on NLU tasks.
The new models vastly outperform their predecessors and even match domain-specific finetunes🧑⚕️.
[5/8]
The new models vastly outperform their predecessors and even match domain-specific finetunes🧑⚕️.
[5/8]
- 32,768 tokens
- addition of newline and tab characters
- support emojis with zero-width-joiner
- numbers are split into two digits tokens
- support French elisions
[3/8]
- 32,768 tokens
- addition of newline and tab characters
- support emojis with zero-width-joiner
- numbers are split into two digits tokens
- support French elisions
[3/8]