Benjamin Warner
banner
benjaminwarner.dev
Benjamin Warner
@benjaminwarner.dev
Research at sophont.med, previously answer.ai

Vaccines save lives.
Some personal news: I've joined sophont.med to help build the next generation of open medical foundation models.

We've relaunched medarc.ai, our open science research community. Join us if you want to help advance open medical AI.

And we are hiring.
October 27, 2025 at 7:08 PM
Reposted by Benjamin Warner
counterpoint: GPT-5 does this, it says it doesn’t know rather than hallucinate, the world hasn’t fallen apart
September 13, 2025 at 11:50 AM
Reposted by Benjamin Warner
ModernBERT goes MULTILINGUAL!

One of the most requested models I've seen, @jhuclsp.bsky.social has trained state-of-the-art massively multilingual encoders using the ModernBERT architecture: mmBERT.

Stronger than an existing models at their sizes, while also much faster!

Details in 🧵
September 9, 2025 at 2:54 PM
ChatGPT has been the best technical search engine since o4-mini.

Thinking Mini still makes for a good faster search if you don’t need the extra reasoning ability.
September 6, 2025 at 10:19 PM
Good LLMs do know and/or can reason about these things. Small, cheap, and often free LLMs are the models which cannot.

Remember the glue on pizza Reddit post that the subpar Google AI cited uncritically? Bing’s then integration of GPT 3.5 recognized the Reddit post as sarcasm.
LLMs don't reason, so it doesnt know that the reddit post is likely an uncritical repost of the other articles. It doesn't follow the discussion to see it is (probably) debunked in the comments. All things a marginally educated human would do. AI summary raises that bar, because of ai hype.
August 24, 2025 at 9:53 PM
Reposted by Benjamin Warner
Writing Speed-of-Light Flash Attention for 5090 in CUDA C++ by Thien Tran

He walkthrough how he learned to implement Flash Attention for 5090 in CUDA C++. The main objective is to learn writing attention in CUDA C++,
August 24, 2025 at 12:45 AM
Reposted by Benjamin Warner
Microsoft made a useful LLM copilot tool that could summarize text in spreadsheets. They provided clear instructions about how to use it and not to use it. In response, journalists are now mocking them for doing exactly the right thing and showing how to use and not use the tools.
August 21, 2025 at 1:50 AM
Reports of AI eating entry level jobs are greatly exaggerated.

My guess is current and near-future LLMs are more likely to increase the demand for programmers, not decrease demand (Jevons Paradox).
But, plot twist:

The much-discussed contraction in entry-level tech hiring appears to have *reversed* in recent months.

In fact, relative to the pre-generative AI era, recent grads have secured coding jobs at the same rate as they’ve found any job, if not slightly higher.
July 18, 2025 at 5:06 PM
One of the questions we debated while training ModernBERT was whether a modern trained encoder would unlock zero-shot reasoning using only it's generative head?

Spoilers: the answer is yes.
February 10, 2025 at 6:13 PM
Reposted by Benjamin Warner
o3-mini is really good at writing internal documentation - feed it a codebase, get back a detailed explanation of how specific aspects of it work simonwillison.net/2025/Feb/5/o...
o3-mini is really good at writing internal documentation
I wanted to refresh my knowledge of how the Datasette permissions system works today. I already have [extensive hand-written documentation](https://docs.datasette.io/en/latest/authentication.html) for...
simonwillison.net
February 5, 2025 at 6:09 AM
Reposted by Benjamin Warner
If you want to quickly catch up on all the open modeling things (DeepSeek, ModernBERT, etc.), this was a great overview, by @natolambert.bsky.social.

I somehow got into an argument last week with someone who was insisting that all models are industrial blackboxes... and I wish I'd had this on hand.
The latest open artifacts (#6): Reasoning models, China's lead in open-source, and a growing multimodal space
Artifacts log 6 The open LM ecosystem yet again accelerates.
www.interconnects.ai
January 27, 2025 at 3:05 PM
In addition to being the best retrieval model under 300M params on METB (without extra work), and top 10 for under 1B, here's a fun tidbit from Alibaba's GTE ModernBERT model card:

gte-modernbert-base beats gte-qwen1.5-7b on LoCo long context retrieval with 7B less parameters.
January 23, 2025 at 7:22 PM
Reposted by Benjamin Warner
The newest extremely strong embedding model based on ModernBERT-base is out: `cde-small-v2`. Both faster and stronger than its predecessor, this one tops the MTEB leaderboard for its tiny size!

Details in 🧵
January 14, 2025 at 1:21 PM
Reposted by Benjamin Warner
ModernBERT-embed-base is awesome because it allows to use ModernBERT-base for various tasks out-of-the-box
But the large variant of ModernBERT is also awesome...
So today, @lightonai.bsky.social is releasing ModernBERT-embed-large, the larger and more capable iteration of ModernBERT-embed!
January 14, 2025 at 3:32 PM
ModernBERT is officially released on Transformers v4.48.0. You no longer need to install from git to use.

If you are plugging ModernBERT into an existing encoder finetuning pipeline, try increasing the learning rate. We've found that ModernBERT tends to prefer a higher LR than older models.
January 10, 2025 at 6:28 PM
The good: 32GB
The bad: $2,000
The Ugly*: PCIe 5 without NVLink
January 7, 2025 at 7:12 AM
Reposted by Benjamin Warner
Via @simonwillison.net's excellent blog, I found this great quote about AI models, from @benjaminwarner.dev et al. www.answer.ai/posts/2024-1...

It seems to me that AI will be most relevant in people's lives because the Honda Civic is ubiquitous, not so much because everyone is driving a Ferrari.
January 1, 2025 at 4:04 PM
Reposted by Benjamin Warner
That didn't take long! Nomic AI has finetuned the new ModernBERT-base encoder model into a strong embedding model for search, classification, clustering and more!

Details in 🧵
December 31, 2024 at 3:43 PM
This week we released ModernBERT, the first encoder to reach SOTA on most common benchmarks across language understanding, retrieval, and code, while running twice as fast as DeBERTaV3 on short context and three times faster than NomicBERT & GTE on long context.
December 22, 2024 at 6:12 AM
Reposted by Benjamin Warner
Great blog post (by a 15-author team!) on their release of ModernBERT, the continuing relevance of encoder-only models, and how they relate to, say, GPT-4/llama. Accessible enough that I might use this as an undergrad reading.
Finally, a Replacement for BERT: Introducing ModernBERT
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
December 19, 2024 at 7:11 PM
I feel the need for speed.
December 13, 2024 at 9:56 PM