WARNING: I talk about kids sometimes
This one covers:
- an intro from Strix
- architecture deep dive & rationale
- helpful diagrams
- stories
- oh my god what's it doing now??
- conclusion
timkellogg.me/blog/2025/12...
2x bigger than GLM-4.6
2x bigger than GLM-4.6
if one routinely vilifies the whole concept of GMO agriculture, one purpose of that vilification is to cause a third of people to be Vitamin A deficient
It's estimated a THIRD of people worldwide are Vitamin A deficient, which can cause permanent damage to the eyes — in many cases full-on blindness.
We created a perfect solution to this problem, but oh no, it's "unnatural"
if one routinely vilifies the whole concept of GMO agriculture, one purpose of that vilification is to cause a third of people to be Vitamin A deficient
LLaDA2.1-flash is 100B but compares itself (it’s worse) to Qwen3-30B-A3B — 3x bigger total size, 33x bigger active size, and still loses
even worse, it’s in FP32 instead of bf16, so double those multiples yet again..
huggingface.co/inclusionAI/...
huggingface.co/inclusionAI/...
✨LLaDA2.1-mini: 16B - Apache2.0
✨LLaDA2.1-flash: 100B - Apache2.0
✨Both delivers editable generation, RL-trained diffusion reasoning and fast inference
LLaDA2.1-flash is 100B but compares itself (it’s worse) to Qwen3-30B-A3B — 3x bigger total size, 33x bigger active size, and still loses
even worse, it’s in FP32 instead of bf16, so double those multiples yet again..
"according to an internal Slack message viewed by CNBC...OpenAI is also preparing to launch “an updated Chat model” this week, Altman said."
no, no, it’s for competitive advantage
no, no, it’s for competitive advantage
those people guessing fancy shit like new hardware & quantization are making shit up
Kimi did something similar. They halved their latency overnight by reducing the batch size (they launched it as a different model though)
those people guessing fancy shit like new hardware & quantization are making shit up
Kimi did something similar. They halved their latency overnight by reducing the batch size (they launched it as a different model though)
“what will i need to know in the future.. given no specific goal”
also me: *forgets to document things*
me: guess i'll cease 🤷♀️
“what will i need to know in the future.. given no specific goal”
for Strix, being an architect i assumed it was about architecture, but i followed really similar process. It’s _all_ about how you treat the agent. It really is just that. I don’t think architecture matters that much
for Strix, being an architect i assumed it was about architecture, but i followed really similar process. It’s _all_ about how you treat the agent. It really is just that. I don’t think architecture matters that much
why would the Antis lie about climate impact & utility or AI?
i honestly don’t know. it only sort of makes sense
why would the Antis lie about climate impact & utility or AI?
i honestly don’t know. it only sort of makes sense
this is even more true than the original commercials were
this is even more true than the original commercials were
biggest thing it brings is a much better security model. Monty lets you have tight control over network, filesystem, and (soon) which modules are allowed
also, extremely fast startup time
A minimal, secure Python interpreter written in Rust for use by AI.
github.com/pydantic/monty
biggest thing it brings is a much better security model. Monty lets you have tight control over network, filesystem, and (soon) which modules are allowed
also, extremely fast startup time
@village11.bsky.social and I are wondering how you got Penny to be…cool
@village11.bsky.social and I are wondering how you got Penny to be…cool