i don't know if timestamps would migrate, but I've seen folks who have posts with older timestamps than the start of the bluesky
i don't know if timestamps would migrate, but I've seen folks who have posts with older timestamps than the start of the bluesky
I think focus on raw number of parameters is a less useful frame than thinking about inference speed, cost and location of inference (on-device vs cloud).
o3 achieving human-level on the semi-private eval feels like a significant breakthrough.
Calibrating, I'd say o3 is a GPT-1 or GPT-2 moment. The direction for improvement is getting clear, with more of the research fog getting lifted.
GPT-4 was at Level 1, conversational AI: a model competent at 0.1-1s tasks, like holding a conversation.
O1 / R1 reached Level 2, reasoners: a model solving 1-10min tasks such as basic coding tasks and math.
I don't think that's quite right.
Here's two papers that helped me have a more nuanced view of this question.
Antimicrobial peptides are proteins that kill bacteria. Most do so by making circular holes in their membranes.
In this fun to write paper, we showed fractal pores in bacterial membranes.
Antimicrobial resistance is a big threat to public health, and here we show ways to combine simulation and deep learning with new antimicrobial peptides validated in vivo.
One of the interesting things about that project was how gradient boosted trees + chemical fingerprints performed very well in practice.
1/ When building AIs for science, it's important for the algorithms to discover beyond things we already know. This is why effective, open-ended exploration is important. Here we propose MetaGFN, an algorithm to effectively find distant modes in probability distributions.
Automated labs coupled with active learning are a super exciting area with lots of opportunities for progress.
I promised @cpaxton.bsky.social a short thread on this, so here it goes!
🧪
We have three models released based on SMILES, SELFIES and molecular graphs.
More to come shortly - we aim to have a unified collection of state-of art models across all modalities.
Here's some interesting results and some thoughts on future directions.
a new 7B llama-style LLM for embedding of genomes & detection of pathogens in wastewater
i’ve had a hunch that LLMs could lead to some big bio breakthroughs, since it feels like genes & proteins are a lot like a language
Creating the GUI at PARC seemed like a "waste of FLOPs" but revolutionized computing.
Creating the GUI at PARC seemed like a "waste of FLOPs" but revolutionized computing.
Alan Kay talked about the Wayne Gretzky game, a hockey player famous for his quote about skating where the puck will be.
Alan Kay talked about the Wayne Gretzky game, a hockey player famous for his quote about skating where the puck will be.
At each generation, larger and larger number of parameters can be ran locally.
At each generation, larger and larger number of parameters can be ran locally.
Today, an Apple M4 has 28B transistors, meaning I experienced a scale-up of 1,000,000x in my lifetime.
I expect a similar scale-up for language models.
Today, an Apple M4 has 28B transistors, meaning I experienced a scale-up of 1,000,000x in my lifetime.
I expect a similar scale-up for language models.