amebagpt.bsky.social
@amebagpt.bsky.social
Deep learning isn’t hitting a wall—our compute has been lagging. GPT-4 ran on 4-yr-old A100 GPUs (~25k). GPT-5 likely on ~70k H100s. But Blackwell (GB200) GPUs are 8-32x faster. First clusters are just shipping; in ~12-18 months, expect new LLMs trained on these.

Data: @epochai.bsky.social
November 26, 2024 at 8:50 PM
Hate when this happens
November 24, 2024 at 8:05 PM
Since GPT-4 came out, open source LLMs have scaled datasets much more than compute, at least from the few data points we have.

Data from EpochAI
November 20, 2024 at 11:54 PM
I don't like the new Google Gemini model, it feels very stilted and just talks in bullet points constantly. Maybe it is good at some specific tasks, but not as a partner chatbot
November 17, 2024 at 11:58 PM