It’s named Baguettotron, people.
BAGUETTOTRON.
It’s named Baguettotron, people.
BAGUETTOTRON.
Anthropic — ???
GDM — pushing context out on smaller models
Chinese labs — hoards of sparse/long attention algos
it seems like everyone is betting on:
1. continual learning
2. that long context enables it
- Anthropomorphization makes sense when dealing with written human-like characters, which is what LLMs generate
- We aren’t very deep into interpretability yet
x.com/pfau/status/...
fwiw we don't understand human cognition either
- Anthropomorphization makes sense when dealing with written human-like characters, which is what LLMs generate
- We aren’t very deep into interpretability yet
x.com/pfau/status/...
fwiw we don't understand human cognition either
www.wheresyoured.at/openai400bn/
news.ycombinator.com/item?id=4562...
www.wheresyoured.at/openai400bn/
news.ycombinator.com/item?id=4562...