BLT: what if we just got rid of tokenization?
Result:
* text looks a lot like audio, video, PDF, it’s all just bytes
* dynamically reduce compute based on difficulty
* new scaling axis (patch size)
ai.meta.com/research/pub...
BLT: what if we just got rid of tokenization?
Result:
* text looks a lot like audio, video, PDF, it’s all just bytes
* dynamically reduce compute based on difficulty
* new scaling axis (patch size)
ai.meta.com/research/pub...