previously @eleutherai.bsky.social
🌐 lintang.sutawika.com
We are thrilled to announce the Common Pile v0.1, an 8TB dataset of openly licensed and public domain text. We train 7B models for 1T and 2T tokens and match the performance similar models like LLaMA 1 & 2
We are thrilled to announce the Common Pile v0.1, an 8TB dataset of openly licensed and public domain text. We train 7B models for 1T and 2T tokens and match the performance similar models like LLaMA 1 & 2
But it’s a nice anecdote.
go.bsky.app/NhTwCVb
go.bsky.app/NhTwCVb