recently, @primeintellect.bsky.social have announced finishing their 10B distributed learning, trained across the world.
what is it exactly?
🧵
recently, @primeintellect.bsky.social have announced finishing their 10B distributed learning, trained across the world.
what is it exactly?
🧵
Deadline: November 25
www.ed.ac.uk/studying/pos...
If you are passionate about:
- adaptive tokenization and memory in foundation models
- modular deep learning
- computational typology
please message me or meet me at #NeurIPS2024!
Deadline: November 25
www.ed.ac.uk/studying/pos...
If you are passionate about:
- adaptive tokenization and memory in foundation models
- modular deep learning
- computational typology
please message me or meet me at #NeurIPS2024!
Piotr Nawrot!
A repo & notebook on sparse attention for efficient LLM inference: github.com/PiotrNawrot/...
This will also feature in my #NeurIPS 2024 tutorial "Dynamic Sparsity in ML" with André Martins: dynamic-sparsity.github.io Stay tuned!
Piotr Nawrot!
A repo & notebook on sparse attention for efficient LLM inference: github.com/PiotrNawrot/...
This will also feature in my #NeurIPS 2024 tutorial "Dynamic Sparsity in ML" with André Martins: dynamic-sparsity.github.io Stay tuned!
👉 github.com/sordonia/pg_mb…
Part of "Dynamic Sparsity in ML" tut#neurips202424, feedback welcome and join for discussions! 😊
👉 github.com/sordonia/pg_mb…
Part of "Dynamic Sparsity in ML" tut#neurips202424, feedback welcome and join for discussions! 😊