At Sakana AI, we remain committed to pioneering novel AI systems by applying nature-inspired principles such as evolution and collective intelligence.
At Sakana AI, we remain committed to pioneering novel AI systems by applying nature-inspired principles such as evolution and collective intelligence.
Algorithm (TreeQuest): github.com/SakanaAI/tre...
Algorithm (TreeQuest): github.com/SakanaAI/tre...
Interactive Paper (with web-demo): pub.sakana.ai/ctm/
Full Paper: arxiv.org/abs/2505.05522
GitHub Project: github.com/SakanaAI/con...
Interactive Paper (with web-demo): pub.sakana.ai/ctm/
Full Paper: arxiv.org/abs/2505.05522
GitHub Project: github.com/SakanaAI/con...
Paper: openreview.net/forum?id=dh4...
Transformer-Squared adapts its weights on the fly for each query, achieving strong performance across tasks and enabling parameter-efficient life-long learning.
Paper: openreview.net/forum?id=dh4...
Transformer-Squared adapts its weights on the fly for each query, achieving strong performance across tasks and enabling parameter-efficient life-long learning.
Paper: openreview.net/forum?id=Kvd...
CycleQD is an ecological niche-inspired model-merging approach that achieves great performance on computer science tasks while retaining language capabilities.
Paper: openreview.net/forum?id=Kvd...
CycleQD is an ecological niche-inspired model-merging approach that achieves great performance on computer science tasks while retaining language capabilities.
Paper: openreview.net/forum?id=s1k...
Neural Attention Memory Models (NAMMs) is an evolved memory system trained to improve performance and efficiency on language transformers, and zero-shot transferring to vision and RL foundation models.
Paper: openreview.net/forum?id=s1k...
Neural Attention Memory Models (NAMMs) is an evolved memory system trained to improve performance and efficiency on language transformers, and zero-shot transferring to vision and RL foundation models.
Paper: openreview.net/forum?id=cqs...
TAID is a novel knowledge distillation method that uses a time-dependent intermediate distribution addressing common challenges in distilling LLMs.
Paper: openreview.net/forum?id=cqs...
TAID is a novel knowledge distillation method that uses a time-dependent intermediate distribution addressing common challenges in distilling LLMs.