A huge shoutout to my incredible co-authors from multiple institutions for their contributions to this work:
Aswathy Ajith, Arham Khan, @nathaniel-hudson.bsky.social , @calebgeniesse.bsky.social, Yaoqing Yang, @kylechard.bsky.social , @ianfoster42.bsky.social, Michael Mahoney
A huge shoutout to my incredible co-authors from multiple institutions for their contributions to this work:
Aswathy Ajith, Arham Khan, @nathaniel-hudson.bsky.social , @calebgeniesse.bsky.social, Yaoqing Yang, @kylechard.bsky.social , @ianfoster42.bsky.social, Michael Mahoney
Our methods aren’t just for small models! We show that they scale effectively to larger LMs, providing robust memorization mitigation without compromising performance across different sizes of models. Exciting progress for real-world applications!
Our methods aren’t just for small models! We show that they scale effectively to larger LMs, providing robust memorization mitigation without compromising performance across different sizes of models. Exciting progress for real-world applications!
Our proposed unlearning method, BalancedSubnet, outperforms others by effectively removing memorized info while maintaining high accuracy.
Our proposed unlearning method, BalancedSubnet, outperforms others by effectively removing memorized info while maintaining high accuracy.
Unlearning-based methods are faster and more effective than regularization or fine-tuning in mitigating memorization.
Unlearning-based methods are faster and more effective than regularization or fine-tuning in mitigating memorization.
We created TinyMem, a suite of small, efficient models designed to help test and benchmark memorization mitigation techniques. TinyMem allows for quick experiments with lower computational costs.
We created TinyMem, a suite of small, efficient models designed to help test and benchmark memorization mitigation techniques. TinyMem allows for quick experiments with lower computational costs.
We test 17 methods—regularization, fine-tuning, and unlearning—5 of which we propose. These methods aim to remove memorized info from LMs while preserving performance.
We test 17 methods—regularization, fine-tuning, and unlearning—5 of which we propose. These methods aim to remove memorized info from LMs while preserving performance.