Our methods aren’t just for small models! We show that they scale effectively to larger LMs, providing robust memorization mitigation without compromising performance across different sizes of models. Exciting progress for real-world applications!
Our methods aren’t just for small models! We show that they scale effectively to larger LMs, providing robust memorization mitigation without compromising performance across different sizes of models. Exciting progress for real-world applications!
Our proposed unlearning method, BalancedSubnet, outperforms others by effectively removing memorized info while maintaining high accuracy.
Our proposed unlearning method, BalancedSubnet, outperforms others by effectively removing memorized info while maintaining high accuracy.
Unlearning-based methods are faster and more effective than regularization or fine-tuning in mitigating memorization.
Unlearning-based methods are faster and more effective than regularization or fine-tuning in mitigating memorization.
We created TinyMem, a suite of small, efficient models designed to help test and benchmark memorization mitigation techniques. TinyMem allows for quick experiments with lower computational costs.
We created TinyMem, a suite of small, efficient models designed to help test and benchmark memorization mitigation techniques. TinyMem allows for quick experiments with lower computational costs.
We test 17 methods—regularization, fine-tuning, and unlearning—5 of which we propose. These methods aim to remove memorized info from LMs while preserving performance.
We test 17 methods—regularization, fine-tuning, and unlearning—5 of which we propose. These methods aim to remove memorized info from LMs while preserving performance.
Language models (LMs) often "memorize" data, leading to privacy risks. This paper explores ways to reduce that!
Paper: arxiv.org/pdf/2410.02159
Code: github.com/msakarvadia/...
Blog: mansisak.com/memorization/
Language models (LMs) often "memorize" data, leading to privacy risks. This paper explores ways to reduce that!
Paper: arxiv.org/pdf/2410.02159
Code: github.com/msakarvadia/...
Blog: mansisak.com/memorization/