https://www.soufianehayou.com/
Paper: arxiv.org/abs/2506.20629
Code: github.com/soufiane001/...
Paper: arxiv.org/abs/2506.20629
Code: github.com/soufiane001/...
✅ Works across different post-training scenarios: supervised fine-tuning, reinforcement learning
✅ Minimal computational overhead
In the worst case, it ties with the best manual approach. Usually, it's better.
✅ Works across different post-training scenarios: supervised fine-tuning, reinforcement learning
✅ Minimal computational overhead
In the worst case, it ties with the best manual approach. Usually, it's better.
Instead of guessing, it automatically identifies the optimal modules for LoRA placement based on a notion of module-data alignment that we call NFN (Normalised Feature Norms).
Instead of guessing, it automatically identifies the optimal modules for LoRA placement based on a notion of module-data alignment that we call NFN (Normalised Feature Norms).
❌ Other papers: "Actually, put them in MLP"
❌ Everyone: just guessing and trying common target modules
❌ Other papers: "Actually, put them in MLP"
❌ Everyone: just guessing and trying common target modules