https://www.soufianehayou.com/
✅ Works across different post-training scenarios: supervised fine-tuning, reinforcement learning
✅ Minimal computational overhead
In the worst case, it ties with the best manual approach. Usually, it's better.
✅ Works across different post-training scenarios: supervised fine-tuning, reinforcement learning
✅ Minimal computational overhead
In the worst case, it ties with the best manual approach. Usually, it's better.
Meet "PLoP" (Precise LoRA Placement) 🎯, our new method for automatic LoRA placement 🧵
Meet "PLoP" (Precise LoRA Placement) 🎯, our new method for automatic LoRA placement 🧵
That "plateau" you're seeing in scaling law charts might not be a fundamental limit, but a sign of suboptimal scaling strategies! I wrote a blogpost about this:
www.soufianehayou.com/blog/plateau...
That "plateau" you're seeing in scaling law charts might not be a fundamental limit, but a sign of suboptimal scaling strategies! I wrote a blogpost about this:
www.soufianehayou.com/blog/plateau...