Project Website: shuhaibm.github.io/refed/
Paper: arxiv.org/abs/2502.04511
Thanks to my incredible co-authors Xiusi Chen, Heng Ji, @dilekh.bsky.social!
Project Website: shuhaibm.github.io/refed/
Paper: arxiv.org/abs/2502.04511
Thanks to my incredible co-authors Xiusi Chen, Heng Ji, @dilekh.bsky.social!
📊 We demonstrate consistent improvement across both base and instruct variants of different model architectures
📊 Analysis of filtering strategies reveals dataset variants that maintain strong performance while reducing costs
📊 We demonstrate consistent improvement across both base and instruct variants of different model architectures
📊 Analysis of filtering strategies reveals dataset variants that maintain strong performance while reducing costs
📊 Llama-3.1-8B-Instruct + REFED achieves SOTA among SFT-based 8B parameter models on AlpacaEval 2.0
📊 Comparisons and ablation studies validate every component of our framework and show advantages over traditional feedback
📊 Llama-3.1-8B-Instruct + REFED achieves SOTA among SFT-based 8B parameter models on AlpacaEval 2.0
📊 Comparisons and ablation studies validate every component of our framework and show advantages over traditional feedback