@Anthropic. Seattle bike lane enjoyer. Opinions my own.
(I'm still based in Seattle 🏔️🌲🏕️; but in SF regularly)
(I'm still based in Seattle 🏔️🌲🏕️; but in SF regularly)
The training data thus contains high (doc, query) word similarity cases, but not obviously irrelevant docs, or relevant docs not found by vector search.
The training data thus contains high (doc, query) word similarity cases, but not obviously irrelevant docs, or relevant docs not found by vector search.
But beware 👻 !
Despite expressivity, top-K re-rankers generalize poorly as K increases.
arxiv.org/pdf/2411.11767
But beware 👻 !
Despite expressivity, top-K re-rankers generalize poorly as K increases.
arxiv.org/pdf/2411.11767
RLHF = 30% *more* copying than base!
Awesome work from the awesome Ximing Lu (gloriaximinglu.github.io) et al. 🤩
arxiv.org/pdf/2410.04265
RLHF = 30% *more* copying than base!
Awesome work from the awesome Ximing Lu (gloriaximinglu.github.io) et al. 🤩
arxiv.org/pdf/2410.04265