With amazing collaborators:
Archie Sravankumar
Lijuan Liu
Yuning Mao
Rui Hou
Sinong Wang
@jfoerst.bsky.social
Madian Khabsa
@lukezettlemoyer.bsky.social
With amazing collaborators:
Archie Sravankumar
Lijuan Liu
Yuning Mao
Rui Hou
Sinong Wang
@jfoerst.bsky.social
Madian Khabsa
@lukezettlemoyer.bsky.social
With low-threshold tuning, we take Llama3-70B from:
➡️ 51% → 87% correctness
➡️ Retaining 53% of the original completeness
With low-threshold tuning, we take Llama3-70B from:
➡️ 51% → 87% correctness
➡️ Retaining 53% of the original completeness
We introduce a threshold that tunes how eagerly the model should respond:
Low threshold = more reliable answers 🔒 (Left box)
High threshold = more detailed answers 📝(Right box)
We introduce a threshold that tunes how eagerly the model should respond:
Low threshold = more reliable answers 🔒 (Left box)
High threshold = more detailed answers 📝(Right box)
1️⃣ Break pretrained LLM responses into factual fragments
2️⃣ Use ground truth to flag incorrect fragments
3️⃣ Modify finetuning responses by removing or replacing errors with “Unsure from here” 🚧
1️⃣ Break pretrained LLM responses into factual fragments
2️⃣ Use ground truth to flag incorrect fragments
3️⃣ Modify finetuning responses by removing or replacing errors with “Unsure from here” 🚧
This leads to partially incorrect outputs in critical domains like Coding, Math, Medicine, and QA.
Why? Standard finetuning ignores what the pretrained model actually knows and pushes it to always complete every prompt.
This leads to partially incorrect outputs in critical domains like Coding, Math, Medicine, and QA.
Why? Standard finetuning ignores what the pretrained model actually knows and pushes it to always complete every prompt.