Lucie Charlotte Magister
charlottemagister.bsky.social
Lucie Charlotte Magister
@charlottemagister.bsky.social
PhD student @ University of Cambridge, focusing on Explainability and Interpretability for GNNs
3/🧵PEFT: We use these QA pairs to finetune a LoRA adapter conversation by conversation. We find that weighting the loss on the QA tokens focuses the model on relevant content, rather than structure. Iterating on each conversation for 10 epochs gives the best results.
November 21, 2024 at 6:03 PM
2/🧵Data Augmentation: We up-sample conversations as positive and negative QA pairs. Positive pairs contain questions about the conversation content, while negative pairs contain questions about related topics not discussed. The negative pairs allow to draw a knowledge boundary.
November 21, 2024 at 6:03 PM
1/🧵PLUM is a 2 stage pipeline performing data augmentation for up-sampling conversations as QA pairs, that we use to finetune a LoRA adapter with a weighted CE loss. We perform competitively with baselines such as RAG, achieving 81.5% accuracy across 100 conversations.
November 21, 2024 at 6:03 PM