Have you experimented with using LLMs to b) synthetically generate QA pairs based off of source documents, b) generate hard negatives, c) generate paraphrases?
Does it help with performance? Mostly considering this for languages other than English.
Have you experimented with using LLMs to b) synthetically generate QA pairs based off of source documents, b) generate hard negatives, c) generate paraphrases?
Does it help with performance? Mostly considering this for languages other than English.
Would be very helpful and appreciated.
Would be very helpful and appreciated.
1. What's your intuition regarding possibility of doing sbert model distillation on models trained with query prefixes? Can we just use regular parallel texts, or would we need prefix parallel texts?
2. Matryoshka aware model distillation should be possible right?
1. What's your intuition regarding possibility of doing sbert model distillation on models trained with query prefixes? Can we just use regular parallel texts, or would we need prefix parallel texts?
2. Matryoshka aware model distillation should be possible right?