Pretrained 1B/8B param models, with controlled insertion of texts designed to emulate key memorization risks: copyright (e.g., book passages), privacy (e.g., synthetic biographies), and test set contamination
Pretrained 1B/8B param models, with controlled insertion of texts designed to emulate key memorization risks: copyright (e.g., book passages), privacy (e.g., synthetic biographies), and test set contamination
Looking forward to what the community builds and studies using these models!!
Pretrained 1B/8B param models, with controlled insertion of texts designed to emulate key memorization risks: copyright (e.g., book passages), privacy (e.g., synthetic biographies), and test set contamination
Looking forward to what the community builds and studies using these models!!
Docs - github.com/sgl-project/...
Blog - aflah02.substack.com/p/multi-node...
Docs - github.com/sgl-project/...
Blog - aflah02.substack.com/p/multi-node...
"... Then they came for me—and there was no one left to speak for me."
"... Then they came for me—and there was no one left to speak for me."
Feeling like I did some real work after a while when all I did was turn flags on and off and look at wandb logs 🤓
Feeling like I did some real work after a while when all I did was turn flags on and off and look at wandb logs 🤓
P.s. I could never get the code to run lol as setting up an env for TF1 was borderline impossible without proper pinning
P.s. I could never get the code to run lol as setting up an env for TF1 was borderline impossible without proper pinning