Ian Magnusson
banner
ianmagnusson.bsky.social
Ian Magnusson
@ianmagnusson.bsky.social
Science of language models @uwnlp.bsky.social and @ai2.bsky.social with @PangWeiKoh and @nlpnoah.bsky.social. https://ianmagnusson.github.io
Come chat with me at #NeurIPS2024 and learn about how to use Paloma to evaluate perplexity over hundreds of domains! ✨We have stickers too✨
December 10, 2024 at 3:54 AM
Further decomposing perplexity, we find that some vocabulary strings get worse as models scale (see examples) ✍️
Again, not always bad, but Paloma reports average loss of each vocabulary string, surfacing strings that behave differently in some domains.
December 20, 2023 at 8:33 PM
We also show that performance improves in almost all domains as models are scaled, but domains improve unequally 📈📉
Differences in improvement, such as these examples, can indicate divergence, stagnation, or saturation—not all bad, but worth investigating!
December 20, 2023 at 8:33 PM
We pretrain six 1B baselines on popular corpora 🤖

With these we find Common-Crawl-only pretraining has inconsistent fit to many domains:
1. C4 and mC4 baselines erratically worse fit than median model
2. C4, mC4, and Falcon baselines sometimes non-monotonic perplexity in Fig 1
December 20, 2023 at 8:32 PM
Along with the datasets we curate, we build eval corpora from held out Dolma data that sample:
💬 top 100 subreddits
🧑‍💻 top 100 programming languages 
Different research may require other domains, but Paloma enables research on 100s of domains from existing metadata.
December 20, 2023 at 8:30 PM
We introduce guidelines and implement controls for LM experiments 📋:
1. Remove contaminated pretraining
2. Fix train order
3. Subsample eval data based on metric variance
4. Fix the vocabulary unless you study changing it
5. Standardize eval format
December 20, 2023 at 8:30 PM
Paloma benchmark results are organized by comparability of:
🧪 controls like benchmark decontamination
💸 measures of cost (parameter and training token count)

Find out more:
📃 arXiv (arxiv.org/pdf/2312.105...)
🤖 data and models (huggingface.co/collections/...)
December 20, 2023 at 8:29 PM
LMs are used to process text from many topics, styles, dialects, etc., but how well do they do?

📈 Evaluating perplexity on just one corpus like C4 doesn't tell the whole story 📉

✨📃✨
We introduce Paloma, a benchmark of 585 domains from NY Times to r/depression on Reddit.
December 20, 2023 at 8:28 PM