📄Paper: arxiv.org/abs/2510.01171
🌐Blog: verbalized-sampling.com
Team: Jiayi Zhang @simon-ycl.bsky.social @derekch.bsky.social Anthony Sicilia, Michael Tomz, @chrmanning.bsky.social @shi-weiyan.bsky.social
@stanfordnlp.bsky.social XNortheasternXWVU
📄Paper: arxiv.org/abs/2510.01171
🌐Blog: verbalized-sampling.com
Team: Jiayi Zhang @simon-ycl.bsky.social @derekch.bsky.social Anthony Sicilia, Michael Tomz, @chrmanning.bsky.social @shi-weiyan.bsky.social
@stanfordnlp.bsky.social XNortheasternXWVU
💻 Quickstart and Colab: github.com/CHATS-lab/ve...
🎮 pip install verbalized-sampling
Package includes LangChain integration + tunable diversity knobs!
#VerbalizedSampling
💻 Quickstart and Colab: github.com/CHATS-lab/ve...
🎮 pip install verbalized-sampling
Package includes LangChain integration + tunable diversity knobs!
#VerbalizedSampling
We prove that human raters give higher scores to boring, predictable answers. So models learn to play it safe.
But this diversity wasn't deleted - just suppressed. One sentence unlocks it all.
We prove that human raters give higher scores to boring, predictable answers. So models learn to play it safe.
But this diversity wasn't deleted - just suppressed. One sentence unlocks it all.
✍️ Creative writing → 2.1× diversity
💬 Dialogue → Matches human behavior
📊 Synthetic training data → +18% better
Emergent trend: Big models gain more than small ones
Tested w/ @stanfordnlp.bsky.social on thousands of outputs
✍️ Creative writing → 2.1× diversity
💬 Dialogue → Matches human behavior
📊 Synthetic training data → +18% better
Emergent trend: Big models gain more than small ones
Tested w/ @stanfordnlp.bsky.social on thousands of outputs
Just paste this line before any creative task. That's it!
Instead of the same "safe" answer five times, you get five completely different ones. Here's the difference:
Just paste this line before any creative task. That's it!
Instead of the same "safe" answer five times, you get five completely different ones. Here's the difference: