Stella Biderman
stellaathena.bsky.social
Stella Biderman
@stellaathena.bsky.social
I make sure that OpenAI et al. aren't the only people who are able to study large scale AI systems.
Small caveat: I misunderstood arXiv's ToS when I wrote this paper. While a large portion of arXiv has an open license, the majority (last time I checked) does not. That shouldn't have a check under "author."

PG-19 lacks one because of how radically technology has changed.

arxiv.org/abs/2101.00027
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
Recent work has demonstrated that increased training dataset diversity improves general cross-domain knowledge and downstream generalization capability for large-scale language models. With this in mi...
arxiv.org
October 20, 2025 at 6:55 AM
In the original Pile paper we talked about various conceptions of consent (though I don't stand by everything I wrote about this topic 5 years ago). None of this data has EIC, though I think that the ones marked "author" in the table are ones where authorial objection would be unreasonable.
October 20, 2025 at 6:55 AM
Adding to what @mmitchell.bsky.social said, EIC cannot be use-agnostic by definition. It must be explicit to the use in question. If you put a notice that says "everyone can use this for every purpose" that's *not* EIC.
October 20, 2025 at 6:55 AM
That was the best I was able to find, I swear I've read others though.
October 4, 2025 at 6:25 AM
It doesn't directly address your original question though... maybe I should write a blog post about it.
October 4, 2025 at 6:24 AM
We have a repo documenting resources to learn about this and all other HPC aspects of LLM training: github.com/EleutherAI/c...
GitHub - EleutherAI/cookbook: Deep learning for dummies. All the practical details and useful utilities that go into working with real models.
Deep learning for dummies. All the practical details and useful utilities that go into working with real models. - EleutherAI/cookbook
github.com
October 4, 2025 at 6:22 AM
You can't train a LLM like that without multiple revolutionary breakthroughs. It's a common talking point from people who are grifters or clueless, but the technology simply doesn't work like that.
October 4, 2025 at 6:09 AM
Obviously she assaulted him /s
September 26, 2025 at 1:20 PM
This is always dire, but especially so when the US is being run by an authoritarian who delights in using state power to go after people he dislikes. The second OpenAI starts asking for your ID, the government will be asking OpenAI for your chats.
September 18, 2025 at 12:40 AM
OpenAI says adult chats deserve confidentiality, then single out teens for surveillance and says that they'll call the cops on people with mental health crises.

This will kill people and not help them get the care they need. It happens all the time

www.vera.org/news/we-need...
We Need to Think Beyond Police in Mental Health Crises
In March of 2020, Joe Prude called 911 for assistance. His brother, Daniel Prude, was behaving erratically and had just bolted out the back door of Joe…
www.vera.org
September 18, 2025 at 12:40 AM
There are some papers demonstrating that this improves performance, especially in translation contexts IIRC.
September 5, 2025 at 9:01 PM
It's also a pretty notable comment about my friend group that when I wrote this comment I was considering "partner" to be the opposite-gender counterpart of "girlfriend"
September 2, 2025 at 2:49 AM
Good luck! Maybe you'll succeed where people have failed for decades.
August 26, 2025 at 5:21 PM
Same

(Since I don't know most of the people in this thread, the joke is that I run one of the servers Naomi mentioned. Except it's not a joke.)
August 26, 2025 at 5:08 PM
I have so few straight friends that "partner" to me is mostly coded as "bi but in a relationship with someone of the opposite gender"
August 26, 2025 at 5:05 PM
You're right that this is an active area of research but I'm unaware of any meaningful successes coming out of it.
August 26, 2025 at 5:01 PM
Can you name an example of an idea that is well-grounded in biology that has proven successful for neural networks? I don't mean "oh DL was inspired by how non-neuroscientists thinks the brain works," I mean an actual case of making a model work better by making it more brain-like
August 26, 2025 at 5:00 PM