I'm excited to be on the faculty job market this fall. I just updated my website with my CV.
stephencasper.com
techcrunch.com/2025/11/06/...
techcrunch.com/2025/11/06/...
www.youtube.com/watch?v=VWk3...
www.youtube.com/watch?v=VWk3...
Here's what I learned from our investigation of over 50 platforms, sites, apps, Discords, etc., while writing this paper.
papers.ssrn.com/sol3/papers...
Here's what I learned from our investigation of over 50 platforms, sites, apps, Discords, etc., while writing this paper.
papers.ssrn.com/sol3/papers...
In most (non-adversarial) cases, I expect the opposite will often apply...
In most (non-adversarial) cases, I expect the opposite will often apply...
papers.ssrn.com/sol3/papers....
papers.ssrn.com/sol3/papers....
www.aisi.gov.uk/careers
www.aisi.gov.uk/careers
This new paper studies how a small number of models power the non-consensual AI video deepfake ecosystem and why their developers could have predicted and mitigated this.
This new paper studies how a small number of models power the non-consensual AI video deepfake ecosystem and why their developers could have predicted and mitigated this.
Shamelessly copied from a slack message.
Shamelessly copied from a slack message.
Here's a roundup of some key papers on data filtering & safety.
Tl;DR -- Filtering harmful training data seems to effectively make models resist attacks (incl. adv. fine-tuning), but only when the filtered content is 'hard to learn' from the non-filtered content
🧵
Here's a roundup of some key papers on data filtering & safety.
Tl;DR -- Filtering harmful training data seems to effectively make models resist attacks (incl. adv. fine-tuning), but only when the filtered content is 'hard to learn' from the non-filtered content
🧵
(1/6)
(1/6)
It appears that state AI bills -- many of which big tech has fought tooth and nail to prevent -- are categorically regulatory capture.
It appears that state AI bills -- many of which big tech has fought tooth and nail to prevent -- are categorically regulatory capture.
But in case it makes your life easier, feel free to copy or adapt my rebuttal template linked here.
docs.google.com/document/d/1...
But in case it makes your life easier, feel free to copy or adapt my rebuttal template linked here.
docs.google.com/document/d/1...
From a technical perspective, safeguarding open-weight model safety is AI safety in hard mode. But there's still a lot of progress to be made. Our new paper covers 16 open problems.
🧵🧵🧵
From a technical perspective, safeguarding open-weight model safety is AI safety in hard mode. But there's still a lot of progress to be made. Our new paper covers 16 open problems.
🧵🧵🧵
From a technical perspective, safeguarding open-weight model safety is AI safety in hard mode. But there's still a lot of progress to be made. Our new paper covers 16 open problems.
🧵🧵🧵
From a technical perspective, safeguarding open-weight model safety is AI safety in hard mode. But there's still a lot of progress to be made. Our new paper covers 16 open problems.
🧵🧵🧵
I'm increasingly persuaded that the only quantitative measures that matter anymore are usage stats & profit.
I'm increasingly persuaded that the only quantitative measures that matter anymore are usage stats & profit.
Now that Moonshot claims Kimi K2 Thinking is SOTA, it seems, uh, less than ideal that it came with zero reporting related to safety/risk.
Now that Moonshot claims Kimi K2 Thinking is SOTA, it seems, uh, less than ideal that it came with zero reporting related to safety/risk.
Is it because more Chinese companies are "fast followers" who find their niche by making open models?
Is it cultural? Do Eastern/Chinese cultures value open tech more?
Is it because more Chinese companies are "fast followers" who find their niche by making open models?
Is it cultural? Do Eastern/Chinese cultures value open tech more?
We’ll host speakers from political theory, economics, mechanism design, history, and hierarchical agency.
post-agi.org
We’ll host speakers from political theory, economics, mechanism design, history, and hierarchical agency.
post-agi.org
unicode.org/L2/L2025/252...
t.co/yJfp8ezU64
unicode.org/L2/L2025/252...
t.co/yJfp8ezU64
Of course -- that's obvious. Nobody would ever dispute that.
So then why are we saying that?
Maybe it's a little too obvious...
Of course -- that's obvious. Nobody would ever dispute that.
So then why are we saying that?
Maybe it's a little too obvious...
In a new paper with @realbrianjudge.bsky.social at #EAAMO25, we pull back the curtain on AI safety's toolkit. (1/n)
arxiv.org/pdf/2509.22872
In a new paper with @realbrianjudge.bsky.social at #EAAMO25, we pull back the curtain on AI safety's toolkit. (1/n)
arxiv.org/pdf/2509.22872
Of course -- that's obvious. Nobody would ever dispute that.
So then why are we saying that?
Maybe it's a little too obvious...
Of course -- that's obvious. Nobody would ever dispute that.
So then why are we saying that?
Maybe it's a little too obvious...
www.nature.com/articles/d41...
www.nature.com/articles/d41...