Jacob Haimes
banner
jacobhaimes.bsky.social
Jacob Haimes
@jacobhaimes.bsky.social
founder of Kairos.fm, host of the Into AI Safety and muckrAIkers podcasts, working with Apart Research and the Odyssean Institute.
All views my own. He/him.
Pinned
A recent @apartresearch.bsky.social Researcher Spotlight featured me! Check it out to hear more about my journey Into AI Safety (pun intended):
www.youtube.com/watch?v=lFAm...
Researcher Spotlight: Jacob Haimes
YouTube video by Apart - Safe AI
www.youtube.com
Reposted by Jacob Haimes
Tune into the latest episode of Into AI Safety, "Against 'The Singularity' w/ Dr. David Thorstad".

In this episode, we cover the varying claims on the likelihood of 'The Singularity' hypothesis, and the risks of unfounded forecasting.

Full episode up now: kairos.fm/intoaisafety...
November 25, 2025 at 7:00 PM
Reposted by Jacob Haimes
🚨 New muckrAIkers: "AI Safety For Who?"

@jacobhaimes.bsky.social & @thegermanpole.bsky.social break down how instruction tuning/RLHF create anthropomorphic chatbots that exploit human empathy—leading to mental health harms. kairos.fm/muckraikers/...

Find us wherever you listen! (links in thread)
AI Safety for Who? | Kairos.fm
AI safety is making you less safe: chatbot anthropomorphization, mental health harms, dark patterns
kairos.fm
October 13, 2025 at 8:42 PM
Reposted by Jacob Haimes
🎉 Just dropped a new Into AI Safety episode! Host @jacobhaimes.bsky.social chats with Alistair Lowe-Norris (ex-Microsoft, now at Iridius) about how responsible AI actually happens in practice.

Check us out on Patreon or wherever you get your podcasts!
(links in thread)
kairos.fm/intoaisafety...
Getting Agentic w/ Alistair Lowe-Norris | Kairos.fm
Responsible AI veteran Alistair Lowe-Norris on ISO standards, compliance frameworks, and building safer AI by design.
kairos.fm
October 20, 2025 at 9:11 PM
Maybe I'm crazy, but this AC review I received from NeurIPS D&B track seems to essentially say "this is great," followed by a comment stating it has been rejected without any context.

Final scores were 4 5 4 4, i.e. all reviewers and the AC agreed the paper was an Accept.

Absolutely wild.
September 26, 2025 at 9:32 PM
Reposted by Jacob Haimes
🚨 New Into AI Safety episode is live!

Li-Lian Ang from BlueDot Impact discusses their evolution from broad AI safety courses to targeted impact acceleration, addressing elitism in the field, and why we need more advocates beyond just technical researchers.

kairos.fm/intoaisafety/e023
September 16, 2025 at 4:05 PM
Super happy to share HumanAgencyBench, which takes steps towards understanding the impact of chatbot interactions on huma agency.

Working with @jacyanthis.bsky.social (and the team) has been fantastic, and I'd happily do it again. If you have the chance to work with him, don't pass it up!
LLM agents are optimized for thumbs-up instant gratification. RLHF -> sycophancy

We propose human agency as a new alignment target in HumanAgencyBench, made possible by AI simulation/evals. We find e.g., Claude most supports agency but also most tries to steer user values 👇 arxiv.org/abs/2509.08494
September 15, 2025 at 5:37 PM
Reposted by Jacob Haimes
🚨 New Into AI Safety episode is live!

I chatted with Andres Sepulveda Morales, founder of Red Mage Creative and organizer of the Fort Collins Rocky Mountain AI Interest Group about surviving the tech layoff cycle, dark patterns in AI, and building inclusive AI communities.
August 5, 2025 at 3:16 AM
Reposted by Jacob Haimes
👀 New Into AI Safety episode is live!

Will Petillo from PauseAI joins to discuss the grassroots movement for pausing frontier AI development, balancing diverse perspectives in activism, and why meaningful AI governance requires both political engagement and public support kairos.fm/intoaisafety...
June 24, 2025 at 12:36 AM
Reposted by Jacob Haimes
🚨 New episode is out: "One Big Bad Bill" - breaking down AI's relevance to Trump's bill. We cover automated fraud detection, government data consolidation, and a 10-year ban on state AI regulation.

Find us on Spotify, Apple Podcasts, YouTube, or wherever you listen (links in thread).
One Big Bad Bill | Kairos.fm
Breaking down Trump's massive bill: AI fraud detection, centralized databases, military integration, and a 10-year ban on state AI regulation.
kairos.fm
June 23, 2025 at 10:23 PM
Reposted by Jacob Haimes
New muckrAIkers episode drops! We're breaking down the wild economic claims around AI into 3 buckets, and digging into what the data actually shows 📊 kairos.fm/muckraikers/...

You can find the show on Spotify, Apple Podcasts, YouTube, or wherever else you listen (links in thread).
Breaking Down the Economics of AI | Kairos.fm
We break down 3 clusters of AI economic hype: automating profit centers, removing cost centers, and explosive growth. Reality check included.
kairos.fm
May 26, 2025 at 6:03 PM
Reposted by Jacob Haimes
NEW EPISODE: "Making Your Voice Heard w/ Tristan Williams & Felix de Simone" - where we explore how everyday citizens can influence AI policy through effective communication with legislators 🎙️ kairos.fm/intoaisafety...

Listen on Spotify, Apple Podcasts, YouTube, or wherever you get your podcasts!
May 19, 2025 at 9:05 PM
Reposted by Jacob Haimes
New muckrAIkers episode! DeepSeek R1 - What is "reasoning" and does it actually change the AI landscape? Industry fallout, billion dollar market crash, and why we're skeptical about the hype. kairos.fm/muckraikers/...

Listen on Spotify, Apple Podcasts, YouTube, or wherever you get your podcasts!
DeepSeek: 2 Months Out | Kairos.fm
Deep dive into DeepSeek; what is reasoning, and does it change the "AI" landscape?
kairos.fm
April 9, 2025 at 4:45 PM
Incredibly disappointing to see the current US administration attempting to make safe and ethical "AI" a partisan issue:

"The US has also demanded that the final statement excludes any mention of the environmental cost of AI, existential risk or the UN." - www.thetimes.com/article/a7ae...

☹️
AI summit draft declaration criticised for lack of safety progress
A leaked version of the Paris AI summit document omits key commitments made at Bletchley Park in 2023 in ‘negligence of an unprecedented magnitude’
www.thetimes.com
February 10, 2025 at 5:28 PM
Reposted by Jacob Haimes
This week's episode of muckrAIkers is a sneak preview at all of the stories we soon are going to tackle in depth on DeepSeek R1.

Developments are ongoing, but if you want a good 15 minute overview of new so far, check out kairos.fm/muckraikers/... or find us wherever you listen!
DeepSeek Minisode | Kairos.fm
A short update on DeepSeek.
kairos.fm
February 10, 2025 at 4:20 PM
Excited to share the first blogpost output from the Apart Lab Studio (@apartresearch.bsky.social) by Reworr, which I had the pleasure of supporting!

Check it out for one way to actively monitor one kind of AI misuse: LLM-based cyberattacks.

www.apartresearch.com/post/hunting...
AI Hackers in the Wild: LLM Agent Honeypot
This Apart Lab Studio research blog attempts to ascertain the current state of AI-powered hacking in the wild through an innovative 'honeypot' system designed to detect LLM-based attackers.
www.apartresearch.com
February 1, 2025 at 12:56 AM
Reposted by Jacob Haimes
Super excited to announce our latest episode of muckrAIkers: Understanding AI World Models w/ Chris Canal! We get into test-time compute, the moving goalposts of “AGI,” and so much more kairos.fm/muckraikers/...

You can find the show on Spotify, Apple Podcasts, YouTube, or wherever else you listen.
Understanding AI World Models w/ Chris Canal | Kairos.fm
Chris Canal, founder of Equistamp, joins muckrAIkers as our first ever podcast guest to discuss AI risks and the world models that inform them.
kairos.fm
January 27, 2025 at 4:19 PM
A recent @apartresearch.bsky.social Researcher Spotlight featured me! Check it out to hear more about my journey Into AI Safety (pun intended):
www.youtube.com/watch?v=lFAm...
Researcher Spotlight: Jacob Haimes
YouTube video by Apart - Safe AI
www.youtube.com
January 26, 2025 at 7:44 PM