thegermanpole.bsky.social
@thegermanpole.bsky.social
Reposted
🚨 New muckrAIkers: "AI Safety For Who?"

@jacobhaimes.bsky.social & @thegermanpole.bsky.social break down how instruction tuning/RLHF create anthropomorphic chatbots that exploit human empathy—leading to mental health harms. kairos.fm/muckraikers/...

Find us wherever you listen! (links in thread)
AI Safety for Who? | Kairos.fm
AI safety is making you less safe: chatbot anthropomorphization, mental health harms, dark patterns
kairos.fm
October 13, 2025 at 8:42 PM
Dangers of being ab auto-didact: I somehow missed the identity $a^b=e^(b ln(a))$ for the last 5 years.

I'm technically semi-autodidact because I did get formal training, but somewhere that very useful and basic thing got lost and I never had enough pain to derive it myself.
February 9, 2025 at 7:57 AM
Kairos.fm Megapack
Join the conversation
go.bsky.app
January 27, 2025 at 4:56 PM
Probably our best episode yet, Chris was an awesome guest and is a true Mensch
Super excited to announce our latest episode of muckrAIkers: Understanding AI World Models w/ Chris Canal! We get into test-time compute, the moving goalposts of “AGI,” and so much more kairos.fm/muckraikers/...

You can find the show on Spotify, Apple Podcasts, YouTube, or wherever else you listen.
Understanding AI World Models w/ Chris Canal | Kairos.fm
Chris Canal, founder of Equistamp, joins muckrAIkers as our first ever podcast guest to discuss AI risks and the world models that inform them.
kairos.fm
January 27, 2025 at 4:50 PM
Reposted

🚨 New Paper Alert: Open Problem in Machine Unlearning for AI Safety 🚨

Can AI truly "forget"? While unlearning promises data removal, controlling emergent capabilities is a inherent challenge. Here's why it matters: 👇

Paper: arxiv.org/pdf/2501.04952
1/8
January 10, 2025 at 4:58 PM