Aran Nayebi
banner
anayebi.bsky.social
Aran Nayebi
@anayebi.bsky.social
Assistant Professor of Machine Learning, Carnegie Mellon University (CMU)

Building a Natural Science of Intelligence 🧠🤖

Prev: ICoN Postdoctoral Fellow @MIT, PhD @Stanford NeuroAILab
Personal Website: https://cs.cmu.edu/~anayebi
In today's Generative AI lecture, we dive into reasoning models by dissecting how DeepSeek-R1 works (GRPO vs. PPO, which removes the need for a separate value network + training with a simpler rule-based reward), and end on mechanistic interpretability to better understand those reasoning traces.
November 10, 2025 at 8:46 PM
In today's Generative AI lecture, we primarily discuss scaling laws and the key factors that go into building large-scale foundation models.

Slides: www.cs.cmu.edu/~mgormley/co...

Full course info: bsky.app/profile/anay...
October 23, 2025 at 1:44 PM
Congratulations to my Ph.D. student Reece Keller for winning the best talk award at #CRSy25 on our project building the first task-optimized autonomous agent that predicts whole-brain data! Check out the post below for other cool talks!!

Detailed summary: bsky.app/profile/reec...
🐟 @reecedkeller.bsky.social @cmu.edu explored autonomous behaviour in virtual zebrafish, where intrinsic motivation drives self-directed exploration.
October 21, 2025 at 2:41 AM
Congrats to this year's Nobel Prize winners!
Philippe's seminal work is in fact what our recent closed form UBI AI capability threshold builds on: bsky.app/profile/anay...
October 13, 2025 at 3:38 PM
My ILIAD ’25 talk, “Intrinsic Barriers & Pathways to Alignment”: why “aligning to all human values” provably can’t work, why reward hacking is inevitable in large state spaces, & how small value sets bypass “no free lunch” limits to yield formal corrigibility.

www.youtube.com/watch?v=Oajq...
Day 5 Aran Formal Guarantees of Corrigibility
YouTube video by ILIAD Conference
www.youtube.com
October 10, 2025 at 3:16 PM
A nice application of our NeuroAI Turing Test! Check out
@ithobani.bsky.social's thread for more details on comparing brains to machines!
1/X Our new method, the Inter-Animal Transform Class (IATC), is a principled way to compare neural network models to the brain. It's the first to ensure both accurate brain activity predictions and specific identification of neural mechanisms.

Preprint: arxiv.org/abs/2510.02523
October 6, 2025 at 3:52 PM
Honored to be quoted in this @newsweek.com article discussing how AI could accelerate the need for UBI.

Read more here: www.newsweek.com/ai-taking-jo...
AI Is taking jobs: Could universal basic income become a reality?
Forecasts that AI could erase tens of millions of jobs by the end of the decade appear to be making the notion of a guaranteed income less radical.
www.newsweek.com
October 5, 2025 at 3:23 PM
In today's Generative AI lecture, we talk about all the different ways to take a giant auto-complete engine like an LLM and turn it into a useful chat assistant.
October 1, 2025 at 7:46 PM
In today's Generative AI lecture, we discuss the 4 primary approaches to Parameter-Efficient Fine-Tuning (PEFT): subset, adapters, Prefix/Prompt Tuning, and Low-Rank Adaptation (LoRA).

We show each of these amounts to finetuning a different aspect of the Transformer.
September 29, 2025 at 8:00 PM
1/6 Recent discussions (e.g. Rich Sutton on @dwarkesh.bsky.social’s podcast) have highlighted why animals are a better target for intelligence — and why scaling alone isn’t enough.
In my recent @cmurobotics.bsky.social seminar talk, “Using Embodied Agents to Reverse-Engineer Natural Intelligence”,
September 29, 2025 at 2:02 PM
Excited to have this work accepted as an *oral* to NeurIPS 2025!
1/ What if we make robots that process touch the way our brains do?
We found that Convolutional Recurrent Neural Networks (ConvRNNs) pass the NeuroAI Turing Test in currently available mouse somatosensory cortex data.
New paper by @Yuchen @Nathan @anayebi.bsky.social and me!
September 18, 2025 at 9:33 PM
Excited to have this work accepted to NeurIPS 2025! See you all in San Diego!
1/ I'm excited to share recent results from my first collaboration with the amazing @anayebi.bsky.social
and @leokoz8.bsky.social !

We show how autonomous behavior and whole-brain dynamics emerge in embodied agents with intrinsic motivation driven by world models.
September 18, 2025 at 9:31 PM
In today's Generative AI lecture, we discuss how to implement Diffusion Models and go through their derivation. Next time, we discuss their deeper relationships with variational inference :)

Slides: www.cs.cmu.edu/~mgormley/co...

Full course info: bsky.app/profile/anay...
September 17, 2025 at 7:51 PM
In today's Generative AI lecture, we discuss Generative Adversarial Networks (GANs) & review probabilistic graphical models (PGMs) as a prelude to Diffusion models and VAEs, which we will discuss next time!

Slides: www.cs.cmu.edu/~mgormley/co...

Full course info: bsky.app/profile/anay...
September 15, 2025 at 9:19 PM
In today's Generative AI lecture, we cover Vision Transformers (as well as the broader notion of Encoder-Only Transformers).

We also explain the historical throughline to some of these ideas, inspired by Nobel-prize-winning observations in neuroscience!
September 11, 2025 at 1:36 AM
In today's Generative AI lecture, we give an overview of the pre-training/post-training pipeline, and discuss modern Transformer implementations, from Rotary Position Embeddings (RoPE) to Grouped Query Attention (GQA) to Sliding Window Attention.
September 8, 2025 at 8:37 PM
Cool UBI simulator made by the AI+Wellbeing Institute based on our paper! www.ai-well-being.com/building-our...
September 8, 2025 at 5:30 PM
In today's Generative AI lecture, we cover how to train a Transformer Language Model, as well as what makes it efficient at learning in order to scale to GPT levels—covering key-value caching & tokenizers, among other things:
September 4, 2025 at 1:16 AM
If you’re attending ILIAD, I’ll be presenting this work online tomorrow from 11 am - 12 pm PT!
www.iliadconference.com
August 28, 2025 at 5:30 PM
This semester, Matt Gormley & I are co-teaching CMU's Generative AI course!

Today we discussed the Transformer architecture & Multi-Headed Attention.

Follow along 👇 if you want to learn more about the tech that's powering today's AI, from ChatGPT to reasoning models to agents!
August 27, 2025 at 8:40 PM
Lab website finally up! anayebi.github.io/group/
NeuroAgents Lab
The NeuroAgents lab studies the science and societal implications of intelligent systems.
anayebi.github.io
August 6, 2025 at 3:47 PM
1/ Updated now with nearly tight lower bounds—i.e., proofs showing when alignment becomes intractable, even for ideal agents.

Key AI safety takeaways:
🧠 Too many values ⇒ makes alignment intractable
👁 Task-space growth ⇒ oversight failure
(continued below
🧵👇)
Are there fundamental barriers to AI alignment once we develop generally-capable AI agents?

We mathematically prove the answer is *yes*, and outline key properties for a "safe yet capable" agent. 🧵👇

Paper: arxiv.org/abs/2502.05934
July 31, 2025 at 3:12 PM
1/ How do we build AI systems that are corrigible—shut down when asked, tell the truth, preserve oversight—and still do something useful?

We give the first provable framework that makes it implementable—unlike RLHF or Constitutional AI, which can fail when goals conflict.

🧵👇
July 29, 2025 at 7:49 PM
🚀 New Open-Source Release! PyTorchTNN 🚀
A PyTorch library for biologically-inspired temporal neural nets: unrolling computation through time. Integrates with our recent Encoder-Attender-Decoder, which flexibly combines models (Transformer, SSM, RNN) since no single one fits all sequence tasks.
🧵👇
July 24, 2025 at 4:21 PM
As someone who loves our National Parks, this is genuinely concerning. Please RT to spread the word!
🚨 There’s a lot of horrific news out of Washington right now—but you might have missed this: Senate Republicans just introduced a plan to sell off 120 million acres of our public lands.

Let me break down what’s in the bill and why it’s a full-scale land grab. 🧵
June 17, 2025 at 6:07 AM