Leading Conceptualization Lab http://conceptualization.ai
Minting new vocabulary to conceptualize generative models.
I think this is why LLMs often feel 'fixated on the wrong thing' or 'overly literal'—they are usually responding using the most relevant single thing they remember, not the aggregate of what was said
We introduce Oolong, a dataset of simple-to-verify information aggregation questions over long inputs. No model achieves >50% accuracy at 128K on Oolong!
I think this is why LLMs often feel 'fixated on the wrong thing' or 'overly literal'—they are usually responding using the most relevant single thing they remember, not the aggregate of what was said
is it awesome or painful or something you live in fear of?
is it awesome or painful or something you live in fear of?
Me: Show me the seahorse emoji
Grok: Here it is: 🦐
Wait, that's a shrimp. My bad—the actual seahorse emoji is: 🦎
No, that's a lizard. Let me get this right: the seahorse is 🦈? Shark? Nope.
Actually, the real seahorse emoji is: 🦑
1. Understanding how LLMs 'see' the world (ex: LMs can't see conspicious omissions, see AbsenceBench)
2. How can we make things with LLMs that have never been made before? (ex: Communnication Games, see 📌)
3. See my other posts :)
Me: Show me the seahorse emoji
Grok: Here it is: 🦐
Wait, that's a shrimp. My bad—the actual seahorse emoji is: 🦎
No, that's a lizard. Let me get this right: the seahorse is 🦈? Shark? Nope.
Actually, the real seahorse emoji is: 🦑
1. Understanding how LLMs 'see' the world (ex: LMs can't see conspicious omissions, see AbsenceBench)
2. How can we make things with LLMs that have never been made before? (ex: Communnication Games, see 📌)
3. See my other posts :)
1. Understanding how LLMs 'see' the world (ex: LMs can't see conspicious omissions, see AbsenceBench)
2. How can we make things with LLMs that have never been made before? (ex: Communnication Games, see 📌)
3. See my other posts :)
Humans pass the mirror test at ~18 months 👶
But what about LLMs? Can they recognize their own writing—or even admit authorship at all?
In our new paper, we put 10 state-of-the-art models to the test. Read on 👇
1/n 🧵
Humans pass the mirror test at ~18 months 👶
But what about LLMs? Can they recognize their own writing—or even admit authorship at all?
In our new paper, we put 10 state-of-the-art models to the test. Read on 👇
1/n 🧵
THE EVERYTHING MACHINE IS HERE NOW USE IT
THE EVERYTHING MACHINE IS HERE NOW USE IT
The dream of “autonomous AI scientists” is tempting:
machines that generate hypotheses, run experiments, and write papers. But science isn’t just automation.
cichicago.substack.com/p/the-mirage...
🧵
The dream of “autonomous AI scientists” is tempting:
machines that generate hypotheses, run experiments, and write papers. But science isn’t just automation.
cichicago.substack.com/p/the-mirage...
🧵
1) How can I change course while maintaining the machine that keeps me alive?
2) Is this machine actually keeping me alive?
1) How can I change course while maintaining the machine that keeps me alive?
2) Is this machine actually keeping me alive?
Link to apply in the next tweet.
Link to apply in the next tweet.
(Gemini integration in docs is abyssmal, same with Notion's AI which I think is ChatGPT but somehow does not condition on enough context to be useful?)
(Gemini integration in docs is abyssmal, same with Notion's AI which I think is ChatGPT but somehow does not condition on enough context to be useful?)
Postdoc: uchicago.infoready4.com#freeformComp...
Assistant Professor: apply.interfolio.com/174766
Associate Professor: apply.interfolio.com/174768
Postdoc: uchicago.infoready4.com#freeformComp...
Assistant Professor: apply.interfolio.com/174766
Associate Professor: apply.interfolio.com/174768