https://linktr.ee/florianmai
Democracy needs large reforms. It's time that alternative ways of selecting representatives get into the overton window, e.g. sortition.
Democracy needs large reforms. It's time that alternative ways of selecting representatives get into the overton window, e.g. sortition.
But of course, DOGE is not interested in doing what's right anyway.
The plan is to migrate all systems off COBOL quickly which would likely require the use of generative AI.
www.wired.com/story/doge-r...
But of course, DOGE is not interested in doing what's right anyway.
openai is culturally fascist as well as functionally fascist
it’s not an accident
🧵
🧵
The same warnings are now appearing with increasing frequency from smart outside observers of the AI industry who do not gain from hype, like Kevin Roose (below) & Ezra Klein
I think ignoring the possibility they are right is a mistake
The same warnings are now appearing with increasing frequency from smart outside observers of the AI industry who do not gain from hype, like Kevin Roose (below) & Ezra Klein
I think ignoring the possibility they are right is a mistake
Our new paper, “In House Evaluation is Not Enough” has 3 calls-to-actions to empower evaluators:
1️⃣ Standardized AI flaw reports
2️⃣ AI flaw disclosure programs + safe harbors.
3️⃣ A coordination center for transferable AI flaws.
1/🧵
Our new paper, “In House Evaluation is Not Enough” has 3 calls-to-actions to empower evaluators:
1️⃣ Standardized AI flaw reports
2️⃣ AI flaw disclosure programs + safe harbors.
3️⃣ A coordination center for transferable AI flaws.
1/🧵
I think it's relatively easy to train this via RL because Gemini Flash 2.0, also a native image generator, can identify its own mistakes.
I think it's relatively easy to train this via RL because Gemini Flash 2.0, also a native image generator, can identify its own mistakes.
LLMs are trained on just one tokenization per word, but they still understand alternative tokenizations. We show that this can be exploited to bypass safety filters without changing the text itself.
#AI #LLMs #tokenization #alignment
LLMs are trained on just one tokenization per word, but they still understand alternative tokenizations. We show that this can be exploited to bypass safety filters without changing the text itself.
#AI #LLMs #tokenization #alignment
When we combine AI Agents, the risks are unlimited, especially as they are tested in isolation
arxiv.org/abs/2502.14143
@garymarcus.bsky.social guess you have seen
No. They're firing women and people of color in the military.
They’re forbidding whole fields of research.
They’re erasing trans people from existence.
No. They're firing women and people of color in the military.
They’re forbidding whole fields of research.
They’re erasing trans people from existence.
www.youtube.com/watch?v=9_l0...
www.youtube.com/watch?v=9_l0...
What world do you want to live in?
bsky.app/profile/fabi...
What world do you want to live in?
bsky.app/profile/fabi...
"Under the partnership, Guardian reporting and archive journalism will be available as a news source within ChatGPT, alongside the publication of attributed short summaries and article extracts."
"Under the partnership, Guardian reporting and archive journalism will be available as a news source within ChatGPT, alongside the publication of attributed short summaries and article extracts."
epoch.ai/gradient-upd...
epoch.ai/gradient-upd...
Our method, MONA, prevents many such hacks, *even if* humans are unable to detect them!
Inspired by myopic optimization but better performance – details in🧵
Our method, MONA, prevents many such hacks, *even if* humans are unable to detect them!
Inspired by myopic optimization but better performance – details in🧵