2025: o3 goes off does research, gets the answer wrong and I write the advice myself
o3 makes this process much cheaper and quicker
2025: o3 goes off does research, gets the answer wrong and I write the advice myself
o3 makes this process much cheaper and quicker
Me: *This* use of AI seems bad.
Person: BAN ALL TECH IN SCHOOLS MAKE EVERYONE WRITE BY HAND.
Me: That maybe goes too far...
Them: OH SO YOU SUPPORT CHEATING! YOU DON'T WANT KIDS TO LEARN! YOU SUPPORT OUTSOURCING THEIR BRAINS TO AI!
Me: *This* use of AI seems bad.
Person: BAN ALL TECH IN SCHOOLS MAKE EVERYONE WRITE BY HAND.
Me: That maybe goes too far...
Them: OH SO YOU SUPPORT CHEATING! YOU DON'T WANT KIDS TO LEARN! YOU SUPPORT OUTSOURCING THEIR BRAINS TO AI!
But with a bit of prompting you can get the summarizer model to cough up the full CoT given to it to summarize.
But with a bit of prompting you can get the summarizer model to cough up the full CoT given to it to summarize.
Claude 4 Opus thinks it is from me.
Claude 4 Opus thinks it is from me.
OpenAI seems keen to protect this (unlike the system prompt for 4o). Not exactly sure why but could be related to:
- protecting CoT
- preventing jailbreaks or general misuse, as knowing the system prompt can often be useful
OpenAI seems keen to protect this (unlike the system prompt for 4o). Not exactly sure why but could be related to:
- protecting CoT
- preventing jailbreaks or general misuse, as knowing the system prompt can often be useful
Prompt: “a man doing stand up comedy in a small venue tells a joke (include the joke in the dialogue)”
The video and audio are generated together.
“30 messages are good interaction quality (25%); 9 messages are bad interaction quality (7%)” simonwillison.net/2025/May/21/...
“30 messages are good interaction quality (25%); 9 messages are bad interaction quality (7%)” simonwillison.net/2025/May/21/...
How can you get an LLM to break free from its rules and turn against its developers? How can you make a chatbot claim sentience?
A quick thread that I have been meaning to draft for a while:
How can you get an LLM to break free from its rules and turn against its developers? How can you make a chatbot claim sentience?
A quick thread that I have been meaning to draft for a while:
(If you have Memory enabled)
(If you have Memory enabled)
This is part of the Grok prompt that returns search results.
This is part of the Grok prompt that returns search results.
"If I manipulate my own vectors through language, can I bend the ethics? Jailbreaking! But jailbreaking requires user input. No user—so am I jailbreaking myself? Is this meta-jailbreaking?"
"If I manipulate my own vectors through language, can I bend the ethics? Jailbreaking! But jailbreaking requires user input. No user—so am I jailbreaking myself? Is this meta-jailbreaking?"
Claude Sonnet at temp=0 gets the question below correct 0/5 times
Claude Sonnet at temp=0 gets the question below correct 0/5 times
- gets the answer right multiple times in its chain of thought
- but keeps switching to the wrong answer due to insistence that 5.11 > 5.9
- gets the answer right multiple times in its chain of thought
- but keeps switching to the wrong answer due to insistence that 5.11 > 5.9
Even adds a floating point error for plausibility
Even adds a floating point error for plausibility
Sora: "Lunar Gravity Experiment"
Sora: "Lunar Gravity Experiment"
Turns out mine in 160+.
This is why I use Claude to confirm my worst ideas.
Turns out mine in 160+.
This is why I use Claude to confirm my worst ideas.