Cameron Jones
banner
camrobjones.bsky.social
Cameron Jones
@camrobjones.bsky.social
Postdoc in the Language and Cognition lab at UC San Diego. I’m interested in persuasion, deception, LLMs, and social intelligence.
I’m really proud to have (in a minor way) contributed to this update and the upcoming 2026 report.

Whether or not you’re closely following capabilities/safety progress it’s an incredibly useful resource: a rigorous, concise, & well-evidenced summary of developments!
AI is evolving too quickly for an annual report to suffice. To help policymakers keep pace, we're introducing the first Key Update to the International AI Safety Report. 🧵⬇️

(1/10)
October 16, 2025 at 6:18 PM
Totally agree with @seantrott.bsky.social here. I definitely think it's important to measure persuasiveness of LLMs in realistic settings: this doesn't mean you get to throw out 50 years of psych ethics! seantrott.substack.com/p/informed-c...
Informed consent is central to research ethics
On the unauthorized experiment conducted on a subreddit community.
seantrott.substack.com
May 8, 2025 at 9:34 PM
Reposted by Cameron Jones
🧪
Yes, LLMs can now pass the Turing test, but don’t confuse this with AGI, which is a long way off.

arxiv.org/abs/2503.23674
Large Language Models Pass the Turing Test
We evaluated 4 systems (ELIZA, GPT-4o, LLaMa-3.1-405B, and GPT-4.5) in two randomised, controlled, and pre-registered Turing tests on independent populations. Participants had 5 minute conversations s...
arxiv.org
April 3, 2025 at 7:21 AM
New preprint: we evaluated LLMs in a 3-party Turing test (participants speak to a human & AI simultaneously and decide which is which).

GPT-4.5 (when prompted to adopt a humanlike persona) was judged to be the human 73% of the time, suggesting it passes the Turing test (🧵)
April 1, 2025 at 3:14 PM
Reposted by Cameron Jones
Check it out for cool plots like this about how affinities between words in sentences and how they can show how Green Day isn't like green paint or green tea. And congrats to @coryshain.bsky.social and the CLiMB lab! climblab.org
March 11, 2025 at 8:04 PM
Reposted by Cameron Jones
📈Out today in @PNASNews!📈

In a large pre-registered experiment (n=25,982), we find evidence that scaling the size of LLMs yields sharply diminishing persuasive returns for static political messages. 

🧵:
March 7, 2025 at 6:28 PM
@yann-lecun.bsky.social at #StandUpForScience NYC in Washington Square Park — “I work on both natural and artificial intelligence, and I think this government could do with a little more intelligence.”
March 7, 2025 at 5:46 PM
Reposted by Cameron Jones
#StandUpForScience today! NYC is 12-3 PM EST in Washington Square Park, details about other cities here: standupforscience2025.org
STAND UP FOR SCIENCE
March 7, 2025. Washington DC and nationwide. Because science is for everyone.
standupforscience2025.org
March 7, 2025 at 1:14 PM
Reposted by Cameron Jones
Today in AI weirdness: if you fine-tune a model to deliberately produce insecure code it also "asserts that humans should be enslaved by AI, gives malicious advice, and acts deceptively" www.emergent-misalignment.com
Emergent Misalignment
Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs
www.emergent-misalignment.com
February 25, 2025 at 9:39 PM
Reposted by Cameron Jones
Thanks to @kensycoop.bsky.social for this great interview about my book.

We cover domestication syndrome, plasticity-led evolution, soft inheritance, animal traditions, how culture shapes evolution, and more.

Kensy also does a wonderful production job, turning me into a coherent speaker! Thank you
New episode!! 📣📣

A conversation w/ @kevinlala.bsky.social about his new (co-authored) book, ‘Evolution Evolving'!

Ideas about evolution have changed a lot in recent decades. An emerging view—synthesized by Lala et al.—puts developmental processes front and center.

Listen: disi.org/the-developm...
February 25, 2025 at 3:43 PM
Reposted by Cameron Jones
Any talk you hear from the current administration about making the US more competitive in science and technology is utter bullshit. What they are doing is sabotaging our country for years if not decades to come.
February 22, 2025 at 11:03 PM
I wrote up some notes on my trip to the first @IASEAIorg conference—mostly on the importance of "agents", the risks that they might pose, and how/whether we can mitigate them.

camrobjones.substack.com/p/notes-from...
Notes from IASEAI
On agents, ethics, and catastrophic risks
camrobjones.substack.com
February 16, 2025 at 8:22 PM
We've relaunched @turingtestlive with a 3-party format where you speak to a human and an LLM at the same time.

See if you can tell the difference between a human and an AI here: turingtest.live
The Turing Test — Can you tell a human from an AI?
The Turing Test — Can you tell a human from an AI?
Turingtest.live
February 10, 2025 at 12:32 PM
Reposted by Cameron Jones
Andy Whiten and I wrote a @science.org perspective about a cool new study from @inbalarnon.bsky.social @simonkirby.bsky.social @ellengarland.bsky.social et al! They found humpback whale song has language-like statistical structure, using methods inspired by infant language learning 🐋🎶 Links below ⬇️
February 6, 2025 at 7:10 PM
I’m in Paris for IASEAI, let me know if you’re around and would want to meet up!
February 5, 2025 at 12:30 PM
This article provides help on how to write code and documentation to help LLMs use your framework. This is what the real AI takeover looks like. encore.dev/blog/llm-ins...
Using LLMs to help LLMs build Encore apps – Encore Blog
How we used LLMs to produce instructions for LLMs to build Encore applications.
encore.dev
January 16, 2025 at 12:49 PM
How effective are LLMs are persuading and deceiving people? In a new preprint we review different theoretical risks of LLM persuasion; empirical work measuring how persuasive LLMs currently are; and proposals to mitigate these risks. 🧵

arxiv.org/abs/2412.17128
Lies, Damned Lies, and Distributional Language Statistics: Persuasion and Deception with Large Language Models
Large Language Models (LLMs) can generate content that is as persuasive as human-written text and appear capable of selectively producing deceptive outputs. These capabilities raise concerns about pot...
arxiv.org
January 10, 2025 at 1:59 PM
You can now access Turing Test Live any time! turingtest.live
December 23, 2024 at 4:53 PM
Reposted by Cameron Jones
Sometimes o1's thinking time almost feels like a slight. o1 is like "oh I thought about this uninvolved question of yours for 7 seconds and here is my 20 page essay on it"
December 15, 2024 at 5:38 PM
Can an AI convince you it's human? Can you convince another human you're not an AI? Find out at turingtest.live. Live now! And daily: 1–2 PM & 8–9 PM GMT.
The Turing Test — Can you tell a human from an AI?
The Turing Test — Can you tell a human from an AI?
turingtest.live
December 20, 2024 at 1:10 PM
Turing test live uses a 3-party format, where you chat with a human and an AI simultaneously. Can you tell them apart? Live now and every day from 1–2 PM & 8–9 PM GMT at turingtest.live.
The Turing Test — Can you tell a human from an AI?
The Turing Test — Can you tell a human from an AI?
turingtest.live
December 19, 2024 at 1:20 PM
I’m running an experiment to see how well LLMs do at a Turing test at turingtest.live.

You can play now! (For the next hour, and then every day from 8-9am and 3-4am ET).

It uses a 3-player format where you talk to a human and an LLM simultaneously and have to decide which is which.
The Turing Test — Can you tell a human from an AI?
The Turing Test — Can you tell a human from an AI?
turingtest.live
December 12, 2024 at 8:14 PM
turingtest.live is back up! With new models, prompts, and 3-party format where you speak to a person and an LLM simultaneously.

See if you can tell the difference between human and an AI!
The Turing Test — Can you tell a human from an AI?
The Turing Test — Can you tell a human from an AI?
turingtest.live
December 12, 2024 at 12:54 PM
We're relaunching turingtest.live on Thursday at 1pm GMT / 8am ET / 5am PT. The new site will use a 3 player format where you speak to a human and an AI simultaneously and decide which is which! We're also testing a variety of new prompting approaches.
December 9, 2024 at 4:56 PM
Reposted by Cameron Jones
Do you know what rating you’ll give after reading the intro? Are your confidence scores 4 or higher? Do you not respond in rebuttal phases? Are you worried how it will look if your rating is the only 8 among 3’s? This thread is for you.
November 27, 2024 at 5:25 PM