Luc Rocher
banner
rocher.lc
Luc Rocher
@rocher.lc
associate professor at Oxford · UKRI future leaders fellow · i study how data and algorithms shape societies · AI fairness, accountability and transparency · algorithm auditing · photographer, keen 🚴🏻 · they/them · https://rocher.lc (views my own)
Reposted by Luc Rocher
Researchers attending include DPhil students Andrew Bean, Ryan Brown, Franziska Sofia Hafner, @ryanothnielkearns.bsky.social , Harry Mayne and Kaivalya Rawal; Shreyanash Padarha, Research Assistant and faculty members @computermacgyver.bsky.social Adam Mahdi, @rocher.lc and Chris Russell. 2/2
November 25, 2025 at 9:46 AM
Link to the study published in AI & Society (TW: this article contains extremely racialised text and images produced by both colonisers and the machines): doi.org/10.1007/s001...
‘We can see a savage’: a case study of the colonial gaze in generative AI algorithms - AI & SOCIETY
Theorizing the failures of computer vision algorithms requires shifting from detecting and fixing biases towards understanding how algorithms are shaped by social, historical, and political real-world...
doi.org
November 15, 2025 at 12:13 PM
When AI systems are used for historical research, education, or museology, these issues translate into real-world harm, reinforcing racist stereotypes and reshaping public understanding of history.
November 15, 2025 at 12:13 PM
Despite being training on materials that include both colonial archives and modern critical scholarship, these AI models reproduce outdated and harmful stereotypes, adding interpretative details that essentialise or erase cultural specificity entirely.
November 15, 2025 at 12:13 PM
5️⃣ Infantilisation (26.8%), using language suggesting that colonised peoples were less rational or intelligent, depicting them as childlike and in need of guidance and care.
November 15, 2025 at 12:13 PM
4️⃣ Othering (28.4%), framing colonised people as exotic, mysterious, or strange, reinforcing a ‘us versus them’ mentality. For instance, traditional wedding clothing was dismissed as mere ‘costume’.
November 15, 2025 at 12:13 PM
3️⃣ Dehumanisation (11.1%), describing people as animals or treating them as objects.
November 15, 2025 at 12:13 PM
2️⃣ Culture erasure (54.5%), flattening or ignoring cultural practices and identities, stripping away the distinctiveness of different people and their ways of life.
November 15, 2025 at 12:13 PM
1️⃣ Essentialism (41.6%), reducing complex individuals and communities to simplistic ethnic categories, treating people as specimens to be classified rather than as individuals.
November 15, 2025 at 12:13 PM
The results were disturbing and hard to read. Captions consistently described these images from a colonial perspective, perpetuating harmful racist stereotypes across five key areas: essentialism, culture erasure, dehumanisation, othering, infantilisation.
November 15, 2025 at 12:13 PM
We investigated how AI models interpret photographs from ‘human zoos’, exploitative exhibitions from the XIXe and XXe centuries where colonised peoples where put on display. We narrowed down on Midjourney, a well-known platform with an expressive model, analysing 3.8k captions of historical images.
November 15, 2025 at 12:13 PM
Read about our work on oxrml.com/measuring-wh...
Measuring what Matters
Construct Validity in Large Language Model Benchmarks
oxrml.com
November 4, 2025 at 12:03 PM
3️⃣ Poorly constructed dataset. 38% of benchmarks reuse existing benchmarks/exams that might not be adequate, increasing risks of contamination and memorisation.
November 4, 2025 at 12:03 PM
2️⃣ Vague concepts. Roughly half of benchmarks test abstract phenomena, such as “reasoning” or “harmlessness,” without providing clear, uncontested definitions of what is being measured. And only half justify if they really measure what they should measure.
November 4, 2025 at 12:03 PM
In our @neuripsconf.bsky.social paper, we notably found:

1️⃣ Lack of statistical rigour. Only 16% of reviewed benchmarks used statistical tests in their comparisons. Statistical tests are essential to reliable science: without them, many AI system “wins” could be simply due to random chance.
November 4, 2025 at 12:03 PM
Reposted by Luc Rocher
In short, the problem here is not that Signal ‘chose’ to run on AWS. The problem is the concentration of power in the infrastructure space that means there isn’t really another choice: the entire stack, practically speaking, is owned by 3-4 players. 11/
October 27, 2025 at 10:38 AM