Andrew Strait
banner
agstrait.bsky.social
Andrew Strait
@agstrait.bsky.social
UK AI Security Institute

Former Ada Lovelace Institute, Google, DeepMind, OII
Man, even the brocast community appears to be reading @shannonvallor.bsky.social 's book.
July 24, 2025 at 8:47 PM
Notable new NBER study on genAI and labor: despite widespread adoption of AI chatbots in Danish workplaces, their impact on earnings and hours worked is negligible. Productivity gains average just 3%, challenging the narrative of AI-driven labor market disruption

www.nber.org/system/files...
May 19, 2025 at 8:01 AM
*sweating in ACT*
March 10, 2025 at 9:24 PM
This is really cool work from @apartresearch.bsky.social - DarkBench to test different models for various dark patterns.

Would love to see more benchmarks like these!

www.apartresearch.com/post/uncover...
February 20, 2025 at 4:43 PM
@lujain.bsky.social has published an excellent new paper exploring anthropomorphic behaviours in LLMs. Notable finding - majority of these behaviours occur after multi-turn interactions

arxiv.org/abs/2502.07077
February 13, 2025 at 2:35 PM
Paris AI action summit kicking off today
February 10, 2025 at 8:19 AM
A friend pointed out the side-by-side comparison of the changes to the Google RAI principles.

First two are the originals, 3/4/5 are the new ones.

Goodbye, redlines on weapons, surveillance, and human-rights abusing tech...

web.archive.org/web/20230804...

ai.google/responsibili...
February 5, 2025 at 12:11 PM
Behold the power of an open-weight model release.
January 27, 2025 at 12:42 PM
This was an interesting blog post about labor and economic impacts of agentic systems. I want to stop on this point about 'collecting job task data', as I'm seeing it as the basis for extrapolations that an o3 + will inevitably be as good at any kind of task

www.strangeloopcanon.com/p/what-would...
January 24, 2025 at 1:53 PM
Something my grandmother wrote in 1973 after the Watergate scandal to her local newspaper (the Washington Post). Feels very relevant for this week.
January 21, 2025 at 6:39 PM
From an excellent CJR report by @klaudia.bsky.social and @aisvarya17.bsky.social on how ChatGPT misrepresents sources.

This feature of LLMs in particular is the root cause of so many problems when adopted for tasks where accuracy matters.

www.cjr.org/tow_center/h...

H/t @bridainep.bsky.social
December 12, 2024 at 11:15 AM
Oh no.
November 25, 2024 at 1:59 PM
Sections on evidence collecting and indicators (9 and 10) look very strong to me. There are also excellent proposals for post-market monitoring, incident reporting, and whistleblowing. And requirements for executive and board level governance look great.
November 14, 2024 at 2:10 PM
That said, this sections on the nature and drivers of risk are brilliant and get us closer clarity on how to define these abstract concepts. I would, however, question how likelihood and severity are determined - who gets to make that call? How will marginalised groups be involved?
November 14, 2024 at 2:10 PM
The COP requires GPAI model providers to test and mitigate systemic risks - so how does the COP define systemic risk?

This is one issue I have - why exclude surveillance, privacy, human agency, quality of service harms, economic security and human rights? These should surely be included.
November 14, 2024 at 2:10 PM
Love the intended use and AUP policy requirements. These look great.
November 14, 2024 at 2:10 PM
First, great to see some stringent requirements in place for copyrighted data and energy disclosure. This does need to be made public if rightsholders are to use this (I also wonder what kinds of mechanisms rightholders need to query if their works are in a training data set?)
November 14, 2024 at 2:10 PM