Working towards the safe development of AI for the benefit of all at Université de Montréal, LawZero and Mila.
A.M. Turing Award Recipient and most-cited AI researcher.
https://lawzero.org/en
https://yoshuabengio.org/profile/ ..
more
Working towards the safe development of AI for the benefit of all at Université de Montréal, LawZero and Mila.
A.M. Turing Award Recipient and most-cited AI researcher.
https://lawzero.org/en
https://yoshuabengio.org/profile/
Yoshua Bengio is a Canadian computer scientist, and a pioneer of artificial neural networks and deep learning. He is a professor at the Université de Montréal and scientific director of the AI institute MILA. .. more
Along with several of my colleagues, I have pledged to help, and I encourage anyone who can to do the same.
openreview.net/donate
www.youtube.com/watch?v=zQ1P...
Thank you to TED for the chance to share why we need to prioritize safety in AI, and my vision for a safer path forward.
www.ted.com/playlists/88...
We've seen major growth in AI capabilities & risks, but as
@law-zero.bsky.social continues its research, I'm increasingly confident that technical solutions are possible to make AI safer.
time.com/7339687/yosh...
As former founder of Korbit, he brings deep expertise in GenAI, software security, and research to our mission.
Full press release: lawzero.org/en/news/lawz...
Reposted by Yoshua Bengio
As former founder of Korbit, he brings deep expertise in GenAI, software security, and research to our mission.
Full press release: lawzero.org/en/news/lawz...
· When given 10 attempts, attackers can use malicious prompts to bypass leading systems' safeguards about half the time.
· Inserting as few as 250 malicious documents into a model's training data can introduce vulnerabilities.
(6/6)
(5/6)
· Technical safeguards are beginning to inform transparency measures in governance frameworks.
(4/6)
· Improvements in adversarial training methods to make models more resistant to potentially harmful requests,
· Better tools for tracking AI-generated content.
(3/6)