Yoshua Bengio
banner
yoshuabengio.bsky.social
Yoshua Bengio
@yoshuabengio.bsky.social

Working towards the safe development of AI for the benefit of all at Université de Montréal, LawZero and Mila.

A.M. Turing Award Recipient and most-cited AI researcher.

https://lawzero.org/en
https://yoshuabengio.org/profile/ .. more

Yoshua Bengio is a Canadian computer scientist, and a pioneer of artificial neural networks and deep learning. He is a professor at the Université de Montréal and scientific director of the AI institute MILA. .. more

Computer science 91%
Physics 3%
Pinned
Today marks a big milestone for me. I'm launching @law-zero.bsky.social, a nonprofit focusing on a new safe-by-design approach to AI that could both accelerate scientific discovery and provide a safeguard against the dangers of agentic AI.
Every frontier AI system should be grounded in a core commitment: to protect human joy and endeavour. Today, we launch LawZero, a nonprofit dedicated to advancing safe-by-design AI. lawzero.org

We're glad to have you at LawZero, Iulian. Bienvenue dans l'équipe!
We are thrilled to welcome Iulian Serban to LawZero as Senior Director, Research and Development.

As former founder of Korbit, he brings deep expertise in GenAI, software security, and research to our mission.

Full press release: lawzero.org/en/news/lawz...

Reposted by Yoshua Bengio

We are thrilled to welcome Iulian Serban to LawZero as Senior Director, Research and Development.

As former founder of Korbit, he brings deep expertise in GenAI, software security, and research to our mission.

Full press release: lawzero.org/en/news/lawz...

Reposted by Yoshua Bengio

For example, research shows that:
· When given 10 attempts, attackers can use malicious prompts to bypass leading systems' safeguards about half the time.
· Inserting as few as 250 malicious documents into a model's training data can introduce vulnerabilities.
(6/6)

Yet significant challenges remain and the real-world effectiveness of many safeguards is uncertain. ⬇️
(5/6)

· A growing number of companies adopting Frontier AI Safety Frameworks, describing safety and security measures they will take as their AI models become more capable,
· Technical safeguards are beginning to inform transparency measures in governance frameworks.
(4/6)

Over the past year, we’ve seen meaningful progress:
· Improvements in adversarial training methods to make models more resistant to potentially harmful requests,
· Better tools for tracking AI-generated content.
(3/6)

This key update, paired with the first one published in October, aims to provide policymakers and researchers with timely information about important trends in frontier AI development.
Read the full Update here: internationalaisafetyreport.org/publication/...
(2/6)
Second Key Update: Technical Safeguards and Risk Management
This Key Update examines developments in technical approaches to general-purpose AI risk management, from training models to refuse harmful requests to watermarking AI-generated content. Since the pub...
internationalaisafetyreport.org

I’m pleased to share the Second Key Update to the International AI Safety Report, which outlines how AI developers, researchers, and policymakers are approaching technical risk management for general-purpose AI systems.
(1/6)

Learn more about the 5 main concrete actions they should pursue in our latest Blueprint for Multinational Advanced AI Development: aigi.ox.ac.uk/publications...
A Blueprint for Multinational Advanced AI Development - Oxford Martin AIGI
Executive Summary The global race to develop advanced AI has entered a newphase marked by staggering investments, rapid technical breakthroughs, and intensifying geopolitical competition. The United S...
aigi.ox.ac.uk

Geopolitical competition leaves AI bridge powers in a difficult situation where they’ll soon likely face insurmountable barriers to independent frontier AI development. To stay relevant and thrive economically, they need to work together and strategically choose their AI development approaches.

Reposted by Yoshua Bengio

Open-weight models are becoming increasingly capable while creating risks beyond those that already exist for closed-weight models.
To continue benefiting from the advantages of open-weight models, we must develop risk mitigation methodologies specifically for them, as discussed in this paper.
🚨New paper🚨

From a technical perspective, safeguarding open-weight model safety is AI safety in hard mode. But there's still a lot of progress to be made. Our new paper covers 16 open problems.

🧵🧵🧵

Reposted by Yoshua Bengio

🚨New paper🚨

From a technical perspective, safeguarding open-weight model safety is AI safety in hard mode. But there's still a lot of progress to be made. Our new paper covers 16 open problems.

🧵🧵🧵

It was equally a pleasure to receive this distinction alongside an exceptional group of colleagues, whose contributions have had a profound impact on the field of AI as we know it today. www.youtube.com/watch?v=0zXS...
The Minds of Modern AI: Jensen Huang, Yann LeCun, Fei-Fei Li & the AI Vision of the Future | FT Live
YouTube video by FT Live
www.youtube.com

I was very honoured to receive the Queen Elizabeth Prize for Engineering from His Majesty King Charles III this week, and pleased to hear his thoughts on AI safety as well as his hopes that we can minimize the risks while collectively reaping the benefits.

Addressing these risks doesn't mean stopping progress.
Innovation & regulation must go hand in hand, notably by building technical safeguards to make AI systems more trustworthy for individuals and businesses. That's at the heart of our work at @law-zero.bsky.social.

Thank you to @financialtimes.com for the invitation to speak at today's FT Summit and to Cristina Criddle for the excellent discussion.
We touched on AI's early signs of self-preservation and deceptive behaviours, as well as the technical and policy solutions on the horizon.

We need innovative technical and societal solutions to mitigate AI risks. I believe liability insurance for AI developers could be an excellent market-based incentive to drive safety standards and accountability, and is an option worth considering.
www.ft.com/content/181f...
Force AI firms to buy nuclear-style insurance, says Yoshua Bengio
Turing Prize winner urges governments to require tech groups to cover catastrophic outcomes and fund safety research
www.ft.com

Thank you to the University of Copenhagen, the European Commission @ec.europa.eu, and my co-panelists — Lene Oddershede, Max Welling, Peter Sarlin & @fabiantheis.bsky.social — for a day of excellent discussions.

I want to congratulate EVP @hennavirkkunen.bsky.social and EU Commissioner Ekaterina Zaharieva on the launch of RAISE, a critical initiative to support AI for Science and Science for AI.