Yoshua Bengio
banner
yoshuabengio.bsky.social
Yoshua Bengio
@yoshuabengio.bsky.social

Working towards the safe development of AI for the benefit of all at Université de Montréal, LawZero and Mila.

A.M. Turing Award Recipient and most-cited AI researcher.

https://lawzero.org/en
https://yoshuabengio.org/profile/ .. more

Yoshua Bengio is a Canadian computer scientist, and a pioneer of artificial neural networks and deep learning. He is a professor at the Université de Montréal and scientific director of the AI institute MILA. .. more

Computer science 91%
Physics 3%
Pinned
Today marks a big milestone for me. I'm launching @law-zero.bsky.social, a nonprofit focusing on a new safe-by-design approach to AI that could both accelerate scientific discovery and provide a safeguard against the dangers of agentic AI.
Every frontier AI system should be grounded in a core commitment: to protect human joy and endeavour. Today, we launch LawZero, a nonprofit dedicated to advancing safe-by-design AI. lawzero.org

Reposted by Yoshua Bengio

Reposted by Yoshua Bengio

At LawZero, we're rethinking the building blocks of frontier AI to create an intelligent machine that is both highly capable and safe-by-design. We’re excited to share our first blog post outlining some of the objectives and core components of our Scientist AI project. 🧵
(1/4)

I’ve found the collaborative spirit of the 100+ contributors heartening, and am grateful to have benefitted from their complementary perspectives.
Thank you to all contributors for their dedication.

(19/19)

With all the noise around AI, I hope this Report provides policymakers, researchers, and the public with the reliable evidence they need to make more informed choices. We also have an “Extended Summary for Policymakers”:
internationalaisafetyreport.org/publication/...

(18/19)
2026 Report: Extended Summary for Policymakers
The Extended Summary for Policymakers of the 2026 International AI Safety Report. The second International AI Safety Report, published in February 2026, is the next iteration of the comprehensive revi...
internationalaisafetyreport.org

Because no single safeguard reliably prevents misuse or malfunctions, developers are converging on "defence-in-depth,” layering multiple measures—model-level training, input/output filters, monitoring, access controls, and governance.

(17/19)

However, safeguards remain imperfect.
Attackers can still often find ways to evade them fairly easily. One initiative crowdsourced over 60,000 successful attacks against state-of-the-art models. When given 10 attempts, testers can still generate harmful responses about half the time.

(16/19)

These safeguards inform institutional risk management approaches. For example, 12 companies published or updated Frontier AI Safety Frameworks in 2025—more than double the prior year.

(15/19)

Many technical safeguards are improving. For example, models hallucinate less and it is harder to elicit dangerous responses.

(14/19)

Even areas of uncertainty carry risks that warrant attention.
For example, in 2025 multiple companies added safeguards after pre-deployment testing could not rule out the possibility that new models could assist novices seeking to develop biological weapons.

(13/19)

Wider adoption is also raising new challenges. For example, this year we discuss early evidence on how “AI companions”, which are now used by tens of millions of people, may affect people’s emotions and social life.

(12/19)

There is little evidence of overall impacts on labour markets so far, though early-career workers in some AI-exposed occupations have seen declining employment compared with late 2022.

(11/19)

Since the last Report, we’ve seen new evidence of many emerging risks. For example, AI-generated content has become highly realistic and more useful for fraud, scams, and non-consensual intimate imagery. There is growing evidence that AI systems help malicious actors carry out cyberattacks.

(10/19)

But new capabilities pose risks, notably 8 emerging risks.
Misuse:
→ AI-generated content & criminal activity
→ Influence & manipulation
→ Cyberattacks
→ Bio & chemical risks
Malfunctions:
→ Reliability issues
→ Loss of control
Systemic risks:
→ Labor market impacts
→ Risks to human autonomy

(9/19)

These capabilities are increasingly translating into real-world impact.
At least 700 million people now use leading AI systems weekly. In the US, use of AI has spread faster than that of computers and the internet.

(8/19)

But capabilities are also “jagged:” the same model may solve complex problems yet fail at some seemingly simple tasks.

(7/19)

On capabilities: AI systems continue to improve significantly.
Leading models now achieve gold-medal performance on the International Mathematical Olympiad.
AI coding agents can complete 30-minute programming tasks with 80% reliability—up from 10-minute tasks a year ago.

(6/19)

2️⃣ Some risks, from deepfakes to cyberattacks, shifted further from theoretical concerns to real-world challenges.

3️⃣ Many safety measures improved, but remain fallible. Developers increasingly implement multiple layers of safeguards to compensate.

(5/19)

This report provides policymakers with the information they need to make these decisions.

In 2025:
1️⃣ Capabilities continued advancing rapidly, especially in coding, science, and autonomous operation.

(4/19)

AI poses an “evidence dilemma” to policymakers—capabilities evolve quickly, but scientific evidence emerges far more slowly.
Acting too early risks entrenching ineffective policies, but waiting for strong evidence may leave society vulnerable to risks.

(3/19)

Over 100 independent experts contributed to the Report, including Nobel laureates and Turing Award winners, along with an advisory panel with nominees from more than 30 countries and international organisations, including the EU, OECD and UN.
internationalaisafetyreport.org/publication/...

(2/19)
International AI Safety Report 2026
The second International AI Safety Report, published in February 2026, is the next iteration of the comprehensive review of latest scientific research on the capabilities and risks of general-purpose ...
internationalaisafetyreport.org
Today we’re releasing the International AI Safety Report 2026: the most comprehensive evidence-based assessment of AI capabilities, emerging risks, and safety measures to date. 🧵

(1/19)

Thank you WEF for the golf cart ride through Davos, but also (more importantly!) for the excellent questions about the future of AI. 🛺

www.weforum.org/videos/davos...
Ideas on the Move: Yoshua Bengio
“We will get the benefits of AI only if we steer it wisely.” On the move at Davos, Professor Yoshua Bengio explains what drives his work — the urgency of aligning AI with human values, and why optimi...
www.weforum.org