To find out more about the podcast go to 'Malicious use is already happening': machine-learning pioneer on making AI safer.
Below is a short summary and detailed review of this podcast written by FutureFactual:
Yoshua Bengio on AI Safety: Guardrails, Scientist AI, and the Future of Work
In this podcast excerpt, leading AI researcher Yoshua Bengio discusses how society can steer artificial intelligence away from dangerous outcomes and toward safety by building guardrails that are both technical and political. The conversation covers the International AI Safety Report, the three main risk categories—malicious use, unintended malfunctions, and systemic risks like livelihoods disruption—and the idea of Scientist AI, a non-agentic predictor designed to align with human values and support safer experimentation. The discussion also touches on howAI safety work feeds into governance, the potential economic effects of AI, and the need for proactive policy to manage labor market changes and democratic resilience.
Introduction and context
In a detailed interview, Nature's Dare with Benjamin Thompson features Joshua (Yoshua) Bengio, a foundational voice in AI research from the University of Montreal. The discussion centers on how rising AI capabilities create both opportunities and risks for society, and how a coordinated, safety‑forward approach can guide development. Bengio notes that the International AI Safety Report has helped formalize what researchers understand from the literature and has spurred the creation of government AI safety institutes that connect researchers with policymakers. He emphasizes that broad, accessible communication about AI safety is essential for civic understanding and democratic accountability.
Risks, guardrails, and current reality
Bengio identifies three primary risk areas for AI: unintended risks from malfunctions, malicious use, and systemic risks such as job displacement and erosion of livelihoods. He argues that malicious use is already occurring in subtler forms, with deep fakes and cyberattacks illustrating the kinds of threats that require robust guardrails that blend technical safeguards with political and regulatory measures. "Malicious use is already happening, but we're only seeing shades of it" - Joshua Bengio
Scientist AI: safety by design
The conversation introduces the Scientist AI concept: an AI system modeled after the scientific method, focused on understanding causal mechanisms and producing high‑quality, truthful predictions. Bengio argues that this non‑agentic AI has no hidden goals and can serve as a trustworthy advisor to guard against the misalignment that plagues current agentic systems. The idea is to create guardrails that allow AI to assist experimentation and decision‑making without pursuing objectives that drift from human values. "The scientists AI is non‑agentic. In other words, it has no goal, it has no intention" - Joshua Bengio
Economic and labor-market implications
Addressing systemic risk, Bengio discusses how AI is likely to boost productivity but could also concentrate wealth and power, potentially worsening inequality if policies do not adapt. He cites the International AI Safety Report as a basis for understanding mitigation approaches and the limitations of existing tools, and he stresses that policymakers must prepare for significant labor‑market shifts. "AI will enable much more productivity across the board" - Joshua Bengio
Policy, governance, and the path forward
The dialogue highlights the need for guardrails that are both technical and political, and for international collaboration to ensure robust AI governance. Bengio argues for proactive policy design to steer developments toward beneficial outcomes, stressing that governments must engage with researchers to preempt risks and protect democracy. He also notes the current optimism around rapid progress should be tempered by imagination and foresight about long‑term consequences.
Quotes and takeaways
"Malicious use is already happening, but we're only seeing shades of it" - Joshua Bengio
"how do we build AI that will not harm humans by design" - Joshua Bengio
"The scientists AI is non‑agentic. In other words, it has no goal, it has no intention" - Joshua Bengio
"AI will enable much more productivity across the board" - Joshua Bengio