Beta
Podcast cover art for: How Is AI Being Used In The Iran War?
Science Friday
Flora Lichtman·12/03/2026

How Is AI Being Used In The Iran War?

This is a episode from podcasts.apple.com.
To find out more about the podcast go to How Is AI Being Used In The Iran War?.

Below is a short summary and detailed review of this podcast written by FutureFactual:

AI in Warfare and Regulation: Claude, Anthropic, and Public Pushback

In this Science Friday episode, Flora Lichtman speaks with AI journalist Karen Howe about the increasing role of AI in military contexts, the controversy surrounding Anthropic's Claude model, and the complex debates over autonomy and accountability in weapon systems. They also explore the tensions between Silicon Valley, Washington, and the Pentagon, the potential for autonomous decision making, and the grassroots public movements opposing rapid deployment and data-center expansion. The discussion culminates in a look at the growing public appetite for AI regulation and safer, more transparent technology development.

AI, Power, and Military Ambitions

The conversation centers on how artificial intelligence has moved from a tech sector headline to a factor in national security and military planning. Flora Lichtman frames the moment as one where the metaphor of empire, once used to describe the concentration of power in AI startups, now feels literal because of the fusion between AI capabilities and military use, and the alliance between Silicon Valley and Washington. Karen Howe, a journalist focused on AI coverage, adds that this convergence has accelerated the pace of change and raised questions about control, transparency, and accountability. The segment also highlights how public perception and political dynamics are shifting as AI pervades decision making in high-stakes settings.

"It's literal." - Flora Lichtman

The Tehran Targeting and Claude in War

The discussion turns to reported use of Anthropic's Claude (CLAUDE) to analyze intelligence data and identify bombing targets in Iran, with claims of around a thousand targets identified. The reliability issues of large language models are front and center, as AI systems can fabricate details and misidentify targets. The conversation emphasizes that the war context does not eliminate human judgment, but may substitute or bias human decision making as analysts act on AI-generated lists. The panel delves into how Claude is being used in a way that some experts call a “decision support” role, while others argue that such a role still embeds automation bias and can lead to catastrophic outcomes if not properly checked by humans and procedures.

"If you think that your technology is not good for autonomous weapons, it should also not be used for decision support systems." - Dr. Heidi Klaff, Chief Scientist at AI Now

Anthropic’s Position and the Moral Debate

The episode scrutinizes Anthropic’s stance on safety and autonomy. Dario Amade acknowledges hesitancy about autonomous weapons in the current CLAUDE iteration, while also signaling openness to future capabilities under certain human oversight. Karen Howe and Flora Lichtman unpack the moral complexity of presenting themselves as a safety-focused company when deployment choices can introduce imperial dynamics and ethical dilemmas. The discussion also touches on the broader tension between safety narratives and actual deployment realities, where the same technology can be used in ways that undermine safety if governance and accountability are weak.

"The moral high ground for Anthropic feels a little suspect" - Flora Lichtman

Grassroots Resistance and the Data-Center Question

The accounting of AI governance shifts from the Pentagon and corporate boardrooms to local communities. Lichtman highlights rising activism around data-center expansions, including NDA deals with cities and street-level protests, as evidence that public scrutiny and democratic processes are beginning to constrain the AI supply chain. Howe emphasizes that activists are connecting data-center activism to broader concerns such as deployment in the military, educational impacts, and copyright issues. The segment contends that public resistance could slow or redirect the pace of AI development toward greater safety and accountability, underscoring a potential countervailing force against empire-building in AI.

"In recent polls, 80% of Americans now believe that there needs to be some form of regulation on the AI industry." - Flora Lichtman

Looking Ahead: Regulation, Accountability, and the Future of AI

The final portion looks at what observers should watch for next. Lichtman sees public resistance as a hopeful sign that citizens will demand governance structures that curb reckless deployment, protect privacy, and ensure fair use of AI. The discussion calls for translating lessons from grassroots data-center activism into broader AI policy, including safeguards against mass copyright infringements, safer military applications, and stronger transparency. While the path to comprehensive regulation remains contested, the podcast argues that a coalition of the public, policymakers, and researchers can push for a more accountable, safer AI ecosystem that complements innovation rather than sacrificing human oversight.

"There is a broad coalition building to hold this industry accountable" - Flora Lichtman

Related posts

featured
Guardian Science Weekly
·03/03/2026

Does Trump want to wage an AI-powered war?

featured
Nature video
·14/01/2026

What the future holds for AI – from the people shaping it

featured
New Scientist
·18/02/2026

AI Isn't as Powerful as We Think | Hannah Fry

featured
Scientific American
·12/12/2025

Karen Hao on Why AI Is Reshaping Society