DiscoverThe Tech Policy Press Podcast
The Tech Policy Press Podcast
Claim Ownership

The Tech Policy Press Podcast

Author: Tech Policy Press

Subscribed: 111Played: 4,946
Share

Description

Tech Policy Press is a nonprofit media and community venture intended to provoke new ideas, debate and discussion at the intersection of technology and democracy.

You can find us at https://techpolicy.press/, where you can join the newsletter.
409 Episodes
Reverse
In their new book, Move Slow and Upgrade: The Power of Incremental Innovation, Evan Selinger, a professor in the Department of Philosophy at Rochester Institute of Technology and Albert Fox Cahn, founder in residence of the Surveillance Technology Oversight Project (STOP), argue that society is over-fixated on disruptive innovation over the kind of steady incrementalism that can deliver sustainable returns over longer time frames. They argue in favor of more careful deliberation and adopting what they call the “upgrader’s mindset,” which should be applied whenever “disruptive changes would pose the greatest social risk.”
The Pentagon wants AI that can fight wars — without limits. One of the United States’ leading AI companies says there are lines it won't cross. And this week, that standoff turned into an all-out confrontation. To discuss the implications of the dispute between Anthropic and the Pentagon, including the determination that the company represents a supply chain risk, Justin Hendrix spoke to two experts:Kat Duffy, senior fellow for digital and cyberspace policy at the Council on Foreign Relations, andAmos Toh, senior counsel in the Liberty and National Security Program at the Brennan Center for Justice.
Concerns about synthetic media and coordinated manipulation of online platforms have moved from theoretical worry to documented reality. Researchers, regulators, and civil society organizations are working to understand how algorithmically driven content recommendation systems can be exploited — not just by ideologically motivated actors, but by ordinary users pursuing financial gain.Fundación Maldita.es is a Spanish nonprofit that has been working on information integrity and fact-checking since 2017. Its most recent investigation focuses on TikTok, and what they found raises pointed questions about the platform's creator monetization program. Researchers at Maldita documented a network of hundreds of accounts — spanning eighteen countries — that were producing AI-generated videos of protests that never happened, and doing so not out of any discernible political motive, but to accumulate followers, qualify for TikTok's revenue-sharing program, and, in some cases, sell the accounts outright. In this episode, Justin Hendrix is joined by Maldita associate director for public policy Carlos Hernández-Echevarría and public policy officer Marina Sacristán.
As AI technologies proliferate, a growing number of people are asking what it means to live in a world dominated by algorithms and automated systems—and what gets lost when those systems optimize human behavior at scale. These questions sit at the intersection of political theory, technology policy, and everyday life, and they are drawing scholars from fields well outside computer science into the conversation.José Marichal is a political scientist at California Lutheran University who has been writing and teaching about technology and politics for more than two decades. Marichal's new book, You Must Become an Algorithmic Problem: Renegotiating the Socio-Technical Contract, considers the age of recommendation systems and large language models. Drawing on political philosophy, he argues that individuals have entered into an implicit bargain with technology companies, trading unpredictability and novelty for the convenience of algorithmically curated experience. The consequences of that bargain, he contends, reach beyond personal preference and into the foundations of liberal democratic citizenship.
This week marks the second DSA and Platform Regulation conference in Amsterdam, where experts will convene to consider the Digital Services Act (DSA) two years after it entered full effect across the European Union. Over that period, the law has been tested by national elections, geopolitical tensions, high-profile enforcement actions, and the rapid rise of generative AI. It has become both a benchmark for platform accountability and a political lightning rod.Ahead of the conference, Tech Policy Press senior editor Ramsha Jahangir spoke with members of the DSA Observatory, which is organizing the conference, to take stock. What have these first years of enforcement clarified? Where does opacity remain? And what does it mean to conduct DSA research in today’s political climate? Guests include:John Albert, associate researcher, DSA Observatory.Paddy Leerssen, postdoctoral researcher at the University of Amsterdam and part of the DSA Observatory.Magdelena Jozwiak, associate researcher at the DSA Observatory.
A wave of lawsuits in the Unites States is targeting tech firms for their product design decisions. Lawyer Carrie Goldberg has played a role in establishing the product liability theory that underlies them. As the founder of C.A. Goldberg, PLLC, in 2017, her firm brought a lawsuit that sought to apply product liability theory to a tech platform — Herrick v. Grindr — arguing that a dangerous app design, not just user behavior, was the source of harm. In 2022, Goldberg was appointed to the Plaintiffs’ Steering Committee in the federal social media multidistrict litigation. She’s led cases against Amazon, Meta, and Omegle, has testified before the Senate Judiciary Committee on child safety issues, and is the author of Nobody's Victim: Fighting Psychos, Stalkers, Pervs, and Trolls. Justin Hendrix spoke to her from her offices in Brooklyn about what she's learned over the last decade, and about some ongoing litigation that remains in dispute.
"Operation Metro Surge" — the massive immigration enforcement operation playing out right now in Minnesota — was billed as a targeted effort to apprehend undocumented immigrants. But what it has exposed goes far beyond immigration enforcement. It has pulled back the curtain on a sprawling surveillance apparatus that incorporates artificial intelligence, facial recognition, and other novel tools — not just to enable the raids that have turned violent and, in some cases, deadly; but also to silence dissent, to intimidate entire communities, and to discourage people from even watching what masked federal agents are doing in their own neighborhoods.To discuss these events and the prospects for reform, Justin Hendrix spoke to Irna Landrum, a senior campaigner at Kairos Fellowship and author of a recent piece on Tech Policy Press, "How ICE Uses AI to Automate Authoritarianism," and Alejandra Montoya-Boyer, vice president for the Center for Civil Rights and Technology at the Leadership Conference on Civil and Human Rights, which has called for reforms at the Department of Homeland Security and its component agencies.
In his forthcoming book, Your Data Will Be Used Against You, George Washington University Law School professor Andrew Guthrie Ferguson explores how the rise of sensor-driven technologies, social media monitoring, and artificial intelligence can be weaponized against democratic values and personal freedoms. Smart cars, smart homes, smart watches—these devices track our most private activities, and that data can be accessed by police and prosecutors looking for incriminating clues. What should legislatures, courts, and individuals do to protect civil liberties?
The killing of 37-year old nurse Alex Pretti by federal agents in Minneapolis was filmed from multiple angles by residents of the city, and local government officials have implored the public to share evidence of immigration enforcement agents committing acts of violence with investigators. But what are the challenges of using such artifacts in the pursuit of accountability? And what is there to learn from other efforts to use video, including from social media platforms, as evidence when seeking justice for crimes by state actors? Inequality.org managing editor and Tech Policy Press fellow Chris Mills Rodrigo joins Justin Hendrix to discuss these questions and more.
Today's guest is Jennifer Lind,  an associate professor of government at Dartmouth, a fellow at Chatham House London, and the author of the new book Autocracy 2.0: How China’s Rise Reinvented Tyranny, just out from Cornell Press. The book introduces the concept of 'smart authoritarianism,' a strategy that seeks to preserve political dominance while minimizing the economic damage of repression. It’s a sharp and unsettling argument—and one that is worth considering as a wave of autocratization continues to sweep across the globe, increasingly enabled by new technologies.
In a forthcoming paper, George Washington University Law School scholar Spencer Overton argues that the Trump administration's AI policy is consistent with its broader efforts to advance ethnonationalism. By eliminating policies intended to ensure safeguards against algorithmic bias—and recasting work on such problems as ideological threats to innovation—Trump's policies embed exclusion into the technological infrastructure of the future. As a growing body of research suggests, when AI systems operate without regulation, they default to dominant patterns that reproduce racial inequality and suppress cultural pluralism.
A new book titled Governing Digital China offers crucial insights into China's governance ecosystem. Written by Daniela Stockmann, a professor at the Hertie School in Berlin and director of the Center for Digital Governance, and Ting Luo, an associate professor in artificial intelligence and government at the University of Birmingham, the book reveals a more complex reality than simple top-down control.The authors show how massive tech companies like Tencent and Alibaba have become essential partners to the Chinese state, blending corporate and government power. At the same time, citizens exercise bottom-up influence, shaping how both platforms and the state respond to their needs. The result is what the authors call "popular corporatism"—a form of digital authoritarianism that operates quite differently than you might expect.
2026 is poised to be another landmark year for the child online safety debate in the United States.In recent years, states have passed dozens of bills aimed at expanding protections for kids as they navigate risks on social media platforms, AI chatbots and other pools, with more likely on the way. Lawmakers in Washington, meanwhile, are considering a flurry of proposals that could set a national standard on the issue. But many of these efforts are facing legal limbo as industry and some digital rights groups allege they violate constitutional rights and trample on privacy.Tech Policy Press senior editor Cristiano Lima-Strong spoke to three experts tracking the issue to assess the current policy landscape in the United States and how it may shift in 2026, particularly as state legislators continue to take up the cause:Amina Fazlullah is head of tech policy advocacy at Common Sense Media, a group that advocates for child online safety measures. She previously served as a tech policy fellow for Mozilla and as director of policy at the Benton Foundation.Joel Thayer is president of the Digital Progress Institute, a think tank that advocates for age verification policies. He previously clerked for Federal Trade Commission official Maureen Ohlhausen and served as policy counsel for the tech trade group The App Association.Kate Ruane is the director of the Free Expression Project at the Center for Democracy and Technology, a nonprofit that advocates for digital rights. She previously served as lead public policy specialist for the Wikimedia Foundation and as senior legislative counsel for the ACLU.
In what Reuters called a "mass digital undressing spree," Elon Musk is provoking outrage after his Grok chatbot responded to user prompts to remove the clothing from images of women and pose them in bikinis and to create "sexualized images of children" and post them on X. To discuss the controversy and the broader policy implications of generative AI with regard to child sexual abuse material and nonconsensual intimate imagery, Justin Hendrix spoke to Riana Pfefferkorn, a policy fellow at the Stanford Institute for Human-Centered AI and author of numerous reports and articles on these subjects, including for Tech Policy Press.
Tech Policy Press fellow Anika Collier Navaroli joined Justin Hendrix to discuss insights from her special 2025 series of podcasts, Through to Thriving. They discussed insights from her interviews over the course of the year with Ellen Pao, Jerrel Peterson, Alice Hunsberger, Vaishnavi J, Desmond Patton, Nora Benavidez, Mimi Ọnụọha, Timnit Gebru, Jasmine McNealy, Naomi Nix, and Chris Gilliard.
 On Thursday, US President Donald Trump invited reporters into the Oval Office to watch him sign an executive order intended to limit state regulation of artificial intelligence. Trump said AI is a strategic priority for the United States, and that there must be a central source of approval for the companies that develop it.  Today's guest is Olivier Sylvain, a professor of law at Fordham Law School and a senior policy research fellow at the Knight First Amendment Institute at Columbia University.  He's the author of "Why Trump’s AI EO Will be DOA in Court," a perspective published on Tech Policy Press.
On Friday, the European Commission fined Elon Musk’s X €120 million for breaching the Digital Services Act, delivering the first-ever non-compliance decision under the European Union’s flagship tech regulation. By Saturday, Elon Musk was calling for no less than the abolition of the EU. To discuss the enforcement action, the politics surrounding it, and a variety of other issues related to digital regulation in Europe, Justin Hendrix spoke to Joris van Hoboken, a professor at the Institute for Information Law (IViR) at the University of Amsterdam, and part of the core team of the Digital Services Act (DSA) Observatory.
On this podcast, for years we’ve discussed issues such as conspiracy theories, mis- and disinformation, polarization, and the ways in which the design and incentives on today’s technology platforms exacerbate them. Today’s guest is Calum Lister Matheson,  associate professor and chair of the Department of Communication at the University of Pittsburgh and a faculty member of the Pittsburgh Psychoanalytic Center. He's the author of Post-Weird: Fragmentation, Community, and the Decline of the Mainstream, a new book from Rutgers University Press that applies a different lens on the question as he searches for insights into the seemingly inexplicable behaviors of communities such as serpent handlers, pro-anorexia groups, believers in pseudoscience, and conspiracy theorists that deny the reality of gun violence in schools.
The past few years have seen a great deal of introspection about a professional field which has come to be known as 'trust and safety,' comprised of the people who develop, oversee, and enforce social media policies and community guidelines. Many scholars and advocates describe it as having reached a turning point, mostly for the worst. Joining Tech Policy Press contributing editor Dean Jackson to discuss the evolution of trust and safety—not coincidentally, the title of their forthcoming article In the Emory Law Journal—are professors of law Danielle Keats Citron and Ari Ezra Waldman. Also joining the conversation is Jeff Allen, the chief research officer at the Integrity Institute, a nonprofit whose membership is composed of trust and safety industry professionals.
This week, the European Commission unveiled a sweeping plan to overhaul how the EU enforces its digital and privacy rules as part of a ‘Digital Omnibus,’ aiming to ease compliance burdens and speed up implementation of the bloc’s landmark laws. Branded as a “simplification” initiative, the omnibus proposal touches core areas of EU tech regulation — notably the AI Act and the General Data Protection Regulation (GDPR).The Commission argues that this update is necessary to ensure practical implementation of the laws, but civil society organizations see the proposed reform as the “biggest rollback of digital fundamental rights in EU history.”At the same time, leaders are talking loudly about digital sovereignty — including at last week’s summit in Berlin. But with the Omnibus appearing to weaken protections and tilt power toward large tech firms, what kind of sovereignty is actually being built?Tech Policy Press associate editor Ramsha Jahangir spoke to two experts to understand what the EU is trying to achieve:Leevi Saari, EU Policy Fellow at AI Now InstituteJulia Smakman, Senior Researcher at the Ada Lovelace Institute
loading
Comments (5)

C muir

do you invite any guests with vaguely right leaning views? or is this just a echo chamber

Mar 23rd
Reply

C muir

far right? oh Lord a silly woke pod bye

Mar 23rd
Reply

C muir

the problem is when moderation is pushed across all sites. it's censorship

Mar 23rd
Reply

C muir

when you start censoring you have lost the argument

Mar 23rd
Reply

C muir

whatever happened to the left? moderation/censorship

Mar 23rd
Reply