Discover
The Tech Policy Press Podcast
The Tech Policy Press Podcast
Author: Tech Policy Press
Subscribed: 105Played: 4,816Subscribe
Share
© Copyright 2026 Tech Policy Press
Description
Tech Policy Press is a nonprofit media and community venture intended to provoke new ideas, debate and discussion at the intersection of technology and democracy.
You can find us at https://techpolicy.press/, where you can join the newsletter.
You can find us at https://techpolicy.press/, where you can join the newsletter.
404 Episodes
Reverse
A wave of lawsuits in the Unites States is targeting tech firms for their product design decisions. Lawyer Carrie Goldberg has played a role in establishing the product liability theory that underlies them. As the founder of C.A. Goldberg, PLLC, in 2017, her firm brought a lawsuit that sought to apply product liability theory to a tech platform — Herrick v. Grindr — arguing that a dangerous app design, not just user behavior, was the source of harm. In 2022, Goldberg was appointed to the Plaintiffs’ Steering Committee in the federal social media multidistrict litigation. She’s led cases against Amazon, Meta, and Omegle, has testified before the Senate Judiciary Committee on child safety issues, and is the author of Nobody's Victim: Fighting Psychos, Stalkers, Pervs, and Trolls. Justin Hendrix spoke to her from her offices in Brooklyn about what she's learned over the last decade, and about some ongoing litigation that remains in dispute.
"Operation Metro Surge" — the massive immigration enforcement operation playing out right now in Minnesota — was billed as a targeted effort to apprehend undocumented immigrants. But what it has exposed goes far beyond immigration enforcement. It has pulled back the curtain on a sprawling surveillance apparatus that incorporates artificial intelligence, facial recognition, and other novel tools — not just to enable the raids that have turned violent and, in some cases, deadly; but also to silence dissent, to intimidate entire communities, and to discourage people from even watching what masked federal agents are doing in their own neighborhoods.To discuss these events and the prospects for reform, Justin Hendrix spoke to Irna Landrum, a senior campaigner at Kairos Fellowship and author of a recent piece on Tech Policy Press, "How ICE Uses AI to Automate Authoritarianism," and Alejandra Montoya-Boyer, vice president for the Center for Civil Rights and Technology at the Leadership Conference on Civil and Human Rights, which has called for reforms at the Department of Homeland Security and its component agencies.
In his forthcoming book, Your Data Will Be Used Against You, George Washington University Law School professor Andrew Guthrie Ferguson explores how the rise of sensor-driven technologies, social media monitoring, and artificial intelligence can be weaponized against democratic values and personal freedoms. Smart cars, smart homes, smart watches—these devices track our most private activities, and that data can be accessed by police and prosecutors looking for incriminating clues. What should legislatures, courts, and individuals do to protect civil liberties?
The killing of 37-year old nurse Alex Pretti by federal agents in Minneapolis was filmed from multiple angles by residents of the city, and local government officials have implored the public to share evidence of immigration enforcement agents committing acts of violence with investigators. But what are the challenges of using such artifacts in the pursuit of accountability? And what is there to learn from other efforts to use video, including from social media platforms, as evidence when seeking justice for crimes by state actors? Inequality.org managing editor and Tech Policy Press fellow Chris Mills Rodrigo joins Justin Hendrix to discuss these questions and more.
Today's guest is Jennifer Lind, an associate professor of government at Dartmouth, a fellow at Chatham House London, and the author of the new book Autocracy 2.0: How China’s Rise Reinvented Tyranny, just out from Cornell Press. The book introduces the concept of 'smart authoritarianism,' a strategy that seeks to preserve political dominance while minimizing the economic damage of repression. It’s a sharp and unsettling argument—and one that is worth considering as a wave of autocratization continues to sweep across the globe, increasingly enabled by new technologies.
In a forthcoming paper, George Washington University Law School scholar Spencer Overton argues that the Trump administration's AI policy is consistent with its broader efforts to advance ethnonationalism. By eliminating policies intended to ensure safeguards against algorithmic bias—and recasting work on such problems as ideological threats to innovation—Trump's policies embed exclusion into the technological infrastructure of the future. As a growing body of research suggests, when AI systems operate without regulation, they default to dominant patterns that reproduce racial inequality and suppress cultural pluralism.
A new book titled Governing Digital China offers crucial insights into China's governance ecosystem. Written by Daniela Stockmann, a professor at the Hertie School in Berlin and director of the Center for Digital Governance, and Ting Luo, an associate professor in artificial intelligence and government at the University of Birmingham, the book reveals a more complex reality than simple top-down control.The authors show how massive tech companies like Tencent and Alibaba have become essential partners to the Chinese state, blending corporate and government power. At the same time, citizens exercise bottom-up influence, shaping how both platforms and the state respond to their needs. The result is what the authors call "popular corporatism"—a form of digital authoritarianism that operates quite differently than you might expect.
2026 is poised to be another landmark year for the child online safety debate in the United States.In recent years, states have passed dozens of bills aimed at expanding protections for kids as they navigate risks on social media platforms, AI chatbots and other pools, with more likely on the way. Lawmakers in Washington, meanwhile, are considering a flurry of proposals that could set a national standard on the issue. But many of these efforts are facing legal limbo as industry and some digital rights groups allege they violate constitutional rights and trample on privacy.Tech Policy Press senior editor Cristiano Lima-Strong spoke to three experts tracking the issue to assess the current policy landscape in the United States and how it may shift in 2026, particularly as state legislators continue to take up the cause:Amina Fazlullah is head of tech policy advocacy at Common Sense Media, a group that advocates for child online safety measures. She previously served as a tech policy fellow for Mozilla and as director of policy at the Benton Foundation.Joel Thayer is president of the Digital Progress Institute, a think tank that advocates for age verification policies. He previously clerked for Federal Trade Commission official Maureen Ohlhausen and served as policy counsel for the tech trade group The App Association.Kate Ruane is the director of the Free Expression Project at the Center for Democracy and Technology, a nonprofit that advocates for digital rights. She previously served as lead public policy specialist for the Wikimedia Foundation and as senior legislative counsel for the ACLU.
In what Reuters called a "mass digital undressing spree," Elon Musk is provoking outrage after his Grok chatbot responded to user prompts to remove the clothing from images of women and pose them in bikinis and to create "sexualized images of children" and post them on X. To discuss the controversy and the broader policy implications of generative AI with regard to child sexual abuse material and nonconsensual intimate imagery, Justin Hendrix spoke to Riana Pfefferkorn, a policy fellow at the Stanford Institute for Human-Centered AI and author of numerous reports and articles on these subjects, including for Tech Policy Press.
Tech Policy Press fellow Anika Collier Navaroli joined Justin Hendrix to discuss insights from her special 2025 series of podcasts, Through to Thriving. They discussed insights from her interviews over the course of the year with Ellen Pao, Jerrel Peterson, Alice Hunsberger, Vaishnavi J, Desmond Patton, Nora Benavidez, Mimi Ọnụọha, Timnit Gebru, Jasmine McNealy, Naomi Nix, and Chris Gilliard.
On Thursday, US President Donald Trump invited reporters into the Oval Office to watch him sign an executive order intended to limit state regulation of artificial intelligence. Trump said AI is a strategic priority for the United States, and that there must be a central source of approval for the companies that develop it. Today's guest is Olivier Sylvain, a professor of law at Fordham Law School and a senior policy research fellow at the Knight First Amendment Institute at Columbia University. He's the author of "Why Trump’s AI EO Will be DOA in Court," a perspective published on Tech Policy Press.
On Friday, the European Commission fined Elon Musk’s X €120 million for breaching the Digital Services Act, delivering the first-ever non-compliance decision under the European Union’s flagship tech regulation. By Saturday, Elon Musk was calling for no less than the abolition of the EU. To discuss the enforcement action, the politics surrounding it, and a variety of other issues related to digital regulation in Europe, Justin Hendrix spoke to Joris van Hoboken, a professor at the Institute for Information Law (IViR) at the University of Amsterdam, and part of the core team of the Digital Services Act (DSA) Observatory.
On this podcast, for years we’ve discussed issues such as conspiracy theories, mis- and disinformation, polarization, and the ways in which the design and incentives on today’s technology platforms exacerbate them. Today’s guest is Calum Lister Matheson, associate professor and chair of the Department of Communication at the University of Pittsburgh and a faculty member of the Pittsburgh Psychoanalytic Center. He's the author of Post-Weird: Fragmentation, Community, and the Decline of the Mainstream, a new book from Rutgers University Press that applies a different lens on the question as he searches for insights into the seemingly inexplicable behaviors of communities such as serpent handlers, pro-anorexia groups, believers in pseudoscience, and conspiracy theorists that deny the reality of gun violence in schools.
The past few years have seen a great deal of introspection about a professional field which has come to be known as 'trust and safety,' comprised of the people who develop, oversee, and enforce social media policies and community guidelines. Many scholars and advocates describe it as having reached a turning point, mostly for the worst. Joining Tech Policy Press contributing editor Dean Jackson to discuss the evolution of trust and safety—not coincidentally, the title of their forthcoming article In the Emory Law Journal—are professors of law Danielle Keats Citron and Ari Ezra Waldman. Also joining the conversation is Jeff Allen, the chief research officer at the Integrity Institute, a nonprofit whose membership is composed of trust and safety industry professionals.
This week, the European Commission unveiled a sweeping plan to overhaul how the EU enforces its digital and privacy rules as part of a ‘Digital Omnibus,’ aiming to ease compliance burdens and speed up implementation of the bloc’s landmark laws. Branded as a “simplification” initiative, the omnibus proposal touches core areas of EU tech regulation — notably the AI Act and the General Data Protection Regulation (GDPR).The Commission argues that this update is necessary to ensure practical implementation of the laws, but civil society organizations see the proposed reform as the “biggest rollback of digital fundamental rights in EU history.”At the same time, leaders are talking loudly about digital sovereignty — including at last week’s summit in Berlin. But with the Omnibus appearing to weaken protections and tilt power toward large tech firms, what kind of sovereignty is actually being built?Tech Policy Press associate editor Ramsha Jahangir spoke to two experts to understand what the EU is trying to achieve:Leevi Saari, EU Policy Fellow at AI Now InstituteJulia Smakman, Senior Researcher at the Ada Lovelace Institute
In the latest episode in her special podcast series, Through to Thriving, Tech Policy Press fellow Anika Collier Navaroli talks about protecting privacy with Chris Gilliard. Gilliard is co-director of the Critical Internet Studies Institute and the author of Luxury Surveillance, a forthcoming book from MIT Press.
To discuss the past, present and future of information integrity work, Tech Policy Press contributing editor Dean Jackson spoke to American University Center for Security, Innovation and New Technology (CSINT) nonresident fellow Adam Fivenson and assistant professor and CSINT director Samantha Bradshaw.
This episode considers whether today’s massive AI investment boom reflects real economic fundamentals or an unsustainable bubble, and how a potential crash could reshape AI policy, public sentiment, and narratives about the future that are embraced and advanced not only by Silicon Valley billionaires, but also by politicians and governments. Justin Hendrix is joined by:Ryan Cummings, chief of staff at the Stanford Institute for Economic Policy Research and coauthor of a recent New York Times opinion on the possibility of an AI bubble;Sarah West, co-director of the AI Now Institute and coauthor of a Wall Street Journal opinion, "You May Already Be Bailing Out the AI Business"; andBrian Merchant, author of the newsletter Blood in the Machine, a journalist in residence at the AI Now Institute, and author of a recent piece in Wired on signals that suggest a bubble.
This episode was recorded in Barcelona at this year’s Mozilla Festival. One session at the festival focused on how to get better access to data for independent researchers to study technology platforms and products and their effects on society. It coincided with the launch of the Knight-Georgetown Institute’s report, “Better Access: Data for the Common Good,” the product of a year-long effort to create “a roadmap for expanding access to high-influence public platform data – the narrow slice of public platform data that has the greatest impact on civic life,” with input from individuals across the research community, civil society, and journalism. In a gazebo near the Mozilla Festival mainstage, Justin Hendrix hosted a podcast discussion with three people working on questions related to data access and advocating for independent technology research:Peter Chapman, associate director of the Knight-Georgetown Institute;Brandi Geurkink, executive director of the Coalition for Independent Tech Research and a former campaigner and fellow at Mozilla; andLK Seiling, a researcher at the Weizenbaum Institute in Berlin and coordinator of the DSA40 Data Access Collaboratory.Thanks to the Mozilla Foundation and to Francisco, the audio engineer on site at the festival.
For her special series of podcasts, Through to Thriving, Tech Policy Press fellow Anika Collier Navaroli spoke to artist Mimi Ọnụọha, whose work "questions and exposes the contradictory logics of technological progress." The discussion ranged across changing trends in nomenclature of data and artificial intelligence, the role of art in bearing witness to authoritarianism, the interventions and projects that Ọnụọha has created about the datafication of society, and why artists and policy practitioners should work more closely together to build a more just and equitable future.





do you invite any guests with vaguely right leaning views? or is this just a echo chamber
far right? oh Lord a silly woke pod bye
the problem is when moderation is pushed across all sites. it's censorship
when you start censoring you have lost the argument
whatever happened to the left? moderation/censorship