
Sign up to save your podcasts
Or


The recent riots in the United Kingdom raise new questions about online free speech and misinformation. Following the murder of three children in Southport, England, false rumors spread across social media about the killer’s identity and religion, igniting simmering resentment over the British government’s handling of immigration in recent years. X, formerly Twitter, has come under fire for allowing the rumors to spread, and the company’s owner Elon Musk has publicly sparred with British politicians and European Union regulators over the issue.
The incident is the latest in an ongoing debate abroad and in the U.S. about free speech and the real-world impact of online misinformation. In the U.S., politicians have griped for years about the content policies of major platforms like YouTube and Facebook—generally with conservatives complaining the companies are too censorious and liberals bemoaning that they don’t take down enough misinformation and hate speech.
Where should the line be? Is it possible for platforms to respect free expression while removing “harmful content” and misinformation? Who gets to decide what is true and false, and what role, if any, should the government play? Evan is joined by Renee Diresta who studies and writes about adversarial abuse online. Previously, she was a research manager at the Stanford Internet Observatory where she researched and investigated online political speech and foreign influence campaigns. She is the author of Invisible Rulers: The People Who Turn Lies into Reality. Read her recent op-ed in the New York Times here.
By Foundation for American Innovation4.8
1111 ratings
The recent riots in the United Kingdom raise new questions about online free speech and misinformation. Following the murder of three children in Southport, England, false rumors spread across social media about the killer’s identity and religion, igniting simmering resentment over the British government’s handling of immigration in recent years. X, formerly Twitter, has come under fire for allowing the rumors to spread, and the company’s owner Elon Musk has publicly sparred with British politicians and European Union regulators over the issue.
The incident is the latest in an ongoing debate abroad and in the U.S. about free speech and the real-world impact of online misinformation. In the U.S., politicians have griped for years about the content policies of major platforms like YouTube and Facebook—generally with conservatives complaining the companies are too censorious and liberals bemoaning that they don’t take down enough misinformation and hate speech.
Where should the line be? Is it possible for platforms to respect free expression while removing “harmful content” and misinformation? Who gets to decide what is true and false, and what role, if any, should the government play? Evan is joined by Renee Diresta who studies and writes about adversarial abuse online. Previously, she was a research manager at the Stanford Internet Observatory where she researched and investigated online political speech and foreign influence campaigns. She is the author of Invisible Rulers: The People Who Turn Lies into Reality. Read her recent op-ed in the New York Times here.

2,452 Listeners

1,942 Listeners

288 Listeners

7,213 Listeners

93 Listeners

8,039 Listeners

2,435 Listeners

3,905 Listeners

389 Listeners

514 Listeners

64 Listeners

15,918 Listeners

399 Listeners

237 Listeners

35 Listeners