
Sign up to save your podcasts
Or
We are in the midst of the first major U.S. election of the generative AI era. The people who want to win your vote have easy access to tools that can create images, video or audio of real people doing or saying things they never did — and slap on weird appendages or other make-believe effects along with targeted slogans. But the potential to deceive has led about two dozen states to enact some form of regulation requiring political ads that use artificial intelligence to include a label. So how do voters respond when they know a campaign has used AI? That’s what Scott Brennen and his team at New York University’s Center on Technology Policy set out to answer in a recent study.
4.5
12361,236 ratings
We are in the midst of the first major U.S. election of the generative AI era. The people who want to win your vote have easy access to tools that can create images, video or audio of real people doing or saying things they never did — and slap on weird appendages or other make-believe effects along with targeted slogans. But the potential to deceive has led about two dozen states to enact some form of regulation requiring political ads that use artificial intelligence to include a label. So how do voters respond when they know a campaign has used AI? That’s what Scott Brennen and his team at New York University’s Center on Technology Policy set out to answer in a recent study.
1,634 Listeners
900 Listeners
4,333 Listeners
1,716 Listeners
8,637 Listeners
30,668 Listeners
1,367 Listeners
32,091 Listeners
2,171 Listeners
5,493 Listeners
1,451 Listeners
9,500 Listeners
3,589 Listeners
5,945 Listeners
6,217 Listeners
163 Listeners
2,775 Listeners
157 Listeners
1,343 Listeners
90 Listeners