Twitter announced today a sweeping set of changes designed to add “extra friction” to the platform to check the spread of political disinformation. This includes requiring users to add their own comments before retweeting others, labeling premature claims of election victories, removing calls for violence in response to the election, and restricting the reach of tweets containing misinformation from political figures with over 100,000 followers.
We applaud these changes, and believe that if Twitter is serious about its stated goal of “protecting the integrity of the election conversation“ there’s another thing the platform should consider: putting a time delay on the tweets of Donald Trump and other political elites.
Mike Ananny is Associate Professor of Communication and Journalism at the University of Southern California’s Annenberg School for Communication and Journalism. Daniel Kreiss is a principal researcher of the UNC Center for Information, Technology, and Public Life and the Cato Distinguished Associate Professor in the Hussman School of Journalism and Media.
Each new broadcast technology has had to find its relationship to liveness. In 1952, when the Federal Communications Commission (FCC) prohibited broadcasting live telephone conversations—but allowed broadcasting taped conversations—radio station call-in shows used a short delay to get around the prohibition, recording conversations to tape and then, six or seven seconds later, playing the tape. The (not always perfect) solution also gave broadcasters a measure of control over live situations, letting them bleep or mute profanity and inappropriate content or anything else they wanted to keep from their audiences. The “bleep censor” quickly became an industry standard.
Why not put Donald Trump’s tweets, and his Facebook posts, as well as those of other political elites, on a time delay? Twitter and Facebook have extensive and well-documented content rules that prohibit everything from electoral to health disinformation. The platforms have singled out these categories of content in particular because they have significant likelihood of causing real world harm, from voter suppression to undermining the Centers for Disease Control and Prevention’s public health guidelines. The FBI found that the plot to kidnap Michigan Governor Gretchen Whitmer was, in part, organized in a Facebook group.
To date, the enforcement of these policies has been spotty at best. Twitter has labeled some of the president’s tweets as “potentially misleading” readers about mail-in ballots. The platform hid a Trump tweet stating “when the looting starts, the shooting starts” for “glorifying violence,” and it recently hid another tweet equating Covid-19 to the flu, claiming that the president was “spreading misleading and potentially harmful information” when he wrote that “we are learning to live with Covid, in most populations far less lethal!!!” Facebook has taken similar actions, providing links to reliable voter and health information and removing posts that it deems violates its policies.
But these actions often take hours to put in place while this content racks up thousands of engagements and shares. In those hours, as recent research from Harvard shows, Trump is a one-man source of disinformation that travels quickly and broadly across Twitter and Facebook. And we know that the mainstream media often picks up on and amplifies Trump’s posts before platforms moderate them. Journalists report on platforms’ treatments of Trump’s tweets, making that and them the story, and giving life to false claims.
What if we never let Trump’s disinformation breathe to begin with, cutting it off from the social media and mainstream journalism oxygen it craves?
We suggest Twitter and Facebook immediately institute the following process for all of Trump’s social media posts, and for those of other political elites. Any time the President taps “Tweet” or “Post” his content is not displayed immediately but sent to a small 24/7 team of elite content moderators who can evaluate whether the content accords with these platforms’ well-established policies. This is especially important in the context of electoral and health disinformation, which all the major platforms have singled out as being of utmost importance. Within, say, three minutes, such a team would decide whether to (a) let the post through, (b) let the post through with restrictions, (c) place a public notice on Trump’s account saying that the platform is evaluating a post and needs more time, or (d) block the post entirely because it breaks the company’s policies. The platforms would publicly announce that such a system was in place, they would provide weekly metrics on how many posts the review system had considered and categorized, they would allow those impacted to appeal any decisions, and they would revisit these systems after an experimental period to evaluate their effectiveness.