‘The graves are only half empty; who will help us fill them?” Twenty years ago, that rallying cry on Rwandan radio helped explode ethnic enmity into one of history’s worst atrocities. In today’s Iraq, another vicious conflict between a formerly empowered ethnic minority and a long-subjugated majority is causing the deaths of thousands. At its heart is another mass-media appeal to bloodlust on radio’s modern-day equivalent: social media. This time, the world may have a chance to stop what it failed to in Rwanda.

The Sunni Islamic State insurgents, locked in a deadly struggle with Iraq’s Shiite majority, excel online. They command a plethora of official and unofficial channels on Facebook, Twitter and YouTube. “And kill them wherever you find them,” commands one recent propaganda reel of firefights and bound hostages, contorting a passage from the Qur’an. “Take up arms, take up arms, O soldiers of the Islamic State. And fight, fight!” adds another, featuring a sermon from the group’s leader, Abu Bakr al-Baghdadi. The material is often slickly produced, like “The Clanging of Swords IV,” a glossy, feature-length film replete with slow-motion action scenes. Much of it is available in English, directly targeting the recruits with Western passports who have become one of the organization’s more dangerous assets. And almost all of it appeals to the young: manipulated images of Islamic State fighters and their grisly massacres with video-game-savvy captions like, “This is our Call of Duty.”

But officials at social media companies are leery of adjudicating what should be taken down and what should be left alone. “One person’s terrorist is another person’s freedom fighter,” one senior executive told me on condition of anonymity. Making that call is “not something we’d want to do.”

So official Islamic State accounts often remain on Twitter for weeks and accumulate tens of thousands of followers before being removed. A few propaganda videos have been taken off YouTube for “violating YouTube’s policy on shocking and disgusting content,” according to the notice on one removed video. But countless others remain, including a recent sermon by Baghdadi, posted through an account claiming to be Islamic State-affiliated.

There are legitimate free-speech questions here: How should media coverage of propaganda be treated? What about peaceful lectures by otherwise violent terrorists? But those gray areas don’t excuse a lack of enforcement against direct calls for murder, which these companies supposedly ban. “I understand there are freedom-of-speech concerns, but I don’t think that describes what’s going on with much of the content on YouTube,” says Evan Kohlmann, a counterterrorism analyst with Flashpoint Partners and NBC News. “No one’s suggesting they remove all journalistic clips. … This is about extremely explicit content, calling for violence.”

Another objection is practical. There’s simply too much content to monitor and too many ways for it to resurface when quashed. An executive at one major social media company described it as the “Whac-A-Mole” phenomenon — take down one video, it springs up elsewhere. But flawed enforcement shouldn’t excuse inaction now any more than it did in Rwanda 20 years ago, when the U.S. government deemed the use of radio-jamming technology too legally complex, too expensive, too impractical. The perfect, then as now, was the enemy of the good.

More troubling still is the fact that these companies already know how to police and remove content that violates the law. Every major social media network employs algorithms that automatically detect and prevent the posting of child pornography. Many, including YouTube, use a similar technique to prevent copyrighted material from hitting the Web. Why not, in overt cases such as beheading videos and calls for blood, employ a similar system?

Indeed, Twitter, YouTube and Facebook all strictly refuse to police content themselves — instead relying on third parties, mostly users around the world, to flag objectionable material. But the constant torrent of new content is not a burden that can be practically managed by the crowd — any more than companies expect users to serve as the prime monitor for child pornography.

As always, beneath legitimate practical and ethical concerns, there is a question about the bottom line. Section 230 of the Telecommunications Act of 1996 inoculates companies from responsibility for content that users post — as long as they don’t know about it. Individuals involved in content removal policies at the major social media companies, speaking to me on condition of anonymity, said that’s a driving factor in their thinking. “We can’t police any content ourselves,” one explained. Added another: “The second we get into reviewing any content ourselves, record labels say, ‘You should be reviewing all videos for copyright violations, too.’ ”

Yet past is prologue. The world, with each lamentation of “never again,” has cursed its failure to stop Rwanda’s deadly broadcasts. A furious Samantha Power, the U.S. ambassador to the United Nations, once complained that the United States “refused to use its technology” when it could have. A Harvard University study found that jamming the broadcasts could have saved tens of thousands of lives.

The Islamic State’s campaign of incitement is “definitely reminiscent of Rwanda,” said John Prendergast, a former Clinton administration official focused on Africa who has studied Rwanda for years. Now as then, exploiting sectarian hatred can quickly turn deadly on a massive scale. Now as then, cracking down on calls to kill is no panacea, but it can help.

These companies have a moral obligation to do more. And U.S. law should not create a legal barrier for them to act when lives are on the line. The current regime — enforced ignorance and half-measures — may be among our apologies when we say “never again” over Iraq.


Ronan Farrow is the anchor of “Ronan Farrow Daily” on MSNBC. This article was initially published on PostEverything, at www.washingtonpost.com/posteverything.