• Lab
  • AndroidForMobile Foundation at
    Pico wants to inject CRM smarts into news sites hungry for reader relationships
    ABOUT                    SUBSCRIBE
    April 15, 2019, 3:13 p.m.

    As Notre Dame burned, an algorithmic error at YouTube put information about 9/11 under news videos

    A reminder that even efforts to limit misinformation can end up spreading it instead — and that human editors watching over the algorithms can be a pretty good thing, too.

    It’s terrible news for anyone who values history, loves Paris, or read Victor Hugo: Notre-Dame Cathedral, the Gothic gem at the historic center of Paris, . It’s obviously far too early for anything conclusive, but early suggestions from officials are that the blaze could be related the ongoing renovations to the roof. There’s no indication at this writing that it’s a terror attack or related in any way to a terrorist group.

    So as people turned to YouTube to see live streams from trusted news organizations of the fire in progress, why was YouTube showing them background information about 9/11?

    I first noticed this when I went to France24, which produces an English-language feed, and .

    Looking around, along with , I also saw the 9/11 info on the streams of and .

    Why is YouTube adding information to these videos that seems tailor-made to make people think it’s a terror attack? I asked Google and got this statement from a spokesperson:

    We are deeply saddened by the ongoing fire at the Notre Dame cathedral. Last year, we launched information panels with links to third party sources like Encyclopedia Britannica and Wikipedia for subjects subject to misinformation. These panels are triggered algorithmically and our systems sometimes make the wrong call. We are disabling these panels for live streams related to the fire.

    You may remember those information panels from when they were , with CEO Susan Wojcicki saying that:

    …when there are videos that are focused around something that’s a conspiracy — and we’re using a list of well-known internet conspiracies from Wikipedia — then we will show a companion unit of information from Wikipedia showing that here is information about the event.

    That well-intentioned effort faced criticism on a couple of fronts: Google’s YouTube would be freeloading on the backs of unpaid Wikipedia editors, and those info boxes (with a link to Wikipedia) risked infecting that comparatively conspiracy-resistant platform with a bunch of YouTube crazies.

    YouTube has expanded that effort in a few ways over time, including (even if they don’t click through to a video) and .

    It’s unclear why a breaking news event — one about which there hasn’t been time for any substantial conspiracies to take root — got the information panel, much less a 9/11 one; Google fixed the problem less than an hour after I noticed it. But it’s a reminder that even efforts to limit misinformation can end up spreading it instead — and that human editors watching over the algorithms can be a pretty good thing, too.

    UPDATE, 5:40 p.m.: A few quick followups since this story has now picked up at other sites. First, here’s a previous example of the 9/11 infobox being added to an unrelated video; KCRW’s noted that some old footage of New York City from 1976 got tagged with it:

    The doesn’t mention anything more 9/11-related than “New York” — no mention of the World Trade Center, for instance. (The infobox has since been removed.)

    Then : a video of his father Ken’s retirement from Michigan State, which somehow got labeled with a “Jew” infobox:

    Waltzer used to at Michigan State, but again nothing in the title or description mentions anything Jewish.

    Google’s says that it places the boxes “alongside videos on a small number of well-established historical and scientific topics that have often been subject to misinformation online, like the moon landing…This information panel will appear alongside videos related to the topic, regardless of the opinions or perspectives expressed in the videos.” I guess the algorithm it’s using considered a video about a Jewish man retiring to be sufficiently about the topic of “Jew” to merit the box, just as it considered random 40-year-old footage of New York to be “related” to 9/11.

    In other words, it isn’t just that the algorithm sometimes completely misses the boat, like confusing Notre-Dame and the World Trade Center. Even when it’s not making a big categorization error, it can still be putting up very inappropriate “information.”

    Some other examples: A got labeled with 9/11 — presumably because it showed two towers and a lot of smoke?

    9/11 also to a promising “College Music · 24/7 Live Radio · Study Music · Chill Music · Calming Music”:

    Same for a video of a :

    A couple other thoughts: that simply linking to accurate information isn’t the best way to battle a conspiracy theory.

    that while human monitoring of every topic on YouTube is obviously impractical, there’s no reason it couldn’t use humans on a first pass for this sort of stuff on the most important stories — especially the big breaking ones.

    (Etim used to lead content moderation at The New York Times, so he knows the value of giving humans oversight over a small subset of the most important information judgment calls, while letting algorithms handle the rest.)

    POSTED     April 15, 2019, 3:13 p.m.
    Show comments  
    Show tags
    Join the 50,000 who get the freshest future-of-journalism news in our daily email.
    Pico wants to inject CRM smarts into news sites hungry for reader relationships
    “They said: ‘We don’t have a CRM. We have a plugin that manages who’s paid, and we have Mailchimp to email people.’ Honestly, that moment when they said ‘send us a CSV,’ we realized this was nuts.”
    Twitter is removing precise-location tagging on tweets — a small win for privacy but a small loss for journalists and researchers
    For the past decade, location-tagged tweets have been a useful (if imperfect) tool for anyone trying to connect time, place, and information in ways that told us something about the world.
    “News unfolds like a saga”: The French news site Les Jours wants to marry narrative, depth, and investigative reporting
    “Serial” isn’t just a podcast: It’s also the format hook Les Jours uses to bring some of the lessons of drama to long-form investigative reporting. It’s a fascinating mish-mash of ideas you’ll recognize from short-run nonfiction audio, Quartz, Epic Magazine, and more.