Hidden in plain sight: Safety innovation in live-streamed video

Who could’ve predicted that “content moderation” would become such a hot topic? It’s everywhere this year — from hearings on Capitol Hill to news leaks, books, podcasts and conferences. It’s clear that everybody — the companies themselves, users, and certainly policymakers — is struggling to understand how social media companies protect users and free speech at the same time. The struggle is real. Moderating social media to everybody’s satisfaction is like trying to get the entire planet to agree on the best flavor of ice cream.

But even as the extreme challenge of this is dawning on all of us, there’s some real innovation going on. It’s not what you’d normally think, and it’s happening where you might least expect it: live-streamed video, projected to be more than a $70 billion market in two years but already mainstream social media for teens and young adults, according to Pew Research (parents may’ve noticed this!).

Two distinct examples are Twitch, the 800-pound gorilla of streaming platforms based in San Francisco, and Yubo, a fast-growing startup based in Paris. Both are innovating in really interesting ways, blending both human norms and tech tools. Both…

  • Innovate around user safety because it’s “good business”
  • Have live video in their favor as they work on user safety
  • Empower their users to help.

To explain, here’s what they’re up to….

Yubo, an app that now has more than 15 million users in more than a dozen countries, reached its first million within 10 months without any marketing (it was then called Yellow). The app’s all about communication — chat via live video-streaming. Somehow, the focus of early news coverage was on one Tinder-like product feature, even though Yubo’s safety features have long included a “firewall” that keeps 13–17 year-old users separate from adults. Virtually all news coverage has been alarmist, especially in Europe, encouraging parents and child advocates to expect the worst.

So they doubled down, with a mix of high- and low-tech “tools.” One of them is surprisingly unusual, and it definitely involves an element of surprise: nearly real-time intervention when a user violates a community rule (the rules are no bullying, hate speech, underwear, nudity or violence).

When the violation happens — maybe a user joins a chat in their underwear — the user almost immediately gets a message like the one in this screenshot:

What’s so unusual about that? you might ask. But this is rare in social media. It’s one of those sudden disruptors that stops the conversation flow in real time, changing the dynamic and helping chatters learn the community standards by seeing them enforced right when things are happening. Yubo can do that because it takes a screenshot of all chat sessions every 10 seconds; so between users’ own reports of annoying behavior and algorithms running in the background that detect and escalate it, the app can respond very quickly. If the violation doesn’t stop, moderators will either disable the chat mid-stream or suspend the user’s account.

You wouldn’t think an app communicating with users in real time is that unusual, but Yubo says it’s the only social video app that intervenes with young users in this way, and I believe it. In more than 20 years of writing about kids and digital media, I haven’t seen another service that does.

The other innovative part of this is the algorithms, and the way Yubo uses them to detect problems proactively for faster response (Twitch doesn’t even have algorithms running proactively in the background, they told me). Proactive is unusual. For various reasons — including the way platforms are used — content moderation has always been about 99% reactive, dependent on users reporting problems and never even close to real-time. Bad stuff comes down later — sometimes without explanation.

Knowing how hard it is for algorithms to detect nudity and underwear, I asked Yubo COO Marc-Antoine Durand how that works on his app. “We spent nearly a year labeling every reported piece of content to create classifiers. Every time we had a report, a dedicated team had to classify the content into multiple categories (ex: underwear, nudity, drug abuse, etc.). With this, our algorithms can learn and find patterns.”

It takes a whole lot of data — a lot of images — to “teach” algorithms what they’re “looking for.” An app that’s all live-streamed video, all the time, very quickly amasses a ton of data for an algorithm to chew on. And Yubo has reached 95% accuracy in the nudity category (its highest priority) with a dedicated data scientist feeding (and tweaking) the algorithms full-time, he said. “We had to make a distinction between underwear and swimsuits,” he added. “We trained our algorithms to know the difference by detecting the presence of sand, water, a swimming pool, etc.”

Not that nudity’s a huge problem on the app — for example, only 0.1–0.2% of users upload suggestive content as profile photos any given day — Yubo says they prioritize nudity and underwear.

But those aren’t the only tools in the toolbox. Probably because live-streamed video hasn’t come up much on Capitol Hill or in news about “content moderation” (yet), we don’t hear much about how live video itself — and the social norms of the people using it — are safety factors too. But it’s not hard for anyone who uses Skype or Google hangouts to understand that socializing in video chat is almost exactly like socializing in person, where it’s extremely unusual for people to show up in their underwear, right? If someone decides to in Yubo, and others in the chat have a problem with that, the person typically gets reported because it’s just plain annoying, and moderators can suspend or delete their account. The every-10-sec. screenshots are backup so moderators can act quickly on anything not reported. On top of that, Yubo provides moderation tools to users who open chats, to the “streamer” of the “live.”

Twitch leverages social norms too. It empowers streamers, or broadcasters, with tools to maintain safety in chat during their live streams. With 44 billion minutes streamed per month by 3.1 million “streamers” (according to TwitchTracker.com) the company really has to harness the safety power of social norms in millions of communities.

“We know through statistics and user studies that, when you have a lot of toxicity or bad behavior online, you lose users,” said Twitch Associate General Counsel Shirin Keen at the CoMo Summit in Washington, D.C., last spring, one of the first conferences on the subject. “Managing your community is core to success on Twitch.” So because streamer success is Twitch’s bread and butter, the platform has to help streamers help their viewers feel safe.

Twitch’s phenomenal growth has been mostly about video gamers streaming their videogame play. But the platform’s branching out — to the vastness of IRL (for “In Real Life”), its biggest new non-gaming channel. And just like in real real life, communities develop their own social norms. The Twitch people know that. The platform lets streamers shape their own norms and rules. Streamers appoint their own content moderators and provides them with moderation tools for maintaining their own community’s standards.

So can you see how online safety innovation is as much about humans as it is about tech — what we find acceptable or feel comfortable with in the moment and in a certain social setting? Take “emotes,” for example. They’re a way users can express their emotions in otherwise dry, impersonal text chat. They too are part of channel culture and community-building at Twitch. So are “bits,” colorful little animated graphics that viewers can buy to cheer a streamer on from the chat window. And these graphical elements help moderators see—at a glance, in a fast-moving chat stream—when things are good or going south. Last summer Twitch launched “Creator Camp” for training streamers and their “mods” (channel moderators) on how to create their community and its unique culture, “all without bug spray, sunburns, and cheesy sing-alongs (ok maybe a few sing-alongs,” Twitch tells its streamers.

So two vastly different live video companies are experimenting with something we’re all beginning to see, that…

  • Live video makes social media a lot like in-person socializing
  • Social media, like social anything, is affected by social norms
  • Social norms help make socializing safer
  • People stick around and have more fun when they feel safe.

The bottom line being that safety — in the extremely participatory business of social media — is good business.

That’s the kind of innovating happening on the inside of industry. Then there’s the kind happening out here in society. If it could be boiled down to a single word, that would be “transparency.” It may never be a glass box, but the black box is slowly becoming less opaque.

While social media companies have to keep innovating around safety, they also have to innovate around how to make the challenge of, and users’ role in, content moderation more transparent. It’s good we out here are pushing for that, but we have to innovate too. As participants in the public discussion about user safety, we all need to understand a couple of things: that effective content moderation is a balance that companies have to strike between protecting users and protecting their freedom of expression, and that, if the platforms are too transparent, their systems can be gamed or hacked more easily, which makes everybody less safe.

So even as we demand free speech, more protection and greater transparency, the public discussion has to keep innovating too. Safety is good business and works from the inside out and the outside in.

Disclosure: I am an Internet safety adviser to Yubo and other social media services, including Facebook, Twitter and Snapchat, but not including Twitch. My bio is below and here.

  • Live-stream video metrics: ”Two-thirds (67%) of consumers globally have streamed live video,” according to a June 2018 report.
  • All negative, all the time not helpful: At a research symposium in Riga, Latvia, last June, psychology professor Sonia Livingstone, a leading researcher in the field of youth online risk, called the current debate about “tech addiction” and “screen time” impossibly “monolithic. We get complete exclusion of any kind of consideration of the opportunities that technologies bring.” She’s no apologist for technology, she said, but she suggested the importance of taking a “critical stance” toward the uniformly negative narrative we’re seeing in tech news. Even professionally neutral researchers are calling for more balance in the public discussion about online wellbeing.
  • A “day” in the life of a content mod: a talk at MIT by volunteer content moderator at Twitch and grad student Claudia Lo, illustrating how it’s proactive and reactive and so much more than that binary (and this shows how important transparency is for anyone talking about regulation)
  • Other tools and measures Yubo uses, a post in my blog a year ago before the app rebranded (and, more recently, Yubo’s own perspective)
  • About content moderation on Reddit: insights from Nate Mathias, PhD
  • Pioneering journalism on the subject at Wired from Adrian Chen in 2014 and from Catherine Buni and Soraya Chemaly in The Verge’s award-winning “Secret Rules of the Internet” in 2016
  • Twitch and Shout — in-depth radio program on Twitch from public radio’s On the Media
  • RadioLab’s in-depth report on content moderation at Facebook
  • “There is no magic bullet for moderating a social media platform,” Techdirt’s report on the first CoMo conference
  • Games Are Taking a Back Seat to Players on Video Game Streaming Sites

Anne Collier is founder and executive director of The Net Safety Collaborative, home of the U.S.’s social media helpline for schools. She has been blogging about youth and digital media at NetFamilyNews.org since before there were blogs.

Youth advocate; blogger, NetFamilyNews.org; founder, The Net Safety Collaborative