Abu Dhabi, UAEThursday 23 January 2020

YouTube's attempts to face down critics have brought even more scrutiny

Video streaming platform has faced challenges over privacy, copyright and harassment, and its market dominance

YouTube, which faced intense criticism for its content cpolicies, says it is working with regulator on new laws. AFP 
YouTube, which faced intense criticism for its content cpolicies, says it is working with regulator on new laws. AFP 

YouTube spent 2019 answering critics with some of the most drastic changes in its 15-year history. With each step, it gave those activists, regulators and lawmakers more reasons to attack its free-wheeling, user-generated business model.

Susan Wojcicki, YouTube’s chief executive, announced her goals in April. “My top priority,” she wrote, “is responsibility.” Her company spent the year trying to traverse an almost impossible tightrope: nurture a growing community of demanding creators, while pledging to police troubling videos and protect millions of underage users who officially shouldn’t even be watching. The efforts pleased almost no one and highlighted an existential quandary. Every time YouTube tries to fix something, the company, an arm of Alphabet’s Google, risks losing the neutrality that it needs to thrive.

“They know that every time they are successful catching problematic content or removing it, this just raises expectations," said Mike Godwin, a senior fellow at think tank R Street Institute and a trustee of the Internet Society. “It’s a never-ending cycle of increasing demands for these dominant platforms to operate fairly.”

As 2020 begins, the largest online video service is being dragged deeper into political fights over privacy, copyright and content moderation. In response, YouTube is trying to preserve the sanctity of its status as an online platform with little liability for what happens on its site. Instead, that burden is increasingly falling on the shoulders of regulators, video creators and other partners.

Nowhere is that more evident than YouTube’s approach to kids. A landmark privacy settlement this year with the Federal Trade Commission (FTC) is forcing YouTube to split its massive site in two. Every clip, starting in January, must be designated as “made for kids” or not. The overhaul puts billions of ad dollars at stake and has sparked panic among creators, who also now face new legal risk. The company isn’t offering creators legal advice or ways to salvage their businesses. It isn’t even defining what a “made for kids” video is on YouTube — and has argued to the government that it shouldn’t have to.

“Creators will make those decisions themselves,” Ms Wojcicki said last week. “Creators know their content best.”

YouTube privately considered taking more control. Earlier this year, it assembled a team of more than 40 employees to brace for the FTC decision. The team was code-named Crosswalk — as in a way to guide kids across YouTube’s chaotic streets. Among its proposals was a radical one, at least by the standards of Silicon Valley: YouTube would screen every video aimed at kids under the age of 8 in its YouTube Kids app, ensuring that no untoward content crept into the feed of millions of tots around the world. A press release was even drafted in which Wojcicki said professional moderators would check each clip, according to people familiar with the plans. Yet at the last minute, the chief executive and her top deputies ditched the plan, said the people, who asked not to be identified.

The rationale was clear to some at YouTube, one person involved in the project recalled. Hand-picking videos, even for kids, made YouTube look too much like a media company, not a neutral platform. A YouTube spokeswoman denied the idea was turned down because it put the company in charge of programming, but she declined to comment further on the decision. In a recent interview, Ms Wojcicki made it clear that her content-moderation push only goes so far, telling CBS News that even being liable for video recommendations would destroy the essence of the service.

“If we were held liable for every single piece of content that we recommended, we would have to review it,” she said. “That would mean there would be a much smaller set of information that people would be finding. Much, much smaller.”

YouTube’s balancing act between media publisher or hands-off internet bulletin board has sparked intense debate internally. For some business partners and employees, this year’s decisions leaves them with the impression that the company is unable to take a serious stand.

“What is the mission of this company? People don’t even know,” said Claire Stapleton, a former YouTube marketing manager who left this year after clashing with Google over employee protests. “YouTube is so ill-equipped to manage these massive challenges.”

The YouTube spokeswoman said the company has made significant investments to better protect its online community. Over the last 18 months, the results of this effort include an 80 per cent reduction in views of videos that violate its policies. YouTube also increased viewership on videos from “authoritative news publishers” by 60 per cent, according to the spokeswoman. “While there will always be healthy debate around this work, we’ll continue to make the hard decisions needed to better protect the openness of the YouTube platform and the community that depends on it,” she added in a statement.

No episode in 2019 typified YouTube’s arduous search for middle ground more than the Maza affair. In June, journalist and YouTube creator Carlos Maza accused Steven Crowder, a conservative YouTuber, of repeated harassment. The Vox reporter put together a montage of clips from Mr Crowder’s YouTube channel to highlight what Mr Maza said were homophobic and racist insults.

After saying it would review Maza’s complaints, YouTube concluded the comments were not in violation of its policies, angering some of its own employees. YouTube staff held a private call to explain its rationale to Mr Maza, who remained unconvinced. “It was very awkward,” he recalled.

Crowder, meanwhile, devoted a 21-minute video to rehashing his comments. After days of criticism, YouTube removed ads from his videos, angering him.

YouTube spent the months after the Maza episode rewriting its harassment policy. The update, announced earlier this month, set new rules that would now treat Mr Crowder’s videos as violations subject to removal. Like clockwork, the decision riled other creators. Felix Kjellberg, YouTube’s biggest star, who posts as PewDiePie, declared he was leaving the video site and blamed the new policy. “We have this anarchy system, okay,” he said.

While criticism comes from all sides, YouTube’s challenge is practically insurmountable: More than 500 hours of footage are uploaded every minute. And the company’s software is still unable to gain a thorough understanding of the content before people start watching.

Ms Wojcicki’s task is set to become even more difficult. The European Parliament has approved rules that make YouTube liable the moment anyone uploads a video that violates a copyright. That could force YouTube to take down content from popular creators, while increasing its legal bills and hurting ad sales. Ms Wojcicki used Google’s political muscle and invited creators to lobby against the regulation, but she has failed to stop it.

Even in the US, the walls are closing in around YouTube. Republican and Democratic lawmakers have proposed peeling back protections that have shielded internet companies from liability for decades. YouTube’s dominance may draw antitrust scrutiny. Lawmakers are also considering tougher copyright laws, egged on by YouTube’s rivals in media and music.

For now, though, YouTube’s biggest challenge is kids’ privacy. In September, the FTC fined Google for illegally tracking children for its ads business, forcing significant changes to YouTube’s operations. On November 13, YouTube sent an email to tens of thousands of creators about the coming “made for kids” designation. If marked as “made for kids," videos will lose lucrative personalised ads and other valuable features, including user comments. If clips aren’t labeled this way, and the government decides the footage is indeed reaching children, creators can be fined thousands of dollars.

Google estimates the changes will mean YouTube creators “who make mostly child-directed content will likely lose a majority of their revenue.”

In contrast, YouTube itself emerged relatively unscathed. Google paid a $170 million fine, a tiny sliver of its profit. The FTC settlement on the Children's Online Privacy Protection Act, or COPPA, focused on YouTube, not other parts of Google, which worked hard to limit any broader impact on the rest of its businesses, according to one former executive.

The FTC is now rewriting its COPPA rules and has invited public comment. In a filing, Google told the agency it was worried about any laws forcing it to “identify and police” videos aimed at kids. The company was, in effect, arguing it couldn’t know for sure the age of its audience and shouldn’t be punished for that.

Critics were appalled. Lindsey Barrett, a staff attorney at Georgetown Law’s Communications & Technology Clinic who worked with complainants in the FTC case, found it hard to imagine the contortions required for Google to make this argument. “Our entire business is based on being able to slice and dice our audience, and see who’s watching what,” she said. “But we couldn't possibly tell you if there's a child here!”

The YouTube spokeswoman said the company has done its best to comply with its COPPA obligations, as it understands them, and has asked the FTC for more clarification on the rules.

The company is “not answering the questions everyone wants,” said Greg Alkalay, chief executive officer of BatteryPOP, a children's media company. “YouTube’s success comes from its creators. They built a beast and don't know how to wrangle it.”

Updated: January 2, 2020 04:23 AM

SHARE

SHARE