YouTube trying to crack down on conspiracy videos
The company will include links to Wikipedia to try to debunk videos espousing conspiracy theories.
YouTube says it’s cracking down on conspiracy videos, though it’s scant on the details.
Conspiracy videos abound on YouTube, whether it’s about the Earth being flat or school shootings being staged.
YouTube, its parent company Google, Facebook and Twitter are all facing challenges with the spread of misinformation, propaganda and fake news.
YouTube CEO Susan Wojcicki said at a conference on Tuesday that the company will include links to the online encyclopedia Wikipedia to try to debunk videos espousing conspiracy theories.
But Wikipedia itself has had its share of credibility issues, as the service lets anyone edit its content, whether that person is a pedigreed expert or an online troll.
In a statement on Wednesday, YouTube said the links will include other “third-party sources” besides Wikipedia, though YouTube isn’t identifying any.
The organisation that runs Wikipedia said on Wednesday that it had no formal partnership with YouTube, but welcomed the use of Wikipedia resources.
YouTube said the move is part of a broader initiative to crack down on misinformation, though it did not give details on what else is in the works.
While conspiracy videos are nothing new on YouTube, the topic received renewed attention in recent weeks as videos falsely claimed that students speaking out about the February 14 Florida school shooting, which killed 17 people, were “crisis actors” who were not really there when it happened.
Conspiracy videos, to be sure, are not against YouTube’s policies. In the “crisis actor” case, the company said it removed the video because it violated its rules against harassment. As such, YouTube is unlikely to ban misinformation entirely.
Instead, it may adopt Facebook’s tactic of de-emphasising such content and making it less likely to be seen. As it stands, critics say YouTube is doing the opposite.
“What keeps people glued to YouTube? Its algorithm seems to have concluded that people are drawn to content that is more extreme than what they started with – or to incendiary content in general,” Zeynep Tufekci, a University of North Carolina professor who studies the social impact of technology, wrote in a recent New York Times essay.
Mr Tufekci said that as users click through video after video, excited by “uncovering more secrets and deeper truths”, YouTube leads us down “a rabbit hole of extremism, while Google racks up the ad sales”.
Sorry, we are not accepting comments on this article.