Placeholder Image

Subtitles section Play video

  • Back when online video was a different shape,

  • searching for videos on YouTube relied on

  • well, honesty, really.

  • Creators would put up a video,

  • they would tell YouTube what was in it,

  • and YouTube would look at the title and the description and the tags

  • and take it from there.

  • And immediately, people tried to cheat.

  • It turned into an arms race.

  • Keyword stuffing, tag spam, the reply girls of 2012.

  • If there was any way to guess the YouTube algorithm's priorities,

  • then spammers exploited it.

  • And desperate video creators changed what they were making

  • to fit what they thought YouTube wanted,

  • based on rumour, speculation, and extrapolating way too much

  • from way too little data.

  • So recently, YouTube has just kept quiet.

  • Their parent company, Google, learned this lesson back when

  • online video was a RealPlayer window

  • and connecting to the internet sounded like the screeching of a robot cat:

  • if Google ever gave a hint about how to boost web sites up their search results,

  • then there would be a rush of spammers trying to exploit that knowledge.

  • So the advice was always:

  • Just make good things.

  • We'll figure it out.

  • But now, on YouTube, it's not that the folks who control the algorithm

  • won't tell people how it works.

  • It's that they can't.

  • And here's the evidence: a paper written by YouTube engineers,

  • explaining they're using Google's research into machine-learning to recommend videos.

  • That is the same kind of software that creates those weird Deep Dream images,

  • that makes their text to speech sound so realistic,

  • and that beat the world's best Go player at his own game.

  • And I know I'm simplifying a bit here,

  • but machine learning, the way Google does it,

  • is basically a black box.

  • You give a neural network some input,

  • like the game board in Go.

  • And it gives you outputs: moves it thinks might work.

  • Those outputs are tested,

  • and the results go back into the box,

  • and then you repeat that process a billion or so times,

  • and it starts to get really good.

  • But no-one can look inside that black box and see how it works:

  • it's designed by a computer, for a computer.

  • And neural networks are great if you're playing a game

  • with an obvious scoring and points system.

  • You win, or you lose.

  • But training that black box on YouTube videos is a bit messier.

  • It's not just that human behaviour is unpredictable and complicated:

  • it's trying to work out what counts as winning in the first place.

  • If YouTube tells the algorithm "show videos that people like",

  • then it'll kill any channel which talks about politics,

  • where people hit dislike if they disagree.

  • And it'll silence anyone who has a small but vocal group

  • loudly disagreeing with them.

  • If YouTube tells the algorithm, "okay, show videos that people share",

  • then videos about private things like medical issues or sex education vanish,

  • and folks who have a small, loyal, but quiet fanbase

  • disappear into their own little world.

  • And YouTube creators, of course,

  • would love the algorithm to recommend only their own videos...

  • even when the rest of the world doesn't actually want to watch them.

  • So YouTube started out, according to the paper,

  • giving their algorithm the reasonable goal of

  • "increase watch time".

  • But that has a few problems.

  • Because there's no way for a computer to determine quality, or truth.

  • At least, not yet.

  • The system doesn't understand context,

  • it can't tell the difference between actual, reliable information

  • and unhinged, paranoid conspiracy clickbait.

  • Although, admittedly, neither can a lot of people,

  • which is why these videos are getting a lot of traction.

  • And it can't tell the difference

  • between videos that are suitable for children,

  • made with education in mind,

  • and creepy, trademark-infringing unofficial efforts.

  • It just knows what kids click on and what they watch.

  • So, sure, the algorithm might increase watch time, in the short term

  • but a lot of the videos it recommends are going to be questionable at best

  • and actively harmful at worst.

  • And they're going to be the sort of thing

  • that advertisers get really nervous about.

  • Remember, the algorithm is a black box.

  • No-one knows what it's doing.

  • All YouTube can do is change the feedback it's getting,

  • change the signals that saythis is goodorthis is bad”.

  • If YouTube wanted a human to watch and categorise every video being uploaded

  • assafeorunsafein real time,

  • they would need about 100,000 employees working shifts round the clock.

  • Plus, that would expose them to legal issues:

  • in most of the countries where YouTube has an office,

  • if you let an algorithm do the filtering

  • and then manually step in when you get a complaint,

  • you're legally fine.

  • But if you approve everything with a human in the loop,

  • you are a publisher,

  • and you're opening yourself up to some very expensive lawsuits.

  • The ideal algorithm, the ideal black box,

  • from YouTube's point of view,

  • would be one with a goal of "increase ad revenue",

  • and which thought about the long-term,

  • which knew about social issues,

  • and potential advertiser boycotts, and financial strategies,

  • and public perception, and what's suitable for kids,

  • and... and about truth.

  • At which point, what you have is something

  • that can do the job of YouTube's senior management team.

  • And artificial intelligence hasn't gotten that good.

  • At least, not yet.

Back when online video was a different shape,

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it