Placeholder Image

Subtitles section Play video

  • Hey, Vsauce. Michael here. Do you want to be infected with Ebola

  • without having to leave your own home or deal with other people?

  • Well, you might be in luck. You can already download an Ebola virus

  • genome.

  • Right here on the Internet, right now. And if you're willing to wait

  • a few years for 3D bioprinting technology to progress

  • a little bit, you can just acquire one then, submit the genome to it

  • and ta da! All you can print Ebola.

  • Or anthrax or whatever it is you wish to mass-produce at home

  • to wipe out humanity.

  • Are humans going to go extinct

  • soon? Will human extinction be

  • anthropogenic? That is the result of human

  • action. Or will it be one of the good old-fashioned kinds

  • of extinction Earth's history knows pretty well?

  • The Global Catastrophic Risks Survey, issued by Oxford University's

  • Future of Humanity Institute placed our risk of extinction

  • before the year 2100 at 19%.

  • Now, you might be thinking "whatever, blah blah blah armageddon".

  • "It'll be okay, humans are too smart

  • to go extinct." Maybe you're right.

  • But it's difficult to predict the distant future

  • with a lot of certainty. What's really cool though

  • is that if you embrace that uncertainty, a simple argument

  • can show that human extinction soon is actually

  • more probable. It's called the Doomsday

  • argument. Imagine a giant

  • urn that contains either 10 balls

  • numbered 1 to 10, or a million balls

  • numbered 1 to a million. Now, you don't know

  • which is the case, but you are allowed to pull out

  • one ball. You go ahead and do that

  • and it is ball number 4.

  • That's pretty strong evidence in favour of the 10 ball condition

  • because drawing a four from a set of 1 through 10

  • is a one in 10 chance. But drawing four from a million different numbers

  • is a one in a million chance.

  • By analogy you are also a numbered

  • ball. You are a human who knows

  • approximately what your birth number is.

  • It's probably somewhere around 100

  • billion. That's how many other humans

  • were most likely born before you were.

  • Importantly, you didn't get to decide which birth number

  • you would have. So, just like the number for a ball,

  • you are a random sample from the set of all humans

  • who will ever live. The Doomsday argument points out

  • that from 200 billion people there's a 50 percent chance that a randomly chosen

  • person,

  • like you, would be born in the first one hundred billion.

  • Whereas if there will be 10 trillion humans,

  • there's only a one percent chance that any given human,

  • say you, would happen to be born within the first

  • 100 billion. Either you are special

  • and lucky to be born so improbably early in the story of humanity

  • or your birth number is to be expected

  • because there will not be tens of trillions of humans.

  • Human extinction will be sooner

  • rather than later. But before you become

  • too convinced that the end is nigh, keep in mind that the Doomsday argument is

  • not

  • uncontroversial. One problem it might have

  • is a reference class problem. Are you really a

  • random sample from the set of all humans who will ever

  • be born? Well, if you believe that in the not so distant future

  • humans will be quite different than they are today.

  • For instance, there'll be full of more 3D printed

  • organs. The mere fact that right now there aren't very many humans

  • with that trait could be evidence that you aren't a random sample from the

  • set of all humans,

  • just from the set of all humans like

  • you, like does around you. Those born

  • earlier in human history. Also

  • the Doomsday argument doesn't consider the likelihoods

  • of actual threats or human advantages

  • over those threats in the future. It just assumes that

  • we don't know which way the balance will lie; that

  • human extinction soon and human extinction

  • later are equally likely. But maybe you don't believe that.

  • Maybe you are convinced that human ingenuity will

  • always stay one step ahead of any extinction event

  • thrown at it. You could be right,

  • but there's reason to doubt that optimism.

  • For example, the Fermi paradox.

  • If it is likely that intelligent life forms in our universe are capable

  • of living for billions

  • and billions of years, where are they?

  • Why are the skies so silent? Perhaps

  • it is because extinction level threat events are just

  • too common for intelligent life anywhere

  • to ever catch up.

  • So,

  • does this mean we should just give up?

  • The Voluntary Human Extinction Movement think so.

  • Founded in 1991, its supporters believe that

  • humans are a negative influence

  • on Earth and always will be. Thus

  • we have a moral obligation to just stop reproducing

  • right now and fade away. But what would

  • a computer do? In a way, that's

  • kind what Tom 7 did. He created a program that plays video games.

  • The program came up with novel techniques and strategies for playing

  • games and even exploited glitches

  • humans didn't know about, or at least

  • hadn't told it about. He also had the program play other games,

  • like Tetris, which I think is relevant

  • to our question. The computer struggled to figure out what to do.

  • You see, the computer wasn't programmed to consider future repercussions far

  • enough ahead

  • to notice that stacking Tetriminos in certain ways

  • made a big difference. On one run, when faced with

  • imminent demise, the computer did something

  • eerie. Rather than

  • lose, and receive a 'game over',

  • it just paused the game. For

  • ever. Tom 7 describes the computer's reasoning

  • like this: "The only winning move

  • is to not play." And that's right.

  • If you pause a game for ever

  • you will never lose that game. But you'll also never

  • win that game or achieve a high score.

  • Now, we might not know what achieving a

  • sentient life high score in this universe

  • means or whether or not we're capable of achieving one.

  • We might also sometimes panic

  • when the future looks bleak. But if we keep playing

  • and keep learning, chances are we could eventually

  • figure it out and start playing

  • really well.

  • So, thanks for continuing to play, for being here.

  • And as always,

  • thanks for watching.

Hey, Vsauce. Michael here. Do you want to be infected with Ebola

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it