The question “Are humans going to go extinct soon?” is getting popular by the day now. It is due to the fact that everything around us keeps changing too soon, which ultimately leads us to believe maybe the doomsday is near. All the things we hear and see from childhood let us believe that whenever things start to change too fast, it’s not good. This theory is true for the most part. So, let’s delve into the question and understand it more deeply.
Are Humans Going To Extinct Soon?
Will human extinction be anthropogenic? That is the result of human action. Or will it be one of the good old-fashioned kinds of extinction Earth’s history knows pretty well?
The Global Catastrophic Risks Survey, issued by Oxford University’s Future of Humanity Institute, placed our risk of extinction before the year 2100 at 19%. Now, you might be thinking, “Whatever, blah blah blah.” It’ll be okay; humans are too smart to go extinct. Maybe you’re right. But it’s difficult to predict the distant future with a lot of certainty. What’s really cool, though, is that if you embrace that uncertainty, a simple argument can show that human extinction soon is actually more probable. It’s called the Doomsday argument.
Doomsday argument
Imagine a giant urn (a vase, typically with a cover, a rounded body, a narrowed neck, and a pedestal) that contains either 10 balls numbered 1 to 10 or a million balls numbered 1 to a million. Now, you don’t know which is the case, but you are allowed to pull out one ball. You go ahead and do that, and it is ball number 4. That’s pretty strong evidence in favor of the 10-ball condition because drawing a four from a set of 1 through 10 is a one in 10 chance. But drawing four from a million different numbers is a one-in-a-million chance. By analogy, you are also a numbered ball. You are a human who knows approximately what your birth number is. It’s probably somewhere around 100 billion. That’s how many other humans were most likely born before you were. Importantly, you didn’t get to decide which birth number you would have. So, just like the number for a ball, you are a random sample from the set of all humans who will ever live. The Doomsday argument points out that from 200 billion people, there’s a 50 percent chance that a randomly chosen person, like you, would be born in the first one hundred billion. Whereas if there will be 10 trillion humans, there’s only a one percent chance that any given human, say you, would happen to be born within the first 100 billion. Either you are special and lucky to be born so improbably early in the story of humanity or your birth number is to be expected because there will not be tens of trillions of humans. Human extinction will be sooner rather than later. But before you become too convinced that the end is nigh, keep in mind that the Doomsday argument is not uncontroversial.
One problem it might have is a reference class problem. Are you really a random sample from the set of all humans who will ever be born? Well, if you believe that in the not-so-distant future humans will be quite different than they are today. For instance, there’ll be full of more 3D printed organs. The mere fact that right now there aren’t very many humans with that trait could be evidence that you aren’t a random sample from the set of all humans, just from the set of all humans like you, like those around you. Those born earlier in human history. Also, the Doomsday argument doesn’t consider the likelihoods of actual threats or human advantages over those threats in the future. It just assumes that we don’t know which way the balance will lie; that human extinction soon and human extinction later are equally likely. But maybe you don’t believe that. Maybe you are convinced that human ingenuity will always stay one step ahead of any extinction event thrown at it. You could be right, but there’s reason to doubt that optimism. For example, the Fermi paradox.
The Fermi paradox
If it is likely that intelligent life forms in our universe are capable of living for billions and billions of years, where are they? Why are the skies so silent? Perhaps it is because extinction-level threat events are just too common for intelligent life anywhere to ever catch up. So, does this mean we should just give up? The Voluntary Human Extinction Movement thinks so. Founded in 1991, its supporters believe that humans are a negative influence on Earth and always will be. Thus, we have a moral obligation to just stop reproducing right now and fade away. But what would a computer do? To learn that an AI was created. The program came up with novel techniques and strategies for playing games and even exploited glitches humans didn’t know about, or at least hadn’t told it about. They also had the program play other games, like Tetris, which I think is relevant to our question. The computer struggled to figure out what to do. As the computer wasn’t programmed to consider future repercussions far enough ahead to notice that stacking Tetriminos in certain ways made a big difference. On one run, when faced with imminent demise, the computer did something eerie. Rather than lose and receive a ‘game over’, it just paused the game. Forever. This was the description of the computer’s reasoning like this: “The only winning move is to not play.” And that’s right. If you pause a game forever, you will never lose that game. But you’ll also never win that game or achieve a high score. Now, we might not know what achieving a sentient life high score in this universe means or whether or not we’re capable of achieving one. We might also sometimes panic when the future looks bleak. But if we keep playing and keep learning, chances are we could eventually figure it out and start playing really well. So, thanks for continuing to play and for being here.