PDF Summary:Brief Answers to the Big Questions, by Stephen Hawking
Book Summary: Learn the key points in minutes.
Below is a preview of the Shortform book summary of Brief Answers to the Big Questions by Stephen Hawking. Read the full comprehensive summary at Shortform.
1-Page PDF Summary of Brief Answers to the Big Questions
Theoretical physicist Stephen Hawking is best known for his expertise on black holes, quantum gravity, and general relativity, but his writing often addresses philosophical questions as well. In Brief Answers to the Big Questions, Hawking explains why he believes science can answer philosophical questions better than religion can. He then discusses where the human race came from and where he thinks we’re going. In particular, he highlights how colonizing outer space is important for the future of humankind.
In this guide, we’ll discuss Hawking’s perspective, explore how others cross-examine his perspective on religion, and compare his predictions about the future of humanity to Yuval Noah Harari’s predictions in Homo Deus.
(continued)...
(Shortform note: Hawking is not the only one to contemplate the challenges that the evolution of humans and human civilization may pose for the future. In Homo Deus, Yuval Noah Harari makes conjectures about future evolution that are parallel to Hawking’s ideas, but from a different perspective. As we examine Hawking’s predictions, we’ll compare them to Harari’s.)
Evolution Through Information
First, Hawking explains that evolution is characterized by information transfer: As each generation evolves and passes on its genes, the information contained in the genetic code of the species grows. But with the invention of written language, humans have developed other ways of passing useful information from one generation to the next. And now, thanks to digital information processing, the amount of information that humans can access and store is increasing exponentially. Hawking sees this as a new mechanism of evolution.
(Shortform note: While Hawking doesn’t say it explicitly, this paints a bleak picture of the future for people who aren’t literate, or who lack the technical competence to access digital information. Consider what happens when you combine his argument that digital information transfer has become a mechanism of human evolution with the Darwinian principle of survival of the fittest. Together, they imply that humans who are more adept at taking advantage of available information in this digital age will be more “fit” to prosper, while others will tend to get weeded out by natural selection.)
Evolution Through Genetic Engineering
Hawking explains that one particularly important piece of information mankind has been able to decipher and pass on is our own genetic code. He predicts that, now that the human genome has been sequenced and increasingly powerful gene-splicing technologies are being developed, you will soon be able to rewrite the genetic code of your offspring. Genetic engineering will lead to much faster genetic advancement of our species than Darwinian evolution.
Hawking anticipates that first, genetic engineering will allow us to cure diseases and syndromes caused by deficiencies in single genes. Then, more cures for more genetically complex conditions will be developed. After that, we’ll be able to upgrade our children’s genetics for increased intelligence, natural immunities, and eventually capabilities we haven’t even imagined yet.
Hawking acknowledges that there are many valid ethical concerns about genetically engineering humans, and that many countries consequently restrict or prohibit it. However, he argues that civil laws cannot prevent this new form of evolution. He expects that, even if it’s illegal in most countries, sooner or later someone will do it. And once the first genetically engineered superhumans appear, the rest of humankind will be faced with a choice: Embrace genetic engineering, or become an increasingly inferior sub-species of humanity. Either way, genetically enhanced humans would become the dominant species by natural selection.
But even before that time comes, we as a society need to be scientifically literate so that we can make intelligent decisions as we are increasingly confronted with issues pertaining to genetic engineering.
Genetic Engineering and Techno-Humanism
Hawking surmises that, as genetic engineering becomes increasingly feasible, people will begin to take advantage of it whether it’s sanctioned by governments or not. Yuval Noah Harari echoes this idea in Homo Deus, in which he describes the rise of “techno-humanism,” or the belief that humans should use technology to upgrade their own mental capabilities.
Based on Harari’s exposition of techno-humanism, it seems likely that Hawking is right that people will improve themselves and their offspring with technology regardless of their governments’ positions. It also seems likely that techno-humanists will be the first to genetically reengineer themselves once the technology becomes available. Genetic engineering is one of the primary mechanisms whereby they hope to upgrade their minds. Other possible mechanisms include nanotechnology, integration of computer chips into the brain, and brain-altering drugs.
Hawking also says that genetic engineering will raise new ethical and technical questions for society to answer. Harari’s discussion of techno-humanism illustrates the type of questions Hawking is referring to.
To begin with, if people start to upgrade their brains, what parts or capabilities of the brain should they focus on, and what effect will it have on society? For example, if we improve people’s ability to make decisions quickly and calmly, will we lessen their ability to empathize and be patient with others? If a military was made up of these super-rational soldiers, would it end up an efficient but unfeeling force?
Evolution Through Artificial Intelligence
According to Hawking, in addition to accelerating the evolution of our own species, we may create an entirely new form of life: electronic life. By Hawking’s definition, “life” requires only two capabilities: reproduction and metabolism.
(Shortform note: Hawking defines “life” more broadly than most biology textbooks do. One common definition requires something to meet five criteria in order to be considered a living thing: 1) It must be made of cells. 2) It must have DNA. 3) It must maintain homeostasis (regulate its internal chemistry). 4) It must be able to reproduce. 5) It must be able to collect energy that it can use for growth and movement (metabolism). Hawking dismisses the first three of these criteria and keeps only the last two. Thus, by the classical textbook definition, viruses are not alive, since they aren’t made of cells. But by Hawking’s definition, viruses are alive.)
Hawking points out that just because some technology is “alive” doesn’t mean it’s self-aware. That said, he also suggests that self-aware computers and robots may not be relegated to science fiction forever. At some point in the evolution of humans, the human brain developed enough information-processing capability to become self-aware. So, as we continue to develop increasingly sophisticated computers and artificial intelligence (AI) algorithms, it is possible that they may become conscious at some point.
Moreover, according to Hawking, once AI algorithms become better at writing algorithms than human programmers are, they will begin to evolve at an exponentially increasing rate. From this point on, AI will be increasingly beyond our control, whether they become self-aware or not. So we need to make sure that any artificial intelligence we create will work for our good and not our harm.
In particular, Hawking warns that the societal consequences of developing autonomous weapon systems could be disastrous. But not all issues involving AI are this clear-cut. Artificial intelligence algorithms could provide powerful tools for solving important technical problems, such as finding cures to disease, or modeling and controlling climate change.
Once again this makes scientific literacy increasingly important for our society, as society is increasingly confronted with questions about how to manage the development of artificial intelligence, which are both ethical and technical in nature.
Is Weaponized Artificial Intelligence Inevitable?
Hawking advises against developing autonomous war machines or otherwise weaponizing artificial intelligence. However, in light of his discussion of genetic engineering, development of intelligent weapons may be unavoidable.
As we have discussed, he predicts that laws against human genetic engineering will not prevent it from happening, because someone will eventually try it anyway. Then the genetically engineered superhumans will have an advantage over regular humans that will force the rest of humanity to either embrace genetic engineering or lose their influence.
A similar argument could be advanced for intelligent weapons. If artificial intelligence has the potential to make weapon systems dramatically more effective, then the first military power to develop these systems will have the advantage. They’ll get to make the rules, at least until their rivals develop comparable technology.
So can we really avoid development of intelligent weapon systems? As much as it might be in everyone’s best interests in the long term, the opportunity these systems provide for an individual nation or faction to gain power in the short term makes it likely that someone will do it anyway. And then other nations will either have to accept their military supremacy or match their technology.
Artificial Intelligence and Dataism
Just as Harari’s discussion of techno-humanism sheds additional light on Hawking’s discussion of genetic engineering, Harari’s exposition of Dataism provides additional perspective on Hawking’s discussion of the social implications of AI.
Harari explains that Dataism is a type of “techno-religion”: a belief system based around technology. Those who believe in Dataism hold that the intrinsic value of any living thing or system is directly proportional to its data processing capacity. Unlike humanists, Dataists don’t attribute any special significance to self-awareness or other, similar qualities that set humans apart from machines and lower animals.
Dataists believe that humans became the dominant species on Earth simply because humans can collect and process more data than animals can. Inventions like written language increased humanity’s importance because they improved humans’ ability to transmit and process data. By the same token, large communities are more important than small ones because more people can process more data. And free societies are more likely to flourish than dictatorships because they facilitate the free flow of information and allow processing to be distributed among a larger number of people.
Dataists also believe that humanity’s tenure as the dominant species on Earth is rapidly coming to an end. Computers are becoming progressively more powerful and humans are struggling to process all the data that is now available. Hawking predicts that AI will eventually become better at writing AI algorithms than human programmers, and then artificial intelligence will take off exponentially. Dataists make the same prediction and believe that will be the turning point. Once AI takes off, it will become the dominant life form on earth, as its data processing capabilities rapidly exceed humans’ capabilities. After that, Dataists believe humans will merely serve the all-powerful AI until the AI finds a way to assimilate any remaining humans into itself.
Dataism resonates with Hawking’s conjectures that self-awareness is probably just a matter of being able to process enough data, and that artificial intelligence may evolve beyond human control, whether it becomes self-aware or not. However, where Hawking raises these points as possibilities that society needs to consider and plan for, Dataists accept them as doctrine. While Hawking warns that we need to make sure AI serves our best interests, Dataists anticipate that we’ll soon be obliged to serve a global AI—and they think that’s a good thing.
Whether or not Dataists’ predictions come true, their techno-religious beliefs illustrate how technology now influences philosophical thinking, and how technological issues like AI are becoming social issues that affect everyone.
The Future of Human Expansion
According to Hawking, not only do we need to cultivate scientific literacy so our society can address the complex questions raised by new types of evolution, but we also need new generations of scientists and engineers to solve the technical problems of colonizing outer space.
Why We Should Colonize Space
Hawking presents two reasons to colonize outer space:
First, our species is running out of space on earth. Earth’s natural resources are limited, and we’re depleting them at an alarming rate. Hawking is all in favor of conserving natural resources by improving how we manage them, but he believes that in the long term, we’re still going to need more room for humanity to expand into.
Second, colonizing space mitigates the risk of extinction. Hawking warns that there is a significant probability that humans could go extinct on earth in the next millennium. But if we’ve colonized other planets by then, at least that wouldn’t be the end of our species.
Hawking mentions several threats that could lead to extinction of humans on earth. Asteroid impacts have caused extinctions in the past, and Earth will eventually be hit by another large asteroid again.
(Shortform note: Statistically speaking, large asteroids impact the earth about every 50 to 60 million years. It’s been about 66 million years since the last impact, so we’re arguably due for another asteroid impact any day now.)
As Hawking points out, if an asteroid doesn’t get us, a global nuclear war could have much the same effect. For that matter, so could climate change. Hawking expresses concern that human activity may already have triggered runaway climate change: Rising carbon dioxide levels in the atmosphere cause global warming. Global warming melts the polar ice caps. The ground underneath the ice caps absorbs sunlight better than the ice did, accelerating global warming. Furthermore, global warming could lead to deforestation of some areas, reducing the planet’s capacity to sequester carbon dioxide and thus increasing the buildup of carbon dioxide, which accelerates global warming.
Connecting Resource Depletion to Risk of Extinction
Harari offers a slightly different perspective on the dangers of ecological collapse that reveals interconnectivity between the risks that Hawking discusses.
Like Hawking, Harari points out that we as a species are using up the earth’s natural resources at an alarming rate. Unlike Hawking, he conjectures that mankind will likely find a solution to resource depletion. Throughout history, as humans have used up their natural resources, they have discovered or developed new resources.
For example, the Iron Age began when cultures that lacked the copper deposits to make bronze-age tools figured out how to refine iron (iron ore is much more common on Earth than copper). Similarly, we now build airplanes out of materials like aluminum and titanium that were unknown just a few hundred years ago, and we can now generate electricity from sunlight or nuclear fission instead of fossil fuels.
That said, Harari also points out that new technology tends to benefit the rich more than the poor because the poor can’t afford to adopt it. If we reach a point where, for example, the poor increasingly suffer from lack of clean air and water, while the rich have the technology to purify their air and water, this could lead to class warfare.
The same principle would apply to climate change: If Earth’s climate becomes inhospitable to human life, the people who can afford to do so will likely construct climate-controlled dwellings and working spaces. If others are suffering from nearly unlivable environmental conditions, this would likewise contribute to class hostility.
As such, depleting Earth’s resources or incurring climate change not only pose a direct risk to the survival of mankind, but they also increase the risk of extinction through war. Thus, based on Harari’s discussion, Hawking’s risks are not separate risks, but rather facets of the same basic problem.
Harari discusses the problem without proposing a definitive solution, while Hawking asserts that colonizing outer space is the clear solution. Yet, this, too, could fit with Harari’s model: Perhaps colonizing space is how mankind will develop new resources in the next millennium.
How We Can Colonize Space
But is colonizing space even possible? And if so, how do we go about it? Hawking thinks it is possible, and he suggests that we start by colonizing the moon, followed by Mars. He argues that the cost of a program to build bases—and ultimately colonies—there would amount to only a few percent of the world GDP.
Furthermore, Hawking expects that the experience that we gain from building colonies on the moon and Mars, combined with future advances in propulsion technology, would eventually enable us to expand beyond our solar system. He concedes that the laws of physics make it impossible for a spacecraft to travel faster than light, but he points out that there are about a thousand stars within 30 light years of earth. Theoretically, these stars will eventually be reachable by spacecraft. If even a small percentage of these stars have habitable planets, that will give mankind a lot of new real estate to settle.
Climbing Out of the Gravity Well
Although Hawking doesn’t mention it explicitly, his strategy for colonizing space appears to be based on solving the “gravity well problem,” which he certainly would have been familiar with as a physicist.
Basically, the “gravity well problem” is that it takes a tremendous amount of energy (usually in the form of a very large rocket) to launch anything into space from Earth’s surface—just like it would take a lot of energy to climb out of a deep well. This becomes particularly problematic if you want to move a large amount of material into space, for example to colonize another planet.
However, the moon’s gravity is weak enough that you can escape it much more easily. If you had a colony on the moon, it would be much easier (and cheaper) for the lunar colonists to launch a mission to Mars or somewhere else than it would be for people on Earth. Furthermore, the moon’s crust is rich in light metals like titanium and magnesium, which could be useful in constructing spaceships.
This is probably why establishing a colony on the moon is the critical first step of Hawking’s strategy. Once colonized, the moon would serve as an important base of operations from which to launch other space missions.
Are There Other Civilizations in Space?
As we start colonizing space, will we have any competition? Are there indigenous peoples on other planets or extraterrestrial races already spreading their empires across the galaxy?
Hawking says he likes the idea of extraterrestrial life, but he doesn’t think we have any direct evidence one way or the other for its existence. He dismisses reported UFO sightings and alien abductions as hoaxes, hallucinations, or misunderstandings—investigators have shown that some of the alleged sightings are indeed hallucinations, and he thinks that is the most reasonable explanation for all of them. Hawking expects that if mankind ever does make contact with technologically advanced extraterrestrials, it will be widely publicized and obvious to everyone.
(Shortform note: Hawking doesn’t elaborate on his reasons for believing that if extraterrestrial spacecraft arrived on Earth, everyone would know, but one reason they probably would know is that the United States maintains a space surveillance network. This network of radar stations, satellites, and other instruments is capable of detecting any object 10 centimeters across or larger entering or leaving low earth orbit. If Earth were ever visited by alien spacecraft, this network would provide official documentation of the craft’s arrival, but no such arrivals have been reported. This also corroborates the idea that reported UFO sightings are most likely to be hallucinations.)
Hawking discusses three possible reasons that we haven’t been contacted by extraterrestrials:
1) Maybe there are intelligent extraterrestrials out there, but we just haven’t made contact with them. The distance between stars and galaxies in our universe is so vast that it would be easy for multiple races to exist in different parts of the universe without ever bumping into each other.
(Shortform note: If there are intelligent extraterrestrials out there, how would we detect them? Even the nearest stars are much too far away for our telescopes to make out details like orbital space stations, cities on their planets, or other evidence of intelligent life. Our best bet might be picking up their radio communications signals. However, distance reduces radio signal strength, limiting the distance at which we can detect signals. Radio telescopes can pick up radio waves from distant stars, but even the dimmest dwarf stars are quadrillions of times more powerful than the most powerful radio transmitters, so it’s unlikely we’d be able to detect the signals any alien life forms might transmit.)
2) Maybe extraterrestrial life is common in the universe, but intelligent life is not. Hawking questions whether intelligence is really beneficial from the standpoint of Darwinian evolution. Maybe species that evolve sufficient intelligence to develop advanced weapons technology tend to bring about their own extinction through war. If intelligence tends to be an evolutionary dead end, there may not be very many intelligent races in the universe at any given time.
(Shortform note: If intelligent life is rare, this would greatly compound the distance problem, because, as we said before, our best shot at detecting extraterrestrial life is picking up their radio signals. But if the extraterrestrials aren’t intelligent enough to build radio transmitters, then we won’t be able to detect them that way.)
3) Maybe the probability of life forming is actually so low that it only happened once in the entire history of the universe—here on Earth. The idea that life is highly improbable seems far-fetched based on our observations of life on Earth, but observers only evolve in places where life actually forms, so our observations may not be representative. As an illustration, a child growing up in a wealthy neighborhood may assume that every kid gets a new car for her sixteenth birthday, just because everyone she knows did. Perhaps we would be similarly naive to assume that life forms on every planet, just because it formed on ours.
Estimating the Probability of Extraterrestrial Life
Various scientists have put forth estimates of the likelihood of extraterrestrial life. Probably the best-known is the Drake equation, which attempts to estimate the number of intelligent races in our galaxy based on the rate of star formation, the fraction of stars that form suitable planets, and the probability of life forming on a habitable planet. Typical assessments based on the Drake equation predict that there should be at least a thousand intelligent races in the galaxy.
However, scientists concede that some of the parameters in the Drake equation, such as the probability of life forming, are just assumptions. Some have also pointed out that there are a large number of physical and astronomical factors that can affect the habitability of a planet, which the Drake equation doesn’t include.
When these factors are included, the expected number of races typically drops to much less than 1, even if the probability of life forming on a habitable planet is assumed to be 100%, and the scope of the equation is broadened from just our galaxy to the entire observable universe.
Admittedly, there is some uncertainty inherent in all estimates of the probability of extraterrestrial life, but if more exhaustive models tend to drive the probability down, that tends to support the third possibility that Hawking presents, namely that we really are alone in the universe.
Want to learn the rest of Brief Answers to the Big Questions in 21 minutes?
Unlock the full book summary of Brief Answers to the Big Questions by signing up for Shortform.
Shortform summaries help you learn 10x faster by:
- Being 100% comprehensive: you learn the most important points in the book
- Cutting out the fluff: you don't spend your time wondering what the author's point is.
- Interactive exercises: apply the book's ideas to your own life with our educators' guidance.
Here's a preview of the rest of Shortform's Brief Answers to the Big Questions PDF summary:
What Our Readers Say
This is the best summary of Brief Answers to the Big Questions I've ever read. I learned all the main points in just 20 minutes.
Learn more about our summaries →Why are Shortform Summaries the Best?
We're the most efficient way to learn the most useful ideas from a book.
Cuts Out the Fluff
Ever feel a book rambles on, giving anecdotes that aren't useful? Often get frustrated by an author who doesn't get to the point?
We cut out the fluff, keeping only the most useful examples and ideas. We also re-organize books for clarity, putting the most important principles first, so you can learn faster.
Always Comprehensive
Other summaries give you just a highlight of some of the ideas in a book. We find these too vague to be satisfying.
At Shortform, we want to cover every point worth knowing in the book. Learn nuances, key examples, and critical details on how to apply the ideas.
3 Different Levels of Detail
You want different levels of detail at different times. That's why every book is summarized in three lengths:
1) Paragraph to get the gist
2) 1-page summary, to get the main takeaways
3) Full comprehensive summary and analysis, containing every useful point and example