The Signal and the Noise : Why So Many Predictions Fail--But Some Don't
The Signal and the Noise : Why So Many Predictions Fail--But Some Don't
Click to enlarge
Author(s): Silver, Nate
ISBN No.: 9780143125082
Pages: 576
Year: 201502
Format: Trade Paper
Price: $ 27.60
Dispatch delay: Dispatched between 7 to 15 days
Status: Available

At about the time  The Signal and the Noise  was first published in September 2012, "Big Data" was on its way becoming a Big Idea. Google searches for the term doubled over the course of a year,1 as did mentions of it in the news media.2 Hundreds of books were published on the subject. If you picked up any business periodical in 2013, advertisements for Big Data were as ubiquitous as cigarettes in an episode of  Mad Men .   But by late 2014, there was evidence that trend had reached its apex. The frequency with which Big Data was mentioned in corporate press releases had slowed down and possibly begun to decline.3 The technology research firm Gartner even declared that Big Data had passed the peak of its "hype cycle."4   I hope that Gartner is right.


Coming to a better understanding of data and statistics is essential to help us navigate our lives. But as with most emerging technologies, the widespread benefits to science, industry, and human welfare will come only after the hype has died down.   FIGURE P-1: BIG DATA MENTIONS IN CORPORATE PRESS RELEASES   I worry that certain events in my life have contributed to the hype cycle. On November 6, 2012, the statistical model at my Web site FiveThirtyEight "called" the winner of the American presidential election correctly in all fifty states. I received a congratulatory phone call from the White House. I was hailed as "lord and god of the algorithm" by  The Daily Show ''s Jon Stewart. My name briefly received more Google search traffic than the vice president of the United States.   I enjoyed some of the attention, but I felt like an outlier--even a fluke.


Mostly I was getting credit for having pointed out the obvious--and most of the rest was luck.*   To be sure, it was reasonably clear by Election Day that President Obama was poised to win reelection. When voters went to the polls on election morning, FiveThirtyEight''s statistical model put his chances of winning the Electoral College at about 90 percent.* A 90 percent chance is not quite a sure thing: Would you board a plane if the pilot told you it had a 90 percent chance of landing successfully? But when there''s only reputation rather than life or limb on the line, it''s a good bet. Obama needed to win only a handful of the swing states where he was tied or ahead in the polls; Mitt Romney would have had to win almost all of them.   But getting every state right was a stroke of luck. In our Election Day forecast, Obama''s chance of winning Florida was just 50.3 percent--the outcome was as random as a coin flip.


Considering other states like Virginia, Ohio, Colorado, and North Carolina, our chances of going fifty-for-fifty were only about 20 percent.5 FiveThirtyEight''s "perfect" forecast was fortuitous but contributed to the perception that statisticians are soothsayers--only using computers rather than crystal balls.   This is a wrongheaded and rather dangerous idea. American presidential elections are the exception to the rule--one of the few examples of a complex system in which outcomes are usually more certain than the conventional wisdom implies. (There are a number of reasons for this, not least that the conventional wisdom is often not very wise when it comes to politics.) Far more often, as this book will explain, we overrate our ability to predict the world around us. With some regularity, events that are said to be certain fail to come to fruition--or those that are deemed impossible turn out to occur.   If all of this is so simple, why did so many pundits get the 2012 election wrong? It wasn''t just on the fringe of the blogosphere that conservatives insisted that the polls were "skewed" toward President Obama.


Thoughtful conservatives like George F. Will6 and Michael Barone7 also predicted a Romney win, sometimes by near-landslide proportions.   One part of the answer is obvious: the pundits didn''t have much incentive to make the right call. You can get invited back on television with a far worse track record than Barone''s or Will''s--provided you speak with some conviction and have a viewpoint that matches the producer''s goals.   An alternative interpretation is slightly less cynical but potentially harder to swallow: human judgment is intrinsically fallible. It''s hard for any of us (myself included) to recognize how much our relatively narrow range of experience can color our interpretation of the evidence. There''s so much information out there today that none of us can plausibly consume all of it. We''re constantly making decisions about what Web site to read, which television channel to watch, and where to focus our attention.


  Having a better understanding of statistics almost certainly helps. Over the past decade, the number of people employed as statisticians in the United States has increased by 35 percent8 even as the overall job market has stagnated. But it''s a necessary rather than sufficient part of the solution. Some of the examples of failed predictions in this book concern people with exceptional intelligence and exemplary statistical training--but whose biases still got in the way.   These problems are not so simple and so this book does not promote simple answers to them. It makes some recommendations but they are philosophical as much as technical. Once we''re getting the big stuff right--coming to a better understanding of probably and uncertainty; learning to recognize our biases; appreciating the value of diversity, incentives, and experimentation--we''ll have the luxury of worrying about the finer points of technique.   Gartner''s hype cycle ultimately has a happy ending.


After the peak of inflated expectations there''s a "trough of disillusionment"--what happens when people come to recognize that the new technology will still require a lot of hard work.   FIGURE P-2: GARTNER''S HYPE CYCLE   But right when views of the new technology have begun to lapse from healthy skepticism into overt cynicism, that technology can begin to pay some dividends. (We''ve been through this before: after the computer boom in the 1970s and the Internet commerce boom of the late 1990s, among other examples.) Eventually it matures to the point when there are fewer glossy advertisements but more gains in productivity--it may even have become so commonplace that we take it for granted. I hope this book can accelerate the process, however slightly.   This is a book about information, technology, and scientific progress. This is a book about competition, free markets, and the evolution of ideas. This is a book about the things that make us smarter than any computer, and a book about human error.


This is a book about how we learn, one step at a time, to come to knowledge of the objective world, and why we sometimes take a step back.   This is a book about prediction, which sits at the intersection of all these things. It is a study of why some predictions succeed and why some fail. My hope is that we might gain a little more insight into planning our futures and become a little less likely to repeat our mistakes.   More Information, More Problems   The original revolution in information technology came not with the microchip, but with the printing press. Johannes Gutenberg''s invention in 1440 made information available to the masses, and the explosion of ideas it produced had unintended consequences and unpredictable effects. It was a spark for the Industrial Revolution in 1775,1 a tipping point in which civilization suddenly went from having made almost no scientific or economic progress for most of its existence to the exponential rates of growth and change that are familiar to us today. It set in motion the events that would produce the European Enlightenment and the founding of the American Republic.


  But the printing press would first produce something else: hundreds of years of holy war. As mankind came to believe it could predict its fate and choose its destiny, the bloodiest epoch in human history followed.2   Books had existed prior to Gutenberg, but they were not widely written and they were not widely read. Instead, they were luxury items for the nobility, produced one copy at a time by scribes.3 The going rate for reproducing a single manuscript was about one florin (a gold coin worth about $200 in today''s dollars) per five pages,4 so a book like the one you''re reading now would cost around $20,000. It would probably also come with a litany of transcription errors, since it would be a copy of a copy of a copy, the mistakes having multiplied and mutated through each generation.   This made the accumulation of knowledge extremely difficult. It required heroic effort to prevent the volume of recorded knowledge from actually  decreasing,  since the books might decay faster than they could be reproduced.


Various editions of the Bible survived, along with a small number of canonical texts, like from Plato and Aristotle. But an untold amount of wisdom was lost to the ages,5 and there was little incentive to record more of it to the page.   The pursuit of knowledge seemed inherently futile, if not altogether vain. If today we feel a sense of impermanence because things are changing so rapidly, impermanence was a far more literal concern for the generations before us. There was "nothing new under the sun," as the beautiful Bible verses in Ecclesiastes put it--not so much because everything had been discovered but because everything would be forgotten.6   The printing press changed that, and did so permanently and pro.


To be able to view the table of contents for this publication then please subscribe by clicking the button below...
To be able to view the full description for this publication then please subscribe by clicking the button below...