Future Babble - BestLightNovel.com
You’re reading novel Future Babble Part 1 online at BestLightNovel.com. Please use the follow button to get notification about the latest chapter next time when you visit BestLightNovel.com. Use F11 button to read novel in full-screen(PC only). Drop by anytime you want to read free – fast – latest novel. It’s great if you could leave a comment, share your opinion about the new chapters, new novel with others on the internet. We’ll do our best to bring you the finest, latest novel everyday. Enjoy
Future babble.
Why expert predictions are next to worthless, and you can do better.
Dan Gardner.
For Mum and Dad, who gave me a future, and for Victoria, Winston, and MacDougall, who are the future.
Preface.
As I read the newspaper this morning, I became engrossed in a story about what would happen in the future.
The story said the latest economic forecast of the Organization for Economic Cooperation and Development (OECD) showed growth over the next couple of years would be stronger than previously projected. I read every word avidly. Like most people these days, I'm worried about the economy. The last few years have not been good ones. There was a calamitous credit crisis, the puncturing of real estate bubbles in one country after another, and a global recession worse than anything since the Second World War. Then came a weak recovery, debt crises, and fears of inflation. Or deflation. Take your pick. Who knows what will come next? Not me, although I very much want to. So when my morning newspaper delivered the knowledge I crave, from the esteemed experts at the OECD no less, it had my full and enthusiastic attention. Growth will be better than expected! Fabulous!
But then I realized something. The phrase better than expected means there was a forecast that preceded this one, and by issuing a new and different forecast, the OECD was conceding that the latest information suggested the earlier forecast was wrong. If the first forecast could fail, so could the second, and yet I had reacted to the second forecast as if it were a sure thing. It was also a little sobering to recall that the OECD had forecast smooth sailing in 2008 and 2009. Hurricane? What hurricane?
And I remembered something else. I had just finished writing a book about how awful experts are at predicting the future and how psychology compels people to take predictions seriously anyway.
Case in point: me.
I mention this because the mistakes and delusions I chronicle in this book are human follies. We are all susceptible to them, even authors who write about the mistakes and delusions of others. The goal of this book is not to mock particular individuals. Nor is it to scorn the category known as "experts." It is to better understand the human desire to know what will happen, why that desire will never be satisfied, and how we can better prepare ourselves for the unknowable future.
Although most of the issues discussed in this book are very much in the news today-from the economy, to oil prices, to environmental catastrophe-I generally don't examine current predictions for the simple reason that no one really knows if they are accurate or not. I may feel that a prediction is sensible or silly, but others may feel differently, and arguing about it will get us nowhere. Only the pa.s.sage of time can settle it. And so I focus on recent history: If, for example, someone predicted a second Great Depression would start in 1990 (as one economist did in a wildly popular book), then I can say with considerable confidence that he was wrong. And no reasonable person will disagree.
I should also note that while I have tried to be fair in portraying what experts predicted and how accurate those predictions were, I often had to condense entire books to a few sentences and sum up complex evidence with equal brevity. Inevitably, these summaries are full of subjective judgments: Do not take what I say as the final word. Most of the papers and articles discussed here are available on the Internet; all the books can be found in libraries and used book stores. Have a look and decide for yourself. I know that reading old predictions about a future that is now the past may sound hopelessly dull, but it's not. In fact, it's surprisingly enjoyable and rewarding. It's raw history, after all. There is no better way to climb inside a moment in time and see the world from that perspective. Or to realize that while the future we face may be disturbingly uncertain, 'twas ever thus-which is oddly rea.s.suring.
My profound thanks to Philip Tetlock, for doing the demanding research that is the foundation of this book and for being generous and helpful at an unimaginably difficult time in his life. A gentleman and a scholar, indeed.
Also thanks to my research a.s.sistant, Courtney Symons, and my marvelous editors, Susan Renouf at McClelland & Stewart and Stephen Morrow at Dutton. Many others contributed one way or another to the book in your hands, including Scott Gilmore, Rudyard Griffiths, Liam Scott, Bruce Schneier, and my longtime friends and colleagues David Watson and Leonard Stern. A special thanks to Paul Slovic, John Mueller, Marc Ramsay, and Barry Dworkin for reading early versions and providing invaluable thoughts and suggestions. Of course, having been written by a fallible human, this book will contain errors-for which the author is solely responsible. All corrections and constructive criticism will be gratefully received.
And finally, thanks to my wife, Sandra, without whom I could do, and would be, nothing.
1.
Introduction.
The end of everything we call life is close at hand and cannot be evaded.
-H. G. WELLS, 1946.
George Edward Scott, my mother's father, was born in an English village near the city of Nottingham. It was 1906. We can be sure that anyone who took notice of George's arrival in the world agreed that he was a very lucky baby.
There was the house he lived in, for one thing. It was the work of his father, a successful builder, and it was, like the man who built it, correct, confident, and proudly Victorian. Middle-cla.s.s prosperity was evident throughout, from the sprawling rooms to the stained-gla.s.s windows and the cast-iron bathtub with a pull-cord that rang a bell downstairs. A maid carrying a bucket of hot water would arrive in due course.
And there was the country and the era. Often romanticized as the "long Edwardian summer," Britain at the beginning of the twentieth century was indeed a land of peace and prosperity, if not strawberries and champagne. Britain led the world in industry, science, education, medicine, trade, and finance. Its empire was vaster than any in history, its navy invincible. The great and terrible war with Napoleon's France was tucked away in dusty history books and few worried that its like would ever come again.
It was a time when "Progress" was capitalized. People were wealthier. They ate better and lived longer. Trade, travel, and communication steadily expanded, a process that would be called, much later, globalization. Science advanced briskly, revealing nature's secrets and churning out technological marvels, each more wonderful than the last, from the train to the telegraph to the airplane. The latest of these arrived only four years before George Scott was born, and in 1912, when George was six, his father gathered the family in a field to witness the miracle of a man flying through the air in a machine. The pilot waved to the gawkers below. "Now I've seen it," George's grandmother muttered. "But I still don't believe it."
And the future? How could it be anything but grand? In 1902, the great American economist John Bates Clark imagined himself in 2002, looking back on the last hundred years. He p.r.o.nounced himself profoundly satisfied. "There is certainly enough in our present condition to make our gladness overflow" and to hope that "the spirit of laughter and song may abide with us through the years that are coming," Clark wrote. The twentieth century had been a triumph, in Clark's imagining. Technology had flourished, conflict between labor and capital had vanished, and prosperity had grown until the slums were "transformed into abodes of happiness and health." Only trade had crossed borders, never armies, and in the whole long century not a shot had been fired in anger. Of course this was only to be expected, Clark wrote, even though some silly people in earlier generations had actually believed war could happen in the modern world-"as if nations bound together by such economic ties as now unite the countries of the world would ever disrupt the great industrial organism and begin fighting."
At the time, Clark's vision seemed as reasonable as it was hopeful, and it was widely shared by eminent persons. "We can now look forward with something like confidence to the time when war between civilized nations will be as antiquated as the duel," wrote the esteemed British historian G. P. Gooch, in 1911. Several years later, the celebrated Manchester Guardian journalist H. N. Norman was even more definitive. "It is as certain as anything in politics can be, that the frontiers of our modern national states are finally drawn. My own belief is that there will be no more wars among the six Great Powers."
One day, a few months after H. N. Norman had declared the arrival of eternal peace, George Scott fetched his father's newspaper. The top story was the latest development in the push for Irish home rule. Below that was another headline. "War Declared," it read.
It was August 1914. What had been considered impossible by so many informed experts was now reality. But still there was no need to despair. It would be "the war to end all wars," in H. G. Wells's famously optimistic phrase. And it would be brief. It has to be, wrote the editors of The Economist, thanks to "the economic and financial impossibility of carrying out hostilities many more months on the present scale."
For more than four years, the industry, science, and technology that had promised a better world slowly ground millions of men into the mud. The long agony of the First World War shattered empires, nations, generations, and hopes. The very idea of progress came to be scorned as a rotten illusion, a raggedy stage curtain now torn down and discarded.
In defeated Germany, Oswald Spengler's dense and dark Decline of the West was the runaway best seller of the 1920s. In victorious Britain, the Empire was bigger but the faith in the future that had sustained it faded like an old photograph left in the sun. The war left crus.h.i.+ng debts and the economy staggered. "Has the cycle of prosperity and progress closed?" asked H. G. Wells in the foreword to a book whose t.i.tle ventured an even bleaker question: Will Civilization Crash? Yes to both, answered many of the same wise men who had once seen only peace and prosperity ahead. "It is clear now to everyone that the suicide of civilization is in progress," declared the physician and humanitarian Albert Schweitzer in a 1922 lecture at Oxford University. It may have been "the Roaring Twenties" in the United States-a time of jazz, bathtub gin, soaring stocks, and real estate speculation-but it was a decade of gloom in Britain. For those who thought about the future, observes historian Richard Overy, "the prospect of imminent crisis, a new Dark Age, became a habitual way of looking at the world."
My grandfather's fortunes followed Britain's. His father's business declined, prosperity seeped away, and the bathtub pull-cord ceased to summon the downstairs maid. In 1922, at the age of fifteen, George was apprenticed to a plumber. A few years later, bowing to the prevailing sense that Britain's decline was unstoppable, he decided to emigrate. A coin toss-heads Canada, tails Australia-settled the destination. With sixty dollars in his pocket, he landed in Canada. It was 1929. He had arrived just in time for the Great Depression.
A horror throughout the industrialized world, the Great Depression was especially savage in North America. Half the industrial production of the United States vanished. One-quarter of workers were unemployed. Starvation was a real and constant threat for millions. Growing numbers of desperate, frightened people sought salvation in fascism or communism. In Toronto, Maple Leaf Gardens was filled to the rafters not for a hockey game but for a Stalinist rally, urging Canadians to follow the glorious example of the Soviet Union. Among the leading thinkers of the day, it was almost a truism that liberal democracy and free-market capitalism were archaic, discredited, and doomed. Even moderates were sure the future would belong to very different economic and political systems.
In 1933, the rise to power of the n.a.z.is added the threat of what H. G. Wells called the "Second World War" in his sci-fi novel The Shape of Things to Come. Published the same year Adolf Hitler became chancellor of Germany, The Shape of Things to Come saw the war beginning in 1940 and predicted it would become a decade-long ma.s.s slaughter, ending not in victory but in the utter exhaustion and collapse of all nations. Military a.n.a.lysts and others who tried to imagine another Great War were almost as grim. The airplanes that had been so wondrous to a young boy in 1912 would fill the skies with bombs, they agreed. Cities would be pulverized. There would be ma.s.s psychological breakdown and social disintegration. In 1934, Britain began a rearmament program it could not afford for a war that, it increasingly seemed, it could not avoid. In 1936, as n.a.z.i Germany grew stronger, the program was accelerated.
A flicker of hope came from the United States, where economic indicators jolted upward, like a flat line on a heart monitor suddenly jumping. It didn't last. In 1937, the American economy plunged again. It seemed nothing could pull the world out of its death spiral. "It is a fact so familiar that we seldom remember how very strange it is," observed the British historian G. N. Clark, "that the commonest phrases we hear used about civilization at the present time all relate to the possibility, or even the prospect, of its being destroyed."
That same year, George Scott's second daughter, June, was born. It is most unlikely that anyone thought my mother was a lucky baby.
The Second World War began in September 1939. By the time it ended in 1945, at least forty million people were dead, the Holocaust had demonstrated that humanity was capable of any crime, much of the industrialized world had been pounded into rubble, and a weapon vastly more destructive than anything seen before had been invented. "In our recent history, war has been following war in ascending order of intensity," wrote the influential British historian Arnold Toynbee in 1950. "And today it is already apparent that the War of 193945 was not the climax of this crescendo movement." Amba.s.sador Joseph Grew, a senior American foreign service officer, declared in 1945 that "a future war with the Soviet Union is as sure as anything in this world." Albert Einstein was terrified. "Only the creation of a world government can prevent the impending self-destruction of mankind," declared the man whose name was synonymous with genius. Some were less optimistic. "The end of everything we call life is close at hand and cannot be evaded," moaned H. G. Wells.
Happily for humanity, Wells, Einstein, and the many other luminaries who made dire predictions in an era W.H. Auden dubbed "the Age of Anxiety" were all wrong. The end of life was not at hand. War did not come. Civilization did not crumble. Against all reasonable expectation, my mother turned out to be a very lucky baby indeed.
Led by the United States, Western economies surged in the postwar decades. The standard of living soared. Optimism returned, and people expressed their hope for a brighter future by getting married earlier and having children in unprecedented numbers. The result was a combination boom-economic and baby-that put children born during the Depression at the leading edge of a wealth-and-population wave. That's the ultimate demographic sweet spot. Coming of age in the 1950s, they entered a dream job market. To be hired at a university in the early 1960s, a professor once recalled to me, you had to sign your name three times "and spell it right twice." Something of an exaggeration, to be sure. But the point is very real. Despite the constant threat of nuclear war, and lesser problems that came and went, children born in the depths of the Great Depression-one of the darkest periods of the last five centuries-lived their adult lives amid peace and steadily growing prosperity. There has never been a more fortunate generation.
Who predicted that? n.o.body. Which is entirely understandable. Even someone who could have foreseen that there would not be a Third World War-which would have been a triumph of prognostication in its own right-would have had to correctly forecast both the baby boom and the marvelous performance of postwar economies. And how would they have done that? The baby boom was caused by a postwar surge in fertility rates that sharply reversed a downward trend that had been in place for more than half a century. Demographers didn't see it coming. No one did. Similarly, the dynamism of the postwar economies was a sharp break from previous trends that was not forecast by experts, whose expectations were much more pessimistic. Many leading economists even worried that demobilization would be followed by ma.s.s unemployment and stagnation. One surprise after another. That's how the years unfolded after 1945. The result was a future that was as unpredictable as it was delightful-and a generation born at what seemed to be the worst possible time came to be a generation born at the most golden of moments.
The desire to know the future is universal and constant, as the profusion of soothsaying techniques in human cultures-from goats' entrails to tea leaves-demonstrates so well. But certain events can sharpen that desire, making it fierce and urgent. Bringing a child into the world is one such force. What will the world be like for my baby? My great-grandfather undoubtedly asked himself that question when his little boy was born in 1906. He was a well-read person, and so he likely paid close attention to what the experts said. George Edward Scott was a very lucky baby, he would have concluded. And any intelligent, informed person would have agreed. Thirty-one years later, when my grandfather held his infant daughter in his arms, he surely asked himself the same question, and he, too, would have paid close attention to what the experts said. And he would have feared for her future, as any intelligent, informed person would have.
My great-grandfather was wrong. My grandfather was wrong. All those intelligent, informed people were wrong. But mostly, the experts were wrong.
They're wrong a lot, those experts. History is littered with their failed predictions. Whole books can be filled with them. Many have been.
Some failed predictions are prophecies of disaster and despair. In the 1968 book The Population Bomb, which sold millions of copies, Stanford University biologist Paul Ehrlich declared "the battle to feed all of humanity is over. In the 1970s, the world will undergo famines-hundreds of millions of people will starve to death in spite of any crash programs embarked upon now." But there weren't ma.s.s famines in the 1970s. Or in the 1980s. Thanks to the dramatic improvements in agriculture collectively known as "the Green Revolution"-which were well under way by the time Ehrlich wrote his book-food production not only kept up with population growth, it greatly surpa.s.sed it. Ehrlich thought that was utterly impossible. But it happened. Between 1961 and 2000, the world's population doubled but the calories of food consumed per person increased 24 percent. In India, calories per person rose 20 percent. In Italy, 26 percent. In South Korea, 44 percent. Indonesia, 69 percent. China had experienced a famine that killed some thirty million people in the dark years between 1959 and 1961, but in the forty years after that horror, China's per capita food consumption rose an astonis.h.i.+ng 73 percent. And the United States? In the decades after The Population Bomb was published, fears that people would not get enough to eat were forgotten as American waistlines steadily expanded. The already-substantial consumption of the average American rose 32 percent, and the United States became the first nation in history to struggle with an epidemic of obesity.
In 1977, President Jimmy Carter called for the "moral equivalent of war" to s.h.i.+ft the American economy off oil because, he said, the production of oil would soon fail to keep up with demand. When that happened, oil prices would soar and never come down again-the American economy would be devastated and the American dream would turn brown and die like an unwatered suburban lawn. Eight years later, oil prices fell through the floor. They stayed low for two decades.
A small library could be filled with books predicting stock market crashes and economic disasters that never happened, but the giant of the genre was published in 1987. The hardcover edition of economist Ravi Batra's The Great Depression of 1990 hit the top spot on The New York Times best-seller list and spent a total of ten months on the chart; the paperback stayed on the list for an astonis.h.i.+ng nineteen months. When the American economy slipped into recession in 1990, Batra looked prophetic. When the recession proved to be mild and brief, he seemed less so. When the 1990s roared, he looked foolish, particularly when he spent the entire decade writing books predicting a depression was imminent.
In 1990, Jacques Attali-intellectual, banker, former adviser to French president Francois Mitterrand-published a book called Millennium, which predicted dramatic change on the other side of the year 2000. Both the United States and the Soviet Union would slowly lose their superpower status, Attali wrote. Their replacements would be j.a.pan and Europe. As for China and India, they "will refuse to fall under the sway of either the Pacific or the European sphere," but it would be hard for these desperately poor countries to resist. Catastrophic war was "possible, even probable." However, Attali cautioned, this future isn't quite chiseled in stone. "If a miracle were to occur" and China and India were to be "integrated into the global economy and market, all strategic a.s.sumptions underpinning my prognostications would be overturned. That miracle is most unlikely." Of course, that "miracle" is precisely what happened. And almost nothing Attali predicted came true.
Even economists who win n.o.bel Prizes have been known to blow big calls. In 1997, as Asian economies struggled with a major currency crisis, Paul Krugman-New York Times columnist and winner of the n.o.bel Prize in 2008-worried that Asia must act quickly. If not, he wrote in Fortune magazine, "we could be looking at a true Depression scenario-the kind of slump that sixty years ago devastated societies, destabilized governments, and eventually led to war." Krugman's prescription? Currency controls. It had to be done or else. But mostly, it wasn't done. And Asia was booming again within two years.
Pessimists have no monopoly on forecasting flops, however. Excited predictions of the amazing technologies to come-Driverless cars! Robot maids! Jet packs!-have been dazzling the public since the late nineteenth century. These old forecasts continue to entertain today, though for quite different reasons. And for every bear prophesying blood in the stock markets, there is a bull who is sure things will only get better. The American economist Irving Fisher was one. "Stock prices have reached what looks like a permanently high plateau," the esteemed economist a.s.sured nervous investors. "I do not feel there will soon be, if ever, a fifty- or sixty-point break from present levels, such as they have predicted. I expect to see the stock market a good deal higher within a few months." That was October 17, 1929. The market crashed the following week. But that crash was none of Britain's concern, the legendary John Maynard Keynes believed. "There will be no serious consequences in London resulting from the Wall Street Slump," Keynes wrote. "We find the look ahead decidedly encouraging." Shortly afterward, Britain sank with the rest of the world into the Great Depression.
Another bull market, this one in the late 1990s, produced a bookshelf full of predictions so giddy they made Irving Fisher sound like Eeyore. The most famous was the 1999 book Dow 36,000 by James Gla.s.sman and Kevin Ha.s.sett. "If you are worried about missing the market's big move upward, you will discover that it's not too late," Gla.s.sman and Ha.s.sett wrote. Actually, it was too late. Shortly after Dow 36,000 was published, the Dow peaked at less than 12,000 and started a long, painful descent.
Paul Ehrlich can also take consolation in the fact that many of the optimists who a.s.sailed his writing were not much better at predicting the future. "The doomsayers who worry about the prospect of starvation for a burgeoning world population" will not see their terrible visions realized, Time magazine reported in 1966. The reason? Aquaculture. "RAND experts visualize fish herded and raised in offsh.o.r.e pens as cattle are today. Huge fields of kelp and other kinds of seaweed will be tended by undersea 'farmers'-frogmen who will live for months at a time in submerged bunkhouses. The protein-rich underseas crop will probably be ground up to produce a dull-tasting cereal that eventually, however, could be regenerated chemically to taste like anything from steak to bourbon." The same RAND Corporation experts agreed that "a permanent lunar base will have been established long before A.D. 2000 and that men will have flown past Venus and landed on Mars." Herman Kahn, a founder of the Hudson Inst.i.tute and a determined critic of Ehrlich, was similarly off the mark in a thick book called The Year 2000, published in 1967. It is "very likely," Kahn wrote, that by the end of the century, nuclear explosives would be used for excavation and mining, "artificial moons" would be used to illuminate large areas at night, and there would be permanent undersea colonies. Kahn also expected that one of the world's fastest-growing economies at the turn of the millennium would be that of the Soviet Union.
So pessimists and optimists both make predictions that look bad in hindsight. What about left versus right? Not much difference there either. There are plenty of examples of liberal experts making predictions that go awry, like Jonathan Sch.e.l.l's belief that Ronald Reagan's arms buildup was putting the world on course for nuclear war. "We have to admit that unless we rid ourselves of our nuclear a.r.s.enals a holocaust not only might occur but will occur-if not today, then tomorrow; if not this year, then the next," Sch.e.l.l wrote in 1982. "One day-and it is hard to believe it will not be soon-we will make a choice. Either we will sink into the final coma and end it all or, as I trust and believe, we will awaken to the truth of our peril . . . and rise up to cleanse the earth of nuclear weapons." The stock of failed predictions on the right is equally rich. It was, for example, a "slam dunk" that Saddam Hussein's weapons of ma.s.s destruction would be discovered following the American invasion of Iraq in 2003 and that, as Vice President d.i.c.k Cheney said, American soldiers would be "greeted as liberators" by the grateful Iraqi people. "A year from now," observed neoconservative luminary Richard Perle in September 2003, "I'll be very surprised if there is not some grand square in Baghdad named after President Bush."
So the inaccuracy of expert predictions isn't limited to pessimists or optimists, liberals or conservatives. It's also not about a few deluded individuals. Over and over in the history of predictions, it's not one expert who tries and fails to predict the future. It's whole legions of experts.
Paul Ehrlich's bleak vision in The Population Bomb was anything but that of alone crank. Countless experts made similar forecasts in the 1950s and 1960s. In 1967, the year before Ehrlich's book appeared, William and Paul Paddock-one an agronomist, the other a foreign service officer-published a book whose t.i.tle said it all: Famine 1975! When biologist James Bonner reviewed the Paddocks' book in the journal Science, he emphasized that "all serious students of the plight of the underdeveloped nations agree that famine among the peoples of the underdeveloped nations is inevitable." The only question was when. "The U.S. Department of Agriculture, for example, sees 1985 as the beginning of the years of hunger. I have guessed publicly that the interval 19771985 will bring the moment of truth, will bring a dividing point at which the human race will split into the rich and the poor, the well-fed and the hungry-two cultures, the affluent and the miserable, one of which must inevitably exterminate the other. . . . I stress again that all responsible investigators agree that the tragedy will occur."
There was also an expert consensus in support of Jimmy Carter's prediction of perpetually rising oil prices. And Jacques Attali's belief that j.a.pan and Europe would eclipse the United States and dominate the world economy in the twenty-first century was standard stuff among strategic thinkers. As for Jonathan Sch.e.l.l's fear that Ronald Reagan's policies would plunge the world into a nuclear inferno, it dominated university faculties and brought millions of protestors to the streets. And, as easy as it is to forget now, support for the invasion of Iraq was widespread among foreign policy a.n.a.lysts and politicians, most of whom were confident weapons of ma.s.s destruction would be uncovered and American forces greeted as liberators.
We are awash in predictions. In newspapers, blogs, and books, on radio and television, everyday, without fail, experts tell us how the economy will perform next year or whether a foreign conflict will flare into war. They tell us who will win the next election, and whether the price of oil will rise or fall, housing sales will grow or shrink, stock markets will soar or dive. Occasionally, the experts lift their eyes to more distant horizons. I recently read a cover story in Time magazine that claimed the first ten years of the twenty-first century were "the decade from h.e.l.l" and went on to explain "why the next one will be better." But what made the "decade from h.e.l.l" what it was? Events that confounded the expectations of most experts. The 9/11 terrorist attacks. The debacle in Iraq. Hurricane Katrina. The financial crisis of 2008 and the global recession. If the previous decade was shaped by uncertainty and surprise-and no one can seriously argue it was not-why would we expect the next ten years to be so much more predictable? But simple questions like that are seldom asked. Instead, the predictions are churned out, one after another, like widgets on an a.s.sembly line. I recently read a description of the Chinese economy in 2040. And American suburbs in 2050. And now I'm reading an article that explains "why Europe will outs.h.i.+ne North America in the twenty-first century." There are apparently no limits to the vision of these wise men and women. Experts peer into the distant future and warn of great wars and conflicts. They tell us what's in store for the climate, globalization, food, energy, and technology. They tell us all about the world of our children and our grandchildren. And we listen.
Economists, in particular, are treated with the reverence the ancient Greeks gave the Oracle of Delphi. But unlike the notoriously vague p.r.o.nouncements that once issued from Delphi, economists' predictions are concrete and precise. Their accuracy can be checked. And anyone who does that will quickly conclude that economists make lousy soothsayers: "The record of failure to predict recessions is virtually unblemished," wrote IMF economist Prakash Loungani in one of many papers demonstrating the near-universal truth that economists' predictions are least accurate when they are most needed. Not even the most esteemed economists can claim significant predictive success. Retired banker and financial writer Charles Morris examined a decade's worth of forecasts issued by the brilliant minds who staff the White House's Council of Economic Advisors. Morris started with the 1997 forecast. There would be modest growth, the council declared; at the end of the year, the American economy had grown at a rate more than double the council's forecast. In 1998, the story was much the same. And in 1999. In 2000, the council "sharply raised both their near- and medium-term outlooks-just in time for the dot-com bust and the 20012002 recession." The record for the Bush years was "no better," Morris writes. But it was the forecast for 2008 that really amazes: "The 2008 report expected slower but positive growth in the first half of the year, as investment s.h.i.+fted away from housing, but foresaw a nice recovery in the second half, and a decent year overall. Their outlook for 2009 and 2010 was for a solid three percent real growth with low inflation and good employment numbers," Morris writes. "In other words, they hadn't a clue."
And they weren't alone. With very few exceptions, economists did not foresee the financial and economic meltdown of 2008. Many economists didn't recognize the crisis for what it was even as it was unfolding. In December 2007-months after the credit crunch began and the very moment that would officially mark the beginning of the recession in the United States-BusinessWeek magazine ran its annual chart of detailed forecasts for the year ahead from leading American a.n.a.lysts. Under the headline "A Slower but Steady Economy," every one of fifty-four economists predicted the U.S. economy wouldn't "sink into a recession" in 2008. The experts were unanimous that unemployment wouldn't be too bad, either, leading to the consensus conclusion that 2008 would be a solid but unspectacular year. One horrible year later-as people watching the evening news experienced the white-knuckle fear of pa.s.sengers in a plunging jet-BusinessWeek turned to the economists who had so spectacularly blown that year's forecast and asked them to tell its readers what would happen in 2009. There was no mention of the previous year's fiasco, only another chart filled with rea.s.suringly precise numbers. The headline: "A Slower but Steady Economy."
By definition, experts know much about their field of expertise. Economists can-usually-look around and tell us a great deal about the economy, political scientists can do the same for politics and government, ecologists for the environment, and so on. But the future? All too often, their crystal b.a.l.l.s work no better than those of fortune-tellers. And since rational people don't take seriously the prognostications of Mysterious Madam Zelda or any psychic, palm reader, astrologer, or preacher who claims to know what lies ahead, they should be skeptical of expert predictions. And yet we are not skeptical. No matter how often expert predictions fail, we want more. This strange phenomenon led Scott Armstrong, an expert on forecasting at the Wharton School of the University of Pennsylvania, to coin his "seer-sucker" theory: "No matter how much evidence exists that seers do not exist, suckers will pay for the existence of seers." Sometimes we even go back to the very people whose predictions failed in the past and listen, rapt, as they tell us how the future will unfold.
This book explains why expert predictions fail and why we believe them anyway.
The first part of the answer lies in the nature of reality and the human brain. The world is complicated-too complicated to be predicted. And while the human brain may be magnificent, it is not perfect, thanks to a jumble of cognitive wiring that makes systematic mistakes. Try to predict an unpredictable world using an error-p.r.o.ne brain and you get the gaffes that litter history.
As for why we believe expert predictions, the answer lies ultimately in our hardwired aversion to uncertainty. People want to know what's happening now and what will happen in the future, and admitting we don't know can be profoundly disturbing. So we try to eliminate uncertainty however we can. We see patterns where there are none. We treat random results as if they are meaningful. And we treasure stories that replace the complexity and uncertainty of reality with simple narratives about what's happening and what will happen. Sometimes we create these stories ourselves, but, even with the human mind's bountiful capacity for self-delusion, it can be hard to fool ourselves into thinking we know what the future holds for the stock market, the climate, the price of oil, or a thousand other pressing issues. So we look to experts. They must know. They have Ph.D.'s, prizes, and offices in major universities. And thanks to the news media's preference for the simple and dramatic, the sort of expert we are likely to hear from is confident and conclusive. They know what will happen; they are certain of it. We like that because that is how we want to feel. And so we convince ourselves that these wise men and women can do what wise men and women have never been able to do before. Fundamentally, we believe because we want to believe.
We need to see this trap for what it is, especially at this moment. Over the last several years, we have experienced soaring prices for commodities, food shortages, talk of an "age of scarcity," the bursting of a real estate bubble that ruined millions of middle-cla.s.s homeowners, growing evidence of environmental catastrophe, a financial crisis that upset conventional economic wisdom, and a global economic recession the like of which has not been seen since the Second World War. Uncertainty? The air is electric with it. It's precisely in times such as these that the desire to know what the future holds becomes a ravenous hunger. We've seen it happen before. The 1970s may be remembered as the era of disco and bad fas.h.i.+on, but it was, in reality, a tumultuous and unsettling time that created an enormous demand to know what lay ahead. The result was a profusion of detailed and compelling expert predictions, many of them involving the very same issues-oil, food, terrorism, recession, unemployment, deficits and debt, inflation, environmental crisis, the decline of the United States-we are grappling with today. Most of them turned out to be wrong, some hilariously so. That doesn't prove that similar predictions in the present will also fall flat, but it does provide a valuable reminder to be skeptical when experts claim to know what lies in our future.
That sort of skepticism doesn't come easily, but it is possible. As natural as it is to want to hear predictions, and to believe them, we do not have to. With effort, we can learn to accept reality when we do not, and cannot, know what lies ahead.
Of course, that still leaves us with a big problem because, in our lives and businesses, we all have to make plans and forecasts. If the future is unpredictable, doesn't that mean all our planning and forecasting is pointless? Not if we go about it the right way. Certain styles of thinking and decision making do a far better job of groping amid the inky blackness of the future to find a path ahead. These styles can be learned and applied, with results that are positive, although far from perfect. And that leads to the ultimate conclusion, which is one we do not want to accept but must: There are no crystal b.a.l.l.s, and no style of thinking, no technique, no model will ever eliminate uncertainty. The future will forever be shrouded in darkness. Only if we accept and embrace this fundamental fact can we hope to be prepared for the inevitable surprises that lie ahead.
ARE EXPERTS REALLY SO BAD?.
But now I have to pause and make an admission: My whole argument is based on the belief that expert predictions have a lousy track record. But I haven't actually proved that, at least not yet.
So far, I've presented a number of expert predictions that failed. Or rather, I've presented a number of expert predictions that I think failed. But not everyone would agree. Many people insist even today that Paul Ehrlich was essentially on the mark in The Population Bomb. One of those people is Paul Ehrlich. In a 2009 essay, Ehrlich acknowledged that the book "underestimated the impact of the Green Revolution" and so the starvation he expected wasn't as bad as he predicted. But the book's grim vision was basically accurate, he insisted. In fact, its "most serious flaw" was that it was "much too optimistic about the future."
I'll take a closer look at Ehrlich's defense of The Population Bomb later. What matters here is that the failure of Ehrlich's prediction is disputed, and untangling that dispute is complicated. That's typical because expert predictions are common and so are failed predictions. But experts who agree that their predictions failed are rare. As Paul Ehrlich did, they will often concede that they were off on some details here and there. But flat-out wrong? No. Never. Unless pinned down by circ.u.mstances as firmly as a b.u.t.terfly in a display case, they will resolutely deny being wrong.
"I was almost right" is a standard dodge. Another is "It would have happened if I hadn't been blindsided by an unforeseeable event." And then there is the claim that the prediction was a "self-negating prophecy," that it caused others to act and it was those actions that prevented the predicted event from happening. Remember Y2K? The more excitable experts claimed the world's computers would crash on January 1, 2000, and take civilization with them. When nothing remotely like that happened, the doomsters boasted that their predictions had prompted ma.s.sive remediation efforts that had saved humanity from certain doom: You're welcome.
Another mental maneuver is the wait-and-see twist. Many predictions have only vague time frames and so, when an observer thinks a forecast has failed, the expert can insist time isn't up yet: Wait and see. A variation on this is the off-on-timing gambit, which is used when a prediction comes with a clear time frame and the prediction clearly fails within the allotted time: The expert grudgingly concedes that the predicted event hasn't happened within the time frame but insists that's a minor detail. What matters is that the prediction will come to pa.s.s. Eventually. Someday. Paul Ehrlich, for example, acknowledges that the famines he predicted for the 1970s didn't happen, at least not to the extent he expected, but he insists that his a.n.a.lysis was sound and the disasters he foresaw are still coming. "The probability of a vast catastrophe looms steadily larger," he wrote in 2009, forty-one years after The Population Bomb warned of imminent peril. Similarly, when a journalist reminded Richard Perle in 2008 that five years earlier he had predicted a grand square in Baghdad would be named after George W. Bush within one year, Perle didn't respond with a forthright admission of error. Instead, he insisted Bush could still get his Baghdad square. It would just take a little longer than antic.i.p.ated.
A third defense involves carefully parsing the language of the forecast so that a statement that was intended to be a rock-solid prediction that Event X would happen-and is taken that way by the media and the public-is shown to be much more elastic. "I didn't actually say Event X would certainly happen," the expert explains. "I said 'It could happen.'" And implicit in the phrase could happen is the possibility that the predicted event may not happen. Thus, the fact that the event did not happen does not mean the prediction was "wrong." This line of reasoning is often heard from liberal experts who claimed, in the early 1980s, that Ronald Reagan's policies put the world in danger. Very few said nuclear war was "inevitable." They only said Reagan's belligerence made war more likely. Does the fact that there was no war prove they were wrong? Not at all, they say. It's like a weather forecaster who says there is a 70 percent chance of rain. He can't be blamed if the sun s.h.i.+nes because an implicit part of his forecast was "30 percent chance of suns.h.i.+ne." People should just be glad they got lucky.
Obviously, I'm being a little sarcastic here, because these arguments are often weak and self-serving. But not always. Sometimes there is real substance in them and they have to be taken seriously. Predictions about the damage a widening hole in the ozone layer would do, for example, did cause governments to make policy changes that would ensure the predictions did not come true: That's a genuine "self-negating prophecy." It is also undeniably true to say, as Paul Ehrlich does, that the failure of population growth to cause famines in the 1970s does not prove population growth will not cause famines sometime in the future. And the fact that there was no nuclear war in the 1980s really does not prove that Ronald Reagan's policies did not raise the risk of war.
Put all this together and it means there are substantial question marks over many of the failed predictions I presented. Did they really fail? I think so. But reasonable people can and do disagree. Sorting out who's right isn't easy. Different observers will come to different conclusions. In some cases, the truth may never be known.
And there's an even bigger objection that can be raised to my claim about the fallibility of expert predictions: Even if we accept that my examples of failed predictions really are failed predictions, they don't actually prove that expert predictions routinely fail. They only prove that some expert predictions have failed. Even if I were to stuff whole chapters with examples, all I would prove is that many expert predictions have failed. What would be missing is what's needed to prove my point: the rate of failure. If, say, ninety-nine out of one hundred predictions fail, we would probably be better off consulting fortune cookies. But if one in one hundred fails, expert predictions really should be treated with hushed reverence.
So how do we figure out the rate of failure? The first thing we would have to do is expand our inquiry beyond the misses to the hits. And there are hits. Here's one: In 1981, energy expert Amory Lovins predicted that sometime between 1995 and 2005 the world would see "the effective collapse of the Soviet Union from internal political stress." That's pretty impressive, and there are plenty of others like it in the pages of books, magazines, and journals.
But still I wouldn't be able to prove much. What's the total number of experts I would be examining? What's the total number of predictions they made? Over what period? Simply adding the hits and misses I collected wouldn't tell me any of that, and so I still wouldn't know the rate at which predictions fail.
And if that's not complicated enough, there's another frustrating problem to contend with: Imagine someone who throws a dart and-smack!-he hits the bull's-eye. Does that prove he is a great dart thrower? Maybe. But there's no way to be sure based on that one dart. If he throws a second, third, and fourth dart and they all hit the bull's-eye, it's increasingly reasonable to think we are witnessing skill, not luck. But what if he throws dozens more darts and not one hits the bull's-eye? What if he often misses the board entirely? What if this person leaves darts scattered around the room, even a few stuck in the ceiling? In that case, his bull's-eye is probably a fluke. Amory Lovins's amazing prediction about the Soviet Union is a case in point. It was only one of dozens of predictions Lovins made in the same forecast, and almost all the others completely missed the board. (By the end of the 1980s, Lovins predicted, nuclear power programs would "persist only in dictators.h.i.+ps," oil and gas would be scarce and fantastically expensive, unemployment would be high and persistent, the unreliability of food supplies in American cities would give rise to "urban farming and forestry" . . . and so on.) Once you know that, you know it probably wasn't keen geopolitical insight that produced Lovins's bull's-eye. It was luck.
By now, I suspect, your head is swimming. That's the point, I'm afraid. Figuring out how good experts are at predicting the future seems like a simple task, but if we take logic and evidence seriously, it's actually very difficult.
The media have occasionally taken a stab at sorting this out. In 1984, The Economist asked sixteen people to make ten-year forecasts of economic growth rates, inflation rates, exchange rates, oil prices, and other staples of economic prognostication. Four of the test subjects were former finance ministers, four were chairmen of multinational companies, four were economics students at Oxford University, and four were, to use the English vernacular, London dustmen. A decade later, The Economist reviewed the forecasts and discovered they were, on average, awful. But some were more awful than others: The dustmen tied the corporate chairmen for first place, while the finance ministers came last. Many other publications have conducted similar exercises over the years, with similarly humiliating results. The now-defunct magazine Brill's Content, for one, compared the predictions of famous American pundits with a chimpanzee named Chippy, who made his guesses by choosing among flash cards. Chippy consistently matched or beat the best in the business.
As suggestive and entertaining as these stunts are, they are not, to say the least, scientifically rigorous. Rising to that level requires much more: It requires an experiment that is elaborate, expensive, and exhausting.
THE EXPERIMENT.
The first thing the experiment needs is a very large group of experts. The group should be as diverse as possible, with experts from different fields, different political leanings, different inst.i.tutional affiliations, and different backgrounds. At the very beginning of the experiment, the experts should answer a battery of questions designed to test political orientation, worldview, personality, and thinking style.
The experts must be asked clear questions whose answers can later be shown to be indisputably true or false. That means vague p.r.o.nouncements about "weakening state authority" or "growing public optimism" won't do. Even a question like "Will relations between India and Pakistan be increasingly strained?"-which is the standard language of TV pundits-isn't good enough. Questions have to be so precise that no reasonable person would argue about what actually happened-which means asking questions like "Will the official unemployment rate be higher, lower, or the same a year from now?" and "Will India and Pakistan go to war within the next five years?"
For each prediction, experts must state how likely they think it is to actually happen. If they are dead certain something will happen, that is a 100 percent probability. If they are sure it won't happen, it's a 0 percent probability. In between these extremes, experts will be required to attach precise percentages to guesses rather than use vague terms like improbable or very likely. There's no room for fudging when someone says, "There is a 30 percent chance India and Pakistan will go to war within the next five years."
The experiment must obtain a very large number of predictions from each expert in order to allow statistical a.n.a.lysis that can expose lucky hits for what they are. It also allows us to get past the problem of judging predictions in which the expert says the chance of something happening is, for example, "70 percent." If the expert is perfectly accurate, then a broad survey of his predictions will show that in 70 percent of the cases in which he said there was a 70 percent chance of something happening, it actually happened. Similarly, 60 percent of the outcomes said to have a 60 percent chance of happening should have happened. This measure of accuracy is called "calibration."
But there's more to the story than calibration. After all, someone who sat on the fence with every prediction-"Will it happen? I think the odds are 50/50"-would likely wind up with a modestly good calibration score. We can get predictions like that from a flipped coin. What we want in a forecaster, ideally, is someone with a G.o.dlike ability to predict the future. The G.o.ds don't bother with middling probabilities and they certainly don't say, "The odds are 50/50." The G.o.ds say, "This will certainly happen" or "This is impossible." So there must be a second measure of accuracy to go along with calibration. Experts should be scored by confidence. This means that an expert who said there is a 100 percent chance of something happening that actually did happen would score more points than another expert who had said there was only a 70 percent chance of it happening. This measure is called "discrimination."
A third measure must also be generated by answering the same questions that are put to the experts using a variety of simple and arbitrary rules. For example, there is the "no change" rule: No matter what the question is, always predict there will be no change. These results will create benchmarks against which the experts' results can be compared.
And finally, the experiment must continue over the course of many years. That will allow for questions involving time frames ranging from the short term-one to two years-to longer-term predictions covering five, ten, even twenty years, ensuring that the experiment will require experts to make predictions in times of stability and surprise, prosperity and recession, peace and war. When the pa.s.sage of time has revealed the correct answers, they should show how well the experts did. And be given the opportunity to explain the results.
It's difficult to exaggerate how demanding this experiment would be. It would be expensive and complicated and would require the patience of Job. But most of all, it would require a skilled and devoted researcher prepared to give a big chunk of his life to answering one question: How accurate are expert predictions?
Fortunately, there is such a researcher. He is Philip Tetlock.
Today, Tetlock is a much-honored psychologist at the University of California's Haas School of Business. In 1984, he was a newly tenured academic who had just been appointed to a new committee of the National Research Council, a branch of the National Academy of Sciences, arguably the most prestigious scientific body in the world. The committee's remit was nothing less than figuring out how social scientists could help avoid nuclear war and the end of civilization.
"It's hard to re-create the tenor of the times," Tetlock recalls, "but there was a lot of uneasiness." It was the height of the "second" Cold War. The Reagan White House was stockpiling nuclear weapons, the Red Army was fighting CIA-backed guerrillas in Afghanistan, and the death of Leonid Brezhnev had put the Soviet regime into transition, though to what no one could be sure. Watching television in living rooms across the United States, Americans were shocked when the evening news reported a Soviet fighter jet had shot down a Korean Airlines pa.s.senger jet that had strayed into Soviet airs.p.a.ce; then they were terrified by a made-for-TV drama, The Day After, about the ash and tears of life following a nuclear exchange.