The intelligence failures surrounding the invasion of Iraq dramatically illustrate the necessity of developing standards for evaluating expert opinion. This book fills that need. Here, Philip E. Tetlock explores what constitutes good judgment in predicting future events, and looks at why experts are often wrong in their forecasts. Tetlock first discusses arguments about whether the world is too complex for people to find the tools to understand political phenomena, let alone predict the future. He evaluates predictions from experts in different fields, comparing them to predictions by well-informed laity or those based on simple extrapolation from current trends. He goes on to analyze which styles of thinking are more successful in forecasting. Classifying thinking styles using Isaiah Berlin's prototypes of the fox and the hedgehog, Tetlock contends that the fox--the thinker who knows many little things, draws from an eclectic array of traditions, and is better able to improvise in response to changing events--is more successful in predicting the future than the hedgehog, who knows one big thing, toils devotedly within one tradition, and imposes formulaic solutions on ill-defined problems. He notes a perversely inverse relationship between the best scientific indicators of good judgement and the qualities that the media most prizes in pundits--the single-minded determination required to prevail in ideological combat. Clearly written and impeccably researched, the book fills a huge void in the literature on evaluating expert opinion. It will appeal across many academic disciplines as well as to corporations seeking to develop standards for judging expert decision-making.
Anyone who forecasts or does formal planning for a living...
Published by Thriftbooks.com User , 17 years ago
...can't afford to do without this book. It is scary to think that many people will be writing PhDs in the Social Sciences, and then be called upon to make or influence policy without being familiar with this book's central arguments and evidence.
A classic of Political Science & Cognitive Psychology
Published by Thriftbooks.com User , 18 years ago
Tetlock shows conclusively two key points: First, the best experts in making political estimates and forecasts are no more accurate than fairly simple mathematical models of their estimative processes. This is yet another confirmation of what Robyn Dawes termed "the robust beauty of simple linear models." The inability of human experts to out-perform models based on their expertise has been demonstrated in over one hundred fields of expertise over fifty years of research; one of the most robust findings in social science. Political experts are no exception. Secondly, Tetlock demonstrates that experts who know something about a number of related topics (foxes) predict better than experts who know a great deal about one thing (hedgehogs). Generalist knowledge adds to accuracy. Tetlock's survey of this research is clear, crisp, and compelling. His work has direct application to world affairs. For example he is presenting his findings to a conference of Intelligence Community leaders next week (Jan 2007) at the invitation of the Director of National Intelligence. "Expert Political Judgment" is recommended to anyone who depends on political experts, which is pretty much all of us. Tetlock helps the non-experts to know more about what the experts know, how they know it, and how much good it does them in making predictions.
brilliant and encouraging
Published by Thriftbooks.com User , 18 years ago
Why isn't this book on the front page of every newspaper everywhere? The author makes a cogent argument that informed amateurs are as good as "experts" in seriously important fields from investing to politics. He provides experimental evidence that this is the case. This is the hypothesis that a bunch of farmers and merchants tested when they sat down and organized the United States of America. This is the hypothesis that every smart investor proves when he meets or beats hedge funds spending millions on experts and computer power. A very empowering book. Well-written, well-argued, well-referenced. I dare you to read this one and not give it to a friend.
Careful, Plodding, Objective
Published by Thriftbooks.com User , 18 years ago
This book is a rather dry description of good research into the forecasting abilities of people who are regarded as political experts. It is unusually fair and unbiased. His most important finding about what distinguishes the worst from the not-so-bad is that those on the hedgehog end of Isaiah Berlin's spectrum (who derive predictions from a single grand vision) are wrong more often than those near the fox end (who use many different ideas). He convinced me that that finding is approximately right, but leaves me with questions. Does the correlation persist at the fox end of the spectrum, or do the most fox-like subjects show some diminished accuracy? How do we reconcile his evidence that humans with more complex thinking do better than simplistic humans, but simple autoregressive models beat all humans? That seems to suggest there's something imperfect in using the hedgehog-fox spectrum. Maybe a better spectrum would use evidence on how much data influences their worldviews? Another interesting finding is that optimists tend to be more accurate than pessimists. I'd like to know how broad a set of domains this applies to. It certainly doesn't apply to predicting software shipment dates. Does it apply mainly to domains where experts depend on media attention? To what extent can different ways of selecting experts change the results? Tetlock probably chose subjects that resemble those who most people regard as experts, but there must be ways of selecting experts which produce better forecasts. It seems unlikely they can match prediction markets, but there are situations where we probably can't avoid relying on experts. He doesn't document his results as thoroughly as I would like (even though he's thorough enough to be tedious in places): I can't find his definition of extremists. Is it those who predict the most change from the status quo? Or the farthest from the average forecast? His description of how he measured the hedgehog-fox spectrum has a good deal of quantitative evidence, but not quite enough for me check where I would be on that spectrum. How does he produce a numerical timeseries for his autoregressive models? It's not hard to guess for inflation, but for the end of apartheid I'm rather uncertain. Here's one quote that says a lot about his results: Beyond a stark minimum, subject matter expertise in world politics translates less into forecasting accuracy than it does into overconfidence
Great Decision Making Evidence
Published by Thriftbooks.com User , 18 years ago
As both a consultant and an investment manager I have spent a lot of years studying decision theory. One limitation in a lot of the work I encountered was its heavy reliance on lab studies using students. You were never quite sure if the conclusions applied in the "real world." This outstanding book puts that concern to rest. It is by far the richest body of evidence I have encountered on decision making in real world situations. Anybody interested in decision making and decision theory will profit from reading it.
ThriftBooks sells millions of used books at the lowest everyday prices. We personally assess every book's quality and offer rare, out-of-print treasures. We deliver the joy of reading in recyclable packaging with free standard shipping on US orders over $15. ThriftBooks.com. Read more. Spend less.