Menu
Sign In Pricing Add Podcast
Podcast Image

3 Takeaways

The Science and Skill of Superforecasting (#230)

Tue, 31 Dec 2024

Description

Imagine how much better your decision making would be if you could better predict the future. It’s possible, with superforecasting. In fact, a team of superforecasters won a forecasting tournament conducted by the U.S. intelligence community. What do superforecasters actually do, and how can you become a better forecaster? Don’t miss this talk with superforecaster, and CEO of Good Judgement Inc, Warren Hatch.

Audio
Featured in this Episode
Transcription

Chapter 1: What is superforecasting and why is it important?

35.135 - 64.415 Lynne Thoman

I'm Lynne Thoman, and this is Three Takeaways. On Three Takeaways, I talk with some of the world's best thinkers, business leaders, writers, politicians, newsmakers, and scientists. Each episode ends with three key takeaways to help us understand the world and maybe even ourselves a little better. Today, I'm excited to be with super forecaster and CEO of the Good Judgment Project, Warren Hatch.

0

65.456 - 96.508 Lynne Thoman

In 2005, the University of Pennsylvania's Philip Tetlock published a study showing that experts performed about as well at making predictions as what he called dart-tossing chimpanzees. and those who were surest of their predictions did much worse than their humbler colleagues. The study caught the eye of the United States intelligence community, which set up a geopolitical forecasting tournament.

0

Chapter 2: How did the Good Judgment Project win the forecasting tournament?

97.148 - 127.553 Lynne Thoman

The undisputed winner of the tournament was the Good Judgment Project, which was led by University of Pennsylvania professors Philip Tetlock and Barbara Mellers. Over four years, their forecasters answered 500 questions and made a million forecasts that were more accurate than even intelligence analysts who had access to classified data. My guest today is Warren Hatch.

0

128.013 - 156.011 Lynne Thoman

He's the CEO of the Good Judgment Project, which is the group of forecasters who won the U.S. intelligence community's forecasting competition. And Warren is not only the CEO of the Good Judgment Project, he's also one of their top super forecasters. I'm excited to find out from Warren how super forecasters forecast and how we can all get better at forecasting.

0

156.831 - 161.193 Lynne Thoman

Welcome, Warren, and thanks so much for joining Three Takeaways today.

0

162.073 - 164.794 Warren Hatch

Thank you, Lynn. It's a pleasure to be here. Thanks for having me on.

0

165.775 - 172.377 Lynne Thoman

It is my pleasure. Warren, let's start with why is forecasting important? Where do we use it?

Chapter 3: What are the common pitfalls in traditional forecasting methods?

173.348 - 191.24 Warren Hatch

In a sense, every decision that we make is a forecast because we're going to be taking action and doing things to improve the odds that we're going to get our desired outcome, whatever that might be. So we're doing that all day, all week, all year long, or thousands of decisions that way. So we want to make the best possible forecast.

0

191.381 - 198.946 Warren Hatch

And super forecasting is a process to get to the best possible forecast and therefore to get to the best possible decision.

0

199.793 - 203.817 Lynne Thoman

And how do most people forecast and what's wrong with their approach?

0

204.778 - 225.869 Warren Hatch

Well, most people, and it's the dominant way, is they'll use language to express their views about the future. Somebody will ask them, well, do you think this will happen? And they'll say, well, maybe, maybe it will. Or they'll use fancier words like, well, there's a possibility or a distinct possibility. Now, here's the problem with that, many problems.

0

226.589 - 250.648 Warren Hatch

One problem is we're all going to understand that in different ways. There is a famous example when Kennedy came into office and inherited a plan to invade Cuba and topple the regime. He asked his advisors, will this succeed? And they said, there's a fair chance it will succeed. Now, it turns out Kennedy had in mind north of 50%. The analysts had in mind something more like 25%.

252.495 - 274.475 Warren Hatch

By using language, there was noise in that decision process. I imagine if Kennedy had known they had in mind 25%, history might have been a little different. So that is true for all kinds of words like that, where we're going to interpret it differently. Another bad thing about it is that it straddles for the 50-50 line. So if it happens, you say, aha, I said there'd be a distinct possibility.

274.935 - 296.196 Warren Hatch

If it doesn't, well, I only said it was a distinct possibility. One other bad thing about it is it's impossible to put different views together. You can't crowdsource language like that because we're all using different words and in different ways. How much better to use a number? We all know what 72% is. We all know what that means.

297.176 - 317.985 Lynne Thoman

For big events like market crashes or looming wars, or for policy decisions like tax cuts or sanctions or tariffs, we turn to the experts, those people that are supposed to be the most knowledgeable. How do experts do as compared to super forecasters?

Chapter 4: How do experts compare to super forecasters in predictions?

319.025 - 344.43 Warren Hatch

That's a great question, because sometimes there's an apparent tension between experts and good forecasters. where it's one is better than the other. We really are of the view that you want both. You want hybrid models when it comes to experts and things like that. And one core reason for that is experts might be good forecasters. But we don't know that until we see their track record.

0

344.79 - 361.084 Warren Hatch

Just being an expert does not make you a good forecaster. We want to see if when you say an 80% probability of something occurring, it occurs eight times out of 10 and doesn't two times out of 10. That's accuracy in a probabilistic sense. Most experts do not go through that process.

0

361.504 - 380.391 Warren Hatch

They're very good at telling us what we need to understand, how we got where we are, some of the things we might want to watch. going forward. But what we've seen time and again, when you ask them for a probabilistic forecast, experts generally tend to assign higher probabilities to events in their area of expertise than actually occur.

0

381.067 - 404.081 Warren Hatch

And so you're better off by going to a crowd who are very skilled at assigning probabilities that occur with that frequency to give a forecast for that particular thing. Now, best of all is if you get an expert who applies themselves and becomes a good forecaster, that's what we really want to see. But experts, not necessarily good forecasters, just by being an expert.

0

405.029 - 417.048 Lynne Thoman

So the first part of a forecast is really the question. Can you give some examples of what good forecastable questions should be or are?

418.052 - 435.24 Warren Hatch

Oh, I love that. That's a great question. And that's half the work is getting the question right. You want to make sure that everybody understands it in the same way. You want to be sure that it's actionable, it's useful. Why go to all this effort if it's just, you know, a parlor game?

Chapter 5: What makes a good forecasting question?

435.7 - 454.471 Warren Hatch

And you also want to be able to say, well, when that date occurs and you look back, we will agree it happened or it didn't. And a lot of forecast questions that are out there don't meet those criteria. There was a great example because in the original part of the project, the government wrote the questions themselves.

0

455.172 - 478.36 Warren Hatch

And in that first year or so, they had questions of the sort like, will construction begin on a canal through Nicaragua? And this was back when there was a guy in Hong Kong who was going to build a big, giant canal for the Supermax boats. And will construction begin there? was the question that they posed. Now, if you reflect a little bit, you can see that there's a problem with that.

0

479.04 - 499.914 Warren Hatch

What counts as construction? Is it built with boats going through? Or is it big roads going through the jungle? Or is it a golden shovel in the ground? Which was my view. And sure enough, there was a golden shovel in the ground. But that was it. So did construction begin? We need to agree on what the question's asking. So the head of our question team

0

500.686 - 522.362 Warren Hatch

The fail rate we have in our questions now is like 0% this year, thanks to his hard work. And what he'll do is in addition to writing a careful question, he'll include, it's almost like a little contract. He's trained as an attorney where this is how the question will resolve under these circuits. And for things that matter, it's worth going to that effort.

0

522.663 - 534.26 Warren Hatch

Not everything matters that much, but for things that do, then you want to make sure it's crisp and tight. It's actionable, it's verifiable, and we all can agree if it happened or it did not.

535.221 - 544.506 Lynne Thoman

So if somebody were wanting to use super forecasting in their personal life, can you give some examples of how to frame questions?

545.547 - 556.293 Warren Hatch

For the higher order questions, things that really impact our lives, it's often not just a single decision. There might be a series of smaller decisions that go into it.

556.963 - 584.387 Warren Hatch

and so breaking that down into what are the pieces that are really going to be impactful for how i think about this issue so where to go to school that's more than just a single decision there's a series of them do they provide the support that i would want as a student do they have the right kind of curriculum what sorts of career trajectory might i get on those are all smaller sub decisions and those are all things that you can then analyze as a forecasting question too

584.967 - 607.878 Warren Hatch

Now, one thing that people tend to do, both when they're thinking about the question as well as the forecast, is focus on the particulars themselves. Like who's going to win the next presidential election? Most people immediately start thinking about the candidates. And what we want to do is instead of going to the particulars right away, we want to zoom out. How do things like this usually go?

Chapter 6: How can super forecasting improve personal decision making?

646.992 - 651.193 Lynne Thoman

What's the first thing that super forecasters do that's different?

0

652.028 - 668.995 Warren Hatch

That's actually step one is defined. Another word for that is base rate. How do things like this usually go? And by going to a base rate or outside view, that was basically synonyms, you'll get an immediate boost just by doing that compared to the rest of the crowd.

0

669.855 - 673.077 Lynne Thoman

What are the other steps to get better forecasts?

0

674.117 - 699.146 Warren Hatch

Another good one for anyone to have on their checklist is to make a comment. When you've come up with an estimate about the future, jot down your rationale. One reason for that is it crystallizes your thinking. It also allows you to, in the future, look back and say, was I right for the right reasons or did I miss something? And it also allows you to share information with others.

0

699.686 - 709.832 Warren Hatch

And that's how you get a high quality forecast from a crowd. Most wisdom of the crowd approaches. You ask everybody. Then you take an average and you're done. That's where we start.

709.952 - 734.923 Warren Hatch

And what we'd like to do is have everyone make a comment and then exchange those ideas and see if somebody had a point that maybe I missed and then make an update, which is the next very crucial thing is expect to change your view, make an update. And you just do those few things. You're going to end up with a better number. Start with a base rate, make a comment, exchange views, update.

735.819 - 742.861 Lynne Thoman

Warren, what have you learned recently from forecasting and events in the real world, if you will?

743.841 - 770.117 Warren Hatch

One area where we learned a lot was during the pandemic. We were talking about experts earlier. And one thing about experts as well as artificial intelligence is they have models. That's the way they work. And if you have a model, then you're using backward-looking data to build your model. And that works great when we're in moments of relative equilibrium.

770.698 - 792.661 Warren Hatch

But when things get upended and there is a lot of flux, those models don't work so well. And almost by definition, the experts are going to be slower to recognize that because their models will be filtering out the subtle changes that are eroding their models. And an excellent example of that was when COVID began to go global.

Chapter 7: What techniques can enhance forecasting accuracy?

856.866 - 863.887 Warren Hatch

That's where this kind of a process can really shine is when things get upended, when there's more uncertainty than we recognized.

0

864.867 - 876.93 Lynne Thoman

So when there are big unexpected events in the world, black swans, if you will, is when your approach of the super forecasting really shines?

0

877.938 - 899.188 Warren Hatch

it's a process that can be useful in everyday decision-making, but I think where it really stands out relative to other ways of thinking about the world is when there is a lot of flux and maybe not so much dark black swans as very dark gray swans. Because some of these things may be small probability, but high impact.

0

899.308 - 921.643 Warren Hatch

And somehow getting those small probability items on your radar can help you be better prepared. And in the case of the pandemic, there were some people who were identifying that as a possibility, not something that they were attaching high probabilities to, but it was definitely on their radar. And I think identifying those kinds of dark gray swans

0

922.105 - 943.694 Warren Hatch

is part of a good, successful forecasting process. But once it hits, you're absolutely right. Now it's here. Now what do we do? Well, one thing we should do is discount how much reliance we have on the static models. We don't want to get rid of them completely, but the importance and the reliance we've previously had on them should be discounted.

944.272 - 956.377 Warren Hatch

And we should instead factor in more weight to probabilistic judgment from people who are recognizing the small subtle factors that will eventually lead to new models.

957.337 - 961.599 Lynne Thoman

And can you give some examples of what you call dark gray swans?

962.799 - 989.464 Warren Hatch

One is, yeah, another pandemic could be a dark gray swan, wildfires. has become such a thing where the frequency is much higher than it was. But where are they going to hit? That's the challenge. I think a lot of the AI developments can also fall into that category, what some of those risks might look like. And also possible global conflicts. One that we're looking at more is in Korea.

989.724 - 1013.851 Warren Hatch

We started looking at this a couple of months ago because there have been some changes. There have been some changes to North Korea's doctrine. North Korea sent troops to fight on Russia's behalf around Kursk. These are, in the global scheme of things, small but significant. So we're paying much more attention to how events on the Korean Peninsula may unfold as a possible dark, gray swan.

Comments

There are no comments yet.

Please log in to write the first comment.