A few non-curmudgeonly issues with the “pro analytics” camp.
As many have noted, Week 15 of the NFL helped bring the sport’s analytics debate to an unfortunate place.
Here are a few non-curmudgeonly issues I have with some in the “pro-analytics” camp.
Communicating model estimates without uncertainty
Implying expected value optimization is the only valid decision making paradigm
For environments where it is hard to anticipate every contingency, the expected value paradigm that dominates the NFL decision making discourse has legitimate alternatives. However, anyone who questions the “analytics experts” is frequently cast as being unsophisticated or unable to grasp “the math.”
Let’s look at an example. Just before the half in Week 15 against the KC Chiefs, Brandon Staley faced the following decision.
The “medium, go for it” recommendation is the result of an expected value calculation based on six estimated parameters.
Two “conversion rate” estimates
Four win probability estimates for the possible resulting game states
What happens if we propagate some uncertainty into the estimates that are inputs into the expected value calculation? (Like a less sophisticated cousin of this awesome Shiny app, only that also considers that the Win Probability model could also have small errors.)
Are errors of such magnitude sufficiently charitable to the model? For a toy exercise, I think so. That would seem like an excellent model to me! Sources of such errors could be attributable to injuries, weather, game plan, etc. that couldn’t be captured by a model.
Some use an expected win probability advantage of at least 1 WP as a threshold to judge when a decisions is “correct.” Assume we accept the expected value paradigm (i.e., ignore my second complaint at present).
I’m not suggesting I’ve done this optimally (or perhaps even acceptably) for this example, but I think it is important to express uncertainty with estimates. In the future, I hope the “pro analytics” side of the debate finds a way to communicate uncertainty with their estimates.
I highly recommend Gary Klein’s Sources of Power. Klein studied how individuals in complex domains (e.g., emergency responders, combatants) actually make decisions. The result was the (descriptive) Recognition-primed Decision (RPD) model.
Here’s my rough summary of the findings relevant to NFL coaching decision discussion.
Here’s a recent example of “expected value optimization is the only valid decision making paradigm” in the analytics debate.
I see a lot of Klein’s descriptive model of expert “naturalistic decision making” from Coach Belichick, including critical reflection on previous decisions to learn like an expert. If I were a Pats fan, I would not be discouraged.
Coaches are trying to find the most probable single trajectory to victory; the expected value of a strategy will not be realized for a single decision. Additionally, it is very plausible to me that an expert coach could identify valid reasons – unique to the specific game situation – why an expected win probability model may not apply perfectly for the specific decision being made.
There is a lot of serious academic and other work that support the view that there are alternatives that can be superior to expected value optimization in some settings. Their discussion appears entirely absent from the slice of the analytics debate to which I am exposed.
If I were to anticipate criticisms, it would be something like the following: win probability models are now of sufficient quality that NFL games are now “smallish” enough words that no biased human expert can systematically be superior to the model. I’m skeptical.
A few other book recommendations that explore alternatives to “rational” optimization in the applicable contexts: