I'd add one or crazy thing. The intercepted ball had 3 readings with the same gauge.
They were wildly variable as well.
On the same ball.
And we know that Exponent claims gauges only cause a tiny amount of air to leak out.
That seems like a key piece of information, but it's also the only time that the gauges seems to be internally inconsistent. Otherwise, the same-gauge variance would seem pretty small, less than .15 PSI. (I'm getting that because the difference between the gauges is consistently between .3 and .45 PSI, so the same-gauge error is likely less than the difference between those two.) I'd like to know more about the conditions behind those three tests, particularly the timing of them relative to each other and to when the ball was brought inside.
Honestly, the variance is the only part of the physics that I'm having trouble explaining. This article kind of hand waves it away–the author explains that variance will be higher because more time passed when the Pats' balls were being measured and it was during a time when they would have been further from equilibrium and thus gaining pressure faster, but the numbers still don't add up. Min/Max difference in the Pats' halftime readings is 1.35 PSI, which is above the effect of the transient warming curve even if you consider the entire halftime (which certainly wasn't the case).
So, yes, I get the logo vs. non-logo gauges, and I get the warming curve, and I get the difference in initial temperature assumptions. I feel safe in saying there are entirely reasonable scenarios where the average pressure of the Pats' balls can be explained perfectly well by physics. That's average, though, and the variance is harder for me to explain. The only thing I can think of that makes that work is that the initial variance was higher, but that directly challenges Anderson's statement that the initial pressure was consistently 12.5 PSI. I don't have a real problem saying that and think it's plausible, particularly since the initial pressures weren't written down; but it's a different type of argument than the rest of the physics corrections since everything else can be explained without challenging someone's testimony, only their assumptions and models.
The high variance of measurements on the intercepted ball is useful, but not entirely consistent with the measurements of the other balls and thus I'm tempted to say it's a result of something like there being a few minutes between tests of the intercepted ball rather than inaccuracy of the gauges (calibration aside).
Is there a mechanism that's been demonstrated that would account for high variance? Maybe something to do with how long the ball was in play or how wet one ball got compared to the rest?