Great talk by the way...
Newton's laws of motion are a good example. I could pull others from Statistical Mechanics and other technical fields if I dusted off my textbooks.
Modeling of traffic is another example, where you have thousands of independent, (some) stupid drivers on the road... And yet congestion and bottlenecks can be calculated using relatively simple formulae in some cases.
But here I agree with you: sometimes those models miss the mark. But their value is not in getting every prediction right, or matching reality exactly
In how many games would you say the xG scores are absolutely wrong?
I'd wager that in most situations the xG total adequately represents the quality of chances available to both teams. I don't think it works completely well in the aggregate for a game (taking 30 poor shots worth 0.1 xG doesn't necessarily mean the score should be 3-0) but for the majority of situations it works. And that's what a great model does. If we had a model that was fit to every data point it would be useless for further use.
Are you saying the xG models are wrong based on your own perspective of watching games?
You cannot invalidate a model by single instances that are off from reality. Because the essence of a model is that you are trying to distill a very complicated existence into something manageable on paper that can be manipulated, and in the process of that distillation, some things will be left out that will result in occasional slips, but those slips don't mean the model is invalidated.
And just because 2 models approximating the same event don't yield the exact same answer, it doesn't mean they are both wrong. It means that model construction was different, however as long as the assumptions were clearly stated and the methodology for each model is uniformly applied to the dataset, depending on my motivations I may choose to select one or the other. Or both!
Yes but I promise you that you don't do this consistently.
The reason why we create models is because no human being has watched all the games in the PL this season and evaluated every chance with the same level of intensity and analysis between every shot.
And the stakes in football are too high for managers to rely solely on human memories and judgements in determining how to approach football games. But if we have a tool that can be used to gauge chance quality in advance, based on the systemic processing of all the chance data we have access to... It won't be perfect but it will fit the data well enough.
Ok, fine, no xG model can tell me how good that shot was as well as
@mctrials23 will, 2 rounds of beer in. However can he also break down to me exactly how much the chance quality would change by if the player changes, or 4 players blocking becomes 3, or Kepa is in goal, or it is raining? And also can he use this model to determine exactly where chances should be created on the pitch? He would rightly tell me to feck off
We need to remember that these models are trained on singular events, shots. If the model is good enough (from a statistical POV), not only are they good enough to be useful for single games, they are good enough to be used for evaluating single shot quality. You would be hard pressed to find a shot scored as xG of 0.7 that would not be a clear chance from the average top flight player.
No one is saying take model outputs as gospel. No one should take the eye test as gospel. No one should take my word as gospel.
And I think if the xG models had these serious flaws in accuracy they would have been highlighted in papers and articles and blogs giving more fuel to the "burn math" crowd (not saying you're one of them ha)
How much data would you need? You can get excellent models on as few as 20-30 data points... We have shots data that I'm sure goes into the 10s of thousands. If you can't build a good model on that dataset then give up
That said there is always room for improvement. And I'm sure if we saw what clubs were using behind closed doors we would be blown away...