Mike G wrote:For a player metric to be credible, it has to accurately describe a given season's team success. It's a necessary, but not sufficient, test of validity. Year to year stability is the rest of the test.
No, such test is neither necessary nor sufficient. It is also 2/3 of Dave Berri's "quality check". I can show you an easy example, why that is basically useless:
Code: Select all
Player Tm MP UNO UND UNR
Mike Scott ATL 178 1.94 0.40 2.34
Kyle Korver ATL 444 1.10 0.55 1.64
Thabo Sefolosha ATL 265 1.25 0.20 1.45
Mike Muscala ATL 34 5.27 -3.92 1.35
Kent Bazemore ATL 82 2.39 -1.38 1.02
Dennis Schröder ATL 196 0.96 -0.64 0.32
Al Horford ATL 383 0.48 -0.35 0.14
Shelvin Mack ATL 159 0.59 -1.53 -0.94
Pero Antic ATL 199 0.29 -1.46 -1.17
DeMarre Carroll ATL 308 -0.03 -1.21 -1.24
Paul Millsap ATL 456 -0.24 -1.10 -1.34
John Jenkins ATL 21 6.72 -8.60 -1.87
Jeff Teague ATL 425 -0.46 -1.43 -1.89
Elton Brand ATL 21 6.42 -8.97 -2.55
These are values for the current season. When you take the minute weighted average of UNO and UND multiplied by 5, you get the Hawks ORtg/DRtg above league average. How did I arrive at those numbers? Well, I took the team-level data (in that case ORtg/DRtg) and then devided those among the players using their jersey numbers (using z-scores here) and their minutes played. Given the fact that jersey numbers are mostly staying constant from season to season as well as rather high season-to-season correlation for the minutes for each player exists, such a metric will have a typical year-to-year consistency. I called that UNR, because that is the Ultimate Nonsense Rating.
What people need to understand is where the "summing up to wins/team-level-play/etc." within a season comes from. It is explicitely a result of either starting at a team-level and then subsequentially devide that value among the team's players or making it fit to the team-level play afterwards. That is not rocket science to create such a metric, in fact I created one which then even fulfills Berri's third part: it must make sense. The "it makes sense part" is given by me declaring "scoring plus defense" as "making sense" (similar arbritarily as Berri). Then I simply took the individual player's scoring per 100 team possession rate and added a "small" defensive adjustment term to it. What makes that metric sum up nearly perfectly to wins is the fact that the minuted weighted scoring per 100 poss for individual players multiplied by 5 will give you the team's ORtg. Setting up the "defensive adjustment" in a fashion that the team's defensive prowess (Drtg over league average) will be distributed among the players accordingly. In that way this metric will completely match the ORtg and DRtg for each team. Then we know that this correlates highly with the win% (we can even improve that by simply using a SOS adjustment as well).
What is the point: Wether we start with a team-level value or let the results fit the team-level at the end, doesn't matter, because we can distribute the value among the players on a specific team in every way we want, it will ALWAYS correlate highly with the team win%/team net rating. But that is not a quality of the individual player values. The year-to-year consisteny is mostly guaranteed anyway, because the players will likely stick with their role from season to season and will get similar minutes. What we need to understand here is, that we are dealing with a sampling bias. We are not working with a random sample, but with players who are selected by skilled people (GM, Scouts) and then used in a fitting fashion again by highly skilled people (coaches), while the overall role determines the boxscore entries as well as skills. Being able to separate the skill part from the role part is what a good metric can do, and at that I'm rather sure PER is better than a lot of metrics, which are summing up to the wins/team-level play in-season (or show a high correlation to that).