Jinxed wrote:permaximum,
I have a bit of a philosophical disagreement with your approach of using high roster turnover rate as way of evaluating a player's skill.
A player who plays well with high RT is not a necessarily more skilled player than one who plays well with normal RT, he's just more skilled at playing with high RT. It's analogous to being good at pickup basketball (when any random person can be on your team) vs being good in a league setting with a coach, a defined role, and familiarity with teammates.
These are different skill sets. And I see no reason why we should call the former "the better evaluator of player skill".
Your tests don't show that these simple metrics Tom Thibs/AWS are better than RPM at evaluating player skill, they just show they are better at evaluating which players are better at playing with high RT than RPM.
What you're saying (being good in a league setting with a coach, a defined role, and familiarity with teammates) can be a small factor but I don't believe it's significant. All NBA-level players should fit into different systems well enough. Any impact change
"that's related to player skill" that happens because a given player is in a different role should be very minimal almost to the extend of non-existence.
It's more about their environments instead of themselves. At this stage, player impact can be limited or maximized by outside interference... not the other way around. The effect of player adaptability in such a huge data shouldn't have changed results this dramatically. Actually it shouldn't have changed anything at all. In this research, 38658 games were predicted by metric values which were affected by thousands of players. In core, metrics were in the race. Not the players.
For example, AWS being better than all metrics at higher roster turnover rates mean that the metric predicted game outcomes better than others because it's more detached from teams' synergistic effects on players.
RPM or any +/- related metric try to distribute "TEAM SUCCESS" to players that take part in those games. What those metrics miss is that TEAM SUCCES is not equal to PLAYERS. The equation should be;
TEAM SUCESS = a.P1 + b.P2 + c.P3 + d.P4 +e.P5 - f.P6 - g.P7 - h.P8 -i.P9 - j.P10 + HCA + REST + CONSTANT
CONSTANT = Lots of things but mainly something like Brand Value, Score Margin, Game Time, Timeout, Momentum, Garbage Time etc.
a,b,c,d,e,f,g,h,i,j = Coefficients: Synergistic effect of lineups, roles, coach, team and the city itself on player impacts.
Compared to other +/- metrics, RPM try to adjust for HCA, REST, Score Margin etc. but it can't adjust for those coefficients since it still has to use the core principle of APM. Deduction. These APM based metrics share those synergistic effects and constants almost equally to players on the court which is unfair and completely misleading. That's why some players in good teams that barely pass the eye-test tend to score very well on PM based metrics. Like Nick Collison. Like Amir Johnson. Like every role player in top teams. On top of that, these metrics are biased by priors and effected by multi-collinearity issues of APM.
Although these linear box-score metrics are very simple, they use the method of induction. It's the valid approach and that's why they're better at evaluating player peformances that are isolated from sygnergistic effects. Because they don't capture those effects at all. Although they suffer the consequences of not using an empirical approach and advanced data of today, their advantages still exceed the disadvantages although even they are not good enough considering HCA alone is 0.615. This only proves the fact that current all-in-one metrics' state is pretty poor.
Long story short, evaluating metric performances at very high roster turnover rates is a must if we want to detach players from synergistic effects and these linear box-score metrics don't even have that problem.