Re: APBR-DraftExpress 2015 NBA Draft Project
Posted: Sat Jun 06, 2015 2:53 am
Go with whatever is your call within your time budget. There can be after-riffs here or on other sites or next season.
Analysis of basketball through objective evidence
http://www.apbr.org/metrics/
I see your point; I guess I hadn't quite thought it through all the way. To do it correctly, all models should be standardized and the Z scores used to do the combination. If we don't have time, that's OK, but ideally Z scores would be used.ampersand5 wrote:This is not necessarily the case. For example, one system could have player 1 rated 99.7 (1) and the 10th ranked player rated 99.3 (10) while a second system has player 1 rated 23 (40) and the other played rated 74 (30).DSMok1 wrote:
It's an issue, for sure. If one rating system has a player #1 and another player #10, and a second system has player 1 #40 and player 2 #30... should they average the same? Unlikely, there's a bell curve distribution usually. Player 1 should be rated higher, possibly much higher, in an average of the two.
This is ultimately my point - we have no idea what the distribution is of the different ratings. By using lots of models, we are trying to find consensus on where a player ranks.
I understand what you were trying to state - in assuming that there is a (potentially) bigger variation in player scores between middle of the pack rankings (30-40) than at the top (1-2).
I find this to be problematic because the models were created to produce a player ranking system, not to create a uniform player score rating. By way of example -
lets look at how two models could rate the top ten players in the draft
Model 1:
1) 100
2) 99
3) 98.5
4) 98
5) 97
6) 96.5
7) 96.4
8) 96
9) 95.8
10) 95.6
Model 2:
1) 100
2) 92
3) 78
4) 55
5) 54
6) 52
7) 50
8) 49.7
9) 47
10) 45
Any model that has a different distribution of player ratings than the rest is going to through off the entire cumulative average for any of the players that do well/poorly in that model.
Because each model has a completely different way of rating players within its own system, I do not think it makes sense to use these comparative ratings in relation to other models.
In comparing out-of-sample retrodictions to actual draft order, EWP does about as well as NBA decision-makers, while my "HUMBLE" model (which integrates scouting consensus) actually does a bit better than either (https://pbs.twimg.com/media/CGljDbwU8AA_j20.png:large). This is good support for the use of draft models, but I still would not advocate drafting simply based on a statistical model, or even an averaged collection of models. What I would advise is a scientific approach to prospect evaluation, and I think these models provide an excellent foundation to begin that process. This does not mean weighing models against your personal subjective sense, but rather understanding where the built-in assumptions might go wrong in a particular case and challenging them with data or strong logical arguments.
For example... None of the models account for progression across the season, should we give Justise Winslow a subjective bump due to his late-season surge? I don't know, but this is a testable hypothesis. Most models account for net strength-of-schedule, but they don't give special weight to individual performances against the stiffest competition. If they did, it might hurt D'Angello Russell's rating. This is a simple question that remains untested. If Coach K's system historically depresses bigmen's defensive rebounding (it does) it should probably be taken into account when evaluating Okafor. Speaking of Okafor, he and Towns offer contrasting “Old school” vs “New school” styles. Identifying whether and how much this matters might be helped through an analytic approach, but it definitely demands the subjective counsel of scouts who understand the complexity of NBA game-planning... I could go on for days with more examples like these, and that is exactly what front offices should be doing when they debate prospects using statistical outputs.
People who fall into arguments of “well the numbers say this” or “well my years of experience watching basketball say this” will always lose to those who constantly challenge their own opinions with the most rigorous methods available. Using models correctly is a lot of work, and it is a process that integrates follow-up analytic research and subjective observations in equal measure. This requires commitment to understanding both the math and the game. This likely turns a lot of people off on either end, but the wisest path usually is not easy.
If we post z-score standardized numbers, people might also interpret that as positive-negative nba imact with the cutoff point coming somewhere in the early second round, when in fact, the overwhelming majority of any draft class will have negative to no nba impact.vjl110 wrote:BTW... regarding the issue of how to standardize the models, I think rank order is going to be the most intuitive for folks to read. Yes it has problems, but it really shouldn't pervert the outcomes much at all.
I need to update the rankings on the spreadsheet (as found here: https://docs.google.com/spreadsheets/d/ ... =865187672 ) to reflect the changes made by posters + add in statmans rankings.nrestifo wrote:Yes, min-max normalization stabilizes variance, rather than make it equal, which is what z-score does. A 0-100 scale, however, is more interpretable, which is why min-max is often used. Z score will give results like 4.17, 0.56, -1.29. Z score is
(xval-mean(x))/stdev(x)
where xval is the raw value, mean(x) is the mean of each draft model's ranking, and stdev(x) is the standard deviation of each model's ranking. But do whatever you want/have time for, because as Crow said, you're the one doing the work to put it together, which is appreciated. Looking forward to the Monday piece.
If I had all the raw numbers and editing privileges, I'd be happy to do so, would only take me a minute or two, just so people had an idea what it would like.ampersand5 wrote: I don't have the time to do them myself, but if someone wanted to create Z Score rankings, I think be happy to see how they look (and obviously be very appreciative).
The biggest asset of doing this is that if we had Z-score ratings, we could then have all of the international prospects fitted into our rankings/article, which is something that we would all benefit from.
I will have a rough draft of the article up later today.
All of the raw numbers are posted in this thread, so its really just an issue of someone inputting all of them. I just asked Jesse if he already has them in a spreadsheet that he could send us, so I'll see what he says.nrestifo wrote:If I had all the raw numbers and editing privileges, I'd be happy to do so, would only take me a minute or two, just so people had an idea what it would like.ampersand5 wrote: I don't have the time to do them myself, but if someone wanted to create Z Score rankings, I think be happy to see how they look (and obviously be very appreciative).
The biggest asset of doing this is that if we had Z-score ratings, we could then have all of the international prospects fitted into our rankings/article, which is something that we would all benefit from.
I will have a rough draft of the article up later today.
71 Jonathan Holmes 21 (+50) - JF
60 Rakeem Christmas 29 (+31) - LV
74 Julian Washburn 44 (+30) - ME
59 Anthony Brown 33 (+26) - NR
33 Devin Booker 9 (+24) - NR
66 Chris Walker 43 (+23) - SS
57 Norman Powell 35 (+22) - BPM
62 Joseph Young 45 (+17) - JF
46 Michael Qualls 32 (+14) - LV
45 Michael Frazier 31 (+14) - SS
Most "Underrated" by DX according to Draft Model Composite:
28 Wesley Saunders 72 ( -44 ) - NR
31 Seth Tuttle 74 ( -43 ) - ME
48 Derrick Marks 69 ( -21 ) - SS
42 Branden Dawson 61 ( -19 ) - LV
30 Vince Hunter 48 ( -18 ) - LV
44 Chasson Randle 62 ( -18 ) - SS
9 Delon Wright 25 ( -16 ) - JF
26 Terry Rozier 40 ( -14 ) - SS
41 Larry Nance 54 ( -13 ) - ME
5 Kevon Looney 17 ( -12 ) - NR
"Glancing at the boxscore, the gap between Hunter and Washburn as prospects seems huge. As a sophomore, Hunter carried an impressive 30.5 usage% compared to Washburns 18% rate which is unusually low for a senior NBA prospect. Not only that but Hunter scored much more efficiently (54.3 eFG% vs. 48 eFG%) in spite of carrying a much higher load. Hunter also added dominant rebounding as a secondary boxscore filler, while Washburn has no standout numbers. However, I think we can start to see why scouts view them a closer prospects when we try to identify "NBA tools" and realistic role at the next level. Washburn is noted as a strong defender (something that often evades statistical models) and shows some promise as an outside shooter. Ideally he could fit nicely into the traditional 3-n-D role. Meanwhile, Hunter does not appear to have much range or creation ability, and thus seems destined to be an undersized paint 4. That is probably the easiest role to fill in the NBA, and rarely costs more than the veteran minimum. This issue of replaceability can be expanded to other players the models appear to "overrate" as well. Branden Dawson projects as an interesting wing-stopper with limited shooting ability. This skillset can influence team differential, but there are guys like Al Farouq Aminu who have demonstrated mastery of that niche available for cheap every offseason. There is limited incentive to target an undervalued skill in the draft."