You are viewing a javascript disabled version of the site. Please enable Javascript for this site to function properly.
Go to headerGo to navigationGo to searchGo to contentsGo to footer
In content section. Select this link to jump to navigation

You can beat the “market”: Estimating the return on investment for NHL team scouting


Scouting is a major part of talent acquisition for any professional sports team. In the National Hockey League (NHL), the ‘market’ for scouting is set by the NHL’s Central Scouting Service which develops a ranking of draft eligible players. In addition to the Central Scouting rankings, NHL teams use their own internal scouting to augment their knowledge of eligible players and develop their own rankings. Using a novel statistical approach we show in this paper that the additional information possessed by teams provides better rankings than those of Central Scouting. Using data from the 1998 to 2002 NHL drafts, we estimate that the average yearly gain per team from their internal scouting is between $1.7MM and $5.2MM. These values are consistent across the three measures of player productivity that we consider: cumulative Games Played, cumulative Time On Ice and cumulative Goals Versus Threshold where we aggregate these metrics across the first seven years post draft. We used this time frame since teams generally retain rights to their draft picks for seven years. Further, we find that no individual team outperformed the others in terms of draft performance.

1Introduction and background

The Edmonton Oilers selected Jordan Eberle, a center, with the 22nd pick in the 2008 National Hockey League (NHL) Entry Draft. Three picks later, at 25, the Calgary Flames selected another center, Greg Nemisz. This was not the obvious order of things at the time. Going into the draft, the NHL’s Central Scouting Service (CSS) ranked Eberle as the 33rd best North American skater while Nemisz ranked 22nd in that same category; in other words, the league’s own amateur scouts ranked Eberle as a second-round pick at best, even if you ignored the dozens of North American goaltenders and Europeans available. Since the draft, Eberle has played 356 NHL games, while Nemisz has played in 15. The Oilers presumably had additional information from their internal scouting staff about why to pick Eberle over Nemisz, as well as other centers that were available with the 22nd pick, including Daultan Leveille, who has yet to play in the NHL, and Derek Stepan, who has played more than 400 regular season and playoff games and represented the United States at the 2014 Olympics.

Sometimes the information teams possess misleads them. Infamously, the New York Rangers used the 12th overall pick in 2003 to select Hugh Jessiman, a right wing from Dartmouth ranked the 20th-best North American skater by CSS. The Rangers preferred Jessiman to a virtual all-star team’s worth of talent (Dustin Brown, ranked the No. 2 North American skater by CSS, along with Brent Seabrook, Zach Parise, Ryan Getzlaf, Brent Burns, Ryan Kesler, Mike Richards and Corey Perry). Jessiman has since played two NHL games, and played seven games in the 2014-15 season with the Capitals, albeit in Vienna, not Washington.

CSS was formed by former NHL general manager Jack Button in 1975 as a service to NHL clubs to scout amateur prospects and later administer the NHL’s annual combine, which invites the top 100 prospects to come to a centralized location for physical examinations, light drills, and interviews with teams. (Malloy, 2011). Other leagues have had some form of centralized scouting as a cost-saving service to their member clubs, such as the Major League Baseball Scouting Bureau, begun in 1974, or the National Football League’s dueling BLESTO and National Football Scouting organizations (dating to the 1960 s, along with a third, the now-defunct Quadra).1 A significant part of CSS’s appeal to the league is the media attention its rankings generate, as no league has placed such a public emphasis on its scouting service’s rankings as the NHL, which publishes various CSS lists several times per year to great fanfare. The CSS employs nearly 30 scouts in the field, more than any team’s in-house department, including six European-based scouts via Goran Stubb and his Finland-based European Scouting Services. (Morreale, 2011; Shoalts, 2010). Each NHL team has available to them information about possible draftees from the CSS as well as information compiled by their own staffs. The outcomes of draft decisions like those of Edmonton and Calgary discussed above are full of variation— more than half of all players selected in the NHL Entry Draft never play a single game in the NHL.

In this paper, we look at the quality of rankings by the CSS and draft order by team so as to evaluate and quantify the value of the additional information that teams have. In doing so, we estimate the average annual return that teams get from their internal scouting. This is analogous to an approach used in evaluating the success of portfolio managers in finance— the search for “alpha”, or risk-adjusted returns in excess of those readily available in a market index. (Jensen, 1967). Were the “index”, in this case the freely available CSS rankings, comparably successful at picking talent after adjusting for the cost of teams running their own scouting departments, it would endorse a more passive approach to talent evaluation, similar to that of John Bogle when he launched the first index mutual fund at Vanguard in 1975. If nothing else, the CSS can serve teams as “a naíve model against which their in-house gunslingers can measure their prowess” as Paul Samuelson (1974) wrote of a market index fund and money managers. The data that we analyze here is from the five NHL Entry Drafts from 1998 to 2002. For each player selected we have their selection number, their position, time on ice (TOI), the games played (GP) in the NHL, the ranking by player type by CSS and their career goals versus threshold (GVT). GVT is an attempt to quantity the contribution of a given player similar to baseball’s VORP (Value Over Replacement Player). Teams have the rights to players for at least the first seven years after they are drafted.2 Consequently, we focus our analysis on player performance during this period after a player is drafted. Below we use the total of each metric for the first seven seasons after they were drafted. Since for all of the players in our sample this includes the 2004-5 lockout season, we use eight total years to capture seven seasons worth of data for each player. These data were compiled from,, and We use the CSS final rankings released prior to the draft. These rankings were obtained via contemporaneous media accounts accessed via Internet searches and the Lexis/Nexis database. A player’s position was recorded as either center (C), defensemen (D), forward (F), goalie (G), left wing (L) or right wing (R). In the analyses that follow we categorize players as forwards (F) which includes C’s, L’s, and R’s, defensemen (D) or goalies (G). The selection number for a player is the place in the draft order that they were selected. That is, the 10th player selected will have a selection number of 10, the 100th player selected will have a selection number of 100, etc.

CSS ranks players by category, either North American or European, and by whether they are a skater or a goalie. Thus, CSS produces four separate rankings without a correspondence between them for comparison purposes. To better utilize the CSS rankings, we employ Iain Fyffe’s “Central Scouting Integratinator” (CESCIN) metric, Fyffe (2011). CESCIN takes the rankings of players by CSS within their given category and multiplies them by a factor based upon historical draft selection records. The CESCIN multiplier is derived by taking the ratio of the expected number of players taken from each of the four categories to the total number selected in a given draft. See Fyffe (2011) for additional details. For players who were drafted but not ranked by CSS, we give those players values of CESCIN that was larger than the maximal values produced by the original CESCIN. Next, we take the values generated by CESCIN and ranked them to produce the CSS orderings we use in the rest of the paper. Three response metrics are considered below: time on ice (TOI), games played (GP) and goals versus threshold (GVT). GVT is a metric of player value created by Tom Awad that allocates value in team performance among the individuals on a given team, Awad (2009b). The units for GVT are goals so that a GVT of 7.5 credits a player with producing 7.5 goals over replacement level. Since players can have negative GVT, we give players who never played in the NHL a value of GVT below the lowest value of GVT in our database. Our focus is on the top 210 selections since that is the length, over seven rounds, of the eight most recent NHL Entry Drafts (2008 to 2014). Thus, for each year we have 210 observations except for 20023.

For the data that we are considering, there were 595 were forwards (F’s), 332 as defensemen (D’s), and 122 as goalies (G’s). For GP, 54% of the players selected never played a game in the NHL. Table 1 has statistical summaries of TOI, GP and GVT. For GVT, there were no values of that metric for players that did not appear in an NHL game. Among those who did play at least one game the worst GVT that was calculated was – 27.7. For completeness and comparability, we choose to give those who did not appear in an NHL game a GVT of – 30. We select this value to be below the other values in our data and to permit a more complete analysis. For goalies, we gave them 20 minutes of TOI for every game in which they appeared. Again this was done to facilitate comparison across positions. We considered alternatives to the 20 minutes for goalies as well as weighting minutes for defensemen differently than forwards; however, the results presented here were not materially changed by these alternatives. To gauge team preferences for players we use the selection at which the player was taken. For assessing how CSS ranked players we use CESCIN values.

There has been some previous work on evaluation of NHL draft picks. Much of this work has focused on the value of an individual selection. Johnson (2006), Tango (2007), Awad (2009a), Gregor (2011), Schuckers (2011) and Tulsky (2013) have all looked at methods for evaluating what an individual draft pick is worth in terms of a measure of value which, in turn, builds on the work of Massey and Thaler (2005) and Borghesi (2008) in football, among others, and, more generally, Scully (1974), among many others, in baseball, over the past four decades. Many of the critiques of team drafting are that for a given selection a player whose future performance exceeded the current selection was often available. See, for example, Tingling (2011). Given the difficulty with projecting the future performance, it is important to focus on the trends rather than individuals. Clearly there is monotonicity in average performance of players versus draft selection and clearly there are long tails to the distributions of player performance for a given selection. The focus of our analyses is the value that teams get from their scouting departments. It is difficult to quantify the value of CSS since teams have information both from CSS and from their own internal scouting departments. Serge Savard, a Hockey Hall of Fame player who won two Stanley Cups as a general manager for the Montreal Canadiens, acknowledged that the CSS rankings heavily influenced their own, and the difficulties of comparing talent across multiple leagues, even within Canada:

quoteWe were wrong on the first round maybe 50 percent of the time. That’s mainly because of Central Scouting. When Central Scouting comes out with their first-round list, all the scouts think, “Oh, Christ, I better get this player in my list or I’ll look bad.” [All the scouts’] lists are similar because of Central Scouting. I only had one guy, Rick Taylor, who didn’t care about Central Scouting’s list and his list was so different than the others.... How come we missed Luc Robitaille? One of my scouts, Rick Taylor, had Luc Robitaille [rated to be drafted] in the first round and nobody else had him in the top five rounds. The other scouts down-played Taylor. They said, “You only see Quebec. You don’t see Ontario. You don’t see the West. You don’t see college. You don’t see Europe.” So scouting is a tough thing to do. (Farris, 2011).

We propose that the CSS represents a suitable, if crude, proxy for a benchmark index such as the Standard and Poor’s 500. Teams could, theoretically, get rid of their amateur scouting department and rely on CSS’s rankings, as it typically ranks more than the 211 players drafted.4 Below we look at the difference in how teams rated players and how CSS rated players to get an idea about the value added by team scouting staffs.

The number of scouts, and thus costs of scouting departments, employed by teams varies considerably. For example, prior to the 2013-14 season the New York Islanders had 11 individuals with scouting responsibilities while the Toronto Maple Leafs listed 23 on their respective webpages. A spring 2014 survey of team websites included a range of 7 to 24 scouts, with little correlation between market size and staffing (for example, Chicago and Los Angeles each had 14 scouts, while Winnipeg had 24). (Wage, 20145). One small market U.S.-based team estimates that they spend approximately $2 million on their annual scouting budget. This is consistent with the Phoenix Coyotes’ 2009 income statement disclosed as part of its bankruptcy, which included a line item of $1.4 million for “scouting operations”, presumably encompassing both amateur and professional scouting, (In re Coyotes Hockey LLC, 2009). The Phoenix Coyotes’ media guide for the following season (2009-10) listed a 10-member scouting department, including management. Teams have attempted heavy cost-cutting measures, such as the Buffalo Sabres’ 2006 overhaul of its well-regarded scouting department to heavily emphasize video scouting, (Joyce 2008). Other teams, both smaller- and larger-budget, have viewed scouting budgets as a competitive advantage. The New Jersey Devils, traditionally a lower- revenue team, have regularly employed more than 20 scouts, while Brian Burke increased the large-revenue Maple Leafs’ scouting budget upon taking over the team in 2009, attempting to “exploit that advantage” that there are no league-mandated constraints on scouting budgets, (Shoalts 2010). Similarly, upon his purchase of the Sabres in 2011, new owner Terry Pegula saw scouting budgets as a point of competitive advantage for larger-budget teams: “There is no salary cap in the National Hockey League on scouting budgets and player-development budgets.” (Klein and Hackel, 2011).

In the rest of the paper we begin by considering how often CSS rankings and team draft order were able to optimally or nearly optimally identify and select the best player at a given selection. We find that teams outperform the CSS rankings. Next we consider a non-parametric LOESS regression following Schuckers (2011) of our performance metrics onto the player orderings, both CSS and team. This approach also finds that, on average, teams outperform the CSS. Finally, we consider a novel approach that looks at the relationship in the rank orderings and relative value of our performance metrics. Since this final approach is conditional at the individual level, it is the most relevant and informative. We then calculate the excess value above the CSS rankings that teams get in terms of GP and GVT and these are roughly $4 million dollars of “alpha” per year.

2Quality of central scouting draft order

In this section we assess the ability of CSS to correctly order the possible draft selection based upon CSS’s own ordering of draftees. To evaluate this we look at the percent of times that the ordering by CSS as reflected in CESCIN resulted in the optimal ordering at a given position, either C, D, F, G, L or R. Note that this differs from the way the CSS ranks players which combines centers, forwards, defensemen, left wings and right wings into skaters. We also consider the percent of times that the CSS came within approximately one-half standard deviation of the optimal choice at a given position. We determine if a selection was optimal or nearly optimal by considering all of the remaining draftees in a given year at the same position as the selected player. If the selected player had the highest metric among all other available players, then that player was considered the optimal selection. To be nearly optimal the player had to be within half of a standard deviation (SD) of the highest metric for all other available players. For GP, the standard deviation (SD) of the players taken in the first 210 selections was 215 games, while the standard deviation of those same players for GVT was 21. Table 2 has the results of this analysis for TOI, GP and GVT. Overall it is clear that team ordering (based upon the actual draft) is better than CSS ordering. Simply choosing the best available player (optimal ordering) happens about 8% of the time, on average, using CSS and about 14% of the time with team ordering. Ordering players so that the current selection is nearly optimal happens about an average of 14% of the time for Central Scouting. For teams this latter value is about 26% . A further analysis of these data indicates that the advantage for team ordering is persistent across rounds. These results can be found in Table 3.

3Comparison of average performance by player ordering

Above, we concentrated on the optimal or nearly optimal decision at a given selection. We next look at the impact of these selection criteria on the average outcome variables per draft selection. To evaluate this impact we examine the relationship between our response metrics (TOI, GP and GVT) and player ordering. As before, ordering for CSS is done based upon CESCIN while ordering for teams is from the actual draft selections. To estimate these relationships we use LOESS regression as was done in Schuckers (2011) for National Football League data. LOESS regression is a flexible non-parametric methodology for locally smoothing the response at each value of the predictor (selection). In the leftmost graph of Fig. 1 we have plots of the LOESS regressions for predicting TOI based upon the draft selection order from CSS (blue) and from the actual draft (red). Our expected or predicted values from both of these curves are very similar. Both start at roughly the same value for the first selections and decrease steeply until approximately the 40th selection, then less steeply until the 120th selection after which they are roughly flat. The difference between the two curves is that the draft selections outperform the CSS rankings from about 40 to 100. Currently this corresponds to roughly the early second round to the end of the third round. These are the locations where, on average, team selections are better than the ordering from Central Scouting. Central Scouting does better than teams over the last 35 selections which might be due to teams taking a more risky drafting strategy at that stage. Moving to GP and GVT, we see a similar shape to the smoothed prediction relationship that we found for TOI. Although somewhat hard to judge due to the different scaling of the y-axes, the overall shape of these curves is quite similar.

4Rank differential comparison

While the above analysis gives a sense of how CSS and team scouting group perform in their rankings on average, those comparisons and rankings do not condition on some important factors such as the individual player and their position. To account for these we next consider an analysis that looks at the differential between each player’s actual selection and their CSS ranking. Below we will refer to this as rank differential (Δ rank). For example, Rico Fata was the 6th overall selection by the Calgary Flames in the 1998 draft. CESCIN has Fata as the 13th ranked player. Consequently our rank differential for Fata would be -7 meaning he was taken seven places ahead of where CSS ranked him. Players with negative rank differential were taken earlier than CSS ranked them and players with positive rank differential were taken later than CSS ranked them. 56% of selections were rank differential positive, 43% were negative and 1% were zero. If team scouting does well then we should expect that players who have negative rank differential will also outperform what we would expect based upon CSS ranking of them and thus have positive metric differential (Δ metric), and vice versa for players with positive rank differential. In general, our approach for the ith draft player can be stated as:

where f metric is a smooth continuous function to be estimated separately for TOI, GP and GVT.

For TOI, GP and GVT we look at what each player achieved relative to what we would have expected from them based upon their CSS ranking and compared that to what we would have expected based upon their actual draft selection. To calculate the expected values we use the LOESS regressions from Fig. 1. Figure 2 has a plot of each player’s first seven season’s total TOI minus the Expected TOI for each player plotted against their rank differential as well as the same plot for GP and GVT. If team scouting was perfect, there would only be players in the upper left and lower right quadrants of these graphs. Players in the upper left are those that exceeded CSS expectations and were taken earlier than CSS had them ranked. Players in the lower left are those that under performed CSS expectations and teams drafted them earlier than CSS had them ranked. Each graph in Fig. 2 also contains a green curve that is the estimated smoothed f metric . It is important to note that all of these curves are negatively sloped and roughly pass through the origin (0,0). This suggests that when teams differentiate from the CSS ordering they are, on average, gaining some value in terms of TOI, GP or GVT. To estimate the average net values that team scouting contributes above and beyond the CSS, we estimate the TOI, GP and GVT gained at each rank differential for all players. We refer to this quantity as the Average Value over Prediction (AVP). That is, we calculate:

AVPmetric=[irmetric(Δranki)1(Δranki>0)-irmetric(Δranki)1(Δranki<0)]/[Total number of selections]
which represents the estimated average value gained on each metric per draft pick. We then scaled the average values of each to seven selections per season which is the average number of picks that a team has in the current NHL Entry Draft. This quantity, AVP, then represents the per-team, per-draft average gain from the information generated by internal scouting. We next use the estimates of the value of a minute on ice, a game played and a goal to estimate the average annual revenue differential due to team scouting for each of our three metrics. Dorish (2011) estimated that the average NHL salary per game was $29,300. Assuming an average of 20 minutes per game, we find that teams generate approximately $1.7 MM using TOI. Likewise, we can use that same figure to say that in terms of GP team scouting adds approximately $2.5 MM with each draft. Using 20 minutes per game as an average, we estimate that teams gained an average of approximately $1.7 MM in value from their internal scouting. Using GVT, we estimate that team scouting gains a team about $5.3 MM in value using the metric that a goal in the NHL is worth approximately $0.33 MM (Vollman, 2012). A further analysis by position (including the estimation of f metric by position) yielded similar relationships and results for each of these three metrics. In the aggregate, teams would produce between $51 and $159 million in value for their teams by this measure per year; it is unlikely that the aggregate teams’ amateur scouting budgets approach even the low end of these figures. In short, by actively scouting young players, teams generate material excess return over what they would they would have generated had they simply passively followed CSS’s recommendations.

We also analyze the value gained over these drafts by team. There are some winners and losers but the average gains per team per draft pick across these five years did not differ significantly from what would have been expected by chance (Shapiro-Wilk p-values: p >  0.1 for all). We use an average here since Atlanta, Columbus and Minnesota entered the league between 1999 and 2000 and participated in fewer drafts and, hence, had fewer draft selections than the other teams. No teams were outliers in this analysis. Further, we looked at the correlation in team average return over CSS for 1998– 2000 and the team average return over CSS for 2001-2002. The correlations were slightly positive, r  0.2 for all three metrics, but not significantly so (p >  0.1 for all). No team over this period outperformed the others and no team got a significantly larger or smaller average return from their internal scouting. We further have no way to estimate each team’s scouting budget, emphasis on particular leagues or countries, or draft strategies based on positional need or best available talent.


In this paper we have looked at the impact of team scouting compared to the NHL’s CSS. The metrics for performance that we have used are TOI, GP and GVT for the first seven seasons of a player’s career. We began by looking at the chance that the ordering by CSS and by team scouting resulted in choosing optimal or nearly optimal players. At a given selection, teams were not substantially better at picking optimal players with respect to TOI, GP, or GVT but were markedly better at selecting players within a half standard deviation of optimal. There is a clear benefit here to the additional information that team’s possess relative to the NHL CSS, even when accounting for costs. To quantify the value of this information we presented a novel methodology that looks at the expected differential from the two rankings relative to the difference in the rankings themselves. This new and comprehensive approach can be applied to any sport and any method for ordering draft selections and evaluating the performance of players. It is clear from this analysis that individualized NHL team rankings consistently outperform CSS rankings.

This should be expected since teams use considerable resources to gain further information about each potential draftee. Of the three measures (TOI, GP, GVT), we prefer TOI as it is the most direct of these measures for assessing the value that teams give to their draft choices. For TOI, GP and GVT, we find that the average value that a team gets from their scouting is between $1.7 MM and $5.2 MM per year. This range represents the average return that teams get on the total amount that they spend on scouting. It also represents a set of bounds on the amounts teams should budget for internal scouting. Further there was variability between teams in their average return from drafting; however, that variability was not beyond what would have been expected by chance. This means that there is not strong enough evidence to suggest that drafting quality is different across teams, at least for the five years that we considered. A larger sample over a longer period of time could provide additional evidence of differences amongst teams.

This analysis could be improved by having better access to team scouting budgets (or proxies thereof) and, most importantly better metrics for career player performance. The three that we have used here, TOI, GVT and GP, are reasonable proxies. As hockey analytics develop, utilizing more advanced methods like the Expected Goals Model (EGM), Macdonald (2012) or the Total Hockey Ratings (THoR) (Schuckers and Curro, 2013) will provide better estimates for player value and the value of player scouting once they are available for historical data. In this analysis wefocused on five years of NHL Draft Entry selections. Additional years of draft data would provide better estimates of player value and of the value of team scouting. We also note that while team scouting outperforms the Central Scouting Service, team scouting is far from optimal. A future analysis might look at the predictive power of analytics such as league equivalencies, see for example Desjardins (2004) or Vollman (2011), to rank players and compare results from that sort of analysis to those given here. Similarly, better data on teams’ front office personnel and duties would provide proxies for team budgets and geographic areas of emphasis. With all that in mind, however, it is clear from this analysis that NHL teams are getting considerable financial benefit from their internal scouting. They are beating the market set by the Central Scouting Service.


We would like to thank the editor and two referees for their helpful comments that improved the manuscript.



Anonymous. The Value of a Draft Pick, March 13, 2011., accessed June 11, 2013.


Awad T. , 2009a. The Top 10 Greatest Drafts Ever, Part One, June 18, 2009., accessed July 31, 2015.


AwadT. , 2009b. Understanding GVT, Part 1, July 30, 2009., accessed April 1, 2013.


BorghesiR. , 2008. Allocation of Scarce Resources: Insight from the NFL Salary Cap. Journal of Economics and Business. 60(6), 536–550.


BuchananJ. , 1965. An Economic Theory of Clubs, Economica. 32(125)1–14.


BurgerJ. ,WaltersS. , 2009. Uncertain Prospects: Rates of return in the baseball draft, Journal of Sports Economics. 10, 485–501.


DesjardinsG. , League Equivalencies, December 30, 2004., accessed June 11, 2013.


DorishJ. , Average Salaries in the NBA, NFL, MLB and NHL, November 12, 2011., accessed June 10, 2013.


FarrisJ. , Behind the Moves: NHL General Managers Tell How Winners are Built. (2011) circaNow (Canada).


Fyffe, I., Evaluating Central Scouting, January 19, 2011., accessed April 1, 2013.


GregorJ. , The Value of a Draft Pick, February 25, 2011., accessed June 11, 2013.


In re Coyotes Hockey, LLC, Case No. 2:09-BK-09491, (Bankr. D. Ariz. 2009), Business and Industry Monthly Operating Report, June 2009.


JensenM. , 1967. The Performance of Mutual Funds In The Period 1945-1964, Journal of Finance. 23(2), 389–416.


JohnsonD. , Draft Schmaft: From draft pick to NHLer, June 24, 2006., accessed June 11, 2013.


JoyceG. , Future Greats and Heartbreaks. (2008) Anchor (Canada).


KleinJ. , HackelS. , Sabres’ Job of No. 1 Fan Is Taken, February 26, 2011., accessed Dec. 23, 2013.


MacdonaldB. , 2012. AnExpected Goals Model for Evaluating NHL Teams and Players, Proceedings of 2012 MIT Sloan Sports Analytics Conference, Boston, MA.


MalloyS. , 2011. The Art of Scouting,Wiley (Mississauga, Ontario).


MasseyC. , ThalerR. , 2013. The Loser’s Curse: Decision Making and Market Efficiency in the National Football League Draft, Management Science. 59(7), 1479–1495.


McGuireM. , 2010. Matt McGuire’s NFL Draftology 321: Revising the trade value chart,, accessed June 13, 2013.


MorrealeM. , Marr hired as new Director of NHL Central Scouting, October 4, 2011., accessed December 23, 2013.

22 (2008) NHL Draft History, accessed June 2011.


SamuelsonP. , 1974. Challenge to judgment, Journal of Portfolio Management. 1(1), 17–19.


ScullyG. , 1974. Pay and Performance in Major League Baseball, The American Economic Review. 64(6), 915–930.


SchuckersM. , 2011. An Alternative to the NFL Draft Pick Value Chart Based upon Player Performance, Journal of Quantitative Analysis in Sports. 7(2), 1–12.


SchuckersM. , CurroJ. , 2013. Total Hockey Rating (THoR): A comprehensive statistical rating of National Hockey League forwards and defensemen based upon all on-ice events, in Proceedings of the 2013 MIT Sloan Sports Analytics Conference,


ShoaltsD. , The haves and have-nots in scouting, January 22, 2010., accessed December 23, 2013.


TangoT. , 2007. Value of draft picks in NHL, May 22, 2007., accessed June 2011.


TinglingP. , et al. 2011. Does order matter? An empirical analysis of NHL draft decisions, Sport, Business and Management: An International Journal. 1(2), 155–171.


TulskyE. , NHL draft: What does it cost to trade up? April 25, 2013, accessed June 10, 2013.


VollmanR. , Howe and Why: AHL Translations Revisited September 20, 2011., accessed June 10, 2013.


VollmanR. , Goals Versus Salary, 2011-12, October 10, 2012., accessed June 10, 2013.


WageB. , NHL Scouting Staffs – Carolina is Shorthanded, April 21, 2014., accessed September 7, 2015.


Figures and Tables


Average Performance Comparison for CSS and Team Draft Ordering: (l to r) TOI, GP, GVT.

Average Performance Comparison for CSS and Team Draft Ordering: (l to r) TOI, GP, GVT.

Relationship between Relative Performance and Rank Differential: (l to r) TOI, GP, GVT.

Relationship between Relative Performance and Rank Differential: (l to r) TOI, GP, GVT.
Table 1

Statistical Summaries of First Seven TOI, GP and GVT

MedianMean75th percentileMaxStd. deviation
Table 2

Comparison of Performance of CSS and Actual Draft Orderings

MetricOrderingPercent ofPercent of
optimalnearly optimal
TOICSS14% 19%
Team20% 32%
GPCSS4% 17%
Team11% 30%
GVTCSS4% 10%
Team10% 14%
Table 3

Comparison of Drafting Performance by Ordering and by Round

ordering byTeamCSS
roundsoptimalNear optimaloptimalNear optimal
1 to 36% 15% 2% 5%
4 to 712% 30% 9% 27%
ordering byTeamCSS
roundsoptimalNear optimaloptimalNear optimal
1 to 36% 10% 3% 4%
4 to 712% 18% 10% 12%
ordering byTeamCSS
roundsoptimalNear optimaloptimalNear optimal
1 to 316% 21% 11% 14%
4 to 721% 39% 18% 37%