(Note: World Rankings have gained much importance over the last decade. To my knowledge there has not been any published research validating the methodology for ranking players. The paper below, written over ten years ago, tried to evaluate competing rankings systems by how well each predicted performance. Neither ranking system did very well. Rankings are not a powerful predictor because of the random variation in scoring (i.e., the best golfer will often lose over one trial), past performance is not perfect indicator of future performance (i.e., a mutual fund that did well last year may not have similar results next year as every prospectus will tell you), and rankings are affected by the attributes of the course being played (i.e, there are horses for courses). Though a ranking system is not a good predictor of performance, it does serve two important functions. First, the ranking system acts as a (hopefully) neutral arbiter in deciding who qualifies for major tournaments. Second, it feeds our desire for lists (i.e., the 10 best places to retire, the 10 best Mexican restaurants), which also makes for endless 19th Hole banter--Luke Donald, are you kidding!)
Golf Digest printed an article (April 1999) critical of the World Golf Rankings (hereinafter termed IMG rankings). The author, Dean Knuth, devised a new system for ranking players. Knuth provided reasoning why the Golf Digest system would do a better job, but supplied no empirical evidence to substantiate his claim. The purpose of this paper is to evaluate the two ranking systems in their ability to explain performance at the Tournament Players Championship (TPC).
Rankings from both systems were available on 59 players who entered the TPC. The rankings were not the most current since the Golf Digest rankings were only available as of January 31, 1999.
The players were ranked from 1 to 59 consistent with the rankings of both systems. First, the rankings were correlated with the standings after the second round. The Spearman coefficient of rank correlation is shown in the table below. A coefficient of 1.0 would mean the rankings and the order of finish were identical. A coefficient of zero would mean there was no relationship between the rankings and performance. The IMG ranking had a slightly higher correlation coefficient (.56>.54) as shown in the table below.
Spearman Correlation Coefficient
Next, rankings were correlated with the final standings. Those players who missed the cut or were disqualified were ranked last. There was no significant difference between the correlation coefficients for the Golf Digest and IMG systems.
Last, a study was limited to only those players who made the cut. The 38 players in the study who made the cut were re-ranked and that ranking was correlated with the final standings. The Golf Digest system did not have a significantly larger correlation coefficient (.54 vs. .53) than the IMG system. A plot of the TPC standing versus the rankings (See the Appendix below) of each system demonstrate the similarity of results in using the two systems. They also demonstrate that neither system explains a great deal of the variance in performance.
In summary, the Golf Digest system does not do any better in explaining performance than the IMG system. There may be elements of the Golf Digest system worthy of adoption. Further research should isolate the effect of each of Golf Digest’s recommended changes to determine if it advances or impedes equity in the rankings.
Like all quasi-experiments there are significant caveats to the results.
· Only One Tournament - There is a large random component in performance among golfers from week to week. A player’s performance will also vary with the type of course (long, tight, etc.). Even in the best case studied here, the rankings only explained 34 percent of the variation in the order of finish. Therefore, the superiority of a ranking system could easily be due to chance when only one tournament is considered. The methodology demonstrated here should be repeated at the four majors and WGC tournaments to determine if the Golf Digest system (or other variation) is consistently better than the IMG system.
· Use of Old Rankings - The only set of both rankings available were as of January 31, 1999. Any definitive study would need rankings current with the tournament under study.
· Missing Rankings - The study was done using only 59 players out of the 144 who played. The research should have had rankings for all players so a more complete evaluation could have been done. It is especially important to see how well the ranking systems do in the lower regions – 40th to 70th – because of the new criterion for admission to tournaments based on the World Rankings. Without more rankings on players in the lower ranges, no judgment on the efficacy of the two systems over various regions could be made.
Rankings have more than just the sole purpose of predicting performance. The rankings clearly will be one factor in determining where and how often the best players play. Tours will also be impacted. Decreasing the points awarded on the Japanese tour may threaten that tour’s financial viability, and harm the popularity of the game in that country (e.g., What if no Japanese player qualified for a major event because of the ranking system?). It is important these incentives be shaped into a framework that will promote and serve the best interests of the game.
Within that framework, however, it is possible a more equitable ranking system could be built. Many of the suggestions put forward by Golf Digest seem reasonable. Policy should not be changed on what appears reasonable, however, but on what has been empirically tested to improve the accuracy of the rankings.
COMPARISON OF RANKINGS VERSUS TPC STANDINGS
 This is consistent with Knuth’s defense of the Slope System – all theory and no empirical verification. In setting forth his credentials for writing this article, Knuth again referenced a 30-year-old test score..
 Kendall, Maurice and Jean Gibbon, Rank Correlation Methods, Oxford University Press, NY, 1990. The Pearson correlation coefficient gave approximately the same results as shown in the table.
 Thirty nine of the players made the cut (i.e., low seventy players and ties). Nick Faldo was disqualified and was treated as a non-cut player.
 The two systems explain approximately 30 percent of the variance in performance. There appears to be a large random component in determining performance. That is why there are so many upsets in match play golf tournaments.
 Golf Digest’s claim of the inferiority of Japanese tour players was not borne out at the TPC. Joe Ozaki led the tournament after two days, and finished well ahead of his ranking. Brian Watts tied for the lead after the first day, but finished slightly lower than his IMG ranking. Shegeki Maruyama and Carlos Franco both missed the cut, but because of their low rankings (43rd and 41st) this was not unexpected. To assess the relative strength of the various tours, the performance of each tour’s players should be evaluated over an extended period. It would then be possible to adjust the points for each tour to bring performance and ranking into parity. This is a much more scientific and equitable approach than just slashing the points for the Japanese and Australian tours as Golf Digest recommends.