Consolidating Computer Rankings

March 16, 2015
With the games of 14 March the field was finally connected, with a games graph diameter of six. 18.4 percent of the 45,150 team-pairs were either opponents or opponents' opponents and two thirds were separated by a path length of three.

Kenneth Massey (http://www.masseyratings.com) compiles a College Baseball Ranking Composite report that includes human and computer rating-based rankings. I publish a consolidated version of the human rankings, and beginning March 17 will do the same for the computer ratings, with a few differences.

Actually, because computer ratings rank all of the teams there are consolidation methods that can be brought to bear that aren't available for the human polls. Mapping multiple ordered lists into a representative one is a voting problem, and there are some relatively simple ones that I apply to Dr. Massey's lists.

Borda
For each ranking, assign each team points equal to the number of teams ranked below the team. For lists 301 teams long this amounts to giving 300 points for first, 299 for second, and so on down to zero for last. A slight twist to this familiar approach is how ties are handled. If there's a two-way tie for second, say, the rankings begin 1, 3, 3 instead of 1, 2, 2. Ties are rare in advanced computer ratings, but can be common when their ordinal rankings are combined.

To calculate the Borda result, add the Borda counts for each team from all rankings, assign 1st to the team with the most points, 2nd to the team with the next-most points and so on, taking into account the convention for ties. Note that a ranking by descending Borda Count will be identical to ranking by increasing average rank.

Bucklin
The Bucklin Majority rank for a team is the one for which a majority of the rankings have the team ranked at least that highly. For the human polls I use a "weak" definition of majority: "at least 50% of the rankings." For computer rankings I use the stronger "50%+1" definition.

This is the only summary that actually uses one of the rankings as its value. If there are an odd number of rankings Bucklin is identically the median rank (for an even number of voters it is the higher/worse rank in the median calculation.) For all the other composites some arithmetic function is applied to the rankings and then the teams are ordered by the values of that function. Bucklin is my preferred ranking for human polls with their truncated ballots, but it doesn't work as well for computer rankings unless there are a lot of those. (It works fine for FBS football, for instance.)

Instant Runoff Voting
IRV is actually a method used to pick a "single winner" (#1 team) using an algorithm that includes votes for #2, #3, etc. until some alternative has 50%+1 of the accumulated votes for first. There isn't an easy way to turn the "true" IRV algorithm into a consolidation of ranked ballots but a tweak to the Bucklin calculation to include the right kind of tiebreaker provides an ordered list that is about the same as what one would get with an IRV for 1st followed by IRV for 2nd from the remaining teams, and so on.

The actual calculation is just Bucklin with tiebreakers. The first tiebreaker for teams with the same majority ranking is the number of votes contributing to the majority. With a small number of "voters" this is not likely to break the tie, so the second tiebreaker is a "truncated" Borda Count - it only considers votes up to and including the majority rank.

Pairwise Winning Percentage
Pwise is based upon the number of other teams for which a team is ranked better than. It is based upon Condorcet methods that use a pairwise matrix where the row corresponding to team A and column corresponding to team B is the number of voters (ratings) that have team A ranked better than team B. The ranking is based upon a "pairwise winning percentage." For each (TeamA,TeamB) pair, count a pairwise win for team A (and pairwise loss for team B) if #votes("team A is better than tea B") is strictly greater than #votes("team B is better than teamA"), and pairwise loss for team A (win for team B) if the #votes("team B is better than team A") is strictly greater than #votes("team A is better than team B"). Pairwise ties for each of A and B occur when those #votes are equal.

To rank the teams, calculate "pairwise winning percentage" in the usual way, using for each team #pairwise wins + ½ #pairwise losses divided by the number of other teams. Note that for any particular pair of teams A and B, team B may have a "pairwise win" over team team A but team A may be ranked higher because it has more pairwise wins against the other N-2 teams than team B.

GeometricMean
No one has ever suggested this as a voting method, but it is included because Dr. Massey and I independently observed that it has some nice properties with regard to consolidating rankings. Specifically it gives more weight to better ranks by any rating compared to the equal weights given by the arithmetic average (Borda).

Sort in ascending order the nth root of the product of n rankings or equivalently the AntiLog of the average Logarithm of the rankings.

Correlating Rankings

There is no "right" computer rating compilation any more than there is a "right" computer rating. There are measurements of how similar two ratings are (this is one of the benefits of considering only rankings that include all teams.)

The simplest and to me the most intuitive is to just count how many of the ½×n×(n-1) pairs of teams are in the same relative order (concordant pairs) in the rankings being compared and how many are in reverse order (discordant pairs.) The number of discordant pairs for two rankings is called the distance between them. It is the number of team-swaps that would be required in either ranking to make it exactly the same as the other.

These can be used in a number of ways to form a ranking correlation with a value ranging from -1 (teams are ranked in exactly opposite order) to +1 (the lists are identical). One such correlation is Kendall's tau (τ) defined by

τ(rankA,RankB) = #concordant pairs(rankA,rankB) - #discordant pairs(rankA,RankB)

½×n×(n-1)
where n is the number of teams ranked.
In week five there are six computers, and their τ-correlations are:

AvgDistCompSAGRTMASKLKMORNOL
3875.2SAG0.87210.85200.84670.80070.7645
4020.4RT0.87210.85290.81550.78650.7765
4080.4MAS0.85200.85290.81890.78490.7815
4716.8KLK0.84670.81550.81890.77380.6935
5404.8MOR0.80070.78650.78490.77380.6490
5987.2NOL0.76450.77650.78150.69350.6490
Obviously τ(rankB,rankA) = τ(rankA,rankB) since it doesn't matter which of RankA or RankB is sorted to match the other. Using average distance between ranks, in week 5 the Sagarin ranking is the most representative of the computer rankings.

When the size of the ranked list (n) is large, the distances may seem "big" but for D1 baseball you have to compare them to a ½×n×(n-1) that is a lot bigger: 45,150 team-pairs for 2015's field of 301 teams. To make it easier to take that into account instead of looking at the distances one could look at the percentage of concordant team pairs.

The percentage of concordant pairs is related to τ by %concordant = ½×(τ+1)
½×(τ+1) serves the same role in ordinal ranking correlations as r2 in correlations of real-valued lists. It is a mapping of values in the [-1,1] interval to [0,1] ( [0,100] when expressed as a per cent.)

Correlating Composites

I do not know if there's a theorem that proves that any "composite" ranking always correlates more closely to the constituents of the lists than any one of the lists does to the others, but that is true for all of the composite rankings I listed above. Every one of the composite rankings has a lower average distance from the six computer rankings than any of the computer rankings has:
Distance%Concordant
AvgDistCompSAGRTMASKLKMORNOL    SAGRTMASKLKMORNOL
2906.50PWISE16591916209029913978480596.2695.6995.3093.2991.0989.24
2930.33BUCK17911947204929733916490695.7695.4295.1993.1391.0388.82
2950.50BORDA18001995212131853991461195.9395.4995.2192.8491.0489.66
2979.00IRV18421988209430283971495195.8995.5795.3393.2591.1588.96
2984.33GMEAN18152031220132294068456295.9595.4795.0992.8090.9389.83
3875.20SAG2868332034384469528193.6192.6092.3390.0488.23
4020.40RT2868329841384787501193.6192.6590.7789.3388.83
4080.40MAS3320329840624823489992.6092.6590.9489.2589.08
4716.80KLK3438413840625073687392.3390.7790.9488.6984.68
5404.80MOR4469478748235073787290.0489.3389.2588.6982.45
5987.20NOL5281501148996873787288.2388.8389.0884.6882.45

For the week 5 rankings the pairwise percentage is the most representative, which is consistent with its construction. It is no surprise that the geometric mean is the least representative (though its still a very good consolidation) given the reason we include it in the first place - it exaggerates the differences in the better ranks.

In memory of
SEBaseball.com

© Copyright 2015 Paul Kislanko