The first things to say are in defence of the ICC ratings. Objective attempts to rank teams based on historical performance are, by definition, historical (and thus backwards looking): the top team at any one moment in time has had the best record (according to the chosen criteria for assessment), this does not mean it will continue to have the best record in future. And either a system weights recent matches very heavily, in which case the ratings will prove very unstable (a great team loses a couple of games and loses its position), or it rates them less heavily (in which case a team in serious decline may stay number one for a while). And sometimes an objective system will agree with one's subjective assessment, and sometimes it will not, and the comparison of the two is itself subjective: has one chosen the right objective criteria, in which case one's subjective assessment is revealed as false, or does the failure of the system to yield an instinctively satisfying result prove that the system is wrong? A system that suggested manifestly absurd conclusions can be rejected; in fact, looking back at the historical record, it's clear that most of the time, the best team in the world according to the ratings was also the team most followers of the game would have picked. But when no side is dominant, supporters will disagree subjectively with each other, and at least some of them will also disagree with the output of any given rating system.
But... read on a little bit more, and the Wikipedia page will also tell you something of how the ICC ratings are calculated, and it's immediately clear: it's a very complicated system. Indeed, it contains some apparent absurdities: the quality of a team's victory is assessed according to the strength of the opposition, unless the opposition is of a very different strength to the team itself, in which case it is assessed according to the team's own strength. This just seems bizarre. When you look at the system in more detail, you can understand how it works to produce mostly sensible outcomes. But it's a hack, not a clean system based on first principles.
But is such a clean system possible? In fact, a much cleaner system is possible, based on principles of the Hungarian mathematician Arpad Elo, used to rank players in chess and other games. What if we apply Elo's system to test cricket?
In fact, the possibly disappointing news is that in general the outcomes are not so different to the outcomes of the ICC's system. The ICC system can in fact be considered a dirty hack to try and generate an Elo-like rating while taking into account certain cricket-specific factors. But the result is a rating that is hard to interpret and subject to a large number of essentially arbitrary parameters. What I have done is to implement a relatively pure Elo system and applied it to a database of test match results. In these posts, I'm going to explain the system, look at the logic of the ICC system in comparison, and also to present some interesting observations about how ratings have changed over time. This will take a while, so before I start, here are the current ratings of the sides under my system, with the scores and rankings of the teams under the ICC's system in brackets:
- Australia 154 (109 - 3rd)
- India 116 (110 - 2nd)
- Pakistan 93 (106 - 4th)
- South Africa 74 (114 - 1st)
- England 73 (99 - 6th)
- New Zealand 67 (99 - 5th)
- Sri Lanka -4 (89 - 7th)
- West Indies -86 (76 - 8th)
- Bangladesh -172 (47 - 9th)
- Zimbabwe -315 (5 - 10th)
But what does this mean? For that, you'll have to wait for the next posts in this series.
No comments:
Post a Comment