|
What is an Elo Rating?An Elo rating is a numercial measure of performance. If all you want to know is how Elo ratings relate to traditional banzuke ranks for the top rikishi, then here is the situation since 2004:
This section covers the basic ideas behind the Elo system and the issues faced by those using it. The next section, The BKQ Model, provides a mathematical account of Elo ratings and highlights relevant issues for those interested in the details of how ratings are calculated. Finally, in Analysis, all remaining matters are reviewed and a summary presented. That said, let me conclude these opening remarks with a confession: I am not an expert on sumo or Elo and I have not spent much time on researching Elo systems beyond the basic definitions. It follows that some of what I say here is opinion and uninformed by academic standards. On the other hand I have spent a great deal of time thinking about what I do know and how that could be creatively used. Tell Me More!In sports like football, all competitors play each other in a given league or division, so a simple tally of wins or points scored is a reasonable way of rating competitors. In chess and Grand Sumo, some competitors never face each other, while others may compete multiple times. Thus, a simple ranking system based on the number of wins is unlikely to represent a player's standing accurately. The Elo rating system was invented by Professor Arpad Elő in 1960 at the request of FIDE to provide an objective numerical rating for chess players. His system is in fact is a way of providing a numerical rating for competitors in any activity in which it is not the case that everyone competes against everyone else. Elo's system took some time to become accepted but has been used by FIDE since 1970 as the primary way of ranking players. Elo ratings for Grand Sumo is not a new idea but it seems be on the fringes of acceptability. These pages address the question of whether Elo's ideas might be usefully applied to sumo ratings. I am certain that I am not the first person to do this but I have been unable to find anything on the internet beyond mentions in online fora, and my requests for more subtantial work have not been successful. (I would welcome references to any work that readers know of.) Here's Elo's idea, as applied to Grand Sumo:
Of course, the ratings this system produces will depend on the starting date, the initial rating and on precisely how much change in rating there is for a given difference in rating. The important point for me is that the calculations are mathematical; i.e. completely transparent and not in any way subjective. This is unlike the banzuke-based ranking system as discussed below. Readers who want to dive straight into the numbers at this point should skip ahead to the The BKQ Model. Here we will continue with discussion of more qualitative issues. Matters ArisingI said above that this site is about whether Elo ratings can be applied to sumo. In fact I am biased and I want Elo to be adopted for sumo – not as a replacement for the traditional banzuke-based ranking system – but as a complementary way to measure ability and progress.I have encountered resistance to the idea of Elo ratings in general and to the idea of applying them to sumo. Here I want to raise these issues and if not entirely counter them, then to explain why I think Elo works for sumo (if it does!) InflationElo ratings have faced a notable issue known as "inflation". In chess, this phenomenon became evident as top players' Elo ratings significantly surpassed those of previous years. For instance, in 1972, Bobby Fischer, widely regarded as one of the greatest chess players of all time, achieved the milestone of breaking the 2700 barrier – an event of considerable sensation at the time. However, in May 2024, this achievement has become commonplace, with dozens of players now exceeding 2700. Some people think this inflation is a result of competitors getting better over time. It seems feasible that the level of competition in chess has indeed improved due to the success and availability of computer chess and the application of sports science techniques, including psychology. However, the idea that today there are dozens of people "as good as" Bobby Fischer would be controversial. An alternative idea is that the inflation is inherent in the mathematics of the Elo system. I have not seen any proof of this per se, but it is certainly the case that my Elo ratings for rikishi in sumo also inflate. For example, one of the best rikishi ever was Taiho who was at his peak in March 1969 after an extraordinary run of wins. His sumo Elo at that time was 2009 and this was significantly higher than any of his contemporaries. However, at the time of writing, 2009 is the rating of a sekiwake and I don't think many people would agree that if Taiho were fighting today, he would only be a sekiwake. Could it be that the level of competition in sumo has risen due to advances in sports sciences? It seems unlikely to me, and opinion (at r/Sumo) is divided. I think it is more likely that inflation is inherent in the Elo ratings system. My take on inflation is:
ConvergenceElo's idea was that a competitor's rating would converge to a measure of his "true ability" relative to his competitors, thus establishing who was "really" the best. However, he was also aware that ratings would only converge if:
It turns out that you can prove that ratings do converge under these conditions, but neither condition holds in sumo or chess. Nevertheless, Elo ratings – or at least, a system based on Elo's ideas – have been adopted by the world's chess authority. Furthermore, whatever other properties Elo ratings have when applied to sumo, it seems obvious to me that a rikishi who has a run of victories will see his Elo ratings rise and they will continue to do so until he reaches the upper limit of his capabilities relative to his co-competitors. Conversely, a rikishi on a losing run will see his ratings drop until he reaches his natural level. For example, compare Midorifuji's Elo record with his story. I think there is an excellent correspondence between the two as he progresses through the ranks, becomes injured, recovers, continues his ascent to the top of maegashira and bobbles about whilst he finds his level. PredictionIf you have a numerical rating for two rikishi then you have data for assessing the probability of one rikishi beating another. In fact Elo's system involves something called the "estimated probability of winning". However, Elo himself noted that his system could not allow for factors that may be relevant such as:
Of course it is difficult to see how any of these ideas can be made measurable, let alone incorporate them into a rating system in an objective way. As to whether previous performance as measured by Elo ratings is any guarantee of future results, this is currently under investigation. Adequacy of Banzuke-Based RankingsThe current system of rating rikishi by awarding them the status of a given chii (rank) such as "Juryo 5 West" has been used since the 1700s. It is traditional, meaningful, and, I suspect, attractively mysterious to non-Japanese fans. This is true of me and I do not want to see the system replaced. However, some people don't see why anyone would need any other way of ranking rikishi.For me, the issue is that assigning chii involves a subjective element due to the traditional idea that there should only be two rikishi at a given rank. As far as I am aware, the motivation for this idea is aesthetic: when there are exactly two rikishi at a given rank the banzuke has a pleasing symmetry. In practice, this "rule" is not strictly followed in the case of san'yaku ranks and in modern Grand Sumo there can be any number of rikishi at any san'yaku rank. However, for maegashira and below the idea is nearly a cast-iron rule. This means that a rikishi might not be promoted to a given rank because there are two other rikishi who are deemed by the JSA to have a stronger claim to holding that rank. A similar remark applies to demotion. The people who make such decisions are highly experienced former rikishi and whilst there are occasional controversial decisions, I think it is reasonable to assume that there are rational arguments behind every decision. Nevertheless, the process contains subjective elements in contast to the objective Elo system. Another issue for me is that the traditional ranks do not tell me as much as I would like about a rikishi's performance. Of course, to be a Yokozuna is to have demonstrated extraordinarily high levels of performance, but the title does not tell me if the rikishi in question hardly ever attends a basho due to injury, or happens to be one of the strongest the sport has ever seen, or is somewehere in between. Similarly, what does being an "M8" (say) mean beyond "the judges think he is better than those ranked M9 and below and worse than those ranked M7 and above". That is useful information to be sure, but it doesn't tell me if, for example, the rikishi normally fights at M5 but was recently injured. In that case, I would think of the guy as being "M5, really". I believe that Elo ratings are more robust than chii but this another subject yet to be investiged. Other ObjectionsSome people like the idea of a numerical measure of rikishi performance but don't understand how Elo works. I have heard people say that a rating of 2100 (say) is a "meaningless number". I think it is no less meaningful as saying that the rikishi is M8 (see above): a rating of 2100 means, at the very least, the rikishi is better than those with a score of below 2100 and worse than those with a rating of more than 2100. However, I think there is more to Elo ratings than this as I will try to explain below.There are any number of ways we could assign numbers to rikishi to reflect their performance. Each idea comes with its pros and cons. For example, if we take a rikishi's rating to be the number of wins he has had, then this system is very simple and clearly objective. However, a disadvantage is that it does not consider the relative strengths of the rikishi involved: a rikishi gets one point for winning irrespective of his opponent. If everybody fought everybody else then this would be fair, but this is not the case in sumo. And, as far as I know, the choice of who gets to fight who again involves subjectivity. The significance of relative strength is certainly something the Judge's Division takes into account but it begs the question: what actually is the "strength" of the rikishi? My guess is that experts would say it is a combination of a rikishi's current rank, his recent performance, past and present injuries and so on, but these are all, essentially, subjective. Elo's system neatly side-steps this issue by not using "strength" at all: way:
I think it is reasonable to say that those rikishi with higher ratings have provided evidence of being "better" than those with lower ratings. It is for this reason that I think it is fair to say that Elo ratings – or, to be precise – the differences between Elo ratings are meaningful. Finally, some people accept that Elo ratings aren't meaningless but object to them because of some perceived infelicity. Some of those people have proposed alternative systems such as Glicko (which, in contrast to Elo, takes absences into account). With sincere apologies to everyone working on such ideas, I have to say I haven't looked into these alternatives because I don't have the time and Elo is what I am interested in. My goal is to present my perspective and hope to convince you that what I say is worth listening to, not to dismiss the validity of other viewpoints. It seems likely to me that some readers won't want to read any further because of this – in which case I'm sorry to disappoint you and I respect your choice. |