As you may have noticed, each FICS player now has a rating and an RD. RD stands for "ratings deviation".
The new system with the RD improves upon the binary categorization that was
used before on fics and elsewhere, where players with fewer than 20 games were
labeled"provisional" and others were labeled "established". Instead of two
separate ratings formulas for the two categories, there is now a single
formula incorporating the two ratings and the two RD's to find the ratings
changes for you and your opponent after a game.
The Ratings Deviation is used to measure how much a player's current rating
should be trusted. A high RD indicates that the player may not be competing
frequently or that the player has not played very many games yet at the
current rating level. A low RD indicates that the player's rating is fairly
well established. This is described in more detail below under "RD
Interpretation".
In general, if your RD is high, then your rating will change a lot each time
you play. As it gets smaller, the ratings change per game will go down.
However, your opponent's RD will have the opposite effect, to a smaller
extent: if his RD is high, then your ratings change will be somewhat smaller
than it would be otherwise.
In this system, the RD will decrease somewhat each time you play a game,
because when you play more games there is a stronger basis for concluding what
your rating should be. However, if you go for a long time without playing any
games, your RD will increase to reflect the increased uncertainty in your
rating due to the passage of time. Also, your RD will decrease more if your
opponent's rating is similar to yours, and decrease less your opponent's
rating is much different.
the other system, except for provisional games, the ratings changes for the two players in a game would balance each other out - if A wins 16 points, B loses 16 points. That is not the case with this system. Here is the explanation I received from Mark Glickman:
The system does not conserve rating points - and with good reason! Suppose two players both have ratings of 1700, except one has not played in awhile and the other playing constantly. In the former case, the player's rating is not a reliable measure while in the latter case the rating is a fairly reliable measure. Let's say the player with the uncertain rating defeats the player with the precisely measured rating. Then I would claim that the player with the imprecisely measured rating should have his rating increase a fair amount (because we have learned something informative from defeating a player with a precisely measured ability) and the player with the precise rating should have his rating decrease by a very small amount (because losing to a player with an imprecise rating contains little information). That's the intuitive gist of my extension to the Elo system.
On average, the system will stay roughly constant (by the
law of large numbers). In other words, the above scenario
in the long run should occur just as often with the
imprecisely rated player losing.
Direct from Mark Glickman:
Each player can be characterized as having a true (but unknown) rating that may be thought of as the player's average ability. We never get to know that value, partly because we only observe a finite number of games, but also because that true rating changes over time as a player's ability changes. But we can *estimate* the unknown rating. Rather than restrict oneself to a single estimate of the true rating, we can describe our estimate as an *interval* of plausible values. The interval is wider if we are less sure about the player's unknown true rating, and the interval is narrower if we are more sure about the unknown rating. The RD quantifies the uncertainty in terms of probability:
The interval formed by Current rating +/- RD contains your true rating with probability of about 0.67.
The interval formed by Current rating +/- 2 * RD contains your true rating with probability of about 0.95.
The interval formed by Current rating +/- 3 * RD contains your true rating with probability of about 0.997.
For those of you who know something about statistics, these are not confidence intervals, but are called "central posterior intervals" because the derivation came from a "Bayesian" analysis of the problem.
These numbers are found from the cumulative distribution function of the
normal distribution with mean = current rating, and standard deviation = RD.
For example, CDF[ N[1600,50], 1550 ] = .159 approximately (that's shorthand
Mathematical notation.)
where c is a numerical constant chosen so that predictions made according to the ratings from this system will be approximately optimal.
Step 2. Calculate the "attenuating factor" due to your OPPONENT's RD, for use in later steps.
A PostScript file containing Mark Glickman's paper discussing this ratings
system may be obtained via ftp. The ftp site is hustat.harvard.edu, the
directory is /pub/glickman, and the file is called "glicko.ps".
.
The Glicko Ratings System was invented by Mark Glickman, Ph.D. who is currently at the Harvard Statistics Department, and who is bound for Boston University.
Vek and Hawk programmed and debugged the new ratings calculations (we may still be debugging it). Helpful assistance was given by Surf, and Shane fixed a heinous bug that Vek invented.
Vek wrote this helpfile and Mark Glickman made some essential
corrections and additions.
C.C.Cafe | Index | D.I.C.S. | Events | Feedback |