===================
== Thomas Pinder ==
===================
Bayesian ML, Causal Inference, and JAX


Proper Scoring Rules

A short post on scoring rules and their connection to a divergence metric.

A scoring rule is a metric that characterises the quality of a probabilistic forecast. If we are interested in forecasting rainfall, then we let the random variable Y denote a future event and Y the set of all possible values that Y could take. In this example, we would be define Y=R0 as it would not make sense to have negative rainfall. Our model is probabilistic and therefore outputs a probability distribution. We use P to denote the set of all valid probability distributions with support on Y. When computing a scoring rule, we seek to compare our model’s forecasted distribution ptP at time t against a true observation yt. In the context of our precipitation example, pt could be an exponential distribution with a rate parameter of 2 and yt would be a real value, such as 2.2mm, that would correspond to the true rainfall amount at time t.

A scoring rule S is then a function S:P×YR. Lower scores are indicative of a higher quality forecast. If we let q be the true probability distribution of rainfall, then for any pP we have (1)S(p,q)=Eq[p,Y]=S(p,y)dq(y) . A scoring rule is defined as a proper scoring rule if for all p (2)S(q,q)S(p,q) . Similarly, the proper scoring rule is strict if the inequality in (2) becomes a strictly less that. In this case, (2) will only achieve equality if p=q.

Scoring rules are a broad family of functions and there are many connections to statistical divergences. One example is the equivalence of the log-scoring rule S(p,y)=logp(y) and the Kullback-Leibler divergence (KLD) $\operatorname{KL}(q, p).Note the order of arguments in the KL-divergence operator matters as the divergence is asymmetric. To see this connection, we can write

S(q,q)S(p,q)=q(y)(logq(y)logp(y))dy=q(y)logq(y)logp(y)dy .

We can see that starting from the log-scoring rule, in just two lines we've arrived at the definition of the KLD.