Scoring rule

From Wikipedia, the free encyclopedia
(Redirected from Scoring function)
Jump to navigation Jump to search
File:Scoring functions.gif
Visualization of the expected score under various predictions from some common scoring functions. Dashed black line: forecaster's true belief, red: linear, orange: spherical, purple: quadratic, green: log.

In decision theory, both a scoring rule[1] as well as a scoring function[2] provide an ex post summary measure for the evaluation of the quality of a prediction or forecast. They assign a numeric score to a single prediction given the actual outcome. Depending on the sign convention, this score can be interpreted as a loss or a reward for the forecaster. Scoring rules assess probabilistic predictions or forecasts, i.e. predictions of the whole probability distribution F of the outcome. On the other hand, scoring functions assess point predictions, i.e. predictions of a property or functional T(F) of the probability distribution F of the outcome. Examples of such a property are the expectation and the median.

File:Logarithmic score.png
The average logarithmic score of 10 points i.i.d. sampled from a standard normal distribution (blue histogram), evaluated on a variety of distributions (red line). Although not necessarily true for individual samples, on average, a proper scoring rule will give the lowest score if the predicted distribution matches the data distribution.
File:Calibration plot.png
A calibration curve allows to judge how well model predictions are calibrated, by comparing the predicted quantiles to the observed quantiles. Blue is the best calibrated model, see calibration (statistics).

Scoring rules answer the question "how good is a predicted probability distribution given the observation of the actual outcome?" Scoring rules that are (strictly) proper are proven to have the lowest expected score if the predicted distribution equals the underlying distribution of the target variable. Although this might differ for individual observations, this should result in a minimization of the expected score if the "correct" distributions are predicted.

In the same way, scoring functions answer the question "how good is a point prediction given the observation of the actual outcome?". Scoring functions that are (strictly) consistent (for the functional T) are proven to have the lowest expected score if the point prediction equals (or is among) the true functional of the underlying distribution of the target variable.

Scoring rules and scoring functions are often used as "cost functions" or "loss functions" of forecasting models. If a sample of forecasts and observations of the outcome is collected, they can be evaluated as the empirical mean of the given sample, often also called the "score". Scores of predictions of different models or forecasters can then be compared to conclude which model or forecaster is best.

For example, consider a probabilistic model that predicts (based on an input x) a gaussian distribution 𝒩(ΞΌ,Οƒ2) with mean ΞΌβˆˆβ„ and standard deviation Οƒβˆˆβ„+. A common interpretation of probabilistic models is that they aim to quantify their own predictive uncertainty. In this example, an observed target variable yβˆˆβ„ is then held compared to the predicted distribution 𝒩(ΞΌ,Οƒ2) and assigned a score 𝐒(𝒩(ΞΌ,Οƒ2),y)βˆˆβ„. When a probabilistic model is trained on a scoring rule, it should "teach" the model to predict when its uncertainty is low, and when its uncertainty is high, and it should result in calibrated predictions, while minimizing the predictive uncertainty.

Although the example given concerns the probabilistic forecasting of a real valued target variable, a variety of different scoring rules have been designed with different target variables in mind. Scoring rules exist for binary and categorical probabilistic classification, as well as for univariate and multivariate probabilistic regression.

Definitions

[edit | edit source]

Consider a sample space or observation domain, Ξ©, which comprises the potential outcomes of a future observation; a Οƒ-algebra π’œ of subsets of Ξ© and a convex class β„± of probability measures on (Ξ©,π’œ). A function defined on Ξ© and taking values in the extended real line, ℝ‾=[βˆ’βˆž,∞], is β„±-quasi-integrable if it is measurable with respect to π’œ and is quasi-integrable with respect to all Fβˆˆβ„±.

A (statistical) functional T is a potentially set-valued mapping from the class of probability distributions β„± to a Euclidean space, i.e. T:ℱ→ℝd with Fβ†’T(F).

Probabilistic forecast

[edit | edit source]

A probabilistic forecast is any probability measure Fβˆˆβ„±, i.e. a distribution of potential future observations.

Point forecast

[edit | edit source]

A point forecast for the functional T is any value xβˆˆβ„d.

Scoring rule

[edit | edit source]

A scoring rule is any extended real-valued function 𝐒:ℱ×Ω→ℝ such that 𝐒(F,β‹…) is β„±-quasi-integrable for all Fβˆˆβ„±. 𝐒(F,y) represents the loss or penalty when the forecast Fβˆˆβ„± is issued and the observation y∈Ω materializes.

Scoring function

[edit | edit source]

A scoring function is any real-valued function S:ℝd×Ω→ℝ where S(x,y) represents the loss or penalty when the point forecast xβˆˆβ„d is issued and the observation y∈Ω materializes.

Orientation / Sign convention

[edit | edit source]

Scoring rules 𝐒(F,y) and scoring functions S(x,y) are negatively (positively) oriented if smaller (larger) values mean better. Changing the convention can be accomplished by multiplying the score by βˆ’1. Here we adhere to the negative orientation, hence the association with "loss".

Expected score

[edit | edit source]

We write for the expected score of a probabilistic prediction Fβˆˆβ„± with respect to the underlying distribution Qβˆˆβ„±:

𝔼Y∼Q[𝐒(F,Y)]=βˆ«π’(F,Ο‰)dQ(Ο‰)

Similar, the expected score of a point prediction xβˆˆβ„d with resprect to the underlying distribution Qβˆˆβ„±:

𝔼Y∼Q[S(x,Y)]=∫S(x,Ο‰)dQ(Ο‰)

Sample average score

[edit | edit source]

A way to estimate the expected score is by means of the sample average score. Given a sample of prediction-observation pairs e.g. (Fi,yi) for probabilistic predictions Fi and observations yi∈Ω, i=1…n, (xi,yi) for point predictions xi, the average score is calculated as

  • for scoring rules:
 E[𝐒]^=1nβˆ‘i=1n𝐒(Fi,yi)
  • for scoring functions:
 E[S]^=1nβˆ‘i=1nS(xi,yi)

By invoking some law of large numbers argument, the sample average scores are consistent estimators of the expectation.

Properties

[edit | edit source]

Propriety and consistency

[edit | edit source]

Strictly proper scoring rules and strictly consistent scoring functions encourage honest forecasts by maximization of the expected reward: If a forecaster is given a reward of βˆ’π’(F,y) if y realizes (e.g. y=rain), then the highest expected reward (lowest score) is obtained by reporting the true probability distribution.[1]

Proper scoring rules

[edit | edit source]

A scoring rule 𝐒 is proper relative to β„± if (assuming negative orientation) its expected score is minimized when the forecasted distribution matches the distribution of the observation.

𝔼Y∼Q[𝐒(Q,Y)]≀𝔼Y∼Q[𝐒(F,Y)] for all F,Qβˆˆβ„±.

It is strictly proper if the above equation holds with equality if and only if F=Q.

Consistent scoring functions

[edit | edit source]

A scoring function S is consistent for the functional T relative to the class β„± if

𝔼Y∼F[S(t,Y)]≀𝔼Y∼F[S(x,Y)] for all Fβˆˆβ„±, all t∈T(F) and all xβˆˆβ„d.

It is strictly consistent if it is consistent and equality in the above equation implies that x∈T(F).

Affine transformation

[edit | edit source]

After an affine transformation a strictly proper scoring rule remains strictly proper, a strictly consistent scoring function (for some functional T) remains strictly consistent.[3] That is, if 𝐒(F,y) is a strictly proper scoring rule then a+b𝐒(F,y) with bβ‰ 0 is also a strictly proper scoring rule, though if b<0 then the optimization sense of the scoring rule switches between maximization and minimization. For scoring functions the same statement applies with the obvious changes.

Locality

[edit | edit source]

A proper scoring rule is said to be local if its estimate for the probability of a specific event depends only on the probability of that event. This statement is vague in most descriptions but we can, in most cases, think of this as the optimal solution of the scoring problem "at a specific event" is invariant to all changes in the observation distribution that leave the probability of that event unchanged. All binary scores are local because the probability assigned to the event that did not occur is determined so there is no degree of flexibility to vary over.

Affine functions of the logarithmic scoring rule are the only strictly proper local scoring rules on a finite set that is not binary.

Decomposition

[edit | edit source]

The expectation value of a proper scoring rule S can be decomposed into the sum of three components, called uncertainty, reliability, and resolution,[4][5] which characterize different attributes of probabilistic forecasts:

E(S)=UNC+RELβˆ’RES.

If a score is proper and negatively oriented (such as the Brier Score), all three terms are positive definite. The uncertainty component is equal to the expected score of the forecast which constantly predicts the average event frequency. The reliability component penalizes poorly calibrated forecasts, in which the predicted probabilities do not coincide with the event frequencies.

The equations for the individual components depend on the particular scoring rule. For the Brier Score, they are given by

UNC=xΒ―(1βˆ’xΒ―)
REL=E(pβˆ’Ο€(p))2
RES=E(Ο€(p)βˆ’xΒ―)2

where xΒ― is the average probability of occurrence of the binary event x, and Ο€(p) is the conditional event probability, given p, i.e. Ο€(p)=P(x=1∣p)

Examples of proper scoring rules

[edit | edit source]

There are an infinite number of scoring rules, including entire parameterized families of strictly proper scoring rules. The ones shown below are simply popular examples.

Categorical variables

[edit | edit source]

For a categorical response variable with m mutually exclusive events, Y∈Ω={1,…,m}, a probabilistic forecaster or algorithm will return a probability vector 𝐩∈[0,1]m with probabilities for each of the m outcomes.

If y=i materializes, one often abbreviates the score as 𝐒(𝐩,i).

Logarithmic score

[edit | edit source]
File:ExpectedLog.png
Expected value of logarithmic rule. When Event 1 is expected to occur with probability of 0.8, the blue line is described by the function 0.8log(x)+(1βˆ’0.8)log(1βˆ’x).

The logarithmic scoring rule is a strictly proper and local scoring rule. This is also the negative of Shannon entropy, which is commonly used as a scoring criterion in Bayesian inference. This scoring rule has strong foundations in information theory.

𝐒(𝐩,i)=ln(pi)

Here, the score is calculated as the logarithm of the probability estimate for the actual outcome. That is, a prediction of 80% that correctly proved true would receive a score of ln(0.8) = βˆ’0.22. This same prediction also assigns 20% likelihood to the opposite case, and so if the prediction proves false, it would receive a score based on the 20%: ln(0.2) = βˆ’1.6. The goal of a forecaster is to maximize the score and for the score to be as large as possible, and βˆ’0.22 is indeed larger than βˆ’1.6.

If one treats the truth or falsity of the prediction as a variable x with value 1 or 0 respectively, and the expressed probability as p, then one can write the logarithmic scoring rule as x ln(p) + (1 βˆ’ x) ln(1 βˆ’ p). Note that any logarithmic base may be used, since strictly proper scoring rules remain strictly proper under linear transformation. That is:

L(𝐩,i)=logb(pi)

is strictly proper for all b>1.

Brier/Quadratic score

[edit | edit source]

The quadratic scoring rule is a strictly proper scoring rule

𝐒Q(𝐩,i)=2piβˆ’π©β‹…π©=2piβˆ’βˆ‘j=1mpj2

where pi is the probability assigned to the correct answer i.

The Brier score, originally proposed by Glenn W. Brier in 1950,[6] can be obtained by an affine transform from the quadratic scoring rule.

𝐒B(𝐩,i)=βˆ‘j=1m(yjβˆ’pj)2

Where yj=1 when the jth event is correct and yj=0 otherwise. It can be thought of as a generalization of mean squared error to probabilistic forecasts.

An important difference between these two rules is that a forecaster should strive to maximize the quadratic score 𝐒Q yet minimize the Brier score 𝐒B. This is due to a negative sign in the linear transformation between them.

Spherical score

[edit | edit source]

The spherical scoring rule is also a strictly proper scoring rule

𝐒(𝐩,i)=pi‖𝐩‖=pip12+β‹―+pm2

Also its generalization with Ξ±>1 is strictly proper

𝐒(𝐩,i)=piΞ±βˆ’1(βˆ‘j=1mpjΞ±)(Ξ±βˆ’1)/Ξ±


Ranked Probability Score

[edit | edit source]

The ranked probability score [7] (RPS) is a strictly proper scoring rule, that can be expressed as:

RPS(𝐩,i)=βˆ‘k=1mβˆ’1(βˆ‘j=1kpjβˆ’yj)2

Where yj=1 when the jth event is correct and yj=0 otherwise, and C is the number of classes. Other than other scoring rules, the ranked probability score considers the distance between classes, i.e. classes 1 and 2 are considered closer than classes 1 and 3. The score assigns better scores to probabilistic forecasts with high probabilities assigned to classes close to the correct class. For example, when considering probabilistic forecasts 𝐩1=(0.5,0.5,0) and 𝐩2=(0.5,0,0.5), we find that RPS(𝐩1,1)=0.25, while RPS(𝐩2,1)=0.5, despite both probabilistic forecasts assigning identical probability to the correct class.

Comparison of categorical strictly proper scoring rules

[edit | edit source]

Shown below on the left is a graphical comparison of the Logarithmic, Quadratic, and Spherical scoring rules for a binary classification problem. The x-axis indicates the reported probability for the event that actually occurred.

It is important to note that each of the scores have different magnitudes and locations. The magnitude differences are not relevant however as scores remain proper under affine transformation. Therefore, to compare different scores it is necessary to move them to a common scale. A reasonable choice of normalization is shown in the picture where all scores intersect the points (0.5,0) and (1,1). This ensures that they yield 0 for a uniform distribution (two probabilities of 0.5 each), reflecting no cost or reward for reporting what is often the baseline distribution. All normalized scores below also yield 1 when the true class is assigned a probability of 1.

File:RawScore.png
Score of a binary classification for the true class showing logarithmic (blue), spherical (green), and quadratic (red)
File:NormalizedScore.png
Normalized score of a binary classification for the true class showing logarithmic (blue), spherical (green), and quadratic (red)

Univariate continuous variables

[edit | edit source]

The scoring rules listed below aim to evaluate probabilistic predictions when the predicted distributions are univariate continuous probability distributions, i.e. the predicted distributions F are defined over a univariate target variable Yβˆˆβ„ and have a probability density function f:ℝ→ℝ+.

Logarithmic score for continuous variables

[edit | edit source]

The logarithmic score is a local strictly proper scoring rule. It is defined as

L(F,y)=βˆ’ln(f(y)).

It is a local, strictly proper scoring rule. The logarithmic score for continuous variables has strong ties to Maximum likelihood estimation. However, in many applications, the continuous ranked probability score is often preferred over the logarithmic score, as the logarithmic score can be heavily influenced by slight deviations in the tail densities of forecasted distributions.[8]

Continuous ranked probability score

[edit | edit source]
File:Illustration CRPS.png
Illustration of the continuous ranked probability score (CRPS). Given a sample y and a predicted cumulative distribution F, the CRPS is given by computing the difference between the curves at each point x of the support, squaring it and integrating it over the whole support.

The continuous ranked probability score (CRPS)[9] is a strictly proper scoring rule much used in meteorology. It is defined as

CRPS(F,y)=βˆ«β„(F(x)βˆ’H(xβˆ’y))2dx

where H is the Heaviside step function and yβˆˆβ„ is the observation. For distributions with finite first moment, the continuous ranked probability score can be written as:[1]

CRPS(F,y)=𝔼X∼F|Xβˆ’y|βˆ’12𝔼X,X∼F|Xβˆ’X|

where X and X are independent random variables, both sampled from the distribution F. This is the energy form of CRPS and opens the door to estimating the CRPS via Monte Carlo sampling (through approximating the expectation value).

Furthermore, when the cumulative probability function F is continuous, the continuous ranked probability score can also be written as[10]

CRPS(F,y)=𝔼X∼F|Xβˆ’y|+𝔼X∼F[X]βˆ’2𝔼X∼F[Xβ‹…F(X)]

The continuous ranked probability score can be seen as both a continuous extension of the ranked probability score, as well as quantile regression. The continuous ranked probability score over the empirical distribution F^q of an ordered set points q1≀…≀qn (i.e. every point has 1/n probability of occurring), is equal to twice the mean quantile loss applied on those points with evenly spread quantiles (Ο„1,…,Ο„n)=(1/(2n),…,(2nβˆ’1)/(2n)):[11]

CRPS(F^q,y)=2nβˆ‘i=1nΟ„i(yβˆ’qi)++(1βˆ’Ο„i)(qiβˆ’y)+

For many popular families of distributions, closed-form expressions for the continuous ranked probability score have been derived. The continuous ranked probability score has been used as a loss function for artificial neural networks, in which weather forecasts are postprocessed to a Gaussian probability distribution.[12][13]

CRPS was also adapted to survival analysis to cover censored events.[14]

The CRPS can be thought of as the generalization of the mean absolute error (MAE) to probabilistic forecasts, and for a single sample is equivalent to the MAE. Another way to think of it is the Brier/quadaratic score of the sampled cumulative distribution function F for the binary event {X≀y}.

CRPS is a special case of the CramΓ©r distance (or CramΓ©r's distance) and can be seen as an improvement of Wasserstein distance often used in machine learning. CramΓ©r distance performed better in ordinal regression than KL distance or the Wasserstein metric.[15]

While CRPS is widely used for evaluating probabilistic forecasts, it has critical theoretical limitations. It has been shown that CRPS can produce systematically misleading evaluations by favoring probabilistic forecasts whose medians are close to the observed outcome, regardless of the actual probability assigned to that region, potentially resulting in higher scores for forecasts that allocate negligible (or even zero) probability mass to the true outcome. Furthermore, CRPS is not invariant under smooth transformations of the forecast variable, and its ranking of forecast systems may reverse under such transformations, raising concerns about its consistency for evaluation purposes.[16]

Multivariate continuous variables

[edit | edit source]

The scoring rules listed below aim to evaluate probabilistic predictions when the predicted distributions are univariate continuous probability distribution's, i.e. the predicted distributions are defined over a multivariate target variable Xβˆˆβ„n and have a probability density function f:ℝn→ℝ+.

Multivariate logarithmic score

[edit | edit source]

The multivariate logarithmic score is similar to the univariate logarithmic score:

L(D,y)=βˆ’ln(fD(y))

where fD denotes the probability density function of the predicted multivariate distribution D. It is a local, strictly proper scoring rule.

HyvΓ€rinen scoring rule

[edit | edit source]

The HyvΓ€rinen scoring function (of a density p) is defined by[17]

s(p)=2Ξ”ylogp(y)+β€–βˆ‡ylogp(y)β€–22

Where Ξ” denotes the Hessian trace and βˆ‡ denotes the gradient. This scoring rule can be used to computationally simplify parameter inference and address Bayesian model comparison with arbitrarily-vague priors.[17][18] It was also used to introduce new information-theoretic quantities beyond the existing information theory.[19]

Energy score

[edit | edit source]

The energy score is a multivariate extension of the continuous ranked probability score:[1]

ESΞ²(D,Y)=𝔼X∼D[β€–Xβˆ’Yβ€–2Ξ²]βˆ’12𝔼X,X∼D[β€–Xβˆ’Xβ€–2Ξ²]

Here, β∈(0,2), β€–β€–2 denotes the n-dimensional Euclidean distance and X,X are independently sampled random variables from the probability distribution D. The energy score is strictly proper for distributions D for which 𝔼X∼D[β€–Xβ€–2] is finite. It has been suggested that the energy score is somewhat ineffective when evaluating the intervariable dependency structure of the forecasted multivariate distribution.[20] The energy score is equal to twice the energy distance between the predicted distribution and the empirical distribution of the observation.

Variogram score

[edit | edit source]

The variogram score of order p is given by:[21]

VSp(D,Y)=βˆ‘i,j=1nwij(|Yiβˆ’Yj|pβˆ’π”ΌX∼D[|Xiβˆ’Xj|p])2

Here, wij are weights, often set to 1, and p>0 can be arbitrarily chosen, but p=0.5,1 or 2 are often used. Xi is here to denote the i'th marginal random variable of X. The variogram score is proper for distributions for which the (2p)'th moment is finite for all components, but is never strictly proper. Compared to the energy score, the variogram score is claimed to be more discriminative with respect to the predicted correlation structure.

Conditional continuous ranked probability score

[edit | edit source]

The conditional continuous ranked probability score (Conditional CRPS or CCRPS) is a family of (strictly) proper scoring rules. Conditional CRPS evaluates a forecasted multivariate distribution D by evaluation of CRPS over a prescribed set of univariate conditional probability distributions of the predicted multivariate distribution:[22]

CCRPS𝒯(D,Y)=βˆ‘i=1kCRPS(PX∼D(Xvi|Xj=Yj for jβˆˆπ’ži),Yvi)

Here, Xi is the i'th marginal variable of X∼D, 𝒯=(vi,π’ži)i=1k is a set of tuples that defines a conditional specification (with vi∈{1,…,n} and π’žiβŠ†{1,…,n}βˆ–{vi}), and PX∼D(Xvi|Xj=Yj for jβˆˆπ’ži) denotes the conditional probability distribution for Xvi given that all variables Xj for jβˆˆπ’ži are equal to their respective observations. In the case that PX∼D(Xvi|Xj=Yj for jβˆˆπ’ži) is ill-defined (i.e. its conditional event has zero likelihood), CRPS scores over this distribution are defined as infinite. Conditional CRPS is strictly proper for distributions with finite first moment, if the chain rule is included in the conditional specification, meaning that there exists a permutation Ο•1,…,Ο•n of 1,…,n such that for all 1≀i≀n: (Ο•i,{Ο•1,…,Ο•iβˆ’1})βˆˆπ’―.

Interpretation of proper scoring rules

[edit | edit source]

All proper scoring rules are equal to weighted sums (integral with a non-negative weighting functional) of the losses in a set of simple two-alternative decision problems that use the probabilistic prediction, each such decision problem having a particular combination of associated cost parameters for false positive and false negative decisions. A strictly proper scoring rule corresponds to having a nonzero weighting for all possible decision thresholds. Any given proper scoring rule is equal to the expected losses with respect to a particular probability distribution over the decision thresholds; thus the choice of a scoring rule corresponds to an assumption about the probability distribution of decision problems for which the predicted probabilities will ultimately be employed, with for example the quadratic loss (or Brier) scoring rule corresponding to a uniform probability of the decision threshold being anywhere between zero and one. The classification accuracy score (percent classified correctly), a single-threshold scoring rule which is zero or one depending on whether the predicted probability is on the appropriate side of 0.5, is a proper scoring rule but not a strictly proper scoring rule because it is optimized (in expectation) not only by predicting the true probability but by predicting any probability on the same side of 0.5 as the true probability.[23][24][25][26][27][28]

Examples of consistent scoring functions

[edit | edit source]

There are an infinite number of scoring functions including entire parameterized families of strictly consistent scoring functions for certain functionals T. The ones shown below are a selection of well known ones.

Expectation

[edit | edit source]

The following scoring functions are strictly consistent for the expected value, i.e. T(F)=EY∼F[Y].

Squared error

[edit | edit source]
S(x,y)=(xβˆ’y)2

Intervals

[edit | edit source]

The point prediction consists of a central (1βˆ’Ξ±) prediction interval, x=(l,u), where the lower endpoint l predicts the Ξ±2quantile and the upper endpoint u predicts the 1Ξ±2quantile.

Interval score

[edit | edit source]

The inverval score is a combination of the two pinball losses for the corresponding quantiles.

SΞ±(l,u;y)=(uβˆ’l)+2Ξ±(lβˆ’y)𝟏{y<l}+2Ξ±(yβˆ’u)𝟏{y>u}

"The forecaster is rewarded for narrow prediction intervals, and he or she incurs a penalty, the size of which de- pends on Ξ±, if the observation misses the interval"[1]

Applications

[edit | edit source]
File:LogScore.png
The logarithmic rule

Meteorological weather forecasts

[edit | edit source]

An example of probabilistic forecasting is in meteorology where a weather forecaster may give the probability of rain on the next day. One could note the number of times that a 25% probability was quoted, over a long period, and compare this with the actual proportion of times that rain fell. If the actual percentage was substantially different from the stated probability we say that the forecaster is poorly calibrated. A poorly calibrated forecaster might be encouraged to do better by a bonus system. A bonus system designed around a proper scoring rule will incentivize the forecaster to report probabilities equal to his personal beliefs.[3]

In addition to the simple case of a binary decision, such as assigning probabilities to 'rain' or 'no rain', scoring rules may be used for multiple classes, such as 'rain', 'snow', or 'clear', or continuous responses like the amount of rain per day.

The image shows an example of a scoring rule, the logarithmic scoring rule, as a function of the probability reported for the event that actually occurred. One way to use this rule would be as a cost based on the probability that a forecaster or algorithm assigns, then checking to see which event actually occurs.

Scoring rules can be used beyond evaluation metrics to directly serve as loss function to construct estimators.[1]

See also

[edit | edit source]

Literature

[edit | edit source]

References

[edit | edit source]
  1. ^ a b c d e f Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  2. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  3. ^ a b Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  4. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  5. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  6. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  7. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  8. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  9. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  10. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  11. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  12. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  13. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  14. ^ Countdown Regression: Sharp and Calibrated Survival Predictions, https://arxiv.org/abs/1806.08324
  15. ^ The Cramer Distance as a Solution to Biased Wasserstein Gradients https://arxiv.org/abs/1705.10743
  16. ^ Beyond Strictly Proper Scoring Rules: The Importance of Being Local https://doi.org/10.1175/WAF-D-19-0205.1
  17. ^ a b Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  18. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  19. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  20. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  21. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  22. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  23. ^ Leonard J. Savage. Elicitation of personal probabilities and expectations. J. of the American Stat. Assoc., 66(336):783–801, 1971.
  24. ^ Schervish, Mark J. (1989). "A General Method for Comparing Probability Assessors", Annals of Statistics 17(4) 1856–1879, https://projecteuclid.org/euclid.aos/1176347398
  25. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  26. ^ Roulston, M. S., & Smith, L. A. (2002). Evaluating probabilistic forecasts using information theory. Monthly Weather Review, 130, 1653–1660. See APPENDIX "Skill Scores and Cost–Loss". [1]
  27. ^ "Loss Functions for Binary Class Probability Estimation and Classification: Structure and Applications", Andreas Buja, Werner Stuetzle, Yi Shen (2005) http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.184.5203
  28. ^ Hernandez-Orallo, Jose; Flach, Peter; and Ferri, Cesar (2012). "A Unified View of Performance Metrics: Translating Threshold Choice into Expected Classification Loss." Journal of Machine Learning Research 13 2813–2869. http://www.jmlr.org/papers/volume13/hernandez-orallo12a/hernandez-orallo12a.pdf
[edit | edit source]