ALL Metrics
-
Views
-
Downloads
Get PDF
Get XML
Cite
Export
Track
Research Note

Optimal threshold estimation for binary classifiers using game theory

[version 1; peer review: 2 approved]
PUBLISHED 25 Nov 2016
Author details Author details
OPEN PEER REVIEW
REVIEWER STATUS

This article is included in the Bioinformatics gateway.

This article is included in the Machine learning: life sciences collection.

Abstract

Many bioinformatics algorithms can be understood as binary classifiers. They are usually trained by maximizing the area under the receiver operating characteristic (ROC) curve. On the other hand, choosing the best threshold for practical use is a complex task, due to uncertain and context-dependent skews in the abundance of positives in nature and in the yields/costs for correct/incorrect classification. We argue that considering a classifier as a player in a zero-sum game allows us to use the minimax principle from game theory to determine the optimal operating point. The proposed classifier threshold corresponds to the intersection between the ROC curve and the descending diagonal in ROC space and yields a minimax accuracy of 1-FPR. Our proposal can be readily implemented in practice, and reveals that the empirical condition for threshold estimation of “specificity equals sensitivity” maximizes robustness against uncertainties in the abundance of positives in nature and classification costs.

Keywords

Binary classifier, ROC curve, accuracy, optimal threshold, optimal cutoff, class imbalance, game theory, minimax principle.

Introduction

Many bioinformatics algorithms can be understood as binary classifiers, as they are used to investigate whether a query entity belongs to a certain class1. Score-based binary classifiers assign a number to the query. If this score surpasses a threshold, the query is assigned to the class under consideration. A minority of users are able to choose a threshold using their understanding of the algorithm, while the majority uses the default threshold.

Binary classifiers are often trained and compared under a unified framework, the receiver operating characteristic (ROC) curve2. Briefly, classifier output is first compared to a training set at all possible classification thresholds, yielding the confusion matrix with the number of true positives (TP), false positives (FP), true negatives (TN) and false negatives (FN) (Table 1). The ROC curve plots the true positive rate (TPR = TP/(TP + FN)), also called sensitivity,) against the false positive rate (FPR = FP/(FP + TN)) , which equals 1-specificity) (Figure 1, continuous line). Classifier training often aims at maximizing the area under the ROC curve, which amounts to maximizing the probability that a randomly chosen positive is ranked before a randomly chosen negative2. This summary statistic measures performance without committing to a threshold.

Table 1. Confusion matrix for training of a binary classifier.

TP: Number of true positives. FP: Number of false positives. FN: Number of false negatives. TN: Number of true negatives.

Training set
pn
Classifier outputp’TPFP
n’FNTN
67f130e7-cd88-4a03-a855-77b060b46464_figure1.gif

Figure 1. Optimal threshold estimation in ROC space for a binary classifier using game theory.

The descending diagonal TPR = 1 – FPR (dashed line) minimizes classifier performance with respect to qP. The intersection between the receiver operating characteristic (ROC) curve (continuous line) and this diagonal maximizes this minimal, worst-case utility and determines the optimal operating point according to the minimax principle (empty circle).

Practical application of a classifier requires using a threshold-dependent performance measure to choose the operating point1,3. This is in practice a complex task because the application domain may be skewed in two ways4. First, for many relevant bioinformatics problems the prevalence of positives in nature qP = (TP + FN)/(TP + TN + FP + FN) does not necessarily match the training set qP and is hard to estimate2,5. Second, the yields (or costs) for correct and incorrect classification of positives and negatives in the machine learning paradigm (YTP, YTN, YFP, YFN) may be different from each other and highly context-dependent1,3. Points in the ROC plane with equal performance are connected by iso-yield lines with a slope, the skew ratio, which is the product of the class skew and the yield skew4:

SKEWRATIO=qN.(YFP+YTN)qP.(YTP+YFN)(1)

The skew ratio expresses the relative importance of negatives and positives, regardless of the source of the skew4. Multiple threshold-dependent performance measures have been proposed and discussed in terms of skew sensitivity3,4, but often not justified from first principles.

Theory

Game theory allows us to consider a binary classifier as a zero-sum game between nature and the classifier6. In this game, nature is a player that uses a mixed strategy, with probabilities qP and qN=1-qP for positives and negatives, respectively. The algorithm is the second player, and each threshold value corresponds to a mixed strategy with probabilities pP and pN for positives and negatives. Two of the four outcomes of the game, TP and TN, favor the classifier, while the remaining two, FP and FN, favor nature. The game payoff matrix (Table 2) displays the four possible outcomes and the corresponding classifier utilities a, b, c and d. The Utility of the classifier within the game is:

UTILITY=a.TP+d.TN+b.FP+c.FNTP+TN+FP+FN(2)

Table 2. Payoff matrix for a zero-sum game between nature and a binary classifier.

a: Player I utility for a true positive. b: Player I utility for a false positive. c: Player I utility for a false negative. d: Player I utility for a true negative.

Player II: Nature
pn
Player I: Classifierp’ab
n’cd

The payoff matrix for this zero-sum game corresponds directly to the confusion matrix for the classifier, and the game utilities a, b, c, d correspond to the machine learning yields YTP, YFP, YFN, YTN, respectively (Table 1). Without loss of generality4, we can study the case a=d=1 and b=c=0. Classifier Utility within the game then reduces to the Accuracy or fraction of correct predictions24. In sum, maximizing the Utility of a binary classifier in a zero-sum game against nature is equivalent to maximizing its Accuracy, a common threshold-dependent performance measure.

We can now use the minimax principle from game theory6 to choose the operating point for the classifier. This principle maximizes utility for a player within a game using a pessimistic approach. For each possible action a player can take, we calculate a worst-case utility by assuming that the other player will take the action that gives them the highest utility (and the player of interest the lowest). The player of interest should take the action that maximizes this minimal, worst-case utility. Thus, the minimax utility of a player is the largest value that the player can be sure to get regardless of the actions of the other player.

In our classifier versus nature game, Utility/Accuracy of the classifier is skew-sensitive, depending on qP for a given threshold3,4:

UTILITY=1FPR+qP.(FPR+TPR1)(3)

The derivative of the Utility with respect to qP is zero along the TPR = 1 − FPR line in ROC space (Figure 1, dashed line). The derivative is negative below this line and positive above it, indicating that points along this line are minima of the Utility function with respect to the strategy qP of the nature player. According to the minimax principle, the classifier player should operate at the point along the TPR = 1 − FPR line that maximizes Utility. In ROC space, this condition corresponds to the intersection between the ROC curve and the descending diagonal (Figure 1, empty circle) and yields a minimax value of 1 − FPR for the Utility. It is worth noting that this analysis regarding class skew is also valid for yield/cost skew4.

Discussion

We showed that binary classifiers may be analyzed in terms of game theory. From the minimax principle, we propose a criterion to choose an operating point for the classifier that maximizes robustness against uncertainties in the skew ratio, i.e., in the prevalence of positives in nature and in yield skew, i.e., the yields/costs for true positives, true negatives, false positives and false negatives. This can be of practical value, since these uncertainties are widespread in bioinformatics and clinical applications.

In machine learning theory, TPR = 1 − FPR is the line of skew-indiference for Accuracy as a performance metric4. This is in agreement with the skew-indifference condition imposed by the minimax principle from game theory. However, to our knowledge, skew-indifference has not been exploited for optimal threshold estimation. Furthermore, the operating point of a classifier is often chosen by balancing sensitivity and specificity, without reference to the rationale behind7. Our game theory analysis shows that this empirical practice can be understood as a maximization of classifier robustness.

Comments on this article Comments (1)

Version 3
VERSION 3 PUBLISHED 08 Feb 2017
Revised
Version 1
VERSION 1 PUBLISHED 25 Nov 2016
Discussion is closed on this version, please comment on the latest version above.
Author details Author details
Competing interests
Grant information
Copyright
Download
 
Export To
metrics
Views Downloads
F1000Research - -
PubMed Central
Data from PMC are received and updated monthly.
- -
Citations
CITE
how to cite this article
Sanchez IE. Optimal threshold estimation for binary classifiers using game theory [version 1; peer review: 2 approved]. F1000Research 2016, 5(ISCB Comm J):2762 (https://doi.org/10.12688/f1000research.10114.1)
NOTE: If applicable, it is important to ensure the information in square brackets after the title is included in all citations of this article.
track
receive updates on this article
Track an article to receive email alerts on any updates to this article.

Open Peer Review

Current Reviewer Status: ?
Key to Reviewer Statuses VIEW
ApprovedThe paper is scientifically sound in its current form and only minor, if any, improvements are suggested
Approved with reservations A number of small changes, sometimes more significant revisions are required to address specific details and improve the papers academic merit.
Not approvedFundamental flaws in the paper seriously undermine the findings and conclusions
Version 1
VERSION 1
PUBLISHED 25 Nov 2016
Views
23
Cite
Reviewer Report 07 Dec 2016
Luis Diambra, Centro Regional de Estudios Genómicos, Universidad Nacional de La Plata (UNLP-CONICET), La Plata, Argentina 
Approved
VIEWS 23
The author presents a criterion to choose the operating point for a binary classifier. This criterion is analyzed in term of the game theory. By using the mininax principle author proposes to use as classifier threshold the intersection between the ROC curve and ... Continue reading
CITE
CITE
HOW TO CITE THIS REPORT
Diambra L. Reviewer Report For: Optimal threshold estimation for binary classifiers using game theory [version 1; peer review: 2 approved]. F1000Research 2016, 5(ISCB Comm J):2762 (https://doi.org/10.5256/f1000research.10895.r17994)
NOTE: it is important to ensure the information in square brackets after the title is included in all citations of this article.
Views
29
Cite
Reviewer Report 06 Dec 2016
Pieter Meysman, Department of Mathematics and Computer Science, University of Antwerp, Edegem, Belgium 
Approved
VIEWS 29
The article by Ignacio Enrique Sanchez concerns a common problem in machine learning, namely the selection of the optimal classification threshold, and provides a mathematical solution based on the principles of game theory. The main concern of the article deals ... Continue reading
CITE
CITE
HOW TO CITE THIS REPORT
Meysman P. Reviewer Report For: Optimal threshold estimation for binary classifiers using game theory [version 1; peer review: 2 approved]. F1000Research 2016, 5(ISCB Comm J):2762 (https://doi.org/10.5256/f1000research.10895.r17996)
NOTE: it is important to ensure the information in square brackets after the title is included in all citations of this article.

Comments on this article Comments (1)

Version 3
VERSION 3 PUBLISHED 08 Feb 2017
Revised
Version 1
VERSION 1 PUBLISHED 25 Nov 2016
Discussion is closed on this version, please comment on the latest version above.
Alongside their report, reviewers assign a status to the article:
Approved - the paper is scientifically sound in its current form and only minor, if any, improvements are suggested
Approved with reservations - A number of small changes, sometimes more significant revisions are required to address specific details and improve the papers academic merit.
Not approved - fundamental flaws in the paper seriously undermine the findings and conclusions
Sign In
If you've forgotten your password, please enter your email address below and we'll send you instructions on how to reset your password.

The email address should be the one you originally registered with F1000.

Email address not valid, please try again

You registered with F1000 via Google, so we cannot reset your password.

To sign in, please click here.

If you still need help with your Google account password, please click here.

You registered with F1000 via Facebook, so we cannot reset your password.

To sign in, please click here.

If you still need help with your Facebook account password, please click here.

Code not correct, please try again
Email us for further assistance.
Server error, please try again.