<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">582928</article-id>
<article-id pub-id-type="doi">10.3389/frai.2021.582928</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artificial Intelligence</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>A Modified AUC for Training Convolutional Neural Networks: Taking Confidence Into Account</article-title>
<alt-title alt-title-type="left-running-head">Namdar et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Modified AUC for Training CNNs</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Namdar</surname>
<given-names>Khashayar</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1028737/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Haider</surname>
<given-names>Masoom A.</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/997906/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Khalvati</surname>
<given-names>Farzad</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/950872/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Department of Medical Imaging, University of Toronto, <addr-line>Toronto</addr-line>, <addr-line>ON</addr-line>, <country>Canada</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>The Hospital for Sick Children (SickKids), <addr-line>Toronto</addr-line>, <addr-line>ON</addr-line>, <country>Canada</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>Lunenfeld-Tanenbaum Research Institute, Sinai Health System, <addr-line>Toronto</addr-line>, <addr-line>ON</addr-line>, <country>Canada</country>
</aff>
<aff id="aff4">
<label>
<sup>4</sup>
</label>Sunnybrook Research Institute, <addr-line>Toronto</addr-line>, <addr-line>ON</addr-line>, <country>Canada</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/531449/overview">Jake Y. Chen</ext-link>, University of Alabama at Birmingham, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/481771/overview">Akram Mohammed</ext-link>, University of Tennessee Health Science Center (UTHSC), United&#x20;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1337637/overview">Rubul Kumar Bania</ext-link>, North-Eastern Hill Uiveristy, Tura Campus, India</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Khashayar Namdar, <email>ernest.namdar@utoronto.ca</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Medicine and Public Health, a section of the journal Frontiers in Artificial Intelligence</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>30</day>
<month>11</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>4</volume>
<elocation-id>582928</elocation-id>
<history>
<date date-type="received">
<day>13</day>
<month>07</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>30</day>
<month>09</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Namdar, Haider and Khalvati.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Namdar, Haider and Khalvati</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Receiver operating characteristic (ROC) curve is an informative tool in binary classification and Area Under ROC Curve (AUC) is a popular metric for reporting performance of binary classifiers. In this paper, first we present a comprehensive review of ROC curve and AUC metric. Next, we propose a modified version of AUC that takes confidence of the model into account and at the same time, incorporates AUC into Binary Cross Entropy (BCE) loss used for training a Convolutional neural Network for classification tasks. We demonstrate this on three datasets: MNIST, prostate MRI, and brain MRI. Furthermore, we have published GenuineAI, a new python library, which provides the functions for conventional AUC and the proposed modified AUC along with metrics including sensitivity, specificity, recall, precision, and F1 for each point of the ROC&#x20;curve.</p>
</abstract>
<kwd-group>
<kwd>AUC</kwd>
<kwd>ROC</kwd>
<kwd>CNN</kwd>
<kwd>binary classification</kwd>
<kwd>loss function</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Classification is an important task in different fields, including Engineering, Social Science, and Medical Science. To evaluate quality of classification, a metric is needed. Accuracy, precision, and F1 score are three popular examples. However, there are other metrics that are more accepted in specific fields. For example, sensitivity and specificity are widely used in Medical Science.</p>
<p>For binary classification, Receiver Operating Characteristic (ROC) curve incorporates different evaluation metrics. The Area Under ROC Curve (AUC) is a widespread metric, especially in Medical Science (<xref ref-type="bibr" rid="B14">Sulam et&#x20;al., 2017</xref>). In engineering, AUC has been used to evaluate the classification models since the early 1990s (<xref ref-type="bibr" rid="B2">Burke et&#x20;al., 1992</xref>), and AUC research has continued ever since. Kottas <italic>et&#x20;al.</italic> proposed a method to report confidence intervals for AUC (<xref ref-type="bibr" rid="B7">Kottas et&#x20;al., 2014</xref>). Yu <italic>et&#x20;al.</italic> proposed a modified AUC which is customized for gene ranking (<xref ref-type="bibr" rid="B17">Yu et&#x20;al., 2018</xref>). Yu also proposed another version of AUC for penalizing regression models used for gene selection with high dimensional data (<xref ref-type="bibr" rid="B18">Yu and Park, 2014</xref>). Rosenfeld <italic>et&#x20;al.</italic> used AUC as a loss function and demonstrated AUC-based training lead to better generalization (<xref ref-type="bibr" rid="B12">Rosenfeld et&#x20;al., 2014</xref>). Their research, however, is not in the context of Neural Networks (NN); instead, they use Support Vector Machines (SVM). Therefore, their method does not address the challenges we address in this paper, including taking confidence of the model into account in calculating AUC and thus, making it a better metric for training neural networks. Zhao et&#x20;al. proposed an algorithm for AUC maximization in online learning (<xref ref-type="bibr" rid="B20">Zhao et&#x20;al., 2011</xref>). A stochastic approach for the same task was introduced by <xref ref-type="bibr" rid="B15">Ying et&#x20;al. (2016)</xref>. Cortes and Mohri studied correlation of AUC, as it is optimized, and error rate (<xref ref-type="bibr" rid="B3">Cortes and Mohri, 2004</xref>). Their research showed that minimizing the error rate may not result in maximizing AUC. Ghanbari and Scheinberg directly optimized error rate and AUC of the classifiers; however, their approach only applies to linear classifiers (<xref ref-type="bibr" rid="B4">Ghanbari and Scheinberg, 2018</xref>).</p>
<p>This paper explains in detail the meaning of AUC, how reliable it is, under which circumstances it should be used, and its limitations. It also proposes a novel approach to eliminate these limitations. Our primary focus is on deep learning and Convolutional Neural Networks (CNNs), which differentiates our work from the previous work in the literature. We propose confidence-incorporated AUC (cAUC) as a modified AUC which directly correlates to Cross-Entropy Loss function and thus, helps to stop CNN training at a more optimum point in terms of confidence. This is not possible with conventional AUC, as not only the minimum of Binary Cross-Entropy loss function may not correlate with the maximum of AUC, but also AUC does not take the confidence of the model into account. We have also published a new library called GeuineAI<xref ref-type="fn" rid="fn1">
<sup>1</sup>
</xref>, which contains our modified AUC and conventional AUC with more features in comparison to the existing standard python libraries.</p>
</sec>
<sec id="s2">
<title>Revisiting the Concept of AUC</title>
<p>In supervised binary classification, each datapoint has a label. Conformed with standards of Machine Learning, labels are either 0/1 or 01/10 or sometimes &#x2b;1/-1 and the model&#x2019;s (classifier&#x2019;s) outputs are usually probabilities. In the case of cancer detection, for example, input data may be CT or MRI images. Cancerous cases will be images labeled with 1 (positive) and normal (healthy) images will have 0 (negative) as their labels. The model returns a probability for each image. In the ideal scenario, the model&#x2019;s output will be 1 for cancerous images and 0 for normal&#x20;ones.</p>
<p>Four possible outcomes of binary classification are True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN). From <xref ref-type="table" rid="T1">Table&#x20;1</xref>, it can be inferred that TX means Truly predicted as X and FX means Falsely predicted as&#x20;X.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Possible outcomes of binary classification.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td align="left"/>
<td align="center">
<bold>Actual value</bold>
</td>
<td align="center">
<bold>Predicted value</bold>
</td>
</tr>
<tr>
<td align="left">TN</td>
<td align="center">0</td>
<td align="center">0</td>
</tr>
<tr>
<td align="left">FP</td>
<td align="center">0</td>
<td align="center">1</td>
</tr>
<tr>
<td align="left">FN</td>
<td align="center">1</td>
<td align="center">0</td>
</tr>
<tr>
<td align="left">TP</td>
<td align="center">1</td>
<td align="center">1</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Defined as the total number of correct predictions out of total cases, Accuracy is calculated by <xref ref-type="disp-formula" rid="e1">Equation (1)</xref>.<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>As it can be seen, accuracy is only concerned about correct versus wrong predictions. In many situations, especially in Medical Science, this is not enough. The consequences of misclassifying a normal case as cancerous and considering a cancerous case as normal are way different. The first one is referred to FP, also known as Type I error, whereas the second one is a FN or Type II error. True Positive Rate (TPR) and False Positive Rate (FPR) are two criterions which distinguish the error types.<disp-formula id="e2">
<mml:math id="m2">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>
<disp-formula id="e3">
<mml:math id="m3">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>TPR is also known as sensitivity and refers to the ratio of correct predictions to total within actual positives. FPR is the ratio of wrong predictions within actual negatives. FPR is related to specificity by <xref ref-type="disp-formula" rid="e4">Eq. 4</xref>, which is used frequently in Medical Science.<disp-formula id="e4">
<mml:math id="m4">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
</p>
<p>As mentioned before, predicted value should be binary, but output of the model is probability. Thresholding is how probabilities are converted to predicted values. As an example, if the output is 0.6 and the threshold is 0.5, predicted value is 1.<disp-formula id="e5">
<mml:math id="m5">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>y in <xref ref-type="disp-formula" rid="e5">Eq. 5</xref> is the predicted value, <italic>p</italic> is the output of the model, which is a probability, and <italic>t</italic> is the threshold. Depending on <italic>t</italic>, TPR and FPR will be different. ROC is the curve formed by plotting TPR versus FPR for all possible thresholds and AUC is the area under that&#x20;curve.</p>
<p>In the following, we take an example-based approach to highlight the fundamentals of&#x20;AUC.</p>
<p>Example 1: <xref ref-type="table" rid="T2">Table&#x20;2</xref> contains the simplest possible example. It should be followed from left to right. <inline-formula id="inf1">
<mml:math id="m6">
<mml:mrow>
<mml:msup>
<mml:mi>y</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> refers to the desired value which is the same as the label (ground truth).</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Example 1.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td rowspan="2" align="left">
<inline-formula id="inf2">
<mml:math id="m7">
<mml:mrow>
<mml:msup>
<mml:mi>y</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="4" align="center">
<inline-formula id="inf3">
<mml:math id="m8">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="4" align="center">
<inline-formula id="inf4">
<mml:math id="m9">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">
<inline-formula id="inf5">
<mml:math id="m10">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="left">
<inline-formula id="inf6">
<mml:math id="m11">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf7">
<mml:math id="m12">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="left">
<inline-formula id="inf8">
<mml:math id="m13">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">
<inline-formula id="inf9">
<mml:math id="m14">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="left">
<inline-formula id="inf10">
<mml:math id="m15">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf11">
<mml:math id="m16">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>It can be seen from <xref ref-type="table" rid="T2">Table&#x20;2</xref> that actual positives and actual negatives are necessary to draw an ROC curve. Although it may seem trivial, lack of one category in one batch leads to NaN in training of Machine Learning (ML) models. Furthermore, if the batch size is equal to one, the batch AUC is always NaN. Consequently, for any NN to be directly trained with a modified AUC, or for any code where AUC is calculated within each batch, batch size of one cannot be used. Furthermore, the sampler should be customized in a way to return samples from both classes in each&#x20;batch.</p>
<p>
<statement>
<p>Example 2: <xref ref-type="table" rid="T3">Table&#x20;3</xref> contains an example of classifying one positive and one negative cases and <xref ref-type="fig" rid="F1">Figure&#x20;1</xref> shows the corresponding ROC curve. There are important points in this example. ROC curves always start from (0,0) and always end at (1,1). The reason is that if threshold is 0, all predicted values are 1. They will be either TP or FP. Therefore, both TPR and FPR are 1. On the other hand, if threshold is 1, everything is predicted as negative. In this case, predictions are all TN or FN. Consequently, TPR and FPR will be both zero. Two things must be taken into account when writing a ML code: <italic>t</italic>&#x20;&#x3d; 0, and <italic>t</italic>&#x20;&#x3d; 1 should be treated separately and <italic>t</italic> should be iterated backward if going from (0, 0) to (1, 1) is desired. Backward iteration necessity comes from the fact that the highest t corresponds to the lowest TPR and FPR. Exceptions of <italic>t</italic>&#x20;&#x3d; 0 and <italic>t</italic>&#x20;&#x3d; 1 are needed for rare cases when the output of the model is exactly 0 or&#x20;1.</p>
<p>Example 3: Our third example is complement of Example 2. As it is indicated in <xref ref-type="table" rid="T4">Table&#x20;4</xref>, output probability for the positive case (<inline-formula id="inf12">
<mml:math id="m17">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>) is higher. Under these conditions, AUC is equal to 1, as depicted in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. In other words, ideal situation for classification of one positive and one negative example in terms of AUC is when output probability of the positive case is higher.</p>
<p>It should be noted that if the two probabilities were slightly different, e.g., <inline-formula id="inf13">
<mml:math id="m18">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.501</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf14">
<mml:math id="m19">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.499</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, AUC would be 1. The separation of probabilities does not have to be at 0.5. <inline-formula id="inf15">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.0002</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf16">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.0001</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> would still result in AUC &#x3d; 1. This leads to an important issue which is confidence. It turns out AUC does not take into account the confidence of the&#x20;model.</p>
</statement>
</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Example 2.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td align="left">
<inline-formula id="inf17">
<mml:math id="m22">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf18">
<mml:math id="m23">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf19">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf20">
<mml:math id="m25">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf21">
<mml:math id="m26">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="3" align="left">
<inline-formula id="inf22">
<mml:math id="m27">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf23">
<mml:math id="m28">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf24">
<mml:math id="m29">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf25">
<mml:math id="m30">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf26">
<mml:math id="m31">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf27">
<mml:math id="m32">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left"/>
</tr>
<tr>
<td align="left">
<inline-formula id="inf28">
<mml:math id="m33">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf29">
<mml:math id="m34">
<mml:mrow>
<mml:mn>0.4</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.6</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf30">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf31">
<mml:math id="m36">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf32">
<mml:math id="m37">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="3" align="left">
<inline-formula id="inf33">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.6</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf34">
<mml:math id="m39">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf35">
<mml:math id="m40">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf36">
<mml:math id="m41">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf37">
<mml:math id="m42">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
</td>
<td align="left">
</td>
<td align="center">
<inline-formula id="inf38">
<mml:math id="m43">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">
</td>
</tr>
<tr>
<td align="left"/>
<td rowspan="4" align="center">
<inline-formula id="inf39">
<mml:math id="m44">
<mml:mrow>
<mml:mn>0.6</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf40">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf41">
<mml:math id="m46">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf42">
<mml:math id="m47">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="center">
<inline-formula id="inf43">
<mml:math id="m48">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td rowspan="2" align="center">
<inline-formula id="inf44">
<mml:math id="m49">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf45">
<mml:math id="m50">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf46">
<mml:math id="m51">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="center">
<inline-formula id="inf47">
<mml:math id="m52">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>ROC of Example 2.</p>
</caption>
<graphic xlink:href="frai-04-582928-g001.tif"/>
</fig>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Example 3.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td align="left">
<inline-formula id="inf48">
<mml:math id="m53">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf49">
<mml:math id="m54">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf50">
<mml:math id="m55">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf51">
<mml:math id="m56">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf52">
<mml:math id="m57">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="3" align="left">
<inline-formula id="inf53">
<mml:math id="m58">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf54">
<mml:math id="m59">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf55">
<mml:math id="m60">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf56">
<mml:math id="m61">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf57">
<mml:math id="m62">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf58">
<mml:math id="m63">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left"/>
</tr>
<tr>
<td align="left">
<inline-formula id="inf59">
<mml:math id="m64">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf60">
<mml:math id="m65">
<mml:mrow>
<mml:mn>0.4</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.6</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf61">
<mml:math id="m66">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf62">
<mml:math id="m67">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf63">
<mml:math id="m68">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="3" align="left">
<inline-formula id="inf64">
<mml:math id="m69">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.6</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf65">
<mml:math id="m70">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf66">
<mml:math id="m71">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf67">
<mml:math id="m72">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf68">
<mml:math id="m73">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
</td>
<td align="left">
</td>
<td align="center">
<inline-formula id="inf69">
<mml:math id="m74">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">
</td>
</tr>
<tr>
<td align="left"/>
<td rowspan="4" align="center">
<inline-formula id="inf70">
<mml:math id="m75">
<mml:mrow>
<mml:mn>0.6</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf71">
<mml:math id="m76">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf72">
<mml:math id="m77">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf73">
<mml:math id="m78">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="center">
<inline-formula id="inf74">
<mml:math id="m79">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td rowspan="2" align="center">
<inline-formula id="inf75">
<mml:math id="m80">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf76">
<mml:math id="m81">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf77">
<mml:math id="m82">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="center">
<inline-formula id="inf78">
<mml:math id="m83">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>ROC of Example 3.</p>
</caption>
<graphic xlink:href="frai-04-582928-g002.tif"/>
</fig>
<p>
<statement>
<p>Example 4: In the fourth example (<xref ref-type="table" rid="T5">Table&#x20;5</xref>), the output probabilities are the same for the two samples. This leads to AUC of 0.50. This example shows that whenever all output probabilities are equal, AUC is 0.50 and ROC is a straight line from (0, 0) to (1, 1) (<xref ref-type="fig" rid="F3">Figure&#x20;3</xref>). This is true for all different values of N where N is batch size or number of samples.</p>
</statement>
</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Example 4.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td align="left">
<inline-formula id="inf79">
<mml:math id="m84">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf80">
<mml:math id="m85">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf81">
<mml:math id="m86">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf82">
<mml:math id="m87">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf83">
<mml:math id="m88">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="left">
<inline-formula id="inf84">
<mml:math id="m89">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf85">
<mml:math id="m90">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf86">
<mml:math id="m91">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf87">
<mml:math id="m92">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf88">
<mml:math id="m93">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf89">
<mml:math id="m94">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left"/>
</tr>
<tr>
<td rowspan="4" align="left">
<inline-formula id="inf90">
<mml:math id="m95">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="4" align="center">
<inline-formula id="inf91">
<mml:math id="m96">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf92">
<mml:math id="m97">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf93">
<mml:math id="m98">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf94">
<mml:math id="m99">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="3" align="center">
<inline-formula id="inf95">
<mml:math id="m100">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf96">
<mml:math id="m101">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf97">
<mml:math id="m102">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf98">
<mml:math id="m103">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf99">
<mml:math id="m104">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>ROC of Example 4.</p>
</caption>
<graphic xlink:href="frai-04-582928-g003.tif"/>
</fig>
<p>
<statement>
<p>Example 5: In example 5, N is equal to 3 and there are 4 points in the ROC curve (<xref ref-type="fig" rid="F4">Figures&#x20;4</xref>, <xref ref-type="fig" rid="F5">5</xref>). The reason for this phenomenon is effective threshold boundaries. As it can be seen in <xref ref-type="table" rid="T6">Table&#x20;6</xref>, up to t &#x3d; 0.4, no value of t changes the model&#x2019;s predictions. It turns out that those effective boundaries are defined by predicted probabilities. It should now be highlighted, in Examples 2 and 3, N was 2 and there were 3 points on the ROC curve. In the general form, for N predictions, there will be N&#x2b;1 points on the ROC curve. For each pair of predictions with equal probabilities, one point is omitted. The extreme case is when all output probabilities are equal. In this case, there will be two points on the ROC curve and AUC is 0.5 (Example&#x20;4).</p>
</statement>
</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>ROC of Example 5.</p>
</caption>
<graphic xlink:href="frai-04-582928-g004.tif"/>
</fig>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>ROC curves for N &#x3d; 3, two actual negative and an actual positive.</p>
</caption>
<graphic xlink:href="frai-04-582928-g005.tif"/>
</fig>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Example 5.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td align="left">
<inline-formula id="inf100">
<mml:math id="m105">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf101">
<mml:math id="m106">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf102">
<mml:math id="m107">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf103">
<mml:math id="m108">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf104">
<mml:math id="m109">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf105">
<mml:math id="m110">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf106">
<mml:math id="m111">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf107">
<mml:math id="m112">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="5" align="left">
<inline-formula id="inf108">
<mml:math id="m113">
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mn>3</mml:mn>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf109">
<mml:math id="m114">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf110">
<mml:math id="m115">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf111">
<mml:math id="m116">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>2</mml:mn>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf112">
<mml:math id="m117">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left"/>
</tr>
<tr>
<td rowspan="3" align="center">
<inline-formula id="inf113">
<mml:math id="m118">
<mml:mrow>
<mml:mn>0.4</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.45</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf114">
<mml:math id="m119">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf115">
<mml:math id="m120">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf116">
<mml:math id="m121">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf117">
<mml:math id="m122">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf118">
<mml:math id="m123">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="1" align="center">
<inline-formula id="inf119">
<mml:math id="m124">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf120">
<mml:math id="m125">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf121">
<mml:math id="m126">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
</td>
<td align="left">
</td>
<td align="left">
</td>
<td align="center">
<inline-formula id="inf122">
<mml:math id="m127">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf123">
<mml:math id="m128">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="3" align="center">
<inline-formula id="inf124">
<mml:math id="m129">
<mml:mrow>
<mml:mn>0.45</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>0.55</mml:mn>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf125">
<mml:math id="m130">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf126">
<mml:math id="m131">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf127">
<mml:math id="m132">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf128">
<mml:math id="m133">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.55</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf129">
<mml:math id="m134">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf130">
<mml:math id="m135">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="6" align="left">
<inline-formula id="inf131">
<mml:math id="m136">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.55</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf132">
<mml:math id="m137">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf133">
<mml:math id="m138">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="1" align="center">
<inline-formula id="inf134">
<mml:math id="m139">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf135">
<mml:math id="m140">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left"/>
</tr>
<tr>
<td rowspan="4" align="center">
<inline-formula id="inf136">
<mml:math id="m141">
<mml:mrow>
<mml:mn>0.55</mml:mn>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf137">
<mml:math id="m142">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf138">
<mml:math id="m143">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf139">
<mml:math id="m144">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf140">
<mml:math id="m145">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf141">
<mml:math id="m146">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="center">
<inline-formula id="inf142">
<mml:math id="m147">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<inline-formula id="inf143">
<mml:math id="m148">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td rowspan="2" align="center">
<inline-formula id="inf144">
<mml:math id="m149">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mn>0</mml:mn>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="center">
<inline-formula id="inf145">
<mml:math id="m150">
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec sec-type="methods" id="s3">
<title>Methods</title>
<p>Inspired by the previous examples, we will now investigate some characteristics of ROC and AUC. We will demonstrate how misclassification of a single data point can decrease AUC, and what extreme scenarios of misclassification look like. We will then provide an example to show a higher AUC does not necessarily correspond to better classification. The section is concluded with introducing cAUC, our proposed modified AUC, and mathematical support for its correlation to Binary Cross Entropy (BCE).</p>
<p>A result of having <italic>N&#x2b;1</italic> points on the ROC curve is that <italic>N&#x2b;1</italic> different effective values can be assigned to threshold <italic>t</italic>. In other words, while infinite values for <italic>t</italic> can be selected, selecting more than <italic>N&#x2b;1</italic> values for <italic>t</italic> would not help to achieve more accurate AUC or &#x201c;smoother&#x201d; ROC curve. Even if calculations are precise, the efficiency will be degraded because if t values are not selected from different effective intervals, they will result in the same point on ROC. In Example 3, <italic>t</italic>&#x20;&#x3d; 0, 0.1, 0.2, 0.3, or any other value below 0.4 will result in (1, 1) on ROC. Furthermore, because continuous variables have to be discretized, selecting fixed step size to increase <italic>t</italic> may result in inaccuracy. It happens almost certainly if two probabilities are highly close to each other and the fixed step is not small enough to land between them. Usually high values of N create such circumstances. Therefore, having a method for selecting optimal threshold is crucial. Changing value of <italic>t</italic> is effective if and only if it affects predictions. Assuming probabilities are sorted, any value of <italic>t</italic> between <inline-formula id="inf146">
<mml:math id="m151">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf147">
<mml:math id="m152">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> does not change predictions. Supported by the same rational, the optimal values of <italic>t</italic> we suggest is given by <xref ref-type="disp-formula" rid="e6">(6)</xref>. An optimum set, based on the rule of having <italic>N&#x2b;1</italic> points in ROC, has to have <italic>N&#x2b;1</italic> members. However, our proposed set has <italic>N&#x2b;2</italic> elements. If <xref ref-type="disp-formula" rid="e5">Eq. 5</xref> is conformed, 1 can be removed from the set. Nevertheless, adding 0 and 1 to the set is a safe approach for avoiding programming errors.<disp-formula id="e6">
<mml:math id="m153">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1,2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<p>
<xref ref-type="fig" rid="F5">Figure&#x20;5</xref> depicts all possible outcomes (except special cases of equal probabilities). It seems ROC is always staircase looking, except for the situations where a pair of predicted probabilities are equal. Thus, using trapezoid integration is the best and most accurate technique to calculate AUC. Furthermore, <xref ref-type="fig" rid="F5">Figure&#x20;5</xref> demonstrates order of predicted probabilities plays a key role in amount of AUC. If there is at least one threshold <italic>t</italic> where the probabilities of all actual positives and negatives are above and below it, respectively, then the AUC is equal to 1. Although the mathematical proof needs more fundamentals, there is one key support: selecting t at the boundary of positive and negative data points results in a perfect classification corresponding to (0, 1) on ROC.<disp-formula id="e7">
<mml:math id="m155">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2203;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x3c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>d</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2200;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>Where <italic>AP</italic> and <italic>AN</italic> are actual positives and actual negatives, respectively.</p>
<p>To be able to separate positive and negative datapoints in a way that probabilities of positive cases are higher, we introduce <italic>&#x3b5;</italic>. In <xref ref-type="table" rid="T7">Table&#x20;7</xref>, <italic>&#x3b5;</italic> is a positive real number which is less than or equal to <italic>p</italic>. This ensures <italic>p-&#x3b5;</italic> is zero or positive and implies that <italic>p-&#x3b5;</italic> is less than <italic>p</italic>. For example, if <italic>p</italic> is 0.8, <italic>&#x3b5;</italic> can be in the range of 0&#x2013;0.8. It also explains why <xref ref-type="disp-formula" rid="e7">(7)</xref> is true. For any <inline-formula id="inf148">
<mml:math id="m156">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#x2208;</mml:mo>
<mml:mo>[</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b5;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mi>p</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, conditions of <xref ref-type="disp-formula" rid="e7">(7)</xref> are met and the AUC is equal to 1. In this case, (0,0), (0,1), and (1,1) are points of&#x20;ROC.</p>
<table-wrap id="T7" position="float">
<label>TABLE 7</label>
<caption>
<p>A group of realizations with N &#x3d; 3, AN &#x3d; 2, and AP &#x3d; 1.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">&#x2014;</th>
<th colspan="2" align="center">t</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Sorted Actual Values</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">1</td>
</tr>
<tr>
<td align="left">Predicted Probabilities</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p</td>
</tr>
<tr>
<td align="left">&#x2014;</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TP</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In <xref ref-type="table" rid="T7">Table&#x20;7</xref>, probabilities of all actual negatives are equal (<italic>p-&#x3b5;</italic>). To be able to sort probabilities within each class of datapoints, <inline-formula id="inf149">
<mml:math id="m157">
<mml:mi>&#x3b4;</mml:mi>
</mml:math>
</inline-formula> is introduced. <xref ref-type="table" rid="T8">Table&#x20;8</xref> extends <xref ref-type="table" rid="T7">Table&#x20;7</xref> scenario to more general cases where probabilities are not necessarily equal. In this case, <inline-formula id="inf150">
<mml:math id="m158">
<mml:mi>&#x3b4;</mml:mi>
</mml:math>
</inline-formula> can be considered as a random noise which is a non-negative real number. It helps to simulate predicted probabilities better. With <inline-formula id="inf151">
<mml:math id="m159">
<mml:mi>&#x3b4;</mml:mi>
</mml:math>
</inline-formula>, the predicted probabilities do not follow a distinct pattern of having a fixed distance.</p>
<table-wrap id="T8" position="float">
<label>TABLE 8</label>
<caption>
<p>A group of realizations with N &#x3d; 8, AN &#x3d; 4, AP &#x3d; 4, and AUC &#x3d; 1.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
<th colspan="2" align="center">t</th>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Sorted Actual Values</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
</tr>
<tr>
<td align="left">Sorted Probabilities</td>
<td align="left">p-&#x3b5;-3&#x3b4;</td>
<td align="left">p-&#x3b5;-2&#x3b4;</td>
<td align="left">p-&#x3b5;-&#x3b4;</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p</td>
<td align="left">p&#x2b;&#x3b4;</td>
<td align="left">p&#x2b;2&#x3b4;</td>
<td align="left">p&#x2b;3&#x3b4;</td>
</tr>
<tr>
<td align="left">&#x2014;</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TP</td>
<td align="left">TP</td>
<td align="left">TP</td>
<td align="left">TP</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>
<xref ref-type="table" rid="T9">Table&#x20;9</xref> shows the other extreme. When there is threshold <italic>t</italic> such that probabilities of all actual positives and negatives are below and above it, respectively, then the AUC is zero.<disp-formula id="e8">
<mml:math id="m160">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2203;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>d</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2200;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x3c;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<table-wrap id="T9" position="float">
<label>TABLE 9</label>
<caption>
<p>A group of realizations with N &#x3d; 8, AN &#x3d; 4, AP &#x3d; 4, and AUC &#x3d; 0.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
<th colspan="2" align="center">t</th>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
<th align="center">&#x2014;</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Sorted actual values</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
</tr>
<tr>
<td align="left">Sorted probabilities</td>
<td align="left">p-3&#x3b4;</td>
<td align="left">p-2&#x3b4;</td>
<td align="left">p-&#x3b4;</td>
<td align="left">p</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p-&#x3b5;&#x2b;&#x3b4;</td>
<td align="left">p-&#x3b5;&#x2b;2&#x3b4;</td>
<td align="left">p-&#x3b5;&#x2b;3&#x3b4;</td>
</tr>
<tr>
<td align="left">&#x2014;</td>
<td align="left">FP</td>
<td align="left">FP</td>
<td align="left">FP</td>
<td align="left">FP</td>
<td align="left">FN</td>
<td align="left">FN</td>
<td align="left">FN</td>
<td align="left">FN</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>
<xref ref-type="table" rid="T10">Table&#x20;10</xref> depicts all remaining possible scenarios where AUC is greater than zero (0 &#x3c; AUC). <xref ref-type="table" rid="T10">Table&#x20;10</xref> gives the big picture. For <inline-formula id="inf152">
<mml:math id="m161">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#x2208;</mml:mo>
<mml:mo>[</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b5;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mtext>&#xa0;&#xa0;</mml:mtext>
<mml:mi>p</mml:mi>
<mml:mo>)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> there will be one FP in predicted values (<xref ref-type="table" rid="T10">Table&#x20;10</xref>), which means TPR is 1 and FPR is positive. For <inline-formula id="inf153">
<mml:math id="m162">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#x2208;</mml:mo>
<mml:mo>[</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>,</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mi>p</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, there will be one FN in predicted values (<xref ref-type="table" rid="T10">Table&#x20;10</xref>), which means FPR &#x3d; 0 and TPR less than one. In other words, in the ROC curve, (<xref ref-type="table" rid="T10">Tables&#x20;10(b),(c)</xref>) correspond to points (<inline-formula id="inf154">
<mml:math id="m163">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>,1) and (0, <inline-formula id="inf155">
<mml:math id="m164">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>), respectively, where <inline-formula id="inf156">
<mml:math id="m165">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf157">
<mml:math id="m166">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are positive real numbers (<xref ref-type="fig" rid="F6">Figure&#x20;6</xref>). Obviously, this causes a reduction in AUC as much as the area of a triangle. (0, <inline-formula id="inf158">
<mml:math id="m167">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>), (<inline-formula id="inf159">
<mml:math id="m168">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>,1), and (0, 1) are vertices of the triangle. FN contributes to TPR whereas FP is part of FPR. Therefore, <inline-formula id="inf160">
<mml:math id="m169">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is influenced by FP and <inline-formula id="inf161">
<mml:math id="m170">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, is a function of FN. Because they both play a role in the triangle&#x2019;s area, it can be concluded that the AUC does not discriminate between FP and FN. All it does is scaling the importance with respect to degree of imbalance. In other words, AUC equalizes importance of positive and negative cases as if the number of <italic>AP</italic>s and <italic>AN</italic>s were the same. In this perspective, ROC has a built-in normalizer mechanism. However, in real world, that may not be desired. In most cancer detection situations, for example, importance of a positive case massively outweighs that of a negative&#x20;case.</p>
<table-wrap id="T10" position="float">
<label>TABLE 10</label>
<caption>
<p>A group of realizations with N &#x3d; 8, AN &#x3d; 4, AP &#x3d; 4, and 0 &#x3c; AUC&#x3c;1.</p>
</caption>
<table>
<tbody valign="top">
<tr>
<td rowspan="4" align="left">(a)</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td colspan="2" align="left">t</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td/>
</tr>
<tr>
<td align="left">sorted Actual values</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td/>
</tr>
<tr>
<td align="left">sorted Probabilities</td>
<td align="left">p-&#x3b5;-2&#x3b4;</td>
<td align="left">p-&#x3b5;-&#x3b4;</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p</td>
<td align="left">p</td>
<td align="left">p&#x2b;&#x3b4;</td>
<td align="left">p&#x2b;2&#x3b4;</td>
<td align="left">p&#x2b;3&#x3b4;</td>
<td/>
</tr>
<tr>
<td/>
<td align="left">&#x2014;</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">
</td>
<td align="left">
</td>
<td align="left">TP</td>
<td align="left">TP</td>
<td align="left">TP</td>
</tr>
<tr>
<td rowspan="4" align="left">(b)</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">t</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left"/>
<td/>
</tr>
<tr>
<td align="left">sorted Actual values</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td/>
</tr>
<tr>
<td align="left">sorted Probabilities</td>
<td align="left">p-&#x3b5;-2&#x3b4;</td>
<td align="left">p-&#x3b5;-&#x3b4;</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p</td>
<td align="left">p</td>
<td align="left">p&#x2b;&#x3b4;</td>
<td align="left">p&#x2b;2&#x3b4;</td>
<td align="left">p&#x2b;3&#x3b4;</td>
<td/>
</tr>
<tr>
<td/>
<td align="left">&#x2014;</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">FP</td>
<td align="left">TP</td>
<td align="left">TP</td>
<td align="left">TP</td>
<td align="left">TP</td>
</tr>
<tr>
<td rowspan="6" align="left">(c)</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left">t</td>
<td align="left">&#x2014;</td>
<td align="left">&#x2014;</td>
<td align="left"/>
<td/>
</tr>
<tr>
<td align="left">sorted Actual values</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">0</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td align="left">1</td>
<td/>
</tr>
<tr>
<td align="left">sorted Probabilities</td>
<td align="left">p-&#x3b5;-2&#x3b4;</td>
<td align="left">p-&#x3b5;-&#x3b4;</td>
<td align="left">p-&#x3b5;</td>
<td align="left">p</td>
<td align="left">p</td>
<td align="left">p&#x2b;&#x3b4;</td>
<td align="left">p&#x2b;2&#x3b4;</td>
<td align="left">p&#x2b;3&#x3b4;</td>
<td/>
</tr>
<tr>
<td align="left">&#x2014;</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">TN</td>
<td align="left">FN</td>
<td align="left">TP</td>
<td align="left">TP</td>
<td align="left">TP</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>ROC of Example of <xref ref-type="table" rid="T10">Table&#x20;10</xref>.</p>
</caption>
<graphic xlink:href="frai-04-582928-g006.tif"/>
</fig>
<p>The fact that the AUC does not discriminate between FP and FN implies that what should be used as a criterion when training a model is ROC curve itself and not the AUC. Hence, in order to translate probabilities to predictions, one specific <italic>t</italic> <inline-formula id="inf162">
<mml:math id="m171">
<mml:mrow>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is needed.</p>
<p>In medical science (e.g., cancer detection), instead of AUC value, the clinical value of a classification method is usually studied in terms of TPR or FPR. For example, for a desired TPR, using the ROC curve, the point with lowest FPR is selected. From there, the desired threshold is derived, and the classification is performed. Thus, to evaluate the performance, confusion matrix is the most informative way of reporting where a model with a lower AUC may be preferred when the specific TPR/FPR are considered. One possible example is illustrated in <xref ref-type="fig" rid="F7">Figure&#x20;7</xref>.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>ROC curves for two different models with N &#x3d; 7.</p>
</caption>
<graphic xlink:href="frai-04-582928-g007.tif"/>
</fig>
<p>In <xref ref-type="fig" rid="F7">Figure&#x20;7</xref>, AUC of the orange line and the blue line are 0.75 and 0.58, respectively. Although the orange line has a higher AUC, if the acceptable sensitivity is set at 1, the blue line corresponds to the best model. In other words, to be able to identify every single positive example, with the orange line we will misclassify 75% of our negative examples compared with 50% of misclassification by the blue&#x20;one.</p>
<sec id="s3-1">
<title>Proposed AUC With Confidence</title>
<p>We call a model confident if it returns probabilities near 1 for all positive cases and probabilities near 0 for all negative examples. In previous section, it was demonstrated that AUC does not provide the confidence of the classification model under study. In other words, whether the predicted probabilities are close to each other or not does not affect the AUC value. As a result, a classification model that is able to separate the positive and negative cases by a small margin (e.g., 5%), has the same AUC as the one that separates the positive and negative cases by a large margin (e.g., 25%). Risk assessment in Medical Science and regression in Statistics are cases where having large margins may not be the target. However, in the context of classification, the margin is a key point. The whole idea of Support Vector Machines (SVM) is formed around large margin classification (<xref ref-type="bibr" rid="B11">Parikh and Shah, 2016</xref>). The ultimate effect of Cross Entropy (CE) loss function on NNs is imposing separation between predicted probability of&#x20;positive and negative examples (<xref ref-type="bibr" rid="B19">Zhang and Sabuncu, 2018</xref>).</p>
<p>To address this issue, we propose a modified AUC (cAUC), which provides a confidence measure for the classification model. To do so, we introduce two coefficients, <inline-formula id="inf163">
<mml:math id="m172">
<mml:mi>&#x3b1;</mml:mi>
</mml:math>
</inline-formula> and <inline-formula id="inf164">
<mml:math id="m173">
<mml:mi>&#x3b2;</mml:mi>
</mml:math>
</inline-formula>.<disp-formula id="e9">
<mml:math id="m174">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>max</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>min</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>
<disp-formula id="e10">
<mml:math id="m175">
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>min</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>max</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>
<disp-formula id="e11">
<mml:math id="m176">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>
</p>
<p>The idea behind <xref ref-type="disp-formula" rid="e11">Eq. 11</xref> is the smaller the range between the probabilities of the two classes, the lower the AUC will be and vice versa. If the range is the maximum possible value (which is 1), the AUC remains unchanged. Otherwise, it is decreased.</p>
<p>In the following, we show that our cAUC local maximums correspond to BCE local minimums. Intuitively, BCE is minimized when the probabilities created by the model are close to 1 for APs and near 0 for ANs. This translates to the concept of confidence we discussed above. Mathematically, BCE is explained through <xref ref-type="disp-formula" rid="e12">Eq. 12</xref>. Using the same separation approach, we have used so far, BCE can be rewritten for APs and ANs as <xref ref-type="disp-formula" rid="e13">Eq.&#x20;13</xref>. From <xref ref-type="disp-formula" rid="e13">Eq. 13</xref>, it can be concluded ideal BCE loss is resulted under conditions of <xref ref-type="disp-formula" rid="e14">Eq. 14</xref>.<disp-formula id="e12">
<mml:math id="m177">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>&#xa0;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mstyle>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>
<disp-formula id="e13">
<mml:math id="m178">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:mstyle>
<mml:mo>&#x2b;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(13)</label>
</disp-formula>
<disp-formula id="e14">
<mml:math id="m179">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2200;</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>d</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2200;</mml:mo>
<mml:msubsup>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>d</mml:mi>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
<label>(14)</label>
</disp-formula>
</p>
<p>If conditions of <xref ref-type="disp-formula" rid="e14">Eq. 14</xref> are met, from <xref ref-type="disp-formula" rid="e7">Eq. 7</xref> it can be inferred AUC is equal to 1 because for any threshold between 0 and 1, all datapoints are correctly classified. In this case <xref ref-type="disp-formula" rid="e9">Eq. 9</xref>, <xref ref-type="disp-formula" rid="e10">10</xref> result in <inline-formula id="inf165">
<mml:math id="m180">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. Ultimately, our definition of cAUC, <xref ref-type="disp-formula" rid="e11">Eq. 11</xref>, returns <inline-formula id="inf166">
<mml:math id="m181">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. Therefore, the ideal cases of cAUC and BCE correspond to each other. Through a similar procedure, it can be proved their worst cases (cAUC &#x3d; 0 and BCE <inline-formula id="inf167">
<mml:math id="m182">
<mml:mrow>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>) correspond too. In the transition between the two extremes, BCE and confidence-related part of cAUC (the exponential coefficients) have a monotonic behavior.</p>
<p>We proved that if AUC is equal to 1, the probability of positive and negative examples can be close to each other and thus, leading to high BCE. Therefore, a high AUC does not necessarily mean low BCE. Thus, instead of AUC, we propose monitoring cAUC, which in global optimums is guaranteed to result in ideal BCE and AUC, and in local optimums has higher potential for stopping the training when the model is confident, not overfit, and achieves a high&#x20;AUC.</p>
</sec>
</sec>
<sec sec-type="results" id="s4">
<title>Results</title>
<p>We will evaluate our confidence-incorporated AUC (cAUC) on 4 different scenarios: random predictions, a customized dataset based on MNIST (<xref ref-type="bibr" rid="B9">LeCun and Cortes, 2010</xref>), our proprietary Prostate Cancer (PCa) dataset, and a dataset based on BraTS19 (<xref ref-type="bibr" rid="B10">Menze et&#x20;al., 2015</xref>; <xref ref-type="bibr" rid="B1">Bakas et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B13">Bakas, 2018</xref>). Our PCa dataset of Diffusion-weighted MRI is described in our previous research (<xref ref-type="bibr" rid="B16">Yoo et&#x20;al., 2019</xref>). The CNN architectures and the utilized settings are similar to our shallow models used in other research projects (<xref ref-type="bibr" rid="B5">Hao et&#x20;al., 2020</xref>). Nonetheless, the details are provided in <xref ref-type="sec" rid="s11">Supplementary Appendix A</xref>. Given the fact that AUC is not differentiable, to train the network we used BCE. The only essential point which should be covered is input channels of our CNN for MNIST classification. Because MNIST is a single channel dataset, we revised the network to be compatible with&#x20;it.</p>
<sec id="s4-1">
<title>cAUC vs AUC on Random Data</title>
<p>To test the proposed AUC, in an N &#x3d; 10 simulation, real values and predicted probabilities were generated randomly using U [0, 1] as <xref ref-type="table" rid="T11">Table&#x20;11</xref>. In case of arbitrary classification, expected value of AUC is 0.5. The goal here is to calculate expected values of cAUC for such conditions. Another point for the presented values in <xref ref-type="table" rid="T11">Table&#x20;11</xref> is to highlight importance of sample size. With the widespread use of AI in Medical Science, researchers must care about sample sizes. Our experiment shows AUC &#x3d; 0.66 is not hard to achieve through chance when N is not high enough.</p>
<table-wrap id="T11" position="float">
<label>TABLE 11</label>
<caption>
<p>Comparison of AUC and the proposed AUC for a random&#x20;case.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Real values</th>
<th align="center">Sorted probabilities</th>
<th align="center">Parameters</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">1</td>
<td>0.803258838</td>
<td align="center">
<inline-formula id="inf168">
<mml:math id="m183">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.80325884</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>0.27759354</mml:mn>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5256653</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="left">0</td>
<td rowspan="2" align="char" char=".">0.517853202</td>
<td align="center">
<inline-formula id="inf169">
<mml:math id="m184">
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.30374599</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>0.69960646</mml:mn>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>0.39586047</mml:mn>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="center">
<inline-formula id="inf170">
<mml:math id="m185">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.6666666666666666</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="left">1</td>
<td rowspan="2" align="char" char=".">0.639592674</td>
</tr>
<tr>
<td rowspan="2" align="center">
<inline-formula id="inf171">
<mml:math id="m186">
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>U</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.1027290563696407</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td rowspan="2" align="left">1</td>
<td rowspan="2" align="char" char=".">0.303745995</td>
</tr>
<tr>
<td rowspan="7" align="center">&#x2014;</td>
</tr>
<tr>
<td align="left">0</td>
<td align="char" char=".">0.699606458</td>
</tr>
<tr>
<td align="left">0</td>
<td align="char" char=".">0.318090495</td>
</tr>
<tr>
<td align="left">0</td>
<td align="char" char=".">0.277593543</td>
</tr>
<tr>
<td align="left">1</td>
<td align="char" char=".">0.421482502</td>
</tr>
<tr>
<td align="left">1</td>
<td align="char" char=".">0.556011119</td>
</tr>
<tr>
<td align="left">1</td>
<td align="char" char=".">0.548716153</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Simulations with N &#x3d; 100 and 10,000 trials show expected value of AUC is 0.50 and expected value of the revised AUC is 0.07. Intuitively, AUC &#x3d; 0.5 happens when everything is by chance. We showed one example is when output of the model is constant. In other words, when variance of the output vector is zero. In this case, coefficients <inline-formula id="inf172">
<mml:math id="m187">
<mml:mi>&#x3b1;</mml:mi>
</mml:math>
</inline-formula> and <inline-formula id="inf173">
<mml:math id="m188">
<mml:mi>&#x3b2;</mml:mi>
</mml:math>
</inline-formula> also are zero in limit [according to <xref ref-type="disp-formula" rid="e9">(9)</xref> and <xref ref-type="disp-formula" rid="e10">(10)</xref>]. Therefore, cAUC will be <inline-formula id="inf174">
<mml:math id="m189">
<mml:mrow>
<mml:mn>0.5</mml:mn>
<mml:mtext>&#x2a;</mml:mtext>
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> which is&#x20;0.07.</p>
</sec>
<sec id="s4-2">
<title>cAUC vs. AUC on an MNIST-Based Dataset</title>
<p>MNIST is a well-known dataset of handwritten digits, including 60,000 train and 10,000 test images (<xref ref-type="bibr" rid="B9">LeCun and Cortes, 2010</xref>). It includes single channel, 28&#x20;&#xd7; 28 pixel, normalized images. The 10 different digits form classes of data in MNIST, by default. Because our ultimate goal was Medical applications, we marked examples of 7 as positive and all other digits as negative to create our imbalanced binary MNIST-based dataset. Our train set included the first 5,000 examples of training cohort of MNIST and our validation set was 1,500 examples (indices: 45,000&#x2013;46,500) of it. Our test set was built from the first 1,000 examples of MNIST test. This was done to ensure our dataset size is reasonable in comparison to Medical ones. To make our data noisy, as it is always seen in Medical datasets, we added uniform random noise to each pixel. For that end, we first scaled MNIST examples in order to have each pixel values in the range of [0, 1]. Then we added 5&#x20;times of a random image to it and scaled the result back to [0, 1] as stated in <xref ref-type="disp-formula" rid="e15">Eq. 15</xref> <disp-formula id="e15">
<mml:math id="m190">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>I</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>T</mml:mi>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>255</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x2b;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>y</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>d</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>m</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>d</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>28,28</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>6</mml:mn>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(15)</label>
</disp-formula>
</p>
<p>
<xref ref-type="fig" rid="F8">Figure&#x20;8</xref> shows results of the classification over 50 epochs of training. In each epoch, average BCE loss, AUC, and cAUC for training, validation, and test cohorts are calculated. This procedure is maintained until the last epoch and then the monitored values are plotted.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Classification results on the MINIST-based dataset.</p>
</caption>
<graphic xlink:href="frai-04-582928-g008.tif"/>
</fig>
</sec>
<sec id="s4-3">
<title>cAUC vs. AUC on a Proprietary PCa Dataset</title>
<p>
<xref ref-type="fig" rid="F9">Figure&#x20;9</xref> depicts the results of classification over our institutional review board approved PCa dataset, which included Diffusion-weighted MRI images of 414 prostate cancer patients (5,706 2D slices). The dataset was divided into training (217 patients, 2,955 slices), validation (102 patients, 1,417 slices), and test sets (95 patients, 1,334 slices). Label for each slice was generated based on the targeted biopsy results where a clinically significant prostate cancer (Gleason score&#x3e;6) was considered a positive label. The golden vertical line is where cAUC guides us to stop and the grey vertical line is where we would stop if AUC was&#x20;used.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Classification results on the PCa dataset.</p>
</caption>
<graphic xlink:href="frai-04-582928-g009.tif"/>
</fig>
</sec>
<sec id="s4-4">
<title>cAUC vs. AUC on a BraTS-Based Dataset</title>
<p>We used the BraTS19 dataset, with the same setting as our previous research (<xref ref-type="bibr" rid="B6">Hao et&#x20;al., 2021</xref>). The dataset contains 335 patients of which 259 patients were diagnosed with high-grade glioma (HGG) and 76 patients had low-grade glioma (LGG). For each patient, we stacked three MRI sequences, which are T1-weighted, post&#x2013;contrast-enhanced T1-weighted (T1C), and T2-weighted (T2) volumes. With the help of BraTS segmentations, we randomly extracted 20 slices per patient with the tumor region in axial plane. Our training dataset contained 203 patients, which corresponds to 2,927 slices (1,377 LGG and 1,550 HGG examples). 66 patients were included in the validation set (970 slices, 450 LGG and 520 HGG examples). Another 66 patients formed our test set (970 slices, 450 LGG and 520 HGG examples). LGG slices were labeled as 0 and HGGs were assigned to be 1. The images were resized to 224&#x20;&#xd7; 224 pixels. <xref ref-type="fig" rid="F10">Figure&#x20;10</xref> illustrates the results of classification over the dataset. cAUC directs the model to stop at epoch number 4 whereas both AUC and BCE would lead to the seventh&#x20;epoch.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Classification results on our BraTS-based dataset.</p>
</caption>
<graphic xlink:href="frai-04-582928-g010.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s5">
<title>Discussion</title>
<p>In this research, we first highlighted several important ROC and AUC characteristics. We demonstrated that to draw ROC curve, both actual positives and actual negatives are needed. Threshold equal to 1 corresponds to (0,0) in the ROC curve and <italic>t</italic>&#x20;&#x3d; 0 appears as (1,1). If a function is to calculate TPR, FPR or other metrics, it should iterate backward on the <italic>t</italic> values. The AUC is not concerned about confidence of the model. Regardless of N, if all the predictions are the same (<inline-formula id="inf175">
<mml:math id="m191">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>N</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>), AUC will be 0.5 and the ROC curve will be a straight line from (0,0) to (1,1). Selecting more thresholds does not result in a smoother ROC or more accurate AUC. Thresholds must be selected from the set of the predicted probabilities plus 0 and 1. The order of predicted probabilities is correlated to the ROC shape and has a major impact on AUC. If there is at least a threshold where the probabilities of all actual positives and all actual negatives are above and below it, respectively, then the AUC is equal to 1. Conversely, the AUC will be 0 for the opposite case. The AUC does not differentiate FP from FN. All it does is scaling actual positive and actual negatives in a way that they have equal contributions to AUC. Therefore, the ROC curve should be used as the criterion and not AUC, if FP and FN have different weights. Because the final goal is classification, what is important is the performance of the model at a specific threshold. Therefore, there may be cases where a model with a lower AUC performs better at one threshold. The right approach is finding the optimum threshold from ROC and reporting the confusion matrix at that threshold.</p>
<p>The core of our research was the amendment of AUC in terms of margins. To add confidence to the optimized model, AUC needs to be refined. Using two coefficients, a revised AUC was proposed. Through simulations and mathematics, we showed the revised AUC reflects confidence of the&#x20;model.</p>
<p>Unlike AUC, through experiments on MNIST, our PCa, and BraTS dataset, we demonstrated that local maximums in the proposed modified AUC correspond to local minimums of cross-entropy loss function. It was shown that selecting the best model based on cAUC is computationally efficient, mathematically reasonable, and it results in avoiding overfitting.</p>
<p>The conventional approach for when to stop training a CNN to achieve the highest AUC is to monitor the AUC while the model is being trained with a loss function such as BCE, and save the model whenever AUC breaks the previous highest score. However, when BCE is set to be used as the loss function, the hypothesis is that the best model has the lowest loss and therefore, the minimum loss is what the model is trained for. Hence, choosing the best model based on the highest AUC is not well rationalized and may not lead to the optimum&#x20;point.</p>
<p>Our proposed metric inherits several limitations of the standard AUC and ROC but does not add any additional restrictions. Similar to AUC, cAUC is not differentiable and cannot be directly used as a loss function for training any NN. Additionally, calculating cAUC for a batch of data, especially if the batch size is small, will not help because it will be a measure of ranking in a small sample of the dataset. Similar to the standard AUC, cAUC does not give more importance to the positive examples.</p>
</sec>
<sec sec-type="conclusion" id="s6">
<title>Conclusion</title>
<p>Our results demonstrate the proposed cAUC is a better metric to choose the best performing model. On our MNIST-based dataset, when training a CNN, it results in stopping earlier which is computationally desirable. Moreover, it has landed in a less overfitting-prone area. Our results on the prostate MRI dataset are particularly interesting. With standard AUC we would stop training the CNN model at a suboptimal point with regards to BCE. With our proposed cAUC, we are able to stop at an optimal point where the training model gives the highest AUC. Our BraTS dataset experiments demonstrate cAUC can indicate optimum points that neither AUC nor BCE would direct the model towards&#x20;them.</p>
</sec>
</body>
<back>
<sec id="s7">
<title>Data Availability Statement</title>
<p>The data analyzed in this study is subject to the following licenses/restrictions: Three datasets have been used for the research. The MNIST and the BraTS datasets are publicly available. The prostate dataset analyzed in this research is available from the corresponding author on reasonable request pending the approval of the institution(s) and trial/study investigators who contributed to the dataset. Requests to access these datasets should be directed to <email>mahaider@radfiler.com</email>.</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>KN and FK contributed to the design of the concept and implementation of the algorithms. MH contributed in collecting and reviewing the data. All authors contributed to the writing and reviewing of the manuscript. All authors read and approved the final manuscript.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s11">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/frai.2021.582928/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/frai.2021.582928/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="Presentation1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<fn-group>
<fn id="fn1">
<label>1</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://pypi.org/project/GenuineAI/">https://pypi.org/project/GenuineAI/</ext-link>
</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bakas</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Akbari</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Sotiras</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bilello</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Rozycki</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kirby</surname>
<given-names>J.&#x20;S.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Advancing the Cancer Genome Atlas Glioma MRI Collections with Expert Segmentation Labels and Radiomic Features</article-title>. <source>Sci. Data</source> <volume>4</volume>, <fpage>170117</fpage>. <pub-id pub-id-type="doi">10.1038/sdata.2017.117</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bakas</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Reyes</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jakab</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bauer</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Rempfler</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Crimi</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Identifying the Best Machine Learning Algorithms for Brain Tumor Segmentation, Progression Assessment, and Overall Survival Prediction in the BRATS Challenge</article-title>. </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bottou</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Curtis</surname>
<given-names>F. E.</given-names>
</name>
<name>
<surname>Nocedal</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Optimization Methods for Large-Scale Machine Learning</article-title>. <source>SIAM Review</source> <volume>60</volume> (<issue>2</issue>), <fpage>223</fpage>&#x2013;<lpage>311</lpage>. <pub-id pub-id-type="doi">10.1137/16M1080173</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Burke</surname>
<given-names>H. B.</given-names>
</name>
<name>
<surname>Ph</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Rosen</surname>
<given-names>D. B.</given-names>
</name>
<name>
<surname>Ph</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Goodman</surname>
<given-names>P. H.</given-names>
</name>
</person-group> (<year>1992</year>). <source>Comparing Artificial Neural Nrworks to Other Statistical Memods for Medical Outcome Prediction</source>, <fpage>2213</fpage>&#x2013;<lpage>2216</lpage>. </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cortes</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Mohri</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>AUC Optimization vs. Error Rate Minimization</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> </citation>
</ref>
<ref id="B4">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Ghanbari</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Scheinberg</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Directly and Efficiently Optimizing Prediction Error and AUC of Linear Classifiers</article-title>. <comment>[Online]. Available: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1802.02535">http://arxiv.org/abs/1802.02535</ext-link>
</comment>. </citation>
</ref>
<ref id="B5">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hao</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Namdar</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Haider</surname>
<given-names>M. A.</given-names>
</name>
<name>
<surname>Khalvati</surname>
<given-names>F. A.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A Comprehensive Study of Data Augmentation Strategies for Prostate Cancer Detection in Diffusion-Weighted MRI Using Convolutional Neural Networks</article-title>. <source>J. Digit Imaging</source> <volume>34</volume> (<issue>4</issue>), <fpage>862</fpage>&#x2013;<lpage>876</lpage>. <pub-id pub-id-type="doi">10.1007/s10278-021-00478-7</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hao</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Namdar</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Khalvati</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A Transfer Learning-Based Active Learning Framework for Brain Tumor Classification</article-title>. <source>Front. Artif. Intell.</source> <volume>4</volume>, <fpage>61</fpage>. <pub-id pub-id-type="doi">10.3389/frai.2021.635766</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kottas</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kuss</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Zapf</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>A Modified Wald Interval for the Area under the ROC Curve (AUC) in Diagnostic Case-Control Studies</article-title>. <source>BMC Med. Res. Methodol</source>. <volume>14</volume> (<issue>1</issue>), <fpage>1</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1186/1471-2288-14-26</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>LeCun</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Cortes</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>{MNIST} Handwritten Digit Database</article-title>. <comment>[Online]. Available: <ext-link ext-link-type="uri" xlink:href="http://yann.lecun.com/exdb/mnist/">http://yann.lecun.com/exdb/mnist/</ext-link>
</comment>. </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Menze</surname>
<given-names>B. H.</given-names>
</name>
<name>
<surname>Jakab</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bauer</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kalpathy-Cramer</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Farahani</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Kirby</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS)</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>34</volume> (<issue>10</issue>), <fpage>1993</fpage>&#x2013;<lpage>2024</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2014.2377694</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Parikh</surname>
<given-names>K. S.</given-names>
</name>
<name>
<surname>Shah</surname>
<given-names>T. P.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Support Vector Machine - A Large Margin Classifier to Diagnose Skin Illnesses</article-title>. <source>Proced. Tech.</source> <volume>23</volume>, <fpage>369</fpage>&#x2013;<lpage>375</lpage>. <pub-id pub-id-type="doi">10.1016/j.protcy.2016.03.039</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rosenfeld</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Meshi</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Tarlow</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Globerson</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Learning Structured Models with the AUC Loss and its Generalizations</article-title>. <source>J.&#x20;Mach. Learn. Res.</source> <volume>33</volume>, <fpage>841</fpage>&#x2013;<lpage>849</lpage>. </citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sulam</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ben-Ari</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Kisilev</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2017</year>). <source>&#x201c;Maximizing AUC with Deep Learning for Classification of Imbalanced Mammogram Datasets,&#x201d; Eurographics Work</source>. <publisher-loc>Bremen, Germany</publisher-loc>: <publisher-name>Vis. Comput. Biol. Med.</publisher-name>. <comment>[Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://www.cs.bgu.ac.il/%7Erba/Papers/MaximizingAUC_MG.pdf">https://www.cs.bgu.ac.il/&#x223c;rba/Papers/MaximizingAUC_MG.pdf</ext-link>
</comment>. </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ying</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Lyu</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Stochastic Online AUC Maximization</article-title>. <source>Adv. Neural Inf. Process. Syst. No. Nips</source>, <fpage>451</fpage>&#x2013;<lpage>459</lpage>. </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yoo</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gujrathi</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Haider</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Khalvati</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Prostate Cancer Detection Using Deep Convolutional Neural Networks</article-title>. <source>Nat. Sci. Rep.</source> <pub-id pub-id-type="doi">10.1038/s41598-019-55972-4</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>Y.-C. I.</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Applying a Modified AUC to Gene Ranking</article-title>. <source>Csam</source> <volume>25</volume> (<issue>3</issue>), <fpage>307</fpage>&#x2013;<lpage>319</lpage>. <pub-id pub-id-type="doi">10.29220/CSAM.2018.25.3.307</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>AucPR: AucPR: An AUC-Based Approach Using Penalized Regression for Disease Prediction with High-Dimensional Omics Data</article-title>. <source>BMC Genomics</source> <volume>15</volume> (<issue>Suppl. 10</issue>), <fpage>1</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1186/1471-2164-15-S10-S1</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>M. R.</given-names>
</name>
</person-group> (<year>2018</year>).<article-title>Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels</article-title>. <source>Adv. Neural Inf. Process. Syst.</source>, <fpage>8778</fpage>&#x2013;<lpage>8788</lpage>. </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Hoi</surname>
<given-names>S. C. H.</given-names>
</name>
<name>
<surname>Jin</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Online AUC Maximization</article-title>. <source>Proc. 28th Int. Conf. Mach. Learn. ICML</source>, <fpage>233</fpage>&#x2013;<lpage>240</lpage>. </citation>
</ref>
</ref-list>
</back>
</article>