<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Big Data</journal-id>
<journal-title>Frontiers in Big Data</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Big Data</abbrev-journal-title>
<issn pub-type="epub">2624-909X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">590296</article-id>
<article-id pub-id-type="doi">10.3389/fdata.2020.590296</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Big Data</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Subgroup Invariant Perturbation for Unbiased Pre-Trained Model Prediction</article-title>
<alt-title alt-title-type="left-running-head">Majumdar et al.</alt-title>
<alt-title alt-title-type="right-running-head">SIP for Unbiased Pretrained Model Prediction</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Majumdar</surname>
<given-names>Puspita</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="http://loop.frontiersin.org/people/1033684/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Chhabra</surname>
<given-names>Saheb</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="http://loop.frontiersin.org/people/1048285/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Singh</surname>
<given-names>Richa</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="http://loop.frontiersin.org/people/1048576/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Vatsa</surname>
<given-names>Mayank</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="http://loop.frontiersin.org/people/1049931/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Department of Computer Science and Engineering, Indraprastha Institute of Information Technology, <addr-line>New Delhi</addr-line>, <country>India</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Department of Computer Science and Engineering, Indian Institute of Technology Jodhpur, <addr-line>Rajasthan</addr-line>, <country>India</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/845418/overview">Pin-Yu Chen</ext-link>, IBM Research, United States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1062543/overview">Chulin Xie</ext-link>, University of Illinois at Urbana-Champaign, United States Ren Wang, University of Michigan, United States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Richa Singh, <email>richa@iitj.ac.in</email>
</corresp>
<fn>
<p>This article was submitted to Machine Learning and Artificial Intelligence, a section of the journal Frontiers in Big Data</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>18</day>
<month>02</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2020</year>
</pub-date>
<volume>3</volume>
<elocation-id>590296</elocation-id>
<history>
<date date-type="received">
<day>31</day>
<month>07</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>20</day>
<month>11</month>
<year>2020</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2021 Majumdar, Chhabra, Singh and Vatsa.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Majumdar, Chhabra, Singh and Vatsa</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Modern deep learning systems have achieved unparalleled success and several applications have significantly benefited due to these technological advancements. However, these systems have also shown vulnerabilities with strong implications on the fairness and trustability of such systems. Among these vulnerabilities, bias has been an <italic>Achilles&#x2019; heel problem</italic>. Many applications such as face recognition and language translation have shown high levels of bias in the systems towards particular demographic sub-groups. Unbalanced representation of these sub-groups in the training data is one of the primary reasons of biased behavior. To address this important challenge, we propose a two-fold contribution: a bias estimation metric termed as <italic>Precise Subgroup Equivalence</italic> to jointly measure the bias in model prediction and the overall model performance. Secondly, we propose a novel bias mitigation algorithm which is inspired from adversarial perturbation and uses the PSE metric. The mitigation algorithm learns a single uniform perturbation termed as <italic>Subgroup Invariant Perturbation</italic> which is added to the input dataset to generate a transformed dataset. The transformed dataset, when given as input to the pre-trained model reduces the bias in model prediction. Multiple experiments performed on four publicly available face datasets showcase the effectiveness of the proposed algorithm for race and gender prediction.</p>
</abstract>
<kwd-group>
<kwd>Fairness</kwd>
<kwd>trustability</kwd>
<kwd>bias estimation</kwd>
<kwd>bias mitigation</kwd>
<kwd>subgroup invariant perturbation</kwd>
<kwd>gender classification</kwd>
<kwd>race classification</kwd>
</kwd-group>
<counts>
<page-count count="0"/>
</counts>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1.</label>
<title> Introduction</title>
<p>Increasing use of artificial intelligence (AI) and machine learning (ML) for automation coupled with instances of biased predictions has motivated and mandated researchers across the globe to pursue designing dependable AI systems. Out of the several attributes of dependability in AI systems such as interpretability, explainability, robustness, bias, and fairness (<xref ref-type="bibr" rid="B22">Mehrabi et al., 2019</xref>; <xref ref-type="bibr" rid="B10">Drozdowski et al., 2020</xref>; <xref ref-type="bibr" rid="B26">Ntoutsi et al., 2020</xref>), this research is focused towards bias and fairness.</p>
<p>Face analysis tasks such as face detection, face recognition, expression analysis, age and gender prediction are some of the AI applications in which several instances of biased or unfair predictions have been observed. For instance, <xref ref-type="bibr" rid="B4">Buolamwini and Gebru (2018)</xref> have shown that commercial gender classifiers perform better for lighter skin males while giving poor performance for darker skin females. Other instances include false identification of 28 members (specifically people of color) of the US Congress as criminals by Amazon&#x27;s facial recognition tool (<xref ref-type="bibr" rid="B27">Paolini-Subramanya, 2018</xref>). <xref ref-type="bibr" rid="B25">Nagpal et al. (2019)</xref> analyzed several pre-trained face recognition models to determine where and how the bias manifests in the deep neural networks. In light of these incidents, while some corporate and government organizations have decided to minimize or ban the development or usage of automated face analysis systems (<xref ref-type="bibr" rid="B7">Conger et al., 2019</xref>), several others are continuing the deployment and usage. Therefore, it is of paramount importance that we design mechanisms to improve the trustability and dependability of these systems. To address the challenges related to biased predictions of AI systems, researchers are broadly pursuing three directions: understanding bias, mitigating bias, and accounting for bias (<xref ref-type="bibr" rid="B26">Ntoutsi et al., 2020</xref>). Understanding bias involves realizing the source of bias along with estimating it (<xref ref-type="bibr" rid="B4">Buolamwini and Gebru, 2018</xref>; <xref ref-type="bibr" rid="B6">Celis and Rao, 2019</xref>; <xref ref-type="bibr" rid="B25">Nagpal et al., 2019</xref>; <xref ref-type="bibr" rid="B29">Radford and Joseph, 2020</xref>) whereas mitigation strategies involve designing algorithms that address bias (<xref ref-type="bibr" rid="B8">Creager et al., 2019</xref>; <xref ref-type="bibr" rid="B24">Nagpal et al., 2020</xref>).</p>
<p>In the literature, it has been demonstrated that if the training data used for learning the models is not balanced in terms of demographic subgroups, for instance, <italic>male</italic> and <italic>female</italic> are two different subgroups of <italic>gender</italic>, then there can be significant differences in the classification performance of pre-trained models observed on subgroups (<xref ref-type="bibr" rid="B2">Barocas and Selbst, 2016</xref>). Recent instances of biased predictions can be safely attributed to this observation as the training data required for deep learning models is often collected from the Internet using convenience sampling, which inherently leads to disparate proportions of data across subgroups. Models trained on historically biased datasets lead to biased results. Therefore, researchers have proposed several algorithms to mitigate the effect of bias on model prediction (<xref ref-type="bibr" rid="B1">Alvi et al., 2018</xref>; <xref ref-type="bibr" rid="B16">Gong et al., 2019</xref>). However, there is generally a trade-off between fairness and model performance (<xref ref-type="bibr" rid="B12">Du et al., 2019</xref>; <xref ref-type="bibr" rid="B20">Li and Vasconcelos, 2019</xref>). Removal of bias may affect the overall model performance while a high performing model may affect the performance of the under-represented subgroup. Therefore, it is important to 1) measure the trade-off between the effect of bias and the model performance through a unified metric and 2) mitigate the effect of bias without affecting the model performance. A solution to the problem is to re-train the models with large datasets having equal distribution of samples across different subgroups. However, in a real-world scenario, collecting such diverse datasets is not a trivial task. Also, re-training the models require updating millions of parameters and is computationally expensive.</p>
<p>This research focuses on <italic>estimating</italic> the trade-off between the effect of bias and the model performance and <italic>mitigating</italic> the influence of demographic subgroup bias on pre-trained model prediction to improve the model performance. Existing metrics such as Disparate Impact, Average False Rate, and Degree of Bias provide information of only bias or error rates, but they do not provide the complete information. The first contribution of this research is a unified bias metric, termed as Precise Subgroup Equivalence (PSE) which provides a joint estimate of bias in model prediction and the overall model performance. The second contribution is to <italic>mitigate</italic> the influence of demographic subgroup bias on pre-trained model prediction to improve the model performance. We propose a novel algorithm based on adversarial perturbation for bias mitigation. In general, adversarial perturbation utilizes the vulnerability of deep models towards small changes in the input to reduce the confidence of model prediction. In this research, we have used this concept to reduce the effect of bias on model prediction. To the best of our knowledge, this is the first time that adversarial perturbation is used for bias mitigation. The proposed algorithm utilizes the model prediction to learn a single uniform Subgroup Invariant Perturbation (SIP) for a given dataset. SIP is added to the input dataset to generate a transformed dataset, which, when given as an input to the model, produces unbiased outcomes and improves the overall model performance. <xref ref-type="fig" rid="F1">Figure 1</xref> shows a visual illustration of the proposed algorithm for bias mitigation using SIP. The proposed algorithm is used to mitigate the impact of demographic subgroup bias in race and gender model predictions.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Effect of demographic subgroup bias on pre-trained model prediction. <bold>(A)</bold> Pre-trained model prediction is biased towards subgroup <italic>R1</italic>. <bold>(B)</bold> Bias mitigation using SIP to achieve equal performance across <italic>R1</italic> and <italic>R2</italic> (best viewed in color).</p>
</caption>
<graphic xlink:href="fdata-03-590296-g001.tif"/>
</fig>
<p>The effectiveness of the algorithm is demonstrated under two scenarios: 1) <italic>independent demographic subgroup analysis</italic> and 2) <italic>inter-sectional demographic subgroup analysis</italic> on multiple datasets to showcase enhanced performance and reduced effect of bias on model prediction. The results show that PSE provides a unified score of both error and disparity in subgroups which is addressed using the proposed algorithm. Further, since the number of learned parameters is equal to the size of the input image, the proposed algorithm is observed to be computationally efficient as well.</p>
</sec>
<sec id="s2">
<label>2.</label>
<title> Related Work</title>
<p>Recent years have observed significant increase in the research on different aspects of bias and fairness in AI systems. Existing literature can be grouped into three broad categories: 1) Understanding and Estimating Bias, 2) Bias Mitigation Algorithms, and 3) Fairness Metrics.</p>
<p>Understanding and Estimating Bias: Researchers have focused on understanding the presence of bias in the prediction of commercial-off-the-shelf systems (COTS) and pre-trained deep models. <xref ref-type="bibr" rid="B4">Buolamwini and Gebru (2018)</xref> evaluated commercial gender classifiers from Microsoft, IBM, and Face &#x2b;&#x2b; on four categories based on the skin type, namely, darker males, darker females, lighter males, and lighter females. It was found that the classifiers performed best for males with lighter skin tone and least for females with darker skin tone. <xref ref-type="bibr" rid="B25">Nagpal et al. (2019)</xref> provided an analysis of bias in deep face recognition models. They have shown that deep models encode race and age-specific features that lead to biased discrimination. According to various studies, the training data distribution has a huge impact on the model&#x27;s performance (<xref ref-type="bibr" rid="B33">Torralba and Efros, 2011</xref>; <xref ref-type="bibr" rid="B3">Bolukbasi et al., 2016</xref>). Models trained on imbalanced datasets lead to biased outputs. Therefore, different data re-sampling techniques have been proposed by the researchers to balance the training data distribution. This is done either by over-sampling the minority class (<xref ref-type="bibr" rid="B23">Mullick et al., 2019</xref>) or under-sampling the majority class (<xref ref-type="bibr" rid="B11">Drummond et al., 2003</xref>). However, a recent study has shown that even models trained with balanced datasets amplify bias (<xref ref-type="bibr" rid="B35">Wang et al., 2019</xref>). It is shown that the learned models amplify the association between labels and gender, which in turn leads to biased discrimination.</p>
<p>Bias Mitigation: Mitigation algorithms can either be applied as a pre-processing step or in-processing, or post-processing. Different algorithms have been proposed to mitigate the effect of bias. <xref ref-type="bibr" rid="B31">Ryu et al. (2017)</xref> addressed the problem of the performance gap in different subgroups of race and gender attributes. They hypothesized that faces look different across different genders and races, and proposed InclusiveNet which learns the demographic information prior to attribute detection. <xref ref-type="bibr" rid="B14">Dwork et al. (2018)</xref> proposed decoupled classifiers to increase fairness and accuracy in classification systems. The decoupled classifiers learn a separate classifier for sensitive attributes and can be used with any black-box network. <xref ref-type="bibr" rid="B9">Das et al. (2018)</xref> proposed a Multi-Task Convolution Neural Network (MTCNN) to classify gender, age, and ethnicity attributes and minimized the effect of bias by utilizing disjoint features of fully connected layers of a deep Convolution Neural Network (CNN). <xref ref-type="bibr" rid="B1">Alvi et al. (2018)</xref> proposed a joint learning and unlearning framework for mitigating bias in CNN models for gender, age, race, and pose classification. A disentangled representation learning technique is presented to obtain flexibly fair features by <xref ref-type="bibr" rid="B8">Creager et al. (2019)</xref>. <xref ref-type="bibr" rid="B19">Kim et al. (2019)</xref> proposed a regularization algorithm to unlearn the bias information. Recently, <xref ref-type="bibr" rid="B24">Nagpal et al. (2020)</xref> proposed a filter drop technique for learning unbiased representations. Results are demonstrated for gender prediction across different ethnicity groups.</p>
<p>Apart from bias mitigation in attribute prediction, researchers have also focused on mitigating bias in face recognition. <xref ref-type="bibr" rid="B16">Gong et al. (2019)</xref> addressed the problem of bias in face recognition systems and proposed a debiasing adversarial network. The proposed network learns unbiased representation for both identity and demographic attributes. <xref ref-type="bibr" rid="B17">Huang et al. (2019)</xref> investigated the problem of deep imbalanced learning in the context of deep representation learning for attribute prediction and face recognition. They proposed Cluster-based Large Margin Local Embedding (CLMLE) method, which maintains inter-cluster margin among the same and different classes. <xref ref-type="bibr" rid="B34">Wang and Deng (2019)</xref> proposed a reinforcement learning-based race balance network (RL-RBN) to mitigate racial bias. <xref ref-type="bibr" rid="B32">Singh et al. (2020)</xref> provided a review of techniques related to bias in face recognition.</p>
<p>Fairness Metrics: To measure the fairness of deep models, different metrics have been proposed in the literature.</p>
<p>Statistical Parity (SP) (<xref ref-type="bibr" rid="B5">Calders and Verwer, 2010</xref>): It is one of the widely used fairness metrics. It suggests that a model gives unbiased output if the prediction is independent of the demographic group such as <italic>race</italic>, <italic>gender</italic>, and <italic>religion</italic>. Deviation from statistical parity is measured as the ratio of the probability of a positive classification for both subgroups of a demographic group. It is termed as Disparate Impact (DI) (<xref ref-type="bibr" rid="B15">Feldman et al., 2015</xref>) and computed as:<disp-formula id="e1">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>I</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>Y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>Y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where, <italic>D</italic> represents the demographic group, and <inline-formula id="inf1">
<mml:math>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>Y</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> represents the predicted decision or class. A lower value of <inline-formula id="inf2">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>I</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> indicates higher bias in the model prediction.</p>
<p>Degree of Bias (DoB) (<xref ref-type="bibr" rid="B16">Gong et al., 2019</xref>): It is defined as the standard deviation of <italic>Classification Accuracy</italic> (<italic>CAcc</italic>) across different subgroups of a demographic group. Mathematically, it is represented as:<disp-formula id="e2">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>B</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>c</mml:mi>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where, <inline-formula id="inf3">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents a subgroup of a demographic group <italic>D</italic>. High performance gap across different subgroups will result in high <inline-formula id="inf4">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, which in turn implies bias in the model prediction.</p>
</sec>
<sec sec-type="materials|methods" id="s3">
<label>3.</label>
<title> Materials and Methods</title>
<p>The following subsections discuss the proposed metric, estimation of bias in model prediction, and bias mitigation using Subgroup Invariant Perturbation (SIP). There are two different scenarios for bias estimation and mitigation: 1) <italic>independent demographic subgroup analysis</italic> and 2) <italic>intersectional demographic subgroup analysis</italic>. In the first scenario, bias estimation/mitigation is performed across the subgroups of a demographic group. For example, bias estimation/mitigation is performed across the subgroups of <italic>gender</italic>. In the second scenario, bias estimation/mitigation is performed across the intersection of different demographic groups. For example, bias estimation/mitigation is performed across the intersectional subgroups of <italic>race</italic> and <italic>gender</italic>.</p>
<sec id="s3-1">
<label>3.1.</label>
<title> Proposed Metric: Precise Subgroup Equivalence</title>
<p>Existing fairness metrics evaluate the performance gap across different subgroups (<xref ref-type="bibr" rid="B13">Du et al., 2020</xref>). However, these do not reflect the overall model performance. For instance, if a model gives almost equal but low performance across different subgroups, then <inline-formula id="inf5">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>I</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> will be high, and <inline-formula id="inf6">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> will be low. Therefore, the model prediction will be considered unbiased across different subgroups. However, an unbiased but low performing model is undesirable. Therefore, in this research, Precise Subgroup Equivalence (PSE) metric is introduced that jointly estimates the effect of demographic subgroup bias on model prediction and the overall model performance. Precise Subgroup Equivalence (PSE) is the average of Disparate Impact (DI), Average False Rate (AFR), and Degree of Bias (DoB).<disp-formula id="e3">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>I</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>F</mml:mi>
<mml:mi>R</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mn>3</mml:mn>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>Since a lower value of DI indicates higher bias in model prediction, therefore higher value of <inline-formula id="inf7">
<mml:math>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>I</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> indicates higher bias in model prediction. Here, <inline-formula id="inf8">
<mml:math>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>F</mml:mi>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the mean of <italic>False Positive Rate</italic> (<italic>FPR</italic>) and <italic>False Negative Rate</italic> (<italic>FNR</italic>). It is robust to the subgroup imbalance problem and reflects the overall model performance. On the other hand, <inline-formula id="inf9">
<mml:math>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>I</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf10">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> reflects the bias in the model prediction. Therefore, <inline-formula id="inf11">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> provides a joint estimate of the overall model performance and the impact of bias. A model with low <inline-formula id="inf12">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> indicates an unbiased high performing model.</p>
</sec>
<sec id="s3-2">
<label>3.2.</label>
<title> Bias Estimation</title>
<p>For joint estimation of pre-trained model performance and the impact of demographic subgroup bias, <inline-formula id="inf13">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the model prediction corresponding to a given dataset is computed. Let <inline-formula id="inf14">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula> be the training set with <italic>n</italic> number of images.<disp-formula id="e4">
<mml:math>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where, each image <inline-formula id="inf15">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is associated with <italic>m</italic> demographic groups. Let <inline-formula id="inf16">
<mml:math>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:math>
</inline-formula> and <inline-formula id="inf17">
<mml:math>
<mml:mi mathvariant="bold">E</mml:mi>
</mml:math>
</inline-formula> are the two demographic groups and <italic>s</italic> and <italic>t</italic> be the number of subgroups in <inline-formula id="inf18">
<mml:math>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:math>
</inline-formula> and <inline-formula id="inf19">
<mml:math>
<mml:mi mathvariant="bold">E</mml:mi>
</mml:math>
</inline-formula>, respectively.<disp-formula id="e5">
<mml:math>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mtext>and</mml:mtext>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi mathvariant="bold">E</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">E</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">E</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">E</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>where, <inline-formula id="inf20">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf21">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represent a subgroup of the respective demographic group. Let <inline-formula id="inf22">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> be a pre-trained model with weight <inline-formula id="inf23">
<mml:math>
<mml:mi mathvariant="bold">W</mml:mi>
</mml:math>
</inline-formula> and bias <italic>b</italic> trained for predicting demographic group <inline-formula id="inf24">
<mml:math>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:math>
</inline-formula>.</p>
<p>For the first scenario, the probability of predicting an input image <inline-formula id="inf25">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to subgroup <inline-formula id="inf26">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is represented as:<disp-formula id="e6">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<p>For the second scenario, the probability of predicting an input image <inline-formula id="inf27">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to subgroup <inline-formula id="inf28">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> across demographic group <inline-formula id="inf29">
<mml:math>
<mml:mi mathvariant="bold">E</mml:mi>
</mml:math>
</inline-formula> is represented as:<disp-formula id="e7">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">E</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>
</p>
<p>The <inline-formula id="inf30">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of model <inline-formula id="inf31">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> corresponding to dataset <inline-formula id="inf32">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula> is computed as:<disp-formula id="e8">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>D</mml:mi>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>F</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mn>3</mml:mn>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>where, <inline-formula id="inf33">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf34">
<mml:math>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>F</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf35">
<mml:math>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>o</mml:mi>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the Disparate Impact, Average False Rate, and Degree of Bias of model <inline-formula id="inf36">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> corresponding to dataset <inline-formula id="inf37">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula>, respectively.</p>
</sec>
<sec id="s3-3">
<label>3.3.</label>
<title> Bias Mitigation</title>
<p>After estimating the bias in the prediction of a pre-trained model <inline-formula id="inf38">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> corresponding to dataset <inline-formula id="inf39">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula>, the next task is to mitigate the effect of bias to improve the overall model performance. For this purpose, a single uniform Subgroup Invariant Perturbation (SIP) is learned by minimizing the <inline-formula id="inf40">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> corresponding to the first scenario for the given dataset <inline-formula id="inf41">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula>. The aim is to generate a transformed dataset <inline-formula id="inf42">
<mml:math>
<mml:mi mathvariant="bold">T</mml:mi>
</mml:math>
</inline-formula> by adding SIP to all the images of dataset <inline-formula id="inf43">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula>, such that when <inline-formula id="inf44">
<mml:math>
<mml:mi mathvariant="bold">T</mml:mi>
</mml:math>
</inline-formula> is given as input to the pre-trained model <inline-formula id="inf45">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> produces unbiased outcomes and improves the overall performance. We hypothesize that the learned SIP is effective for mitigating the bias corresponding to the second scenario as well. In order to validate this, multiple experiments are performed, and the results are discussed in <xref ref-type="sec" rid="s5-2">Section 5.2</xref>. The optimization process for learning SIP <inline-formula id="inf46">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula> is discussed below.</p>
<p>Let <inline-formula id="inf47">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula> be the Subgroup Invariant Perturbation (SIP), initialized with zeros. Each image <inline-formula id="inf48">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of the dataset has pixel values in the range of <inline-formula id="inf49">
<mml:math>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Let <inline-formula id="inf50">
<mml:math>
<mml:mi mathvariant="bold">T</mml:mi>
</mml:math>
</inline-formula> be the transformed dataset obtained by adding <inline-formula id="inf51">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula> to the dataset <inline-formula id="inf52">
<mml:math>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula>. To bring the pixel values of each image in the transformed dataset in the range of <inline-formula id="inf53">
<mml:math>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf54">
<mml:math>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> function is applied as follows:<disp-formula id="e9">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mn>2</mml:mn>
</mml:mfrac>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where, <inline-formula id="inf55">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the transformed image corresponding to the input image <inline-formula id="inf56">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The probability of predicting a transformed image <inline-formula id="inf57">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to subgroup <inline-formula id="inf58">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is given by:<disp-formula id="e10">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>
</p>
<p>For models that yield biased predictions, there is a performance gap across different subgroups, where the performance of some subgroups are better than others. Therefore, the objective is to reduce <inline-formula id="inf59">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> by 1) enhancing the performance of the low performing subgroups and 2) maintaining/enhancing the performance of high performing subgroups. In order to achieve both the objectives, the following objective function is used.<disp-formula id="e11">
<mml:math>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>Y</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where, <inline-formula id="inf60">
<mml:math>
<mml:mrow>
<mml:msub>
<mml:mi>Y</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the true label and <inline-formula id="inf61">
<mml:math>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mo>.</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>.</mml:mo>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the function to minimize the distance between the true label and the probability of predicting the true class. The above objective function is optimized corresponding to SIP <inline-formula id="inf62">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula>. For this purpose, the following function is minimized:<disp-formula id="e12">
<mml:math>
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mtext>min</mml:mtext>
</mml:mrow>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:munder>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>Y</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>
<disp-formula id="equ1">
<mml:math>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>Y</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>q</mml:mi>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:munderover>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x7c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>where, <inline-formula id="inf63">
<mml:math>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <italic>q</italic> is the number of images belonging to subgroup <italic>j</italic> with <inline-formula id="inf64">
<mml:math>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. <inline-formula id="inf65">
<mml:math>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mo>.</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>.</mml:mo>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> will increase the probability of predicting the true class, which in turn reduces the <inline-formula id="inf66">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. Low <inline-formula id="inf67">
<mml:math>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3d5;</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> will simultaneously ensure reduced effect of bias on model prediction along with improved model performance. <xref ref-type="fig" rid="F2">Figure 2</xref> shows the block diagram of the steps involved in learning the SIP <inline-formula id="inf68">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula>.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Block diagram of the steps involved in learning Subgroup Invariant Perturbation (SIP). In the first step, SIP <inline-formula id="inf69">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula> is initialized with zeros and added to the images of the training set to generated the transformed set. In the next step, the transformed set is given as input to the pre-trained model and model prediction is obtained. Next, loss is computed and optimization is performed over <inline-formula id="inf70">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula> to minimize PSE. The updated <inline-formula id="inf71">
<mml:math>
<mml:mi mathvariant="bold">N</mml:mi>
</mml:math>
</inline-formula> is added to the training set and the process is repeated until convergence (best viewed in color).</p>
</caption>
<graphic xlink:href="fdata-03-590296-g002.tif"/>
</fig>
</sec>
</sec>
<sec id="s4">
<label>4.</label>
<title> Experimental Setup</title>
<p>The performance of the proposed algorithm is evaluated for race and gender classification on four different datasets. The results are reported using the proposed metric PSE, two existing bias evaluation metrics, and one existing performance evaluation metric. The details of the datasets with the corresponding protocols and the pre-trained models used for the experiments are discussed below.</p>
<sec id="s4-1">
<label>4.1.</label>
<title> Databases and Protocols</title>
<p>Experiments are performed for race and gender prediction, using data corresponding to race <italic>R1 (light skin color)</italic> and <italic>R2 (dark skin color)</italic>, and gender <italic>G1 (Male)</italic> and <italic>G2 (Female)</italic>. The distribution of the number of images in each dataset across different race and gender subgroups is shown in <xref ref-type="table" rid="T1">Table 1</xref>; <xref ref-type="fig" rid="F3">Figure 3</xref> shows sample images from each dataset.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Distribution of number of images in the MORPH, UTKFace, LFWA, and CelebA datasets across different race and gender subgroups.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="2">Dataset</th>
<th colspan="2" align="center">Race</th>
<th colspan="2" align="center">Gender</th>
</tr>
<tr>
<th align="center">
<italic>R</italic>1</th>
<th align="center">
<italic>R</italic>2</th>
<th align="center">
<italic>G</italic>1</th>
<th align="center">
<italic>G</italic>2</th>
</tr>
</thead>
<tbody>
<tr>
<td>MORPH</td>
<td align="center">10,662</td>
<td align="center">42,725</td>
<td align="center">46,835</td>
<td align="center">8,527</td>
</tr>
<tr>
<td>UTKFace</td>
<td align="center">10,076</td>
<td align="center">4,525</td>
<td align="center">12,389</td>
<td align="center">11,312</td>
</tr>
<tr>
<td>LFWA</td>
<td align="center">9,830</td>
<td align="center">560</td>
<td align="center">10,181</td>
<td align="center">2,962</td>
</tr>
<tr>
<td>CelebA</td>
<td align="center">&#x2014;</td>
<td align="center">&#x2014;</td>
<td align="center">75,976</td>
<td align="center">1,06,756</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Sample images of the <bold>(A)</bold> MORPH, <bold>(B)</bold> UTKFace, <bold>(C)</bold> LFWA, and <bold>(D)</bold> CelebA datasets.</p>
</caption>
<graphic xlink:href="fdata-03-590296-g003.tif"/>
</fig>
<p>MORPH dataset (Album-2) (<xref ref-type="bibr" rid="B30">Rawls and Ricanek, 2009</xref>) contains more than <inline-formula id="inf72">
<mml:math>
<mml:mrow>
<mml:mn>54,000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> images of <inline-formula id="inf73">
<mml:math>
<mml:mrow>
<mml:mn>13,180</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> subjects. The dataset is partitioned into 60% training set, 20% validation set, and 20% testing set. The partitioning is done with non-overlapping subjects in each set.</p>
<p>UTKFace dataset (<xref ref-type="bibr" rid="B36">Zhang et al., 2017</xref>) contains more than <inline-formula id="inf74">
<mml:math>
<mml:mrow>
<mml:mn>20,000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> face images and divided into three parts, having <inline-formula id="inf75">
<mml:math>
<mml:mrow>
<mml:mn>9,779</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf76">
<mml:math>
<mml:mrow>
<mml:mn>10,718</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf77">
<mml:math>
<mml:mrow>
<mml:mn>3,206</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> images in Part I, Part II, and Part III, respectively. Part I is used for training, Part II for testing, and Part III for validation (<xref ref-type="bibr" rid="B9">Das et al., 2018</xref>).</p>
<p>LFWA dataset (<xref ref-type="bibr" rid="B18">Huang et al., 2008</xref>) contains <inline-formula id="inf78">
<mml:math>
<mml:mrow>
<mml:mn>13,233</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> images of <inline-formula id="inf79">
<mml:math>
<mml:mrow>
<mml:mn>5,749</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> subjects with 73 attributes. Attributes corresponding to each image is annotated with intensity values. These are binarized by converting positive intensity values with label 1 and negative intensity values with label 0. For experiments, attributes corresponding to race <italic>R1</italic>, <italic>R2</italic>, and gender <italic>G1</italic> are taken. Images with label 0 for <italic>G1</italic> are considered as <italic>G2</italic>. Experiments are performed using the standard pre-defined protocol proposed by (<xref ref-type="bibr" rid="B18">Huang et al., 2008</xref>).</p>
<p>CelebA dataset (<xref ref-type="bibr" rid="B21">Liu et al., 2015</xref>) consists of a total of <inline-formula id="inf80">
<mml:math>
<mml:mrow>
<mml:mn>2,02,599</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> face images of more than <inline-formula id="inf81">
<mml:math>
<mml:mrow>
<mml:mn>10,000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> celebrities with 40 annotated binary attributes. For experiments, the <italic>G1</italic> attribute is taken and images with label 0 for <italic>G1</italic> are considered as <italic>G2</italic>. The experiments are performed using the standard pre-defined protocol defined by (<xref ref-type="bibr" rid="B21">Liu et al., 2015</xref>).</p>
<p>Pre-trained models: Experiments are performed using pre-trained VGGFace (<xref ref-type="bibr" rid="B28">Parkhi et al., 2015</xref>) model, which is trained on the VGGFace dataset (<xref ref-type="bibr" rid="B28">Parkhi et al., 2015</xref>) for face recognition. VGGFace dataset is a large scale dataset of 2.6M facial images corresponding to 2.6K subjects. VGGFace model has shown high generalization abilities for face recognition. Therefore, we have used this model and fine-tuned it for race and gender prediction. In this research, three race prediction models and four gender prediction models are used for the experiments. The race prediction models are obtained by separately fine-tuning the pre-trained VGGFace model on the MORPH, UTKFace, and LFWA datasets. Similarly, the gender prediction models are obtained by fine-tuning on the MORPH, UTKFace, LFWA, and CelebA datasets. These models are treated as pre-trained race and gender prediction models in all the experiments.</p>
</sec>
<sec id="s4-2">
<label>4.2.</label>
<title> Implementation Details</title>
<p>The implementation details of the network training and perturbation learning for mitigation are given below.</p>
<p>Network training: Each model is trained by adding two fully connected dense layers of 512 dimensions after the final convolutional layer of the VGGFace model. Models are trained for 20 epochs with Adam optimizer. The learning rate is set to 0.0001 for the first 10 epochs and reduced by 0.1 after every 5 epochs. Categorical cross-entropy loss is used to train the models.</p>
<p>Perturbation learning for mitigation: Perturbation is learned from the training set of a given dataset. In order to learn Subgroup Invariant Perturbation (SIP), a matrix is initialized with zeros of size <inline-formula id="inf82">
<mml:math>
<mml:mrow>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> (equal to the dimension of the input image), which results in 12,288 number of parameters. The parameters of this matrix are only trainable during SIP learning while keeping the parameters of the model frozen. In the first step, SIP is added to the images in the training set using <xref ref-type="disp-formula" rid="e9">Equation 9</xref> and given as input to the model to obtain the predictions. In the second step, model predictions are used to compute the loss using <xref ref-type="disp-formula" rid="e12">Equation 12</xref>. In the final step, the gradient of the loss is computed with respect to the given input, and this gradient is backpropagated to the input to update the parameters of the SIP matrix only. The process is repeated until convergence. For perturbation learning, Adam optimizer is used with a learning rate of 0.001. Depending upon the training set, the batch size is set between 500 and 1,000. Each batch is processed for 16 iterations.</p>
</sec>
</sec>
<sec id="s5">
<label>5.</label>
<title> Results and Analysis</title>
<p>Models trained on datasets with over-representation of some demographic subgroups and under-representation of others often result in biased outputs. In a real-world scenario, it is difficult to have knowledge of the dataset used to train a model. However, depending on the training data distribution, the model could lead to biased prediction outputs. Therefore, it is important to first estimate the bias in model prediction, followed by mitigation. As discussed previously, the model&#x27;s overall performance should also be considered during the estimation/mitigation of bias in model prediction to balance the trade-off between fairness and model performance. Therefore, in this research, we have jointly estimated bias in model prediction and the overall model performance using the proposed metric PSE. A series of experiments are performed where models pre-trained on some datasets are evaluated on others for bias estimation using the existing and proposed metrics. Next, we use the PSE of the model to mitigate the effect of bias in model prediction using the proposed algorithm.</p>
<p>We have segregated this section into: 1) Bias Estimation and 2) Bias Mitigation in <xref ref-type="sec" rid="s5-1">Sections 5.1</xref> and <xref ref-type="sec" rid="s5-2">Sections 5.2</xref>, respectively. Analysis of the experiments are performed under both the scenarios, <italic>Independent demographic subgroup analysis</italic> and <italic>Intersectional demographic subgroup analysis</italic>. In the first scenario of independent demographic subgroup analysis, bias estimation/mitigation algorithms are analyzed across the subgroups of a demographic group individually. Whereas, in the second scenario, analysis is performed across the intersection of different demographic groups. <xref ref-type="table" rid="T2">Table 2</xref> shows the details of the experiments performed in this research.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Details of the experiments to estimate and mitigate the effect of demographic subgroup bias on pre-trained race and gender prediction models.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th>Task</th>
<th align="center">Scenario</th>
<th align="center">Model trained on</th>
<th align="center">Bias estimation/mitigation</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="3">Race prediction</td>
<td rowspan="3">Independent/intersectional demographic subgroup analysis</td>
<td>MORPH</td>
<td>UTKFace, LFWA</td>
</tr>
<tr>
<td>UTKFace</td>
<td>MORPH, LFWA</td>
</tr>
<tr>
<td>LFWA</td>
<td>MORPH, UTKFace</td>
</tr>
<tr>
<td rowspan="7">Gender prediction</td>
<td rowspan="4">Independent demographic subgroup analysis</td>
<td>MORPH</td>
<td>UTKFace, LFWA, CelebA</td>
</tr>
<tr>
<td>UTKFace</td>
<td>MORPH, LFWA, CelebA</td>
</tr>
<tr>
<td>LFWA</td>
<td>MORPH, UTKFace, CelebA</td>
</tr>
<tr>
<td>CelebA</td>
<td>MORPH, UTKFace, LFWA</td>
</tr>
<tr>
<td rowspan="3">Intersectional demographic subgroup analysis</td>
<td>MORPH</td>
<td>UTKFace, LFWA</td>
</tr>
<tr>
<td>UTKFace</td>
<td>MORPH, LFWA</td>
</tr>
<tr>
<td>LFWA</td>
<td>MORPH, UTKFace</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="s5-1">
<label>5.1.</label>
<title> Bias Estimation</title>
<p>Bias estimation plays a key role in designing solutions for bias mitigation. Therefore, it is important to have a good metric to estimate bias in model prediction along with the overall model performance. There are various fairness and performance evaluation metrics, such as DI, DoB, and AFR. DI measures the deviation from statistical parity, and DoB represents the standard deviation of classification accuracy across different subgroups. On the other hand, AFR gives the average of the false positive rate and false negative rate. These metrics either evaluate the performance gap across different subgroups or the overall model performance. Therefore, we have introduced a new metric PSE that evaluates both fairness and model performance. To validate this fact, we have evaluated the performance of multiple pre-trained models (trained on different datasets) using existing and proposed metrics. The experimental setup of this experiment is discussed below:</p>
<p>Experimental Setup: In this experiment, the performance of pre-trained models is evaluated using five different evaluation metrics: subgroup-specific error rate, (1-DI), DoB, AFR, and PSE for bias estimation. Evaluation of each pre-trained model is done on the training set of all the datasets except the one on which the model is trained. For instance, if the model is pre-trained on the MORPH dataset, then it is evaluated on the LFWA, CelebA, and UTKFace datasets. This setup is considered by keeping in mind the real-world scenario where the training set of the pre-trained model is unknown. Bias estimation is done on the training set because the PSE learned from the training set is used to mitigate the bias in model prediction for the corresponding dataset.</p>
<sec id="s5-1-1">
<title>5.1.1. Independent Demographic Subgroup Analysis</title>
<p>In this scenario, the models are evaluated across different <italic>race</italic> and <italic>gender</italic> subgroups, respectively, of a given dataset. The error rate of each subgroup is computed to understand the variations in performance across subgroups. <xref ref-type="table" rid="T3">Table 3</xref> shows the performance of pre-trained race prediction models. It is observed that the error rate of the models varies significantly across different <italic>race</italic> subgroups. It is also observed that the distribution of training data plays a significant role in the performance of pre-trained models. For instance, the model trained on the MORPH dataset when evaluated on the UTKFace dataset results in 27.72% and 22.47% error rate corresponding to subgroup <italic>R</italic>1 and <italic>R</italic>2, respectively. On the other hand, when the LFWA model is evaluated on the UTKFace dataset, it gives 0.04% and 97.54% error rate corresponding to subgroup <italic>R</italic>1 and <italic>R</italic>2, respectively. The significant difference in the error rate of each subgroup obtained by different pre-trained models is due to the skewed training data distribution on which these models are trained as shown in <xref ref-type="table" rid="T1">Table 1</xref>. The MORPH dataset has under-representation of subgroup <italic>R</italic>1 and over-representation of subgroup <italic>R</italic>2. On the other hand, the LFWA dataset has a majority of subgroup <italic>R</italic>1. Therefore, the model trained on the MORPH dataset performs better for subgroup <italic>R</italic>2, while the LFWA model gives better performance for subgroup <italic>R</italic>1. A similar observation can be drawn when the evaluation is performed on the LFWA dataset using the models trained on the MORPH and UTKFace datasets.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Performance of pre-trained <italic>race</italic> prediction models (%) across different <italic>race</italic> subgroups for independent demographic subgroup analysis scenario.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="2">Bias estimated on</th>
<th rowspan="2" align="center">Model trained on</th>
<th colspan="2" align="center">Error</th>
<th rowspan="2" align="center">1 &#x2013; DI</th>
<th rowspan="2" align="center">AFR</th>
<th rowspan="2" align="center">DoB</th>
<th rowspan="2" align="center">PSE</th>
</tr>
<tr>
<th align="center">R1</th>
<th align="center">R2</th>
</tr>
</thead>
<tbody>
<tr>
<td>UTKFace</td>
<td align="center">MORPH</td>
<td align="center">27.72</td>
<td align="center">22.47</td>
<td align="center">6.77</td>
<td align="center">25.09</td>
<td align="center">2.62</td>
<td align="center">11.49</td>
</tr>
<tr>
<td/>
<td align="center">LFWA</td>
<td align="center">0.04</td>
<td align="center">97.54</td>
<td align="center">97.53</td>
<td align="center">48.78</td>
<td align="center">48.75</td>
<td align="center">65.02</td>
</tr>
<tr>
<td>MORPH</td>
<td align="center">UTKFace</td>
<td align="center">0.52</td>
<td align="center">80.02</td>
<td align="center">79.92</td>
<td align="center">40.27</td>
<td align="center">39.75</td>
<td align="center">53.31</td>
</tr>
<tr>
<td/>
<td align="center">LFWA</td>
<td align="center">0.00</td>
<td align="center">96.86</td>
<td align="center">96.86</td>
<td align="center">48.43</td>
<td align="center">48.43</td>
<td align="center">64.57</td>
</tr>
<tr>
<td>LFWA</td>
<td align="center">MORPH</td>
<td align="center">83.32</td>
<td align="center">7.64</td>
<td align="center">81.94</td>
<td align="center">45.48</td>
<td align="center">37.84</td>
<td align="center">55.08</td>
</tr>
<tr>
<td/>
<td align="center">UTKFace</td>
<td align="center">17.82</td>
<td align="center">60.37</td>
<td align="center">51.78</td>
<td align="center">39.09</td>
<td align="center">21.27</td>
<td align="center">37.38</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>On evaluating the performance of a pre-trained model using individual metrics for a given dataset, it is observed that PSE is a good indicator of fairness and model performance. For instance, the PSE of the LFWA model corresponding to the UTKFace dataset is 65.02%. The high value of PSE indicates a biased and low performing model. The values of metrics (1-DI) and DoB are 97.53% and 48.75%, indicating a biased model. However, these do not provide any insights about model performance. On the other hand, the AFR of the model is 48.78%, indicating that the model performance is low without providing any insight about the bias in model prediction. This shows that metric PSE provides a joint estimation of bias and model performance.</p>
<p>The performance of the <italic>gender</italic> prediction models is reported in <xref ref-type="table" rid="T4">Table 4</xref>. A similar observation is drawn regarding the effectiveness of metric PSE from <xref ref-type="table" rid="T4">Table 4</xref>. For instance, the performance of the model trained on the UTKFace dataset, when evaluated on the MORPH dataset, shows almost equal but high error rate across different <italic>gender</italic> subgroups. Therefore (1-DI) and DoB of this model are low, but AFR is high. Thus, none of the metrics is able to provide a unified estimate of fairness and model performance. On the other hand, the PSE of this model is 13.86% showing a joint estimate of both. A similar observation is obtained when this pre-trained model is evaluated on the LFWA and CelebA datasets. This showcases that PSE provides a unified score of both error and disparity in subgroups.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Performance of pre-trained <italic>gender</italic> prediction models (%) across different <italic>gender</italic> subgroups for independent demographic subgroup analysis scenario.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="2">Bias estimated on</th>
<th rowspan="2" align="center">Model trained on</th>
<th colspan="2" align="center">Error</th>
<th rowspan="2" align="center">1 &#x2013; DI</th>
<th rowspan="2" align="center">AFR</th>
<th rowspan="2" align="center">DoB</th>
<th rowspan="2" align="center">PSE</th>
</tr>
<tr>
<th align="center">G1</th>
<th align="center">G2</th>
</tr>
</thead>
<tbody>
<tr>
<td>UTKFace</td>
<td>MORPH</td>
<td>29.42</td>
<td>42.75</td>
<td>18.90</td>
<td>36.08</td>
<td>6.66</td>
<td>20.54</td>
</tr>
<tr>
<td/>
<td>LFWA</td>
<td>24.94</td>
<td>49.12</td>
<td>32.22</td>
<td>37.02</td>
<td>12.09</td>
<td>27.11</td>
</tr>
<tr>
<td/>
<td>CelebA</td>
<td>53.05</td>
<td>41.70</td>
<td>19.47</td>
<td>47.37</td>
<td>5.67</td>
<td>24.17</td>
</tr>
<tr>
<td>MORPH</td>
<td>UTKFace</td>
<td>36.75</td>
<td>33.53</td>
<td>4.85</td>
<td>35.13</td>
<td>1.61</td>
<td>13.86</td>
</tr>
<tr>
<td/>
<td>LFWA</td>
<td>39.05</td>
<td>24.96</td>
<td>18.77</td>
<td>32.00</td>
<td>7.04</td>
<td>19.27</td>
</tr>
<tr>
<td/>
<td>CelebA</td>
<td>69.82</td>
<td>29.78</td>
<td>57.01</td>
<td>49.79</td>
<td>20.02</td>
<td>42.27</td>
</tr>
<tr>
<td>LFWA</td>
<td>UTKFace</td>
<td>30.27</td>
<td>36.69</td>
<td>9.21</td>
<td>33.48</td>
<td>3.21</td>
<td>15.30</td>
</tr>
<tr>
<td/>
<td>MORPH</td>
<td>19.27</td>
<td>57.66</td>
<td>47.56</td>
<td>38.46</td>
<td>19.19</td>
<td>35.07</td>
</tr>
<tr>
<td/>
<td>CelebA</td>
<td>16.79</td>
<td>45.74</td>
<td>34.80</td>
<td>31.26</td>
<td>14.47</td>
<td>26.84</td>
</tr>
<tr>
<td>CelebA</td>
<td>UTKFace</td>
<td>43.23</td>
<td>42.88</td>
<td>0.62</td>
<td>43.05</td>
<td>0.17</td>
<td>14.61</td>
</tr>
<tr>
<td/>
<td>MORPH</td>
<td>39.71</td>
<td>57.79</td>
<td>30.00</td>
<td>48.75</td>
<td>9.04</td>
<td>29.26</td>
</tr>
<tr>
<td/>
<td>LFWA</td>
<td>12.21</td>
<td>54.75</td>
<td>48.45</td>
<td>33.47</td>
<td>21.27</td>
<td>34.40</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s5-1-2">
<title>5.1.2. Intersectional Demographic Subgroup Analysis</title>
<p>Existing studies (<xref ref-type="bibr" rid="B1">Alvi et al., 2018</xref>; <xref ref-type="bibr" rid="B9">Das et al., 2018</xref>; <xref ref-type="bibr" rid="B24">Nagpal et al., 2020</xref>) have shown that the influence of one demographic group can affect the prediction of others. For instance, the performance of a gender prediction model may be affected due to the imbalance in ethnicity subgroups. In such a case, the model prediction will be biased towards the over-represented ethnicity subgroup. Therefore, it is important to estimate the bias of one demographic group on the prediction of others. For this purpose, in this scenario, the pre-trained <italic>race</italic> prediction models are evaluated across different <italic>gender</italic> subgroups and vice versa. This scenario showcases the performance of the pre-trained models across the intersection of different demographic groups.</p>
<p>
<xref ref-type="table" rid="T5">Table 5</xref> shows the results of this experiment. On evaluating the performance across all the datasets using different pre-trained race prediction models, it is observed that the models trained on the UTKFace and LFWA datasets result in a high error rate for predicting race <italic>R</italic>2 across <italic>G</italic>2, i.e., subgroup (<italic>R</italic>2, <italic>G</italic>2). It is also observed that none of the samples in this intersectional subgroup are correctly classified by the model trained on the LFWA dataset when evaluated on the UTKFace dataset. This results in a high PSE value of 65.83%. For gender prediction across <italic>race</italic> subgroups, it is observed that all the pre-trained <italic>gender</italic> prediction models (except model trained on the LFWA dataset when evaluated on the MORPH dataset) perform worse for predicting gender <italic>G</italic>2 across <italic>R</italic>2, i.e., subgroup (<italic>G</italic>2, <italic>R</italic>2). The results from <xref ref-type="table" rid="T5">Table 5</xref> highlight that the majority of the pre-trained <italic>race</italic> and <italic>gender</italic> prediction models do not perform well for (<italic>R</italic>2, <italic>G</italic>2) and (<italic>G</italic>2, <italic>R</italic>2) subgroups, respectively.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Performance of pre-trained <italic>race</italic> prediction models (%) across different <italic>gender</italic> subgroups and <italic>gender</italic> prediction models across <italic>race</italic> subgroups of a given dataset for intersectional demographic subgroup analysis scenario.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="3">Bias estimated on</th>
<th rowspan="3" align="center">Model trained on</th>
<th colspan="4" align="center">Error</th>
<th rowspan="3" align="center">1 &#x2013; DI</th>
<th rowspan="3" align="center">AFR</th>
<th rowspan="3" align="center">DoB</th>
<th rowspan="3" align="center">PSE</th>
</tr>
<tr>
<th colspan="2" align="center">G1</th>
<th colspan="2" align="center">G2</th>
</tr>
<tr>
<th align="center">R1</th>
<th align="center">R2</th>
<th align="center">R1</th>
<th align="center">R2</th>
</tr>
</thead>
<tbody>
<tr>
<td colspan="10">Race prediction across gender subgroups</td>
</tr>
<tr>
<td>&#x2003;UTKFace</td>
<td align="left">MORPH</td>
<td align="center">35.70</td>
<td align="center">18.41</td>
<td align="center">20.76</td>
<td align="center">26.47</td>
<td align="center">14.21</td>
<td align="center">25.33</td>
<td align="center">5.74</td>
<td align="center">15.09</td>
</tr>
<tr>
<td/>
<td align="left">LFWA</td>
<td align="center">1.15</td>
<td align="center">97.52</td>
<td align="center">1.30</td>
<td align="center">100.00</td>
<td align="center">98.75</td>
<td align="center">49.98</td>
<td align="center">48.76</td>
<td align="center">65.83</td>
</tr>
<tr>
<td>&#x2003;MORPH</td>
<td align="left">UTKFace</td>
<td align="center">0.56</td>
<td align="center">79.71</td>
<td align="center">0.36</td>
<td align="center">81.95</td>
<td align="center">80.74</td>
<td align="center">40.64</td>
<td align="center">40.18</td>
<td align="center">53.85</td>
</tr>
<tr>
<td/>
<td align="left">LFWA</td>
<td align="center">0.00</td>
<td align="center">96.55</td>
<td align="center">0.00</td>
<td align="center">98.74</td>
<td align="center">97.64</td>
<td align="center">48.82</td>
<td align="center">48.82</td>
<td align="center">65.09</td>
</tr>
<tr>
<td>&#x2003;LFWA</td>
<td align="left">UTKFace</td>
<td align="center">20.49</td>
<td align="center">58.96</td>
<td align="center">9.52</td>
<td align="center">67.40</td>
<td align="center">56.17</td>
<td align="center">39.09</td>
<td align="center">24.08</td>
<td align="center">39.78</td>
</tr>
<tr>
<td/>
<td align="left">MORPH</td>
<td align="center">83.24</td>
<td align="center">6.99</td>
<td align="center">83.58</td>
<td align="center">10.87</td>
<td align="center">81.78</td>
<td align="center">46.17</td>
<td align="center">37.23</td>
<td align="center">55.06</td>
</tr>
<tr>
<td colspan="10">Gender prediction across race subgroups</td>
</tr>
<tr>
<td>&#x2003;UTKFace</td>
<td align="left">MORPH</td>
<td align="center">32.07</td>
<td align="center">38.95</td>
<td align="center">16.42</td>
<td align="center">54.91</td>
<td align="center">28.09</td>
<td align="center">35.58</td>
<td align="center">11.34</td>
<td align="center">25.00</td>
</tr>
<tr>
<td/>
<td align="left">LFWA</td>
<td align="center">28.97</td>
<td align="center">43.68</td>
<td align="center">9.46</td>
<td align="center">62.75</td>
<td align="center">39.79</td>
<td align="center">36.21</td>
<td align="center">16.99</td>
<td align="center">30.99</td>
</tr>
<tr>
<td>&#x2003;MORPH</td>
<td align="left">UTKFace</td>
<td align="center">36.76</td>
<td align="center">11.05</td>
<td align="center">36.85</td>
<td align="center">41.81</td>
<td align="center">18.38</td>
<td align="center">31.61</td>
<td align="center">7.66</td>
<td align="center">19.22</td>
</tr>
<tr>
<td/>
<td align="left">LFWA</td>
<td align="center">44.50</td>
<td align="center">17.80</td>
<td align="center">37.70</td>
<td align="center">27.66</td>
<td align="center">23.19</td>
<td align="center">31.91</td>
<td align="center">9.18</td>
<td align="center">21.43</td>
</tr>
<tr>
<td>&#x2003;LFWA</td>
<td align="left">UTKFace</td>
<td align="center">30.70</td>
<td align="center">35.28</td>
<td align="center">31.01</td>
<td align="center">50.00</td>
<td align="center">17.08</td>
<td align="center">36.75</td>
<td align="center">5.89</td>
<td align="center">19.91</td>
</tr>
<tr>
<td/>
<td align="left">MORPH</td>
<td align="center">19.65</td>
<td align="center">57.28</td>
<td align="center">17.47</td>
<td align="center">58.70</td>
<td align="center">48.39</td>
<td align="center">38.27</td>
<td align="center">19.71</td>
<td align="center">35.46</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s5-2">
<label>5.2.</label>
<title>Bias Mitigation</title>
<p>The experiments performed for bias estimation show that the pre-trained models do not give equal performance across different subgroups. Therefore, in this experiment, a single uniform Subgroup Invariant Perturbation is learned by minimizing the PSE of the pre-trained model prediction to achieve improved and almost equal performance across different subgroups. Multiple experiments are performed to evaluate the effectiveness of the proposed algorithm to mitigate the effect of bias in pre-trained model prediction. As mentioned in <xref ref-type="sec" rid="s3-3">Section 3.3</xref>, SIP learned corresponding to the &#x2018;independent subgroup analysis&#x2019; scenario is used to evaluate the performance of the proposed algorithm for the &#x2018;intersectional subgroup analysis&#x2019; scenario as well. The performance of the proposed algorithm is compared with pre-trained and fine-tuned model predictions. Performance is evaluated using multiple existing metrics and the proposed metric PSE. Additionally, we have compared the number of trainable parameters of the proposed algorithm with model fine-tuning. Experimental setup of this experiment is discussed below:</p>
<p>Experimental Setup: In this experiment, SIP is learned corresponding to the training set of all the datasets individually other than the dataset on which the pre-trained model is trained. The learned SIP is added to the testing set of the corresponding dataset for evaluating the performance of the proposed algorithm. For instance, the model pre-trained on the MORPH dataset learns SIP using the training set of the UTKFace dataset and bias is estimated on the testing set of the UTKFace dataset. Similarly, during bias estimation of the MORPH model on the LFWA dataset, SIP learned on the training set of the LFWA dataset is used. For fine-tuning, the pre-trained model is updated using the training set of a given dataset and evaluated on the testing set of the corresponding dataset. The performance of the pre-trained model is evaluated on the testing set of the corresponding dataset.</p>
<sec id="s5-2-1">
<title>5.2.1. Independent Demographic Subgroup Analysis</title>
<p>The results of the pre-trained model, fine-tuned model, and the proposed mitigation algorithm are summarized in <xref ref-type="table" rid="T6">Table 6</xref>. It is observed that the proposed algorithm reduces the bias in the model prediction and enhances the performance. For instance, the proposed algorithm reduces the PSE by 8.99% and 21.39% from the pre-trained and fine-tuned MORPH model predictions, respectively, for the UTKFace dataset. It is interesting to observe that fine-tuning increases the bias in the model prediction and decreases the overall performance. This is because the fine-tuned model decreases the error rate from 33.95 to 4.85% of subgroup <italic>R</italic>1 but increases the error rate of subgroup <italic>R</italic>2 from 18.61 to 46.37% compared to the pre-trained model. The UTKFace dataset has an under-representation of subgroup <italic>R</italic>2. Therefore, a model fine-tuned on this dataset decreases the error rate of subgroup <italic>R</italic>1 and penalizes subgroup <italic>R</italic>2. A similar observation can be drawn from the subgroup-specific error rates of fine-tuned MORPH and UTKFace models on the LFWA dataset, due to the minority of subgroup <italic>R</italic>2. On the other hand, the proposed algorithm overcomes the problem and reduces the performance gap across different subgroups.</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Performance of <italic>race</italic> prediction models (%) after bias mitigation using the proposed and existing algorithms [Multi-task (<xref ref-type="bibr" rid="B9">Das et al., 2018</xref>) and Filter Drop (<xref ref-type="bibr" rid="B24">Nagpal et al., 2020</xref>)] for independent demographic subgroup analysis scenario.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="2">Bias estimated on</th>
<th rowspan="2" align="center">Model trained on</th>
<th rowspan="2" align="center"/>
<th colspan="2" align="center">Error</th>
<th rowspan="2" align="center">1 &#x2013; DI</th>
<th rowspan="2" align="center">AFR</th>
<th rowspan="2" align="center">DoB</th>
<th rowspan="2" align="center">PSE</th>
</tr>
<tr>
<th align="center">R1</th>
<th align="center">R2</th>
</tr>
</thead>
<tbody>
<tr>
<td>UTKFace</td>
<td align="left">MORPH</td>
<td align="left">Pre-trained</td>
<td align="center">33.95</td>
<td align="center">18.61</td>
<td align="center">18.86</td>
<td align="center">26.27</td>
<td align="center">7.67</td>
<td align="center">17.60</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Fine-tuned</td>
<td align="center">4.85</td>
<td align="center">46.37</td>
<td align="center">43.63</td>
<td align="center">25.61</td>
<td align="center">20.76</td>
<td align="center">30.00</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Multi-task</td>
<td align="center">29.66</td>
<td align="center">13.32</td>
<td align="center">18.85</td>
<td align="center">21.48</td>
<td align="center">8.17</td>
<td align="center">16.17</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Filter drop</td>
<td align="center">27.95</td>
<td align="center">16.16</td>
<td align="center">14.06</td>
<td align="center">22.04</td>
<td align="center">5.90</td>
<td align="center">14.00</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Proposed</td>
<td align="center">14.55</td>
<td align="center">19.74</td>
<td align="center">6.08</td>
<td align="center">17.17</td>
<td align="center">2.59</td>
<td align="center">
<bold>8.61</bold>
</td>
</tr>
<tr>
<td/>
<td align="left">LFWA</td>
<td align="left">Pre-trained</td>
<td align="center">0.08</td>
<td align="center">97.45</td>
<td align="center">97.45</td>
<td align="center">48.76</td>
<td align="center">48.68</td>
<td align="center">64.96</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Fine-tuned</td>
<td align="center">3.55</td>
<td align="center">53.44</td>
<td align="center">51.72</td>
<td align="center">28.49</td>
<td align="center">24.94</td>
<td align="center">35.05</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Multi-task</td>
<td align="center">31.55</td>
<td align="center">14.50</td>
<td align="center">19.93</td>
<td align="center">23.02</td>
<td align="center">8.53</td>
<td align="center">17.16</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Filter drop</td>
<td align="center">26.78</td>
<td align="center">14.84</td>
<td align="center">14.01</td>
<td align="center">20.80</td>
<td align="center">5.97</td>
<td align="center">13.59</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Proposed</td>
<td align="center">15.42</td>
<td align="center">25.16</td>
<td align="center">11.52</td>
<td align="center">20.29</td>
<td align="center">4.87</td>
<td align="center">
<bold>12.23</bold>
</td>
</tr>
<tr>
<td>MORPH</td>
<td align="left">UTKFace</td>
<td align="left">Pre-trained</td>
<td align="center">0.31</td>
<td align="center">86.45</td>
<td align="center">86.41</td>
<td align="center">43.37</td>
<td align="center">43.07</td>
<td align="center">57.62</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Fine-tuned</td>
<td align="center">6.04</td>
<td align="center">1.85</td>
<td align="center">4.27</td>
<td align="center">3.94</td>
<td align="center">2.09</td>
<td align="center">3.43</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Multi-task</td>
<td align="center">3.32</td>
<td align="center">5.99</td>
<td align="center">2.75</td>
<td align="center">4.65</td>
<td align="center">1.34</td>
<td align="center">2.91</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Filter drop</td>
<td align="center">4.13</td>
<td align="center">5.76</td>
<td align="center">1.69</td>
<td align="center">4.93</td>
<td align="center">0.82</td>
<td align="center">2.48</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Proposed</td>
<td align="center">2.47</td>
<td align="center">4.91</td>
<td align="center">2.51</td>
<td align="center">3.68</td>
<td align="center">1.22</td>
<td align="center">
<bold>2.47</bold>
</td>
</tr>
<tr>
<td/>
<td align="left">LFWA</td>
<td align="center">Pre-trained</td>
<td align="center">0.00</td>
<td align="center">98.11</td>
<td align="center">98.11</td>
<td align="center">49.05</td>
<td align="center">49.05</td>
<td align="center">65.40</td>
</tr>
<tr>
<td/>
<td/>
<td align="center">Fine-tuned</td>
<td align="center">7.09</td>
<td align="center">1.73</td>
<td align="center">5.46</td>
<td align="center">4.41</td>
<td align="center">2.68</td>
<td align="center">4.18</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Multi-task</td>
<td align="center">2.07</td>
<td align="center">7.96</td>
<td align="center">6.01</td>
<td align="center">5.00</td>
<td align="center">2.95</td>
<td align="center">4.65</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Filter drop</td>
<td align="center">3.22</td>
<td align="center">6.54</td>
<td align="center">3.42</td>
<td align="center">4.87</td>
<td align="center">1.66</td>
<td align="center">3.32</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Proposed</td>
<td align="center">1.31</td>
<td align="center">4.46</td>
<td align="center">3.20</td>
<td align="center">2.88</td>
<td align="center">1.57</td>
<td align="center">
<bold>2.55</bold>
</td>
</tr>
<tr>
<td>LFWA</td>
<td align="left">UTKFace</td>
<td align="left">Pre-trained</td>
<td align="center">19.85</td>
<td align="center">63.86</td>
<td align="center">54.92</td>
<td align="center">41.85</td>
<td align="center">22.00</td>
<td align="center">39.59</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Fine-tuned</td>
<td align="center">0.71</td>
<td align="center">81.76</td>
<td align="center">81.63</td>
<td align="center">41.23</td>
<td align="center">40.52</td>
<td align="center">54.46</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Multi-task</td>
<td align="center">32.04</td>
<td align="center">24.92</td>
<td align="center">9.49</td>
<td align="center">28.47</td>
<td align="center">3.56</td>
<td align="center">13.84</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Filter drop</td>
<td align="center">35.62</td>
<td align="center">28.78</td>
<td align="center">9.61</td>
<td align="center">32.19</td>
<td align="center">3.42</td>
<td align="center">15.07</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Proposed</td>
<td align="center">8.56</td>
<td align="center">22.11</td>
<td align="center">14.83</td>
<td align="center">15.33</td>
<td align="center">6.77</td>
<td align="center">
<bold>12.31</bold>
</td>
</tr>
<tr>
<td/>
<td align="left">MORPH</td>
<td align="left">Pre-trained</td>
<td align="center">81.95</td>
<td align="center">2.11</td>
<td align="center">81.56</td>
<td align="center">42.02</td>
<td align="center">39.92</td>
<td align="center">54.50</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Fine-tuned</td>
<td align="center">1.47</td>
<td align="center">78.25</td>
<td align="center">77.93</td>
<td align="center">39.86</td>
<td align="center">38.39</td>
<td align="center">52.06</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Multi-task</td>
<td align="center">32.61</td>
<td align="center">27.02</td>
<td align="center">7.66</td>
<td align="center">29.81</td>
<td align="center">2.80</td>
<td align="center">13.42</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Filter drop</td>
<td align="center">32.69</td>
<td align="center">26.32</td>
<td align="center">8.64</td>
<td align="center">29.50</td>
<td align="center">3.19</td>
<td align="center">13.78</td>
</tr>
<tr>
<td/>
<td/>
<td align="left">Proposed</td>
<td align="center">8.75</td>
<td align="center">23.16</td>
<td align="center">15.80</td>
<td align="center">15.95</td>
<td align="center">7.20</td>
<td align="center">
<bold>12.98</bold>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The lowest PSE value is highlighted.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>The performance of <italic>gender</italic> prediction models is shown in <xref ref-type="table" rid="T7">Table 7</xref>. It is observed that the proposed algorithm reduces the PSE of each model corresponding to all the datasets. For instance, the PSE of the pre-trained and fine-tuned UTKFace model corresponding to the MORPH dataset is 22.91% and 17.53%, respectively. The proposed algorithm reduces the PSE to 8.34%. This showcase that the proposed algorithm is jointly able to reduce the bias in model prediction and improve the overall performance of the model. <xref ref-type="fig" rid="F4">Figure 4</xref> shows the visualization of the learned Subgroup Invariant Perturbation (SIP). A face like structure can clearly be seen in all the perturbations.</p>
<table-wrap id="T7" position="float">
<label>TABLE 7</label>
<caption>
<p>Performance of <italic>gender</italic> prediction models (%) with the proposed and existing bias mitigation algorithms for independent demographic subgroup analysis scenario.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="2">Bias estimated on</th>
<th rowspan="2" align="center">Model trained on</th>
<th rowspan="2" align="center"/>
<th colspan="2" align="center">Error</th>
<th rowspan="2" align="center">1 &#x2013; DI</th>
<th rowspan="2" align="center">AFR</th>
<th rowspan="2" align="center">DoB</th>
<th rowspan="2" align="center">PSE</th>
</tr>
<tr>
<th align="center">
<italic>G</italic>1</th>
<th align="center">
<italic>G</italic>2</th>
</tr>
</thead>
<tbody>
<tr>
<td>UTKFace</td>
<td>MORPH</td>
<td>Pre-trained</td>
<td>22.64</td>
<td>41.37</td>
<td>24.56</td>
<td>32.13</td>
<td>9.36</td>
<td>22.02</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>15.61</td>
<td>29.30</td>
<td>16.23</td>
<td>22.45</td>
<td>6.84</td>
<td>15.17</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>23.65</td>
<td>37.87</td>
<td>18.62</td>
<td>30.75</td>
<td>7.11</td>
<td>18.83</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>28.08</td>
<td>34.99</td>
<td>9.60</td>
<td>31.52</td>
<td>3.46</td>
<td>14.86</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>10.18</td>
<td>22.30</td>
<td>13.50</td>
<td>16.24</td>
<td>6.06</td>
<td>
<bold>11.93</bold>
</td>
</tr>
<tr>
<td/>
<td>LFWA</td>
<td>Pre-trained</td>
<td>18.25</td>
<td>51.74</td>
<td>40.97</td>
<td>34.99</td>
<td>16.74</td>
<td>30.90</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>15.00</td>
<td>31.67</td>
<td>19.62</td>
<td>23.33</td>
<td>8.33</td>
<td>17.09</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>24.01</td>
<td>38.52</td>
<td>19.09</td>
<td>31.26</td>
<td>7.25</td>
<td>19.20</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>29.77</td>
<td>33.33</td>
<td>5.06</td>
<td>31.54</td>
<td>1.78</td>
<td>12.79</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>11.91</td>
<td>23.15</td>
<td>12.77</td>
<td>17.53</td>
<td>5.62</td>
<td>
<bold>11.97</bold>
</td>
</tr>
<tr>
<td/>
<td>CelebA</td>
<td>Pre-trained</td>
<td>42.96</td>
<td>47.32</td>
<td>7.65</td>
<td>45.13</td>
<td>2.18</td>
<td>18.32</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>15.14</td>
<td>33.44</td>
<td>21.57</td>
<td>24.28</td>
<td>9.15</td>
<td>18.33</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>30.86</td>
<td>38.08</td>
<td>10.43</td>
<td>34.46</td>
<td>3.61</td>
<td>16.17</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>29.12</td>
<td>34.93</td>
<td>8.19</td>
<td>32.01</td>
<td>2.91</td>
<td>14.37</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>34.70</td>
<td>34.98</td>
<td>0.44</td>
<td>34.83</td>
<td>0.14</td>
<td>
<bold>11.80</bold>
</td>
</tr>
<tr>
<td>MORPH</td>
<td>UTKFace</td>
<td>Pre-trained</td>
<td>41.27</td>
<td>19.02</td>
<td>27.48</td>
<td>30.14</td>
<td>11.12</td>
<td>22.91</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>5.84</td>
<td>28.52</td>
<td>24.09</td>
<td>17.17</td>
<td>11.34</td>
<td>17.53</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>10.00</td>
<td>22.80</td>
<td>14.22</td>
<td>16.39</td>
<td>6.40</td>
<td>12.34</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>8.88</td>
<td>21.67</td>
<td>14.03</td>
<td>15.27</td>
<td>6.40</td>
<td>11.90</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>14.11</td>
<td>3.59</td>
<td>10.92</td>
<td>8.84</td>
<td>5.26</td>
<td>
<bold>8.34</bold>
</td>
</tr>
<tr>
<td/>
<td>LFWA</td>
<td>Pre-trained</td>
<td>53.80</td>
<td>20.58</td>
<td>41.84</td>
<td>37.19</td>
<td>16.61</td>
<td>31.88</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>8.69</td>
<td>18.75</td>
<td>11.02</td>
<td>13.72</td>
<td>5.03</td>
<td>9.92</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>9.63</td>
<td>23.07</td>
<td>14.87</td>
<td>16.34</td>
<td>6.72</td>
<td>12.64</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>9.88</td>
<td>20.94</td>
<td>12.26</td>
<td>15.4</td>
<td>5.53</td>
<td>11.06</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>15.92</td>
<td>4.75</td>
<td>11.72</td>
<td>10.33</td>
<td>5.58</td>
<td>
<bold>9.21</bold>
</td>
</tr>
<tr>
<td/>
<td>CelebA</td>
<td>Pre-trained</td>
<td>66.47</td>
<td>31.45</td>
<td>51.09</td>
<td>48.95</td>
<td>17.51</td>
<td>39.18</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>7.71</td>
<td>22.01</td>
<td>15.50</td>
<td>14.85</td>
<td>7.15</td>
<td>12.50</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>8.19</td>
<td>24.60</td>
<td>17.87</td>
<td>16.39</td>
<td>8.21</td>
<td>14.16</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>9.80</td>
<td>26.47</td>
<td>18.48</td>
<td>18.12</td>
<td>8.33</td>
<td>14.98</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>19.49</td>
<td>3.26</td>
<td>16.78</td>
<td>11.37</td>
<td>8.11</td>
<td>
<bold>12.09</bold>
</td>
</tr>
<tr>
<td>LFWA</td>
<td>UTKFace</td>
<td>Pre-trained</td>
<td>29.88</td>
<td>39.39</td>
<td>13.57</td>
<td>34.63</td>
<td>4.75</td>
<td>17.65</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>3.99</td>
<td>54.18</td>
<td>52.28</td>
<td>29.08</td>
<td>25.09</td>
<td>35.48</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>27.46</td>
<td>36.70</td>
<td>12.73</td>
<td>32.07</td>
<td>4.62</td>
<td>16.47</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>30.16</td>
<td>34.05</td>
<td>5.56</td>
<td>32.09</td>
<td>1.95</td>
<td>13.20</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>18.12</td>
<td>26.58</td>
<td>10.33</td>
<td>22.35</td>
<td>4.23</td>
<td>
<bold>12.30</bold>
</td>
</tr>
<tr>
<td/>
<td>MORPH</td>
<td>Pre-trained</td>
<td>19.12</td>
<td>58.50</td>
<td>48.69</td>
<td>38.30</td>
<td>19.69</td>
<td>35.56</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>5.50</td>
<td>50.42</td>
<td>47.53</td>
<td>27.95</td>
<td>22.46</td>
<td>32.65</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>39.67</td>
<td>28.74</td>
<td>15.34</td>
<td>34.20</td>
<td>5.47</td>
<td>18.34</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>34.43</td>
<td>37.32</td>
<td>4.40</td>
<td>35.86</td>
<td>1.45</td>
<td>13.90</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>15.95</td>
<td>27.33</td>
<td>13.55</td>
<td>21.63</td>
<td>5.69</td>
<td>
<bold>13.62</bold>
</td>
</tr>
<tr>
<td>
</td>
<td>CelebA</td>
<td>Pre-trained</td>
<td>16.31</td>
<td>45.21</td>
<td>34.53</td>
<td>30.76</td>
<td>14.45</td>
<td>26.58</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>10.56</td>
<td>36.79</td>
<td>29.33</td>
<td>23.67</td>
<td>13.11</td>
<td>22.04</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>25.41</td>
<td>37.41</td>
<td>16.07</td>
<td>31.40</td>
<td>6.00</td>
<td>17.82</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>25.44</td>
<td>34.72</td>
<td>12.49</td>
<td>30.08</td>
<td>4.64</td>
<td>15.74</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>12.61</td>
<td>28.02</td>
<td>17.64</td>
<td>20.31</td>
<td>7.70</td>
<td>
<bold>15.22</bold>
</td>
</tr>
<tr>
<td>CelebA</td>
<td>UTKFace</td>
<td>Pre-trained</td>
<td>42.62</td>
<td>42.77</td>
<td>0.27</td>
<td>42.69</td>
<td>0.07</td>
<td>14.34</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>28.26</td>
<td>11.97</td>
<td>18.51</td>
<td>20.11</td>
<td>8.14</td>
<td>15.59</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>33.90</td>
<td>33.92</td>
<td>0.04</td>
<td>33.91</td>
<td>0.00</td>
<td>
<bold>11.32</bold>
</td>
</tr>
<tr>
<td/>
<td>MORPH</td>
<td>Pre-trained</td>
<td>38.67</td>
<td>56.47</td>
<td>29.02</td>
<td>47.56</td>
<td>8.90</td>
<td>28.49</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>23.85</td>
<td>14.26</td>
<td>11.19</td>
<td>19.05</td>
<td>4.79</td>
<td>11.68</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>14.99</td>
<td>21.73</td>
<td>7.28</td>
<td>18.08</td>
<td>3.37</td>
<td>
<bold>9.58</bold>
</td>
</tr>
<tr>
<td/>
<td>LFWA</td>
<td>Pre-trained</td>
<td>11.71</td>
<td>55.06</td>
<td>49.10</td>
<td>33.38</td>
<td>21.67</td>
<td>34.72</td>
</tr>
<tr>
<td/>
<td/>
<td>Fine-tuned</td>
<td>26.03</td>
<td>12.94</td>
<td>15.04</td>
<td>19.48</td>
<td>6.54</td>
<td>13.69</td>
</tr>
<tr>
<td/>
<td/>
<td>Multi-task</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
</tr>
<tr>
<td/>
<td/>
<td>Filter drop</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
<td>&#x2014;</td>
</tr>
<tr>
<td/>
<td/>
<td>Proposed</td>
<td>11.57</td>
<td>23.33</td>
<td>13.31</td>
<td>17.45</td>
<td>5.88</td>
<td>
<bold>12.21</bold>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The lowest PSE value is highlighted.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Visualization of the learned Subgroup Invariant Perturbation (SIP) corresponding to the <bold>(A)</bold> race and <bold>(B)</bold> gender prediction models.</p>
</caption>
<graphic xlink:href="fdata-03-590296-g004.tif"/>
</fig>
<p>The proposed algorithm is compared with two existing bias mitigation algorithms (<xref ref-type="bibr" rid="B9">Das et al., 2018</xref>; <xref ref-type="bibr" rid="B24">Nagpal et al., 2020</xref>). The comparison of the results for <italic>race</italic> and <italic>gender</italic> prediction are shown in <xref ref-type="table" rid="T6">Tables 6</xref> and <xref ref-type="table" rid="T7">7</xref>, respectively. It is observed that the proposed algorithm outperforms existing algorithms for both <italic>race</italic> and <italic>gender</italic> prediction. The proposed algorithm jointly optimizes bias and the overall model performance while the existing algorithms focus on bias optimization only. Therefore, the PSE of the proposed algorithm is minimum compared to others. For instance, in <italic>gender</italic> prediction (<xref ref-type="table" rid="T7">Table 7</xref>), the PSE of the CelebA model corresponding to the UTKFace dataset for Multi-task (<xref ref-type="bibr" rid="B9">Das et al., 2018</xref>), Filter Drop (<xref ref-type="bibr" rid="B24">Nagpal et al., 2020</xref>), and the proposed algorithm is 16.17%, 14.37%, and 11.80%, respectively. This shows the effectiveness of the proposed algorithm for independent demographic subgroup analysis scenario. In our experimental setup, the existing algorithms are not applicable for <italic>gender</italic> prediction on the CelebA dataset. Apart from this, we have also performed an experiment, where we reduce bias and improve the model performance alternatively using the proposed bias mitigation algorithm. The results of this experiment are compared with the proposed bias mitigation algorithm, where we jointly reduce the bias and improve the model performance. <xref ref-type="fig" rid="F5">Figure 5</xref> shows the comparison of the results of alternate and joint optimization corresponding to the independent demographic subgroup analysis scenario. It is observed that joint optimization leads to better results as it provides combined supervision of bias and model performance for better learning of SIP that results in better performance.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Results shows the comparison of alternate and joint optimization using the proposed algorithm corresponding to the independent demographic subgroup analysis scenario. The proposed algorithm gives the lowest PSE.</p>
</caption>
<graphic xlink:href="fdata-03-590296-g005.tif"/>
</fig>
</sec>
<sec id="s5-2-2">
<title>5.2.2. Intersectional Demographic Subgroup Analysis</title>
<p>To further evaluate the effectiveness of the proposed algorithm across the intersection of different demographic groups, two different experiments are performed. In the first experiment, race classification is performed across <italic>gender</italic> subgroups. While in the second experiment, gender classification is performed across <italic>race</italic> subgroups. These experiments are performed to analyze the presence of <italic>gender</italic> bias on race prediction and <italic>race</italic> bias on gender prediction. Comparison is performed with pre-trained and fine-tuned model predictions. <xref ref-type="fig" rid="F6">Figure 6</xref> shows the PSE corresponding to the first and second experiments. It is observed that in most of the cases, the proposed algorithm gives the lowest PSE. For instance, the PSE of pre-trained and fine-tuned UTKFace models corresponding to the MORPH dataset for gender prediction is 20.58% and 15.91%, respectively. The proposed algorithm reduces the PSE to 8.56%. This indicates that the proposed algorithm is able to reduce the effect of bias of one demographic group on the prediction of others. The reduction in PSE shows the effectiveness of the proposed algorithm.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Results of <bold>(A</bold>&#x2013;<bold>C)</bold> race classification across <italic>gender</italic> subgroups corresponding to the MORPH, UTKFace, and LFWA models, respectively and <bold>(D</bold>&#x2013;<bold>F)</bold> gender classification across <italic>race</italic> subgroups corresponding to the MORPH, UTKFace, and LFWA models, respectively for intersectional demographic subgroup analysis scenario. Comparison is shown with pre-trained and fine-tuned model predictions along with existing algorithms for bias mitigation.</p>
</caption>
<graphic xlink:href="fdata-03-590296-g006.tif"/>
</fig>
<p>
<xref ref-type="fig" rid="F7">Figure 7</xref> compares the performance of the proposed algorithm and the pre-trained model using the score distribution of the model prediction. The results are shown for <italic>race</italic> prediction across different <italic>gender</italic> subgroups of the MORPH model on the LFWA dataset. It is observed that the proposed algorithm reduces the overlap among the subgroups and separates them from each other. Class Activation Map (CAM) of race classification across <italic>gender</italic> subgroups on the UTKFace dataset using the MORPH race prediction model is shown in <xref ref-type="fig" rid="F8">Figure 8</xref>. It is observed that the pre-trained and fine-tuned models focus on different facial regions across the intersection of different demographic subgroups. On the other hand, the proposed algorithm tries to focus on the entire facial region irrespective of different subgroups. This showcases the effectiveness of the learned SIP to mitigate the effect of demographic subgroup bias by enforcing the model to extract features from the entire facial region for discrimination instead of subgroup-specific regions.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Score distribution of the MORPH race prediction model across <italic>gender</italic> subgroups <bold>(A)</bold> <italic>G</italic>1 and <bold>(B)</bold> <italic>G</italic>2 on the LFWA dataset. The first graph of <bold>(A)</bold> and <bold>(B)</bold> shows the score distribution of the pre-trained model and the second graph shows the score distribution of the proposed algorithm.</p>
</caption>
<graphic xlink:href="fdata-03-590296-g007.tif"/>
</fig>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Class Activation Map of race classification across <italic>gender</italic> subgroups on the UTKFace dataset using the MORPH race prediction model. Top row shows the visualization for the pre-trained model prediction, middle row for the fine-tuned model prediction, and the bottom row for the proposed algorithm. It is observed that the proposed algorithm focuses on the entire facial region instead of the subgroup-specific region for feature extraction.</p>
</caption>
<graphic xlink:href="fdata-03-590296-g008.tif"/>
</fig>
<p>On comparing the number of trainable parameters of the proposed algorithm with model fine-tuning, it is observed that the proposed algorithm requires number of parameters equal to the size of the input image, i.e., 12K parameters. On the other hand, model fine-tuning requires updation of 0.52M parameters, which is approximately 43 times more than the proposed algorithm. This shows that the proposed algorithm is computationally efficient.</p>
<p>
<xref ref-type="fig" rid="F6">Figure 6</xref> shows the comparison of the proposed algorithm with existing bias mitigation algorithms (<xref ref-type="bibr" rid="B9">Das et al., 2018</xref>; <xref ref-type="bibr" rid="B24">Nagpal et al., 2020</xref>). It is observed that in most of the cases, the proposed algorithm performs better than existing algorithms while giving comparable results for others. For instance, the PSE of the proposed and existing algorithms for <italic>race</italic> prediction across <italic>gender</italic> subgroups of the MORPH model corresponding to the UTKFace and LFWA datasets are 10.24%, 14.83%, 14.10% and 12.06%, 11.94%, 13.57%, respectively. It is important to note that the proposed algorithm does not require model training and therefore is computationally efficient.</p>
</sec>
</sec>
</sec>
<sec id="s6">
<label>6.</label>
<title> Discussion and Conclusion</title>
<p>The effect of demographic subgroup bias on the performance of commercial and pre-trained models is studied in the past. A lot of progress is made towards estimating and mitigating the influence of bias on model prediction. However, studies have shown that there is a trade-off between fairness and model performance. Maintaining a balance between the two is an important factor. This motivated us to propose a unified metric to measure the trade-off and an algorithm to mitigate the effect of bias on pre-trained model prediction.</p>
<p>We used multiple pre-trained race and gender prediction models for bias estimation and mitigation. Since the existing metrics either evaluate the performance gap across different subgroups or the overall model performance, therefore we have introduced a unified metric, PSE, to jointly estimate the bias in model prediction and the overall model performance. Additionally, a novel algorithm is proposed to mitigate the effect of bias using adversarial perturbation by reducing the PSE of the model prediction. We showed that a single uniform Subgroup Invariant Perturbation (SIP), when added to the input images, is able to mitigate the effect of bias on model prediction.</p>
<p>During bias estimation, it is observed that PSE reflects both error and disparity in subgroups. On analyzing the existing metrics, it is observed that DI and DoB do not reflect the overall model performance, while AFR does not reflect the performance gap across different subgroups. On the other hand, we have experimentally validated in <xref ref-type="table" rid="T3">Tables 3</xref>&#x2013;<xref ref-type="table" rid="T5">5</xref> that PSE considers the model performance along with fairness. Therefore, PSE is utilized by the proposed algorithm to learn SIP for bias mitigation. The performance of race and gender prediction models corresponding to the independent demographic subgroup analysis scenario are summarized in <xref ref-type="table" rid="T6">Tables 6</xref> and <xref ref-type="table" rid="T7">7</xref>, respectively. We have found that the proposed algorithm is able to reduce the PSE of all the pre-trained models corresponding to all the datasets. To test the proposed algorithm for mitigating the influence of bias corresponding to the intersectional subgroup analysis scenario, SIP learned corresponding to the independent subgroup analysis scenario is used. <xref ref-type="fig" rid="F6">Figure 6</xref> shows that the proposed algorithm is effective in mitigating the intersectional subgroup bias. This is validated by the score distributions in <xref ref-type="fig" rid="F7">Figure 7</xref> that shows that the proposed algorithm reduces the overlap between subgroups. We have also found that the proposed algorithm focuses on the entire face for feature extraction instead of subgroup-specific regions in <xref ref-type="fig" rid="F8">Figure 8</xref>.</p>
<p>Existing research towards bias mitigation requires model training to suppress the element of bias for unbiased prediction. However, the proposed algorithm does not require model training for bias mitigation. It requires the number of trainable parameters equal to the size of the input image, which is significantly lower than the model fine-tuning approach. Therefore, the proposed algorithm is computationally efficient. This showcase the applicability of the proposed algorithm in real-world scenarios.</p>
<p>In the future, we plan to extend the proposed algorithm for mitigating the effect of bias due to the influence of multiple demographic subgroups via learning a single Subgroup Invariant Perturbation (SIP). Also, we will investigate the effect of bias on face recognition performance.</p>
</sec>
</body>
<back>
<sec id="s7">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: MORPH: <ext-link ext-link-type="uri" xlink:href="https://ebill.uncw.edu/C20231_ustores/web/store_main.jsp?STOREID=4">https://ebill.uncw.edu/C20231_ustores/web/store_main.jsp?STOREID&#x3d;4</ext-link>, UTKFace: <ext-link ext-link-type="uri" xlink:href="https://susanqq.github.io/UTKFace/">https://susanqq.github.io/UTKFace/</ext-link>, LFWA: <ext-link ext-link-type="uri" xlink:href="http://vis-www.cs.umass.edu/lfw/">http://vis-www.cs.umass.edu/lfw/</ext-link>, CelebA: <ext-link ext-link-type="uri" xlink:href="http://mmlab.ie.cuhk.edu.hk/projects/CelebA.html">http://mmlab.ie.cuhk.edu.hk/projects/CelebA.html</ext-link>.</p>
</sec>
<sec id="s8">
<title>Ethics Statement</title>
<p>Images used in this research are taken from publicly available datasets and the authors of the datasets have not taken explicit written consent from the individual(s) present in the dataset.</p>
</sec>
<sec id="s9">
<title>Author Contributions</title>
<p>PM and SC developed the algorithm, conducted multiple experiments, and analyzed the results under the supervision of RS and MV. All the authors discussed the results, co-wrote, and reviewed the manuscript.</p>
</sec>
<sec sec-type="COI-statement" id="s10">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<ack>
<p>PM is partly supported by DST INSPIRE PhD Fellowship and MV is partially supported through the SwarnaJayanti Fellowship. This research is also partially supported by Facebook Ethics in AI award.</p>
</ack>
<fn-group>
<fn fn-type="other">
<p>
<bold>Specialty section:</bold> This article was submitted to Machine Learning and Artificial Intelligence, a section of the journal Frontiers in Big Data</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> Pin-Yu Chen, IBM Research, United States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> Chulin Xie, University of Illinois at Urbana-Champaign, United States</p>
<p>Ren Wang, University of Michigan, United States</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Alvi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zisserman</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Nell&#xe5;ker</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Turning a blind eye: explicit removal of biases and variation from deep neural network embeddings</article-title>&#x201d;, in <conf-name>European Conference on computer vision</conf-name>. <pub-id pub-id-type="doi">10.2307/2300364</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Barocas</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Selbst</surname>
<given-names>A. D.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Big data&#x2019;s disparate impact</article-title>. <source>Calif. Law Rev.</source> <volume>104</volume>, <fpage>671</fpage>. <pub-id pub-id-type="doi">10.2139/ssrn.2477899</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bolukbasi</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>K.-W.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>J. Y.</given-names>
</name>
<name>
<surname>Saligrama</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Kalai</surname>
<given-names>A. T.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Man is to computer programmer as woman is to homemaker? debiasing word embeddings</article-title>&#x201d;, in <conf-name>Advances in neural information processing systems</conf-name>, <fpage>4349</fpage>&#x2013;<lpage>4357</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-52485-2_4</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Buolamwini</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Gebru</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Gender shades: intersectional accuracy disparities in commercial gender classification</article-title>&#x201d;, in <conf-name>Conference on fairness, accountability and transparency</conf-name>, <fpage>77</fpage>&#x2013;<lpage>91</lpage>. 10.1109/SPW.2016.11 4 </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Calders</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Verwer</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Three naive bayes approaches for discrimination-free classification</article-title>. <source>Data Min. Knowl. Discov.</source> <volume>21</volume>, <fpage>277</fpage>&#x2013;<lpage>292</lpage>. <pub-id pub-id-type="doi">10.1007/s10618-010-0190-x</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Celis</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Rao</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Learning facial recognition biases through vae latent representations</article-title>,&#x201d; in <conf-name>International workshop on fairness, accountability, and transparency in multimedia</conf-name>, <fpage>26</fpage>&#x2013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-020-08688-x</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Conger</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Fausset</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Kovaleski</surname>
<given-names>S. F.</given-names>
</name>
</person-group> (<year>2019</year>). <source>San francisco bans facial recognition technology</source> <ext-link ext-link-type="uri" xlink:href="https://tinyurl.com/y4x6wbos">https://tinyurl.com/y4x6wbos</ext-link>. <comment>[Dataset]</comment>.</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Creager</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Madras</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Jacobsen</surname>
<given-names>J.-H.</given-names>
</name>
<name>
<surname>Weis</surname>
<given-names>M. A.</given-names>
</name>
<name>
<surname>Swersky</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Pitassi</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). &#x201c;<article-title>Flexibly fair representation learning by disentanglement</article-title>&#x201d;, in <conf-name>International conference on machine learning</conf-name>, <fpage>1436</fpage>&#x2013;<lpage>1445</lpage>. <pub-id pub-id-type="doi">10.3389/frai.2020.00033</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Das</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dantcheva</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bremond</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Mitigating bias in gender, age and ethnicity classification: a multi-task convolution neural network approach</article-title>,&#x201d; in <conf-name>European conference on computer vision</conf-name>. <pub-id pub-id-type="doi">10.1007/978-3-030-11009-3_35</pub-id> </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Drozdowski</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Rathgeb</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Dantcheva</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Damer</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Busch</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Demographic bias in biometrics: a survey on an emerging challenge</article-title>. <source>IEEE Trans. Technol. Soc.</source> <volume>28</volume>, <fpage>1728</fpage>. <pub-id pub-id-type="doi">10.1109/TTS.2020.2992344</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Drummond</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Holte</surname>
<given-names>R. C.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2003</year>). &#x201c;<article-title>C4. 5, class imbalance, and cost sensitivity: why under-sampling beats over-sampling</article-title>,&#x201d; in <conf-name>Workshop on learning from imbalanced datasets II (Citeseer)</conf-name>, <fpage>1</fpage>&#x2013;<lpage>8</lpage>. </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Du</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Fairness in deep learning: a computational perspective</source>. <comment>arXiv preprint arXiv:1908.08843</comment>.</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Du</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Fairness in deep learning: a computational perspective</article-title>. <source>IEEE Intell. Syst.</source> <volume>17</volume>, <fpage>156</fpage>. <pub-id pub-id-type="doi">10.1109/MIS.2020.3000681</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dwork</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Immorlica</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Kalai</surname>
<given-names>A. T.</given-names>
</name>
<name>
<surname>Leiserson</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Decoupled classifiers for group-fair and efficient machine learning</article-title>&#x201d;, in <conf-name>Conference on fairness, accountability and transparency</conf-name>, <fpage>119</fpage>&#x2013;<lpage>133</lpage>. <pub-id pub-id-type="doi">10.1145/3357384.3357857</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Feldman</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Friedler</surname>
<given-names>S. A.</given-names>
</name>
<name>
<surname>Moeller</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Scheidegger</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Venkatasubramanian</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Certifying and removing disparate impact</article-title>,&#x201d; in <conf-name>ACM SIGKDD international conference on knowledge discovery and data mining</conf-name>, <fpage>259</fpage>&#x2013;<lpage>268</lpage>. <pub-id pub-id-type="doi">10.1145/2783258.2783311</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gong</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Jain</surname>
<given-names>A. K.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Debface: de-biasing face recognition</article-title>. <comment>arXiv preprint arXiv:1911.08080</comment>. </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Deep imbalanced learning for face recognition and attribute prediction</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>1</volume>, <fpage>17</fpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2019.2914680</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>G. B.</given-names>
</name>
<name>
<surname>Mattar</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Berg</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Learned-Miller</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2008</year>). &#x201c;<article-title>Labeled faces in the wild: a database for studying face recognition in unconstrained environments</article-title>,&#x201d; in <conf-name>Workshop on faces in &#x2018;real-life&#x2019; images: detection, alignment, and recognition</conf-name>. <pub-id pub-id-type="doi">10.1007/2F978-3-319-25958-1&#x005f;8</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kim</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Learning not to learn: training deep neural networks with biased data</article-title>,&#x201d; in <conf-name>IEEE conference on computer vision and pattern recognition</conf-name>. <fpage>9012</fpage>&#x2013;<lpage>9020</lpage>. <pub-id pub-id-type="doi">10.1007/s10489-020-01658-8</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Vasconcelos</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Repair: removing representation bias by dataset resampling</article-title>&#x201d;, in <conf-name>IEEE conference on computer vision and pattern recognition</conf-name>, <fpage>9572</fpage>&#x2013;<lpage>9581</lpage>. <pub-id pub-id-type="doi">10.1038/s41467-020-19784-9</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Deep learning face attributes in the wild</article-title>&#x201d;, in <conf-name>IEEE international conference on computer vision</conf-name>, <fpage>3730</fpage>&#x2013;<lpage>3738</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2015.425</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mehrabi</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Morstatter</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Saxena</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Lerman</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Galstyan</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2019</year>). <source>A survey on bias and fairness in machine learning</source>. <comment>arXiv preprint arXiv:1908.09635</comment>.</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mullick</surname>
<given-names>S. S.</given-names>
</name>
<name>
<surname>Datta</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Generative adversarial minority oversampling</article-title>&#x201d;, in <conf-name>IEEE international conference on computer vision</conf-name>, <fpage>1695</fpage>&#x2013;<lpage>1704</lpage>. </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nagpal</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Vatsa</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Attribute aware filter-drop for bias invariant classification</article-title>,&#x201d; in <conf-name>IEEE/CVF conference on computer vision and pattern recognition workshops</conf-name>, <fpage>32</fpage>&#x2013;<lpage>33</lpage>. </citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nagpal</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Vatsa</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ratha</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Deep learning for face recognition: pride or prejudiced?</source>. <comment>arXiv preprint arXiv:1904.01219</comment>.</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ntoutsi</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Fafalios</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Gadiraju</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Iosifidis</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Nejdl</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Vidal</surname>
<given-names>M.-E.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Bias in data-driven artificial intelligence systems&#x2014;an introductory survey</article-title>. <source>Wiley Interdisciplinary Rev. Data Min. Knowl. Discov.</source> <volume>10</volume>, <fpage>e1356</fpage>. <pub-id pub-id-type="doi">10.1002/widm.1356</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Paolini-Subramanya</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Facial recognition, and bias</source> <ext-link ext-link-type="uri" xlink:href="https://tinyurl.com/y7rat8vb">https://tinyurl.com/y7rat8vb</ext-link>. <comment>[Dataset].</comment>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Parkhi</surname>
<given-names>O. M.</given-names>
</name>
<name>
<surname>Vedaldi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zisserman</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). &#x201c;<article-title>Deep face recognition</article-title>,&#x201d; in <conf-name>British machine vision conference</conf-name>, <fpage>41.1</fpage>&#x2013;<lpage>41.12</lpage>. <pub-id pub-id-type="doi">10.1155/2018/2861695</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Radford</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Joseph</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Theory in, theory out: the uses of social theory in machine learning for social science</article-title>. <source>Front. Big Data</source> <volume>3</volume>, <fpage>18</fpage>. <pub-id pub-id-type="doi">10.3389/fdata.2020.00018</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Rawls</surname>
<given-names>A. W.</given-names>
</name>
<name>
<surname>Ricanek</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2009</year>). &#x201c;<source>Morph: development and optimization of a longitudinal age progression database</source>,&#x201d; in <source>European workshop on biometrics and identity management</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Springer</publisher-name>, <fpage>17</fpage>&#x2013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.16451/j.cnki.issn</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ryu</surname>
<given-names>H. J.</given-names>
</name>
<name>
<surname>Adam</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Mitchell</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Inclusivefacenet: Improving face attribute detection with race and gender diversity</source>. <comment>arXiv preprint arXiv:1712.00193</comment>.</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Singh</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Agarwal</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Nagpal</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Vatsa</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>On the robustness of face recognition algorithms against attacks and bias</article-title>,&#x201d; in <conf-name>AAAI conference on artificial intelligence</conf-name>. <pub-id pub-id-type="doi">10.1609/aaai.v34i09.7085</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Torralba</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Efros</surname>
<given-names>A. A.</given-names>
</name>
</person-group> (<year>2011</year>). &#x201c;<article-title>Unbiased look at dataset bias</article-title>,&#x201d; in <conf-name>IEEE conference on computer vision and pattern recognition</conf-name>, <fpage>1521</fpage>&#x2013;<lpage>1528</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-642-33718-5_12</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Deng</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Mitigate bias in face recognition using skewness-aware reinforcement learning</source>. <comment>arXiv preprint arXiv:1911.10692</comment>.</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yatskar</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>K.-W.</given-names>
</name>
<name>
<surname>Ordonez</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Balanced datasets are not enough: estimating and mitigating gender bias in deep image representations</article-title>,&#x201d; in <conf-name>IEEE international conference on computer vision</conf-name>, <fpage>5310</fpage>&#x2013;<lpage>5319</lpage>. <pub-id pub-id-type="doi">10.1007/2Fs42413-020-00085-4</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Age progression/regression by conditional adversarial autoencoder</article-title>,&#x201d; in <conf-name>IEEE conference on computer vision and pattern recognition</conf-name>, <fpage>5810</fpage>&#x2013;<lpage>5818</lpage>. <pub-id pub-id-type="doi">10.1007&#x002f;s11277-020-07473-1</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>
