<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Med.</journal-id>
<journal-title>Frontiers in Medicine</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Med.</abbrev-journal-title>
<issn pub-type="epub">2296-858X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmed.2024.1443151</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Medicine</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>DeepCGAN: early Alzheimer&#x00027;s detection with deep convolutional generative adversarial networks</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Ali</surname> <given-names>Imad</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2751664/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Saleem</surname> <given-names>Nasir</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Alhussein</surname> <given-names>Musaed</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zohra</surname> <given-names>Benazeer</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Aurangzeb</surname> <given-names>Khursheed</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1322190/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Haq</surname> <given-names>Qazi Mazhar ul</given-names></name>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Computer Science, University of Swat</institution>, <addr-line>Swat, KP</addr-line>, <country>Pakistan</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Electrical Engineering, Faculty of Engineering &#x00026; Technology (FET), Gomal University</institution>, <addr-line>Dera Ismail Khan</addr-line>, <country>Pakistan</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Computer Engineering, College of Computer and Information Sciences, King Saud University</institution>, <addr-line>Riyadh</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Anatomy, School of Medical Sciences and Research, Sharda University</institution>, <addr-line>Greater Noida, UP</addr-line>, <country>India</country></aff>
<aff id="aff5"><sup>5</sup><institution>Department of Anatomy, Noida International Institute of Medical Sciences, Noida International University</institution>, <addr-line>Greater Noida, UP</addr-line>, <country>India</country></aff>
<aff id="aff6"><sup>6</sup><institution>Department of International Bachelor Program in Informatics and Computer Science and Engineering, Yuan Ze University</institution>, <addr-line>Taoyuan City</addr-line>, <country>Taiwan</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Amin Ul Haq, University of Electronic Science and Technology of China, China</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Tassadaq Hussain, King&#x00027;s College London, United Kingdom</p>
<p>Jan Kubicek, VSB-Technical University of Ostrava, Czechia</p>
<p>Fatma Elsarkawy, Suez Canal University, Egypt</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Imad Ali <email>imad.ali&#x00040;uswat.edu.pk</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>29</day>
<month>08</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>11</volume>
<elocation-id>1443151</elocation-id>
<history>
<date date-type="received">
<day>03</day>
<month>06</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>06</day>
<month>08</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2024 Ali, Saleem, Alhussein, Zohra, Aurangzeb and Haq.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Ali, Saleem, Alhussein, Zohra, Aurangzeb and Haq</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Alzheimer&#x00027;s disease (AD) is a neurodegenerative disorder and the most prevailing cause of dementia. AD critically disturbs the daily routine, which usually needs to be detected at its early stage. Unfortunately, AD detection using magnetic resonance imaging is challenging because of the subtle physiological variations between normal and AD patients visible on magnetic resonance imaging.</p></sec>
<sec>
<title>Methods</title>
<p>To cope with this challenge, we propose a deep convolutional generative adversarial network (DeepCGAN) for detecting early-stage AD in this article. The DeepCGAN is an unsupervised generative model that expands the dataset size in addition to its diversity by utilizing the generative adversarial network (GAN). The Generator of GAN follows the encoder-decoder framework and takes cognitive data as inputs, whereas the Discriminator follows a structure similar to the Generator&#x00027;s encoder. The last dense layer uses a softmax classifier to detect the labels indicating the AD.</p></sec>
<sec>
<title>Results</title>
<p>The proposed model attains an accuracy rate of 97.32%, significantly surpassing recent state-of-the-art models&#x00027; performance, including Adaptive Voting, ResNet, AlexNet, GoogleNet, Deep Neural Networks, and Support Vector Machines.</p></sec>
<sec>
<title>Discussion</title>
<p>The DeepCGAN significantly improves early AD detection accuracy and robustness by enhancing the dataset diversity and leveraging advanced GAN techniques, leading to better generalization and higher performance in comparison to traditional and contemporary methods. These results demonstrate the ecacy of DeepCGAN in enhancing early AD detection, thereby potentially improving patient outcomes through timely intervention.</p></sec></abstract>
<kwd-group>
<kwd>GAN</kwd>
<kwd>CNN</kwd>
<kwd>Alzheimer&#x00027;s disease</kwd>
<kwd>deep learning</kwd>
<kwd>cognitive features</kwd>
</kwd-group>
<contract-sponsor id="cn001">King Saud University<named-content content-type="fundref-id">10.13039/501100002383</named-content></contract-sponsor>
<counts>
<fig-count count="6"/>
<table-count count="3"/>
<equation-count count="11"/>
<ref-count count="35"/>
<page-count count="11"/>
<word-count count="7357"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Precision Medicine</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1 Introduction</title>
<p>Alzheimer&#x00027;s disease (AD) is a neurodegenerative condition primarily affecting the elderly, characterized by memory, behavioral, and cognitive impairments that disrupt daily life (<xref ref-type="bibr" rid="B1">1</xref>). This devastating disease is projected to have a staggering impact on global health in the coming decades. Epidemiological studies indicate a disturbing trend, with expectations of a fourfold surge in the worldwide prevalence of AD by 2050, potentially exceeding 100 million cases (<xref ref-type="bibr" rid="B2">2</xref>). The impending prevalence of AD raises critical concerns for individuals, families, and healthcare systems worldwide. The burden of AD extends beyond the individual, affecting the very fabric of society. Some studies employ mathematical modeling to forecast the trends and growth of AD, considering factors such as increasing life expectancy, shifting mortality patterns, and the prevalence of cardiovascular diseases. Unfortunately, these projections collectively suggest a growing proportion of the population will be impacted by AD in the future (<xref ref-type="bibr" rid="B3">3</xref>).</p>
<p>Detecting AD in its early stages is of paramount importance for effective intervention and treatment. AD diagnosis is a complex endeavor, demanding the accurate identification of different dementia subtypes (<xref ref-type="bibr" rid="B4">4</xref>). While the challenges are substantial, recent research highlights the central role of AD in dementia cases, constituting approximately two-thirds of all diagnoses (<xref ref-type="bibr" rid="B5">5</xref>). One of the pressing issues in AD management is the lack of effective pharmacological treatments in clinical practice. This shortfall has prompted a paradigm shift in therapeutic strategies, emphasizing the early-stage detection of AD as a promising avenue for intervention (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B7">7</xref>). Identifying individuals in the early stages of cognitive decline or Mild Cognitive Impairment, whether stable or progressive, is pivotal for understanding high-risk populations and potentially delaying AD progression. The combination of the increasing prevalence of AD and the complexity of its diagnosis underscores the urgent need for advancements in early detection methods and comprehensive care strategies to address the growing global challenge of AD.</p>
<p>The AD research landscape has shifted significantly due to deep learning (DL) models, including stacked auto-encoders, recurrent neural networks, support vector machines, and convolutional neural networks (CNN). The bi-directional gated recurrent units (BiGRUs) layers consist of 2,048 units, with 1,024 units in each direction. The BiGRUs capture long-term temporal cues from the cognitive data, which is crucial for identifying patterns and changes indicative of early AD which have emerged as potent tools in this endeavor (<xref ref-type="bibr" rid="B8">8</xref>). However, limitations exist in feature quality, especially from image processing (<xref ref-type="bibr" rid="B9">9</xref>), requiring DL models adaptable to diverse data types. Simultaneously, generative adversarial networks (GANs), (which is a class of machine learning frameworks where two neural networks, a generator, and a discriminator, compete against each other to produce more accurate results) originally designed for images, have found their place in AD classification (<xref ref-type="bibr" rid="B9">9</xref>). DL models with GANs are proficient in classifying AD states and enhancing image-based AD tasks, like denoising images and precise brain segmentation (<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>). These advances drive understanding, detection, and treatment of AD, a pressing neurodegenerative disease. Although, these architectures have made sufficient advancement in AD detection; however, these existing AD detection models have primarily focused on neuroimaging data, resulting in the underutilization of critical cognitive features. Moreover, temporal information, which is highly relevant for understanding AD progression, has been largely neglected in the literature. Additionally, the well-known challenge of training instability in these models remains a significant concern.</p>
<p>Existing models often struggle with limited dataset sizes and lack diversity, leading to overfitting and poor generalization. Traditional GAN-based methods, primarily designed for image data, fail to leverage cognitive data crucial for early AD detection. This article introduces a groundbreaking method for the early detection of AD&#x02014;the deep convolutional generative adversarial network (DeepCGAN), which is an unsupervised generative model designed to leverage cognitive (clinical) data for AD detection. DeepCGAN addresses these issues by using a deep convolutional GAN framework to expand and diversify the dataset, generating high-quality synthetic data that improves detection accuracy and robustness. DeepCGANs generate high-quality synthetic medical images, crucial for augmenting limited datasets like MRI and PET scans and enhancing model generalization. They create diverse synthetic samples, augmenting training data in medical imaging where labeled data is scarce, improving model performance. DeepCGANs&#x00027; convolutional layers learn complex features for accurate early Alzheimer&#x00027;s detection, and their flexibility across imaging modalities makes them versatile beyond disease detection, which makes DeepCGANs a powerful and effective choice for early AD detection.</p>
<p>To address the aforementioned gaps, the proposed model effectively incorporates and analyzes cognitive data, offering a more comprehensive understanding of AD. Also, the proposed model integrates temporal information using BiGRUs to capture long-term patterns and introduces mechanisms like gradient penalty and relativistic average loss to stabilize training, thereby enhancing the stability and reliability of AD detection with GANs. Operating through a dual structure, the Generator follows an encoder-decoder framework that takes cognitive data as input, while the Discriminator mirrors the architecture of the Generator&#x00027;s encoder. Moreover, the proposed model employs two distinct loss functions, Wasserstein and Relativistic loss, ensuring stable training and improved performance. The pivotal component of the model is the last dense layer, employing a softmax classifier to detect AD labels. The proposed DeepCGAN undergoes comprehensive training using cognitive data, demonstrating promising results in the early prediction of AD, achieving a remarkable 97.32% accuracy on cognitively labeled data from the ADNI dataset, surpassing recent state-of-the-art models. The contributions of this article include:</p>
<list list-type="order">
<list-item><p>For detecting early-stage AD, this article proposes a DeepCGAN, an unsupervised generative model that extends the cognitive features of the data and its diversity by utilizing the GAN framework.</p></list-item>
<list-item><p>To optimize the detection performance of DeepCGAN, a novel convolutional encoder-decoder-based GAN is proposed and trained on the cognitive features.</p></list-item>
<list-item><p>Our comprehensive experiments on the ADNI dataset show that the proposed DeepCGAN performs better in detecting early-stage AD compared to start-of-the-art models.</p></list-item>
</list>
<p>The remainder of this article is structured as follows: Section 2 reviews related work. Section 3 presents the proposed DeepCGAN. Section 4 describes the experimental setup, and Section 5 discusses the evaluation of the proposed model. Finally, Section 6 concludes the article.</p></sec>
<sec id="s2">
<title>2 Related work</title>
<p>The current gold standard for detecting and prognosing neurodegenerative AD relies on clinical assessments of symptoms and their severity. However, early disease detection before clinical symptoms manifest is critical for effective disease management and timely therapeutic intervention. Research indicates that analyzing structural and functional changes in patients during the early stages of AD can provide valuable insights (<xref ref-type="bibr" rid="B12">12</xref>). Machine learning approaches offer a rapid and robust means to interpret medical examinations, aiding in the early detection of AD. Early detection is paramount, allowing for proactive intervention and potentially improving patient outcomes. Machine learning enhances the diagnostic process by uncovering subtle patterns and anomalies that may precede clinical symptoms. It transcends the limitations of conventional clinical assessments, which often rely on symptomatic markers that become evident at later disease stages. Integrating machine learning into AD detection represents a paradigm shift, emphasizing the significance of early and accurate diagnosis in transforming AD research and treatment strategies.</p>
<p>CNNs are deep learning models (<xref ref-type="bibr" rid="B13">13</xref>) known for their ability to extract complex patterns (<xref ref-type="bibr" rid="B14">14</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>). They excel in body part segmentation, surpassing traditional methods like logistic regression and support vector machines (<xref ref-type="bibr" rid="B17">17</xref>). CNN-based computer-aided diagnosis (CAD) systems are effective in neurodegenerative disease detection (<xref ref-type="bibr" rid="B18">18</xref>). In AD detection, methods combining the dual-tree complex wavelet transform with neural networks show promise (<xref ref-type="bibr" rid="B19">19</xref>). Architectures like GoogleNet and ResNet deliver strong results in distinguishing healthy subjects from those with AD and mild cognitive impairment (<xref ref-type="bibr" rid="B20">20</xref>). LeNet-5 has been effectively employed for AD vs. normal control (NC) brain classification (<xref ref-type="bibr" rid="B21">21</xref>). Hosseini et al. extended previous work by proposing a Deeply Supervised Adaptive 3D-CNN (DSA-3D-CNN) for AD prediction (<xref ref-type="bibr" rid="B22">22</xref>). They trained this model on the CAD-Dementia dataset without skull stripping preprocessing and rigorously evaluated its performance through 10-fold cross-validation. In addition to CNNs, ensemble learning (EL) has proven invaluable in the detection and prognosis of neurodegenerative diseases. Given the often limited availability and the inherent 3D nature of medical imaging data, training classifiers can be a challenge (<xref ref-type="bibr" rid="B23">23</xref>). EL offers a promising solution by combining the strengths of multiple trained models, making it particularly useful for classification tasks involving heterogeneous datasets. To harness the power of ensemble learning, individual classifiers are trained on various subsets of the data and subsequently combined. EL with bootstrapping techniques becomes especially beneficial when relevant data is scarce, such as cognitive features. Additionally, when dealing with limited data, common practices include data augmentation to enhance the performance of ensemble models. This combined approach of CNNs and ensemble learning offers a robust and adaptable framework for tackling the complexities of neurodegenerative disease detection and prognosis.</p>
<p>GANs are a prominent method for enhancing imaging data by creating synthetic data that competes with a discriminator aiming to distinguish real from synthetic data (<xref ref-type="bibr" rid="B24">24</xref>). When generative networks excel, they can replicate data based on the inherent structure of real data. In the field of medical imaging, GANs have found success in tasks like MRI and CT reconstruction and unconditional image synthesis (<xref ref-type="bibr" rid="B25">25</xref>, <xref ref-type="bibr" rid="B26">26</xref>). Furthermore, GANs exhibit a wide array of applications in AD-related image processing. They are proficient in denoising low-dose positron emission tomography (PET) scans to yield high-quality images (<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B27">27</xref>). Accurate brain image segmentation, facilitated by GANs, aids in locating features critical for AD diagnosis and research across various image modalities (<xref ref-type="bibr" rid="B28">28</xref>&#x02013;<xref ref-type="bibr" rid="B30">30</xref>). Despite the promise of GANs in AD image processing, the existing models for detecting AD have predominantly centered around neuroimaging data, leading to the insufficient utilization of vital cognitive features. Furthermore, the valuable temporal dimension, crucial for comprehending the progression of AD, has been notably overlooked in the existing literature. Additionally, the persisting issue of training instability in these models continues to pose a noteworthy challenge.</p></sec>
<sec sec-type="materials and methods" id="s3">
<title>3 Materials and methods</title>
<p>The DeepCGAN model, proposed in this study, is designed for AD detection. It leverages a Generative Adversarial Network architecture, specifically tailored to the analysis of cognitive features and temporal information, which are often overlooked in existing AD detection models.</p>
<sec>
<title>3.1 Generative adversarial networks</title>
<p>GAN is a fundamental architecture in machine learning, composed of two primary components: the Generator <italic>G</italic>(<italic>z</italic>) and the Discriminator <italic>D</italic>(<italic>x</italic>), as shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. The GAN framework is designed for generative tasks, aiming to produce synthetic data that closely resembles real data distributions. The Generator <italic>G</italic>(<italic>z</italic>) is responsible for creating new data samples. It takes random noise <italic>N</italic>(<italic>z</italic>) as input, typically drawn from a uniform or normal distribution. Through a learned transformation process, the Generator generates data that mimics real training data. This process relies on adjusting internal parameters to produce data samples that are increasingly realistic.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>GAN framework with generator and discriminator.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-11-1443151-g0001.tif"/>
</fig>
<p>The Discriminator <italic>D</italic>(<italic>x</italic>) acts as an adversary to the Generator. Its primary role is to differentiate between genuine data from the training set and data generated by the Generator. The Discriminator evaluates each input and assigns a probability score, indicating the likelihood of the input being real. If an input is genuine, <italic>D</italic>(<italic>x</italic>) approaches 1, whereas if it is generated, <italic>D</italic>(<italic>x</italic>) tends toward 0. The GAN operates as a two-player minimax game, optimizing the value function <italic>V</italic>(<italic>G, D</italic>). The objective function is given in <xref ref-type="disp-formula" rid="E1">Equation 1</xref>:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo class="qopname">min</mml:mo></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo class="qopname">max</mml:mo></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>V</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">data</mml:mtext></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x02003;&#x02003;&#x02003;&#x02003;&#x02003;&#x02003;</mml:mo><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>z</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>D</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Here, <italic>D</italic>(&#x000B7;) provides the probability that a given sample belongs to the training data <italic>X</italic>. The Generator aims to minimize log(1&#x02212;<italic>D</italic>(<italic>G</italic>(<italic>z</italic>))), making <italic>D</italic>(<italic>G</italic>(<italic>z</italic>)) as high as possible, essentially fooling the Discriminator into considering <italic>G</italic>(<italic>z</italic>) as real data. Conversely, the Discriminator seeks to maximize <italic>D</italic>(<italic>X</italic>) and 1&#x02212;<italic>D</italic>(<italic>G</italic>(<italic>z</italic>)), driving its optimal state toward <italic>P</italic>(<italic>x</italic>) &#x0003D; 0.5. In practice, GANs continually refine the Generator to produce data that is indistinguishable from real data, representing a powerful framework for generating synthetic data in various domains.</p>
</sec>
<sec>
<title>3.2 DeepCGAN for AD detection</title>
<p>The architecture of our proposed GAN model for AD detection is illustrated in <xref ref-type="fig" rid="F2">Figure 2</xref>. This model is carefully designed to effectively utilize cognitive features in the detection process. The Generator component of our model is based on an encoder-decoder framework, optimized for processing cognitive features as inputs. The encoder in our model is designed to extract meaningful features from the input data through a series of convolutional layers. The encoder comprises five 2-D convolutional layers, strategically placed to extract local correlations within the input features. A reshape layer is employed to appropriately format the encoded features. These layers progressively downsample the input, capturing local correlations and essential patterns. Each convolutional layer is followed by batch normalization and Leaky Rectified Linear Unit (ReLU) activation functions to stabilize training and introduce non-linearity. Positioned in the middle of the Generator architecture, the BiGRU layers are crucial for capturing long-term dependencies and temporal dynamics in the cognitive features. Each BiGRU layer consists of 2,048 units (1,024 in each direction), enabling the model to learn bidirectional temporal patterns that are significant for early Alzheimer&#x00027;s detection.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Proposed CNN-based GAN model (DeepCGAN).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-11-1443151-g0002.tif"/>
</fig>
<p>The decoder mirrors the encoder&#x00027;s structure but performs the inverse operation. It utilizes deconvolutional (transposed convolution) layers to reconstruct the input data from the encoded features. The skip connections between corresponding layers of the encoder and decoder facilitate fine-grained feature integration, enhancing the model&#x00027;s ability to preserve important information during reconstruction. The Generator&#x00027;s primary function is to produce synthetic data that closely resembles the real cognitive feature data. By transforming random noise inputs through the encoder-BiGRU-decoder pipeline, the Generator learns to create realistic data samples that help augment the training set and improve the robustness of the Discriminator. The input to our DeepCGAN model consists of cognitive features derived from the ADNI dataset. The input to the model is a three-dimensional tensor with a batch size of 32, 50-time steps, and 128 features. Thus, the input shape is [32, 50, 128], specifically tailored to capture the temporal and cognitive aspects critical for Alzheimer&#x00027;s detection.</p>
<p>The data preprocessing steps include normalization and sequence padding to ensure uniform input dimensions. The preprocessing steps include: Normalization: The cognitive features are normalized to ensure consistent scales and improve model training stability. Padding: Sequences are padded to a fixed length (e.g., 50 time steps) to ensure uniform input dimensions across different samples. Handling Missing Values: Features with more than 40% missing values are removed. For the remaining features, missing values are imputed using appropriate statistical methods (e.g., mean imputation). In addition, two BiGRUs layers are thoughtfully inserted in the middle of the Generator architecture, which enhances the model&#x00027;s ability to capture long-term temporal cues from the cognitive data. This integration addresses a critical gap in existing models that primarily focus on neuroimaging data, thereby improving the detection of early AD. The decoder of our model mirrors the encoder&#x00027;s structure and consists of five 2-D deconvolutional layers, also known as transposed convolution layers. Batch normalization is consistently applied following each convolutional and deconvolutional operation. ReLU functions are used as activation functions within the hidden layers, while a sigmoid activation function is applied to the output layer. To facilitate fine-grained feature integration, we have incorporated skip connections within the Generator. These skip connections concatenate the outputs of each convolutional layer in the encoder with the corresponding inputs of the deconvolutional layers in the decoder. This design element enhances the model&#x00027;s ability to capture intricate feature cues effectively.</p>
<p>The Discriminator component, denoted as <italic>D</italic>, shares a similar structure with the encoder of the Generator. However, a flattened layer is introduced after the fifth convolutional layer to streamline feature processing. Finally, a fully connected dense layer with softmax activation is integrated into the Discriminator to enable classification tasks. Notably, the Discriminator provides two types of outputs, <italic>D</italic>(<italic>y</italic>) and <italic>Dk</italic>(<italic>y</italic>), with D(y) representing sigmoidal output and Dk(y) signifying linear output, linked by the sigmoid non-linearity function &#x003BB;(<italic>Dk</italic>(<italic>y</italic>)) &#x0003D; <italic>D</italic>(<italic>y</italic>). The proposed GAN model for AD detection leverages cognitive features and exhibits a sophisticated architecture, comprising convolutional, deconvolutional, and recurrent layers, skip connections, and a dual-output Discriminator. These design innovations collectively contribute to the model&#x00027;s efficacy in AD detection. Most AD detection models predominantly focus on neuroimaging data, neglecting cognitive features. Our model efficiently incorporates and exploits these underutilized data sources. By including BiGRUs, our model accounts for long-term temporal cues, a crucial aspect often overlooked in AD progression analysis. The Discriminator&#x00027;s architecture, featuring dual output types and skip connections, introduces novel enhancements to improve the model&#x00027;s performance in distinguishing real and synthetic data.</p>
<p>The Discriminator is designed to differentiate between real and synthetic data. It shares a similar structure with the encoder and includes an additional fully connected layer with softmax activation for classification. The dual outputs of the Discriminator, D(y) and Dk(y), provide sigmoidal and linear outputs, respectively, enhancing the model&#x00027;s ability to distinguish between genuine and generated data. The DeeCGAN model is specifically tailored for Alzheimer&#x00027;s detection by focusing on cognitive features and temporal information, which are often underutilized in traditional models. By leveraging the strengths of DeeCGANs in generating realistic synthetic data and incorporating bidirectional GRUs for temporal analysis, our model is able to achieve high accuracy in early Alzheimer&#x00027;s detection.</p>
</sec>
<sec>
<title>3.3 Loss function</title>
<p>In the realm of GANs, choosing appropriate loss functions plays a pivotal role in achieving stable training and optimal performance. Our proposed GAN model incorporates and thoroughly investigates two distinct loss functions to determine the one that yields superior results.</p>
<sec>
<title>3.3.1 Wasserstein loss</title>
<p>The Wasserstein loss, denoted as <italic>L</italic><sub><italic>D</italic></sub> for the Discriminator and <italic>L</italic><sub><italic>G</italic></sub> for the Generator, offers significant advantages in stabilizing and enhancing the robustness of GAN models (<xref ref-type="bibr" rid="B18">18</xref>). The Wasserstein loss function is used to train the DeepCGAN model due to its ability to provide a smoother gradient, leading to more stable training. This stability is crucial for AD detection, as it ensures that the model effectively learns from the subtle and complex patterns in the cognitive data. These loss functions are defined in <xref ref-type="disp-formula" rid="E2">Equations 2</xref>&#x02013;<xref ref-type="disp-formula" rid="E4">4</xref>:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>y</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E3"><label>(3)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E4"><label>(4)</label><mml:math id="M5"><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>G</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mstyle mathvariant='double-struck'><mml:mtext>E</mml:mtext></mml:mstyle><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover><mml:mo>~</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mo>&#x02016;</mml:mo><mml:mrow><mml:msub><mml:mo>&#x02207;</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x000AF;</mml:mo></mml:mover></mml:msub><mml:msub><mml:mi>D</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mo>&#x02016;</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M6"><mml:msub><mml:mrow><mml:mo>&#x02207;</mml:mo></mml:mrow><mml:mrow><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x01EF9;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> represents the gradient of the Discriminator output with respect to <inline-formula><mml:math id="M7"><mml:mover accent="false" class="mml-overline"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo accent="true">&#x000AF;</mml:mo></mml:mover></mml:math></inline-formula>.</p></sec>
<sec>
<title>3.3.2 Relativistic loss</title>
<p>The second loss function incorporated into our GAN model is the Relativistic loss. It computes the probability of real data features being classified as real and the probability of synthetic data features being classified as real. This is achieved by considering the difference between the Discriminator&#x00027;s outputs for real and synthetic input features. The loss functions for the Discriminator and Generator are given by <xref ref-type="disp-formula" rid="E5">Equations 5</xref>, and <xref ref-type="disp-formula" rid="E6">6</xref>, respectively.</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003C5;</mml:mi><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(6)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0007E;</mml:mo><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>&#x003BD;</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>However, the the Relativistic loss in <xref ref-type="disp-formula" rid="E5">Equations 5</xref> and <xref ref-type="disp-formula" rid="E6">6</xref> exhibits high variance, primarily when the Generator significantly influences the Discriminator. To address this, we consider the average loss functions for the Discriminator and Generator are given by <xref ref-type="disp-formula" rid="E7">Equations 7</xref>, and <xref ref-type="disp-formula" rid="E8">8</xref>, respectively.</p>
<disp-formula id="E7"><label>(7)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>y</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x00233;</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo class="qopname">&#x00304;</mml:mo></mml:mover></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E8"><label>(8)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo class="qopname">&#x00304;</mml:mo></mml:mover></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>&#x01D53C;</mml:mtext></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x00233;</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>D</italic><sub>&#x00233;</sub>(<italic>y</italic>) and <inline-formula><mml:math id="M12"><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> represent the relativistic Discriminator outputs for real and synthetic data, respectively. Thus, our GAN model incorporates both Wasserstein and Relativistic loss functions, each with its distinct advantages. These loss functions are carefully chosen and utilized to optimize the model&#x00027;s training stability and performance in AD detection.</p>
<p>We selected the Wasserstein loss and Relativistic loss due to their efficacy in stabilizing GAN training and enhancing the quality of generated data. The Wasserstein loss addresses mode collapse and provides meaningful gradients for GAN convergence. The Relativistic loss improves the model&#x00027;s discriminative power by comparing real and generated data in a relativistic manner, aligning with the goal of distinguishing subtle differences in medical data. These loss functions balance sensitivity and specificity in Alzheimer&#x00027;s detection, with the Wasserstein loss ensuring high-quality synthetic data and the Relativistic loss enhancing the discriminator&#x00027;s accuracy and robustness.</p></sec></sec>
</sec>
<sec id="s4">
<title>4 Experiments</title>
<p>This section provides insights into the dataset, experimental settings, and an evaluation of the proposed model.</p>
<sec>
<title>4.1 Dataset</title>
<p>We utilized the Alzheimer&#x00027;s Disease Neuroimaging Initiative (ADNI) dataset (<xref ref-type="bibr" rid="B20">20</xref>), consisting of three distinct stages. The ADNI dataset encompasses cognitive test scores and records of 5,013 instances, corresponding to 819 different AD patients. The cognitive features selected for this study include memory recall tests, attention assessments, and executive function evaluations, which are clinically relevant as they have been shown to be significant indicators of early cognitive decline associated with AD. Patients visited the clinic multiple times during clinical trials, resulting in new cognitive test scores generated and stored as additional records in the dataset for each visit. Among these records, there are 1,643 belonging to cognitively normal individuals and 3,370 related to AD patients. However, the dataset exhibited missing values and underwent initialization through an Iterative Imputer technique to impute the missing values using a round-robin method. This ensures that the most clinically significant features are retained and accurately represented in the dataset. The irrelevant features were removed during the data cleaning and preprocessing.</p>
<p>In the ADNI1 dataset, each record comprises 113 features. The data includes various cognitive assessments (e.g., MMSE scores, ADAS-Cog scores), biomarkers (e.g., cerebrospinal fluid biomarkers, amyloid-beta levels), and potentially neuroimaging features (e.g., MRI and PET scan data). These features are chosen for their relevance in assessing cognitive decline and AD progression. The input data is organized as a temporal sequence, capturing changes in cognitive features over time. This is crucial for modeling the progression of AD, which involves gradual cognitive decline. The dataset was divided into 80% for training and 20% for testing, resulting in 5,000 samples for training and 1,250 samples for testing. Some of these features had excessive missing values, prompting the removal of those with more than 40% missing values. The remaining features underwent initialization through an Iterative Imputer technique to impute the missing values using a round-robin method. After preprocessing, the final dataset comprised 4,500 samples. Additionally, the dataset contained features with varying value ranges, which were normalized to a range of 0&#x02013;1 using the min-max scaling method. Primary filtering of cognitive features was performed using Pearson&#x00027;s correlation coefficient to identify those most correlated with AD diagnosis. Features with a correlation coefficient above a predefined threshold were selected for further analysis. Performance evaluation utilized metrics including Accuracy, Sensitivity, and F1-Score, which are computed by <xref ref-type="disp-formula" rid="E9">Equations 9</xref>, <xref ref-type="disp-formula" rid="E10">10</xref>, and <xref ref-type="disp-formula" rid="E11">11</xref>, respectively.</p>
<disp-formula id="E9"><label>(9)</label><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>A</mml:mi><mml:mi>c</mml:mi><mml:mi>c</mml:mi><mml:mi>u</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E10"><label>(10)</label><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E11"><label>(11)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>S</mml:mi><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Here, TP represents True Positives, TN stands for True Negatives, FP is False Positives, and FN represents False Negatives.</p>
</sec>
<sec>
<title>4.2 Network settings</title>
<p>The DeepCGAN model architecture for AD detection incorporates carefully chosen parameters to optimize performance across multiple metrics. The feature maps in the Generator&#x00027;s encoder are structured with fixed sizes of 16, 32, 64, 128, and 256 in successive convolutional layers, with specific kernel sizes and strides tailored to enhance feature extraction efficiency. The kernel size is set to (1,3) for the first 2D-Conv layer and (2,3) for subsequent layers, all with a stride of (1,2). Convolutional layers are utilized for their strength in extracting local patterns and hierarchical features from the cognitive data, which are essential for distinguishing subtle differences between normal and AD-affected individuals.Similarly, the BiGRU layers are configured with 2,048 units, with 1,024 units in each direction, split into forward and backward directions, operating over a fixed time step of 50. The BiGRUs were selected for their ability to capture long-term dependencies and temporal patterns in cognitive data, which are crucial for accurately modeling the progression of AD over time. For the Generator&#x00027;s decoder, these parameters are inversely set to reconstruct the input features faithfully. Moreover, the Discriminator (<italic>D</italic>) employs deconvolutional layers with gradually increasing feature maps from 4 to 64, designed to discriminate between real and generated samples effectively. The proposed AD detection models, incorporating these two distinct losses, undergo training and optimization using the Adam optimizer for 1,000 epochs, with a learning rate of 0.005 and a batch size of 32 samples. The combination of convolutional layers and BiGRUs in the DeepCGAN architecture leverages both spatial and temporal features, providing a robust framework for early AD detection by capturing complex patterns in cognitive data that simpler architectures might miss. This setup ensures robust optimization and convergence of the DeepCGAN model. To assess the performance of our proposed model, we conducted a comprehensive comparison with several other models, including DeciTree, RanForest, KNN, Linear Regression (LR), SVM, DNN, AdaBoost, and Adaptive Voting, utilizing various metrics such as Accuracy, Precision, Recall, and F1-Score.</p></sec>
</sec>
<sec id="s5">
<title>5 Results and analysis</title>
<p>In this section, we present the results of our experiments and provide a comprehensive analysis of the findings.</p>
<sec>
<title>5.1 Model performance comparison</title>
<p><xref ref-type="table" rid="T1">Table 1</xref> displays the results obtained from our proposed DeepCGAN model along with other DL models trained on similar cognitive features for detecting cognitive normal and AD. We measure the model&#x00027;s performance using Accuracy, Precision, Recall, and F1-Score as evaluation metrics. Notably, the results demonstrate that our proposed DeepCGAN outperforms all other competing models in terms of these metrics. The DeepCGAN achieved an Accuracy of 97.32%, Precision of 95.31%, Recall of 95.43%, and F1-Score of 95.61%, respectively. In contrast, the lowest-performing model, linear regression, achieved only 82.05% Accuracy, 81.83% Precision, and 81.45% F1-Score.</p>
 <table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Performance analysis using various measures for ADNI cognitive features dataset.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>DL model</bold></th>
<th valign="top" align="center"><bold>Acc. (%)</bold></th>
<th valign="top" align="center"><bold>Pre. (%)</bold></th>
<th valign="top" align="center"><bold>Rec. (%)</bold></th>
<th valign="top" align="center"><bold>F1-S. (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">DeciTree</td>
<td valign="top" align="center">88.93</td>
<td valign="top" align="center">88.93</td>
<td valign="top" align="center">88.93</td>
<td valign="top" align="center">88.93</td>
</tr> <tr>
<td valign="top" align="left">RanForest</td>
<td valign="top" align="center">90.33</td>
<td valign="top" align="center">90.30</td>
<td valign="top" align="center">90.33</td>
<td valign="top" align="center">90.31</td>
</tr> <tr>
<td valign="top" align="left">KNN</td>
<td valign="top" align="center">85.14</td>
<td valign="top" align="center">85.40</td>
<td valign="top" align="center">85.14</td>
<td valign="top" align="center">85.24</td>
</tr> <tr>
<td valign="top" align="left">LR</td>
<td valign="top" align="center">82.05</td>
<td valign="top" align="center">81.83</td>
<td valign="top" align="center">82.05</td>
<td valign="top" align="center">81.45</td>
</tr> <tr>
<td valign="top" align="left">SVM</td>
<td valign="top" align="center">85.84</td>
<td valign="top" align="center">85.68</td>
<td valign="top" align="center">85.84</td>
<td valign="top" align="center">85.71</td>
</tr> <tr>
<td valign="top" align="left">DNN</td>
<td valign="top" align="center">90.53</td>
<td valign="top" align="center">90.67</td>
<td valign="top" align="center">90.53</td>
<td valign="top" align="center">90.59</td>
</tr> <tr>
<td valign="top" align="left">AdaBoost</td>
<td valign="top" align="center">86.64</td>
<td valign="top" align="center">86.61</td>
<td valign="top" align="center">86.64</td>
<td valign="top" align="center">86.31</td>
</tr> <tr>
<td valign="top" align="left">AdapVoting</td>
<td valign="top" align="center">93.92</td>
<td valign="top" align="center">93.89</td>
<td valign="top" align="center">93.92</td>
<td valign="top" align="center">93.89</td>
</tr> <tr>
<td valign="top" align="left">DeepCGAN</td>
<td valign="top" align="center">97.32</td>
<td valign="top" align="center">95.31</td>
<td valign="top" align="center">95.43</td>
<td valign="top" align="center">95.61</td>
</tr></tbody>
</table>
</table-wrap>
<p>To highlight the improvements made by our proposed model, we chose linear regression as a reference model. DeepCGAN substantially improved Accuracy by 15.27%, Precision by 13.48%, and F1-Score by 14.16% compared to linear regression. Moreover, when compared to the second-best model, Adaptive Voting, DeepCGAN showed a 3.40% improvement in Accuracy. It also outperformed DNN and Random Forest by 6.79 and 6.99% in Accuracy, respectively, which is a significant performance gain. Furthermore, our DeepCGAN model demonstrated substantial improvements in Recall and F1-Score compared to competing models. The Recall increased from 88.93% (DeciTree) to 95.43% with DeepCGAN, and the F1-Score increased from 86.31% (AdaBoost) to 95.61%. These results signify the superior ability of DeepCGAN to correctly identify AD cases while minimizing false negatives. The errors are vastly improved over other models. To highlight the effectiveness of the proposed model, we present the overall improvements depicted in <xref ref-type="fig" rid="F3">Figure 3</xref>. The linear regression is the reference model that has achieved the lowest performance among DL models. <xref ref-type="fig" rid="F3">Figure 3</xref> indicates the best performance of the proposed DeepCGAN.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Percentage improvements in accuracy, precision, recall, and F1-score with linear regression as reference lowest model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-11-1443151-g0003.tif"/>
</fig>
</sec>
<sec>
<title>5.2 Loss function analysis</title>
<p>Our DeepCGAN model was trained and optimized using two different loss functions: Wasserstein and Relativistic loss. The Wasserstein loss function was chosen for its ability to provide a smoother gradient, thereby stabilizing the training process. This stability is particularly important for AD detection, where the model must accurately learn from subtle variations in cognitive data. <xref ref-type="fig" rid="F4">Figure 4</xref> presents the confusion matrices for both losses, revealing that the Wasserstein loss function results in better performance. <xref ref-type="fig" rid="F4">Figure 4A</xref> illustrates the predicted labels when trained with Wasserstein loss, while <xref ref-type="fig" rid="F4">Figure 4B</xref> shows the outcomes with Relativistic loss. It is evident that the model trained with Wasserstein loss provides more accurate predictions.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Confusion matrix heat-map for DeepCGAN. <bold>(A)</bold> Wasserstein loss. <bold>(B)</bold> Relativistic loss.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-11-1443151-g0004.tif"/>
</fig>
<p>To further assess the performance, we compared the model&#x00027;s Accuracy on training and validation data. The Wasserstein loss outperformed the relativistic loss by a significant margin, indicating faster convergence and better Accuracy. <xref ref-type="fig" rid="F5">Figure 5</xref> displays the loss curves over 1,000 epochs, illustrating the superior performance of the DeepCGAN model in achieving its detection task. We also evaluated the Area Under the Curve (AUC), which measures the model&#x00027;s ability to differentiate between labels. DeepCGAN exhibited higher AUC values compared to models trained with relativistic loss, further confirming its superior discriminatory capability. <xref ref-type="fig" rid="F6">Figure 6</xref> illustrates the AUC curves for both loss functions.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Loss curves for DeepCGAN. <bold>(A)</bold> Wasserstein loss. <bold>(B)</bold> Relativistic loss.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-11-1443151-g0005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>AUC curves for DeepCGAN.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-11-1443151-g0006.tif"/>
</fig>
</sec>
<sec>
<title>5.3 Comparison against state-of-the-art models</title>
<p>In this subsection, we compare our proposed DeepCGAN model with state-of-the-art (SOTA) models in the literature, including AlexNet (<xref ref-type="bibr" rid="B31">31</xref>), VGG-16 (<xref ref-type="bibr" rid="B32">32</xref>), GoogleNet (<xref ref-type="bibr" rid="B33">33</xref>), and ResNet (<xref ref-type="bibr" rid="B34">34</xref>), using cognitive features from the ADNI dataset. This comparison aims to benchmark the performance of DeepCGAN under similar experimental settings and datasets. <xref ref-type="table" rid="T2">Table 2</xref> presents the results in terms of Accuracy, Precision, Recall, F1-scores, and AUC.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Performance analysis (in %) for SOTA using ADNI Cognitive features dataset.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
<th valign="top" align="center"><bold>F1-score</bold></th>
<th valign="top" align="center"><bold>AUC</bold></th>
</tr>
</thead>
<tbody>
 <tr>
<td valign="top" align="left">AlexNet</td>
<td valign="top" align="center">93.75</td>
<td valign="top" align="center">94.98</td>
<td valign="top" align="center">92.28</td>
<td valign="top" align="center">93.61</td>
<td valign="top" align="center">93.68</td>
</tr> <tr>
<td valign="top" align="left">VGG-16</td>
<td valign="top" align="center">94.96</td>
<td valign="top" align="center">94.02</td>
<td valign="top" align="center">95.43</td>
<td valign="top" align="center">94.97</td>
<td valign="top" align="center">94.96</td>
</tr> <tr>
<td valign="top" align="left">GoogleNet</td>
<td valign="top" align="center">91.73</td>
<td valign="top" align="center">90.20</td>
<td valign="top" align="center">93.50</td>
<td valign="top" align="center">91.82</td>
<td valign="top" align="center">91.79</td>
</tr> <tr>
<td valign="top" align="left">ResNet</td>
<td valign="top" align="center">94.96</td>
<td valign="top" align="center">93.00</td>
<td valign="top" align="center">97.15</td>
<td valign="top" align="center">95.03</td>
<td valign="top" align="center">94.98</td>
</tr> <tr>
<td valign="top" align="left">DeepCGAN</td>
<td valign="top" align="center">97.32</td>
<td valign="top" align="center">95.31</td>
<td valign="top" align="center">95.43</td>
<td valign="top" align="center">95.61</td>
<td valign="top" align="center">99.51</td>
</tr></tbody>
</table>
</table-wrap>
<p>DeepCGAN surpassed all SOTA models in terms of Accuracy, achieving an Accuracy of 97.32%, which is a 5.59% improvement over GoogleNet. Similarly, Precision improved from 90.20% (GoogleNet) to 95.31%, reflecting a 5.11% boost in Precision. When compared to AlexNet, DeepCGAN demonstrated a 3.57% increase in Accuracy. Furthermore, DeepCGAN achieved the highest AUC among all models, with a 99.51% AUC, outperforming ResNet by 4.53% and VGG-16 by 4.55%, highlighting its superior discriminatory power. Regarding Recall, DeepCGAN exhibited substantial improvements over SOTA models except for ResNet, where the results were marginally lower. Specifically, the Recall increased from 92.28% (AlexNet) to 95.43% with DeepCGAN. The F1-Score achieved with GoogleNet was 91.82%. DeepCGAN&#x00027;s superior performance can be attributed to its novel architecture, which combines convolutional layers for effective feature extraction and BiGRUs for capturing temporal dependencies. This dual approach enables the model to detect subtle changes and patterns in cognitive data more accurately than models that rely solely on neuroimaging data or simpler architectures. Additionally, the use of GANs for data augmentation increases the dataset&#x00027;s size and diversity, enhancing the model&#x00027;s generalizability and robustness. The core innovation lies in expanding the cognitive features dataset and enhancing its diversity through GANs.</p>
<p>DeepCGANs significantly enhance Alzheimer&#x00027;s detection due to their ability to generate realistic synthetic images, crucial for augmenting limited MRI and PET scan datasets. Their deep convolutional layers extract complex features, improving diagnostic accuracy by capturing subtle disease indicators. Adversarial training refines synthetic images iteratively, ensuring they closely resemble real patient data. DCGANs&#x00027; adaptability across imaging modalities and superior performance in comparative evaluations underline their transformative role in improving diagnostic accuracy and clinical outcomes for AD.</p>
</sec>
<sec>
<title>5.4 Comparison with existing techniques</title>
<p>In this section, we compare our proposed DeepCGAN model with a recently reported technique by Gill et al. (<xref ref-type="bibr" rid="B35">35</xref>) that used cognitive features for AD detection. Both studies utilized the same ADNI dataset, and the results are presented in <xref ref-type="table" rid="T3">Table 3</xref>. Our proposed DeepCGAN model outperformed the model proposed by Gill et al. (<xref ref-type="bibr" rid="B35">35</xref>) and Adaptive Voting using cognitive features from the ADNI dataset. DeepCGAN achieved the highest Accuracy of 97.32%, representing a 15.52% improvement over Adaptive Voting and a 3.4% improvement over Gill et al.&#x00027;s technique for early AD detection. This improvement is due to its ability to generate synthetic data that closely resembles the actual cognitive features, thus reducing overfitting and improving the model&#x00027;s ability to generalize to new, unseen data.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Performance comparison of the proposed model on ADNI dataset cognitive features.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Records</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
<th valign="top" align="center"><bold>AUC</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Gill et al. (<xref ref-type="bibr" rid="B35">35</xref>)</td>
<td valign="top" align="center">600</td>
<td valign="top" align="center">81.80%</td>
<td valign="top" align="center">79.0%</td>
</tr> <tr>
<td valign="top" align="left">AdaptiveVoting</td>
<td valign="top" align="center">5013</td>
<td valign="top" align="center">93.92%</td>
<td valign="top" align="center">99.3%</td>
</tr> <tr>
<td valign="top" align="left">DeepCGAN</td>
<td valign="top" align="center">5013</td>
<td valign="top" align="center">97.32%</td>
<td valign="top" align="center">99.5%</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="conclusions" id="s6">
<title>6 Conclusion</title>
<p>In this study, we propose a novel convolutional encoder-decoder-based GAN for early AD detection using cognitive features. This model leverages a Generator module with Conv2D and Deconv2D layers in an encoder-decoder architecture to optimize Accuracy, Precision, Recall, F1-Score, and AUC metrics. Our experimental results demonstrate the superior performance of DeepCGAN, which significantly advances early AD detection, and outperforms several state-of-the-art models and benchmarks across various measures, achieving an outstanding 97.32% Accuracy compared to most other DL models in this study&#x00027;s SOTA comparison. Moreover, We find that using the Wasserstein loss is superior for training the proposed GAN. While our GAN excels, it is important to acknowledge the potential of SOTA DL models for early AD detection, which offer advantages over non-DL techniques like Gill&#x00027;s study. These DL models can expedite diagnosis, making them valuable tools in the detection of neurodegenerative diseases like Alzheimer&#x00027;s. The unique contribution of DeepCGAN lies in its novel use of GANs to enhance the dataset&#x00027;s size and diversity, coupled with a sophisticated architecture that integrates convolutional layers and BiGRUs. This approach significantly improves accuracy, precision, and overall performance metrics in detecting AD at early stages, demonstrating the model&#x00027;s superior capability in distinguishing subtle cognitive changes indicative of early AD. In the future, we aim to develop even more robust and streamlined DL models for detecting early and various stages of AD. Our proposed DeepCGAN model significantly advances early AD detection by leveraging a convolutional encoder-decoder-based GAN with Wasserstein loss, achieving superior performance metrics compared to SOTA models such as AlexNet, VGG-16, GoogleNet, and ResNet. This novel approach enhances the diversity and richness of cognitive features, resulting in a remarkable improvement in accuracy, precision, and discriminatory power, as demonstrated through comprehensive comparisons with existing techniques and models.</p></sec>
</body>
<back>
<sec sec-type="data-availability" id="s7">
<title>Data availability statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://adni.loni.usc.edu/data-samples/adni-data/">https://adni.loni.usc.edu/data-samples/adni-data/</ext-link>.</p>
</sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>IA: Writing &#x02013; review &#x00026; editing, Writing &#x02013; original draft, Visualization, Validation, Methodology, Conceptualization. NS: Writing &#x02013; review &#x00026; editing, Resources, Project administration, Methodology, Investigation. MA: Writing &#x02013; review &#x00026; editing, Validation, Supervision, Software, Funding acquisition. BZ: Writing &#x02013; review &#x00026; editing, Visualization, Validation, Software, Methodology, Investigation. KA: Writing &#x02013; review &#x00026; editing, Supervision, Project administration, Funding acquisition, Formal analysis, Data curation, Conceptualization. QH: Writing &#x02013; review &#x00026; editing, Visualization, Validation, Software, Methodology, Formal analysis.</p>
</sec>
<sec sec-type="funding-information" id="s9">
<title>Funding</title>
<p>The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This Research was funded by Researchers Supporting Project Number (RSPD2024R553), King Saud University, Riyadh, Saudi Arabia.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Pan</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>M</given-names></name> <name><surname>Lian</surname> <given-names>C</given-names></name> <name><surname>Zhou</surname> <given-names>T</given-names></name> <name><surname>Xia</surname> <given-names>Y</given-names></name> <name><surname>Shen</surname> <given-names>D</given-names></name></person-group>. <article-title>Synthesizing missing PET from MRI with cycle-consistent generative adversarial networks for Alzheimer&#x00027;s disease diagnosis</article-title>. In: <source>Medical Image Computing and Computer Assisted Intervention-MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part III 11</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <fpage>455</fpage>&#x02013;<lpage>63</lpage>.<pub-id pub-id-type="pmid">34355223</pub-id></citation></ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Yan</surname> <given-names>Y</given-names></name> <name><surname>Lee</surname> <given-names>H</given-names></name> <name><surname>Somer</surname> <given-names>E</given-names></name> <name><surname>Grau</surname> <given-names>V</given-names></name></person-group>. <article-title>Generation of amyloid PET images via conditional adversarial training for predicting progression to Alzheimer&#x00027;s disease</article-title>. In: <source>Predictive Intelligence in Medicine: First International Workshop, PRIME 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Proceedings 1</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <fpage>26</fpage>&#x02013;<lpage>33</lpage>.</citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Zhou</surname> <given-names>L</given-names></name> <name><surname>Yu</surname> <given-names>B</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Zu</surname> <given-names>C</given-names></name> <name><surname>Lalush</surname> <given-names>DS</given-names></name> <etal/></person-group>. <article-title>3D auto-context-based locality adaptive multi-modality GANs for PET synthesis</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2019</year>) <volume>38</volume>:<fpage>1328</fpage>&#x02013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2018.2884053</pub-id><pub-id pub-id-type="pmid">30507527</pub-id></citation></ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Yu</surname> <given-names>B</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Zu</surname> <given-names>C</given-names></name> <name><surname>Lalush</surname> <given-names>DS</given-names></name> <name><surname>Lin</surname> <given-names>W</given-names></name> <etal/></person-group>. <article-title>3D conditional generative adversarial networks for high-quality PET image estimation at low dose</article-title>. <source>Neuroimage</source>. (<year>2018</year>) <volume>174</volume>:<fpage>550</fpage>&#x02013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2018.03.045</pub-id><pub-id pub-id-type="pmid">29571715</pub-id></citation></ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ouyang</surname> <given-names>J</given-names></name> <name><surname>Chen</surname> <given-names>KT</given-names></name> <name><surname>Gong</surname> <given-names>E</given-names></name> <name><surname>Pauly</surname> <given-names>J</given-names></name> <name><surname>Zaharchuk</surname> <given-names>G</given-names></name></person-group>. <article-title>Ultra-low-dose PET reconstruction using a generative adversarial network with feature matching and task-specific perceptual loss</article-title>. <source>Med Phys</source>. (<year>2019</year>) <volume>46</volume>:<fpage>3555</fpage>&#x02013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1117/12.2512946</pub-id></citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Islam</surname> <given-names>J</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name></person-group>. <article-title>GAN-based synthetic brain PET image generation</article-title>. <source>Brain Inf</source> . (<year>2020</year>) <volume>7</volume>:<fpage>1</fpage>&#x02013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1186/s40708-020-00104-2</pub-id><pub-id pub-id-type="pmid">32232602</pub-id></citation></ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>HW</given-names></name> <name><surname>Lee</surname> <given-names>HE</given-names></name> <name><surname>Lee</surname> <given-names>S</given-names></name> <name><surname>Oh</surname> <given-names>KT</given-names></name> <name><surname>Yun</surname> <given-names>M</given-names></name> <name><surname>Yoo</surname> <given-names>SK</given-names></name></person-group>. <article-title>Slice-selective learning for Alzheimer&#x00027;s disease classification using a generative adversarial network: a feasibility study of external validation</article-title>. <source>Eur J Nucl Med Mol Imaging</source>. (<year>2020</year>) <volume>47</volume>:<fpage>2197</fpage>&#x02013;<lpage>206</lpage>. <pub-id pub-id-type="doi">10.1007/s00259-019-04676-y</pub-id><pub-id pub-id-type="pmid">31980910</pub-id></citation></ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kang</surname> <given-names>H</given-names></name> <name><surname>Park</surname> <given-names>JS</given-names></name> <name><surname>Cho</surname> <given-names>K</given-names></name> <name><surname>Kang</surname> <given-names>D-Y</given-names></name></person-group>. <article-title>Visual and quantitative evaluation of amyloid brain PET image synthesis with the generative adversarial network</article-title>. <source>Appl Sci</source>. (<year>2020</year>) <volume>10</volume>:<fpage>2628</fpage>. <pub-id pub-id-type="doi">10.3390/app10072628</pub-id></citation>
</ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kang</surname> <given-names>SK</given-names></name> <name><surname>Seo</surname> <given-names>S</given-names></name> <name><surname>Shin</surname> <given-names>SA</given-names></name> <name><surname>Byun</surname> <given-names>MS</given-names></name> <name><surname>Lee</surname> <given-names>DY</given-names></name> <name><surname>Kim</surname> <given-names>YK</given-names></name> <etal/></person-group>. <article-title>Adaptive template generation for amyloid PET using a deep learning approach</article-title>. <source>Hum Brain Mapp</source>. (<year>2018</year>) <volume>39</volume>:<fpage>3769</fpage>&#x02013;<lpage>78</lpage>. <pub-id pub-id-type="doi">10.1002/hbm.24210</pub-id><pub-id pub-id-type="pmid">29752765</pub-id></citation></ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jha</surname> <given-names>D</given-names></name> <name><surname>Kim</surname> <given-names>JI</given-names></name> <name><surname>Kwon</surname> <given-names>GR</given-names></name></person-group>. <article-title>Diagnosis of Alzheimer&#x00027;s disease using dual-tree complex wavelet transform, PCA, and feedforward neural network</article-title>. <source>J Healthc Eng</source>. (<year>2017</year>) <volume>2017</volume>:<fpage>9060124</fpage>. <pub-id pub-id-type="doi">10.1155/2017/9060124</pub-id><pub-id pub-id-type="pmid">29065663</pub-id></citation></ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sarraf</surname> <given-names>S</given-names></name> <name><surname>Tofighi</surname> <given-names>G</given-names></name></person-group>. <article-title>Classification of Alzheimer&#x00027;s disease structural MRI data by deep learning convolutional neural networks</article-title>. <source>arXiv</source> [preprint] (<year>2016</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1607.06583</pub-id></citation>
</ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaiswal</surname> <given-names>S</given-names></name> <name><surname>Gupta</surname> <given-names>P</given-names></name> <name><surname>Prasad</surname> <given-names>LN</given-names></name> <name><surname>Kulkarni</surname> <given-names>R</given-names></name></person-group>. <article-title>An empirical model for the classification of diabetes and diabetes_types using ensemble approaches</article-title>. <source>J Artif Intell Technol</source>. (<year>2023</year>) <volume>3</volume>:<fpage>181</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.37965/jait.2023.0220</pub-id></citation>
</ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ganaie</surname> <given-names>MA</given-names></name> <name><surname>Hu</surname> <given-names>M</given-names></name> <name><surname>Tanveer</surname> <given-names>M</given-names></name> <name><surname>Suganthan</surname> <given-names>PN</given-names></name></person-group>. <article-title>Ensemble deep learning: a review</article-title>. <source>arXiv</source> [preprint] (<year>2021</year>). <pub-id pub-id-type="doi">10.1016/j.engappai.2022.105151</pub-id></citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>X</given-names></name> <name><surname>Xu</surname> <given-names>K</given-names></name> <name><surname>Hall</surname> <given-names>P</given-names></name></person-group>. <article-title>A survey of image synthesis and editing with generative adversarial networks</article-title>. <source>Tsinghua Sci Technol</source>. (<year>2017</year>) <volume>22</volume>:<fpage>660</fpage>&#x02013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.23919/TST.2017.8195348</pub-id></citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wolterink</surname> <given-names>JM</given-names></name> <name><surname>Dinkla</surname> <given-names>AM</given-names></name> <name><surname>Savenije</surname> <given-names>MHF</given-names></name> <name><surname>Seevinck</surname> <given-names>PR</given-names></name> <name><surname>van den Berg</surname> <given-names>CAT</given-names></name> <name><surname>I&#x00161;gum</surname> <given-names>I</given-names></name></person-group>. <article-title>Deep MR to CT synthesis using unpaired data</article-title>. In:<person-group person-group-type="editor"><name><surname>Gooya</surname> <given-names>A</given-names></name> <name><surname>Frangi</surname> <given-names>AF</given-names></name> <name><surname>Tsaftaris</surname> <given-names>SA</given-names></name> <name><surname>Prince</surname> <given-names>JL</given-names></name></person-group>, editors. <source>Simulation and Synthesis in Medical Imaging</source>. Vol. 10557. Cham: Springer (<year>2017</year>). p. <fpage>14</fpage>&#x02013;<lpage>23</lpage>.<pub-id pub-id-type="pmid">31121961</pub-id></citation></ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yi</surname> <given-names>X</given-names></name> <name><surname>Walia</surname> <given-names>E</given-names></name> <name><surname>Babyn</surname> <given-names>P</given-names></name></person-group>. <article-title>Generative adversarial network in medical imaging: a review</article-title>. <source>Med Image Anal</source>. (<year>2019</year>) <volume>58</volume>:<fpage>101552</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2019.101552</pub-id><pub-id pub-id-type="pmid">31521965</pub-id></citation></ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Bjorck</surname> <given-names>N</given-names></name> <name><surname>Gomes</surname> <given-names>CP</given-names></name> <name><surname>Selman</surname> <given-names>B</given-names></name> <name><surname>Weinberger</surname> <given-names>KQ</given-names></name></person-group>. <article-title>Understanding batch normalization</article-title>. In: <source>Advances in Neural Information Processing Systems</source>. <publisher-loc>Vol. 31. Red Hook, NY</publisher-loc>: <publisher-name>Curran Associates, Inc</publisher-name> (<year>2018</year>).</citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Dukler</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>W</given-names></name> <name><surname>Lin</surname> <given-names>A</given-names></name> <name><surname>Mont&#x000FA;far</surname> <given-names>G</given-names></name></person-group>. <article-title>Wasserstein of Wasserstein loss for learning generative models</article-title>. In: <source>International Conference on Machine Learning</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>PMLR</publisher-name> (<year>2019</year>). p. <fpage>1716</fpage>&#x02013;<lpage>25</lpage>.</citation>
</ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jolicoeur-Martineau</surname> <given-names>A</given-names></name></person-group>. <article-title>The relativistic discriminator: a key element missing from standard GAN</article-title>. <source>arXiv</source> [preprint] (<year>2018</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1807.00734</pub-id></citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>de Vent</surname> <given-names>NR</given-names></name> <name><surname>Agelink van Rentergem</surname> <given-names>JA</given-names></name> <name><surname>Schmand</surname> <given-names>BA</given-names></name> <name><surname>Murre</surname> <given-names>JM</given-names></name> <name><surname>Huizenga</surname> <given-names>HM</given-names></name> <name><surname>Consortium</surname> <given-names>A</given-names></name></person-group>. <article-title>Advanced Neuropsychological Diagnostics Infrastructure (ANDI): a normative database created from control datasets</article-title>. <source>Front Psychol</source>. (<year>2016</year>) <volume>7</volume>:<fpage>1601</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2016.01601</pub-id><pub-id pub-id-type="pmid">27812340</pub-id></citation></ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Weiner</surname> <given-names>MW</given-names></name> <name><surname>Veitch</surname> <given-names>DP</given-names></name> <name><surname>Miller</surname> <given-names>MJ</given-names></name> <name><surname>Aisen</surname> <given-names>PS</given-names></name> <name><surname>Albala</surname> <given-names>B</given-names></name> <name><surname>Beckett</surname> <given-names>LA</given-names></name> <etal/></person-group>. <article-title>Increasing participant diversity in AD research: plans for digital screening, blood testing, and a community-engaged approach in the Alzheimer&#x00027;s Disease Neuroimaging Initiative 4</article-title>. <source>Alzheimers Dement</source>. (<year>2023</year>) <volume>19</volume>:<fpage>307</fpage>&#x02013;<lpage>17</lpage>. <pub-id pub-id-type="doi">10.1002/alz.12797</pub-id><pub-id pub-id-type="pmid">36209495</pub-id></citation></ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hosseini-Asl</surname> <given-names>E</given-names></name> <name><surname>Ghazal</surname> <given-names>M</given-names></name> <name><surname>Mahmoud</surname> <given-names>A</given-names></name> <name><surname>Aslantas</surname> <given-names>A</given-names></name> <name><surname>Shalaby</surname> <given-names>AM</given-names></name> <name><surname>Casanova</surname> <given-names>MF</given-names></name> <etal/></person-group>. <article-title>Alzheimer&#x00027;s disease diagnostics by a 3D deeply supervised adaptable convolutional network</article-title>. <source>Front Biosci</source>. (<year>2018</year>) <volume>23</volume>:<fpage>584</fpage>&#x02013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.2741/4606</pub-id><pub-id pub-id-type="pmid">28930562</pub-id></citation></ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kingma</surname> <given-names>DP</given-names></name> <name><surname>Ba</surname> <given-names>J</given-names></name></person-group>. <article-title>Adam: a method for stochastic optimization</article-title>. <source>arXiv</source> [preprint] (<year>2014</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1412.6980</pub-id></citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hosseini-asl</surname> <given-names>E</given-names></name> <name><surname>Keynton</surname> <given-names>R</given-names></name> <name><surname>El-baz</surname> <given-names>A</given-names></name></person-group>. <article-title>Alzheimer&#x00027;s disease diagnostics by adaptation of 3D convolutional network</article-title>. In: <source>Proc. - Int Conf Image Process ICIP.</source> Piscataway, NJ (<year>2016</year>). 502 p.</citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhan</surname> <given-names>B</given-names></name> <name><surname>Li</surname> <given-names>D</given-names></name> <name><surname>Wu</surname> <given-names>X</given-names></name> <name><surname>Zhou</surname> <given-names>J</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name></person-group>. <article-title>Multi-modal MRI image synthesis via GAN with multi-scale gate mergence</article-title>. <source>IEEE J Biomed Health Inf</source> . (<year>2021</year>) <volume>26</volume>:<fpage>17</fpage>&#x02013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1109/JBHI.2021.3088866</pub-id><pub-id pub-id-type="pmid">34125692</pub-id></citation></ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Suganthi</surname> <given-names>K</given-names></name></person-group>. <article-title>Review of medical image synthesis using GAN techniques</article-title>. In: <source>ITM Web of Conferences</source>. Vol. 37. Les Ulis: EDP Sciences (<year>2021</year>). p. 01005.</citation>
</ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>J</given-names></name> <name><surname>Shi</surname> <given-names>M</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Yuan</surname> <given-names>Y</given-names></name> <name><surname>Yang</surname> <given-names>C</given-names></name> <etal/></person-group>. <article-title>Incremental data stream classification with adaptive multi-task multi-view learning</article-title>. <source>Big Data Mining Anal</source>. (<year>2023</year>) <volume>7</volume>:<fpage>87</fpage>&#x02013;<lpage>106</lpage>. <pub-id pub-id-type="doi">10.26599/BDMA.2023.9020006</pub-id></citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>KV</surname> <given-names>S</given-names></name> <name><surname>Ravi</surname> <given-names>V</given-names></name> <name><surname>Almeshari</surname> <given-names>M</given-names></name> <name><surname>Alzamil</surname> <given-names>Y</given-names></name> <name><surname>DS</surname> <given-names>SK</given-names></name> <etal/></person-group>. <article-title>A new deep learning model based on neuroimaging for predicting Alzheimer&#x00027;s disease</article-title>. <source>Open Bioinf J</source>. (<year>2023</year>) <volume>16</volume>:<fpage>1</fpage>&#x02013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.2174/0118750362260635230922051326</pub-id></citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dong</surname> <given-names>X</given-names></name> <name><surname>Yu</surname> <given-names>Z</given-names></name> <name><surname>Cao</surname> <given-names>W</given-names></name> <name><surname>Shi</surname> <given-names>Y</given-names></name> <name><surname>Qianli</surname> <given-names>MA</given-names></name></person-group>. <article-title>A survey on ensemble learning</article-title>. <source>Front Comput Sci</source>. (<year>2020</year>) <volume>14</volume>:<fpage>241</fpage>&#x02013;<lpage>58</lpage>. <pub-id pub-id-type="doi">10.1007/s11704-019-8208-z</pub-id></citation>
</ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Prakash</surname> <given-names>D</given-names></name> <name><surname>Madusanka</surname> <given-names>N</given-names></name> <name><surname>Bhattacharjee</surname> <given-names>S</given-names></name> <name><surname>Park</surname> <given-names>HG</given-names></name> <name><surname>Kim</surname> <given-names>CH</given-names></name> <name><surname>Choi</surname> <given-names>HK</given-names></name> <etal/></person-group>. <article-title>comparative study of Alzheimer&#x00027;s disease classification using multiple transfer learning models</article-title>. <source>J Multim Inf Syst</source>. (<year>2019</year>) <volume>6</volume>:<fpage>209</fpage>&#x02013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.33851/JMIS.2019.6.4.209</pub-id></citation>
</ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krizhevsky</surname> <given-names>A</given-names></name> <name><surname>Sutskever</surname> <given-names>I</given-names></name> <name><surname>Hinton</surname> <given-names>GE</given-names></name></person-group>. <article-title>Imagenet classification with deep convolutional neural networks</article-title>. <source>Adv Neural Inf Process Syst</source>. (<year>2012</year>) <volume>25</volume>:<fpage>1</fpage>&#x02013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1145/3065386</pub-id></citation>
</ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname> <given-names>K</given-names></name> <name><surname>Zisserman</surname> <given-names>A</given-names></name></person-group>. <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <source>arXiv</source> [preprint] (<year>2014</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1409.1556</pub-id></citation>
</ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szegedy</surname> <given-names>C</given-names></name> <name><surname>Liu</surname> <given-names>W</given-names></name> <name><surname>Jia</surname> <given-names>Y</given-names></name> <name><surname>Sermanet</surname> <given-names>P</given-names></name> <name><surname>Reed</surname> <given-names>S</given-names></name> <name><surname>Anguelov</surname> <given-names>D</given-names></name> <etal/></person-group>. <article-title>Going deeper with convolutions</article-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.</source> Piscataway, NJ (<year>2015</year>). p. <fpage>1</fpage>&#x02013;<lpage>9</lpage>.</citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>K</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Ren</surname> <given-names>S</given-names></name> <name><surname>Sun</surname> <given-names>J</given-names></name></person-group>. <article-title>Deep residual learning for image recognition</article-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.</source> Piscataway, NJ (<year>2016</year>). p. <fpage>770</fpage>&#x02013;<lpage>8</lpage>.</citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gill</surname> <given-names>S</given-names></name> <name><surname>Mouches</surname> <given-names>P</given-names></name> <name><surname>Hu</surname> <given-names>S</given-names></name> <name><surname>Rajashekar</surname> <given-names>D</given-names></name> <name><surname>MacMaster</surname> <given-names>FP</given-names></name> <name><surname>Smith</surname> <given-names>EE</given-names></name> <etal/></person-group>. <article-title>Using machine learning to predict dementia from neuropsychiatric symptom and neuroimaging data</article-title>. <source>J Alzheimers Dis</source>. (<year>2020</year>) <volume>75</volume>:<fpage>277</fpage>&#x02013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.3233/JAD-191169</pub-id><pub-id pub-id-type="pmid">32250302</pub-id></citation></ref>
</ref-list>
</back>
</article>