<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurorobot.</journal-id>
<journal-title>Frontiers in Neurorobotics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurorobot.</abbrev-journal-title>
<issn pub-type="epub">1662-5218</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnbot.2023.1203962</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Palmprint recognition based on gating mechanism and adaptive feature fusion</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Zhang</surname> <given-names>Kaibi</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Xu</surname> <given-names>Guofeng</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Jin</surname> <given-names>Ye Kelly</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Qi</surname> <given-names>Guanqiu</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1664211/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Yang</surname> <given-names>Xun</given-names></name>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Bai</surname> <given-names>Litao</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/525320/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>School of Automation, Chongqing University of Posts and Telecommunications</institution>, <addr-line>Chongqing</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Integrated Chinese and Western Medicine, The Second Affiliated Hospital of Chongqing Medical University</institution>, <addr-line>Chongqing</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>College of Business and Economics, California State University</institution>, <addr-line>Los Angeles, CA</addr-line>, <country>United States</country></aff>
<aff id="aff4"><sup>4</sup><institution>Double Deuce Sports</institution>, <addr-line>Bowling Green, KY</addr-line>, <country>United States</country></aff>
<aff id="aff5"><sup>5</sup><institution>Computer Information Systems Department, State University of New York at Buffalo State</institution>, <addr-line>Buffalo, NY</addr-line>, <country>United States</country></aff>
<aff id="aff6"><sup>6</sup><institution>China Merchants Chongqing Communications Research and Design Institute Co., Ltd.</institution>, <addr-line>Chongqing</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Qian Jiang, Yunnan University, China</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Alwin Poulose, Indian Institute of Science Education and Research, Thiruvananthapuram, India; Shuaiqi Liu, Hebei University, China; Xinghua Feng, Southwest University of Science and Technology, China</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Litao Bai <email>blt&#x00040;cqmu.edu.cn</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>26</day>
<month>05</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>17</volume>
<elocation-id>1203962</elocation-id>
<history>
<date date-type="received">
<day>11</day>
<month>04</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>08</day>
<month>05</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2023 Zhang, Xu, Jin, Qi, Yang and Bai.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Zhang, Xu, Jin, Qi, Yang and Bai</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract>
<p>As a type of biometric recognition, palmprint recognition uses unique discriminative features on the palm of a person to identify his/her identity. It has attracted much attention because of its advantages of contactlessness, stability, and security. Recently, many palmprint recognition methods based on convolutional neural networks (CNN) have been proposed in academia. Convolutional neural networks are limited by the size of the convolutional kernel and lack the ability to extract global information of palmprints. This paper proposes a framework based on the integration of CNN and Transformer-GLGAnet for palmprint recognition, which can take advantage of CNN&#x00027;s local information extraction and Transformer&#x00027;s global modeling capabilities. A gating mechanism and an adaptive feature fusion module are also designed for palmprint feature extraction. The gating mechanism filters features by a feature selection algorithm and the adaptive feature fusion module fuses them with the features extracted by the backbone network. Through extensive experiments on two datasets, the experimental results show that the recognition accuracy is 98.5% for 12,000 palmprints in the Tongji University dataset and 99.5% for 600 palmprints in the Hong Kong Polytechnic University dataset. This demonstrates that the proposed method outperforms existing methods in the correctness of both palmprint recognition tasks. The source codes will be available on <ext-link ext-link-type="uri" xlink:href="https://github.com/Ywatery/GLnet.git">https://github.com/Ywatery/GLnet.git</ext-link>.</p></abstract>
<kwd-group>
<kwd>palmprint recognition</kwd>
<kwd>convolutional neural networks (CNN)</kwd>
<kwd>gate control mechanism</kwd>
<kwd>adaptive feature fusion</kwd>
<kwd>deep learning-based artificial neural networks</kwd>
</kwd-group>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content></contract-sponsor>
<contract-sponsor id="cn002">Natural Science Foundation of Chongqing<named-content content-type="fundref-id">10.13039/501100005230</named-content></contract-sponsor>
<counts>
<fig-count count="7"/>
<table-count count="3"/>
<equation-count count="5"/>
<ref-count count="51"/>
<page-count count="11"/>
<word-count count="6771"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>With the rapid development of information security, various identification technologies using human-specific biometric features have attracted widespread attention. Human-specific biometric features include face, gait, and fingerprints (Shi et al., <xref ref-type="bibr" rid="B32">2020</xref>; Yang, <xref ref-type="bibr" rid="B42">2020</xref>; Zhang et al., <xref ref-type="bibr" rid="B45">2021</xref>). As an emerging recognition technique of biometric features, palmprint recognition can analyze various recognizable features, including main lines, ridges, minutiae, and textures (Zhang et al., <xref ref-type="bibr" rid="B44">2018</xref>). Palmprint features are considered to be unique biometric information for each person, with a high degree of security and reliability. The research on palmprint recognition techniques is also receiving more and more attention from academia (Wang et al., <xref ref-type="bibr" rid="B37">2019</xref>; Zhong et al., <xref ref-type="bibr" rid="B47">2019</xref>; Fei et al., <xref ref-type="bibr" rid="B6">2020a</xref>). Traditional palmprint recognition methods are mainly developed based on both texture an structure features (Fei et al., <xref ref-type="bibr" rid="B5">2018</xref>). Many effective methods have been applied, such as apparent latent direction code (ALDC; Fei et al., <xref ref-type="bibr" rid="B9">2019</xref>) and joint multi-view feature learning (JMvFL; Fei et al., <xref ref-type="bibr" rid="B7">2020b</xref>), have been applied to palmprint recognition. Some researchers also proposed techniques such as 2D palmprint recognition (Korichi et al., <xref ref-type="bibr" rid="B22">2021</xref>) and 3D palmprint recognition (Yang et al., <xref ref-type="bibr" rid="B41">2020</xref>) for different palmprint morphologies. Recently, new breakthroughs in palmprint recognition performance have been achieved by using deep learning methods (Ungureanu et al., <xref ref-type="bibr" rid="B36">2020</xref>). Deep learning-based palmprint recognition methods can avoid information loss caused by traditional manual features and significantly improve recognition performance (Zhong et al., <xref ref-type="bibr" rid="B47">2019</xref>). For example, Liu et al. (<xref ref-type="bibr" rid="B29">2021</xref>) proposed a novel and effective end-to-end algorithm for palmprint recognition, known as the similarity metric hash network.</p>
<p>However, most of the current deep learning-based palmprint recognition methods have certain shortcomings. First, general image classification is applied to recognize palmprints (Zhong and Zhu, <xref ref-type="bibr" rid="B48">2019</xref>), which is not specially designed for palmprint features. In other words, the relationship between palmprint features involving fine details, texture and global mainline information is not considered. Some researchers designed specialized palmprint recognition neural networks based on the characteristics of palmprints (Genovese et al., <xref ref-type="bibr" rid="B11">2019</xref>). Most of these specially designed neural networks are developed based on CNN. Affected by the size of the convolution kernel, CNN can only extract local features of palmprints in a limited way, and cannot extract global feature information of palmprints. For palmprint recognition, it is necessary to study the neural network design of palmprint global feature information extraction. The Transformer, a framework based entirely on a self-attentive mechanism with powerful modeling capabilities for global information, has been widely used in image feature extraction (Zhu et al., <xref ref-type="bibr" rid="B51">2022</xref>, <xref ref-type="bibr" rid="B50">2023</xref>; He et al., <xref ref-type="bibr" rid="B13">2023</xref>).</p>
<p>This paper designs a network structure GLGAnet based on CNN and Transformer. CNN is used to extract local features such as texture and fine details of the palmprint. The Transformer module is used to establish the connection between main lines and ridges in the palmprint features. In the network structure, the features extracted by each component in the GLGAnet backbone network are controlled by the designed gating mechanism. The proposed network structure is able to extract global features such as main lines and ridges of the palmprint, local features such as palm texture and fine details. This paper has three main contributions as follows.</p>
<p>1. A new lightweight network-GLGANet is designed to process local features (e.g., texture) and global features (e.g., mainline) of palmprints. The proposed network extracts local features of palmprints by deep convolutional layers and global feature information of palmprints by using Transformer.</p>
<p>2. A gating mechanism is designed based on the different features extracted by the deep convolutional layer and the Transformer module. The proposed gating mechanism first selects the features of the palmprint by using a feature selection algorithm, and then multiplies them with the features extracted by the backbone network to achieve control over the features extracted by the backbone network.</p>
<p>3. An adaptive convolutional fusion module is designed. This module performs dimensionality reduction and fusion of multi-level features extracted by the gating mechanism through multi-level convolution and single convolution.</p>
<p>The rest of this paper is organized as follows. Section 2 briefly introduces the related topics. Section 3 elaborates on the deep learning network structure we proposed for palmprint recognition. Section 4 presents the experimental results. Section 5 offers the concluding remarks.</p></sec>
<sec id="s2">
<title>2. Related work</title>
<sec>
<title>2.1. Traditional palmprint recognition methods</title>
<p>After more than two decades of development, traditional palmprint recognition methods have been developed based on hand-crafted and machine learning methods. The most typical hand-crafted feature recognition methods include Competitive code (Kong and Zhang, <xref ref-type="bibr" rid="B21">2004</xref>), Palm code (Zhang et al., <xref ref-type="bibr" rid="B43">2003</xref>), Fusion code (Kong et al., <xref ref-type="bibr" rid="B20">2006</xref>), RLOC (Jia et al., <xref ref-type="bibr" rid="B15">2008</xref>), and BOCV (Guo et al., <xref ref-type="bibr" rid="B12">2009</xref>). These methods are characterized by fast recognition, and a distance measure is often used in the matching process, which may cause a decrease in the correct rate. For example, the competitive coding scheme (CompCode) proposed by Kong and Zhang (<xref ref-type="bibr" rid="B21">2004</xref>) uses several Gabor filters in different directions to encode the palmprint features and applies the Hamming distance to match the palmprint. Traditional machine learning recognition methods typically include PCA (Wang et al., <xref ref-type="bibr" rid="B38">2022</xref>), SVM (Yaddaden and Parent, <xref ref-type="bibr" rid="B39">2022</xref>), histogram (Jia et al., <xref ref-type="bibr" rid="B14">2013</xref>), etc. Compared with hand-crafted feature recognition-based methods, these methods are stabler, which extract more palmprint features with a subsequent increase in computational complexity. For example, Patil et al. (<xref ref-type="bibr" rid="B31">2015</xref>) extracted palmprint features by some discrete transforms, dimensionalized features using principal component analysis (PCA) and finally used the dimensionalized features for matching.</p>
</sec>
<sec>
<title>2.2. Deep learning based palmprint recognition methods</title>
<p>In recent years, CNN-based approaches have achieved remarkable performance on many computer vision tasks. Many deep learning-based recognition methods have also been proposed in the field of palmprint recognition. For example, Tarawneh et al. (<xref ref-type="bibr" rid="B35">2018</xref>) first used a pre-trained deep neural network VGG-19 (Simonyan and Zisserman, <xref ref-type="bibr" rid="B33">2014</xref>) to extract palmprint features and then used SVM to match and classify palmprints. Dian and Dongmei (<xref ref-type="bibr" rid="B4">2016</xref>) used AlexNet (Krizhevsky et al., <xref ref-type="bibr" rid="B24">2017</xref>) to extract palmprint and then combined hausdorff distance to match and recognize palmprint images. Yang et al. (<xref ref-type="bibr" rid="B40">2017</xref>) proposed a deep learning and local coding approach. Zhang et al. (<xref ref-type="bibr" rid="B44">2018</xref>) proposed an improved Inception-ResNet-v1 network structure and introduced a central loss function for feature extraction PalmRCNN. Matkowski et al. (<xref ref-type="bibr" rid="B30">2019</xref>) proposed an end-to-end palmprint recognition network (EE-PRnet). The network consists of two main networks, the ROI localization and alignment network (ROI-lanet) and the Feature Extraction and Recognition Network (FERnet). Most of these deep learning-based palmprint recognition methods are designed for image classification tasks. But there is no design that studies the characteristics of the main lines, fine details, and textures of palmprints. Other researchers have studied the characteristics of palmprints and have integrated deep learning into traditional hand-based design methods. As a typical approach, Genovese et al. (<xref ref-type="bibr" rid="B11">2019</xref>) proposed an unsupervised model-PalmNet-which trains palmprint-specific filters through an unsupervised process based on Gabor response and principal component analysis (PCA). The proposed method was experimented on several datasets and achieved remarkable results. Also, Liang et al. (<xref ref-type="bibr" rid="B28">2021</xref>) proposed a competitive convolutional neural network (CompNet) with constrained learnable Gabor filters, which extracts palmprint features for recognition by training learnable Gabor convolutional kernels. In addition, some other researchers have considered making neural networks for lightweight palmprint recognition after considering the extent to which palmprint functionality can be run on edge devices. For example, Jia et al. (<xref ref-type="bibr" rid="B17">2022</xref>) proposed a lightweight palmprint recognition neural network (EEPNet), which incorporates five strategies such as image stitching and image downscaling to improve palmprint recognition performance. These deep learning-based methods mentioned above are limited by the convolutional kernel perceptual field when extracting features and cannot effectively extract the global contextual information of palmprints, and can only achieve matching by extracting local features for recognition. The proposed GLGAnet integrating CNN and Transformer architecture can not only model the relationship between main lines and ridges of palmprints, but also effectively extract texture information of palmprints.</p></sec></sec>
<sec id="s3">
<title>3. The proposed method</title>
<p>In this section first introduces the general framework of the improved network architecture integrating CNN and Transformer. The functions of the individual modules are then described.</p>
<sec>
<title>3.1. Network structure</title>
<p>The proposed palmprint recognition framework is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, which mainly consists of a multilayer depth convolution module, a residual Vision Transformer module, a gating mechanism module, and an adaptive convolutional fusion module. The multi-layer depth convolution module is composed of depth-separable convolutional layers, which extract various local features of palmprints, such as ridge features and texture features, from the original palmprint image. The residual VIT module uses Transformer as the backbone structure to model the palmprint image globally and extract global contextual features of the palmprint, such as the orientation features of the three main lines in the palmprint. The gating mechanism module is dominated by the feature selection algorithm, which controls features extracted by the multilayer depth convolution module and the residual VIT module through the feature selection algorithm. The adaptive convolution fusion module performs feature fusion through the gate mechanism of multi-convolution and single convolution.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>The structure of the proposed network. The <italic>X</italic><sub><italic>i</italic></sub> expressed the output of the gating mechanism.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0001.tif"/>
</fig>
</sec>
<sec>
<title>3.2. Multilayer deep convolution block and res-VIT block</title>
<p>Multiple multi-layer depth convolution modules (MDCB) are used to extract local information from palmprints. The multi-layer depth convolution module consists of a depth-separable convolution layer, two normal convolution layers, a regularization layer and two activation layers. The depth-separable convolution learns the correlation of spatial features across channels of the palmprint image, such as local textures of the palmprint. Specifically, when the multilayer deep convolution module receives the input palmpArint image, it first expands the feature volume to multiple channels by ordinary convolution to obtain the output features. Further, the features obtained by ordinary convolution are output to a depth-separable convolution layer to obtain more local feature information of the palm print in a recursive manner across multiple depth-separable convolution layers. The feature volume is then compressed by ordinary convolution layers to the same channel as the input features. Finally, to avoid gradient disappearance during the computation, the feature are superimposed on the original input to obtain the depth convolution feature output <italic>F</italic><sub><italic>map</italic></sub>:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>X</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>D</mml:mi><mml:mi>W</mml:mi><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where the <italic>X</italic> is the input of the multi-layer depth convolution module.</p>
<p>The most important structure in Res-VIT is the Visual Transformer. The Transformer is developed entirely based on a self-attentive mechanism and has good global modeling capabilities. The input <italic>y</italic> is linearized as <italic>Q, K, andV</italic>, denote key, value, and dimension, respectively. The self-attentive mechanism can be obtained by Equation (2):</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;Atten</mml:mtext><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Q</mml:mi><mml:mo>,</mml:mo><mml:mi>K</mml:mi><mml:mo>,</mml:mo><mml:mi>V</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext>Soft</mml:mtext><mml:mo class="qopname">max</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>Q</mml:mi><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>V</mml:mi></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Since the Transformer structure is designed based on a fully self-attentive mechanism, it lacks the inductive bias in convolution. Therefore, in the Res-VIT module, the input <italic>Y</italic> is first expanded into <italic>M</italic> non-overlapping flat blocks <inline-formula><mml:math id="M3"><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, where <italic>P</italic> &#x0003D; <italic>wh</italic>, and <inline-formula><mml:math id="M4"><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>H</mml:mi><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi></mml:mrow></mml:mfrac></mml:math></inline-formula>, and then encoded between the blocks by applying the Transformer to obtain <inline-formula><mml:math id="M5"><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>M</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> as:</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext>&#x00A0;</mml:mtext><mml:mo>&#x000D7;</mml:mo><mml:mi>&#x003BB;</mml:mi><mml:mtext class="textrm" mathvariant="normal">Transformer</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x02264;</mml:mo><mml:mi>p</mml:mi><mml:mo>&#x02264;</mml:mo><mml:mi>P</mml:mi></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where &#x003BB; is the number of Transformer blocks.</p>
<p>After encoding, Res-VIT loses neither the order of the individual blocks nor the order of the pixel space of the palmprint image within each block. Thus, by a collapse operation, <italic>Y</italic><sub><italic>G</italic></sub> is transformed to <italic>Y</italic><sub><italic>U</italic></sub>, and <italic>Y</italic><sub><italic>U</italic></sub> is then projected into the lower dimensional space by point-by-point convolution to obtain the features.Finally, the input features of the original palmprint are spliced with this feature by means of the channel adaptive convolution module.</p>
</sec>
<sec>
<title>3.3. Gate control mechanism</title>
<p>The multi-layer depth convolution module extracts local feature information of the palmprint, and Res-VIT extracts global feature information of the palmprint. A gating mechanism is designed to process the extracted information by the convolution and feature selection algorithm. Its structure is schematically shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. Specifically, the palmprint image is first convolved to preliminary palmprint features, and then these features are processed by a feature selection algorithm to select more prominent features of the palmprint. These features are then multiplied with the features extracted from the multi-layer depth convolution and Res-VIT structures, allowing the features extracted from the multi-layer depth convolution and Res-VIT structures to be enhanced.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Diagram of the gating mechanism.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0002.tif"/>
</fig>
<p>Feature selection algorithm: First, each input palmprint image is convolved, and then the average value of the input features is calculated. Second, the features in each feature image pixel that are higher than the mean of the feature image pixel are retained and the features that are lower than the mean of the feature image pixel are set to zero to obtain a feature map vector. Finally, this feature map is multiplied with component <italic>i</italic> to control the features obtained from component <italic>i</italic> organization so that the features obtained from component <italic>i</italic> are more prominent. The convolution kernel is updated once for each execution of the feature selection algorithm. The algorithm flow is shown in <xref ref-type="table" rid="T4">Algorithm 1</xref>.</p>
<table-wrap position="float" id="T4">
<label>Algorithm 1</label>
<caption><p>Feature selection algorithm.</p></caption>
<table frame="box" rules="all">
<tbody>
<tr>
<td valign="top" align="left"><bold>Input:</bold>Original image <italic>Y</italic>, control matrix <italic>X</italic>, original image size <italic>N</italic>&#x0002A;<italic>N</italic>, cascade output <italic>X</italic></td>
</tr>
<tr>
<td valign="top" align="left"><bold>Output:</bold>Eigenvector G</td>
</tr>
<tr>
<td valign="top" align="left">1:Function:Feature selection(Y,K,N)</td>
</tr>
<tr>
<td valign="top" align="left">2:initial X=0 and N=0;</td>
</tr>
<tr>
<td valign="top" align="left">3:compute component feature <italic>C</italic><sub><italic>i</italic></sub> &#x0003D; <italic>conponent</italic><sub><italic>i</italic></sub>(<italic>Y</italic>)</td>
</tr>
<tr>
<td valign="top" align="left">4:compute Original feature mean <inline-formula><mml:math id="M7"><mml:mover accent="true"><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mo>*</mml:mo><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x02211;</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td valign="top" align="left">5:<bold>while</bold> <italic>i</italic>&#x0003C;<italic>N</italic>&#x0002A;<italic>N</italic> <bold>do</bold></td>
</tr>
<tr>
<td valign="top" align="left">6: <bold>for</bold> <italic>j</italic> &#x0003D; 1, <italic>j</italic>&#x0003C;<italic>N</italic>&#x0002A;<italic>N</italic>; <italic>j</italic>&#x0002B;&#x0002B; <bold>do</bold></td>
</tr>
<tr>
<td valign="top" align="left">7:<bold>if</bold> <inline-formula><mml:math id="M8"><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0003C;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo>&#x00304;</mml:mo></mml:mover></mml:math></inline-formula> <bold>then</bold></td>
</tr>
<tr>
<td valign="top" align="left">8: <italic>C</italic><sub><italic>i</italic></sub> &#x0003D; 0;</td>
</tr>
<tr>
<td valign="top" align="left">9: <bold>else</bold></td>
</tr>
<tr>
<td valign="top" align="left">10: <italic>C</italic><sub><italic>i</italic></sub> &#x0003D; <italic>C</italic><sub><italic>i</italic></sub></td>
</tr>
<tr>
<td valign="top" align="left">11: <bold>end if</bold></td>
</tr>
<tr>
<td valign="top" align="left">12: <italic>G</italic><sub><italic>i</italic></sub> &#x0003D; <italic>C</italic><sub><italic>i</italic></sub></td>
</tr>
<tr>
<td valign="top" align="left">13:Updated convolution kernel for (4);</td>
</tr>
<tr>
<td valign="top" align="left">14:<bold>end while</bold></td>
</tr>
<tr>
<td valign="top" align="left">15: return <italic>G</italic><sub><italic>i</italic></sub> </td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>3.4. Adaptive convolution fusion module</title>
<p>The feature maps extracted by the gating mechanism are too large to be effectively fused with the feature maps of the backbone network, and an adaptive convolutional fusion module is designed. A schematic diagram of the adaptive convolutional fusion module is shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. The adaptive convolution fusion module consists of a combination of multi-level convolution and single-level convolution. The fusion module efficiently fuses the feature maps of each level from the output of the gating mechanism with the feature maps extracted from the backbone network. The single-level convolution consists of a convolutional layer, a combined pooling layer and an activation layer. Specifically, in <xref ref-type="fig" rid="F2">Figure 2</xref>, the output of each layer is finally fed into the adaptive convolution module by the gating mechanism, and the first two layers of features can be resized by multi-layer convolution in the adaptive convolution module. Further extraction of local features is also performed. The last three layers have Res-VIT, which extracts the global contextual information of the palmprint, compresses the extracted information by single layer convolution in the adaptive convolution layer, and finally fuses the features of the five layers. The process can be represented by Equation (4).</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:mi>F</mml:mi><mml:mo>=</mml:mo><mml:mi>M</mml:mi><mml:mtext class="textrm" mathvariant="normal">ul</mml:mtext><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>p</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mtext>Sin</mml:mtext><mml:mi>g</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mo>,</mml:mo><mml:mn>5</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <italic>X</italic><sub><italic>i</italic></sub> denotes the features output by the gating mechanism and <italic>F</italic> denotes the output features of the adaptive convolution module.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Adaptive convolution module architecture diagram.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0003.tif"/>
</fig>
</sec>
<sec>
<title>3.5. Loss function</title>
<p>Palmprint recognition is essentially a classification problem. To enable the network to achieve good recognition performance, the proposed method is optimized during training in this paper using a cross-entropy loss function, which can be expressed by Equation (5).</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:mi>L</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:mi>s</mml:mi><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x00177;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Where <italic>M</italic> denotes the number of categories of palmprints; <italic>y</italic><sub><italic>ij</italic></sub>is a signed function (0 or 1) that takes 1 if the true category of palmprint sample <italic>i</italic> is equal to <italic>j</italic> and 0 otherwise; <italic>y</italic><sub><italic>ij</italic></sub> denotes the predicted probability that the observed palmprint sample <italic>i</italic> belongs to sample <italic>j</italic>.</p></sec></sec>
<sec id="s4">
<title>4. Experiments</title>
<sec>
<title>4.1. Experimental preparation</title>
<p>To verify the validity of the proposed method, experiments were conducted on the Tongji University palmprint dataset (Tongji), and the Hong Kong Polytechnic University palmprint dataset (Poly-U). The Tongji dataset collected contactless palmprint images from 300 volunteers, including 192 males and 108 females. All volunteers were from Tongji University in Shanghai, China. There were two collections in total. The average time interval between the first and second acquisition collections was &#x0007E;2 months. In each collection, 10 palmprint images were collected from each palm of each volunteer, and 6,000 palmprint images were collected from 600 palms. A total of 12,000 (300 &#x000D7; 2 &#x000D7; 10 &#x000D7; 2) high quality non-contact palmprint images were collected in the Tongji dataset. A schematic diagram of the Tongji palmprint dataset is shown in <xref ref-type="fig" rid="F4">Figure 4A</xref>. The Poly-U dataset collected contact palmprint images from 50 volunteers on two occasions. The Poly-U palmprint dataset is shown in <xref ref-type="fig" rid="F4">Figure 4B</xref>.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Schematic diagram of the two palmprint datasets. <bold>(A)</bold> Tongji palmprint dataset. <bold>(B)</bold> Poly-U palmprint dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0004.tif"/>
</fig>
<p>The proposed model framework was implemented in the Pytorch environment. The input Tongji dataset has an image size of 128*128 and a batch size of 8. The input Poly-U dataset has an image size of 128*128 and a batch size of 8. Two NVIDIA TITAN RTX graphics cards were used for training. In the experiment, the number of Transformer blocks &#x003BB; was set to 4. The optimizer used in this paper is Adam, with all parameters at default values. The initial learning rate and weight decay for model training were 1e-4 and 1e-5, respectively, and the experimental dataset was divided into a training and the testing set uses a 6:4 ratio. (7,200 sheets in the Tongji dataset are used as the training set and 4,800 sheets are used as the testing; 360 sheets in the Poly-U dataset are used as the training set and 240 sheets are used as the testing).</p>
</sec>
<sec>
<title>4.2. Comparative experiments</title>
<p>To validate the effectiveness of the proposed method in palmprint recognition, several state-of-the-art recognition methods that were tested on the Tongji palmprint dataset and the Poly-U palmprint dataset benchmarks are compared. This includes the proposed traditional descriptor-based palmprint recognition methods (Fei et al., <xref ref-type="bibr" rid="B8">2020c</xref>, <xref ref-type="bibr" rid="B10">2022</xref>; Jia et al., <xref ref-type="bibr" rid="B18">2020</xref>; Zhou et al., <xref ref-type="bibr" rid="B49">2020</xref>; Kusban, <xref ref-type="bibr" rid="B25">2021</xref>; Li et al., <xref ref-type="bibr" rid="B27">2022</xref>; Zhao et al., <xref ref-type="bibr" rid="B46">2022</xref>) and the deep learning-based palmprint recognition methods (Alrahawe et al., <xref ref-type="bibr" rid="B1">2021</xref>; Jia et al., <xref ref-type="bibr" rid="B16">2021</xref>; Jing et al., <xref ref-type="bibr" rid="B19">2021</xref>; Fei et al., <xref ref-type="bibr" rid="B10">2022</xref>). A brief description of these methods is given in <xref ref-type="table" rid="T1">Table 1</xref>. The source codes of many existing palmprint recognition methods are not publicly available, and to avoid biases introduced when models are retrained, publications are directly referred to obtain the results of the corresponding methods. This is a common approach in palmprint recognition research. The recognition accuracy results of the different methods for the Tongji University palmprint dataset and the Hong Kong Polytechnic University palmprint dataset are shown in <xref ref-type="fig" rid="F5">Figures 5</xref>, <xref ref-type="fig" rid="F6">6</xref>, respectively. The method with the highest recognition accuracy in each dataset is bolded in the corresponding bar.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Methods used for performance comparison in experiments.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:#8f9496">
<th valign="top" align="left"><bold>Methods</bold></th>
<th valign="top" align="left"><bold>Description</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">COSDISH (<xref ref-type="bibr" rid="B3">2020</xref>), Jia et al. (<xref ref-type="bibr" rid="B18">2020</xref>)</td>
<td valign="top" align="left">A discrete supervised hashing method based on column sampling is proposed for palmprint recognition.</td>
</tr> <tr>
<td valign="top" align="left">BIT (<xref ref-type="bibr" rid="B2">2020</xref>), Zhou et al. (<xref ref-type="bibr" rid="B49">2020</xref>)</td>
<td valign="top" align="left">A biologically based palmprint recognition transform feature extractor (BIT) is proposed.</td>
</tr> <tr>
<td valign="top" align="left">LCMFC (<xref ref-type="bibr" rid="B26">2020</xref>), Fei et al. (<xref ref-type="bibr" rid="B8">2020c</xref>)</td>
<td valign="top" align="left">It proposes a multi-feature learning method to extract directional and textural features of palmprint for palmprint recognition.</td>
</tr> <tr>
<td valign="top" align="left">Jia et al. (<xref ref-type="bibr" rid="B16">2021</xref>)</td>
<td valign="top" align="left">A new deep palmprint hash recognition method is proposed.</td>
</tr> <tr>
<td valign="top" align="left">Alrahawe et al. (<xref ref-type="bibr" rid="B1">2021</xref>)</td>
<td valign="top" align="left">A contactless biometric system for palmprint recognition based on convolutional neural network (ConvNet) was designed.</td>
</tr> <tr>
<td valign="top" align="left">KPCA (<xref ref-type="bibr" rid="B23">2021</xref>), Kusban (<xref ref-type="bibr" rid="B25">2021</xref>)</td>
<td valign="top" align="left">It designs a palmprint recognition method using Gabor technique and kernel principal component analysis.</td>
</tr> <tr>
<td valign="top" align="left">Jing et al. (<xref ref-type="bibr" rid="B19">2021</xref>)</td>
<td valign="top" align="left">A supervised group sparse regularizer is designed to optimize the localization and group sparsity of extracted palmprint features</td>
</tr> <tr>
<td valign="top" align="left">Zhao et al. (<xref ref-type="bibr" rid="B46">2022</xref>)</td>
<td valign="top" align="left">It proposes a new multi-view discriminative palmprint recognition method based on double cohesive learning.</td>
</tr> <tr>
<td valign="top" align="left">SLCN (<xref ref-type="bibr" rid="B34">2022</xref>), Fei et al. (<xref ref-type="bibr" rid="B10">2022</xref>)</td>
<td valign="top" align="left">It proposes a collaborative palmprint-specific binary feature learning method and a network consisting of a single convolutional layer.</td>
</tr> <tr>
<td valign="top" align="left">Li et al. (<xref ref-type="bibr" rid="B27">2022</xref>)</td>
<td valign="top" align="left">A method for localized triple pattern recognition of palmprints based on line features is proposed.</td>
</tr></tbody>
</table>
</table-wrap>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Comparison of the correct recognition rates of different methods on the Tongji dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Comparison of the correct recognition rates of different methods on the Poly-U dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0006.tif"/>
</fig>
<p><xref ref-type="table" rid="T2">Table 2</xref> shows the correct recognition rates of the different methods on the two datasets. According to the results in <xref ref-type="table" rid="T2">Table 2</xref>, it is clear that the proposed method achieves high performance on both datasets compared with the traditional and deep learning based methods in the existing work. According to the results in the table that the proposed method performs better on the PolyU dataset because of its higher requirement for image alignment accuracy during the image acquisition phase. On the tongji dataset, in addition to the excellent performance of the proposed method in this paper, the next best method is a new multi-view discriminative palmprint recognition method based on double cohesive learning proposed by Zhao et al. (<xref ref-type="bibr" rid="B46">2022</xref>), which improves the palmprint recognition performance by using palmprint features from multiple views presented in a low-dimensional subspace. The next best performer compared with the proposed method on the PolyU dataset is a biologically based transform feature extractor (BIT) for palmprint recognition proposed by Zhou et al. (<xref ref-type="bibr" rid="B49">2020</xref>). This method recognizes palmprint by simulating the visual perception mechanism of simple cells, which is computationally complex, not robust and cannot be directly applied to palmprint. The proposed method achieves stable recognition performance by controlling the transfer of features through a gating mechanism.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Comparison of recognition accuracy of different methods in PolyU and Tongji data set.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:#8f9496">
<th valign="top" align="left"><bold>Methods</bold></th>
<th valign="top" align="center"><bold>Tongji</bold></th>
<th valign="top" align="center"><bold>PolyU</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">COSDISH (<xref ref-type="bibr" rid="B3">2020</xref>), Jia et al. (<xref ref-type="bibr" rid="B18">2020</xref>)</td>
<td valign="top" align="center">96.38</td>
<td valign="top" align="center">95.88</td>
</tr> <tr>
<td valign="top" align="left">BIT (<xref ref-type="bibr" rid="B2">2020</xref>), Zhou et al. (<xref ref-type="bibr" rid="B49">2020</xref>)</td>
<td valign="top" align="center">94.69</td>
<td valign="top" align="center">99.25</td>
</tr> <tr>
<td valign="top" align="left">LCMFC (<xref ref-type="bibr" rid="B26">2020</xref>), Fei et al. (<xref ref-type="bibr" rid="B8">2020c</xref>)</td>
<td valign="top" align="center">97.12</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Jia et al. (<xref ref-type="bibr" rid="B16">2021</xref>)</td>
<td valign="top" align="center">95.36</td>
<td valign="top" align="center">98.81</td>
</tr> <tr>
<td valign="top" align="left">Alrahawe et al. (<xref ref-type="bibr" rid="B1">2021</xref>)</td>
<td valign="top" align="center">96.88</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Jing et al. (<xref ref-type="bibr" rid="B19">2021</xref>)</td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center">97.65</td>
</tr> <tr>
<td valign="top" align="left">KPCA (<xref ref-type="bibr" rid="B23">2021</xref>), Kusban (<xref ref-type="bibr" rid="B25">2021</xref>)</td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center">98.81</td>
</tr> <tr>
<td valign="top" align="left">SLCN (<xref ref-type="bibr" rid="B34">2022</xref>), Fei et al. (<xref ref-type="bibr" rid="B10">2022</xref>)</td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center">97.66</td>
</tr> <tr>
<td valign="top" align="left">Zhao et al. (<xref ref-type="bibr" rid="B46">2022</xref>)</td>
<td valign="top" align="center">97.69</td>
<td valign="top" align="center">&#x02013;</td>
</tr> <tr>
<td valign="top" align="left">Li et al. (<xref ref-type="bibr" rid="B27">2022</xref>)</td>
<td valign="top" align="center">&#x02013;</td>
<td valign="top" align="center">97.47</td>
</tr> <tr>
<td valign="top" align="left"><bold>GLGAnet</bold></td>
<td valign="top" align="center"><bold>98.50</bold></td>
<td valign="top" align="center"><bold>99.50</bold></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Bold value indicates the best experimental results.</p>
</table-wrap-foot>
</table-wrap>
<p>According to the results reported in <xref ref-type="fig" rid="F5">Figures 5</xref>, <xref ref-type="fig" rid="F6">6</xref>, the method proposed in this paper obtained a more competitive performance compared with other methods. Specifically, on the Tongji University palmprint dataset, the recognition accuracy of the method proposed in this paper was 98.5%, outperforming other methods by 2.12, 3.81, 1.38, 3.14, 1.62, and 0.81%, respectively. On the Hong Kong Polytechnic University palmprint dataset, achieved 99.8% recognition rate, outperforming other methods by 3.62, 0.25, 0.69, 0.81, 1.84, 1.85, and 2.03%, respectively. On the Tongji University palmprint dataset, the existing methods of convolutional neural networks were not effective for palmprint recognition. For example, Jia et al. (<xref ref-type="bibr" rid="B16">2021</xref>) used a deep palmprint hash recognition method which neglected to extract the global features of palmprints and obtained the features of palmprints through an ordinary convolutional neural network. However, the ordinary convolutional neural network was limited by the size of the perceptual field and could not achieve good results even when using a hash layer and multiple loss functions to jointly supervise the training process. Alrahawe et al. (<xref ref-type="bibr" rid="B1">2021</xref>) designed a palmprint recognition system based on convolutional neural network, which is limited by the size. The designed convolutional neural network can only extract some shallow information and cannot achieve high recognition results. On the PolyU dataset, some researchers have designed convolutional neural networks for palmprint recognition. For example, a collaborative palmprint-specific binary feature learning method and a compact network consisting of a single convolutional layer were proposed by Fei et al. (<xref ref-type="bibr" rid="B10">2022</xref>). The palmprint feature learning method in this method uses a feature projection function, which loses some information of the palmprint and cannot achieve good recognition results. In contrast, the proposed method uses a deep convolutional neural network to extract local features of the palmprint image in separate channels. The global features of the palmprint are then obtained using a Res-Vit structure based on a self-attentive mechanism, allowing the designed network to avoid the shortcomings of the above study. It is possible to model and extract not only the local but also the global information of the palmprint. In particular, the proposed method also designs a gating mechanism for the selection control of the extracted features according to the different features extracted by the deep convolutional neural network and Res-Vit, making the palmprint recognition the best possible.</p>
</sec>
<sec>
<title>4.3. Ablation experiments</title>
<p>To further validate the importance and practical contribution of the backbone network and the designed modules (including the gating mechanism) used in this paper, relevant ablation experiments were conducted.</p>
<p>In this experiment, GLGANet backbone is used as the baseline model and then its effectiveness is verified by adding different gating mechanisms one by one. Specifically, the performance of the following five models is compared: GLGANet backbone: a good pre-trained model for palmprint recognition using the GLGANet backbone network, which is the baseline model. GLGANet backbone&#x0002B;G1: the first gating mechanism is added on top of the GLNet backbone network to control the features extracted by the first deep convolutional layer. GLGANet backbone&#x0002B;G1&#x0002B;G2: A second gating mechanism is added to GLGANet&#x0002B;G1 to control the features extracted by the first and second deep convolutional layers. GLGANet backbone&#x0002B;G1&#x0002B;G2&#x0002B;G3&#x0002B;G4:On top of the above model, a fourth gating mechanism is added to control the global features extracted by the second Transformer structure to a deeper level of the palmprint. Full model: The proposed full model (GLGANet &#x0002B;G1&#x0002B;G2&#x0002B;G3&#x0002B;G4&#x0002B;G5).</p>
<p>A comparison of GLGANet backbone and GLGANet backbone&#x0002B;G1 demonstrates the effectiveness of the gating mechanism used in GLGANet backbone to control the selection of the local features extracted from the first depth convolutional layer for palm prints. Comparison of GLGANet backbone&#x0002B;G1 and GLGANet backbone&#x0002B;G1&#x0002B;G2 shows that the gating mechanism is effective in controlling the selection of features extracted from the deep convolutional layer. Comparing GLGANet backbone&#x0002B;G1&#x0002B;G2 with GLGANet backbone&#x0002B;G1&#x0002B;G2&#x0002B;G3 and GLGANet backbone&#x0002B;G1&#x0002B;G2&#x0002B;G3&#x0002B;G4, it verifies that the gating mechanism is effective in controlling the local features of palmprints extracted from the depth convolution layer. It also has a control effect on the global features of the palmprint extracted by the Transformer structure. Finally, by comparing GLGANet backbone&#x0002B;G1&#x0002B;G2&#x0002B;G3&#x0002B;G4 with the full model, the effectiveness of the gating mechanism in controlling the local features and the fusion of the local and global features of the palmprint by the adaptive convolution fusion module is verified.</p>
<p><xref ref-type="table" rid="T3">Table 3</xref> lists the objective performance exhibited by the different models. Each of the gating mechanisms described above has improved the correctness of the palmprint recognition results. The results did not improve when two gating mechanisms were added. The reason is that the features extracted by adding the first two gating mechanisms were fused with the layer-level features extracted by the backbone network and then convolved and pooled several times, losing some key information. Starting with the addition of the third gating mechanism, the results were significantly improved. The recognition results are more obvious with the addition of the last five gating mechanisms and the adaptive convolution module.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>The results of the ablation experiment.</p></caption> 
<table frame="box" rules="all">
<thead>
<tr style="background-color:#8f9496">
<th valign="top" align="left"><bold>Models</bold></th>
<th valign="top" align="center"><bold>Percentage of results correct (<italic>%</italic>)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">GLGANet backbone</td>
<td valign="top" align="center">98.00</td>
</tr> <tr>
<td valign="top" align="left">GLGANet backbone&#x0002B;g1</td>
<td valign="top" align="center">98.40</td>
</tr> <tr>
<td valign="top" align="left">GLGANet backbone&#x0002B;g1&#x0002B;g2</td>
<td valign="top" align="center">98.80</td>
</tr> <tr>
<td valign="top" align="left">GLGANet backbone&#x0002B;g1&#x0002B;g2&#x0002B;g3</td>
<td valign="top" align="center">99.00</td>
</tr> <tr>
<td valign="top" align="left">GLGANet backbone&#x0002B;g1&#x0002B;g2&#x0002B;g3&#x0002B;g4</td>
<td valign="top" align="center">99.20</td>
</tr> <tr>
<td valign="top" align="left">GLGANet backbone&#x0002B;g1&#x0002B;g2&#x0002B;g3&#x0002B;g4&#x0002B;g5</td>
<td valign="top" align="center"><bold>99.50</bold></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Bold value indicates the best experimental results.</p>
</table-wrap-foot>
</table-wrap>
<p><xref ref-type="fig" rid="F7">Figure 7</xref> shows the convergence of the loss function during the training of different models. As shown in <xref ref-type="fig" rid="F7">Figure 7</xref>, the proposed method does not achieve the best result when the first three gating mechanisms are added, which also corresponds to the results of the ablation experiments. (the effect is significantly improved when the third gating mechanism is added). As the number of gating mechanisms increases, the convergence of the loss function also verifies the effectiveness of the gating mechanisms. In particular, the loss function converges most significantly and the best recognition is achieved when all five gating mechanisms are added.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Convergence of the loss function during the training of different models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnbot-17-1203962-g0007.tif"/>
</fig></sec></sec>
<sec id="s5">
<title>5. Summary</title>
<p>This paper proposes a framework for palmprint recognition based on CNN and Transformer. The method extracts the local information of palmprint by using CNN and the global contextual information of palmprint by Transformer. In addition, a gating mechanism is designed based on the difference in the features extracted by the CNN and Transformer, and the gating mechanism is used to control the selection of the extracted local and global features. In addition, an adaptive convolutional fusion module is designed to effectively fuse the features obtained by the gating mechanism with those of the backbone network through single or multiple convolutions. Experimental results verify the effectiveness of the key components extracted by the proposed method. In addition, the method achieved better performance in benchmark tests on the Tongji palmprint dataset and Poly-U dataset compared with state-of-the-art methods. Future work may focus on exploring the use of fewer gating mechanisms to achieve high accuracy in palmprint recognition and extending the proposed method to other biometric recognition problems.</p></sec>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: <ext-link ext-link-type="uri" xlink:href="https://cslinzhang.github.io/ContactlessPalm/">https://cslinzhang.github.io/ContactlessPalm/</ext-link>.</p></sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.</p></sec>
</body>
<back>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>This research was jointed sponsored by National Natural Science Foundation of China (82205049), Natural Science Foundation of Chongqing (cstc2020jcyj-msxmX0259), Chongqing Medical Scientific Research Project (Joint project of Chongqing Health Commission and Science and Technology Bureau, 2022MSXM184), and Kuanren Talents Program of the Second Affiliated Hospital of Chongqing Medical University, China, Research on intelligent control technology of highway tunnel operation based on digital twin, the Key Project of Science and Technology of Ministry of Transport of China (Grant No. 2021-MS6-145).</p>
</sec>

<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>XY was employed by China Merchants Chongqing Communications Research and Design Institute Co., Ltd. YKJ was employed by company Double Deuce Sports.</p>
<p>The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alrahawe</surname> <given-names>E. A.</given-names></name> <name><surname>Humbe</surname> <given-names>V. T.</given-names></name> <name><surname>Shinde</surname> <given-names>G.</given-names></name></person-group> (<year>2021</year>). <article-title>A contactless palmprint biometric system based on CNN</article-title>. <source>Turkish J. Comput. Math. Educ</source>. <volume>12</volume>, <fpage>6344</fpage>&#x02013;<lpage>6356</lpage>. <pub-id pub-id-type="doi">10.1109/eSmarTA52612.2021.9515726</pub-id><pub-id pub-id-type="pmid">35009614</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><collab>BIT</collab></person-group>. (<year>2020</year>). <source>A biologically inspired transform feature extractor for palmprint recognition, namely BIT.</source></citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><collab>COSDISH</collab></person-group>. (<year>2020</year>). <source>A supervised hashing method for palmprint recognition- column sampling based discrete supervised hashing (COSDISH)</source>.</citation>
</ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dian</surname> <given-names>L.</given-names></name> <name><surname>Dongmei</surname> <given-names>S.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Contactless palmprint recognition based on convolutional neural network,&#x0201D;</article-title> in <source>2016 IEEE 13th International Conference on Signal Processing (ICSP)</source> (<publisher-loc>Chengdu</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1363</fpage>&#x02013;<lpage>1367</lpage>. <pub-id pub-id-type="doi">10.1109/ICSP.2016.7878049</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Lu</surname> <given-names>G.</given-names></name> <name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Teng</surname> <given-names>S.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Feature extraction methods for palmprint recognition: a survey and evaluation</article-title>. <source>IEEE Trans. Syst. Man Cybern</source>. <volume>49</volume>, <fpage>346</fpage>&#x02013;<lpage>363</lpage>. <pub-id pub-id-type="doi">10.1109/TSMC.2018.2795609</pub-id></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Wen</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2020a</year>). <article-title>Feature extraction for 3-d palmprint recognition: a survey</article-title>. <source>IEEE Trans. Instrum. Meas</source>. <volume>69</volume>, <fpage>645</fpage>&#x02013;<lpage>656</lpage>. <pub-id pub-id-type="doi">10.1109/TIM.2020.2964076</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Teng</surname> <given-names>S.</given-names></name> <name><surname>Guo</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>S.</given-names></name> <name><surname>Jia</surname> <given-names>W.</given-names></name></person-group> (<year>2020b</year>). <article-title>Joint multiview feature learning for hand-print recognition</article-title>. <source>IEEE Trans. Instrum. Meas</source>. <volume>69</volume>, <fpage>9743</fpage>&#x02013;<lpage>9755</lpage>. <pub-id pub-id-type="doi">10.1109/TIM.2020.3002463</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Wen</surname> <given-names>J.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name></person-group> (<year>2020c</year>). <article-title>Learning compact multifeature codes for palmprint recognition from a single training image per palm</article-title>. <source>IEEE Trans. Multim</source>. <volume>23</volume>, <fpage>2930</fpage>&#x02013;<lpage>2942</lpage>. <pub-id pub-id-type="doi">10.1109/TMM.2020.3019701</pub-id></citation>
</ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Zhang</surname> <given-names>W.</given-names></name> <name><surname>Teng</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Local apparent and latent direction extraction for palmprint recognition</article-title>. <source>Inform. Sci</source>. <volume>473</volume>, <fpage>59</fpage>&#x02013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1016/j.ins.2018.09.032</pub-id></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhao</surname> <given-names>S.</given-names></name> <name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Wen</surname> <given-names>J.</given-names></name> <name><surname>Xu</surname> <given-names>Y.</given-names></name></person-group> (<year>2022</year>). <article-title>Toward efficient palmprint feature extraction by learning a single-layer convolution network</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst</source>. <pub-id pub-id-type="doi">10.1109/TNNLS.2022.3160597</pub-id><pub-id pub-id-type="pmid">35349454</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Genovese</surname> <given-names>A.</given-names></name> <name><surname>Piuri</surname> <given-names>V.</given-names></name> <name><surname>Plataniotis</surname> <given-names>K. N.</given-names></name> <name><surname>Scotti</surname> <given-names>F.</given-names></name></person-group> (<year>2019</year>). <article-title>Palmnet: Gabor-PCA convolutional networks for touchless palmprint recognition</article-title>. <source>IEEE Trans. Inform. Forens. Sec</source>. <volume>14</volume>, <fpage>3160</fpage>&#x02013;<lpage>3174</lpage>. <pub-id pub-id-type="doi">10.1109/TIFS.2019.2911165</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>Z.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Zuo</surname> <given-names>W.</given-names></name></person-group> (<year>2009</year>). <article-title>Palmprint verification using binary orientation co-occurrence vector</article-title>. <source>Pattern Recogn. Lett</source>. <volume>30</volume>, <fpage>1219</fpage>&#x02013;<lpage>1227</lpage>. <pub-id pub-id-type="doi">10.1016/j.patrec.2009.05.010</pub-id></citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>X.</given-names></name> <name><surname>Qi</surname> <given-names>G.</given-names></name> <name><surname>Zhu</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Cong</surname> <given-names>B.</given-names></name> <name><surname>Bai</surname> <given-names>L.</given-names></name></person-group> (<year>2023</year>). <article-title>Medical image segmentation method based on multi-feature interaction and fusion over cloud computing</article-title>. <source>Simul. Model. Pract. Theory</source> <volume>126</volume>:<fpage>102769</fpage>. <pub-id pub-id-type="doi">10.1016/j.simpat.2023.102769</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Hu</surname> <given-names>R.-X.</given-names></name> <name><surname>Lei</surname> <given-names>Y.-K.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Gui</surname> <given-names>J.</given-names></name></person-group> (<year>2013</year>). <article-title>Histogram of oriented lines for palmprint recognition</article-title>. <source>IEEE Trans. Syst. Man Cybern</source>. <volume>44</volume>, <fpage>385</fpage>&#x02013;<lpage>395</lpage>. <pub-id pub-id-type="doi">10.1109/TSMC.2013.2258010</pub-id></citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Huang</surname> <given-names>D.-S.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2008</year>). <article-title>Palmprint verification based on robust line orientation code</article-title>. <source>Pattern Recogn</source>. <volume>41</volume>, <fpage>1504</fpage>&#x02013;<lpage>1513</lpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2007.10.011</pub-id></citation>
</ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Huang</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>B.</given-names></name> <name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Min</surname> <given-names>H.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Deep multi-loss hashing network for palmprint retrieval and recognition,&#x0201D;</article-title> in <source>2021 IEEE International Joint Conference on Biometrics (IJCB)</source> (<publisher-loc>Shenzhen</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1109/IJCB52358.2021.9484403</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Ren</surname> <given-names>Q.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>S.</given-names></name> <name><surname>Min</surname> <given-names>H.</given-names></name> <name><surname>Chen</surname> <given-names>Y.</given-names></name></person-group> (<year>2022</year>). <article-title>EEPNet: an efficient and effective convolutional neural network for palmprint recognition</article-title>. <source>Pattern Recogn. Lett</source>. <volume>159</volume>, <fpage>140</fpage>&#x02013;<lpage>149</lpage>. <pub-id pub-id-type="doi">10.1016/j.patrec.2022.05.015</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jia</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>B.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Min</surname> <given-names>H.</given-names></name> <name><surname>Feng</surname> <given-names>H.</given-names></name></person-group> (<year>2020</year>). <article-title>A performance evaluation of hashing techniques for 2D and 3D palmprint retrieval and recognition</article-title>. <source>IEEE Sensors J</source>. <volume>20</volume>, <fpage>11864</fpage>&#x02013;<lpage>11873</lpage>. <pub-id pub-id-type="doi">10.1109/JSEN.2020.2973357</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jing</surname> <given-names>K.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Xu</surname> <given-names>X.</given-names></name></person-group> (<year>2021</year>). <article-title>Double-Laplacian mixture-error model-based supervised group-sparse coding for robust palmprint recognition</article-title>. <source>IEEE Trans. Circuits Syst. Video Technol</source>. <volume>32</volume>, <fpage>3125</fpage>&#x02013;<lpage>3140</lpage>. <pub-id pub-id-type="doi">10.1109/TCSVT.2021.3103941</pub-id></citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kong</surname> <given-names>A.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Kamel</surname> <given-names>M.</given-names></name></person-group> (<year>2006</year>). <article-title>Palmprint identification using feature-level fusion</article-title>. <source>Pattern Recogn</source>. <volume>39</volume>, <fpage>478</fpage>&#x02013;<lpage>487</lpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2005.08.014</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kong</surname> <given-names>A.-K.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2004</year>). <article-title>&#x0201C;Competitive coding scheme for palmprint verification,&#x0201D;</article-title> in <source>Proceedings of the 17th International Conference on Pattern Recognition, Vol. 1</source> (<publisher-loc>Cambridge</publisher-loc>), <fpage>520</fpage>&#x02013;<lpage>523</lpage>. <pub-id pub-id-type="doi">10.1109/ICPR.2004.1334184</pub-id></citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Korichi</surname> <given-names>M.</given-names></name> <name><surname>Samai</surname> <given-names>D.</given-names></name> <name><surname>Meraoumia</surname> <given-names>A.</given-names></name> <name><surname>Benlamoudi</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Towards effective 2D and 3D palmprint recognition using transfer learning deep features and Reliff method,&#x0201D;</article-title> in <source>2021 International Conference on Recent Advances in Mathematics and Informatics (ICRAMI)</source> (<publisher-loc>Tebessa</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/ICRAMI52622.2021.9585911</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><collab>KPCA</collab></person-group>. (<year>2021</year>). <source>A dimensionality reduction method for palmprint image processing-kernel principal component analysis (KPCA)</source>.</citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krizhevsky</surname> <given-names>A.</given-names></name> <name><surname>Sutskever</surname> <given-names>I.</given-names></name> <name><surname>Hinton</surname> <given-names>G. E.</given-names></name></person-group> (<year>2017</year>). <article-title>ImageNet classification with deep convolutional neural networks</article-title>. <source>Commun. ACM</source> <volume>60</volume>, <fpage>84</fpage>&#x02013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1145/3065386</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kusban</surname> <given-names>M.</given-names></name></person-group> (<year>2021</year>). <article-title>Improvement palmprint recognition system by adjusting image data reference points</article-title>. <source>J. Phys</source>. 1858, 012077. <pub-id pub-id-type="doi">10.1088/1742-6596/1858/1/012077</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><collab>LCMFC</collab></person-group>. (<year>2020</year>). <source>A multifeature learning method to jointly learn compact multifeature codes (LCMFCs) for palmprint recognition with a single training sample per palmprint</source>.</citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>M.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Liu</surname> <given-names>H.</given-names></name> <name><surname>Meng</surname> <given-names>Q.</given-names></name></person-group> (<year>2022</year>). Palmprint recognition based on the line feature local tri-directional patterns <italic>IET Biometr</italic>. <volume>11</volume>, <fpage>570</fpage>&#x02013;<lpage>580</lpage>. <pub-id pub-id-type="doi">10.1049/bme2.12085</pub-id></citation>
</ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liang</surname> <given-names>X.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Lu</surname> <given-names>G.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name></person-group> (<year>2021</year>). <article-title>CompNet: competitive neural network for palmprint recognition using learnable Gabor kernels</article-title>. <source>IEEE Signal Process. Lett</source>. <volume>28</volume>, <fpage>1739</fpage>&#x02013;<lpage>1743</lpage>. <pub-id pub-id-type="doi">10.1109/LSP.2021.3103475</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>C.</given-names></name> <name><surname>Zhong</surname> <given-names>D.</given-names></name> <name><surname>Shao</surname> <given-names>H.</given-names></name></person-group> (<year>2021</year>). <article-title>Few-shot palmprint recognition based on similarity metric hashing network</article-title>. <source>Neurocomputing</source> <volume>456</volume>, <fpage>540</fpage>&#x02013;<lpage>549</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2020.07.153</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matkowski</surname> <given-names>W. M.</given-names></name> <name><surname>Chai</surname> <given-names>T.</given-names></name> <name><surname>Kong</surname> <given-names>A. W. K.</given-names></name></person-group> (<year>2019</year>). <article-title>Palmprint recognition in uncontrolled and uncooperative environment</article-title>. <source>IEEE Trans. Inform. Forens. Secur</source>. <volume>15</volume>, <fpage>1601</fpage>&#x02013;<lpage>1615</lpage>. <pub-id pub-id-type="doi">10.1109/TIFS.2019.2945183</pub-id></citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patil</surname> <given-names>J. P.</given-names></name> <name><surname>Nayak</surname> <given-names>C.</given-names></name> <name><surname>Jain</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Palmprint recognition using DWT, DCT and PCA techniques,&#x0201D;</article-title> in <source>2015 IEEE International Conference on Computational Intelligence and Computing Research (ICCIC)</source> (<publisher-loc>IEEE</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/ICCIC.2015.7435677</pub-id></citation>
</ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>Y.</given-names></name> <name><surname>Yu</surname> <given-names>X.</given-names></name> <name><surname>Sohn</surname> <given-names>K.</given-names></name> <name><surname>Chandraker</surname> <given-names>M.</given-names></name> <name><surname>Jain</surname> <given-names>A. K.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Towards universal representation learning for deep face recognition,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE)</source>, 6817&#x02013;6826. <pub-id pub-id-type="doi">10.1109/CVPR42600.2020.00685</pub-id></citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname> <given-names>K.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2014</year>). <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <source>arXiv [Preprint]</source>. arXiv: 1409.1556. <pub-id pub-id-type="doi">10.48550/arXiv.1409.1556</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><collab>SLCN</collab></person-group>. (<year>2022</year>). <source>A single-layer convolution network (SLCN) for palmprint recognition.</source></citation>
</ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tarawneh</surname> <given-names>A. S.</given-names></name> <name><surname>Chetverikov</surname> <given-names>D.</given-names></name> <name><surname>Hassanat</surname> <given-names>A. B.</given-names></name></person-group> (<year>2018</year>). <article-title>Pilot comparative study of different deep features for palmprint identification in low-quality images</article-title>. <source>arXiv [Preprint]</source>. arXiv: 1804.04602. <pub-id pub-id-type="doi">10.48550/arXiv.1804.04602</pub-id></citation>
</ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ungureanu</surname> <given-names>A.-S.</given-names></name> <name><surname>Salahuddin</surname> <given-names>S.</given-names></name> <name><surname>Corcoran</surname> <given-names>P.</given-names></name></person-group> (<year>2020</year>). <article-title>Toward unconstrained palmprint recognition on consumer devices: a literature review</article-title>. <source>IEEE Access</source> <volume>8</volume>, <fpage>86130</fpage>&#x02013;<lpage>86148</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.2992219</pub-id></citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Gong</surname> <given-names>W.</given-names></name> <name><surname>Xu</surname> <given-names>X.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;A summary of palmprint recognition technology,&#x0201D;</article-title> in <source>2019 4th International Conference on Control, Robotics and Cybernetics (CRC)</source> (<publisher-loc>Tokyo</publisher-loc>), <fpage>91</fpage>&#x02013;<lpage>97</lpage>. <pub-id pub-id-type="doi">10.1109/CRC.2019.00027</pub-id></citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>J.</given-names></name> <name><surname>Zhao</surname> <given-names>M.</given-names></name> <name><surname>Xie</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Zhu</surname> <given-names>W.</given-names></name></person-group> (<year>2022</year>). <article-title>Fusion of bilateral 2DPCA information for image reconstruction and recognition</article-title>. <source>Appl. Sci</source>. 12, 12913. <pub-id pub-id-type="doi">10.3390/app122412913</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yaddaden</surname> <given-names>Y.</given-names></name> <name><surname>Parent</surname> <given-names>J.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;An efficient palmprint authentication system based on one-class SVM and hog descriptor,&#x0201D;</article-title> in <source>2022 2nd International Conference on Advanced Electrical Engineering</source> (<publisher-loc>ICAEE</publisher-loc>) (Constantine), <fpage>1</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1109/ICAEE53772.2022.9962020</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>A.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Sun</surname> <given-names>Q.</given-names></name> <name><surname>Zhang</surname> <given-names>Q.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Palmprint recognition based on CNN and local coding features,&#x0201D;</article-title> in <source>2017 6th International Conference on Computer Science and Network Technology (ICCSNT)</source> (<publisher-loc>Dalian</publisher-loc>), <fpage>482</fpage>&#x02013;<lpage>487</lpage>. <pub-id pub-id-type="doi">10.1109/ICCSNT.2017.8343744</pub-id></citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>B.</given-names></name> <name><surname>Xiang</surname> <given-names>X.</given-names></name> <name><surname>Yao</surname> <given-names>J.</given-names></name> <name><surname>Xu</surname> <given-names>D.</given-names></name></person-group> (<year>2020</year>). <article-title>3D palmprint recognition using complete block wise descriptor</article-title>. <source>Multim. Tools Appl</source>. <volume>79</volume>, <fpage>21987</fpage>&#x02013;<lpage>22006</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-020-09000-7</pub-id></citation>
</ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Fingerprint recognition system based on big data and multi-feature fusion,&#x0201D;</article-title> in <source>2020 International Conference on Culture-Oriented Science &#x00026; Technology (ICCST)</source> (<publisher-loc>Beijing</publisher-loc>), <fpage>473</fpage>&#x02013;<lpage>475</lpage>. <pub-id pub-id-type="doi">10.1109/ICCST50977.2020.00097</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Kong</surname> <given-names>W.-K.</given-names></name> <name><surname>You</surname> <given-names>J.</given-names></name> <name><surname>Wong</surname> <given-names>M.</given-names></name></person-group> (<year>2003</year>). <article-title>Online palmprint identification</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell</source>. <volume>25</volume>, <fpage>1041</fpage>&#x02013;<lpage>1050</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2003.1227981</pub-id><pub-id pub-id-type="pmid">21245008</pub-id></citation></ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Cheng</surname> <given-names>Z.</given-names></name> <name><surname>Shen</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Palmprint and palmvein recognition based on DCNN and a new large-scale contactless palmvein dataset</article-title>. <source>Symmetry</source> <volume>10</volume>:<fpage>78</fpage>. <pub-id pub-id-type="doi">10.3390/sym10040078</pub-id></citation>
</ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Cross-view gait recognition with deep universal linear embeddings,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>Nashville, TN</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>9095</fpage>&#x02013;<lpage>9104</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR46437.2021.00898</pub-id></citation>
</ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>S.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Fei</surname> <given-names>L.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Zhao</surname> <given-names>P.</given-names></name></person-group> (<year>2022</year>). <article-title>Double-cohesion learning based multiview and discriminant palmprint recognition</article-title>. <source>Inform. Fus</source>. <volume>83</volume>, <fpage>96</fpage>&#x02013;<lpage>109</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2022.03.005</pub-id></citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhong</surname> <given-names>D.</given-names></name> <name><surname>Du</surname> <given-names>X.</given-names></name> <name><surname>Zhong</surname> <given-names>K.</given-names></name></person-group> (<year>2019</year>). <article-title>Decade progress of palmprint recognition: a brief survey</article-title>. <source>Neurocomputing</source> <volume>328</volume>, <fpage>16</fpage>&#x02013;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2018.03.081</pub-id></citation>
</ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhong</surname> <given-names>D.</given-names></name> <name><surname>Zhu</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>Centralized large margin cosine loss for open-set deep palmprint recognition</article-title>. <source>IEEE Trans. Circuits Syst. Video Technol</source>. <volume>30</volume>, <fpage>1559</fpage>&#x02013;<lpage>1568</lpage>. <pub-id pub-id-type="doi">10.1109/TCSVT.2019.2904283</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>X.</given-names></name> <name><surname>Zhou</surname> <given-names>K.</given-names></name> <name><surname>Shen</surname> <given-names>L.</given-names></name></person-group> (<year>2020</year>). <article-title>Rotation and translation invariant palmprint recognition with biologically inspired transform</article-title>. <source>IEEE Access</source> <volume>8</volume>, <fpage>80097</fpage>&#x02013;<lpage>80119</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.2990736</pub-id></citation>
</ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>Z.</given-names></name> <name><surname>He</surname> <given-names>X.</given-names></name> <name><surname>Qi</surname> <given-names>G.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Cong</surname> <given-names>B.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>Brain tumor segmentation based on the fusion of deep semantics and edge information in multimodal MRI</article-title>. <source>Inform. Fusion</source> <volume>91</volume>, <fpage>376</fpage>&#x02013;<lpage>387</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2022.10.022</pub-id><pub-id pub-id-type="pmid">34899175</pub-id></citation></ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>Z.</given-names></name> <name><surname>Lei</surname> <given-names>Y.</given-names></name> <name><surname>Qi</surname> <given-names>G.</given-names></name> <name><surname>Chai</surname> <given-names>Y.</given-names></name> <name><surname>Mazur</surname> <given-names>N.</given-names></name> <name><surname>An</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>A review of the application of deep learning in intelligent fault diagnosis of rotating machinery</article-title>. <source>Measurement</source> <volume>2022</volume>:<fpage>112346</fpage>. <pub-id pub-id-type="doi">10.1016/j.measurement.2022.112346</pub-id></citation>
</ref>
</ref-list> 
</back>
</article> 