Skip to main content

ORIGINAL RESEARCH article

Front. Neurorobot., 11 January 2023
Volume 16 - 2022 | https://doi.org/10.3389/fnbot.2022.1065099

A deep ensemble learning method for single finger-vein identification

Chongwen Liu1,2 Huafeng Qin1,2* Qun Song1,2 Huyong Yan1,2 Fen Luo1,2
  • 1College of Artificial Intelligence, Chongqing Technology and Business University, Chongqing, China
  • 2Chongqing Key Laboratory of Intelligent Perception and BlockChain Technology, Chongqing Technology and Business University, Chongqing, China

Finger-vein biometrics has been extensively investigated for personal verification. Single sample per person (SSPP) finger-vein recognition is one of the open issues in finger-vein recognition. Despite recent advances in deep neural networks for finger-vein recognition, current approaches depend on a large number of training data. However, they lack the robustness of extracting robust and discriminative finger-vein features from a single training image sample. A deep ensemble learning method is proposed to solve the SSPP finger-vein recognition in this article. In the proposed method, multiple feature maps were generated from an input finger-vein image, based on various independent deep learning-based classifiers. A shared learning scheme is investigated among classifiers to improve their feature representation captivity. The learning speed of weak classifiers is also adjusted to achieve the simultaneously best performance. A deep learning model is proposed by an ensemble of all these adjusted classifiers. The proposed method is tested with two public finger vein databases. The result shows that the proposed approach has a distinct advantage over all the other tested popular solutions for the SSPP problem.

1. Introduction

With the wide application of the internet, information security has become increasingly critical. Traditional personal identification technique, such as key and password, is difficult to meet people's needs. For example, the key is easily copied and missed, and the password is usually forgotten, especially for older people. Biometric technique as a solution has been widely investigated in recent years. Compared to the traditional identification approach, physiological characteristics to identify or verify a person has the following advantages (Albrecht et al., 2009): (1) Difficult to be missed; (2) Difficult to be forged; (3) Easy to use; (4) Easy to carry. Various traits, such as the face iris, fingerprint, and vein, have been employed for the recognition of a person and are broadly split into two categories (Vodinh, 2012; Kuzu et al., 2020): (1) Extrinsic characteristics, e.g., face, iris, and fingerprint. (2) Intrinsic characteristics, e.g., finger-vein, hand-vein, and palm-vein. Extrinsic characteristics are suspected to be copied and forged, and their fake version has been proven to be successfully employed to attack the recognition system (Khan et al., 2015), and improve the quality of service (Wu et al., 2022). On the contrary, the intrinsic characteristics are concealed in our bodies and are very difficult to be copied without the user's willingness. In addition, only the vein in a living body can be captured effectively and further used for identification. Thus, intrinsic traits provide high privacy and security in practical applications.

1.1. Related work

Despite the recent advances in finger-vein biometric recognition, it is still a challenging task in practical application since vein-capturing results are easily affected by many factors, such as illumination, environment temperature, and the behavior of the user. These factors cannot avoid during the capturing process, in this case, the training dataset may contain a large number of low-quality finger-vein images, which may decrease the recognition accuracy. To achieve robust recognition, various approaches are proposed for vein recognition in recent years and are broadly categorized into the following categories (Hou et al., 2022; Shaheed et al., 2022b).

(1) Local descriptor-based approaches: Local descriptor-based approaches mainly consist of local statistical information-based methods and local invariant-based methods. Local statistical information-based methods include local binary pattern (LBP) (Lee et al., 2010, 2011; Yang et al., 2014; Kang et al., 2015), local line binary pattern (LLBP) (Rosdi et al., 2011; Yang et al., 2013), efficient local binary pattern (ELBP) (Liu and Kim, 2016), discriminative binary codes (DBC) (Xi et al., 2016), and fuzzy images (Qin et al., 2022). A typical representation of local invariant-based methods is the scale-invariant feature transform (SIFT) (Qin et al., 2013; Wang et al., 2015).

(2) Superpixel-based feature extraction approach: The representative approaches in this category are the superpixel-based feature (Liu et al., 2014), vein textons map (Dong et al., 2014), hyper information feature (Xi et al., 2014), and personalized feature (Xi et al., 2013). Superpixel-based feature extraction methods have achieved a high recognition rate in some public databases (Kirchgasser et al., 2020).

(3) Subspace-learning-based approaches: Subspace learning as a powerful technique have been widely used in pattern recognition task such as vein identification. A projection matrix computed from training data is employed to map the finger-vein images into subspace, and the resulting features are further used for recognition. The typical methods include principal component analysis (PCA) (Wu and Liu, 2011a), two dimensional principal component analysis (2DPCA) (Qiu et al., 2016), two-directional and two-dimensional principal component analysis ((2D)2PCA) (Yang et al., 2012; Li et al., 2017; Zhang et al., 2021; Ban et al., 2022; She et al., 2022), linear discriminant analysis (LDA) (Wu and Liu, 2011b), high-dimensional state space (Zhang et al., 2022), self-feature-based method (Xie et al., 2022), and latent factor model (Wu et al., 2022).

(4) Deep learning-based approaches: The deep learning-based approaches have been directly used to learn robust features from original images and successfully applied for computation vision tasks. Some researchers brought them into finger-vein recognition. For example, deep learning approaches are employed for vein image segmentation (Liskowski and Krawiec, 2016; Qin et al., 2019; Yang et al., 2019; Shaheed et al., 2022a), quality assessment of vein image (Qin and Yacoubi, 2015; Qin and El-Yacoubi, 2018), fuzzy networks (Liu H. et al., 2022; Lu et al., 2022; Muthusamy and Rakkimuthu, 2022), and finger-vein recognition (Wang et al., 2017; Avci et al., 2019; Gumusbas et al., 2019; Zhang J. et al., 2019).

1.2. Motivation

As discussed in the related works, the handcrafted-based approaches are proposed based on prior human knowledge, some vein features related to recognition may be missed during the feature extraction process. On the contrary, the deep learning-based extraction approaches without any prior assumption can automatically extract high-level features by representation learning that are objectively related to vein recognition. The deep learning-based extraction approach takes the original image pixels as input and iterative uncovers hierarchical features. In this way, the decision errors on recognition are minimized. The need for voids is explicitly extracting some image processing-based features that might discard relevant information about image classification. Therefore, the deep learning-based approaches are capable of extracting more complete vein features for recognition, compared to handcrafted approaches. Currently, deep learning-based methods, such as deep neural networks, show high recognition performance because they harness rich prior knowledge acquired by training them on a huge training dataset. However, in finger-vein recognition, it is impossible to capture a lot of vein samples from the same finger, so the training sample of each class is generally limited. For example, the samples from each finger are < 12 in existing databases (Miura et al., 2004, 2007; Das et al., 2019). Especially, there is only one single sample per finger for single sample per person (SSPP) problem when considering their limited storage and privacy policy. Therefore, it becomes particularly intractable for such an identification system with SSPP when within-class information is not available to predict the unknown variations in query samples. Currently, various approaches have been proposed for biometrics identification with SSPP such as face (Wu and Deng, 2016; Wang et al., 2018), palmprint (Shao and Zhong, 2021), and fingerprint (Chatterjee et al., 2017). Similarly, the corresponding finger-vein recognition systems still suffer from the SSPP problem due to the following facts: 1) There is only one enrollment sample in some recognition systems, such as credit card and large-scale recognition, because of the limitation of storage capability. 2) To achieve online identification, the recognition systems usually store one sample per subject to improve processing speed and time. 3) Some users are not willing to cooperatively capture sufficient samples for their personal privacy. Providing one enrollment sample is convenient, simple, and acceptable for users.

As described in previous works, deep learning techniques show a powerful capacity for feature representation, but they generally require sufficient training samples to train a large number of network parameters. Therefore, their learning capacity may have not been well exploited for finger-vein SSPP due to the limited training data of each class (a person or subject) for SSPP. Besides, the deep learning model is easy to suffer from over-fitting on a small amount of dataset. As a result, deep learning-based vein identification approaches may not achieve high performance for finger-vein SSPP.

Ensemble learning aims at combining multiple learners to obtain a more robust representation of the object and is successfully applied for vision tasks such as SAR image category (Zhao et al., 2016), fault diagnosis (Liu et al., 2021), image cluster (Tsai et al., 2014), and human activity recognition (Jethanandani et al., 2020). In addition, some researchers applied it to biometrics, e.g., classification tasks such as fingerprint classification (Zhang et al., 2011b), palm-vein recognition (Joardar et al., 2017), and face recognition (Bhatt et al., 2014; Ding and Tao, 2018). As the features from different learners can achieve a complementary representation for the input image, their combination performs well for identification. To extract enough features from a single finger-vein training sample, in this article, we further research our work on Liu C. et al. (2022) and proposed a deep ensemble learning approach for finger-vein identification.

The rest of the paper is organized as follows. The methodology and significant contributions are listed in Section 2. In Section 3, the proposed method is described in detail. The contract experiment and the results are reported in Section 4. In Section 5, the full research work is concluded for easy reading.

2. Proposed method system and our contributions

Motivated by the success of ensemble learning and driven by the SSPP finger-vein recognition, we propose an ensemble deep neural network to learn robust representation from a single sample for SSPP finger-vein recognition. In our work, multiple deep learning classifiers are employed to extract robust features from different feature maps generated from an original input image, and then a robust deep ensemble learning approach is generated by combining all weak classifiers. To further improve the performance of our approach, a shared learning approach is investigated during the training process. Besides, we proposed a learning speed adjustment approach so that all weak classifiers can achieve the best performance at the same time. As each classifier can capture robust features, our ensemble learning approach produces a more complete representation of a finger-vein image for identification. The main contributions of this article are summarized as follows:

(1) This work makes the first attempt at SSPP finger-vein identification. In this work, a deep ensemble learning model is proposed for identification with a single finger-vein training sample. First, we employ three baselines to generate multiple feature maps from an original finger-vein image. With these maps, we train multiple convolutional neural networks (CNNs) in parallel to obtain weak classifiers. Second, all classifiers are combined to obtain an ensemble classifier for vein identification. The experimental results imply that the proposed system achieves state-of-the-art recognition results.

(2) We proposed a shared learning scheme to share representations from multiple feature maps. Generally, the classifier of CNN is easily overfitting for the SSPP problem, so a shared learning scheme is employed to improve their performance. We employ a feature map to train a classifier. In this way, multiple classifiers are obtained to extract features. To improve the performance of classifiers, the knowledge is transferred among classifiers if their input maps are similar. Specifically, for a given classifier, we stop to train it at a fixed number of iterative steps and fine-tune it by relative feature maps. As the knowledge of input maps is exploited by multiple classifiers, combining them can achieve a robust representation of a finger-vein image. The experimental results show that the shared learning scheme can improve the performance of the deep ensemble learning model.

(3) We develop a dynamic speed adjustment scheme to automatically control the learning speed of each classifier. The learning speeds of all classifiers are generally different, so it is difficult for them to achieve the best performance at the same time, which results in poor performance for ensemble classifiers. To solve this problem, a learning speed adjustment approach is proposed to improve the performance of our deep assemble learning approach. For example, we divide the whole training process into several phases. After training a classifier in the current phase, the number of training steps in the next phase is computed by the proposed speed adjustment scheme. In this way, the learning speed of each classifier is adjusted dynamically so that they achieve optimal results at the same time, which enables the ensemble classifier to achieve the best performance for vein identification.

The framework of our work is shown in Figure 1. First, we generate feature maps from an input image based on several baselines. With the resulting map and input image, we train several weak classifiers separately. We use CNN models as weak classifiers. Second, we ensemble all classifiers to obtain one ensemble classifier for identification. After training, an input image is subject to prepossessing base on baselines and the resulting maps are taken as the input of our deep learning ensemble mode to compute its probability of being to a class. The testing process is shown in the red line in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Framework.

3. Ensemble learning for SSPP finger-vein recognition

3.1. Feature extraction

To achieve robust performance in the SSPP problem, it is necessary to generate multiple feature maps from the single training sample of each class. The feature maps represent different aspects of the object (Felzenszwalb et al., 2010), so their combination can achieve better performance (Li et al., 2018). Here, three baselines (e.g., CNN Das et al., 2019, Gabor filters Zhang et al., 2019a, and LBP Kang et al., 2015) have achieved promising performance for finger-vein recognition, so we employ them to produce feature maps, which are taken as the input of ensemble model for classification.

3.1.1. Finger-vein images

Generally, the vein vessel is difficult to be observed in visible light, but it is captured by near-infrared light with 760 nm wavelength. The vein pattern appears darker than the other regions of the finger because only the blood vessels absorb the infrared rays. Some studies (Kumar and Zhou, 2012) have shown that using extracting vein patterns from vein images achieves promising performance for identification. The finger-vein image sample is shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Finger-vein image.

3.1.2. Segmentation

As shown in Figure 2, the contrast of the vein pattern in the original finger-vein image is poor, which results in low recognition accuracy. Previous studies showed that image segmentation as a solution has been employed for feature extraction and shows good performance (Miura et al., 2007; Song et al., 2011; Qin and El-Yacoubi, 2017). In recent years, deep learning-based methods have shown a more robust capacity for feature representation compared to handcrafted approaches. So, some researchers bring deep learning-based methods to vein segmentation (Liskowski and Krawiec, 2016; Qin and El-Yacoubi, 2017; Qin et al., 2019; Yang et al., 2019). To train a good weak classifier, the CNN-based model is used to extract robust vein texture patterns, which are input into the weak classifier. First, a CNN-based approach is developed to predict the probability of pixels belonging to veins or backgrounds by learning a deep feature representation. As the finger-vein consists of clear regions and ambiguous regions, several baselines are employed to automatically label pixels as veins or backgrounds in the image's clear regions, thus avoiding the tedious and prone-to-error manual labeling. Then, a CNN is trained to extract the vein patterns from any image region. Second, to improve the performance, an original method based on an FCN was to recover missing finger-vein patterns in the binary image. Figure 3 illustrates the segmentation image from a gray-scale image.

FIGURE 3
www.frontiersin.org

Figure 3. Segmentation of finger-vein image.

3.1.3. Local binary pattern

The Local binary pattern describes the relationship between the neighborhood points and the corresponding center point, with the features of constant rotation and grayscale. It is widely used to extract finger vein features and shows good performance (Lee et al., 2010, 2011; Rosdi et al., 2011; Yang et al., 2013, 2014; Kang et al., 2015). Therefore, we employ LBP to local information of finger-vein images. The LBP is computed by

LBPP,R=P=0P-1s(gp-gc)2ps(x)={1,x>=00,x<0.    (1)

Where gc is the gray value of the central pixel, gp is the value of its neighbors, P is the number of neighbors, and R is the radius of the neighborhood.

In Kocher et al. (2016), the author empirically evaluates different features obtained by using these more recent LBP-related feature extraction techniques for finger-vein spoofing detection. The LBP feature map provides a local representation of a finger-vein image. LBP patterns are extracted from the original image, and segmented image are shown in Figure 4.

FIGURE 4
www.frontiersin.org

Figure 4. Both original image and segmented image transform by LBP.

3.1.4. Gabor

The Gabor filter is a type of wavelet, it has good time-domain and frequency-domain transform characteristics. Gabor functions are used to construct filters with different scaling directions caused by different parameters (e.g., spatial position, frequency, phase, and direction). Furthermore, the Gabor filter is widely used to capture texture information, and it adapts to extract features from a finger-vein image. In finger vein recognition, there have been more studies using Gabor as a feature, such as Yang et al. (2009), Cho et al. (2012), and Zhang et al. (2019b). The Gabor filter is defined as follows:

G=exp(-x2+γ2y22δ2)exp(i(2πxλ+ψ))    (2)

where

x=xcosθ+ysinθy=-xsinθ+ycosθ    (3)

In the Gabor function, λ is the wavelength of the cosine factor; θ is the orientation of the normal to the parallel stripes, ψ is the phase offset of the cosine factor, δ is the standard deviation of the Gaussian envelope, and γ is the spatial aspect ratio.

For the practical application of finger-vein recognition, the real part of the Gabor filter is used. In Zhang et al. (2019a)'s study, the Garbo filter is used to extract texture information of a finger-vein image, and a CNN network is employed to finish the recognition work. The original image and Gabor feature are shown in Figure 5.

FIGURE 5
www.frontiersin.org

Figure 5. Both the original image and segmented image transform by the Gabor filter.

We employ the three baselines to extract five different features from the original images. Including the original finger-vein images, there are a total of six feature maps, as shown in Table 1. The six different feature maps are shown in Figure 6.

TABLE 1
www.frontiersin.org

Table 1. Feature maps.

FIGURE 6
www.frontiersin.org

Figure 6. Samples of 6 different feature maps. (A) 1st feature map, (B) 2nd feature map, (C) 3rd feature map, (D) 4th feature map, (E) 5th feature map, and (F) 6th feature map.

3.2. Convolutional neural networks

The ensemble learning model is constructed from the combination of weak classifiers (Sagi and Rokach, 2018). The choice of a weak classifier affects the performance of the ensemble model. In previous work, CNNs were widely used in finger vein recognition and achieved good recognition performance, such as Wang et al. (2017), Gumusbas et al. (2019), Avci et al. (2019), Gumusbas et al. (2019), and Zhang J. et al. (2019). In this article, we choose the CNN model as the independent weak classifier, and each CNN is trained by one feature map, as shown in Figure 1. Each trained CNN model describes one aspect of the finger vein.

Convolutional neural network is a multi-layer perception network with hidden layers. A traditional recognition model for a classifier can be formulated by minimizing the error function. During training a deep network, the gradient descent method is used to update network parameters. The details of the network structure are shown in Tables 2, 3.

TABLE 2
www.frontiersin.org

Table 2. CNN parameters for the first and the second feature maps.

TABLE 3
www.frontiersin.org

Table 3. CNN parameters for the third, the fourth, the fifth, and the sixth feature maps.

In each CNN, the parameters, such as the number of layers and size of the kernel, are different for the different classifiers. In general, the closer relative classifiers have more same parameters. The first feature map and the second feature map are basic features to generate the third feature map, the fourth feature map, the fifth feature map, and the sixth feature map, and a convolutional neural network (CNN) with five layers are employed to extract their feature. As listed in Table 2, the CNN consists of three convolutional layers of 64 kernels with the size of 2 × 2, a convolutional layer of 128 kernels with the size of 2 × 2, and a convolutional layer of 256 kernels with the size of 5 × 5. For the remaining four classifiers, we employ a CNN with six convolutional layers for feature extraction, as shown in Table 3. Specifically, there are 64 kernels with the size of 2 × 2 in the first three convolutional layers and 128 kernels with the size of 2 × 2 in the fourth convolutional layer. The last two convolutional layers include 128 kernels with the size of 5 × 5 and 256 kernels with the size of 5 × 5, respectively. Combing all classifiers, we built an ensemble learning model to identify a subject with a single training sample.

3.3. Shared learning

If the feature maps are correlative, we can utilize the knowledge from other weak classifiers to improve the current classifier (Lou et al., 2017). In general, features with higher correlation provides more positive knowledge, which results in the improvement of the weak classifier. So, to achieve the best-shared representations, we compute the similarity among feature maps, which determines the result of shared learning.

3.3.1. Similarity of feature maps

In Zhang et al. (2011a)'s study, a feature similarity index (FSIM) is proposed for the similarity of feature maps.

We compute the FSIM of feature1(image) and feature2(image) to express the correlation of feature 1 and feature 2 of the current image. The FSIM measurement between feature1(image) and feature2(image) are separated into two components, each for PC or GM. The PC value of images can be considered as a dimensionless measure for the significance of a local structure, which is defined in Morrone et al. (1986) and Kovesi (1999). We define 2 features PC and GM to compute the similarity. The PC features PC(image) is computed as follows:

PC(image)=E(image)ϵ+nAn(image)    (4)

Where E(image)=F2(image)+H2(image) and ϵ is a small positive constant, An(image)=en2(image)+on2(image), F(image)=nen(image), and H(image)=non(image), where en(image) and on(image) the even- and odd-symmetric filters on scale n of the image.

The GM feature G(image) is compute as follows:

G(image)=Gh2(image)+Gv2(image)    (5)

Where Gh(image) and Gv(image) describe the horizontal and vertical directions gradient operators on the image, respectively.

We use I1 and I2 to express image1 and image2, the FSIM is computed as follows:

FSIM(I1,I2)=xΩSL(I1,I2)·PCm(I1,I2)xΩPCm(I1,I2)    (6)

Where Ω means the whole image spatial domain.

The SL similarity combines the similarity SPC of PC and the similarity SG of G:

SL(I1,I2)=[SPC(I1,I2)]α·[SG(I1,I2)]β    (7)

where α and β are parameters used to adjust the relative importance of PC and G features, which is computed as follows:

SPC(I1,I2)=2PC(I1)·PC(I2)+T1PC2(I1)+PC2(I2)+T1SG(I1,I2)=2G(I1)·G(I2)+T2G2(I1)+G2(I2)+T2    (8)

where T1 is a positive constant to increase the stability of SPC, T2 is a positive constant depending on the dynamic range of GM values.

The PCm means the maximum value of PC(I1) and PC(I2), which is PCm(I1, I2) = max(PC(I1), PC(I2)).

After the similarity of different features of a single image is computed, we summarize the similarity of feature maps by the average similarity of all images. We use Fj St(Fi, Fj) to express the feature map similarity of feature map i(Fi) and feature map j(Fj), which is computed as follows:

St(Fi,Fj)=FSIM(Ii,Ij)n    (9)

where n is the number of training samples, and Ii and Ij are feature maps in Fi and Fj, respectively, but generate from the same finger-vein image.

3.3.2. Shared learning scheme

We divide the training process of the weak classifier into K steps, and each step contains P epochs. After training the current classifier in each step, the training samples from other feature maps are used to adjust the parameter by shared learning. In this work, we use the feature map similarity St(Fi, Fj) to compute the number of epochs used in shared learning EcFi, Fj. The EcFi, Fj is set as follows:

EcFi,Fj=2P4×St(Fi,Fj)-1    (10)

where Fi is the feature maps used to train the current classifier and Fj expresses other feature maps.

The performance can be improved by sharing knowledge when high feature correlation (Misra et al., 2016). In the training process, train step u is divided into two parts: training the current classifier and shared learning of other classifiers. The shared learning process is followed by the train current classifier process. The network structure is shown in Figure 7. In one train step of CNNi, the data of featurei are used to train CNNi by P epochs, and then the data of featurej are used to train CNNi by EcFi, Fj epochs. After the data of all high correlative features are used to train CNNi by shared learning, CNNi enters the next training step.

FIGURE 7
www.frontiersin.org

Figure 7. Shared learning by high correlative feature maps.

3.4. Learning speed adjustment

During training weak classifiers, the learning speeds of classifiers are different, because the input data are different. When all weak classifiers achieve their best performance, the ensemble model is best. We propose a learning speed adjustment method to control the learning speeds of weak classifiers. The parameter α is used to adjust the learning speed of each classifier. The whole training process is divided into K steps and P epochs in each step. The total number of epochs R is as follows:

R=K*P    (11)

The parameter αi, u is used to adjust the number of epochs dynamically for classifieri of train step u. When the classifier had achieved well performance, the number of epochs will decrease in the next step. The αi, u is computed as follows:

αi,u=11+e-Li,u-1    (12)

Where the Li, u−1 is the loss function value after train step u−1. To ensure the program runs smoothly, a round down is used to ensure the epoch time is an integer and more than 0. So, the number of epochs pi, u for train step u of classifieri is as follows:

pi,u={1,αi,u·P<1αi,u·P,other.    (13)

3.5. Ensemble classifier

There are two problems during ensemble weak classifiers: (1) How to enhance good weak classifiers while weakening poor weak classifiers; (2) How to make all weak classifiers perform best at the same time. The second problem has been discussed in the previous subsection, in this subsection, we discuss the first problem. To tackle the first problem, an access weight can be set to increase the weight of a good classifier and decrease the weight of the pool classifier. The E+ is computed as

Ei,u+=11+exp(-Scorei,u)i=1K11+exp(-Scorei,u)    (14)

where Scorei, u is the close test accuracy of each classifieri after train step u. The E is computed as

Ei,u-=11+exp(-Li,u)i=1K11+exp(-Li,u)    (15)

where Li, u is the loss function value after train step u.The ensemble weight W after train step u is updated as follows:

Wi,u*={|Ei,u+-Ei,u-|i=1K|Ei,u+-Ei,u-|,Ei,u+0.50,Ei,u+<0.5    (16)

where K is the number of classifiers. The ensemble weight W is updated after each training step. Considering the efficiency of the ensemble model, if the classifier plays poor performance, this classifier could not join the ensemble step. In our model, we use the close test of classifiers to measure their efficiency; if the close test scores more than 0.5, this classifier will join the ensemble step and vice versa.

3.6. Structure of training process

Three different approaches are proposed in this article.

3.6.1. Basic approach

The training process of the basic approach is divided into three stages, which are shown as the training part in Figure 1. We extract feature maps as described in Section 3.1, then we train a weak classifier (CNN) as described in Section 3.2, finally, we ensemble all classifiers by vote as described in Section 3.5. Although this approach simply trains the classifier separately, when the ensemble of all classifiers by vote weight during training, both positive contribution and negative effect are considered by employing Equations (14)–(16). To avoid the poor weak classifier play gadfly in the ensemble model, we set a threshold to accept the classifiers and the vote weight by employing Equation (16) in the ensemble model. This structure is shown in Figure 8. The weak classifiers can be trained in parallel.

FIGURE 8
www.frontiersin.org

Figure 8. Training structure of basic approach.

3.6.2. The shared learning method

The shared learning method inherits the basic approach. During training classifiers, other feature maps are used to train classifiers in different ‘levels' in each training step. These ‘levels' correlate by employing Equation (9). The train network structure of shared learning is shown in Figure 9.

FIGURE 9
www.frontiersin.org

Figure 9. Training structure of the shared learning method.

3.6.3. Shared learning with adjust learning speed method

The third method inherits the second one which is shared learning high relative feature maps, in addition, we add the parameter of adjusting learning speed. In this method, the performance of each classifier maintains well after 35 train steps, so the ensemble classifier plays better performance than the other 2 methods proposed before. The parameter of learning speed adjustment by employing Equation (12) to determine the number of epochs of the next train step by employing Equation (13). This parameter is updated at the end of each training step and feedback to train classifiers before the next training step. The train network structure of the shared learning and adjusted learning speed method is shown in Figure 10.

FIGURE 10
www.frontiersin.org

Figure 10. Training structure of the shared learning with adjust learning speed method.

4. Experiments

To estimate the performance of our approach, we carry out experiments on two public finger-vein databases, namely the Hong Kong Polytechnic University (HKPU) (Kumar and Zhou, 2012) and the University Sains Malaysia (FV-USM) (Asaari et al., 2014). In experiments, we show the experimental results of each classifier and objective task, respectively. Also, some existing approaches, such as Das et al. (2019), have been employed for finger-vein identification in comparable experiments. All the experiments have been performed in Python 3.7, with a system configuration of 128Gb RAM, Tesla P40 graphics card, and two processors both with Intel(R) Xeon(R) Gold 5118 @ 2.30GHz and Linux version 5.3.0 operating system.

4.1. Finger-vein database

In this article, we conduct experiments on the HKPU database and FV-USM database.

HKPU database: The HKPU finger-vein image database (Kumar and Zhou, 2012) consists of images from 156 male and female volunteers. It has been acquired between April 2009 and March 2010 using a contactless imaging device at the HKPU campus. It is composed of 3,132 images from 156 subjects, all of them in a BMP format with a resolution of 513 × 256 pixels. In this dataset, about 93% of the subjects are younger than 30 years, and finger-vein images from 105 subjects have been acquired in two separate sessions with a minimum interval of 1 month and a maximum of over 6 months, with an average of 66.8 days. In each session, every subject has provided 6 image samples from the index and middle finger of the left hand. Other 51 subjects have one single session of acquired data. In the experiment, 2,520 images (105 subjects × 2 fingers × 2 sessions × 6 samples) from two separate sessions are employed to test our approach.

FV-USM database: The FV-USM database (Asaari et al., 2014) is from University Sains Malaysia. It consists of left and right-hand index and middle fingers' vein images from 123 subjects. Among them, 83 are male and 40 are female, with an age range of 20 − 52 years. All images have been acquired in two different sessions with six images per finger in every session. There are 2,952 images (123 subjects × 2 fingers × 2 sessions × 6 samples) All images are in gray level BMP format with a resolution of 640 × 480 pixels.

The details of both datasets are described in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Finger-vein databases.

4.2. Experiment setup

We aim at solving the finger-vein SSPP problem, so the first image from the first session is selected for training and the three images from the second session are employed for testing. As a result, there are 210 samples in the training set and 630 samples (1,260 fingers × 6 samples) in the test set for the HKPU database. Similarly, there are 246 training samples and 1,476 samples (246 fingers × 6 samples) for the FV-USM database. We train our model on the training set and compute the identification accuracy on the test set to estimate the performance of the proposed method.

In our article, the six feature maps are generated by three baselines (e.g., LBP, Gabor, and CNN), as shown in Table 1, and their parameters are determined based on setting in existing works (Kang et al., 2015; Qin and El-Yacoubi, 2017; Zhang et al., 2019a). For LBP descriptors, the radius is set to 1 which implies that only the one layer around the center pixel and eight pixels around the center pixel are taken into account for the feature's next action. The second baseline e.g., Gabor filter has six scales, namely 7, 9, 11, 13, 15, and 17, the wavelength λ is determined by π/2, and the directions θ are set to 0o, 45o, 90o, and 135o. For CNN, the network structures and parameters are presented in Tables 2, 3.

4.3. Performance impacted by shared learning

In this section, we carry out experiments to verify whether the shared learning scheme improves identification accuracy. As described in Section 3, we proposed a shared learning scheme to learn the knowledge among different feature maps, as shown in Figure 6. Generally, if two feature maps are more relative, the performances of both feature maps are improved by the shared learning scheme. In our article, the similarity/relativity between two feature maps is computed by Equation (9) and further taken as an input of Equation (10) to compute the iteration number of shared learning. As shown in Table 5, the first feature map is highly related to the second feature map, the third feature map, and the fourth feature map on both datasets based on Equation (9). Then, the numbers of interaction steps for there relative shared learning are determined to 4, 1, and 2 by Equation (10), respectively. To achieve shared learning, we employ the second feature map to fine-tune the CNN trained by the first feature map at in four steps, and the resulting CNN model is further trained based on the third feature map, followed by the fourth feature map. In this way, the remaining classifiers are trained for identification. In experiments, we test the six classifiers on both databases mentioned in Section 4.2. For example, there are 210 samples from 210 fingers in the training set and 1,260 samples in the testing set for the HKPU database, and 492 training samples and 2,952 testing samples for the FV-USM database. The identification accuracies of six classifiers with shared learning are listed in Table 6. Also, the performance of each classifier without shared learning is reported in Table 6 for comparison. From the experimental results, we observe that the performance of all classifiers is significantly improved after shared learning. Specifically, the identification accuracy increases by an average of 2% for all weak classifiers, which implies that learning knowledge from other relative feature maps by our shared learning approach can improve the performance of the current classifier. This may be explained by the following facts. Using different features to train weak classifiers can transfer to form a more complementary representation of the original finger-vein image.

TABLE 5
www.frontiersin.org

Table 5. Performance comparisons of shared learning.

TABLE 6
www.frontiersin.org

Table 6. Performance comparisons of shared learning in different CNNs.

The experiment also shows the relation between relativity and performance, as shown in Table 5. The experimental results show that the higher relativity between the two classifiers brings more improvement in identification accuracy. For example, the first feature map achieves the largest improvement in identification performance by transferring the knowledge of the second feature map which has the highest relativity with the first feature map. On the contrary, less improvement is achieved based on the shared learning between the first feature map and the third feature map because the third feature map shows less similarity to the first feature map. The good performance attributes to the fact that the knowledge is easier to be transferred if multiple feature maps have good relativity.

4.4. Performance impacted by learning speed

As described in the subsection, our approach is proposed by combining six weak classifiers. In general, the learning speed of each classifier is different. Therefore, the object task can achieve the highest identification accuracy only if each classifier achieves optimal performance at the same time. In this section, the experiments are carried out to evaluate how to impact the performance of object tasks by employing Equation (12) and Equation (13) to adjust the learning speed of each classifier. Figures 11A, B show the identification accuracy of six classifiers and our approach to the HKPU database before and after adjusting the learning speed. In Figure 11A, the performance of all approaches is improved when the number of iteration steps is < 30. However, two classifiers achieve significant degradation in identification accuracy after 35 iterations. The reason is that the two classifiers have higher speed than the remaining classifiers so they cannot achieve optimal performance at the same time. Therefore, it is difficult for object tasks to achieve good performance. In contrast, the learning speed of all weak classifiers is adjusted dynamically, they show the best performance after about 50 iterations (as shown in Figure 11B), which brings the improvement of the objection classifier. Therefore, our approach achieves higher identification accuracy after employing Equation (12) and Equation (13) for learning speed adjustment. The experiments on FV-USM (as shown in Figures 11C, D) show consistent trends that all weak classifiers can achieve optimal performance at the same time (about 40 iterations). In addition, observed training curves (Figures 11AD), we see that all approaches show good stability after employing the learning speed adjustment scheme.

FIGURE 11
www.frontiersin.org

Figure 11. Performance of various classifiers with learning speed adjustment. (A) The performance of all our approaches on the HKPU database without adjusting learning speed. (B) The performance of all our approaches+adjust learning speed on the HKPU database. (C) The performance of all our approaches on the FV-USM database without adjusting learning speed. (D) The performance of all our approaches+adjust learning speed on the FV-USM database.

Table 7 lists the identification accuracy of the basic model, basic approach + shared learning, basic approach + speed adjustment, and basic approach + shared learning + speed adjustment on the HKPU database and FV-USM database. The experimental results have shown that the basic model and basic approach + shared learning achieve a significant improvement in identification accuracy after adjusting learning speed by Equation (13). For example, the basic approach + shared learning and basic approach + shared learning + learning speed adjustment achieve 82.60% and 92.10% identification accuracy on HKPU database and FV-USM database, respectively, and improves identification accuracy by about 7% and 1% compared to basic approach + shared learning. The experimental results in Table 7 imply that our learning speed adjustment scheme is effective to improve the performance of object tasks for an ensemble learning problem.

TABLE 7
www.frontiersin.org

Table 7. Performance comparisons of the adjust learning speed method.

4.5. Performance comparisons

In this section, we compare our approach with existing approaches to evaluate the performance of our method in terms of improving identification accuracy. In experiments, state-of-art methods, such as Das et al. (2019), Kumar and Zhou (2012), and Song et al. (2011), are employed for finger-vein identification with a single training sample per finger. As described in the section, we select the first image for training and six images collected in the second session for testing. As a result, there are 210 training samples and 1,260 testing samples for the FV-USM database and 496 training samples and 2,952 testing sample for the FV-USM database. The identification accuracies of various approaches have been listed in Table 8 for comparison. From the experimental results, we can observe that the proposed approach (basic approach + shared learning + learning speed adjustment) outperforms the existing approaches and achieves the highest identification accuracy, e.g., 92.11 and 94.17% on the HKPU database and FV-USM database, respectively. Also, we observed that the hand-crafted approaches (Song et al., 2011; Kumar and Zhou, 2012; Qin et al., 2013), considered our work achieve < 91.00% identification accuracy on both databases. Such a poor performance may be attributed following facts: (1) The handcrafted segmentation-based approaches assume that the cross-sectional profile of a vein pattern shows a valley (Miura et al., 2007) or line-like texture (Miura et al., 2004) and proposed various mathematical models to extract vein patterns. However, the vein pixels create more complex distributions instead of the valley or straight lines and the pixels in the non-vein region also show valley or line-like attributes, so their performance is limited. (2) The handcrafted segmentation-based approaches usually match vein networks stored in testing samples and enrolment samples for identification. Therefore, it is difficult for such a matching scheme to achieve good performance when there are larger variations such as rotation, scaling, and translation between two images. Compared to handcrafted segmentation-based approaches, the deep learning-based approach (Qin and El-Yacoubi, 2017) is capable of extracting robust vein networks from a law image because it harnesses rich prior knowledge from huge training samples without any assumption. Therefore, the deep learning-based approach (Qin and El-Yacoubi, 2017) achieves better performance, e.g., 91.75 and 92.29%, identification accuracies on both databases. Similarly, its matching scheme is not robust for image samples with large rotation and translation variations. As a solution, a convolutional neural network (Das et al., 2019) is proposed to extract high-level features instead of a vein network by representation learning that is objectively related to vein identification and achieves promising performance. However, it does not perform well for the SSPP problem. This is explained by the following facts. Deep learning-based approaches generally require a large number of training samples to estimate a usually huge number of deep network parameters. In the SSPP configuration, there is only one training sample for each class, so the learning capacity becomes weak and subject to string overfitting, which leads to low identification performance. Ensemble learning is a solution to the SSPP problem because it can exploit the knowledge from different feature maps to improve identification accuracy. Therefore, our basic approach + shared learning achieves 85.63 and 92.31% accuracies on both datasets, which are further improved to 92.11 and 94.17% by adjusting the learning speed of all weak classifiers. Such a good performance may be attributed to these facts. Each classifier includes different discriminate features. As shown in Table 5, the relative feature maps can make a positive contribution to classification, so the knowledge related to classification is transferred among feature maps by shared learning, which effectively improves the feature representation capacity of object tasks. Meanwhile, the learning speed adjustment ensures each classifier achieves the best performance at the same time so that the object task performs the best representation for vein identification.

TABLE 8
www.frontiersin.org

Table 8. Performance comparisons with state-of-art methods.

5. Conclusion

In this article, the authors proposed a new deep ensemble learning approach for SSPP finger-vein recognition. Multiple improvements have been made in this work. A schema of generating multiple feature maps from a single training image is proposed to enhance the performance of the training section. A shared learning schema is applied to the classifier training section. A new learning speed adjustment approach is proposed to improve the performance of the weak classifiers. With a solid comparative experiment, it is very convincing that the proposed method is an outstanding solution for the SSPP finger-vein recognition problem. First, we generate feature maps from training images. Second, we propose a shared learning scheme during the training of weak classifiers. Third, an approach is proposed to adjust the learning speed of weak classifiers. In the experimental part, we compare the performance of our method with the state-of-art method. The final result of the whole model can achieve state-of-the-art recognition results.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found at: The Hong Kong Polytechnic University Finger Image Database, Version 1.0: http://www4.comp.polyu.edu.hk/~csajaykr/fvdatabase.htm Vein USM (FV-USM) Database: drfendi.com/fv_usm_database/.

Author contributions

CL contributed to the conception of the study, performed the experiment, and wrote the manuscript. HQ performed the data analyses and wrote the manuscript. QS contributed significantly to analysis and manuscript preparation. HY performed the experiment. FL helped significantly to check manuscript preparation. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the National Natural Science Foundation of China under (Grant 61976030), the Scientific and Technological Research Program of Chongqing Municipal Education Commission under (Grants KJQN202000841 and KJQN201900848), the funds for creative research groups of Chongqing Municipal Education Commission under (Grant CXQT21034), Chongqing Talent Program under (Grant CQYC201903246), the Fellowship of China Post-Doctoral Science Foundation under (Grant 59676651E), the Science Fund for Creative Research Groups of Chongqing Universities under (Grant CXQT21034), the Chongqing Technology and Business University Research Funds (413/950322019 and 413/1952037), and the Scientific Innovation 2030 Major Project for New Generation of AI under (Grant No. 2020AAA0107300).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Albrecht, T., Lüthi, M., Vetter, T., Chen, H., and Houmani, N. (2009). Encyclopedia of Biometrics. Springer US.

Google Scholar

Asaari, M. S. M., Suandi, S. A., and Rosdi, B. A. (2014). Fusion of band limited phase only correlation and width centroid contour distance for finger based biometrics. Expert. Syst. Appl. 41, 3367–3382. doi: 10.1016/j.eswa.2013.11.033

CrossRef Full Text | Google Scholar

Avci, A., Kocakulak, M., and Acir, N. (2019). “Convolutional neural network designs for finger-vein-based biometric identification,” in 2019 11th International Conference on Electrical and Electronics Engineering (ELECO) (Bursa), 580–584.

Google Scholar

Ban, Y., Wang, Y., Liu, S., Yang, B., Liu, M., Yin, L., et al. (2022). 2D/3D multimode medical image alignment based on spatial histograms. Appl. Sci. 12, 13. doi: 10.3390/app12168261

CrossRef Full Text | Google Scholar

Bhatt, H. S., Singh, R., Vatsa, M., and Ratha, N. K. (2014). Improving cross-resolution face matching using ensemble-based co-transfer learning. IEEE Trans. Image Process. 23, 5654–5669. doi: 10.1109/TIP.2014.2362658

PubMed Abstract | CrossRef Full Text | Google Scholar

Chatterjee, A., Bhatia, V., and Prakash, S. (2017). Anti-spoof touchless 3d fingerprint recognition system using single shot fringe projection and biospeckle analysis. Opt. Lasers Eng. 95, 1–7. doi: 10.1016/j.optlaseng.2017.03.007

CrossRef Full Text | Google Scholar

Cho, S. R., Park, Y. H., Nam, G. P., Shin, K. Y., Lee, H. C., Park, K. R., et al. (2012). Enhancement of finger-vein image by vein line tracking and adaptive gabor filtering for finger-vein recognition. Appl. Mech. Mater. 145, 219–223. doi: 10.4028/www.scientific.net/AMM.145.219

CrossRef Full Text | Google Scholar

Das, R., Piciucco, E., Maiorana, E., and Campisi, P. (2019). Convolutional neural network for finger-vein-based biometric identification. IEEE Trans. Inf. Forensics Security 14, 360–373. doi: 10.1109/TIFS.2018.2850320

CrossRef Full Text | Google Scholar

Ding, C., and Tao, D. (2018). Trunk-branch ensemble convolutional neural networks for video-based face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40, 1002–1014. doi: 10.1109/TPAMI.2017.2700390

PubMed Abstract | CrossRef Full Text | Google Scholar

Dong, L., Yang, G., Yin, Y., Liu, F., and Xi, X. (2014). “Finger vein verification based on a personalized best patches map,” in IEEE International Joint Conference on Biometrics (Clearwater, FL: IEEE), 1–8.

Google Scholar

Felzenszwalb, P. F., Girshick, R. B., Mcallester, D., and Ramanan, D. (2010). Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 32, 1627–1645. doi: 10.1109/TPAMI.2009.167

PubMed Abstract | CrossRef Full Text | Google Scholar

Gumusbas, D., Yildirim, T., Kocakulak, M., and Acir, N. (2019). “Capsule network for finger-vein-based biometric identification,” in 2019 IEEE Symposium Series on Computational Intelligence (SSCI) (Xiamen: IEEE), 437–441.

Google Scholar

Hou, B., Zhang, H., and Yan, R. (2022). Finger-vein biometric recognition: a review. IEEE Trans. Instrum. Meas. 71, 1–26. doi: 10.1109/TIM.2022.3200087

CrossRef Full Text | Google Scholar

Jethanandani, M., Sharma, A., Perumal, T., and Chang, J.-R. (2020). Multi-label classification based ensemble learning for human activity recognition in smart home. Internet Things 12, 100324. doi: 10.1016/j.iot.2020.100324

CrossRef Full Text | Google Scholar

Joardar, S., Chatterjee, A., Bandyopadhyay, S., and Maulik, U. (2017). Multi-size patch based collaborative representation for palm dorsa vein pattern recognition by enhanced ensemble learning with modified interactive artificial bee colony algorithm. Eng. Appl. Artif. Intell. 60, 151–163. doi: 10.1016/j.engappai.2017.02.002

CrossRef Full Text | Google Scholar

Kang, W., Chen, X., and Wu, Q. (2015). The biometric recognition on contactless multi-spectrum finger images. Infrared Phys. Technol. 68, 19–27. doi: 10.1016/j.infrared.2014.10.007

CrossRef Full Text | Google Scholar

Khan, S. H., Ali Akbar, M., Shahzad, F., Farooq, M., and Khan, Z. (2015). Secure biometric template generation for multi-factor authentication. Pattern Recognit. 48, 458–472. doi: 10.1016/j.patcog.2014.08.024

CrossRef Full Text | Google Scholar

Kirchgasser, S., Kauba, C., Lai, Y.-L., Zhe, J., and Uhl, A. (2020). Finger vein template protection based on alignment-robust feature description and index-of-maximum hashing. IEEE Trans. Biometr. Behav. Identity Sci. 2, 337–349. doi: 10.1109/TBIOM.2020.2981673

CrossRef Full Text | Google Scholar

Kocher, D., Schwarz, S., and Uhl, A. (2016). “Empirical evaluation of lbp-extension features for finger vein spoofing detection,” in 2016 International Conference of the Biometrics Special Interest Group (BIOSIG) (Darmstadt), 1–5.

Google Scholar

Kovesi, P. (1999). Image features from phase congruency. videre: J. Comp. Vis. Res. 1, 1–26.

Google Scholar

Kumar, A., and Zhou, Y. (2012). Human identification using finger images. IEEE Trans. Image Process. 21, 2228–2244. doi: 10.1109/TIP.2011.2171697

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuzu, R. S., Piciucco, E., Maiorana, E., and Campisi, P. (2020). On-the-fly finger-vein-based biometric recognition using deep neural networks. IEEE Trans. Inf. Forensics Security 15, 2641–2654. doi: 10.1109/TIFS.2020.2971144

CrossRef Full Text | Google Scholar

Lee, E. C., Jung, H., and Kim, D. (2011). New finger biometric method using near infrared imaging. Sensors 11, 2319–2333. doi: 10.3390/s110302319

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, H. C., Kang, B. J., Lee, E. C., and Park, K. R. (2010). Finger vein recognition using weighted local binary pattern code based on a support vector machine. J. Zhejiang Univer. Sci. Comput. Electron. Comput. Electron. 11, 514–524. doi: 10.1631/jzus.C0910550

CrossRef Full Text | Google Scholar

Li, J., Xu, K., Chaudhuri, S., Yumer, E., Zhang, H., and Guibas, L. (2017). Grass: Generative recursive autoencoders for shape structures. ACM Trans. Graph. 36, 52. doi: 10.1145/3072959.3073637

CrossRef Full Text | Google Scholar

Li, Q., Yang, Z., Luo, L., Wang, L., Zhang, Y., Lin, H., et al. (2018). “A multi-task learning based approach to biomedical entity relation extraction,” in 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) (Madrid: IEEE), 680–682.

PubMed Abstract | Google Scholar

Liskowski, P., and Krawiec, K. (2016). Segmenting retinal blood vessels with deep neural networks. IEEE Trans. Med. Imaging 35, 2369–2380. doi: 10.1109/TMI.2016.2546227

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, C., and Kim, Y. H. (2016). “An efficient finger-vein extraction algorithm based on random forest regression with efficient local binary patterns,” in 2016 IEEE International Conference on Image Processing (ICIP) (Phoenix, AZ: IEEE).

Google Scholar

Liu, C., Qin, H., Yang, G., Shen, Z., and Wang, J. (2022). Ensemble Deep Learning Based Single Finger-Vein Recognition. Singapore: Springer.

Google Scholar

Liu, F., Yin, Y., Yang, G., Dong, L., and Xi, X. (2014). “Finger vein recognition with superpixel-based features,” in IEEE International Joint Conference on Biometrics (Clearwater, FL: IEEE), 1–8.

Google Scholar

Liu, H., Liu, M., Li, D., Zheng, W., Yin, L., and Wang, R. (2022). Recent advances in pulse-coupled neural networks with applications in image processing. Electronics 11, 264. doi: 10.3390/electronics11203264

CrossRef Full Text | Google Scholar

Liu, Z., Zhang, M., Liu, F., and Zhang, B. (2021). Multidimensional feature fusion and ensemble learning-based fault diagnosis for the braking system of heavy-haul train. IEEE Trans. Ind. Inform. 17, 41–51. doi: 10.1109/TII.2020.2979467

CrossRef Full Text | Google Scholar

Lou, Y., Fu, G., Jiang, Z., Men, A., and Zhou, Y. (2017). “Improve object detection via a multi-feature and multi-task cnn model,” in 2017 IEEE Visual Communications and Image Processing (VCIP) (St. Petersburg, FL: IEEE), 1–4.

Google Scholar

Lu, H., Zhu, Y., Yin, M., Yin, G., and Xie, L. (2022). Multimodal fusion convolutional neural network with cross-attention mechanism for internal defect detection of magnetic tile. IEEE Access 10, 60876–60886. doi: 10.1109/ACCESS.2022.3180725

CrossRef Full Text | Google Scholar

Misra, I., Shrivastava, A., Gupta, A., and Hebert, M. (2016). “Cross-stitch networks for multi-task learning,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Las Vegas, NV: IEEE).

Google Scholar

Miura, N., Nagasaka, A., and Miyatake, T. (2004). Feature extraction of finger-vein patterns based on repeated line tracking and its application to personal identification. Mach. Vis. Appl. 15, 194–203. doi: 10.1007/s00138-004-0149-2

CrossRef Full Text | Google Scholar

Miura, N., Nagasaka, A., and Miyatake, T. (2007). Extraction of finger-vein patterns using maximum curvature points in image profiles. ICE Trans. Inf. Syst. 90, 1185–1194. doi: 10.1093/ietisy/e90-d.8.1185

CrossRef Full Text | Google Scholar

Morrone, M. C., Ross, J., Burr, D. C., and Owens, R. (1986). Mach bands are phase dependent. Nature 324, 250–253. doi: 10.1038/324250a0

CrossRef Full Text | Google Scholar

Muthusamy, D., and Rakkimuthu, P. (2022). Trilateral filterative hermitian feature transformed deep perceptive fuzzy neural network for finger vein verification. Expert. Syst. Appl. 196, 116678. doi: 10.1016/j.eswa.2022.116678

CrossRef Full Text | Google Scholar

Qin, H., El Yacoubi, M. A., Lin, J., and Liu, B. (2019). An iterative deep neural network for hand-vein verification. IEEE Access 7, 34823–34837. doi: 10.1109/ACCESS.2019.2901335

CrossRef Full Text | Google Scholar

Qin, H., and El-Yacoubi, M. A. (2017). Deep representation-based feature extraction and recovering for finger-vein verification. IEEE Trans. Inf. Forensics Security 12, 1816–1829. doi: 10.1109/TIFS.2017.2689724

CrossRef Full Text | Google Scholar

Qin, H., and El-Yacoubi, M. A. (2018). Deep representation for finger-vein image-quality assessment. IEEE Trans. Circ. Syst. Video Technol. 28, 1677–1693. doi: 10.1109/TCSVT.2017.2684826

CrossRef Full Text | Google Scholar

Qin, H., Qin, L., Xue, L., He, X., Yu, C., and Liang, X. (2013). Finger-vein verification based on multi-features fusion. Sensors 13, 15048–15067. doi: 10.3390/s131115048

PubMed Abstract | CrossRef Full Text | Google Scholar

Qin, H., Qin, L., and Yu, C. (2011). Region growth-based feature extraction method for finger-vein recognition. Optical Eng. 50, 214–229. doi: 10.1117/1.3572129

CrossRef Full Text | Google Scholar

Qin, H., and Yacoubi, M. A. E. (2015). “Finger-vein quality assessment by representation learning from binary images,” in International Conference on Neural Information Processing (Istanbul).

Google Scholar

Qin, X., Ban, Y., Wu, P., Yang, B., Liu, S., Yin, L., et al. (2022). Improved image fusion method based on sparse decomposition. Electronics 11, 321. doi: 10.3390/electronics11152321

CrossRef Full Text | Google Scholar

Qiu, S., Liu, Y., Zhou, Y., Huang, J., and Nie, Y. (2016). Finger-vein recognition based on dual-sliding window localization and pseudo-elliptical transformer. Expert Syst Appl. 64, 618–632. doi: 10.1016/j.eswa.2016.08.031

CrossRef Full Text | Google Scholar

Rosdi, B. A., Chai, W. S., and Suandi, S. A. (2011). Finger vein recognition using local line binary pattern. Sensors 11, 11357–11371. doi: 10.3390/s111211357

PubMed Abstract | CrossRef Full Text | Google Scholar

Sagi, O., and Rokach, L. (2018). Ensemble learning: a survey. Wiley Interdisc. Rev. Data Min. Knowl. Discov. 8, e1249. doi: 10.1002/widm.1249

CrossRef Full Text | Google Scholar

Shaheed, K., Mao, A., Qureshi, I., Kumar, M., Hussain, S., Ullah, I., et al. (2022a). Ds-cnn: A pre-trained xception model based on depth-wise separable convolutional neural network for finger vein recognition. Expert. Syst. Appl. 191, 116288. doi: 10.1016/j.eswa.2021.116288

CrossRef Full Text | Google Scholar

Shaheed, K., Mao, A., Qureshi, I., Kumar, M., Hussain, S., and Zhang, X. (2022b). Recent advancements in finger vein recognition technology: methodology, challenges and opportunities. Inf. Fusion 79, 84–109. doi: 10.1016/j.inffus.2021.10.004

CrossRef Full Text | Google Scholar

Shao, H., and Zhong, D. (2021). One-shot cross-dataset palmprint recognition via adversarial domain adaptation. Neurocomputing 432, 288–299. doi: 10.1016/j.neucom.2020.12.072

CrossRef Full Text | Google Scholar

She, Q., Hu, R., Xu, J., Liu, M., Xu, K., and Huang, H. (2022). Learning high-dof reaching-and-grasping via dynamic representation of gripper-object Interaction. ACM Trans. Graphics 41, 1–14. doi: 10.1145/3528223.3530091

CrossRef Full Text | Google Scholar

Song, W., Kim, T., Kim, H. C., Choi, J. H., Kong, H. J., and Lee, S. R. (2011). A finger-vein verification system using mean curvature. Pattern Recognit. Lett. 32, 1541–1547. doi: 10.1016/j.patrec.2011.04.021

CrossRef Full Text | Google Scholar

Tsai, J., Lin, Y., and Liao, H. M. (2014). Per-cluster ensemble kernel learning for multi-modal image clustering with group-dependent feature selection. IEEE Trans. Multimedia 16, 2229–2241. doi: 10.1109/TMM.2014.2359769

CrossRef Full Text | Google Scholar

Vodinh, T. (2012). Biomedical Photonics Handbook, 2nd Edn. Boca Raton, FL: CRC Press.

Google Scholar

Wang, J., Wang, G., and Zhou, M. (2017). Bimodal vein data mining via cross-selected-domain knowledge transfer. IEEE Trans. Inf. Forensics Security 13, 733–744. doi: 10.1109/TIFS.2017.2766039

CrossRef Full Text | Google Scholar

Wang, J., Xiao, J., Lin, W., and Luo, C. (2015). Discriminative and generative vocabulary tree: with application to vein image authentication and recognition. Image Vis. Comput. 34, 51–62. doi: 10.1016/j.imavis.2014.10.014

CrossRef Full Text | Google Scholar

Wang, L., Li, Y., and Wang, S. (2018). “Feature learning for one-shot face recognition,” in 2018 25th IEEE International Conference on Image Processing (ICIP) (Athens: IEEE), 2386–2390.

Google Scholar

Wu, D., Luo, X., He, Y., and Zhou, M. (2022). A prediction-sampling-based multilayer-structured latent factor model for accurate representation to high-dimensional and sparse data. IEEE Trans. Neural Netw. Learn. Syst. 1–14. doi: 10.1109/TNNLS.2022.3200009

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, J.-D., and Liu, C.-T. (2011a). Finger-vein pattern identification using principal component analysis and the neural network technique. Expert. Syst. Appl. 38, 5423–5427. doi: 10.1016/j.eswa.2010.10.013

CrossRef Full Text | Google Scholar

Wu, J.-D., and Liu, C.-T. (2011b). Finger-vein pattern identification using svm and neural network technique. Expert. Syst. Appl. 38, 14284–14289. doi: 10.1016/j.eswa.2011.05.086

CrossRef Full Text | Google Scholar

Wu, Z., and Deng, W. (2016). “One-shot deep neural network for pose and illumination normalization face recognition,” in 2016 IEEE International Conference on Multimedia and Expo (ICME) (Seattle, WA: IEEE), 1–6.

Google Scholar

Xi, X., Lu, Y., and Yin, Y. (2016). Learning discriminative binary codes for finger vein recognition. Pattern Recognit. 66, 26–33. doi: 10.1016/j.patcog.2016.11.002

CrossRef Full Text | Google Scholar

Xi, X., Yang, G., Yin, Y., and Meng, X. (2013). Finger vein recognition with personalized feature selection. Sensors 13, 11243–11259. doi: 10.3390/s130911243

PubMed Abstract | CrossRef Full Text | Google Scholar

Xi, X., Yang, G., Yin, Y., and Yang, L. (2014). Finger vein recognition based on the hyperinformation feature. Opt. Eng. 53, 013108–013108. doi: 10.1117/1.OE.53.1.013108

PubMed Abstract | CrossRef Full Text | Google Scholar

Xie, L., Zhu, Y., Yin, M., Wang, Z., Ou, D., Zheng, H., et al. (2022). Self-feature-based point cloud registration method with a novel convolutional siamese point net for optical measurement of blade profile. Mech. Syst. Signal Process. 178, 109243. doi: 10.1016/j.ymssp.2022.109243

CrossRef Full Text | Google Scholar

Yang, G., Xi, X., and Yin, Y. (2012). Finger vein recognition based on (2d) 2 pca and metric learning. Biomed. Res. Int. 2012, 324249. doi: 10.1155/2012/324249

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, G., Xiao, R., Yin, Y., and Lu, Y. (2013). Finger vein recognition based on personalized weight maps. Sensors 13, 12093–12112. doi: 10.3390/s130912093

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, J., Yang, J., and Shi, Y. (2009). “Combination of gabor wavelets and circular gabor filter for finger-vein extraction,” in International Conference on Emerging Intelligent Computing Technology and Applications (Ulsan).

Google Scholar

Yang, L., Yang, G., Yin, Y., and Xi, X. (2014). Exploring soft biometric trait with finger vein recognition. Neurocomputing 135, 218–228. doi: 10.1016/j.neucom.2013.12.029

CrossRef Full Text | Google Scholar

Yang, W., Hui, C., Chen, Z., Xue, J., and Liao, Q. (2019). Fv-gan: finger vein representation using generative adversarial networks. IEEE Trans. Inf. Forensics Security 14, 2512–2524. doi: 10.1109/TIFS.2019.2902819

CrossRef Full Text | Google Scholar

Zhang, J., Lu, Z., Li, M., and Wu, H. (2019). Gan-based image augmentation for finger-vein biometric recognition. IEEE Access 7, 183118–183132. doi: 10.1109/ACCESS.2019.2960411

CrossRef Full Text | Google Scholar

Zhang, J., Tang, Y., Wang, H., and Xu, K. (2022). Asro-dio: active subspace random optimization based depth inertial odometry. IEEE Trans. Rob. 1–13. doi: 10.1109/TRO.2022.3208503

CrossRef Full Text | Google Scholar

Zhang, J., Zhu, C., Zheng, L., and Xu, K. (2021). Rosefusion: random optimization for online dense reconstruction under fast camera motion. ACM Trans. Graph. 40, 1–17. doi: 10.1145/3478513.3480500

CrossRef Full Text | Google Scholar

Zhang, L., Zhang, L., Mou, X., and Zhang, D. (2011a). Fsim: A feature similarity index for image quality assessment. IEEE Trans. Image Process. A 20, 2378–2386. doi: 10.1109/TIP.2011.2109730

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, L., Zhang, L., Zhang, D., and Zhu, H. (2011b). Ensemble of local and global information for finger-knuckle-print recognition. Pattern Recognit. 44, 1990–1998. doi: 10.1016/j.patcog.2010.06.007

CrossRef Full Text | Google Scholar

Zhang, Y., Li, W., Zhang, L., and Lu, Y. (2019a). “Adaptive gabor convolutional neural networks for finger-vein recognition,” in 2019 International Conference on High Performance Big Data and Intelligent Systems (HPBD IS) (Shenzhen), 219–222.

Google Scholar

Zhang, Y., Li, W., Zhang, L., Ning, X., Sun, L., and Lu, Y. (2019b). Adaptive learning gabor filter for finger-vein recognition. IEEE Access 7, 159821–159830. doi: 10.1109/ACCESS.2019.2950698

CrossRef Full Text | Google Scholar

Zhao, Z., Jiao, L., Liu, F., Zhao, J., and Chen, P. (2016). Semisupervised discriminant feature learning for sar image category via sparse ensemble. IEEE Trans. Geosci. Remote Sens. 54, 3532–3547. doi: 10.1109/TGRS.2016.2519910

CrossRef Full Text | Google Scholar

Keywords: finger-vein recognition, single sample per person, deep learning, ensemble learning, pattern recognition

Citation: Liu C, Qin H, Song Q, Yan H and Luo F (2023) A deep ensemble learning method for single finger-vein identification. Front. Neurorobot. 16:1065099. doi: 10.3389/fnbot.2022.1065099

Received: 09 October 2022; Accepted: 14 December 2022;
Published: 11 January 2023.

Edited by:

Xiaoyu Shi, Chongqing Institute of Green and Intelligent Technology (CAS), China

Reviewed by:

Yang Chen, Microsoft Research Asia, China
Jie Chen, Northwestern Polytechnical University, China
Huiwu Luo, University of Macau, China

Copyright © 2023 Liu, Qin, Song, Yan and Luo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Huafeng Qin, yes liuchongwen@ctbu.edu.cn; yes qinhuafengfeng@163.com

Download