Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell., 27 January 2026

Sec. Pattern Recognition

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1714882

CBAM-DenseNet with multi-feature quality filtering: advancing accuracy in small-sample iris recognition


Yongheng Pang,
Yongheng Pang1,2*Zishen WangZishen Wang2Nan JiangNan Jiang2Jia QinJia Qin2Suyuan LiSuyuan Li2
  • 1Shanghai Key Laboratory of Forensic Medicine and Key Laboratory of Forensic Science, Ministry of Justice, Shenyang, Liaoning, China
  • 2School of Public Security Information Technology and Intelligence, Criminal Investigation Police University of China, Shenyang, Liaoning, China

In the context of the information age, traditional password and key-based authentication mechanisms are no longer sufficient to meet the growing demands for information security. Iris recognition technology has garnered attention due to its high security and uniqueness. Current iris recognition methods based on single feature extraction are prone to loss of feature information, which affects recognition rates. To address this, this paper proposes a multi-feature fusion-based iris recognition method. The method employs a comprehensive quality evaluation scheme to filter iris images, ensuring the quality of the input images. An improved CAN network is used to effectively remove image noise, and a DenseNet network-based iris feature extraction method is combined with a fusion space and attention mechanism (CBAM) to enhance the expressiveness of features. Through experiments with small sample sizes and testing on various public iris databases, the proposed method has been validated for significant improvements in recognition accuracy and robustness.

1 Introduction

With the rapid advancement of informatization and networking, the demand for information security and personal identity verification in society is growing rapidly, posing an urgent need for high-standard identity authentication mechanisms. Traditional methods of password and key authentication are no longer sufficient to meet current security challenges. Biometric recognition technology, with its efficiency and reliability, has gradually become a hot topic in research and application in the field of identity authentication to meet the increasing security needs. The widespread adoption of biometric recognition technology in identity frameworks across various industries is becoming more prevalent (Gupta, 2023). Iris recognition technology (IR), with its high level of security and difficulty to forge, has become one of the most watched and researched technologies in recent years (He and Li, 2024). The iris is the annular pigmented membrane located between the pupil and the sclera of the eyeball, which appears to have a radially textured pattern from the inside out, along with interspersed spot textures (Kim et al., 2023). Through extensive analysis of the anatomical structure of the iris by ophthalmologists and anatomists, it has been found that the iris, similar to fingerprints, possesses a unique individuality that is distinct from others, and even identical twins have different irises (Wei et al., 2022). Based on the inherent structure of the iris, it offers more accurate and reliable advantages compared to other biometric features such as facial, fingerprint, voice, gait, and vein recognition (Zambrano et al., 2022). Therefore, iris recognition has become a hot topic in biometric recognition research in recent years and is one of the most promising identification technologies. The first step in iris recognition is to evaluate the quality of the collected iris images and select qualified images. Daugman (2001) judged the clarity of the image based on the high-frequency energy of the 2D Fourier of the iris. Wildes (1997) evaluated the roughness of the image by measuring the grayscale changes at the contour of the iris and sclera. However, a unidimensional iris image quality evaluation method obviously cannot meet the application standards of iris quality evaluation in practical applications. Therefore, (Gao et al., 2015) adopted a coarse and fine combined indicator, based on Support Vector Machine (SVM) to achieve the fusion of image quality indicators, and used this to distinguish images of different qualities; (Wang et al., 2020) proposed a recognition-oriented quality assessment method, using the distance of the image embedded in the feature space as a quality indicator, and predicted based on a deep neural network with an attention mechanism. This method significantly improved the performance of the recognition algorithm. Iris localization is also key to iris recognition. In the early days, (Daugman, 2003) used a circular template matching method to achieve the localization of the iris in the human eye. With the development of computer vision theory, (Wildes, 1997) proposed using Hough transform to outline the iris contour boundary to achieve iris localization in the human eye, which has been used to this day. Subsequently, with the continuous development of machine learning and deep learning theories, (Ishikawa, 2004) proposed an active appearance model-based eye localization algorithm, which determines the position of the iris in the human eye by describing the entire external structure of the face. Jan et al. (2021) proposed an iris localization algorithm based on four steps: preprocessing, marking rough pupil area, extracting rough eye contour, and refining eye contour, which effectively localized the iris area and removed noise. With the development of deep learning, significant breakthroughs have been made in both application and theoretical aspects of iris recognition technology over the past few decades. Currently, mainstream feature extraction algorithms for iris feature extraction and matching are based on the description of iris texture. Boles utilized an algorithm based on wavelet transform and zero-crossing point detection for feature extraction of irises to be matched, and encoded by the average of the integral between adjacent zero-crossing points to obtain the corresponding iris feature template (Boles and Boashash, 1998). Wildes et al. (1994) to obtain iris texture images at different scales as features, constructing iris feature templates. However, traditional single-feature iris features have the problem of insufficient feature extraction and low accuracy. With the development of deep learning, deep learning-based iris feature extraction can extract iris features very well. Gangwar and Joshi (2016) proposed a DeepIrisNet neural network, which has a deeper network structure and can be well applied to large-scale iris datasets, and based on the simulation of iris microstructure, it achieves high recognition accuracy. Al-Waisy et al. (2018) proposed a deep learning neural network to distinguish between left and right irises, and the features learned by this neural network model are fused based on the hierarchical fusion method, which improves the accuracy and speed of iris recognition. A bit-parallel event matching framework leveraging AVX-512 vectorization has demonstrated over 35 × speedup by encoding event streams into time-sliced bit sequences and exploiting SIMD parallelism for complex temporal pattern evaluation (Qiu et al., 2025). However, in the specific practical application of iris recognition systems, existing iris recognition systems still face many adjustments in key links such as iris image quality evaluation, iris image preprocessing, and iris image feature extraction (Tan et al., 2012; Kaur et al., 2010; Ahmadi and Akbarizadeh, 2018; Zhou et al., 2020). The existing research difficulties mainly include how to ensure the improvement of iris recognition model performance, how to build efficient and universal recognition models based on different iris databases, and how to adjust and transform the internal structure of neural network recognition models to adapt to the specific needs of iris recognition, and other series of problems. Therefore, in response to the above problems, this paper proposes a multi-dimensional feature fusion-based iris recognition method framework, and the main contributions are as follows:

1. A method for iris image quality evaluation and screening is proposed. This includes five processes: liveness detection of iris images, clarity evaluation, squint detection, annular area detection, and clarity evaluation of the annular area, and the evaluation results of each part are input into the SA-SVM classifier for quality screening.

2. A method for preprocessing of iris images is proposed. Through four steps of SE-CAN network noise reduction, iris image localization, normalization and ROI area selection, and iris image enhancement, the iris image is preprocessed to ensure accurate extraction of iris features.

3. A multi-dimensional feature fusion-based iris recognition method framework is proposed, covering the entire process of image evaluation screening, image preprocessing, feature extraction and multi-dimensional feature fusion, and identity recognition. It has been verified on public datasets such as CASIA v4.0, NICE1.0, and JLU-6.0, confirming the effectiveness, robustness, and generalization ability of the method.

The subsequent research work of this paper is arranged as follows: In the second section, the related indicators of iris image quality evaluation and preprocessing methods are proposed, and the evaluation results are input into the SA-SVM classifier for screening. In the third section, the strategy plan for iris image preprocessing is given. In the fourth section, the iris feature extraction method based on CBAM-DenseNet is introduced, and the multi-feature fusion strategy is designed. In the fifth section, the classification effect of the SA-SVM classifier and the fusion effect of the multi-feature fusion scheme are analyzed on multiple datasets. The proposed CBAM-DenseNet is compared with multiple networks in the experimental comparison, and the recognition effect of the network is evaluated. In the sixth part, a summary of the iris recognition process proposed in this paper is given.

2 Iris image quality evaluation and screening

2.1 Iris liveness detection

The dynamic characteristics of the natural constriction and dilation of the pupil provide a theoretical basis for liveness detection. This study employs linear arithmetic subtraction of iris images to verify the presence of dynamic expression within a sequence of iris images. By selecting the same visual organ from a specific individual and comparing differences in pupil coordinates and radius parameters under different intensities of illumination, the dynamic changes between adjacent images are assessed to achieve detection of iris vitality.

2.2 Iris image clarity assessment

The use of the Benner function calculates the overall clarity value of an iris image. The specific formula is as follows:

F=xy{f(x+2,y)-f(x,y)}2    (1)

Accumulate the gradient values of all pixel points in the image to obtain the evaluation value F Where x, y are the grayscale values of adjacent pixels in two directions of the image. This evaluation value is used to screen iris images that meet quality standards. When F exceeds the preset threshold, the image is determined to be clearly captured; otherwise, it is determined to be a defocused and blurred iris image.

2.3 Iris image gaze detection

Calculate the ratio of the distance between the center of the iris pupil and the center of the iris image to the overall diagonal length of the iris image to reflect the degree of gaze deviation of the iris image. The detailed calculation method is as follows:

L=(x0-x1)2+(y0-y1)2m2+n2    (2)

The ratio L reflects the degree of gaze deviation in the iris image, (x0, y0) represents the center position of the pupil in the iris image, (x1, y1) represents the center point position of the iris image, and m and n denote the length and width of the evaluated iris image, respectively. The greater the distance between the pupil center and the center of the iris image, the larger the corresponding ratio L, indicating a higher degree of iris deviation.

2.4 Iris image circular region detection

Iris image annular area detection is divided into two key parts: the detection of the area ratio of the iris annular region and the detection of the closure obstruction degree of the iris image. The area ratio of the iris annular region is represented by the ratio of the area of the pupil region in the iris image to the area ratio of the iris outer edge contour, reflecting the area ratio of the iris annular region. The relevant calculations are as follows:

t1=1-Δt=1-12    (3)

Where ∂1 represents the area of the outer circular contour of the iris, ∂2 represents the size of the pupil area in the iris image, and t1 represents the quality evaluation coefficient. The detection of closure obstruction in iris images is primarily based on the assessment of the integrity of the pupil in the iris image. The formula is as follows:

t2=β1β2    (4)

Where β1 represents the total number of pixel points in the pupil area of the iris image, β2 represents the number of pixels with a grayscale value of 0 in the pupil area of the iris image, and t2 represents the degree of occlusion of the human eye open and close in the iris image. By performing the detection of the iris annular area ratio and the degree of eye closure occlusion on the iris image in sequence, the corresponding iris image quality evaluation coefficients t1 and t2 are obtained. These two evaluation coefficients are given different weight parameters u1 and u2 to be added and merged as the comprehensive quality evaluation index coefficient Tr of the iris annular area, see formula:

Tr=u1×t1+u2×t2    (5)

2.5 Iris image annular region clarity evaluation

Firstly, the clarity of the annular region of the normalized iris image is evaluated using the Benner (Equation 1) function-based image clarity evaluation function, obtaining the clarity function value f1. Secondly, the normalized iris image is processed for ROI (Region of Interest) extraction, denoted as f2, and the clarity of the extracted iris image ROI is evaluated using the Benner function, yielding the clarity function value. Finally, these two evaluation coefficients are weighted differently by parameters ω1 and ω2 and summed to form the clarity evaluation index coefficient for the iris annular region, denoted as Fh. The specific calculation is as follows:

Fh=ω1×f1+ω2×f2    (6)

2.6 Iris image quality assessment based on simulated annealing-optimized support vector machine

Integrate assessments of the iris image's off-angle, annular region, and ROI to obtain corresponding iris image quality evaluation metrics. Fusion of related iris quality rating coefficients establishes the iris quality evaluation vector {L, T, Fh}. To reduce the impact of magnitude differences on classification accuracy and to enhance the speed of model training, standardize the image quality evaluation vector to obtain the standardized sample vector {L,T,Fh}. Input the obtained sample vector into the SA-SVM classifier for classification processing. Select appropriate kernel functions and optimal balance coefficients to improve classification accuracy. The Gaussian kernel is particularly suitable for iris image data with a small number of samples and low-dimensional sample vectors due to its mapping advantages in feature space. Furthermore, parameter optimization is performed using the Simulated Annealing (SA) algorithm to avoid falling into local optima and enhance the generalization capability of the SVM. This method can improve the performance of the SVM classifier and achieve better generalization under suitable penalty coefficients and kernel parameters. The specific process is depicted in Figure 1.

Figure 1
Flowchart illustrating a process starting with “Start”, initializing variables x0 and K. It generates xi and computes E(xi), checks if E(xi) is less than E(x0). If yes, it updates V to xi; if not, follows “Metropolis Accept”. Checks if K is reached, and if not, goes back. If convergence is met, it outputs x*, otherwise cools temperature and resets K. Ends with “END”.

Figure 1. Flowchart of parameter optimization with the SA algorithm.

3 Iris image preprocessing

3.1 Iris image denoising based on the SE-CAN network

The CAN network is a fully convolutional neural network designed for learning and training variable resolution images, represented by multiple consecutive layers: {I0, I1, I2, …, Ii, …, Ij}. Here, I0 and Ij represent the input and output layers of the network, respectively, with the image dimensions at the input and output layers being the same. In the CAN network architecture, the spatial dimensions of each intermediate layer Ii{0 < i < j} are m × n × w, where m × n is the resolution size of the image, and w is the number of mapped features. The image input to the network, after being processed through calculations, is passed to the next layer. The content transmitted to intermediate layer Ii is obtained by computing the output of layer Ii−1. For specific computations, see the formula:

Iia=Φ(Ψi(gia+b=1j-1Ii-1braKija))    (7)
Ψi(x)=λix+μiBN(X)    (8)

In the given formulas, Iia and Ii-1b represent the feature maps a and b of the intermediate layer Ii and the previous layer Ii−1, respectively. gia denotes the offset scalar, and Kija represents a 3x3 convolution kernel. The asterisk ⊕ra signifies a dilated convolution. ψi indicates an adaptive normalization function, where λi and μi are learnable weights that are adjusted through backpropagation. ϕ denotes a nonlinear activation function operation. The processing procedure for bilateral filtering based on the CAN network includes: first, applying a bilateral filter to the training image samples and storing the results. Second, randomly extracting patches from the original sample images and those processed by the bilateral filter, and extracting data from the random patches to input into the network for training. Through the preset multi-scale CAN layers, the loss between the images processed by conventional filters and those processed by the CAN network is calculated. Finally, the CAN network is trained to approximate the bilateral filter operator based on the loss function. To enhance the correlation of feature information between different channels, a channel attention mechanism Squeeze-and-Excitation(SE) is introduced to the existing CAN network model, as shown in Figure 2.

Figure 2
Illustration of a data transformation process. On the left, a block labeled \( X_1 \) with height \( H_1 \), channels \( C_1 \), and width \( w_1 \) undergoes transformation to form \( U_1 \) with height \( H \) and channels \( C \). The middle section shows compression and extraction resulting in a vector. On the right, the weighted result \( X \) with height \( H \), channels \( C \), and width \( w \). Arrows indicate the transformation and weighting steps.

Figure 2. Workflow of the channel attention mechanism.

The integration of the channel attention mechanism forms the SE-CAN network, which ensures the preservation of important image features during processing and enhances the quality of reconstructed images after noise processing. Details of the SE-CAN network are shown in Figure 3.

Figure 3
Diagram of a neural network architecture featuring a CAN module connected to a scale unit. The CAN outputs to a Global Pooling layer, feeding into an Input Layer, followed by a Hidden Layer, and then an Output Layer. An SE-Net module also influences the network.

Figure 3. Structure of the improved SE-CAN network.

The original image input to the convolutional layer is processed to generate feature map X1, which is transformed to obtain a new feature map U1 with a size of H × W. U1 contains C channels, and the importance of the images output by each channel varies, which may affect the learning and training effectiveness of the image. An attention mechanism module is introduced to perform global average pooling on each channel, and each channel obtains a scalar. Based on the channel attention mechanism, it is processed through two fully connected layers to obtain a scalar between 0 and 1 as the weight of that channel. Each element of the original image H × W is multiplied by the corresponding channel weight to obtain a new feature image X. Compared with the original network structure, the introduction of the channel attention mechanism module can effectively improve the network performance.

3.2 Iris image localization

To address the common noise types present in real-world iris images, such as Gaussian noise and salt-and-pepper noise, we propose an enhanced SE-CAN network. The improvement lies in the integration of a SE block within the CAN framework, which adaptively recalibrates channel-wise feature responses by explicitly modeling inter-channel dependencies. This attention mechanism allows the network to emphasize informative channels while suppressing those corrupted by noise, thereby enhancing the robustness of the extracted features. Specifically, the SE block performs a global averaging operation on the feature maps to generate a channel descriptor, followed by two fully connected layers that learn a nonlinear transformation to produce channel-wise weights. These weights are then applied to rescale the original feature maps, effectively refining the representation for subsequent processing stages. The iris image localization algorithm aims to identify the inner and outer edges of the iris, integrating morphological and Hough circle transform methods. The specific steps are as follows:

• Convert the iris image to a grayscale image and calculate its grayscale histogram. By analyzing the peaks and valleys of the histogram, determine the threshold for binarization to reduce interference from low grayscale values in the pupil area.

• Perform morphological circular closing operations on the iris image to remove interference such as eyelashes and eyelids, achieving noise reduction processing.

• Binarize the denoised iris image, use the Canny operator to perform edge detection on the image to extract the edge information of the pupil, and use the Hough circle transform to determine the radius size of the inner edge contour of the iris.

• Perform ROI cropping and binarization on the original image, set the outer edge threshold, and enhance the recognition of the outer edge contour through morphological circular closing operations.

• Integrate the above steps to accurately locate the iris, extract the inner and outer edge contour information, and provide an accurate iris area for subsequent image processing.

3.3 Iris image normalization and ROI area selection

Iris image normalization typically employs an elastic rubber circle model for iris normalization expansion. The steps are as follows:

• Utilize iris localization techniques to determine boundary information.

• Transform the iris boundary into a fixed-size rectangular area through polar coordinate mapping and radial transformation. Select areas that are rich in texture features and less affected by noise density.

• Determine the size and position of the ROI area based on the type of iris data.

3.4 Enhancement of iris images

Iris image enhancement typically utilizes histogram equalization technology, which involves converting the iris image to a grayscale image, computing its histogram, generating and normalizing the cumulative histogram, and applying a histogram equalization function to enhance contrast and details.

4 Iris feature extraction and fusion

4.1 Iris feature extraction based on CBAM-DenseNet

Four different types of traditional single feature extraction algorithms (multi-channel Gabor filter algorithm, GLCM algorithm, Haar wavelet transform, and LBP algorithm) are used for feature extraction and classification. To effectively address the challenges of overfitting and poor generalization inherent in small-sample learning, our method incorporates three key strategies. First, the DenseNet architecture is inherently suitable for limited data due to its dense connectivity, which promotes feature reuse and mitigates the vanishing gradient problem, thereby enhancing model stability. Second, Dropout layers are strategically placed after each “BN-ReLU-Conv” block within the DenseBlocks and Transition-layers to serve as a regularization mechanism, randomly deactivating neurons during training to prevent co-adaptation. Third, systematic data augmentation is applied to the Region of Interest (ROI) images prior to training, including operations such as blurring, brightness adjustment, rotation, contrast variation, and flipping, to artificially expand the dataset and improve the model's robustness and generalization capability. Furthermore, a deep network model DenseNet and its convolutional network are employed for further feature extraction and fusion. DenseNet is composed of two types of core modules: DenseBlock and Transition-layer. The DenseBlock module facilitates inter-layer feature transfer through dense connections of multiple bottleneck modules. The Transition-layer module is used for dimensionality reduction and to decrease the size of the feature maps to prevent overfitting. Additionally, DenseNet incorporates convolutional and fully connected layers at the front and back ends, respectively, to expand the scope of feature extraction and optimize classification performance. For detailed structural information, please refer to Figure 4.

Figure 4
Flowchart of a neural network architecture starting with a normalized image. It progresses through a 2D convolution, multiple DenseBlock layers including Bneck Blocks A1, B1, C1, A2, B2, and C2, followed by a transition layer. The architecture continues with pooling and a linear layer, concluding with the output.

Figure 4. DenseNet network architecture.

To enhance the model's ability to recognize key features, a Convolutional Block Attention Module (CBAM) fusion mechanism is introduced. The CBAM module optimizes feature retention and interference suppression by adaptively learning the channel and spatial features of the image. CBAM is integrated in front of each DenseBlock and after each Transition-layer to reduce computational complexity and the number of parameters. Figure 5 illustrates the DenseNet structure integrated with CBAM. To prevent overfitting, a Dropout layer is introduced after the “BN-ReLU-Conv” structure in DenseNet. Since each DenseBlock contains two such structures, two Dropout layers are added within each DenseBlock. However, as a dimension reduction layer, the Transition-layer only contains one “BN-ReLU-Conv” structure, hence only one Dropout layer is added.

Figure 5
Diagram of a neural network architecture featuring CBAM modules. It includes an input layer followed by repeated sequences of Batch Normalization, ReLU, and Convolution layers. Colored arrows illustrate connections between layers, emphasizing interactions and transitions through the network. CBAM modules are applied after specific layers, enhancing attention mechanisms.

Figure 5. CBAM-densenet network architecture.

Furthermore, to maintain the size of the feature map while introducing non-linearity and reducing the number of parameters, a 1x1 convolutional layer is added after the DenseBlock and Transition-layer. This not only enhances the network's discrimination and expression but also reduces computational load due to the small number of parameters, accelerating training. For details of the optimized network structure, please refer to Figure 6.

Figure 6
Schematic diagrams of two neural network structure optimizations. (a) DenseBlock structure includes input, batch normalization, ReLU, 1x1 convolution, dropout, average pooling, and output layers. (b) Transition-layer structure has input, batch normalization, ReLU, 1x1 convolution, dropout, additional batch normalization, ReLU, 3x3 convolution, dropout, and output layers. Both depict dimensional changes in width (w), height (h), and channels.

Figure 6. Optimization details of DenseNet network architecture. (a) Schematic diagram of DenseBlock structure optimization. (b) Schematic diagram of Transition-layer structure optimiztion.

4.2 Assessment of multi-feature fusion strategies

The multi-feature fusion strategy for iris images aims to integrate various feature extraction techniques to generate a comprehensive feature vector, thereby optimizing iris recognition performance. When fusing, it is necessary to consider the characteristics of each feature method, as well as their intercorrelations and complementarities. Based on the characteristics of individual features, Gabor, CS-LBP, Haar wavelet features are combined with GLCM features to form four fusion schemes, as shown in Table 1. The proposed multi-feature fusion strategy integrates complementary information from four distinct texture descriptors. Specifically, Gabor filters are employed to capture frequency-domain textural patterns, CS-LBP (Completed Local Binary Pattern) and Haar wavelet transform are utilized to extract spatial-domain local micro-textures and multi-scale edge/contour features, respectively, while GLCM (Gray-Level Co-occurrence Matrix) is used to characterize global statistical textural properties such as homogeneity and contrast. To combine these heterogeneous features effectively, we adopt a channel concatenation mechanism. Each feature extraction method generates a single-channel feature map of size 32 × 256. These four maps are then stacked along the channel dimension to form a unified 4-channel fused feature tensor of size 32 × 256 × 4. This concatenated feature representation preserves the original characteristics of each individual descriptor without loss of information and serves as the input to the subsequent CBAM-DenseNet network for deep feature learning and classification. After feature extraction from iris images, a channel connection method is used according to the established fusion strategy to synthesize new feature images, which are then input into the CBAM-DenseNet model for training. Concurrently, the training process is statistically summarized to select the optimal fusion strategy based on relevant conditions during the learning process. Through the experimental analysis in Section 5.4, Scheme 4 demonstrated faster convergence, higher accuracy, and more stable performance during the training process, thus being chosen as the final feature fusion strategy.

Table 1
www.frontiersin.org

Table 1. Comparison of multi-feature fusion schemes.

4.3 Recognition process based on multi-feature fusion

The fused iris features are used for recognition, requiring the extracted features to be employed for iris recognition matching. By comparing the extracted feature vectors with known individual feature vectors to determine whether the similarity or distance between them reaches a certain threshold, authentication or verification can be carried out. Gabor frequency domain and CS-LBP spatial domain feature images, Haar wavelet mapped images, and GLCM single features from each iris are used. After channel splicing, they are adjusted to a size of 32 × 256 to form a comprehensive iris feature image, which is input into the CBAM-DenseNet network model for further feature extraction and fusion. The extracted feature vector is a multidimensional vector output by the fully connected layer. By calculating the Euclidean distance between the feature vectors of the known category label iris images and the feature vectors of the iris images to be recognized, and comparing it with a preset threshold, the feature similarity between the two images is judged to complete image classification. The multi-iris feature recognition matching process is shown in Figure 7.

Figure 7
Flowchart of an iris recognition system showing three categories of iris images. Each undergoes single iris feature extraction and channel concatenation. Outputs are fed into CABM-DenseNet, producing fusion feature vectors. These vectors undergo recognition comparison using Euclidean distance, leading to a threshold decision, resulting in categorization as Category 1.

Figure 7. Multi-feature fusion iris recognition process.

For the extracted iris fusion feature vectors, the similarity between the iris templates to be matched and the templates in the iris image database is calculated based on the Euclidean distance, and a preset threshold is compared to determine whether the two iris images come from the same eye. The calculation of Euclidean distance is given by the formula:

Eudist=i=1n(ai-bi)2    (9)

where ai and bi represent points in feature spaces X and Y, respectively. If the distance exceeds the threshold, the iris to be recognized does not match the target class; on the contrary, if the distance is less than or equal to the threshold, it is considered to match the target class.

4.4 Design of feature fusion recognition algorithm

For the problem of iris recognition with multi-feature fusion, a CBAM-DenseNet network structure is proposed, which integrates frequency domain, wavelet decomposition mapping, global texture analysis, and spatial domain features of iris images to enrich the image information input to the network model. The existing DenseNet network structure has been optimized to be more suitable for multi-feature feature extraction and recognition of irises. Figure 8a shows the proposed CBAM-DenseNet network model and its detailed parameter settings. The relevant hyperparameters of the CBAM-DenseNet network are set using the Adam optimizer as shown in Table 2. The network model structure settings adopted are shown in Figure 8b.

Figure 8
Diagram of the CBAM-DenseNet Network Model and its parameters. The model flow includes Gabor, CS-LBP, Haar, and GLCM components, concatenated to input dimensions [32, 256, 4]. It outlines layers like Convert, Maxpooling, Dense blocks, Transitions, and Global-average-pooling, leading to output. The right table details layer operations with corresponding strides, including convolution and pooling types, such as Conv (5x5), Max-pooling (3x3, stride 2), and Dense-block configurations.

Figure 8. CBAM-DenseNet network model and structural parameters. (a) CBAM-DenseNet network model. (b) Parameters of the CBAM-DenseNet network structure.

Table 2
www.frontiersin.org

Table 2. Hyperparameter settings.

5 Experiment

5.1 Data description

The experiments utilize the CASIAv4.0, NICE1.0, and JLU-6.0 iris datasets as data sources. Additionally, it incorporates different structural iris types from the CASIAv4.0, including CASIA-IrisV4-Interval (interval) and CASIA-IrisV3-Lamp (visible light). By employing the aforementioned iris image quality assessment methods and preprocessing techniques, iris images from multiple datasets are selected and normalized to construct the iris image dataset, as detailed in Table 3. Additionally, this paper performs data augmentation on iris ROI images, including methods such as image blurring, brightness alteration, image rotation, contrast alteration, and image flipping. The specific operations involve: applying 20% image blurring, brightness variation (25% increase and decrease), image rotation (2° and 4°), contrast variation (25% increase and decrease), and image flipping (horizontal and vertical directions) to augment the iris ROI images. The experimental dataset after augmentation is shown in Figure 9.

Table 3
www.frontiersin.org

Table 3. Custom iris dataset details.

Figure 9
A series of ten iris ROI images with different modifications. (a) Original iris ROI image. (b) Image with 25 percent blur. (c) Image with 25 percent brightness enhancement. (d) Image with 25 percent brightness reduction. (e) Image rotated by 2 degrees. (f) Image rotated by 4 degrees. (g) Image with 25 percent contrast increase. (h) Image with 25 percent contrast decrease. (i) Image with horizontal flip. (j) Image with vertical flip.

Figure 9. Augmented iris ROI region images.

5.2 Analysis of iris image quality evaluation and screening effects

To verify the quality evaluation and screening capabilities of the multi-metric iris image evaluation coefficients and the SA-SVM classifier, the CASIAv4.0 iris image database is utilized. Quality grading is performed through subjective pre-classification by personnel, and the proposed quality evaluation scheme is validated. The main verification includes two aspects: First, the capability of the iris image quality evaluation scheme is analyzed by assessing the accuracy of iris image quality classification; second, the impact of the evaluation scheme on iris recognition accuracy is explored through a pilot experiment with a small sample. The specific steps are as follows:

1. Randomly select about 100 classes of left and right iris images, totaling 2000 images, from the CASIAv4.0 iris image database.

2. Apply the Benner function to 1000 iris images for clarity evaluation, using the median value of the clarity function of these 1000 images as the threshold for classification: images above the threshold are categorized as clear images, and those below are categorized as blurry images, resulting in two main categories including 846 clear images and 154 blurry images. Blurry iris images that do not meet the clarity quality requirements are excluded, yielding 846 clear iris images to construct the experimental dataset IRISQuality.

3. The remaining 846 iris images are categorized according to iris quality evaluation standards into 214 high-quality images, 454 acceptable-quality iris images, and 178 poor-quality iris images.

Based on the experimental dataset IRIS-Quality, iris images of different quality levels are randomly assigned to the training and testing sets at a ratio of 4:1 and input into the SA-SVM classification model for training and learning. Meanwhile, the classification accuracy of this image quality evaluation scheme is compared with other image quality evaluation schemes, with the results shown in Table 4. To evaluate the impact of iris image quality assessment schemes on recognition performance, recognition rate is adopted as the evaluation criterion. Multiple quality assessment methods are used to select iris images for recognition, and it is verified whether the selected images improve matching accuracy. In the preprocessing stage, the Gabor algorithm is used to enhance the ROI area and extract features. The Hamming distance between sample feature images is calculated to determine iris matching. For each method, the Correct Recognition Rate (CRR) and Equal Error Rate (EER) metrics are used for testing, with results detailed in Table 5.

Table 4
www.frontiersin.org

Table 4. Comparison of classification performance of image quality evaluation schemes.

Table 5
www.frontiersin.org

Table 5. Comparison of computational efficiency and resource consumption of different quality assessment schemes.

Analysis results indicate that the multi-metric iris image quality assessment scheme based on the SA-SVM model can effectively evaluate iris quality and filter out low-quality irises, thereby enhancing recognition accuracy. Moreover, after screening, the match between iris images and feature extraction and recognition algorithms is improved, significantly increasing the success rate of iris recognition, thus having a positive effect on iris identification.

5.3 Evaluation of multi-feature fusion effects

In the preliminary experiment, 100 classes of iris images from different eyes in the CASIA-v4.0 iris database were selected. After image quality assessment and screening, 20 iris images per class were obtained, totaling 2,000 iris images. These images were preprocessed to yield 1,000 iris ROI images of size 32 × 256 as the source of experimental data. The experimental data were divided into training and testing sets with a ratio of 8/2. Different iris feature fusion schemes were set up and the iris features were re-extracted and fused based on the optimized CBAM-DenseNet network. Further analysis was conducted on the impact of different iris feature fusions on the recognition accuracy of the network and the loss of data during the training process to select the appropriate iris feature fusion scheme. The experimental data were subjected to feature extraction and fusion according to Schemes 1 to 4. The resulting fused iris feature images were input into the CBAM-DenseNet network for training, and the results obtained from the training data are summarized in Figure 10.

Figure 10
Two line graphs comparing feature fusion schemes post-training. Graph (a) shows loss across 30 epochs with four lines: green for Gabor+Haar+GLCM, blue for Haar+CS-LBP+GLCM, dark blue for Gabor+CS-LBP+GLCM, and red for Gabor+Haar+CS-LBP+GLCM. Loss decreases rapidly toward zero. Graph (b) shows accuracy across 30 epochs, with accuracy rising toward one. The color codes remain the same. The red line reaches optimal performance fastest in both graphs.

Figure 10. Comparison of loss and accuracy after training with different feature fusion schemes. (a) Comparison of loss for different feature fusion schemes post-training. (b) Comparison of accuracy for different feature fusion schemes post - training.

Summary of the experimental results indicates that the iris feature fusion images from Scheme 4 exhibit faster convergence, higher recognition accuracy, and better stability within neural networks. This is attributed to the integration of various iris feature extraction methods in Scheme 4, which enhances the accuracy and stability of iris recognition, thus being selected as the optimal fusion scheme.

5.4 Evaluation of multi-feature fusion recognition effects

In addressing the problem of image classification and recognition, this experiment compares the performance of various mainstream iris image recognition and classification models currently available. The focus of this experiment is to explore the recognition accuracy (CRR) of different iris recognition and classification models on different types of iris databases. By comparing the experimental results with other models, the advantages of the proposed iris recognition model in the field of iris recognition are discussed. The proposed iris multi-feature fusion recognition model was validated on the four datasets mentioned in Section 5.1: CASIAIrisV4-Interval, CASIA-IrisV3-Lamp, NICE1.0, and JLU-6.0, with the results presented in Table 6. Comparative analysis indicates that deep learning-based iris recognition algorithms outperform in accuracy. Despite the CASIA and NICE databases being larger, their performance was lower than expected when compared to the JLU6.0 dataset with a larger sample size. The proposed multi-feature fusion iris model in this study achieved superior recognition accuracy across four datasets, thereby validating its effectiveness. Image quality and size from different databases affect accuracy rates. DenseNet and its variants excel in small-sample sets, underscoring the importance of dataset selection and quality. Moreover, the proposed model leverages the feature extraction advantages of DenseNet and integrates multiple stable iris features, enriching input data and enhancing classification performance.

Table 6
www.frontiersin.org

Table 6. Average recognition accuracy (CRR) based on different experimental datasets.

5.5 Ablation study

To validate the contribution of each component in our proposed framework, we conduct a systematic ablation study on the NICE1.0 dataset–the same benchmark used in Section 5.4. We evaluate four progressively enhanced configurations: (1) a baseline DenseNet-121 model; (2) DenseNet-121 augmented with Dropout and standard data augmentation for regularization; (3) the model further integrated with the CBAM attention module at the final dense block; and (4) the full proposed model incorporating multi-feature fusion (combining Gabor, LBP, Haar, and CNN features). All variants are trained and evaluated under identical settings to ensure a fair comparison. The results are summarized in Table 7.

Table 7
www.frontiersin.org

Table 7. Ablation study on the NICE1.0 dataset.

As shown, the baseline DenseNet-121 achieves an accuracy of 74.63%, reflecting the challenging nature of the NICE1.0 dataset due to significant noise and occlusion. Introducing regularization improves performance by 5.2 percentage points (to 79.85%), demonstrating its effectiveness in mitigating overfitting under limited training data. The addition of the CBAM attention mechanism yields a substantial gain of 5.50%, raising accuracy to 85.37%, which highlights its ability to suppress irrelevant background regions and enhance discriminative iris textures. Finally, integrating handcrafted features via multi-feature fusion further boosts performance to 89.74%, aligning closely with the 89.74% reported for CBAM-DenseNet in Table 4. This consistency confirms the reliability of our experimental analysis and underscores the complementary benefits of combining deep and traditional features.

6 Conclusion

This paper conducts an in-depth study on iris recognition technology, aiming to design a multi-feature fusion-based iris recognition algorithm, and validates its effectiveness and performance. The focus is on the quality assessment and preprocessing of iris images, proposing a feature re-extraction method based on CBAM-Densenet. Experimental results demonstrate that the multi-feature fusion scheme effectively enhances iris image quality, thereby significantly improving the accuracy and robustness of iris recognition.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

YP: Writing – original draft, Writing – review & editing. ZW: Writing – original draft, Writing – review & editing. NJ: Writing – original draft, Writing – review & editing. JQ: Writing – original draft, Writing – review & editing. SL: Writing – original draft, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the Open Research Fund of the Shanghai Key Laboratory of Forensic Medicine and the Key Laboratory of Forensic Science, Ministry of Justice (Grant No. KF202415), the National Natural Science Foundation of China (Grant No. 62406342), the Liaoning Provincial Natural Science Foundation of China (Grant No. 2024-BS-260), the Shenyang Science and Technology Plan Project (Grant No. 24-213-3-43), Project of the Ministry of Public Security Science and Technology Program (Grant No. 2024JSYJC05), and the Research Project of the Liaoning Provincial Department of Education (Grant No. LJ232410175015).

Acknowledgments

We acknowledge the support provided by the Criminal Investigation Police University of China and the Shanghai Key Laboratory of Forensic Medicine.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ahmadi, N., and Akbarizadeh, G. (2018). Hybrid robust iris recognition approach using iris image pre-processing, two-dimensional gabor features and multi-layer perceptron neural network/PSO. IET Biometrics 7, 153–162. doi: 10.1049/iet-bmt.2017.0041

Crossref Full Text | Google Scholar

Al-Waisy, A. S., Qahwaji, R., Ipson, S., Al-Fahdawi, S., and Nagem, T. A. (2018). A multi-biometric iris recognition system based on a deep learning approach. Pattern Analy. Appl. 21, 783–802. doi: 10.1007/s10044-017-0656-1

Crossref Full Text | Google Scholar

Boles, W. W., and Boashash, B. (1998). A human identification technique using images of the iris and wavelet transform. IEEE Transact. Signal Proc. 46, 1185–1188. doi: 10.1109/78.668573

Crossref Full Text | Google Scholar

Daugman, J. (2001). Statistical richness of visual phase information: update on recognizing persons by iris patterns. Int. J. Comput. Vis. 45, 25–38. doi: 10.1023/A:1012365806338

Crossref Full Text | Google Scholar

Daugman, J. (2003). The importance of being random: statistical principles of iris recognition. Pattern Recognit. 36, 279–291. doi: 10.1016/S0031-3203(02)00030-4

Crossref Full Text | Google Scholar

Gangwar, A., and Joshi, A. (2016). “DeepIrisNet: Deep iris representation with applications in iris recognition and cross-sensor iris recognition,” in 2016 IEEE International Conference on Image Processing (ICIP) (Phoenix, AZ: IEEE), 2301–2305.

Google Scholar

Gao, S., Zhu, X., Liu, Y., He, F., and Huo, G. (2015). A quality assessment method of iris image based on support vector machine. J. Fiber Bioeng. Informat. 8, 293–330. doi: 10.3993/jfbim00114

Crossref Full Text | Google Scholar

Gupta, M. (2023). Biometric authentication using gait recognition. Univ. Res. Reports 10, 1–9. doi: 10.36676/urr.2023-v10i4-001

Crossref Full Text | Google Scholar

He, S., and Li, X. (2024). EnhanceDeepIris model for iris recognition applications. IEEE Access. 13, 6154–6154. doi: 10.1109/ACCESS.2024.3388169

Crossref Full Text | Google Scholar

Ishikawa, T. (2004). “Passive driver gaze tracking with active appearance models,” in IEEE Transactions on Intelligent Transportation Systems.

Google Scholar

Jan, F., Min-Allah, N., Agha, S., Usman, I., and Khan, I. (2021). A robust iris localization scheme for the iris recognition. Multimed. Tools Appl. 80, 4579–4605. doi: 10.1007/s11042-020-09814-5

Crossref Full Text | Google Scholar

Kaur, G., Girdhar, A., and Kaur, M. (2010). Enhanced iris recognition system-an integrated approach to person identification. Int. J. Comp. Appl. 975:8887. doi: 10.5120/1182-1630

Crossref Full Text | Google Scholar

Kim, J. S., Lee, Y. W., Hong, J. S., Kim, S. G., Batchuluun, G., and Park, K. R. (2023). LRFID-Net: a local-region-based fake-iris detection network for fake iris images synthesized by a generative adversarial network. Mathematics 11:4160. doi: 10.3390/math11194160

Crossref Full Text | Google Scholar

Qiu, T., Liu, Y., Zong, C., Yang, X., Wang, B., and Wang, M. (2025). Exploiting simd-ified bit-parallelism for high-performance complex event matching. IEEE Trans. Knowl. Data Eng. 38, 1054–1069.

Google Scholar

Tan, T., Zhang, X., Sun, Z., and Zhang, H. (2012). Noisy iris image matching by using multiple cues. Pattern Recognit. Lett. 33, 970–977. doi: 10.1016/j.patrec.2011.08.009

Crossref Full Text | Google Scholar

Wang, L., Zhang, K., Ren, M., Wang, Y., and Sun, Z. (2020). “Recognition oriented iris image quality assessment in the feature space,” in 2020 IEEE International Joint Conference on Biometrics (IJCB) (Houston, TX: IEEE),1–9.

Google Scholar

Wei, J., Wang, Y., Huang, H., He, R., Sun, Z., and Gao, X. (2022). Contextual measures for iris recognition. IEEE Trans. Inform. Forens. Secur. 18, 57–70. doi: 10.1109/TIFS.2022.3221897

Crossref Full Text | Google Scholar

Wildes, R. P. (1997). Iris recognition: an emerging biometric technology. Proc. IEEE 85, 1348–1363. doi: 10.1109/5.628669

Crossref Full Text | Google Scholar

Wildes, R. P., Asmuth, J. C., Green, G. L., Hsu, S. C., Kolczynski, R. J., Matey, J. R., and McBride, S. E. (1994). “A system for automated iris recognition,” in Proceedings of 1994 IEEE Workshop on Applications of Computer Vision (Sarasota, FL: IEEE), 121–128.

Google Scholar

Zambrano, J. E., Benalcazar, D. P., Perez, C. A., and Bowyer, K. W. (2022). Iris recognition using low-level CNN layers without training and single matching. IEEE Access 10, 41276–41286. doi: 10.1109/ACCESS.2022.3166910

Crossref Full Text | Google Scholar

Zhou, W., Ma, X., and Zhang, Y. (2020). Research on image preprocessing algorithm and deep learning of iris recognition. J. Phys.: Conf. Series 1621:012008. doi: 10.1088/1742-6596/1621/1/012008

Crossref Full Text | Google Scholar

Keywords: attention mechanism, CBAM, deep learning, DenseNet, image quality assessment, iris recognition, multi-feature fusion

Citation: Pang Y, Wang Z, Jiang N, Qin J and Li S (2026) CBAM-DenseNet with multi-feature quality filtering: advancing accuracy in small-sample iris recognition. Front. Artif. Intell. 8:1714882. doi: 10.3389/frai.2025.1714882

Received: 28 September 2025; Revised: 27 December 2025;
Accepted: 30 December 2025; Published: 27 January 2026.

Edited by:

Milan Tuba, Singidunum University, Serbia

Reviewed by:

Lei Zhong, Xi'an Technological University, China
Mehmet Sezgin, Istanbul Commerce University, Türkiye

Copyright © 2026 Pang, Wang, Jiang, Qin and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yongheng Pang, cGFuZ3lvbmdoZW5nQGNpcHVjLmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.