Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurorobot., 06 August 2025

Volume 19 - 2025 | https://doi.org/10.3389/fnbot.2025.1630281

Fine-grained image classification using the MogaNet network and a multi-level gating mechanism

Dahai LiDahai Li1Su Chen
Su Chen2*
  • 1School of Electronics and Electrical Engineering, Zhengzhou University of Science and Technology, Zhengzhou, China
  • 2Department of Mechanical and Electrical Engineering, Henan Vocational College of Water Conservancy and Environment, Zhengzhou, China

Fine-grained image classification tasks face challenges such as difficulty in labeling, scarcity of samples, and small category differences. To address this problem, this study proposes a novel fine-grained image classification method based on the MogaNet network and a multi-level gating mechanism. A feature extraction network based on MogaNet is constructed, and multi-scale feature fusion is combined to fully mine image information. The contextual information extractor is designed to align and filter more discriminative local features using the semantic context of the network, thereby strengthening the network’s ability to capture detailed features. Meanwhile, a multi-level gating mechanism is introduced to obtain the saliency features of images. A feature elimination strategy is proposed to suppress the interference of fuzzy class features and background noise. A loss function is designed to constrain the elimination of fuzzy class features and classification prediction. Experimental results demonstrate that the new method can be applied to 5-shot tasks across four public datasets: Mini-ImageNet, CUB-200-2011, Stanford Dogs, and Stanford Cars. The accuracy rates reach 79.33, 87.58, 79.34, and 83.82%, respectively, which shows better performance than other state-of-the-art image classification methods.

1 Introduction

Fine-grained image classification is a crucial task in the field of computer vision (CV), aiming to solve the problem of subdividing subclasses within the same general class (He and Peng, 2017). Fine-grained image classification has garnered the attention of researchers due to its wide-ranging application potential in areas such as smart transportation (Ke and Zhang, 2020), smart retail, and biodiversity monitoring. However, the uniqueness of fine-grained image classification lies in the significant differences within a class and the small differences between classes, which require the model to extract more subtle feature representations to achieve accurate image classification (Zhang W. et al., 2024). Therefore, compared with traditional image recognition tasks, fine-grained image classification is more challenging (Ke et al., 2023).

According to the type of label used, existing fine-grained image classification methods can be divided into two categories: one is a strong supervision method that relies on additional manual annotation information, and the other is a weak supervision method that relies only on image-level labels. The majority of early research focused on strong supervision methods. For example, Liu et al. (2016) proposed a Fully Convolutional Network (FCN) based on image masks. During the training process, the model accurately localized the head and torso of the object using manually marked key point information and then generated an image mask to exclude convolutional descriptors that contained the object region. Finally, the pooling operation integrated the features of the entire object, including the head and trunk, to achieve efficient classification. However, strong supervision methods require a significant amount of annotation information, such as object bounding boxes and key point information. The acquisition of this annotation information usually consumes a considerable amount of time and energy. Therefore, the method based on weak supervision has gradually become a new focus of research. Progressive multi-granularity training (PMG), proposed by Du et al. (2020), is a representative of the weak supervision method. The method utilizes the Jigsaw Puzzle Generator Module (JPGM) to segment the original image into blocks of different granularities and rearranges them into the network. In view of the sensitivity of the shallow layer network to the texture details of the object, PMG inputs fine-grained image blocks into the shallow layer network to extract the features of the object detail region. For coarse-grained image blocks, they are fed into the deep network to capture high-level semantic information of the image. Thus, the feature information of different granularities can be effectively fused, thereby improving the classification effect.

The early research is mainly based on convolutional neural network (CNN) technology to design fine-grained image classification models, and remarkable results have been achieved. However, due to the inherent limitation of the convolution operation, namely the receptive field constraint, these models tend to pay too much attention to local image details, and their global modeling ability for objects is relatively limited, which gradually leads to a performance bottleneck. At the same time, with Transformer excelling in the field of natural language processing, its powerful global information modeling capability has garnered widespread concern. Researchers have begun to introduce the vision transformer (ViT) (Yin et al., 2022) into various computer vision tasks, including pedestrian re-identification, object detection, and image classification, and have made positive progress. However, unlike other visual tasks, fine-grained image classification relies mainly on subtle inter-class differences. It therefore requires the network to be able to discover more subtle discriminative regional features. On the other hand, the ViT model relies on multi-head self-attention (MSA) to mine the correlation between each image block, paying more attention to the integration of global information and less attention to local key areas (Yuan et al., 2021). Therefore, applying the ViT directly to fine-grained image classification tasks may result in poor classification performance.

In response to the above problems, the researchers aim to improve the ViT, seeking to maintain its strong global modeling capabilities while enhancing its ability to capture local features, thereby improving ViT’s performance in fine-grained image classification tasks. For example, to address the ViT’s limitation in ignoring local regional information, Zhao et al. (2024) proposed a TransFG network comprising a Part Selection Module (PSM) and Contrastive Feature Learning (CFL).

The model used the PSM module to integrate the self-attention weights from the first 11 layers of the ViT to obtain the final attention map and then selected the PatchToken containing the locally important region of the object from the last layer of the self-attention encoding, compensating for the lack of detailed features in the network. In addition, the CFL module instructed the network to learn the characteristics shared among the same species and the differences between different ones, thereby amplifying the inter-class differences and narrowing the intra-class differences. Although TransFG considered the importance of local discriminative features to the network, it ignored the local, low, and high-level information contained in different levels, resulting in a partial loss of discriminative information. Therefore, Wang et al. (2021) proposed a new FFVT network based on TransFG. It filtered the most critical tokens in each self-attention encoding layer based on the size of the self-attention weight and integrated them. This method effectively integrated the subtle clues from different regions of the object and enriched the discriminant information of the network at different levels. In addition, to help the network identify the most discriminative area in the image, Xu et al. (2023) clipped and scaled the object area in the original image according to the attention weight and re-input it into the network, thereby improving the network’s recognition ability for images of different scales. Moreover, they also adopt distillation learning, which utilizes the features extracted from the CNN network to guide the ViT network in learning multi-source semantic information from different structures, thereby improving the network’s generalization ability.

Although existing ViT-based methods have made significant progress, several problems remain to be addressed. First, since images need to be divided into image blocks of a certain size before being input into the ViT network, and the size of these image blocks remains constant throughout the entire forward propagation process, the network can only process image information of the same granularity. However, for fine-grained image classification, features of different granularities have varying identification capabilities within the network, so the information learned by the network under a single granularity is incomplete and insufficient. Second, in fine-grained image classification, existing methods based on distillation learning often rely on a teacher model with a large number of parameters to guide a student model with a small number of parameters. However, due to the differences between the models, the teacher network with a better model effect is not necessarily conducive to the learning of the student network (Lu and Han, 2024); that is, the “knowledge” in the teacher network cannot be transferred to the student network. When the performance difference between the teacher network and the student network is too large, the guidance provided by the teacher network can lead to contradictions in the student network’s optimization. Third, different layers of the network have distinct areas of concern for the object; that is, each encoding layer has a different degree of influence on the network classification results. However, the existing method (Wan et al., 2021) simply fuses the self-attention weights of all layers when mining local details, thereby ignoring the differences between each self-attention encoding layer, which affects image classification. In addition, these methods only improve the large prediction region generated by the network, paying more attention to global information and paying less attention to local and underlying features. Background noise in complex scenes can significantly impair the network’s ability to focus on subtle local features.

Based on the above analysis, this study proposes a novel fine-grained image classification method utilizing the MogaNet network and a multi-level gating mechanism. First, using MogaNet (efficient multi-order gated aggregation network) as the backbone network, in the training stage, to extract multi-scale features, a contextual information extractor (CIE) is designed to remove the redundant information of the context, extract fine-grained local features with spatial context from the deep features, and capture the subtle changes in each region. Second, the feature elimination strategy (FES) is designed to mitigate the impact of fuzzy class features and backgrounds on classification, thereby improving the quality of processing ambiguous objects and complex backgrounds. Meanwhile, a multi-level gating mechanism is introduced to obtain the saliency features of images. Finally, features of different depths are fused to produce a refined prediction region. Experiments demonstrate that the proposed method exhibits favorable classification characteristics.

2 Related research

2.1 Fine-grained image classification

Currently, fine-grained image classification methods encompass local localization, feature coding, and attention mechanisms. For example, Ge et al. (2019) proposed a weakly supervised supplementary component model for fine-grained image classification. This method extracted rough object instances through mask R-CNN and CRF segmentation, then searched for the best component model, and fused this component information using bidirectional LSTM, which significantly improved classification performance. In order to obtain highly identifiable local discriminant regions, Liu et al. (2021) proposed a multi-regional attention method. The attention mechanism was repeatedly applied to focus on secondary features, combined with background removal and up-sampling techniques to obtain accurate local image features. Finally, a rectangular box was used to count positions, thereby reducing detail loss. In addition, combining multiple methods is also an effective strategy for improving performance and solving problems more comprehensively. Zhang et al. (2016) proposed a multi-scale collaborative differential network. They employed different kernels and pooling methods to obtain images of varying scales and then input these images into the basic network to generate multi-scale feature blocks. These blocks were fused to create features with more comprehensive information. Finally, these features were fed into a collaborative differential network to capture second-order information about the image. However, the effectiveness of the model largely depends on the adequacy of the data, and it is challenging to obtain data for rare species. Feature learning plays a crucial role in nearly all image tasks, including retrieval, detection, and classification. The performance of deep convolutional networks is particularly outstanding, and their success is mainly attributed to the ability to learn discriminative deep features. In the early days of deep learning development, the features (activations) of fully connected layers were typically used as image representations. Subsequently, with the continuous development of deep learning, researchers discovered that the feature maps of deeper convolutional layers contain intermediate and advanced information, such as parts of an object or the entire object. This has led to the widespread use of convolutional features/descriptors. Moreover, compared with the output of the fully connected layer, the application of encoding techniques to these local convolutional descriptors has brought significant improvements.

To some extent, the improvement of encoding techniques comes from the high-order statistics encoded in the final features. Especially for fine-grained recognition, when the Fisher vector encoding of SIFT features (Scale-Invariant Feature Transform) outperforms the fine-tuned AlexNet in several fine-grained tasks, the demand for end-to-end modeling of high-order statistics becomes apparent. Bilinear convolutional neural networks represent images as the mixed outer product of features derived from two deep convolutional neural networks, thereby encoding the second-order statistics of convolutional activation and significantly improving fine-grained recognition. However, bilinear mixed features generate extremely high-dimensional features, resulting in a significant increase in the number of parameters in the deep network classification module. This may lead to overfitting and make it impractical in practical applications, especially in large-scale applications. Therefore, this study focuses on exploring fine-grained image classification methods in a few-shot learning environment.

2.2 Few-shot learning

In a few-shot learning environment, due to the limited number of samples, the model is prone to overfitting, which reduces its generalization ability to new samples (Song et al., 2023). Few-shot learning strategies can be divided into two main categories: optimization-based meta-learning methods and metric-based learning methods. The core of meta-learning lies in enhancing the ability of models to adapt quickly to new tasks, although challenges arise when dealing with domain shifts. Metric learning is another widely used strategy that measures similarity by calculating the distance between sample features. For example, the prototype network proposed by Xu et al. (2020) calculates the average vector of samples within a class as the class prototype and compares the new sample with the class prototype to achieve fast classification. Yan et al. (2023) proposed the discriminant spatial metric network, aiming to maximize inter-class dispersion and minimize intra-class dispersion, and effectively use the geometric structure of samples to enhance the discriminant ability. Wei et al. (2019) introduced the concept of few-shot fine-grained recognition and designed a piecewise smart mapping function to analyze bilinear features by generating decision boundaries through a set of sub-classifiers. Folkesson et al. (2008) proposed a method for the adaptive selection of representative local descriptors, modifying the traditional KNN classification model to adjust the weights based on the distance between neighboring points. In addition, other studies (Nie et al., 2023) have used attention mechanisms to enhance semantic information and focus on image-rich regions.

2.3 Frequency domain learning

As an effective tool, frequency analysis has been widely utilized in deep learning to extract frequency features from signals. Mitsuhata et al. (2001) employed the Fourier transform and its inverse to reduce the discrepancy between the source distribution and object distribution by adjusting the low-frequency components. Zhong et al. (2022) converted RGB color images into the frequency domain and developed a frequency selection strategy to select frequencies with abundant information. In addition, Chen and Wang (2021) discussed the influence of different frequency components on few-shot learning tasks and proposed a new frequency-guided few-shot learning framework. The framework could adaptively mask relevant information in images according to specific task requirements. Unlike the above work, this study identifies the overall structure of the image by extracting features in the frequency domain, thereby avoiding disturbances from local noise. Local changes in the image are then captured in the airspace. The combination of the two provides better resistance to noise and other interference, and parallel processing can also provide higher degrees of freedom for subsequent operations.

3 Proposed image classification

The proposed fine-grained image classification method, based on the MogaNet network and a multi-level gating mechanism, is illustrated in Figure 1.

Figure 1
Flowchart illustrating a neural network model for image classification. An image of a bird is processed through four layers in the MogaNet-L module. Each layer involves operations: CIE, addition, MGM, and FES. Features are fused, and predictions are made via classification with L_drop and L_final influencing the output.

Figure 1. Proposed image classification structure.

First, the MogaNet backbone network is utilized to extract multi-scale feature maps. Then, a contextual information extractor is proposed to extract both global features, including shapes and textures, and local features, including edges and corners, using multi-scale sliding windows. The attention mechanism is used to suppress low-context areas and enhance the ability to capture detailed features. Meanwhile, a multi-level gating mechanism (MGM) is introduced to obtain the saliency features of images. Then, a feature elimination module is designed to reduce the impact of fuzzy class features and background features on network classification by combining multi-modal feature interaction. Finally, a full connection layer is used to fuse features of different scales, and different input images are classified through the classification layer.

3.1 MogaNet

The backbone network determines the feature extraction capability of the algorithm and its application potential in downstream tasks. Selecting a suitable backbone can help the network focus on improving fine-grained problems. Recent studies have shown that, with advanced training settings and updated structures, convolutional neural networks can achieve comparable or better performance than the ViT without increasing computational requirements. The MogaNet, designed by Li et al. (2023), utilizes an improved convolutional macro architecture and optimization strategy to achieve faster inference speed and higher accuracy with the same number of parameters. MogaNet performs information mining and channel aggregation in pure convolutional networks and is significantly superior to current mainstream backbone networks in general image classification, image retrieval, and semantic segmentation at a lower computational cost. This success demonstrates that the MogaNet network can extract highly powerful features and has significant potential for transfer learning in downstream tasks. The selected MogaNet-L in this study consists of four stages. After the input image passes through different stages, the feature map F i R H i × W i × C i at different scales can be obtained. i indicates the number of phases of the backbone network. H i , W i , and C i represent the height, width, and number of channels of the feature map. After the input image passes through stage 1, the height and width of the feature map are reduced to one-quarter of the original size. Thereafter, after each stage, the height and width of the feature map are halved, and the number of channels is increased to four times that of the previous stage.

3.2 Contextual information extractor

Compared to global features, local features are abundant in the image, the correlation between features is low, and the influence of occlusion is minimal. Previous fine-grained image classification methods typically employ global features to generate large prediction regions, making it challenging for the network to discern local, detailed features in unclear objects and complex scenes. Moreover, the detection objects in the image are often mixed with the background or occupy only a small part of the image, and the local features directly obtained by the network from the local region are very limited. Since each object always exists in a specific environment or coexists with other objects, a good representation of contextual information of objects/scenes can integrate global and local features, help the network detect objects, and play a key role in fine-grained classification tasks. Based on the above analysis, a 3-layer contextual information extractor CIE is designed. Each layer has the same structure and is composed of a multi-scale sliding window and attention guide mechanism. Through a multi-scale sliding window, spatial contextual information of deep features is captured. Then, an attention-directing mechanism is used to suppress low-contextual information areas of shallow features and up-sample deep features to provide an adequate representation of important targets.

The single-layer structure of CIE is shown in Figure 2. Its input comes from the output feature map F i and F i + 1 of any adjacent layer 2 of the backbone network. F i R H i × W i × C i , F i + 1 R H i + 1 × W i + 1 × C i + 1 , where F i is a shallow feature and F i + 1 is a deep feature.

Figure 2
Flowchart diagram of a neural network process. The top section shows input feature maps \(F_{i+1}\) processed through convolution and sliding window operations, resulting in intermediate feature maps \(F_{i+1}^1\) and \(F_{i+1}^2\), then combined into \(f_g\). The lower section shows another input \(F_i\) interacting with the outputs from the top through multiple operations including element-wise multiplication, convolution, and addition, resulting in output \(f_{i+1}\). Labels indicate dimensions and operations like \(f^T_g\) and \(f_a\).

Figure 2. Framework of a single-layer contextual information extractor.

First, a two-dimensional convolution is used to change the number of channels of the deep feature F i + 1 into C i + 1 / 16 and C i + 1 / 36 , and the spatial contextual information of the deep feature is mined through a sliding window operation. The expression is shown in Equations 1 and 2.

F i + 1 1 = Unfold ( Conv 2 d ( F i + 1 ) , s 1 )     (1)
F i + 1 2 = Unfold ( Conv 2 d ( F i + 1 ) , s 2 )     (2)

Unfold shows the sliding window operation. s 1 and s 2 are the sizes of the two sliding windows, respectively. F i + 1 1 and F i + 1 2 respectively represent the feature map obtained after the sliding window operation. The number of channels of F i + 1 1 and F i + 1 2 is determined by the size of the sliding window. Set s 1 = 2 × 2 and s 2 = 3 × 3 , the number of channels of the two feature graphs after sliding window operation is C i + 1 / 4 , which is consistent with the number of channels C i of shallow feature F i . By concatenating F i + 1 1 and F i + 1 2 together, it can obtain the high-level descriptor f g R C i + 1 4 × s , where s’ is the sum of the widths of the two features after the sliding window operation. f g aggregates features from multiple local parts of a deep feature and their contextual information and can emphasize the importance of the local. The f g is computed is as Equation 3.

f g = concate ( F i + 1 1 , F i + 1 2 )     (3)
where concate represents the feature concatenation function.

Inspired by the attention approach, we align the high-level descriptor f g with the shallow feature F i to calculate the similarity between them and obtain the repeated contextual information f a in f g through the vector inner product operation. f a R H i × W i × C i . The f a expression is as Equation 4.

f a = ( f g F i ) f g T     (4)

The deep layers of the network perceive more comprehensive semantic information, while the shallow layers focus on details such as edges and corners. That is, deep features have more semantic context, while shallow features have more detailed context, so for ambiguous targets in complex backgrounds, repeated contexts will contain more useless semantics. Therefore, redundant features are removed from shallow features to extract fine-grained local feature f A with spatial context from global information. The f A expression is as Equation 5.

f A = F i f a     (5)

The feature f i + 1 containing rich contextual information is obtained by downsampling the local feature to the deep feature. f i + 1 R H i + 1 × W i + 1 × C i + 1 , i [ 1 , 3 ] . The expression for f i + 1 is as Equation 6.

f i + 1 = F i + 1 + f fpn ( f A )     (6)
where f fpn represents the feature pyramid sub-sampling function.

3.3 Multi-level gating mechanism (MGM)

The multi-level gating mechanism is derived from MogaNet, a pure convolutional neural network inspired by Vision Transformers (ViT). Its characteristic is to promote the extraction of middle-order features, thereby improving the model’s performance. The multi-level gating mechanism is shown in Figure 3. The structure of the multi-level gating mechanism consists of two branches, and the left branch can be regarded as a residual connection. The branch on the right is mainly used for feature extraction. For the feature extraction of the input feature map, a depth-separable convolution operation is first performed using a 5 × 5 convolution kernel. Then the feature graph is sliced into three sub-feature graphs. For two sub-feature graphs, convolution kernels with dimensions of 5 × 5 and 7 × 7 are used for feature extraction. In the MGM model, the design of multi-branch convolution (5 × 5, 7 × 7) is primarily adopted to capture image features at different scales, thereby better handling various structural and textural information within the image. The features in an image can be classified into fine-grained features (such as texture and detail) and coarse-grained features (such as shape and structure). Convolution kernels of different sizes can extract these features at different scales, thereby providing richer information for the model.

Figure 3
Diagram illustrating a neural network architecture involving input passing through a depthwise convolution block. It splits and processes using dilations of one, two, and three, then concatenates results. The flow includes convolution layers, SiLU activation, and bypass connections, leading to output.

Figure 3. MGM structure.

Dilated convolution is a special convolution operation. It expands the receptive field of the convolution kernel by inserting a dilation rate between the elements of the convolution kernel without the need to add additional parameters. This design enables the model to capture a wider range of contextual information without increasing the computational complexity.

The other one is directly mapped to the next layer, and the three sub-feature graphs are concatenated using a 1 × 1 convolution for fusion. Then, the SiLU function (Elfwing et al., 2018) is used for feature activation. Finally, the features on the two branches are added, and the output is adjusted using a 1 × 1 convolution. The advantage of a multi-level gating mechanism is that the use of large convolution kernels and expansive convolution of different sizes can significantly enhance the receptive field of the network, thereby capturing multi-level features and improving the model’s performance (Yin et al., 2024a; Yin et al., 2024b).

3.4 Feature elimination strategy

A fuzzy category refers to the result of predicting similar classification scores, which is one of the primary causes of misclassification. After obtaining the features containing contextual information through the above CIE, a feature elimination strategy is designed to further eliminate the influence of background and fuzzy classes on the network, allowing it to focus on more discriminative features. By eliminating similar regions and background features between classes, forcing the network to focus on other discriminant features, ViT-SAC (self-assessment classifier) examined the fuzziness in the first K prediction classes and utilized the images and the first K prediction results to reevaluate the classification. This research method is inspired by ViT-SAC, but the proposed method is different from it. (1) This method does not need to crop the original image through the generated feature map and re-input it into the network, achieving end-to-end training and avoiding the impact of training errors and errors in the previous stage on the network training in the later stage; (2) This method uses prediction scores to divide the feature map into object candidate regions and background candidate regions and deletes the corresponding fuzzy class feature points and background feature points directly at the pixel level. Since the output f i + 1 of CIE is context-sensitive, and deleting certain features does not lose the contextual information of the image.

The feature elimination strategy (FES) consists of four modules, each with the same structure. The specific structure of a single module is shown in Figure 4. X is the input of this module. X’ represents the output of this module. Input X 1 of module 1 is the feature graph F 1 in stage 1. Inputs X 2 ~ X 4 of modules 2–4 are outputs f i + 1 of CIE, i [ 1 , 3 ] . Each FES module contains a fuzzy class feature elimination branch and a background elimination branch. The double-branch structure comprehensively considers the contribution degree of fuzzy class features and background features to network classification. It generates discriminative features by eliminating the feature points that are less helpful to classification. For any module of FES, the feature points in the input feature graph X are numbered one by one. X j represents the j-th feature point in X , j { 1 , 2 , , H × W ) . The feature number set Z contains the numbers of all the feature points in X . Z can be expressed is as Equation 7.

Z = { 1 , 2 , , H × W }     (7)
Figure 4
Flowchart depicting a neural network model for image processing. An input bird image is processed through multiple layers, including Backbone+CIE, resulting in a feature matrix X. This flows through V, E, Top-K prediction, BAN, and M layers, incorporating loss functions \(L_{joint}\) and \(L_{back}\). The output is a matrix labeled \(X\tilde{D}\).

Figure 4. Framework of single-layer feature elimination strategy.

The feature elimination branch of the fuzzy class is used to eliminate the feature points of common concern of the first K terms fuzzy class. First, the visual feature V R d V is extracted by the fully connected layer, and the first K prediction results are obtained, where d V represents the dimension of V . Then, the Glove word embedding method is used to learn the language modal information of class tags E = [ E 1 , , E k , , E K ] , E k R d e . This article uses the default d e =1,024 in the Glove to represent the dimension of each class label information. Bilinear attention network (BAN) (Kim et al., 2018) is used to fuse the two-modal information X and E , and the joint feature J R d e and fuzzy attention map M R H × W are obtained, with the expression is as Equations 810.

m = σ ( Linear ( X ) × E )     (8)
M = sum ( m ) . resize ( H , W )     (9)
J = j H × W 1 K X j T × M × E k     (10)

where σ ( ) represents the Softmax function. Linear ( ) indicates the fully connected layer. m represents K feature maps obtained by BAN. m R H × W × K . sum ( ) means add by column, and resize ( H , W ) means return the feature vector of size ( H × W ). The response of M reflects the region of common concern among the first K fuzzy classes, and the higher the response of the feature points, the more likely it is to cause incorrect classification. The H × W feature points in M are sorted according to the response size, and the feature points of common concern of the first τ fuzzy classes are discarded, and the fuzzy cancellation feature number set Z max is obtained. Z max is a subset of the set Z of all feature numbers.

Background elimination is used to eliminate feature points with fewer classes, thereby reducing the influence of complex backgrounds on classification results. The classification score X of all feature points in x is extracted by the full connection layer, x R H × W × K . Each feature point is divided into class P , where P is the total number of classes. The maximum prediction probability is used as the classification score of the feature point, and a low classification score indicates that the feature point belongs to the background feature. The first δ feature points are selected according to the classification scores from high to low, and the background elimination feature number set Z min is obtained. Z min is a subset of the set z of all feature numbers.

Finally, it takes the intersection of the two numbered sets Z final = Z min Z max . By gating the feature points in Z final , a new feature graph X’ is obtained, and the expression is as Equation 11.

X ' = { X j j Z final }     (11)
where X ' R N × C , N is the total number of extracted feature numbers, and C is the number of channels in the feature map.

Since the feature number g ( Z final ) in Z final is not fixed, to determine the classifier parameter of the X’ connection, it takes N = δ τ . When g ( Z final ) > N , delete the feature numbers with low classification scores in Z final until g ( Z final ) = N ; When g ( Z final ) < N , add the number of features with high classification scores to Z final from Zmax until g ( Z final ) = N .

3.5 Feature fusion

Using single-scale feature maps for classification training, the network generates large prediction regions and cannot adequately characterize image details. Therefore, the multi-scale feature fusion layer is introduced to reduce the loss of detail features and fully mine multi-scale spatial information. After the FES strategy, four feature maps of different scales are obtained namely, X 1 = R N 1 × C 1 , X 2 = R N 2 × C 2 , X 3 = R N 3 × C 3 , X 4 = R N 4 × C 4 . First, the dimensions of four feature graphs are unified by using one-dimensional convolution with a convolution kernel size 1, and they are joined together along the channel dimension to obtain X final RN × C 1 . Here, N ' = N 1 + N 2 + N 3 + N 4 . X final contains all the feature points mined by the network that are useful for classification. The selected local features are then recombined into global features that can represent the entire image through a fully connected layer. Thus, a feature graph of dimension RN × C 1 can produce a prediction of dimension R P after passing through the fully connected layer.

3.6 Loss function

Due to small differences among the subclasses in the image classification task, the cross-entropy loss function alone is insufficient to completely supervise the network’s iterative learning. The loss functions designed in this study include prediction loss L final , background elimination loss L back , and fuzzy class elimination loss L joint and drop feature point loss L drop . The total loss L can be expressed is as Equation 12.

L = L final + L back + L joint + L drop     (12)

L final is calculated by the cross-entropy loss function L ce , which can be expressed is as Equation 13.

L final = L ce ( X final , y )     (13)
where y indicates the sample label.

L back is used to oversee the background elimination branch of the FES policy. The average classification score l i R P of all feature points is used to calculate the overall loss of X i with the Equations 14 and 15.

l i = 1 H × W j = 1 H × W x j i     (14)
L back = 1 H × W i = 1 4 p = 1 P y p log l p i     (15)
where x j i represents the classification score of the feature point whose feature number is j in the i th stage. y p represents the real label of the sample. l p i represents the prediction probability that l i belongs to class p .

L joint is used to supervise the fuzzy class feature elimination branch of FES policy, which can be expressed as Equation 16.

L joint = i = 1 4 p = 1 P y p log J p i     (16)

where J p i represents the prediction probability that the feature vector J i of i th stage belongs to class p .

L drop is calculated by the sum h i of the discarded features in the FES policy. The expression with the Equations 17 and 18.

h i = j Z Z final X j i     (17)
L drop = i = 1 4 p = 1 P y p log ( 1 h p i )     (18)

where h p i indicates that h i belongs to the prediction probability of class p .

4 Experimental results and analysis

In order to verify the effectiveness of the proposed method, the model training is carried out on four publicly available fine-grained datasets: Mini-ImageNet (Oreshkin et al., 2018), CUB-200-2011 (CUB) (Zhai and Wu, 2018), Stanford Dogs (Dogs) (Khosla et al., 2011), and Stanford Cars (Cars) (Jiang and Yin, 2023), and the experimental comparison is made with other few-shot fine-grained image classification methods. In addition, ablation experiments are performed to verify the performance advantages of individual modules, the complexity of multi-level gating mechanism modules in backbone networks is analyzed, and the spatial effect of feature visualization is demonstrated.

4.1 Experiment set

The standard fine-grained datasets CUB, Dogs, and Cars, and the few-shot dataset Mini-ImageNet, are used for comparison experiments. CUB is a representative bird dataset containing 11,788 images from 200 different bird species. The Cars dataset comprises 16,185 images of 196 different car brands. The Dogs dataset collects 120 different dog breeds and contains a total of 20,580 images. Mini-ImageNet is the most widely used dataset in the field of few-shot learning, comprising 100 categories with 600 images each, totaling 60,000 color images. The above datasets are shown in Table 1.

Table 1
www.frontiersin.org

Table 1. Dataset information statistics.

Although the data volume division varies under different tasks, in this study, we uniformly divide the numbers of the training set, validation set, and test set of these four datasets into 60:20:20%.

4.2 Experiment settings

This experiment was conducted on a desktop computer equipped with a GTX 1660S GPU, utilizing the PyTorch 1.10 framework and Python 3.7 version to complete the experiment. To evaluate the performance of a fine-grained image classification method, the accuracy index is used as the criterion for judgment. The model conducts 400 rounds of epoch training without using any additional data. The few-shot learning configurations are 5-way 1-shot and 5-way 5-shot. Among them, only 1 or 5 samples are extracted from each class as a support set so as to verify the classification ability of fine-grained images in the case of a few-shot sample. The image samples for training and testing are adjusted to 84 × 84 pixels. The Adam optimizer is used for model training, with a learning rate of 0.001. In the final testing phase, the classification accuracy of the test sample and its 95% confidence interval are evaluated by calculating the average of 600 randomly generated scenarios.

4.3 Experimental comparison

4.3.1 Experimental comparison on a few-shot datasets

To evaluate the performance of the proposed method in the classification task, an experimental comparison is conducted on the Mini-ImageNet dataset, as presented in Table 2.

Table 2
www.frontiersin.org

Table 2. Experimental comparative analysis on the mini-ImageNet dataset.

Table 2 shows the experimental results of the 5-way 1-shot and 5-way 5-shot classification tasks on the Mini-ImageNet dataset. The classification accuracy rates reach 65.87 and 79.33%, respectively. In the 1-shot setting, the accuracy performance of the proposed method is superior to other methods, which is attributed to two main factors: (1) The advantage of multi-scale feature fusion enhances the ability of information capture and (2) accelerated model adaptation to new categories by setting a multi-level gating mechanism.

To evaluate the performance of the model under low-light conditions, we use the low-light dataset for testing. These images simulate scenes at night or in low-light environments, accurately reflecting the model’s performance under such conditions. Table 3 presents the experimental results obtained under low-light conditions.

Table 3
www.frontiersin.org

Table 3. Experimental comparative analysis on the low-light dataset.

Medical images usually have high resolution and complex textures and structures, and the annotation cost is relatively high. Remote sensing images have extensive geographical coverage, multispectral information, and complex backgrounds. The images in these fields have significant differences in data distribution, feature space, and annotation methods. We conduct experiments using public medical image datasets (the ISIC skin lesion dataset). The dataset comprises high-resolution images of various skin lesions, accompanied by detailed and accurate annotations. We employ the transfer learning method and utilize a pre-trained model for the task of medical image classification. The result is shown in Table 4.

Table 4
www.frontiersin.org

Table 4. Experimental comparative analysis on the ISIC skin lesion dataset.

The experimental results show that through transfer learning. However, the results have declined somewhat compared with those of other datasets; the model can still achieve good performance in the task of medical image classification.

4.3.2 Experimental comparison on other datasets

To evaluate the classification performance on fine-grained datasets, this experiment is compared with several methods. As shown in Tables 57, the tables present the performance of different models for the 5-way 1-shot and 5-way 5-shot classification tasks on three fine-grained datasets.

Table 5
www.frontiersin.org

Table 5. Experimental comparative analysis on the CUB dataset.

Table 6
www.frontiersin.org

Table 6. Experimental comparative analysis on the dogs dataset.

Table 7
www.frontiersin.org

Table 7. Experimental comparative analysis on the cars dataset.

As shown in the table, in the CUB dataset, the proposed model achieves the best performance in the 1-shot task, with an accuracy that is 5.46% higher than that of the sub-optimal model FFSNet, and the accuracy is 2.23% higher than that of the sub-optimal model dSE in the 5-shot task. On the Dogs dataset, the 1-shot task and 5-shot task increase by 2.26 and 0.98%, respectively, compared with the dSE method. However, on the Cars dataset, the accuracy of the proposed model is also higher than that of most methods in the 5-way 1-shot setting; however, it is slightly insufficient in the 5-way 5-shot classification task. This may be due to the lack of local diversity in the dataset, which leads to the overlapping of local areas of identification, thereby limiting the improvement of classification performance. In summary, the proposed method’s accuracy is superior to that of other methods on fine-grained datasets.

4.4 Ablation experiment

To illustrate the effectiveness of each module, we conduct ablation experiments on three fine-grained datasets: CUB, Dogs, and Cars. Under the two settings of 5-way 1-shot and 5-way 5-shot, modules are added sequentially for training, and results are obtained. The evaluation index is Top-1 accuracy. The specific experimental results are shown in Tables 810.

Table 8
www.frontiersin.org

Table 8. Ablation experiments with different modules on CUB.

Table 9
www.frontiersin.org

Table 9. Ablation experiments with different modules on dogs.

Table 10
www.frontiersin.org

Table 10. Ablation experiments with different modules on cars.

According to the table results, in the 5-way 1-shot classification task, the accuracy of the benchmark network increases by 3.08, 4.13, and 8.64%, respectively, after the CIE method is added. Subsequently, with the introduction of the MGM method, the performance further improves by 3.27, 7.12, and 9.55% compared to the previous results. In addition, the inclusion of the FES module enhances model performance by approximately 1.5%. The same experiment is performed on the 5-way 5-shot task, achieving classification accuracies of 87.58, 79.34, and 83.82% on three fine-grained datasets, respectively.

4.5 Complexity analysis of the feature fusion module

In order to verify the lightness of FES, this study uses the Thop method to calculate the number of parameters and the structure calculation amount of the model, where the network structure computation refers to the number of floating-point operations per image (GFLOPI) (Meng et al., 2023) that occur during a single forward propagation when one image is input.

In this study, MogaNet is selected as the backbone network of this method. From the perspective of lightweight, compared to the ResNet series networks, the MogaNet network has a lower number of parameters and calculation amount, as shown in Table 11. It can be seen that the parameter number of MogaNet is 226,177, while the parameter number of ResNet12 is 10,737,672, indicating a significant difference in calculation amount. From the point of view of efficiency, the computing resources of few-shot learning tasks are often limited, and the lightweight MogaNet network can be used to train and reason more efficiently. After adding the proposed FES module based on MogaNet, the number of parameters increases by only 11,536 (approximately 5.1%), which is a small increase and does not significantly increase model complexity. Based on MogaNet, the calculation amount is increased by 1.56 GFLOPI/MB, which indicates that the calculation cost is lower. In summary, it demonstrates that FES can significantly enhance network performance while maintaining a minimal increase in parameters and computational requirements.

Table 11
www.frontiersin.org

Table 11. Complexity comparison analysis.

4.6 Loss function change curve

To evaluate the performance of the proposed method, loss convergence experiments are conducted on the CUB dataset for 5-way 1-shot and 5-way 5-shot settings in this section. As shown in Figure 5, the model loss changes with the increase of the training rounds under different training settings. In the first 50 rounds, the loss value of the two training settings decreases significantly, indicating that the model’s learning effect is obvious. During the intermediate period, the loss continues to decline, with the 1-shot task losing approximately 0.12 at 200 rounds and the 5-shot task losing approximately 0.1 at the same number of rounds. At this point, the model’s learning rate starts to slow down, but it continues to optimize. After 300 rounds, the loss tends to flatten out and reach a stable state. From the overall trend, the loss value of the 5-way 5-shot task is consistently lower than that of the 5-way 1-shot task, indicating that under the same number of training rounds, the model for the 5-shot task outperforms that for the 1-shot task, and its learning ability and stability are also stronger.

Figure 5
Line graph showing loss versus epochs for 5-way 1-shot and 5-way 5-shot learning. Both curves decline sharply initially, then gradually decrease, stabilizing around a loss of 0.1 after 400 epochs. Solid line represents 5-way 1-shot; dashed line represents 5-way 5-shot.

Figure 5. Loss change curve.

4.7 Hyperparameter analysis

In the FES model, parameters τ and δ are two key hyperparameters, which, respectively, control the sensitivity of feature extraction and the weight of feature fusion. To better understand the influence of parameters τ and δ on model performance, we conducted a parameter sensitivity analysis. We set different value ranges for the parameters τ and δ, respectively. τ∈[0.1,1.0], δ∈[0.01,0.1], and the step size is 0.05. We use accuracy to evaluate the model’s performance under different parameter settings on the CUB dataset, as shown in Tables 12, 13. It can be seen from the experimental results that when τ = 0.4 and δ = 0.08, the model yields better results.

Table 12
www.frontiersin.org

Table 12. The effect of τ on the performance of the model.

Table 13
www.frontiersin.org

Table 13. The effect of δ on the performance of the model.

4.8 Visualization experiment

By comparing the response region visualizations of different models, this study demonstrates the degree to which each model focuses on key features in the classification task, aiming to highlight the advantages of the model presented in this study. The original images of two bird species and two dog species are processed using the Grad-CAM algorithm. As shown in Figure 6, the proposed model in this study presents clearer and more concentrated features in the response area compared with Relation Network and BSNet. In the bird dataset, the high response area of the model is concentrated in the beak, wing, and other fine-grained features. In the dog dataset, the model responds strongly to key features of the dog, such as patterns and ears, which are crucial for distinguishing between different dog breeds. In contrast, although the response regions of Relation Network and BSNet cover some key features, the distribution is relatively scattered, lacking the high response concentration and clarity of the method presented in this study. The proposed method in this study effectively combines MGM and FES, improving the model’s response concentration in important regions and enhancing its ability to discriminate subtle image features.

Figure 6
Comparison of four images in three rows. Each row shows: the original photo of an animal on the left, followed by thermal-like visualizations from three different methods labeled as (b) relation network, (c) BSNet, and (d) Proposed. Top row features birds, and the middle and bottom rows feature dogs.

Figure 6. Visual result. (a) original image, (b) relation network method, (c) BSNet method, (d) proposed method.

5 Conclusion

To overcome the limitations of the traditional MogaNet model in capturing local details and the shortcomings of existing image classification methods, this study proposes a novel fine-grained image classification network. The network makes full use of the multi-scale features of the backbone to enhance its ability to capture local details. A contextual information extractor is designed, which uses contextual information and an attention mechanism to guide feature transformation, suppress low-information regions, and enhance important regions. The class prediction score is used to eliminate background features and fuzzy class features, allowing the network to focus on discriminative regions. A variety of loss functions are designed, and the constraint feature extraction and fuzzy feature elimination further improve the network’s performance. The model is evaluated on a few-shot fine-grained benchmark dataset, achieving classification accuracy superior to the majority of state-of-the-art image classification methods. In future research, the combination of transfer learning and fine-grained image recognition methods will be further explored to enhance the model’s performance across various application scenarios.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

DL: Writing – review & editing, Formal analysis, Investigation, Methodology, Conceptualization. SC: Writing – review & editing, Project administration, Software, Writing – original draft, Methodology, Validation, Investigation, Supervision, Conceptualization.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Chen, X., and Wang, G. (2021). Few-shot learning by integrating spatial and frequency representation. 2021 18th Conference on Robots and Vision (CRV). IEEE: 49–56.

Google Scholar

Chen, H., Zhang, H., Liu, C., An, J., Gao, Z., and Qiu, J. (2024). FET-FGVC: feature-enhanced transformer for fine-grained visual classification. Pattern Recogn. 149:110265. doi: 10.1016/j.patcog.2024.110265

Crossref Full Text | Google Scholar

Du, R., Chang, D., Ayan, K., Xie, J., Ma, Z., Song, Y, et al. (2020). Fine-grained visual classification via progressive multi-granularity training of jigsaw patches. European conference on computer vision. Cham: Springer International Publishing, 153–168.

Google Scholar

Elfwing, S., Uchibe, E., and Doya, K. (2018). Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 107, 3–11. doi: 10.1016/j.neunet.2017.12.012

PubMed Abstract | Crossref Full Text | Google Scholar

Folkesson, J., Samset, E., Kwong, R. Y., and Westin, C. F. (2008). Unifying statistical classification and geodesic active regions for segmentation of cardiac MRI. IEEE Trans. Inf. Technol. Biomed. 12, 328–334. doi: 10.1109/TITB.2007.912179

PubMed Abstract | Crossref Full Text | Google Scholar

Ge, W., Lin, X., and Yu, Y. (2019). Weakly supervised complementary parts models for fine-grained image classification from the bottom up. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 3034–3043.

Google Scholar

He, X., and Peng, Y. (2017). Fine-grained image classification via combining vision and language. Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 5994–6002. doi: 10.1109/CVPR.2017.775

Crossref Full Text | Google Scholar

Hossain, S., Umer, S., Rout, R. K., and Tanveer, M. (2023). Fine-grained image analysis for facial expression recognition using deep convolutional neural networks with bilinear pooling. Appl. Soft Comput. 134:109997. doi: 10.1016/j.asoc.2023.109997

Crossref Full Text | Google Scholar

Jiang, Y., and Yin, S. (2023). Heterogenous-view occluded expression data recognition based on cycle-consistent adversarial network and K-SVD dictionary learning under intelligent cooperative robot environment. Comput. Sci. Inf. Syst. 20, 1869–1883. doi: 10.2298/CSIS221228034J

Crossref Full Text | Google Scholar

Kang, D., Kwon, H., Min, J., and Cho, M. (2021). Relational embedding for few-shot classification. Proceedings of the IEEE/CVF International Conference on Computer Vision: 8822–8833.

Google Scholar

Ke, X., Cai, Y., Chen, B., Liu, H., and Guo, W. (2023). Granularity-aware distillation and structure modeling region proposal network for fine-grained image classification. Pattern Recogn. 137:109305. doi: 10.1016/j.patcog.2023.109305

Crossref Full Text | Google Scholar

Ke, X., and Zhang, Y. (2020). Fine-grained vehicle type detection and recognition based on dense attention network. Neurocomputing 399, 247–257. doi: 10.1016/j.neucom.2020.02.101

Crossref Full Text | Google Scholar

Khosla, A., Jayadevaprakash, N., Yao, B., and Li, F. (2011). Novel dataset for fine-grained image categorization: Stanford dogs. Proc. CVPR workshop on fine-grained visual categorization (FGVC), 2(1).

Google Scholar

Kim, J. H., Jun, J., and Zhang, B. T. (2018). Bilinear attention networks. Adv. Neural Inf. Proces. Syst. 31, 1–11. doi: 10.48550/arXiv.1805.07932

Crossref Full Text | Google Scholar

Li, S., Wang, Z., Liu, Z., Tan, C., Lin, H., Wu, D., et al. (2023). Moganet: multi-order gated aggregation network. The Twelfth International Conference on Learning Representations.

Google Scholar

Li, Q., and Wu, W. (2024). Attention-based supervised contrastive learning on fine-grained image classification. Pattern. Anal. Applic. 27:96. doi: 10.1007/s10044-024-01317-5

Crossref Full Text | Google Scholar

Liu, X., Xia, T., Wang, J., Yang, Y., Zhou, F., and Lin, Y. (2016). Fully convolutional attention networks for fine-grained recognition. arxiv. doi: 10.48550/arXiv.1603.06765

Crossref Full Text | Google Scholar

Liu, M., Zhang, C., Bai, H., Zhang, R., and Zhao, Y. (2021). Cross-part learning for fine-grained image classification. IEEE Trans. Image Process. 31, 748–758. doi: 10.1109/TIP.2021.3135477

Crossref Full Text | Google Scholar

Lu, H., and Han, M. (2024). MT-DSNet: mix-mask teacher-student strategies and dual dynamic selection plug-in module for fine-grained image recognition. Comput. Vis. Image Underst. 249:104201. doi: 10.1016/j.cviu.2024.104201

Crossref Full Text | Google Scholar

Ma, L., Hong, H., Meng, F., Wu, Q., and Wu, J. (2023). Deep progressive asymmetric quantization based on causal intervention for fine-grained image retrieval. IEEE Trans. Multimed. 26, 1306–1318. doi: 10.1109/TMM.2023.3279990

Crossref Full Text | Google Scholar

Meng, X., Wang, X., Yin, S., and Li, H. (2023). Few-shot image classification algorithm based on attention mechanism and weight fusion. J. Eng. Appl. Sci. 70:14. doi: 10.1186/s44147-023-00186-9

Crossref Full Text | Google Scholar

Mitsuhata, Y., Uchida, T., Murakami, Y., and Amano, H. (2001). The fourier transform of controlled-source time-domain electromagnetic data by smooth spectrum inversion. Geophys. J. Int. 144, 123–135. doi: 10.1046/j.1365-246x.2001.00324.x

Crossref Full Text | Google Scholar

Nie, X., Chai, B., Wang, L., Liao, Q., and Xu, M. (2023). Learning enhanced features and inferring twice for fine-grained image classification. Multimed. Tools Appl. 82, 14799–14813. doi: 10.1007/s11042-022-13619-z

Crossref Full Text | Google Scholar

Oreshkin, B., Rodríguez López, P., and Lacoste, A. (2018). Tadam: task dependent adaptive metric for improved few-shot learning. Adv. Neural Inf. Proces. Syst. 31, 1–13. doi: 10.48550/arXiv.1805.10123

Crossref Full Text | Google Scholar

Shen, W., and Qin, Y. (2025). LCNet-ViT-FG: a product recognition method based on the fusion of self-supervised pretrained CNN and transformer. J. Electron. Imaging 34:033026. doi: 10.1117/1.JEI.34.3.033026

Crossref Full Text | Google Scholar

Song, Y., Wang, T., Cai, P., Mondal, S. K., and Sahoo, J. P. (2023). A comprehensive survey of few-shot learning: evolution, applications, challenges, and opportunities. ACM Comput. Surv. 55, 1–40. doi: 10.1145/3582688

Crossref Full Text | Google Scholar

Wan, Q., Ji, H., and Shen, L. (2021). Self-attention based text knowledge mining for text detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5983–5992.

Google Scholar

Wang, X., Shi, J., Fujita, H., and Zhao, Y. (2023). Aggregate attention module for fine-grained image classification. J. Ambient. Intell. Humaniz. Comput. 14, 8335–8345. doi: 10.1007/s12652-021-03599-7

Crossref Full Text | Google Scholar

Wang, J., Yu, X., and Gao, Y. (2021). Feature fusion vision transformer for fine-grained visual categorization. arxiv. doi: 10.48550/arXiv.2107.02341

Crossref Full Text | Google Scholar

Wei, X. S., Wang, P., Liu, L., Shen, C., and Wu, J. (2019). Piecewise classifier mappings: learning fine-grained learners for novel categories with few examples. IEEE Trans. Image Process. 28, 6116–6125. doi: 10.1109/TIP.2019.2924811

Crossref Full Text | Google Scholar

Wu, J., Chang, D., Sain, A., Li, X., Ma, Z., Cao, J., et al. (2023). Bi-directional feature reconstruction network for fine-grained few-shot image classification. Proc. AAAI Conf. Artif. Intell. 37, 2821–2829. doi: 10.1609/aaai.v37i3.25383

Crossref Full Text | Google Scholar

Xu, Q., Wang, J., Jiang, B., and Luo, B. (2023). Fine-grained visual classification via internal ensemble learning transformer. IEEE Trans. Multimedia 25, 9015–9028. doi: 10.1109/TMM.2023.3244340

Crossref Full Text | Google Scholar

Xu, W., Xian, Y., Wang, J., Bernt, S., and Akata, Z. (2020). Attribute prototype network for zero-shot learning. Adv. Neural Inf. Proces. Syst. 33, 21969–21980. doi: 10.48550/arXiv.2008.08290

Crossref Full Text | Google Scholar

Yan, L., Li, F., Zhang, L., and Zheng, X. (2023). Discriminant space metric network for few-shot image classification. Appl. Intell. 53, 17444–17459. doi: 10.1007/s10489-022-04413-3

Crossref Full Text | Google Scholar

Yin, S., Li, H., Laghari, A. A., Gadekallu, T. R., Sampedro, G. A., and Almadhor, A. (2024b). An anomaly detection model based on deep auto-encoder and capsule graph convolution via sparrow search algorithm in 6G internet-of-everything. IEEE Internet Things J. 11, 29402–29411. doi: 10.1109/JIOT.2024.3353337

Crossref Full Text | Google Scholar

Yin, S., Li, H., Sun, Y., and Muhammad, I. (2024a). Data visualization analysis based on explainable artificial intelligence: a survey. IJLAI Trans. Sci. Eng. 2, 13–20.

Google Scholar

Yin, H., Arash, V., Jose, A., Arun, M., Jan, K., and Pavlo, M. (2022). A-vit: adaptive tokens for efficient vision transformer. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition: 10809–10818.

Google Scholar

Yuan, L., Chen, Y., Wang, T., Yu, W., Shi, Y., Jiang, Z, et al. (2021). Tokens-to-token vit: training vision transformers from scratch on imagenet. Proceedings of the IEEE/CVF international conference on computer vision. 558–567.

Google Scholar

Zhai, A., and Wu, H. Y. (2018). Classification is a strong baseline for deep metric learning. arxiv. doi: 10.48550/arXiv.1811.12649

Crossref Full Text | Google Scholar

Zhang, Y., Cui, G., Wu, J., Pan, W.-T., and He, Q. (2016). A novel multi-scale cooperative mutation fruit fly optimization algorithm. Knowl.-Based Syst. 114, 24–35. doi: 10.1016/j.knosys.2016.09.027

Crossref Full Text | Google Scholar

Zhang, L., Fu, M., and Wang, Y. (2024). SProtoNet: self-supervised ProtoNet for plant leaf disease few-shot classification. Fourth International Conference on Computer Vision and Data Mining (ICCVDM 2023). SPIE, 13063: 138–141.

Google Scholar

Zhang, Z., Liu, Y., Wang, X., Li, B., and Hu, W. (2021). Learn to match: automatic matching network design for visual tracking. Proceedings of the IEEE/CVF international conference on computer vision. 13339–13348.

Google Scholar

Zhang, W., Zhao, Y., Gao, Y., and Sun, C. (2024). Re-abstraction and perturbing support pair network for few-shot fine-grained image classification. Pattern Recogn. 148:110158. doi: 10.1016/j.patcog.2023.110158

Crossref Full Text | Google Scholar

Zhao, H., Ren, K., Yue, T., Zhang, C., and Yuan, S. (2024). TransFG: a cross-view geo-localization of satellite and UAVs imagery pipeline using transformer-based feature aggregation and gradient guidance. IEEE Trans. Geosci. Remote Sens. 62, 1–12. doi: 10.1109/TGRS.2024.3352418

Crossref Full Text | Google Scholar

Zhong, Y., Li, B., Tang, L., Kuang, S., Wu, S., and Ding, S. (2022). Detecting camouflaged object in frequency domain. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition: 4504–4513.

Google Scholar

Keywords: fine-grained image classification, MogaNet network, multi-level gating mechanism, feature elimination strategy, loss function

Citation: Li D and Chen S (2025) Fine-grained image classification using the MogaNet network and a multi-level gating mechanism. Front. Neurorobot. 19:1630281. doi: 10.3389/fnbot.2025.1630281

Received: 17 May 2025; Accepted: 14 July 2025;
Published: 06 August 2025.

Edited by:

Xianmin Wang, Guangzhou University, China

Reviewed by:

Weichuan Zhang, Griffith University, Australia
Tianhao Gu, Qingdao University, China

Copyright © 2025 Li and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Su Chen, Y2hlbnN1bGhAcXEuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.