ORIGINAL RESEARCH article
Front. Plant Sci.
Sec. Sustainable and Intelligent Phytoprotection
Volume 16 - 2025 | doi: 10.3389/fpls.2025.1588901
DualCMNet: A Lightweight Dual-branch Network for Maize Variety Identification based on Multi-modal Feature Fusion
Provisionally accepted- 1College of Information Technology, Jilin Agriculture University, Changchun, China
- 2School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, Jilin Province, China
- 3Jilin Zhongnong Sunshine Data Company Limited, Changchun, Hebei Province, China
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
The accurate identification of maize varieties is of great significance to modern agricultural management and breeding program. However, traditional maize seed classification methods mainly rely on single modal data, which limits the accuracy and robustness of classification. In addition, existing multimodal methods face high computational complexity, making it difficult to balance accuracy and efficiency. Based on multi-modal data from 11 maize varieties, this paper presents DualCMNet, a novel dual-branch deep learning framework that achieved accurate maize seed variety classification by utilizing a one-dimensional convolutional neural network (1D-CNN) for hyperspectral data processing and a MobileNetV3 network for spatial feature extraction from images. Furthermore, to enhance feature expression capability and reduce computational overhead, the framework introduces three key improvements: the HShuffleBlock feature transformation module implements feature dimension alignment and information interaction; the Channel and Spatial Attention Mechanism (CBAM) enhances the expression of key features; and the lightweight gated fusion module dynamically adjusts feature weights through a single gate value to improve fusion effects. During the training process, this study utilizes pre-trained 1D-CNN and MobileNetV3 models for network initialization and implements a staged training strategy, first optimizing non-pre-trained layers, then unfreezing pre-trained layers and applying differentiated learning rates for fine-tuning. Through 5-fold cross-validation evaluation, the method achieved a classification accuracy of 98.75% on the validation set, significantly outperforming single-modal methods, while the total model parameters are only 2.53M, achieving low computational overhead while ensuring high accuracy. This lightweight design enables the model to be deployed in edge computing devices, allowing for real-time identification in the field, thus meeting the practical application requirements in agricultural Internet of Things and smart agriculture scenarios. This study not only provides an accurate and efficient solution for maize seed variety identification but also establishes a universal framework that can be extended to variety classification tasks of other crops.
Keywords: Maize variety classification, Dual-branch network, Lightweight Network, Hyperspectral and image data, Multi-modal fusion
Received: 06 Mar 2025; Accepted: 27 Apr 2025.
Copyright: © 2025 Bi, Xie, Song, Li, Liu, Zhou, Yu, Bi and Zhao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Chunguang Bi, College of Information Technology, Jilin Agriculture University, Changchun, China
Ming Zhao, Jilin Zhongnong Sunshine Data Company Limited, Changchun, Hebei Province, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.