ORIGINAL RESEARCH article

Front. Plant Sci.

Sec. Sustainable and Intelligent Phytoprotection

Volume 16 - 2025 | doi: 10.3389/fpls.2025.1588901

DualCMNet: A Lightweight Dual-branch Network for Maize Variety Identification based on Multi-modal Feature Fusion

Provisionally accepted
Xinhua  BiXinhua Bi1Hao  XieHao Xie1Ziyi  SongZiyi Song2Jinge  LiJinge Li1Chang  LiuChang Liu1Xiaozhu  ZhouXiaozhu Zhou1Helong  YuHelong Yu1Chunguang  BiChunguang Bi1*Ming  ZhaoMing Zhao3*
  • 1College of Information Technology, Jilin Agriculture University, Changchun, China
  • 2School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, Jilin Province, China
  • 3Jilin Zhongnong Sunshine Data Company Limited, Changchun, Hebei Province, China

The final, formatted version of the article will be published soon.

The accurate identification of maize varieties is of great significance to modern agricultural management and breeding program. However, traditional maize seed classification methods mainly rely on single modal data, which limits the accuracy and robustness of classification. In addition, existing multimodal methods face high computational complexity, making it difficult to balance accuracy and efficiency. Based on multi-modal data from 11 maize varieties, this paper presents DualCMNet, a novel dual-branch deep learning framework that achieved accurate maize seed variety classification by utilizing a one-dimensional convolutional neural network (1D-CNN) for hyperspectral data processing and a MobileNetV3 network for spatial feature extraction from images. Furthermore, to enhance feature expression capability and reduce computational overhead, the framework introduces three key improvements: the HShuffleBlock feature transformation module implements feature dimension alignment and information interaction; the Channel and Spatial Attention Mechanism (CBAM) enhances the expression of key features; and the lightweight gated fusion module dynamically adjusts feature weights through a single gate value to improve fusion effects. During the training process, this study utilizes pre-trained 1D-CNN and MobileNetV3 models for network initialization and implements a staged training strategy, first optimizing non-pre-trained layers, then unfreezing pre-trained layers and applying differentiated learning rates for fine-tuning. Through 5-fold cross-validation evaluation, the method achieved a classification accuracy of 98.75% on the validation set, significantly outperforming single-modal methods, while the total model parameters are only 2.53M, achieving low computational overhead while ensuring high accuracy. This lightweight design enables the model to be deployed in edge computing devices, allowing for real-time identification in the field, thus meeting the practical application requirements in agricultural Internet of Things and smart agriculture scenarios. This study not only provides an accurate and efficient solution for maize seed variety identification but also establishes a universal framework that can be extended to variety classification tasks of other crops.

Keywords: Maize variety classification, Dual-branch network, Lightweight Network, Hyperspectral and image data, Multi-modal fusion

Received: 06 Mar 2025; Accepted: 27 Apr 2025.

Copyright: © 2025 Bi, Xie, Song, Li, Liu, Zhou, Yu, Bi and Zhao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Chunguang Bi, College of Information Technology, Jilin Agriculture University, Changchun, China
Ming Zhao, Jilin Zhongnong Sunshine Data Company Limited, Changchun, Hebei Province, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.