Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci.

Sec. Sustainable and Intelligent Phytoprotection

Adaptive Multi-Scale Feature Refinement for Wheat Phenology Recognition Using Cross-Scale Attention Mechanisms

Provisionally accepted
Haifang  SunHaifang Sun1Liang  HouLiang Hou1Xiaorui  GuoXiaorui Guo2Yanan  WangYanan Wang3Jianan  MinJianan Min1Xiaoliu  ZhengXiaoliu Zheng1Zhehao  TianZhehao Tian1Donghui  ZhangDonghui Zhang4Xinshi  ZhangXinshi Zhang1Senhao  LiuSenhao Liu5Yu  GaoYu Gao6Zhaoli  AnZhaoli An1Hao  QiHao Qi7*Liangjie  LvLiangjie Lv2*
  • 1Institute of Agricultural Information and Economy, Hebei Academy of Agriculture and Forestry Sciences, Shijiazhuang, China
  • 2Institute of Cereal and Oil Crops, Hebei Academy of Agriculture and Forestry Sciences, Shijiazhuang, China
  • 3Hebei Agricultural Technology Extension Station, Shijiazhuang, China
  • 4Institute of Remote Sensing Satellite, China Academy of Space Technology, Beijing, China
  • 5Chongqing University, Chongqing, China
  • 6Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China
  • 7Hebei Academy of Agriculture and Forestry Sciences, Shijiazhuang, China

The final, formatted version of the article will be published soon.

Accurate delineation of crop growth stages under real-world field conditions remains a long-standing challenge in computational phenotyping, particularly for wheat whose developmental phases are characterized by subtle, continuous morphological transitions and environmental noise. In this study, we propose AMFR-Net, an Adaptive Multi-Scale Feature Refinement Network tailored for fine-grained wheat stage identification using ground-level RGB imagery. Unlike conventional architectures that struggle with ambiguous inter-stage boundaries and rigid receptive structures, AMFR-Net leverages a ResNet-101 backbone augmented by a novel Adaptive Multi-Scale Attention Fusion (AMSAF) module—comprising cross-scale interaction blocks and confidence-weighted feature aggregation—to hierarchically recalibrate spatial–semantic representations. This design enables the network to adaptively amplify phenologically salient cues while suppressing irrelevant context, ensuring robust generalization under constrained annotation and deployment conditions. Evaluated on the expert-labeled CGIAR benchmark, AMFR-Net achieves state-of-the-art performance across all major metrics (Top-1 Accuracy: 89.10%; Macro-F1: 89.10%; AUC: 97.88%) and demonstrates superior discriminability in phenologically adjacent stages compared to lightweight and deep CNN baselines. Ablation studies validate the synergistic effect of multi-level attention and scale-aware refinement. The proposed framework offers a scalable, interpretable, and field-deployable solution for in-situ phenology monitoring, and sets a foundation for future integration of multimodal sensing, weak supervision, and cross-seasonal adaptation.

Keywords: AMFR-Net6, Deep Visual Recognition3, Edge Deployment4, Field-Based RGB Imagery5, Growth Stage Classification1, Multi-Scale Attention2, Precision Agriculture7

Received: 23 Oct 2025; Accepted: 09 Feb 2026.

Copyright: © 2026 Sun, Hou, Guo, Wang, Min, Zheng, Tian, Zhang, Zhang, Liu, Gao, An, Qi and Lv. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Hao Qi
Liangjie Lv

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.