Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci.

Sec. Sustainable and Intelligent Phytoprotection

This article is part of the Research TopicAdvanced Imaging and Phenotyping for Sustainable Plant Science and Precision Agriculture 4.0View all 11 articles

MDFE-Net: A Multiscale Dilated Feature Enhancement Network for Small Object Detection

Provisionally accepted
Tianzhe  LiuTianzhe Liu1Shihang  LinShihang Lin2Jiayi  ZhangJiayi Zhang2Bin  LiBin Li2*Junyan  ZhuJunyan Zhu2*
  • 1Fujian Police College, Fuzhou, China
  • 2Fujian Agriculture and Forestry University, Fuzhou, China

The final, formatted version of the article will be published soon.

Due to insufficient feature representation and complex backgrounds, small object detection remains a significant challenge. To solve these problems, this paper proposes a novel small object detection framework named Multiscale Dilated Feature Enhancement Network (MDFE-Net), comprising two innovative plug-and-play modules: Multi-scale Dilated Feature Aggregation (MDFA) and Context Feature Enhancement (CFE). Specifically, the MDFA module efficiently integrates multi-scale features via dilated convolution, capturing rich contextual information and enhancing underlying feature representations. CFE improves the local feature perception and preserves and extracts informative cues for small objects to the greatest extent. Consequently, the proposed network significantly improves small object perception and effectively mitigates interference from complex backgrounds. Comprehensive experiments were conducted on two public datasets (VisDrone and GTSDB) and one self-constructed agricultural dataset (PSD-Node). On the above three datasets, the AP50 of MDFE-Net reached 0.304, 0.952, and 0.895, and the AP is 0.172, 0.805, and 0.476, respectively, which exceeded the benchmark model and the current SOTA method.

Keywords: Context feature, Dilated Convolution, Feature enhancement, Multiscale, Small object detection

Received: 31 Dec 2025; Accepted: 05 Feb 2026.

Copyright: © 2026 Liu, Lin, Zhang, Li and Zhu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Bin Li
Junyan Zhu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.