DATA REPORT article

Front. Plant Sci., 22 August 2022

Sec. Sustainable and Intelligent Phytoprotection

Volume 13 - 2022 | https://doi.org/10.3389/fpls.2022.936748

Multi-modal and multi-view image dataset for weeds detection in wheat field

  • 1. College of Agriculture, Nanjing Agricultural University, Nanjing, China

  • 2. National Engineering and Technology Center for Information Agriculture, Nanjing, China

  • 3. Engineering Research Center of Smart Agriculture, Ministry of Education, Nanjing, China

  • 4. Collaborative Innovation Center for Modern Crop Production Co-sponsored by Province and Ministry, Nanjing, China

  • 5. College of Artificial Intelligence, Nanjing Agricultural University, Nanjing, China

Article metrics

View details

9

Citations

4k

Views

1,1k

Downloads

Introduction

Weeds in wheat fields compete with wheat for light, water, fertilizer, and growth space, and therefore are one of the main biohazards that limit the yield and quality formation of wheat (Shiferaw et al., 2013; Singh et al., 2013). Obtaining weed species and location information quickly and accurately is the first step in precise weed control. Existing methods for detecting weeds in wheat fields based on machine learning are highly dependent on the scale and quality of datasets. In particular, deep learning methods usually require massive training samples. However, Multi-modal and Multi-View Image of weed datasets in natural wheat field are very rare currently.

Existing public weed datasets in crop field are shown in Table 1. Compared with other crops such as cotton and sugar beet, the plant space for wheat crop is smaller. Therefore, the background in wheat field is more complex, and the labeling process is more difficult. In addition, the existing weed datasets still have the following drawbacks. First, the number of samples is small, which is difficult to meet the requirements of deep learning. The mode of information is very single and limited. The existing datasets are constructed based on RGB images, which cannot provide a more complete feature space for weed detection. Especially detecting weeds in wheat field, the features extracted from RGB images are easy to detect broadleaf weeds, but it is difficult to detect weeds with similar appearance to wheat. Studies have shown that because of growth competition, height information is an important parameter to discriminate wheat from weeds (Fahad et al., 2015; Xu et al., 2020), and the fusion of RGB image features and height features has become a new method used to improve the efficiency of weed detection. Finally, the existing datasets are mostly based on vertical views for image acquisition (Wu et al., 2021), which is difficult to be applied to weed detection under a complex field background. Leaf overlap and occlusion will have a great impact on detection.

Table 1

DatasetPurposePlantsDescriptionReferences
Dataset of annotated food crops and weed imagesWeeds detection and controlCommon beet, carrot, zucchini, pumpkin, radish, radish and 8 weed species1,118 images with 7,853 XML manually annotated annotationsSudars et al., 2020
A crop/weed field image datasetInstance segmentation for weeds and plantsCarrot and common weeds in North Germany60 images with annotationsHaug and Ostermann, 2015
2016 sugar beets datasetClassification of weeds and plantsSugar Beet and common weeds in Germany4-channel multi-spectral imagesChebrolu et al., 2017
Early-crop-weedClassification of weeds and plantstomato, cotton, velvetleaf and black nightshade766 field images of crops in early stageEspejo-Garcia et al., 2020
Deep weedsClassification of multiple weeds speciesEight nationally significant weed species17,509 images with annotationsOlsen et al., 2019
Plant seedlings datasetClassification of weeds and cropsMaize, wheat, sugar beet and nine weed species5,539 images with annotationsGiselsson et al., 2017
CNU weeds datasetClassification of multiple weeds species21 weeds species in the Republic of Korea208,477 images with annotationsVo Hoang et al., 2020
Carrot-weedsWeeds detectionCarrots and unspecified weeds39 images with annotationsLameski et al., 2017
Lincoln BeetWeeds detectionSugar beet and unspecified weeds4,402 images with annotationsSalazar-Gomez et al., 2021
Cobbity WheatWeeds detectionWheat and two weed species101 images with annotationsColeman, 2021
Radish Wheat DatasetWeeds detectionFour growth stages wheat and four weed species552 images with annotationsRayner, 2022
Crop and weedInstance segmentation for weeds and plantsMaize, the common bean and a variety of weeds2,489 images with annotationsChamp et al., 2020

Public weed datasets in crop field.

Therefore, we proposed Multi-modal and Multi-view Image Dataset for Weeds Detection in Wheat Field (MMIDDWF) that can be used for deep learning. The dataset contains wheat, broad-leaf weed, and grass weed images of two modes and nine views, and aims to provide a public weed dataset to promote the development of weed detection methods in wheat field.

Value of the data

  • A multimodal image dataset will be provided for weed detection in open wheat field, including an RGB image and a depth image of the same scene. Compared with a single RGB image, a depth image can provide three-dimensional structure features for weed detection in wheat field, which is helpful to solve the problem of detecting grass weeds.

  • The dataset also contains multi-view images. Images from nine views can provide a more complete feature space for weed detection in open wheat field, thus helping to solve weed detection problems under a complex background such as leaf occlusion and overlapping.

Materials and methods

Experiment design and image acquisition

Experiments on wheat and weeds were carried out from December 2017 to April 2021 at the demonstration base of the National Engineering and Technology Center for Information Agriculture in Rugao county, Nantong City, Jiangsu province, China. Weeds were not controlled during field management, and seeds of six weed species commonly associated with wheat were randomly sown to simulate weed growth in the open field. Alopecurus aequalis, Poa annua, Bromus japonicus, and E. crusgalli are grass weeds; Amaranthus retroflexus and C. bursa-pastoris are broad-leaf weeds; species composition was similar to that of actual weed species in wheat fields.

The data were collected at the peak of weed occurrence in wheat fields, i.e., in the tillering and jointing stages. The acquisition equipment is shown in Figure 1. The RGB and depth images were acquired using Intel® RealSense™ Depth Camera D415 (Integrated Electronics Corporation, Santa Clara, CA, United States), an RGB-D camera that adopts active infrared stereo vision technology. Infrared stereo cameras generate depth images, and the color sensor generates RGB images, both with a resolution of 1,280 × 720.

Figure 1

Figure 1

(A) Image acquisition equipment, (B) Intel® RealSense™ Depth Camera D415, and (C) TL-IPC44AN-4camera.

RGB and depth field images under natural conditions were obtained during the wheat tillering and jointing stages. Multi-view images were collected with TL-IPC44AN-4(TP-Link Corporation, Shenzhen, China) cameras in four positions at angles of 15 and 30° horizontally. Image collection was conducted from 9 a.m. to 4 p.m. under clear and windless weather conditions. The camera was 70 cm above the crop canopy. Images were transmitted to a computer in real time via USB 3.0.

Image annotation and dataset production

The original depth information is less representative. In particular, feature extraction from depth images with a convolutional neural network generates feature maps of distance rather than geometric structures with physical significance. Therefore, single-channel depth images were transformed to three-channel images by re-encoding the original images to make them more representative and structurally similar to RGB images (Gupta et al., 2014). The three channels of re-encoded images are phase, height above ground, and angle with gravity, and re-encoded images are referred to as PHA images. For the image re-encoding method, refer to Xu et al. (2021). Therefore, the multimodal image dataset in MMIDDWF includes three parts: RGB images, single-channel depth images corresponding to RGB images, and PHA images obtained by recoding depth images. In MMIDDWF, each type includes 1,288 images measuring 500 × 500 pixels. The multi-view image dataset contains 692 images, including 79 RGB images from vertical view, 79 vertical depth images corresponding to vertical view RGB images, and 534 images from eight other views. The details of the dataset are shown in Table 2. LabelImg is employed to annotate broadleaf and grass weeds in images as shown in Figure 2A, and the annotation information of RGB images corresponds to that of depth and PHA images. Figure 2B shows the detection results of weeds in wheat fields based on the multimodal dataset (Xu et al., 2021), which achieved precise detection results in an open wheat field by proposing a dual-channel convolutional neural network and fusing multimodal information.

Table 2

Wheat varieties informationPlanting row spacingWeeds species information
Experiment design
Shengxuan No.6, Sumai No.8, Yangmai No.16 and Yangmai No.2320, 35,and 50 cmFour grass weeds, two broadleaf weeds and other native weeds in wheat fields
Multi-modal image dataset
CameraAngleType and number of imagesImage size
Intel® RealSense™ Depth Camera D415vertical horizontal plane 90°1,288 RGB images and 1,288 PHA images500 × 500
Multi-view image dataset
CameraAngleType and number of imagesImage size
TL-IPC44AN-4 camerawith the horizontal plane is 15° and 30°534 RGB images2,560 × 1,440
Intel® RealSense™ Depth Camera D415vertical horizontal plane 90°79 RGB images and 79 depth images720 × 1280

Details about MMIDDWF.

Figure 2

Figure 2

(A) Labeling of grass and broadleaf weeds in wheat fields using LabelImg and (B) weed detection result in wheat field.

Funding

This study was supported by the National Natural Science Foundation of China (Grant No: 31871524), Modern Agricultural Machinery Equipment and Technology Demonstration and Promotion of Jiangsu Province (Grant No: NJ2021-58), Primary Research and Development Plan of Jiangsu Province of China (Grant No: BE2021304), and Six Talent Peaks Project in Jiangsu Province (Grant No: XYDXX-049).

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Statements

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://github.com/cocococoxu/MMDWWF.

Author contributions

JN, YZ, KX, and WC designed the research. KX, ZJ, QL, and QX conducted the experiment. KX and ZJ analyzed the data and wrote the manuscript. All authors have read and approved the final version of the manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    ChampJ.Mora-FallasA.GoëauH.Mata-MonteroE.BonnetP.JolyA. (2020). Instance segmentation for the fine detection of crop and weed plants by precision agricultural robots. Appl. Plant Sci.8, e11373. 10.1002/aps3.11373

  • 2

    ChebroluN.LottesP.SchaeferWinterhalter, Burgard Stachniss. (2017). Agricultural robot dataset for plant classification, localization and mapping on sugar beet fields. Int. J. Robot Res. 36, 10451052. 10.1177/0278364917720510

  • 3

    ColemanG. (2021). 20201014 - Cobbity Wheat BFLY. Weed-AI. Available online at: https://weed-ai.sydney.edu.au/datasets/73468c19-b098-406a-86fa-df172caaec16.

  • 4

    Espejo-GarciaB.MylonasN.AthanasakosL.FountasS.VasilakoglouI. (2020). Towards weeds identification assistance through transfer learning. Comput. Electron. Agric.171, 105306. 10.1016/j.compag.2020.105306

  • 5

    FahadS.HussainS.ChauhanB. S.SaudS.WuC.HassanS.et al. (2015). Weed growth and crop yield loss in wheat as influenced by row spacing and weed emergence times. Crop Protect.71, 101108. 10.1016/j.cropro.2015.02.005

  • 6

    GiselssonT. M.JørgensenR. N.JensenP. K.DyrmannM.MidtibyH. S. (2017). A public image database for benchmark of plant seedling classification algorithms. arXiv. 1711.05458.

  • 7

    GuptaS.GirshickR.ArbelaezP.MalikJ. (2014). “Learning rich features from RGB-D images for object detection and segmentation,”13th European Conference on Computer Vision (ECCV). Zurich, Switzerland (Cham: Springer). p. 345360.

  • 8

    HaugS.OstermannJ. (2015). “A crop/weed field image dataset for the evaluation of computer vision based precision agriculture tasks,”13th European Conference on Computer Vision (ECCV). Zurich, Switzerland (Cham: Springer). p.105116.

  • 9

    LameskiP.ZdravevskiE.TrajkovikV.KulakovA. (2017). “Weed detection dataset with RG, images taken under variable light conditions,”International Conference on IC, Innovations. (Cham: Springer). p.112119.

  • 10

    OlsenA.KonovalovD. A.PhilippaB.RiddP.WoodJ. C.JohnsJ.et al. (2019). deepweeds: a multiclass weed species image dataset for deep learning. Sci. Rep.9. 10.1038/s41598-018-38343-3

  • 11

    RaynerG. (2022). RadishWheatDataset. Weed-AI. Available online at: https://weed-ai.sydney.edu.au/datasets/8b8f134f-ede4-4792-b1f7-d38fc05d8127.

  • 12

    Salazar-GomezA.DarbyshireM.GaoJ.SklarE. I.ParsonsS. (2021). Towards practical object detection for weed spraying in precision agriculture. arXiv. 2109.11048. 10.48550/arXiv.2109.11048

  • 13

    ShiferawB.SmaleM.BraunH. J.DuveillerE.MurichoG. (2013). Crops that feed the world 10. Past successes and future challenges to the role played by wheat in global food security. Food Sec.5. 10.1007/s12571-013-0263-y

  • 14

    SinghV.SinghaH.RaghubanshiA. S. (2013). Competitive interactions of wheat with Phalaris minor or Rumex dentatus: a replacement series study. Pans Pest Articles News Summaries.59, 245258. 10.1080/09670874.2013.845320

  • 15

    SudarsK.JaskoJ.NamatevsI.OzolaL.BadaukisN. (2020). Dataset of annotated food crops and weed images for robotic computer vision control. Data in Brief.31, 105833. 10.1016/j.dib.2020.105833

  • 16

    Vo HoangT.Gwang-HyunY.Dang ThanhV.Jin-YoungK. (2020). Late fusion of multimodal deep neural networks for weeds classification. Comput. Electron. Agric.175. 10.1016/j.compag.2020.105506

  • 17

    WuZ.ChenY.ZhaoB.KangX.DingY. (2021). Review of weed detection methods based on computer vision. Sensors.21, 3647. 10.3390/s21113647

  • 18

    XuK.LiH.CaoW.ZhuY.NiJ. (2020). Recognition of Weeds in Wheat Fields Based on the Fusion of RGB Images and Depth Images. IEEE Access.11. 10.1109/ACCESS.2020.3001999

  • 19

    XuK.ZhuY.CaoW.JiangX.JiangZ.LiSet al. (2021). Multi-modal deep learning for weeds detection in wheat field based on RGB-D images. Front. Plant Sci.12. 10.3389/fpls.2021.732968

Summary

Keywords

multi-modal image, multi-view image, grass weeds detection, wheat field, machine learning, deep learning

Citation

Xu K, Jiang Z, Liu Q, Xie Q, Zhu Y, Cao W and Ni J (2022) Multi-modal and multi-view image dataset for weeds detection in wheat field. Front. Plant Sci. 13:936748. doi: 10.3389/fpls.2022.936748

Received

05 May 2022

Accepted

20 July 2022

Published

22 August 2022

Volume

13 - 2022

Edited by

Yuzhen Lu, Mississippi State University, United States

Reviewed by

Borja Espejo-Garca, Agricultural University of Athens, Greece; Fengying Dang, Michigan State University, United States; Chetan Badgujar, Kansas State University, United States

Updates

Copyright

*Correspondence: Jun Ni

†These authors have contributed equally to this work and share first authorship

This article was submitted to Sustainable and Intelligent Phytoprotection, a section of the journal Frontiers in Plant Science

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics