<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Plant Sci.</journal-id>
<journal-title>Frontiers in Plant Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Plant Sci.</abbrev-journal-title>
<issn pub-type="epub">1664-462X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpls.2020.534853</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Plant Science</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Sorghum Panicle Detection and Counting Using Unmanned Aerial System Images and Deep Learning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Lin</surname>
<given-names>Zhe</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/908968"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Guo</surname>
<given-names>Wenxuan</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/487447"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Department of Plant and Soil Science, Texas Tech University</institution>, <addr-line>Lubbock, TX</addr-line>, <country>United States</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Department of Soil and Crop Sciences, Texas A&amp;M AgriLife Research</institution>, <addr-line>Lubbock, TX</addr-line>, <country>United States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Spyros Fountas, Agricultural University of Athens, Greece</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Zhanguo Xin, United States Department of Agriculture, United States; Lingfeng Duan, Huazhong Agricultural University, China; Wanneng Yang, Huazhong Agricultural University, China</p>
</fn>
<fn fn-type="corresp" id="fn001">
<p>*Correspondence: Wenxuan Guo, <email xlink:href="mailto:wenxuan.guo@ttu.edu">wenxuan.guo@ttu.edu</email>
</p>
</fn>
<fn fn-type="other" id="fn002">
<p>This article was submitted to Technical Advances in Plant Science, a section of the journal Frontiers in Plant Science</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>02</day>
<month>09</month>
<year>2020</year>
</pub-date>
<pub-date pub-type="collection">
<year>2020</year>
</pub-date>
<volume>11</volume>
<elocation-id>534853</elocation-id>
<history>
<date date-type="received">
<day>14</day>
<month>02</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>17</day>
<month>08</month>
<year>2020</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2020 Lin and Guo</copyright-statement>
<copyright-year>2020</copyright-year>
<copyright-holder>Lin and Guo</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Machine learning and computer vision technologies based on high-resolution imagery acquired using unmanned aerial systems (UAS) provide a potential for accurate and efficient high-throughput plant phenotyping. In this study, we developed a sorghum panicle detection and counting pipeline using UAS images based on an integration of image segmentation and a convolutional neural networks (CNN) model. A UAS with an RGB camera was used to acquire images (2.7&#xa0;mm resolution) at 10-m height in a research field with 120 small plots. A set of 1,000 images were randomly selected, and a mask was developed for each by manually delineating sorghum panicles. These images and their corresponding masks were randomly divided into 10 training datasets, each with a different number of images and masks, ranging from 100 to 1,000 with an interval of 100.&#xa0;A U-Net CNN model was built using these training datasets. The sorghum panicles were detected and counted by a predicted mask through the algorithm. The algorithm was implemented using Python with the Tensorflow library for the deep learning procedure and the OpenCV library for the process of sorghum panicle counting. Results showed the accuracy had a general increasing trend with the number of training images. The algorithm performed the best with 1,000 training images, with an accuracy of 95.5% and a root mean square error (RMSE) of 2.5. The results indicate that the integration of image segmentation and the U-Net CNN model is an accurate and robust method for sorghum panicle counting and offers an opportunity for enhanced sorghum breeding efficiency and accurate yield estimation.</p>
</abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>computer vision</kwd>
<kwd>sorghum panicle</kwd>
<kwd>unmanned aerial systems</kwd>
<kwd>convolutional neural networks</kwd>
<kwd>python</kwd>
<kwd>TensorFlow</kwd>
<kwd>image segmentation</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="2"/>
<equation-count count="6"/>
<ref-count count="75"/>
<page-count count="13"/>
<word-count count="6306"/>
</counts>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<title>Introduction</title>
<p>Sorghum (<italic>Sorghum bicolor</italic> L. Moench) is the fifth top cereal crop in the world, which provides nutrition to humans and livestock, particularly in warm and arid climates (<xref ref-type="bibr" rid="B19">FAO, 1999</xref>). Sorghum is one of the most e&#xfb03;cient crops in the conversion of solar energy and the use of water. It has numerous varieties, including grain sorghums used for human food, and forage sorghum for livestock hay and fodder (<xref ref-type="bibr" rid="B14">Dahlberg et&#xa0;al., 2015</xref>). By measuring the plant population and the weight per panicle, growers can estimate the potential final grain yield (<xref ref-type="bibr" rid="B48">Norman et&#xa0;al., 1995</xref>). However, it is challenging to determine plant population by manually counting sorghum panicles, especially for large fields. Traditional counting methods for yield estimation are mainly focused on hand-sampling in the field, which is tedious, time-consuming, labor-intensive, and prone to human errors. Therefore, it is critical to develop alternative methods to efficiently and accurately count sorghum panicles for determining population and estimating yield.</p>
<p>Technological innovations in platforms and advanced sensors such as unmanned aerial systems (UAS) and efficient image processing capabilities provide opportunities to automate high-throughput plant phenotyping through computer vision. UAS imaging has been widely used in plant phenotyping and precision agriculture-related research. Many low-cost sensors onboard UAS can provide aerial images with centimeter-level spatial resolutions. Further, UAS allows for more flexibility in image acquisition in terms of flight height, flight area, and weather conditions. Different sensors onboard the UAS offer various ways for researchers and growers to characterize plant attributes at different scales. As a result, UAS has become a useful platform for crop growers and researchers to acquire aerial images with high spatial and temporal resolutions for quantifying within-field variations (<xref ref-type="bibr" rid="B24">G&#xf3;mez-Cand&#xf3;n et&#xa0;al., 2014</xref>). For example, RGB (red, green, and blue bands) cameras, multispectral and thermal sensors were applied to estimate LAI (<xref ref-type="bibr" rid="B30">Hunt et&#xa0;al., 2010</xref>; <xref ref-type="bibr" rid="B67">Verger et&#xa0;al., 2014</xref>), biomass (<xref ref-type="bibr" rid="B29">Hunt et&#xa0;al., 2005</xref>; <xref ref-type="bibr" rid="B7">Bendig et&#xa0;al., 2015</xref>), water stress (<xref ref-type="bibr" rid="B21">Gago et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B5">Ballester et&#xa0;al., 2018</xref>), plant height (<xref ref-type="bibr" rid="B7">Bendig et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B16">D&#xed;az-Varela et&#xa0;al., 2015</xref>), plant density (<xref ref-type="bibr" rid="B32">Jin et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B40">Liu et&#xa0;al., 2017</xref>), plant counts (<xref ref-type="bibr" rid="B12">Chen et&#xa0;al., 2017b</xref>; <xref ref-type="bibr" rid="B23">Gn&#xe4;dinger and Schmidhalter, 2017</xref>; <xref ref-type="bibr" rid="B27">Guo et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B50">Olsen et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B49">Oh et&#xa0;al., 2019</xref>), plant and soil temperature (<xref ref-type="bibr" rid="B25">G&#xf3;mez-Cand&#xf3;n et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B75">Zhang et&#xa0;al., 2018</xref>), and plant nitrogen status (<xref ref-type="bibr" rid="B29">Hunt et&#xa0;al., 2005</xref>; <xref ref-type="bibr" rid="B62">Tokekar et&#xa0;al., 2016</xref>). <xref ref-type="bibr" rid="B72">Yang et&#xa0;al. (2017)</xref> provided a review on how UAS remote sensing and multiple sensors were applied in field-based plant phenotyping.</p>
<p>Image segmentation is commonly the first step to extract information of targets from an image by separating a set of pixels containing the objects of interest (<xref ref-type="bibr" rid="B46">Mochida et&#xa0;al., 2018</xref>). The application of image segmentation for plant phenotyping is typically implemented at small scales because the input requires detailed information with accurate labels, which is time-consuming and labor-intensive. Machine learning, together with computer vision, offer opportunities for high-throughput plant phenotyping in recent years. Machine learning can be broadly defined as computational methods to make accurate predictions or improve performance using experience (<xref ref-type="bibr" rid="B47">Mohri et&#xa0;al., 2018</xref>). Deep learning refers to a class of machine learning techniques that leverage multiple layers of non-linear information processing for unsupervised or supervised feature extraction and transformation, and for classification and pattern analysis (<xref ref-type="bibr" rid="B15">Deng et&#xa0;al., 2014</xref>). Deep learning algorithms learn high-level features in an incremental way, which eliminates the need for feature identification and extraction (<xref ref-type="bibr" rid="B38">LeCun et&#xa0;al., 2015</xref>). The deep networks have the capacity to learn complex models that involve crop phenotypic attributes. A variety of vision-based algorithms have been proven effective with high accuracy in plant phenotyping, such as crop or leaf recognition (<xref ref-type="bibr" rid="B58">Sankaran et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B25">G&#xf3;mez-Cand&#xf3;n et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B60">Sladojevic et&#xa0;al., 2016</xref>), disease detection (<xref ref-type="bibr" rid="B6">Barbedo, 2014</xref>; <xref ref-type="bibr" rid="B52">P&#xe9;rez-Ortiz et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B63">Too et&#xa0;al., 2019</xref>), crop classification (<xref ref-type="bibr" rid="B42">Makantasis et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B17">Dyrmann et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B37">Kussul et&#xa0;al., 2017</xref>), and crop or fruit counting (<xref ref-type="bibr" rid="B51">Pape and Klukas, 2015</xref>; <xref ref-type="bibr" rid="B12">Chen et&#xa0;al., 2017b</xref>; <xref ref-type="bibr" rid="B55">Qureshi et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B27">Guo et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B28">Hasan et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B50">Olsen et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B65">Ubbens et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B41">Madec et al., 2019</xref>; <xref ref-type="bibr" rid="B49">Oh et&#xa0;al., 2019</xref>; <xref ref-type="bibr" rid="B71">Xiong et&#xa0;al., 2019</xref>). In recent years, traditional machine learning and deep learning algorithms were used on image segmentation, especially in the areas of biomedical and object detection. For example, <xref ref-type="bibr" rid="B11">Chen et&#xa0;al. (2017a)</xref> developed the Deeplab system and Fully Convolutional Network for semantic image segmentation. <xref ref-type="bibr" rid="B57">Ronneberger et&#xa0;al. (2015)</xref> used a U-Net convolutional neural networks (CNN) algorithm with limited training images for the segmentation of neuronal structures in electron microscopic images. Few studies integrated image segmentation in traditional machine learning or deep learning models for plant phenotyping applications. <xref ref-type="bibr" rid="B31">Islam et&#xa0;al. (2017)</xref> detected potato diseases on individual leaves using image segmentation and the multiclass support vector machine. <xref ref-type="bibr" rid="B70">Wu et&#xa0;al. (2019)</xref> combined image segmentation with VGG-16 CNN on automatic counting of rice seedlings from UAS images. Traditional machine learning and deep learning architectures have been applied to sorghum panicle detection and counting. <xref ref-type="bibr" rid="B27">Guo et&#xa0;al. (2018)</xref> used a two-step, decision-tree-based pixel segmentation model (DTSM), and Support Vector Machine (SVM) method with the Classification Learner in sorghum panicle detection. <xref ref-type="bibr" rid="B50">Olsen et&#xa0;al. (2018)</xref> developed a machine learning algorithm using image annotation to detect and count sorghum panicles with a mean absolute error of 2.66. <xref ref-type="bibr" rid="B22">Ghosal et&#xa0;al. (2019)</xref> proposed a weakly supervised semi-trained CNN model using UAS images for sorghum panicle detection and rough localization. Therefore, image segmentation, together with machine learning, has the potential to detect sorghum panicles and estimate the panicle shape, which can further improve the accuracy of yield prediction.</p>
<p>For effective deep learning algorithms in agricultural applications, model selection and feature definition are critical, which heavily rely on specialized knowledge in both plant phenotyping and computer science (<xref ref-type="bibr" rid="B59">Singh et&#xa0;al., 2016</xref>). Environmental factors such as cloud and windy weather can significantly impact the quality of agricultural images (<xref ref-type="bibr" rid="B22">Ghosal et&#xa0;al., 2019</xref>). In addition, plant phenotyping based on UAS images is also sensitive to plant genotypes, sensor-target angles, overlap among leaves and panicles, panicle damages, and field conditions. As a result, a large number of training images are required to accommodate various environmental conditions to obtain robust and accurate machine learning algorithms for plant phenotyping tasks. However, building a large number of training samples requires a long time and heavy labor. As a result, datasets of crop images are not yet available on a large scale due to the expenses involved in collecting and preparing the corresponding training data. Therefore, it is critical to develop algorithms that determine the appropriate number of images to meet the requirement of accurate plant phenotyping, such as sorghum panicle counting. The objectives of this study were to 1)&#xa0;develop a deep learning CNN image segmentation algorithm to detect and quantify sorghum panicles; 2) evaluate the performance of this algorithm with respect to the number of training images.</p>
</sec>
<sec id="s2" sec-type="materials|methods">
<title>Materials and Methods</title>
<sec id="s2_1">
<title>Experimental Sites</title>
<p>This study was conducted in a research field (33&#xb0; 35&#x2019; 50.53&#x2019;&#x2019; N, 101&#xb0; 54&#x2019; 27.30&#x2019;&#x2019; W) in Lubbock, Texas, in 2018. The climate in this region is semiarid, with an average annual rainfall of 487&#xa0;mm, mostly falling between May and September, frequently as the result of convective thunderstorms (<xref ref-type="bibr" rid="B66">US Climate Data, 2019</xref>). Three sorghum varieties, including NK180, XM217, and KS585 (S&amp;W Seed Company, Sacramento, CA) with two seed populations of 120 and 180, were planted on May 26, 2018. In total, there were 120 plots, each of 6&#xa0;m long and eight rows wide in an east-west direction. A 1.5-m alley was arranged between plots. The NK180 is a drought-tolerant, bird resistant, and early-maturity variety. The NK180 has a whitish color and a relatively large sorghum panicle. The average measured panicle length for this variety in this study was 22&#xa0;cm. The XM217 has a red color and a relatively small sorghum panicle. The average panicle length was 13&#xa0;cm. The KS585 is a drought-tolerant, medium height, and medium-maturity variety. The KS585 has a light brown color, which is close to the soil color, and a relatively small sorghum panicle. The average panicle length was 14&#xa0;cm. A subsurface drip irrigation system was used for irrigation in this field during the growing season.</p>
</sec>
<sec id="s2_2">
<title>UAS Image Collection</title>
<p>A DJI Phantom 4 Pro UAS (DJI, Shenzhen, China) with a 4K RGB camera was applied in image acquisition. The UAS has a 2-axis gimbal that can maintain the orientation of the camera independently from the movement. The UAS is controlled with a 2.4 GHz frequency bidirectional transmission that receives data of the battery voltage, Global Positioning System (GPS) reception, the distance, and the height differences from the home point. The maximum flight duration of the UAS is about 30&#xa0;min. The flight plan was created using the Pix4Dcapture software (Pix4D S.A., Switzerland). The flight plan included 80% front overlap and 80% side overlap. The angle of the camera was set at 90 degrees to the land surface during flight. The UAS was flying at an altitude of 10&#xa0;m at 2.7&#xa0;m s<sup>-1</sup> speed. The spatial resolution was 2.7&#xa0;mm for 10&#xa0;m altitude. Two image datasets were acquired on August 24 and September 10, 2018. All image acquisitions were completed under sunny conditions with light to moderate wind around local solar noon. Raw images were stitched into a whole image using the Pix4DMapper software (Pix4D S.A., Switzerland).</p>
<p>This study applied an integrated method of image segmentation and deep learning for sorghum panicle detection and counting. <xref ref-type="fig" rid="f1">
<bold>Figure 1</bold>
</xref> shows the steps of the algorithm for sorghum panicle detection and counting. The U-Net CNN (<xref ref-type="bibr" rid="B57">Ronneberger et&#xa0;al., 2015</xref>) was adopted as the deep learning framework to train and test the image data.</p>
<fig id="f1" position="float">
<label>Figure 1</label>
<caption>
<p>Flow chart of a sorghum panicle detection and counting algorithm using a U-net Convolutional Neural Networks model on unmanned aerial system images.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g001.tif"/>
</fig>
</sec>
<sec id="s2_3">
<title>Preparing Training Images and Masks</title>
<p>The training images were prepared by randomly cropping the raw UAS images using the Microsoft Paint 3D software (Microsoft Corporation, Redmond, WA). To accurately separate sorghum panicles from other objects in the image, a segmentation mask for each training image was created by encircling the sorghum panicle pixels using the Adobe Photoshop CC software (Adobe Systems Inc., San Jose, CA). Specifically, for each training image of 1024 x 1024 pixels, pixels were divided into two classes, the sorghum panicle class and the non-panicle class. In the mask, the pixels containing sorghum panicles were digitized as white and assigned a value of 1, while the other pixels were set black and assigned the value of 0 (<xref ref-type="fig" rid="f2">
<bold>Figure 2</bold>
</xref>). These mask images were saved separately to ensure that each mask matched its corresponding training image when running the U-Net CNN model. The full training dataset contained 1,000 images. To test the model performance as a function of the number of training images, a series of 10 randomly selected sub-datasets, ranging from 100 to 1,000 with an interval of 100 images (i.e., 100, 200, &#x2026;, 1000 images), were generated from the full training dataset. Each sub-dataset was used to train a model and tested for the accuracy of the panicle count for the corresponding number of training images.</p>
<fig id="f2" position="float">
<label>Figure 2</label>
<caption>
<p>Examples of training images (Top) and corresponding masks (Bottom) for a sorghum panicle detection and counting algorithm using a Convolutional Neural Networks model on unmanned aerial system images (Left to right: NK180, XM217, and KS585).</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g002.tif"/>
</fig>
</sec>
<sec id="s2_4">
<title>U-Net Convolutional Neural Networks</title>
<p>The general procedure of the U-Net CNN in this study is described as follows. The U-Net architecture consists of three sections: the contraction, the bottleneck and the expansion. In this study, there were six blocks in the contraction and the expansion sections. The kernel size was 3 x 3 and the strides were 1 x 1 in the contraction section. In the expansion section, the kernel size was 2 x 2 and the strides were 2 x 2. No padding was applied in either section. In the contraction section, each block contained two convolution layers, followed by a down-sampling layer. Once every pixel was processed after the convolution layers, the result was saved into a new feature map in the same arrangement as the input image. The down-sampling layer was used to reduce the feature map dimension, so only the most essential bits of the feature map were kept. The reduced feature map was then utilized as an input to the next contraction block. The spatial dimensions of the feature maps were halved and the number of feature maps was doubled repeatedly through the down-sampling layer (<xref ref-type="bibr" rid="B26">Guan et&#xa0;al., 2019</xref>; <xref ref-type="bibr" rid="B69">Weng et&#xa0;al., 2019</xref>). The&#xa0;bottleneck layer, which contained two convolution layers but without max pooling, mediated the contraction section and the expansion layer. The data at the bottleneck had the spatial dimension of 32 x 32 with 2048 feature maps. In the expansion section, the block contained two convolution layers followed by an up-sampling layer. After each up-sampling layer, the number of feature maps was halved and the spatial dimensions of the feature maps were doubled to maintain the whole architecture symmetry. In the meantime, the input from the corresponding contraction block was appended to the feature maps. After running all the expansion blocks, the final output feature map with the same spatial dimension as the original input image included the sorghum panicle class and the non-panicle class.</p>
</sec>
<sec id="s2_5">
<title>Segmentation Model Training and Validation</title>
<p>For each training dataset, 90% of the images were set as training, and the rest 10% was used as validation for the training models. For example, in a 500-image dataset, 450 images were trained through the model, and the rest 50 images were used as validation. Before training the segmentation model, all training images and masks went through the image augmentation processes. The hue of each RGB training image was adjusted by a factor of 0.1. Both the training images and corresponding masks were flipped horizontally along the central axis with a 0.5 probability. The training images and corresponding masks were randomly shifted either horizontally or vertically. Finally, both training images and the corresponding masks were rescaled by a factor of 1/1023. The purpose of image augmentation was to increase the amount of training data by applying some transformations to the original training images. This helps the model to generalize better to unseen data and prevent overfitting (<xref ref-type="bibr" rid="B68">Wang and Perez, 2017</xref>; <xref ref-type="bibr" rid="B20">Frid-Adar et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B45">Miko&#x142;ajczyk and Grochowski, 2018</xref>). After the image augmentation process, a two-channel segmentation model was generated from these training images and masks using the U-Net algorithm.</p>    <p>The pixel-wise cross-entropy loss function was used to evaluate the training models of the U-Net CNN algorithm using the 10% validation images in the training datasets. The cross-entropy loss is commonly used as a loss function for training in deep learning networks, especially in image segmentation tasks (<xref ref-type="bibr" rid="B57">Ronneberger et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B61">Sudre et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B44">Martins and Zaglia, 2019</xref>). Cross-entropy loss measures the probability difference between the observed and the predicted values in a classification model (<xref ref-type="bibr" rid="B9">Buja et&#xa0;al., 2005</xref>). The cross-entropy loss (CE) for the binary classification in this study is defined as,</p>
<disp-formula>
<label>(1)</label>
<mml:math display="block" id="M1">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo>-</mml:mo>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>g</mml:mi>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:mo stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>-</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>g</mml:mi>
<mml:mo stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>-</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <italic>y<sub>i</sub>
</italic> represents the labeled value for that sample in the mask, and <bold>
<italic>p<sub>i</sub>
</italic>
</bold> represents the predicted probability being the sorghum panicle in the output feature maps.</p>
</sec>
<sec id="s2_6">
<title>Counting and Evaluations</title>
<p>A test dataset containing 120 images was selected from the fully stitched image for accuracy assessment. The images in the test dataset were different from the images in the training dataset. Each test image was corresponding to two rows of sorghum plants randomly selected from a plot. Sorghum panicles in each test image were manually counted, and the number of sorghum panicles in these test images varied from 95 to 188. The size of each test image was 3800 x 1280. We found out that it was difficult for the U-Net CNN model to process the high-resolution test images directly. In this case, we horizontally split each test image into four non-overlapped subtest images. Before running the model on the test images, the subtest images for each test image were resized to dimensions of 1024 x 1024 pixels. Then the test images were run through the trained segmentation model to perform the panicle detection. Each sorghum panicle detected was treated as a contour using the <italic>findContours</italic> function of the OpenCV library in the prediction output feature map. Our initial assessment indicated that contours with less than six pixels were mainly noise related and not classified as panicles. A bounding polygon was applied around each panicle contour using the <italic>drawContours</italic> function for each subtest image. Therefore, the number of bounding polygons represented the number of predicted sorghum panicles in each subtest image. The summation of sorghum panicles of the four subtest images equaled the total number of sorghum panicles in each test image.</p>
<p>The mean absolute error (MAE), mean absolute percentage error (MAPE), accuracy (ACC), coefficient of the determination (R<sup>2</sup>), and the root mean squared error (RMSE) were used as evaluation metrics to assess the performance of the sorghum panicle counting algorithm.</p>
<disp-formula>
<label>(2)</label>
<mml:math display="block" id="M2">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>|</mml:mo>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(3)</label>
<mml:math display="block" id="M3">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>A</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mrow>
<mml:mo>|</mml:mo> <mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow> <mml:mo>|</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(4)</label>
<mml:math display="block" id="M4">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mi>C</mml:mi>
<mml:mi>C</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo stretchy="false">)</mml:mo>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(5)</label>
<mml:math display="block" id="M5">
<mml:mrow>
<mml:msup>
<mml:mi>R</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="true">&#xaf;</mml:mo>
</mml:mover>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(6)</label>
<mml:math display="block" id="M6">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>=</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where m<sub>i</sub>, <inline-formula>
<mml:math display="inline" id="im1">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi>m</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and c<sub>i</sub> represent the manually counted sorghum panicles for the i<sup>th</sup> image, the mean manual counts, and the predicted count for the i<sup>th</sup> image, respectively. n is the number of test images.</p>
</sec>
<sec id="s2_7">
<title>Hardware and Libraries Used</title>
<p>The algorithm was implemented using the Python programming language (<xref ref-type="bibr" rid="B54">Python Software Foundation, 1995</xref>). The model was trained on a computer with 192 GB of memory at the Texas Tech University High Performance Computing Center (HPCC). Training, evaluation, and testing were performed using the Keras (<xref ref-type="bibr" rid="B13">Chollet, 2015</xref>) high-level neural networks application programming interface (API), running on top of the TensorFlow package (<xref ref-type="bibr" rid="B1">Abadi et&#xa0;al., 2016</xref>). The model in this study was trained using the Adam (<xref ref-type="bibr" rid="B35">Kingma and Ba, 2015</xref>) optimizer with a learning rate of 0.001. Fifteen epochs were performed in the training process. The number of epochs was determined based on the training image size, training required time, and the overall performance of the model. In this study, the cross-entropy loss value did not decrease significantly beyond 15 epochs. The OpenCV-Python library (<xref ref-type="bibr" rid="B8">Bradski, 2000</xref>) was used in model testing.</p>
</sec>
</sec>
<sec id="s3" sec-type="results">
<title>Results</title>
<sec id="s3_1">
<title>Training Model and the Number of Training Images</title>
<p>
<xref ref-type="table" rid="T1">
<bold>Table 1</bold>
</xref> shows the overall decreasing trend of the cross-entropy loss with the number of training images, indicating an increasing accuracy in model performance with the number of training images. The value of cross-entropy loss did not decrease rapidly from 100 to 500 training images. On the other hand, from 600 training images, every 100 more training images resulted in a decrease of more than 0.10 in cross-entropy loss. The value of cross-entropy loss and the trend indicated that there could be potential to improve the performance of the segmentation model by increasing the number of training images. However, due to the restriction of the training period in our HPCC and the CPU memory, the value of the cross-entropy loss used here was based on 15 epochs of training for all different numbers of training images. Many studies have shown that the cross-entropy loss value could be close to 0 with a large number of training epochs (<xref ref-type="bibr" rid="B4">Alom et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B74">Zhang and Sabuncu, 2018</xref>). However, with the number of training epochs, the results of cross-entropy values showed a clear negative trend as the number of training images increased.</p>
<table-wrap id="T1" position="float">
<label>Table 1</label>
<caption>
<p>Cross-entropy loss values for 10 sets of training images for a sorghum panicle detection and counting algorithm using a U-Net Convolutional Neural Networks model on unmanned aerial system images.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">
<bold>No. of Images</bold>
</th>
<th valign="top" align="center">
<bold>100</bold>
</th>
<th valign="top" align="center">
<bold>200</bold>
</th>
<th valign="top" align="center">
<bold>300</bold>
</th>
<th valign="top" align="center">
<bold>400</bold>
</th>
<th valign="top" align="center">
<bold>500</bold>
</th>
<th valign="top" align="center">
<bold>600</bold>
</th>
<th valign="top" align="center">
<bold>700</bold>
</th>
<th valign="top" align="center">
<bold>800</bold>
</th>
<th valign="top" align="center">
<bold>900</bold>
</th>
<th valign="top" align="center">
<bold>1000</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Cross-entropy loss</td>
<td valign="top" align="center">0.71</td>
<td valign="top" align="center">0.66</td>
<td valign="top" align="center">0.63</td>
<td valign="top" align="center">0.59</td>
<td valign="top" align="center">0.53</td>
<td valign="top" align="center">0.50</td>
<td valign="top" align="center">0.39</td>
<td valign="top" align="center">0.34</td>
<td valign="top" align="center">0.20</td>
<td valign="top" align="center">0.11</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3_2">
<title>Sorghum Panicle Counting Performance and the Number of Training Images</title>
<p>
<xref ref-type="fig" rid="f3">
<bold>Figure 3</bold>
</xref> presents the accuracy and coefficient of determination (R<sup>2</sup>) of the model performance in relation to the number of training images. In general, the sorghum panicle count accuracy and R<sup>2</sup> values increased with the number of training images. This trend, however, was not consistent for the cases with training images below 500. The accuracy was low at 59% with 100 training images, increased to 75% and 80% with 200 and 300 training images, respectively, but dropped slightly to 78% with 400 training images. Similarly, the R<sup>2</sup> value was lowest at 0.01 with 100 training images, increased to 0.09 and 0.17 with 200 and 300 training images, but dropped to 0.08 with 400 images. For cases with more than 500 training images, the accuracies and R<sup>2</sup> values consistently increased with the number of training images. With 1000 training images, the highest accuracy of 95.5%, and the highest R<sup>2</sup> value of 0.90 were achieved. In addition, the rate of change in relation to the number of images for R<sup>2</sup> was greater than that for the accuracy. From 500 to 1,000 images, the accuracy increased by 16.5% from 82% to 95.5%, while the R<sup>2</sup> value increased by 900% from 0.09 to 0.90. This indicates the accuracy is a better parameter for evaluating this type of algorithm performance. In summary, the algorithm performance was not stable with less than 500 training images. With more than 500 training images, the algorithm performance steadily improved with respect to accuracy.</p>
<fig id="f3" position="float">
<label>Figure 3</label>
<caption>
<p>Trends of accuracy, Cross-Entropy Loss, and coefficient of determination (R<sup>2</sup>) with the number of training images in a sorghum panicle detection and counting algorithm using a Convolutional Neural Networks model on unmanned aerial system images.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g003.tif"/>
</fig>
<p>MAE, MAPE, and RMSE consistently decreased with the increasing number of training images (<xref ref-type="table" rid="T2">
<bold>Table 2</bold>
</xref>). These trends were not consistent with the trends for the accuracy and R<sup>2</sup> values, which had fluctuations in the relation between the magnitude and number of training images. For a low number of training images, the MAE was relatively large; it was 53.1 for 100 training images and 35.2 for 200 training images. This value dropped to 6.3 for 1,000 images. MAPE was 0.41 for 100 training images and 0.25 for 200 training images, then it dropped to 0.05 for 1,000 training images. A similar trend was observed for the RMSE values. Its change, however, was not as extreme as those MAE values. RMSE was 7.3 for 100 training images, and it gradually dropped to 2.5 for 1,000 training images. Considering the range of sorghum panicles (95 to 188) in the test dataset, the MAE and RMSE values for 1,000 training images are within an acceptable range.</p>
<table-wrap id="T2" position="float">
<label>Table 2</label>
<caption>
<p>Mean absolute error (MAE), mean absolute percentage error (MAPE), and root mean squared error (RMSE) for 10 sets of training images for a sorghum panicle detection and counting algorithm using a U-Net Convolutional Neural Networks model on unmanned aerial system images.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">
<bold>No. of Images</bold>
</th>
<th valign="top" align="center">
<bold>100</bold>
</th>
<th valign="top" align="center">
<bold>200</bold>
</th>
<th valign="top" align="center">
<bold>300</bold>
</th>
<th valign="top" align="center">
<bold>400</bold>
</th>
<th valign="top" align="center">
<bold>500</bold>
</th>
<th valign="top" align="center">
<bold>600</bold>
</th>
<th valign="top" align="center">
<bold>700</bold>
</th>
<th valign="top" align="center">
<bold>800</bold>
</th>
<th valign="top" align="center">
<bold>900</bold>
</th>
<th valign="top" align="center">
<bold>1000</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MAE</td>
<td valign="top" align="center">53.1</td>
<td valign="top" align="center">35.2</td>
<td valign="top" align="center">28.1</td>
<td valign="top" align="center">27.2</td>
<td valign="top" align="center">23.7</td>
<td valign="top" align="center">19.6</td>
<td valign="top" align="center">16.9</td>
<td valign="top" align="center">11.4</td>
<td valign="top" align="center">9.9</td>
<td valign="top" align="center">6.3</td>
</tr>
<tr>
<td valign="top" align="left">MAPE</td>
<td valign="top" align="center">0.41</td>
<td valign="top" align="center">0.25</td>
<td valign="top" align="center">0.20</td>
<td valign="top" align="center">0.21</td>
<td valign="top" align="center">0.18</td>
<td valign="top" align="center">0.15</td>
<td valign="top" align="center">0.13</td>
<td valign="top" align="center">0.09</td>
<td valign="top" align="center">0.07</td>
<td valign="top" align="center">0.05</td>
</tr>
<tr>
<td valign="top" align="left">RMSE</td>
<td valign="top" align="center">7.3</td>
<td valign="top" align="center">5.9</td>
<td valign="top" align="center">5.3</td>
<td valign="top" align="center">5.2</td>
<td valign="top" align="center">4.9</td>
<td valign="top" align="center">4.4</td>
<td valign="top" align="center">4.1</td>
<td valign="top" align="center">3.4</td>
<td valign="top" align="center">3.2</td>
<td valign="top" align="center">2.5</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To better evaluate the algorithm performance with respect to the patterns of over- and under-estimations, <xref ref-type="fig" rid="f4">
<bold>Figure 4</bold>
</xref> shows the error, the difference in sorghum panicles between the model prediction and the manual count result, in relation to the number of training images. If the error is positive, then the algorithm overestimates sorghum panicles; otherwise, the algorithm underestimates. For the cases of 100, 400, 600, 700, and 900 training images, the results represented a mean overestimation of 50.1, 23.7, 20.1, 16.9, and 10.1 panicles, respectively. For 200, 300, and 800 training images, the results represented a mean underestimation of 35.2, 28.0, and 11.4 panicles, respectively. The mean errors were 3.9 and 2.6 for the cases of 500 and 100 images, respectively. However, the variance of prediction results for the 500 training images was larger than that for the 1000 training images. A key to the success of deep learning in object detection tasks is abundant training images. A larger number of training images results in better accuracy and performance (<xref ref-type="bibr" rid="B34">Kamnitsas et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B2">Aggarwal et&#xa0;al., 2019</xref>). Therefore, the accuracy and robustness of this algorithm increased with the number of training images, with 1,000 images providing the best performance.</p>
<fig id="f4" position="float">
<label>Figure 4</label>
<caption>
<p>Distributions of counting errors between predicted and observed sorghum panicles for ten sets of training images using a Convolutional Neural Networks model on unmanned aerial system images.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g004.tif"/>
</fig>
<p>It appeared that the counting accuracy was related to the soil background. <xref ref-type="fig" rid="f5">
<bold>Figures 5</bold>
</xref>
<bold>&#x2013;</bold>
<xref ref-type="fig" rid="f7">
<bold>7</bold>
</xref> show the examples of sorghum panicle detection results for the three varieties with 100, 500, and 1,000 training images. For the case of 100 training images, the prediction was overestimated by 53 on average. This substantial overestimation was due to some soil pixels between plots being counted as sorghum panicles, especially for XM217 and KS585. The sorghum panicle colors of these two varieties were similar to the soil background. Therefore, with only 100 training images, the U-Net CNN algorithm was not able to distinguish the soil and sorghum panicles with similar colors. For the case of 500 training images, both overestimation and underestimation were observed. For XM217, 141 sorghum panicles were predicted compared to 175 values observed. For KS585, the predicted number of sorghum panicles was 163 compared with the observed number of 215. For NK180, 99 sorghum panicles were detected, while the observed was 114. Overlapping sorghum panicles and the misclassification between white soil background and sorghum panicles caused the overestimation and underestimation with 500 training images.</p>
<fig id="f5" position="float">
<label>Figure 5</label>
<caption>
<p>Sample results of sorghum panicle detection for Variety XM217 with 100, 500, and 1000 training images using a Convolutional Neural Networks model and UAS images. Red circles represent underestimation; blue circles represent overestimation compared to the manual masks.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g005.tif"/>
</fig>
<fig id="f6" position="float">
<label>Figure 6</label>
<caption>
<p>Sample results of sorghum panicle detection for Variety KS585 with 100, 500, and 1000 training images using a Convolutional Neural Networks model on UAS images. Red circles represent underestimation; blue circles represent overestimation compared to the manual masks.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g006.tif"/>
</fig>
<fig id="f7" position="float">
<label>Figure 7</label>
<caption>
<p>Sample results of sorghum panicle detection for Variety NK180 with 100, 500, and 1000 training images using a Convolutional Neural Networks model on UAS images. Red circles represent underestimation and blue circles represent overestimation compared to the manual masks.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g007.tif"/>
</fig>
<p>For the case of 1,000 training images, sorghum panicles were&#xa0;overestimated for variety XM217 (168 predicted vs. 175 observed). For KS585, 199 sorghum panicles were predicted compared to 215 observed. For NK180, 106 panicles were predicted compared to 114 observed. The errors in these cases were mainly caused by overlapping sorghum panicles. For some&#xa0;images with bright soil background and leaves, the algorithm could not perfectly separate sorghum panicles from surroundings, which led to the underestimation errors. This situation was more widespread with a small number of training images, especially for sorghum varieties KS585 and NK180, which had bright panicles similar to the soil and shiny leaves.</p>
</sec>
</sec>
<sec id="s4" sec-type="discussion">
<title>Discussion</title>
<p>Previous studies on sorghum panicle detection and counting used points or rectangular bounding boxes to label sorghum panicles for preparing the training datasets and outputting the predicted results (<xref ref-type="bibr" rid="B27">Guo et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B22">Ghosal et&#xa0;al., 2019</xref>; <xref ref-type="bibr" rid="B49">Oh et&#xa0;al., 2019</xref>). For example, <xref ref-type="bibr" rid="B70">Wu et&#xa0;al. (2019)</xref> combined the image segmentation technique and basic CNN algorithm to create a density map of sorghum panicles. The application of image segmentation can exclude the areas that are not directly involved in training dataset preparation and final output. <xref ref-type="bibr" rid="B43">Malambo et&#xa0;al. (2019)</xref> applied a semantic segmentation-based CNN algorithm to separate sorghum panicles from the soil and other parts in images. These machine learning algorithms for sorghum panicle detection were mainly based on image classification. The use of points or bounding boxes does not provide direct information about the sorghum panicle shape and size. Compared to previous similar studies, the U-Net CNN segmentation adopted in this study not only detect but also localize and delineate individual sorghum panicles. Therefore, the use of sorghum panicle masks and deep learning from this study enables the characterization of individual sorghum panicles, leading to more accurate yield estimation. This, however, does not mean sorghum yield can be directly calculated from the images because the sorghum panicles are typically not orthogonal to the UAS sensor during image acquisition. Further research is required to more accurately determine the size and shape of the sorghum panicle if yield prediction is needed.</p>
<p>By using masks, our algorithm also minimized the errors in sorghum panicle detection due to panicle overlaps and mixing with other elements in the image. Agricultural images acquired using UAS typically have a mixture of target items and background elements due to the deformation caused by camera angel and other factors (<xref ref-type="bibr" rid="B33">Kamilaris and Prenafeta-Bold&#xfa;, 2018</xref>; <xref ref-type="bibr" rid="B53">Pradeep et&#xa0;al., 2018</xref>). This makes object detection in computer vision tasks challenging, especially for the multiple overlapping panicles and panicles that are obscured partially by plant leaves (<xref ref-type="bibr" rid="B27">Guo et&#xa0;al., 2018</xref>). <xref ref-type="bibr" rid="B11">Chen et&#xa0;al. (2017a)</xref> used the DeepLabv3+ to detect object boundaries, with a high accuracy using 11,530 high quality pixel-level annotated images. However, this proposed algorithm was only able to separate the object boundaries between two different classes, but could not detect overlapping sorghum panicles described in this study. Similar methods all required a large number of well labeled training images. Compared with these methods, our algorithm was able to separate and count sorghum panicles individually. For example, <xref ref-type="fig" rid="f8">
<bold>Figure 8</bold>
</xref> shows the overlapping panicles situation and the prediction results using training images and corresponding masks. As shown, the algorithm was able to detect overlapping sorghums by providing masks that mark overlapping panicles. This algorithm, however, could not detect all overlapping panicles due to the lack of training masks in such cases. We believe the performance in detecting overlapping panicles can be improved by increasing the number of overlapping training images.</p>
<fig id="f8" position="float">
<label>Figure 8</label>
<caption>
<p>Sample images showing the minimization of errors in sorghum panicle detection due to overlaps using masks and deep learning. Upper images are raw images. Middle images are manually training masks. Bottom ones are predicted masks.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-11-534853-g008.tif"/>
</fig>
<p>One of the limitations encountered in this study was the split of a full image into pieces for counting sorghum panicles due to computation restrictions. Previous studies have also shown such challenges in machine learning and deep learning algorithms to directly process high-resolution images. It is common to crop or split the original large dimension images to smaller images for detecting and counting objects (<xref ref-type="bibr" rid="B3">Aich et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B70">Wu et&#xa0;al., 2019</xref>; <xref ref-type="bibr" rid="B10">Chandra et&#xa0;al., 2020</xref>). This potentially leads to overestimation. In our study, some sorghum panicles were cut into two parts and counted twice because we horizontally split the test image into four sub-images. However, our visual check indicated that most of panicles were not split evenly, resulting in the smaller pieces with less than six pixels not being counted. Therefore, the double counting issue had no significant effect on the accuracy of the algorithm. Future studies are required to address this limitation by adopting more efficient image processing algorithms to avoid potential double counting.</p>
<p>In this study, sorghum panicles with greater contrast in color and brightness with surrounding elements were easily detected and counted, while some other panicles, especially for the variety KS585, were challenging to detect due to their similarity to the surrounding features, including soil and dry leaves. Environmental factors, such as wind and clouds, have a significant impact on UAS image quality, which can affect the performance of deep learning algorithms. Field condition and plant genotypes also affect the accuracy of machine learning tasks (<xref ref-type="bibr" rid="B64">Torres-S&#xe1;nchez et&#xa0;al., 2013</xref>; <xref ref-type="bibr" rid="B56">Rasmussen et&#xa0;al., 2019</xref>). The similar colors between soil background and crops could also cause errors in computer vision tasks (<xref ref-type="bibr" rid="B18">El-Faki et&#xa0;al., 2000</xref>; <xref ref-type="bibr" rid="B39">Lee et&#xa0;al., 2018</xref>). In this study, it appears some soil clusters and leaves were mislabeled as sorghum panicles, probably due to strong sunlight conditions. We acquired most of the images around local noon time. As a result, both the soil surface and some sorghum panicles were relatively bright in full sunlight. In future studies, users may consider acquiring UAS images under relatively soft light environments, such as late afternoon or early morning. For the improvement of the algorithm performance, adding a separate mask for soil pixels can be an effective alternative to separate the sorghum panicles from soil background.</p>
<p>Abundant training datasets are critical for effective deep learning tasks (<xref ref-type="bibr" rid="B15">Deng et&#xa0;al., 2014</xref>; <xref ref-type="bibr" rid="B38">LeCun et&#xa0;al., 2015</xref>), especially for complex computer vision tasks such as sorghum panicle detection and counting. This study provides useful information regarding the number of training images required for such deep learning tasks. The algorithm produces inconsistent predictions and low accuracy with below 500 training images. It is reasonably accurate with 1,000 training images. It is expected that with more training images, the accuracy and robustness can be further improved. <xref ref-type="bibr" rid="B2">Aggarwal et&#xa0;al. (2019)</xref> demonstrated that a large number of training images could improve the performance of the U-Net CNN model, especially in complex models. However, there are not enough public ready-to-use data as training datasets for specific crops and their phenotypic traits. The development of large training datasets for plant phenotyping is time-consuming and labor-intensive. The drawback of the pre-label based algorithm lies in the fact that it is time consuming to prepare these training masks of sorghum panicles. In this study, 1,000 training images and masks were manually prepared and applied to develop the algorithm. It took a considerably longer time to prepare the training datasets compared with previous studies that used dot-labeled training images. The automatic annotation technique has shown its potential in similar algorithms (<xref ref-type="bibr" rid="B73">Zahavy et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B36">Komura and Ishikawa, 2018</xref>; <xref ref-type="bibr" rid="B22">Ghosal et&#xa0;al., 2019</xref>). Predicted outputs from automatic annotation can be used as new training input, which reduces the workload of manual preparation and can improve the efficiency and the robustness of the algorithm.</p>
</sec>
<sec id="s5" sec-type="conclusions">
<title>Conclusions</title>
<p>In this study, we developed an algorithm to integrate deep learning and segmentation to detect and count sorghum panicles using high-resolution UAS images. A dataset of 1,000 randomly selected images and their corresponding manually labeled masks were constructed for training this algorithm. The performance and efficacy of the algorithm were assessed with a different number of subset training images. The performance of the algorithm improved with the number of training images. The performance of the algorithm was not stable with less than 500 training images. With 1,000 training images, the algorithm had the best performance, with an accuracy of 95.5% and an RMSE of 2.5. The algorithm is sufficiently accurate for varying orientations and sizes of three sorghum varieties. Therefore, future studies are required to test the robustness of our algorithm with other varieties. In addition, compared to previous similar studies, our algorithm integrated image segmentation and CNN deep learning, which not only detect but also localize and delineate individual sorghum panicles. The algorithm is also capable of detecting overlapping sorghum panicles. This offers an opportunity for enhanced sorghum breeding efficiency and accurate yield estimation. To achieve this, however, further research is needed to improve the algorithm to quantify panicle dimension in relation to yield.</p>
<p>The development of large training datasets for plant phenotyping is time consuming and labor intensive. Therefore, this study provides a benchmark for the requirement for the number of training images for such phenotyping tasks. On the other hand, a more effective method, such as automatic annotation, is needed to prepare reliable training images. The performance of this algorithm was evaluated at the small-plot scale. Further studies are required to expand this algorithm to detect and count sorghum panicles at the commercial field scale. In addition, sorghum panicle detection accuracy as influenced by environmental factors, including image resolution, soil background, and illumination levels, requires further evaluation.</p>
</sec>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The datasets generated for this study are available on request to the corresponding author.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>Conceptualization: WG and ZL. Methodology: ZL and WG. Software: ZL. Validation: ZL and WG. Formal analysis: ZL and WG. Investigation: ZL and WG. Resources: WG. Data curation: ZL. Writing&#x2014;original draft preparation: ZL. Writing&#x2014;review and editing: ZL and WG. Visualization: ZL and WG. Supervision: WG. Project administration: WG. Funding acquisition: WG.</p>
</sec>
<sec id="s8">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<ack>
<title>Acknowledgments</title>
<p>We would like to acknowledge the Texas Tech University High Performance Computing Center for support of the study.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Abadi</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Agarwal</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Barham</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Brevdo</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Citro</surname> <given-names>C.</given-names>
</name>
<etal/>
</person-group>. (<year>2016</year>). <source>TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems</source>. Available at: <uri xlink:href="http://www.tensorflow.org">www.tensorflow.org</uri> (Accessed <access-date>November 20, 2019</access-date>).</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aggarwal</surname> <given-names>H. K.</given-names>
</name>
<name>
<surname>Mani</surname> <given-names>M. P.</given-names>
</name>
<name>
<surname>Jacob</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>MoDL: Model-Based Deep Learning Architecture for Inverse Problems</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>38</volume>, <fpage>394</fpage>&#x2013;<lpage>405</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TMI.2018.2865356</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Aich</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Josuttes</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Ovsyannikov</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Strueby</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Ahmed</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Duddu</surname> <given-names>H. S.</given-names>
</name>
<etal/>
</person-group>. (<year>2018</year>). &#x201c;<article-title>DeepWheat: Estimating Phenotypic Traits from Crop Images with Deep Learning</article-title>,&#x201d; <source>2018 IEEE Winter Conference on Applications of Computer Vision, WACV 2018</source>, <fpage>323</fpage>&#x2013;<lpage>332</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/WACV.2018.00042</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Alom</surname> <given-names>M. Z.</given-names>
</name>
<name>
<surname>Hasan</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Yakopcic</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Taha</surname> <given-names>T. M.</given-names>
</name>
<name>
<surname>Asari</surname> <given-names>V. K.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Recurrent Residual Convolutional Neural Network based on U-Net (R2U-Net) for Medical Image Segmentation</source>. Available at: <uri xlink:href="http://arxiv.org/abs/1802.06955">http://arxiv.org/abs/1802.06955</uri> (Accessed <access-date>April 27, 2020</access-date>). arXiv Prepr. arXiv1802.06955.</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ballester</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Zarco-Tejada</surname> <given-names>P. J.</given-names>
</name>
<name>
<surname>Nicol&#xe1;s</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Alarc&#xf3;n</surname> <given-names>J. J.</given-names>
</name>
<name>
<surname>Fereres</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Intrigliolo</surname> <given-names>D. S.</given-names>
</name>
<etal/>
</person-group>. (<year>2018</year>). <article-title>Evaluating the performance of xanthophyll, chlorophyll and structure-sensitive spectral indices to detect water stress in five fruit tree species</article-title>. <source>Precis. Agric.</source> <volume>19</volume>, <fpage>178</fpage>&#x2013;<lpage>193</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-017-9512-y</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Barbedo</surname> <given-names>J. G. A.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Using digital image processing for counting whiteflies on soybean leaves</article-title>. <source>J. Asia Pac. Entomol.</source> <volume>17</volume>, <fpage>685</fpage>&#x2013;<lpage>694</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/J.ASPEN.2014.06.014</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bendig</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Yu</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Aasen</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Bolten</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Bennertz</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Broscheit</surname> <given-names>J.</given-names>
</name>
<etal/>
</person-group>. (<year>2015</year>). <article-title>Combining UAV-based plant height from crop surface models, visible, and near infrared vegetation indices for biomass monitoring in barley</article-title>. <source>Int. J. Appl. Earth Obs. Geoinf.</source> <volume>39</volume>, <fpage>79</fpage>&#x2013;<lpage>87</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/J.JAG.2015.02.012</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bradski</surname> <given-names>G.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>The OpenCV Library</article-title>. <source>Dr Dobbs J. Software Tools</source> <volume>25</volume>, <fpage>120</fpage>&#x2013;<lpage>125</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1111/0023-8333.50.s1.10</pub-id>
</citation>
</ref>
<ref id="B9">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Buja</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Stuetzle</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Shen</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2005</year>).<article-title>Loss functions for binary class probability estimation and classification: structure and application</article-title> (Accessed <access-date>April 27, 2020</access-date>).</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chandra</surname> <given-names>A. L.</given-names>
</name>
<name>
<surname>Desai</surname> <given-names>S. V.</given-names>
</name>
<name>
<surname>Balasubramanian</surname> <given-names>V. N.</given-names>
</name>
<name>
<surname>Ninomiya</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Guo</surname> <given-names>W.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Active learning with point supervision for cost-effective panicle detection in cereal crops</article-title>. <source>Plant Methods</source> <volume>16</volume>, <fpage>1</fpage>&#x2013;<lpage>16</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13007-020-00575-8</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname> <given-names>L. C.</given-names>
</name>
<name>
<surname>Papandreou</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Kokkinos</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Murphy</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Yuille</surname> <given-names>A. L.</given-names>
</name>
</person-group> (<year>2017</year>a). <article-title>DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>40</volume>, <fpage>834</fpage>&#x2013;<lpage>848</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2017.2699184</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname> <given-names>S. W.</given-names>
</name>
<name>
<surname>Shivakumar</surname> <given-names>S. S.</given-names>
</name>
<name>
<surname>Dcunha</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Das</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Okon</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Qu</surname> <given-names>C.</given-names>
</name>
<etal/>
</person-group>. (<year>2017</year>b). <article-title>Counting Apples and Oranges with Deep Learning: A Data-Driven Approach</article-title>. <source>IEEE Robot. Autom. Lett.</source> <volume>2</volume>, <fpage>781</fpage>&#x2013;<lpage>788</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/LRA.2017.2651944</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Chollet</surname> <given-names>F.</given-names>
</name>
</person-group> (<year>2015</year>). <source>Keras: Deep learning library for theano and tensorflow</source>. (GitHub).</citation>
</ref>
<ref id="B14">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Dahlberg</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Hutmacher</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Wright</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2015</year>).<article-title>Sorghum: an alternative feed, hay and forage</article-title>. In: <source>Proceedings, 2015 Weastern Alfalfa &amp; Forage Symposium, Reno, NV</source>. Available at: <uri xlink:href="https://alfalfa.ucdavis.edu/+symposium/2015/PDFfiles/DahlbergJeff.pdf">https://alfalfa.ucdavis.edu/+symposium/2015/PDFfiles/DahlbergJeff.pdf</uri> (Accessed <access-date>October 10, 2019</access-date>).</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Yu</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Deng</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Yu</surname> <given-names>D.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>the essence of knowledge Deep Learning Methods and Applications Foundations and Trends &#xae; in Signal Processing Deep Learning Methods and Applications Deep Learning: Methods and Applications</article-title>. <source>Found. Trends R Signal Process</source> <volume>7</volume>, <fpage>197</fpage>&#x2013;<lpage>387</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1561/2000000039</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>D&#xed;az-Varela</surname> <given-names>R. A.</given-names>
</name>
<name>
<surname>de la Rosa</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Le&#xf3;n</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Zarco-Tejada</surname> <given-names>P. J.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>High-resolution airborne UAV imagery to assess olive tree crown parameters using 3D photo reconstruction: Application in breeding trials</article-title>. <source>Remote Sens.</source> <volume>7</volume>, <fpage>4213</fpage>&#x2013;<lpage>4232</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs70404213</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dyrmann</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Karstoft</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Midtiby</surname> <given-names>H. S.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Plant species classification using deep convolutional neural network</article-title>. <source>Biosyst. Eng.</source> <volume>151</volume>, <fpage>72</fpage>&#x2013;<lpage>80</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.biosystemseng.2016.08.024</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>El-Faki</surname> <given-names>M. S.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Peterson</surname> <given-names>D. E.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>Factors affecting color-based weed detection</article-title>. <source>Trans. Am. Soc Agric. Eng.</source> <volume>43</volume>, <fpage>1001</fpage>&#x2013;<lpage>1009</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.13031/2013.2968</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="web">
<person-group person-group-type="author">
<collab>FAO</collab>
</person-group> (<year>1999</year>). <source>SORGHUM: Post-harvest Operations</source>. Available at: <uri xlink:href="http://www.fao.org/in-action/inpho/crop-compendium/cereals-grains/en/">http://www.fao.org/in-action/inpho/crop-compendium/cereals-grains/en/</uri> (Accessed <access-date>October 10, 2019</access-date>).</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Frid-Adar</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Diamant</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Klang</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Amitai</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Goldberger</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Greenspan</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification</article-title>. <source>Neurocomputing</source> <volume>321</volume>, <fpage>321</fpage>&#x2013;<lpage>331</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.neucom.2018.09.013</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gago</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Douthe</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Coopman</surname> <given-names>R. E.</given-names>
</name>
<name>
<surname>Gallego</surname> <given-names>P. P.</given-names>
</name>
<name>
<surname>Ribas-Carbo</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Flexas</surname> <given-names>J.</given-names>
</name>
<etal/>
</person-group>. (<year>2015</year>). <article-title>UAVs challenge to assess water stress for sustainable agriculture</article-title>. <source>Agric. Water Manage.</source> <volume>153</volume>, <fpage>9</fpage>&#x2013;<lpage>19</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.agwat.2015.01.020</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ghosal</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Zheng</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Chapman</surname> <given-names>S. C.</given-names>
</name>
<name>
<surname>Potgieter</surname> <given-names>A. B.</given-names>
</name>
<name>
<surname>Jordan</surname> <given-names>D. R.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>X.</given-names>
</name>
<etal/>
</person-group>. (<year>2019</year>). <article-title>A Weakly Supervised Deep Learning Framework for Sorghum Head Detection and Counting</article-title>. <source>Plant Phenomics</source> <volume>2019</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/2019/1525874</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gn&#xe4;dinger</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Schmidhalter</surname> <given-names>U.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Digital counts of maize plants by Unmanned Aerial Vehicles (UAVs)</article-title>. <source>Remote Sens.</source> <volume>9</volume>, <elocation-id>544</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs9060544</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>G&#xf3;mez-Cand&#xf3;n</surname> <given-names>D.</given-names>
</name>
<name>
<surname>De Castro</surname> <given-names>A.</given-names>
<suffix>II</suffix>
</name>
<name>
<surname>L&#xf3;pez-Granados</surname> <given-names>F.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Assessing the accuracy of mosaics from unmanned aerial vehicle (UAV) imagery for precision agriculture purposes in wheat</article-title>. <source>Precis. Agric.</source> <volume>15</volume>, <fpage>44</fpage>&#x2013;<lpage>56</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-013-9335-4</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>G&#xf3;mez-Cand&#xf3;n</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Virlet</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Labb&#xe9;</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Jolivot</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Regnard</surname> <given-names>J. L.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Field phenotyping of water stress at tree scale by UAV-sensed imagery: new insights for thermal acquisition and calibration</article-title>. <source>Precis. Agric.</source> <volume>17</volume>, <fpage>786</fpage>&#x2013;<lpage>800</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-016-9449-6</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Guan</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Khan</surname> <given-names>A. A.</given-names>
</name>
<name>
<surname>Sikdar</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Chitnis</surname> <given-names>P. V.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Fully Dense UNet for 2-D Sparse Photoacoustic Tomography Artifact Removal</article-title>. <source>IEEE J. Biomed.Health Informatics</source> <volume>24</volume>, <fpage>568</fpage>&#x2013;<lpage>576</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/JBHI.2019.2912935</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Guo</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Zheng</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Potgieter</surname> <given-names>A. B.</given-names>
</name>
<name>
<surname>Diot</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Watanabe</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Noshita</surname> <given-names>K.</given-names>
</name>
<etal/>
</person-group>. (<year>2018</year>). <article-title>Aerial imagery analysis &#x2013; Quantifying appearance and number of sorghum heads for applications in breeding and agronomy</article-title>. <source>Front. Plant Sci.</source> <volume>9</volume>, <elocation-id>1544</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2018.01544</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hasan</surname> <given-names>M. M.</given-names>
</name>
<name>
<surname>Chopin</surname> <given-names>J. P.</given-names>
</name>
<name>
<surname>Laga</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Miklavcic</surname> <given-names>S. J.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Detection and analysis of wheat spikes using Convolutional Neural Networks</article-title>. <source>Plant Methods</source> <volume>14</volume>, <fpage>100</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13007-018-0366-8</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hunt</surname> <given-names>E. R.</given-names>
</name>
<name>
<surname>Cavigelli</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Daughtry</surname> <given-names>C. S. T.</given-names>
</name>
<name>
<surname>McMurtrey</surname> <given-names>J. E.</given-names>
</name>
<name>
<surname>Walthall</surname> <given-names>C. L.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>Evaluation of digital photography from model aircraft for remote sensing of crop biomass and nitrogen status</article-title>. <source>Precis. Agric.</source> <volume>6</volume>, <fpage>359</fpage>&#x2013;<lpage>378</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-005-2324-5</pub-id>
</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hunt</surname> <given-names>E. R.</given-names>
</name>
<name>
<surname>Hively</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Fujikawa</surname> <given-names>S. J.</given-names>
</name>
<name>
<surname>Linden</surname> <given-names>D. S.</given-names>
</name>
<name>
<surname>Daughtry</surname> <given-names>C. S. T.</given-names>
</name>
<name>
<surname>McCarty</surname> <given-names>G. W.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Acquisition of NIR-green-blue digital photographs from unmanned aircraft for crop monitoring</article-title>. <source>Remote Sens.</source> <volume>2</volume>, <fpage>290</fpage>&#x2013;<lpage>305</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs2010290</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Islam</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Dinh</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Wahid</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Bhowmik</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Detection of potato diseases using image segmentation and multiclass support vector machine</article-title>,&#x201d; <source>2017 IEEE 30th Canadian Conference on Electrical and Computer Engineering (CCECE)</source>. (Windsor, ON), <fpage>1</fpage>&#x2013;<lpage>4</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CCECE.2017.7946594</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jin</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Baret</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Hemerl&#xe9;</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Comar</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Estimates of plant density of wheat crops at emergence from very low altitude UAV imagery</article-title>. <source>Remote Sens. Environ.</source> <volume>198</volume>, <fpage>105</fpage>&#x2013;<lpage>114</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.rse.2017.06.007</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kamilaris</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Prenafeta-Bold&#xfa;</surname> <given-names>F. X.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Deep learning in agriculture: A survey</article-title>. <source>Comput. Electron. Agric.</source> <volume>147</volume>, <fpage>70</fpage>&#x2013;<lpage>90</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2018.02.016</pub-id>
</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kamnitsas</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Ledig</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Newcombe</surname> <given-names>V. F. J.</given-names>
</name>
<name>
<surname>Simpson</surname> <given-names>J. P.</given-names>
</name>
<name>
<surname>Kane</surname> <given-names>A. D.</given-names>
</name>
<name>
<surname>Menon</surname> <given-names>D. K.</given-names>
</name>
<etal/>
</person-group>. (<year>2017</year>). <article-title>Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation</article-title>. <source>Med. Image Anal.</source> <volume>36</volume>, <fpage>61</fpage>&#x2013;<lpage>78</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.media.2016.10.004</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Kingma</surname> <given-names>D. P.</given-names>
</name>
<name>
<surname>Ba</surname> <given-names>J. L.</given-names>
</name>
</person-group> (<year>2015</year>). <source>Adam: A method for stochastic optimization. in 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings</source> (<publisher-name>International Conference on Learning Representations, ICLR</publisher-name>). Available at: <uri xlink:href="https://arxiv.org/abs/1412.6980v9">https://arxiv.org/abs/1412.6980v9</uri> (Accessed <access-date>July 2, 2020</access-date>).</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Komura</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Ishikawa</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Machine Learning Methods for Histopathological Image Analysis</article-title>. <source>Comput. Struct. Biotechnol. J.</source> <volume>16</volume>, <fpage>34</fpage>&#x2013;<lpage>42</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.csbj.2018.01.001</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kussul</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Lavreniuk</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Skakun</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Shelestov</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Deep Learning Classification of Land Cover and Crop Types Using Remote Sensing Data</article-title>. <source>IEEE Geosci. Remote Sens. Lett.</source> <volume>14</volume>, <fpage>778</fpage>&#x2013;<lpage>782</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/LGRS.2017.2681128</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>LeCun</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Bengio</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Hinton</surname> <given-names>G.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Deep learning</article-title>. <source>Nature</source> <volume>521</volume>, <fpage>436</fpage>&#x2013;<lpage>444</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/nature14539</pub-id>
</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lee</surname> <given-names>U.</given-names>
</name>
<name>
<surname>Chang</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Putra</surname> <given-names>G. A.</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>D. H.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>An automated, high-throughput plant phenotyping system using machine learning-based plant segmentation and image analysis</article-title>. <source>PloS One</source> <volume>13</volume>, <fpage>e0196615</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1371/journal.pone.0196615</pub-id>
</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Baret</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Andrieu</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Burger</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Hemmerl&#xe9;</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Estimation of wheat plant density at early stages using high resolution imagery</article-title>. <source>Front. Plant Sci.</source> <volume>8</volume>, <elocation-id>739</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2017.00739</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Madec</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Jin</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Lu</surname> <given-names>H.</given-names>
</name>
<name>
<surname>De Solan</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Duyme</surname> <given-names>F.</given-names>
</name>
<etal/>
</person-group>. (<year>2019</year>). <article-title>Ear density estimation from high resolution RGB imagery using deep learning technique</article-title>. <source>Agric. Forest Meteorol.</source> <volume>264</volume>, <fpage>225</fpage>&#x2013;<lpage>234</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.agrformet.2018.10.013</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Makantasis</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Karantzalos</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Doulamis</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Doulamis</surname> <given-names>N.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Deep supervised learning for hyperspectral data classification through convolutional neural networks</article-title>. <source>2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS)</source>. (<publisher-loc>Milan</publisher-loc>), <fpage>4959</fpage>&#x2013;<lpage>4962</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/IGARSS.2015.7326945</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Malambo</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Popescu</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Ku</surname> <given-names>N. W.</given-names>
</name>
<name>
<surname>Rooney</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Moore</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>A Deep Learning Semantic Segmentation-Based Approach for Field-Level Sorghum Panicle Counting</article-title>. <source>Remote Sens.</source> <volume>11</volume>, <elocation-id>2939</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs11242939</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Martins</surname> <given-names>F. F.</given-names>
</name>
<name>
<surname>Zaglia</surname> <given-names>M. C.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Application of convolutional neural network to pixel-wise classification in deforestation detection using PRODES data. GEOINFO, 20 Years After!, 57</source>. Available at: <uri xlink:href="http://www.obt.inpe.br/OBT/assuntos/programa/amazonia/prodes">http://www.obt.inpe.br/OBT/assuntos/programa/amazonia/prodes</uri> (Accessed <access-date>April 30, 2020</access-date>).</citation>
</ref>
<ref id="B45">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Miko&#x142;ajczyk</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Grochowski</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Data augmentation for improving deep learning in image classification problem</article-title>,&#x201d; <source>2018 International Interdisciplinary PhD Workshop (IIPhDW)</source>. (<publisher-name>Swinouj&#x15b;cie</publisher-name>), <fpage>117</fpage>&#x2013;<lpage>122</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/IIPHDW.2018.8388338</pub-id>
</citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mochida</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Koda</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Inoue</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Hirayama</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Tanaka</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Nishii</surname> <given-names>R.</given-names>
</name>
<etal/>
</person-group>. (<year>2018</year>). <article-title>Computer vision-based phenotyping for improvement of plant productivity: A machine learning perspective</article-title>. <source>Gigascience</source> <volume>8</volume>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/gigascience/giy153</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mohri</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Afshin</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Ameet</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Foundations of Machine Learning</source>. (<publisher-loc>MIT Press</publisher-loc>). doi:&#xa0;<pub-id pub-id-type="doi">10.2139/ssrn.3399990</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Norman</surname> <given-names>D. W.</given-names>
</name>
<name>
<surname>Worman</surname> <given-names>F. D.</given-names>
</name>
<name>
<surname>Siebert</surname> <given-names>J. D.</given-names>
</name>
<name>
<surname>Modiakgotla</surname> <given-names>E.</given-names>
</name>
</person-group> (<year>1995</year>). <source>The farming systems approach to development and appropriate technology generation</source> (<publisher-name>Food and Agriculture Organization of the United Nations</publisher-name>). Available at: <uri xlink:href="http://www.fao.org/3/v5330e/V5330e00.htm#Contents">http://www.fao.org/3/v5330e/V5330e00.htm#Contents</uri> (Accessed <access-date>October 10, 2019</access-date>).</citation>
</ref>
<ref id="B49">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Oh</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Olsen</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Ramamurthy</surname> <given-names>K. N.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Counting and Segmenting Sorghum Heads</source>. Available at: <uri xlink:href="https://arxiv.org/abs/1905.13291">https://arxiv.org/abs/1905.13291</uri> (Accessed <access-date>October 10, 2019</access-date>). arxiv.org.</citation>
</ref>
<ref id="B50">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Olsen</surname> <given-names>P. A.</given-names>
</name>
<name>
<surname>Natesan Ramamurthy</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Ribera</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Thompson</surname> <given-names>A. M.</given-names>
</name>
<name>
<surname>Luss</surname> <given-names>R.</given-names>
</name>
<etal/>
</person-group>. (<year>2018</year>). &#x201c;<article-title>Detecting and counting panicles in sorghum images</article-title>,&#x201d; in <source>Proceedings - 2018 IEEE 5th International Conference on Data Science and Advanced Analytics, DSAA 2018</source> (<publisher-loc>Turin, Italy</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>400</fpage>&#x2013;<lpage>409</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/DSAA.2018.00052</pub-id>
</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pape</surname> <given-names>J.-M.</given-names>
</name>
<name>
<surname>Klukas</surname> <given-names>C.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Utilizing machine learning approaches to improve the prediction of leaf counts and individual leaf segmentation of rosette plant images. in</article-title>. <source>bmva.org</source> <volume>3</volume>, <fpage>1</fpage>&#x2013;<lpage>3.12</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.5244/c.29.cvppp.3</pub-id>
</citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>P&#xe9;rez-Ortiz</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Pe&#xf1;a</surname> <given-names>J. M.</given-names>
</name>
<name>
<surname>Guti&#xe9;rrez</surname> <given-names>P. A.</given-names>
</name>
<name>
<surname>Torres-S&#xe1;nchez</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Herv&#xe1;s-Mart&#xed;nez</surname> <given-names>C.</given-names>
</name>
<name>
<surname>L&#xf3;pez-Granados</surname> <given-names>F.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>A semi-supervised system for weed mapping in sunflower crops using unmanned aerial vehicles and a crop row detection method</article-title>. <source>Appl. Soft Comput. J.</source> <volume>37</volume>, <fpage>533</fpage>&#x2013;<lpage>544</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.asoc.2015.08.027</pub-id>
</citation>
</ref>
<ref id="B53">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Pradeep</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>S. G.</given-names>
</name>
<name>
<surname>Wei</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Trajectory optimization of multirotor agricultural UAVs</article-title>,&#x201d; in <source>IEEE Aerospace Conference Proceedings</source> (<publisher-loc>Big Sky, MT</publisher-loc>: <publisher-name>IEEE Computer Society</publisher-name>), <fpage>1</fpage>&#x2013;<lpage>7</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/AERO.2018.8396617</pub-id>
</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<collab>Python Software Foundation</collab>
</person-group> (<year>1995</year>). <article-title>Python tutorial</article-title>. <source>CWI Rep.</source>, <fpage>1</fpage>&#x2013;<lpage>65</lpage>. <uri xlink:href="http://oai.cwi.nl/oai/asset/5007/05007D.pdf">http://oai.cwi.nl/oai/asset/5007/05007D.pdf</uri>. CS-R9526.</citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qureshi</surname> <given-names>W. S.</given-names>
</name>
<name>
<surname>Payne</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Walsh</surname> <given-names>K. B.</given-names>
</name>
<name>
<surname>Linker</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Cohen</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Dailey</surname> <given-names>M. N.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Machine vision for counting fruit on mango tree canopies</article-title>. <source>Precis. Agric.</source> <volume>18</volume>, <fpage>224</fpage>&#x2013;<lpage>244</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-016-9458-5</pub-id>
</citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rasmussen</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Nielsen</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Streibig</surname> <given-names>J. C.</given-names>
</name>
<name>
<surname>Jensen</surname> <given-names>J. E.</given-names>
</name>
<name>
<surname>Pedersen</surname> <given-names>K. S.</given-names>
</name>
<name>
<surname>Olsen</surname> <given-names>S.</given-names>
<suffix>II</suffix>
</name>
</person-group> (<year>2019</year>). <article-title>Pre-harvest weed mapping of Cirsium arvense in wheat and barley with off-the-shelf UAVs</article-title>. <source>Precis. Agric.</source> <volume>20</volume>, <fpage>983</fpage>&#x2013;<lpage>999</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-018-09625-7</pub-id>
</citation>
</ref>
<ref id="B57">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ronneberger</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Fischer</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Brox</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>U-net: Convolutional networks for biomedical image segmentation</article-title>,&#x201d; in <source>International Conference on Medical image computing and computer-assisted intervention</source>. (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>234</fpage>&#x2013;<lpage>241</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id>
</citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sankaran</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Khot</surname> <given-names>L. R.</given-names>
</name>
<name>
<surname>Carter</surname> <given-names>A. H.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Field-based crop phenotyping: Multispectral aerial imaging for evaluation of winter wheat emergence and spring stand</article-title>. <source>Comput. Electron. Agric.</source> <volume>118</volume>, <fpage>372</fpage>&#x2013;<lpage>379</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2015.09.001</pub-id>
</citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Singh</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Ganapathysubramanian</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Singh</surname> <given-names>A. K.</given-names>
</name>
<name>
<surname>Sarkar</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Machine Learning for High-Throughput Stress Phenotyping in Plants</article-title>. <source>Trends Plant Sci.</source> <volume>21</volume>, <fpage>110</fpage>&#x2013;<lpage>124</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.tplants.2015.10.015</pub-id>
</citation>
</ref>
<ref id="B60">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sladojevic</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Arsenovic</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Anderla</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Culibrk</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Stefanovic</surname> <given-names>D.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification</article-title>. <source>Comput. Intell. Neurosci.</source> <volume>2016</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.1155/2016/3289801</pub-id>
</citation>
</ref>
<ref id="B61">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sudre</surname> <given-names>C. H.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Vercauteren</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Ourselin</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Jorge Cardoso</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations</article-title>,&#x201d; in <source>Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)</source>. (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>240</fpage>&#x2013;<lpage>248</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-319-67558-9_28</pub-id>
</citation>
</ref>
<ref id="B62">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tokekar</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Hook</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Mulla</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Isler</surname> <given-names>V.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Sensor Planning for a Symbiotic UAV and UGV System for Precision Agriculture</article-title>. <source>IEEE Trans. Robot.</source> <volume>32</volume>, <fpage>1498</fpage>&#x2013;<lpage>1511</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TRO.2016.2603528</pub-id>
</citation>
</ref>
<ref id="B63">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Too</surname> <given-names>E. C.</given-names>
</name>
<name>
<surname>Yujian</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Njuki</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Yingchun</surname> <given-names>L.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>A comparative study of fine-tuning deep learning models for plant disease identification</article-title>. <source>Comput. Electron. Agric.</source> <volume>161</volume>, <fpage>272</fpage>&#x2013;<lpage>279</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2018.03.032</pub-id>
</citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Torres-S&#xe1;nchez</surname> <given-names>J.</given-names>
</name>
<name>
<surname>L&#xf3;pez-Granados</surname> <given-names>F.</given-names>
</name>
<name>
<surname>De Castro</surname> <given-names>A.</given-names>
<suffix>II</suffix>
</name>
<name>
<surname>Pe&#xf1;a-Barrag&#xe1;n</surname> <given-names>J. M.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Configuration and Specifications of an Unmanned Aerial Vehicle (UAV) for Early Site Specific Weed Management</article-title>. <source>PloS One</source> <volume>8</volume>, <elocation-id>e58210</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1371/journal.pone.0058210</pub-id>
</citation>
</ref>
<ref id="B65">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ubbens</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Cieslak</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Prusinkiewicz</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Stavness</surname> <given-names>I.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>The use of plant models in deep learning: An application to leaf counting in rosette plants</article-title>. <source>Plant Methods</source> <volume>14</volume>, <elocation-id>6</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13007-018-0273-z</pub-id>
</citation>
</ref>
<ref id="B66">
<citation citation-type="web">
<person-group person-group-type="author">
<collab>US Climate Data</collab>
</person-group> (<year>2019</year>). <source>Weather history Lubbock</source> (<publisher-loc>Texas</publisher-loc>: <publisher-name>U.S. Clim. Data</publisher-name>). Available at: <uri xlink:href="https://www.usclimatedata.com/climate/lubbock/texas/united-states/ustx2745/2014/1">https://www.usclimatedata.com/climate/lubbock/texas/united-states/ustx2745/2014/1</uri> (Accessed <access-date>February 25, 2019</access-date>).</citation>
</ref>
<ref id="B67">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Verger</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Vigneau</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Ch&#xe9;ron</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Gilliot</surname> <given-names>J. M.</given-names>
</name>
<name>
<surname>Comar</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Baret</surname> <given-names>F.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Green area index from an unmanned aerial system over wheat and rapeseed crops</article-title>. <source>Remote Sens. Environ.</source> <volume>152</volume>, <fpage>654</fpage>&#x2013;<lpage>664</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.rse.2014.06.006</pub-id>
</citation>
</ref>
<ref id="B68">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Perez</surname> <given-names>L.</given-names>
</name>
</person-group> (<year>2017</year>).<article-title>The Effectiveness of Data Augmentation in Image Classification using Deep Learning</article-title>. Available at: <uri xlink:href="http://arxiv.org/abs/1712.04621">http://arxiv.org/abs/1712.04621</uri> (Accessed <access-date>April 29, 2020</access-date>). arXiv Prepr. arXiv1712.04621</citation>
</ref>
<ref id="B69">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weng</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Qiu</surname> <given-names>X.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>NAS-Unet: Neural architecture search for medical image segmentation</article-title>. <source>IEEE Access</source>. <volume>7</volume>, <fpage>44247</fpage>&#x2013;<lpage>44257</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ACCESS.2019.2908991</pub-id>
</citation>
</ref>
<ref id="B70">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Xu</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Han</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Zhu</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Automatic Counting of in situ Rice Seedlings from UAV Images Based on a Deep Fully Convolutional Neural Network</article-title>. <source>Remote Sens.</source> <volume>11</volume>, <elocation-id>691</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs11060691</pub-id>
</citation>
</ref>
<ref id="B71">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiong</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Cao</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Lu</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Madec</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Shen</surname> <given-names>C.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>TasselNetv2: In-field counting of wheat spikes with context-augmented local regression networks</article-title>. <source>Plant Methods</source> <volume>15</volume>, <fpage>150</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13007-019-0537-2</pub-id>
</citation>
</ref>
<ref id="B72">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Huang</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Yu</surname> <given-names>H.</given-names>
</name>
<etal/>
</person-group>. (<year>2017</year>). <article-title>Unmanned aerial vehicle remote sensing for field-based crop phenotyping: Current status and perspectives</article-title>. <source>Front. Plant Sci.</source> <volume>8</volume>, <elocation-id>1111</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2017.01111</pub-id>
</citation>
</ref>
<ref id="B73">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zahavy</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Sivak</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Kang</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Feng</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Xu</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Mannor</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Ensemble robustness and generalization of stochastic deep learning algorithms</article-title>,&#x201d; <source>arXiv preprint arXiv:1602.02389</source>.</citation>
</ref>
<ref id="B74">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Sabuncu</surname> <given-names>M. R.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Generalized cross entropy loss for training deep neural networks with noisy labels</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>, <fpage>8778</fpage>&#x2013;<lpage>8788</lpage>.</citation>
</ref>
<ref id="B75">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Meng</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Ding</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Ma</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>A method for deriving plant temperature from UAV TIR Image</article-title>,&#x201d; <source>2018 7th International Conference on Agro-geoinformatics (Agro-geoinformatics)</source>. (<publisher-loc>Hangzhou</publisher-loc>), <fpage>1</fpage>&#x2013;<lpage>5</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/Agro-Geoinformatics.2018.8475995</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>