Efficient identification, localization and quantification of grapevine inflorescences and flowers in unprepared field images using Fully Convolutional Networks
Yield and its prediction is one of the most important tasks in grapevine breeding purposes and vineyard management. Commonly, this trait is estimated manually right before harvest by extrapolation, which mostly is labor-intensive, destructive and inaccurate. In the present study an automated image-based workflow was developed for quantifying inflorescences and single flowers in unprepared field images of grapevines, i.e. no artificial background or light was applied. It is a novel approach for non-invasive, inexpensive and objective phenotyping with high-throughput.
First, image regions depicting inflorescences were identified and localized. This was done by segmenting the images into the classes "inflorescence" and "non-inflorescence" using a Fully Convolutional Network (FCN). Efficient image segmentation hereby is the most challenging step regarding the small geometry and dense distribution of single flowers (several hundred single flowers per inflorescence), similar color of all plant organs in the fore- and background as well as the circumstance that only approximately 5 % of an image show inflorescences. The trained FCN achieved a mean Intersection Over Union (IOU) of 87.6 % on the test data set. Finally, single flowers were extracted from the "inflorescence"-areas using Circular Hough Transform. The flower extraction achieved a recall of 80.3 % and a precision of 70.7 % using the segmentation derived by the trained FCN model.
Summarized, the presented approach is a promising strategy in order to predict yield potential automatically in the earliest stage of grapevine development which is applicable for objective monitoring and evaluations of breeding material, genetic repositories or commercial vineyards.
Copyright (c) 2019 The Author(s)
This work is licensed under a Creative Commons Attribution 4.0 International License.
The content of VITIS is published under a Creative Commons Attribution 4.0 license. Any user is free to share and adapt (remix, transform, build upon) the content as long as the original publication is attributed (authors, title, year, journal, issue, pages) and any changes to the original are clearly labeled. We do not prohibit or charge a fee for reuse of published content. The use of general descriptive names, trade names, trademarks, and so forth in any publication herein, even if not specifically indicated, does not imply that these names are not protected by the relevant laws and regulations. The submitting author agrees to these terms on behalf of all co-authors when submitting a manuscript. Please be aware that this license cannot be revoked. All authors retain the copyright on their work and are able to enter into separate, additional contractual arrangements.