4.6 Article

Beet seedling and weed recognition based on convolutional neural network and multi-modality images

Journal

MULTIMEDIA TOOLS AND APPLICATIONS
Volume 81, Issue 4, Pages 5239-5258

Publisher

SPRINGER
DOI: 10.1007/s11042-021-11764-5

Keywords

Object detection; Beets and weeds; Multi-modality images; Deformable convolution; Deep learning

Funding

  1. Priority Academic Program Development of Jiangsu Higher Education Institutions [PAPD-2018-87]
  2. Synergistic Innovation Center of Jiangsu Modern Agricultural Equipment and Technology [4091600002]
  3. Project of Faculty of Agricultural Equipment of Jiangsu University [4121680001]

Ask authors/readers for more resources

This study proposed a novel depth fusion algorithm based on visible and near-infrared imagery to improve the recognition of beet seedlings and weeds. By utilizing an improved region-based fully convolutional network (R-FCN) model, deformable convolution, and online hard example mining, the average precision of the optimal model was significantly enhanced. The study can serve as a theoretical basis for the development of intelligent weed control robots under weak light conditions.
Difficulties in the recognition of beet seedlings and weeds can arise from a complex background in the natural environment and a lack of light at night. In the current study, a novel depth fusion algorithm was proposed based on visible and near-infrared imagery. In particular, visible (RGB) and near-infrared images were superimposed at the pixel-level via a depth fusion algorithm and were subsequently fused into three-channel multi-modality images in order to characterize the edge details of beets and weeds. Moreover, an improved region-based fully convolutional network (R-FCN) model was applied in order to overcome the geometric modeling restriction of traditional convolutional kernels. More specifically, for the convolutional feature extraction layers, deformable convolution was adopted to replace the traditional convolutional kernel, allowing for the entire network to extract more precise features. In addition, online hard example mining was introduced to excavate the hard negative samples in the detection process for the retraining of misidentified samples. A total of four models were established via the aforementioned improved methods. Results demonstrate that the average precision of the improved optimal model for beets and weeds were 84.8% and 93.2%, respectively, while the mean average precision was improved to 89.0%. Compared with the classical R-FCN model, the performance of the optimal model was not only greatly improved, but the parameters were also not significantly expanded. Our study can provide a theoretical basis for the subsequent development of intelligent weed control robots under weak light conditions.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available