首页 | 本学科首页   官方微博 | 高级检索  
     检索      

多通道深度可分离卷积模型实时识别复杂背景下甜菜与杂草
引用本文:孙俊,谭文军,武小红,沈继锋,芦兵,戴春霞.多通道深度可分离卷积模型实时识别复杂背景下甜菜与杂草[J].农业工程学报,2019,35(12):184-190.
作者姓名:孙俊  谭文军  武小红  沈继锋  芦兵  戴春霞
作者单位:江苏大学电气信息工程学院,镇江 212013,江苏大学电气信息工程学院,镇江 212013,江苏大学电气信息工程学院,镇江 212013,江苏大学电气信息工程学院,镇江 212013,江苏大学电气信息工程学院,镇江 212013,江苏大学电气信息工程学院,镇江 212013
基金项目:国家自然科学基金资助项目(No.31471413);江苏高校优势学科建设工程资助项目PAPD(苏政办发2011 6号);江苏省六大人才高峰资助项目(ZBZZ-019)。
摘    要:针对实际复杂田间环境下杂草与作物识别精度低和实时性差的问题,为减少弱光环境对分割识别效果的影响,实现甜菜与杂草的实时精确分割识别,该文首先将可见光图像进行对比度增强,再将近红外与可见光图像融合为4通道图像;将深度可分离卷积以及残差块构成分割识别模型的卷积层,减少模型参数量及计算量,构建编码与解码结构并融合底层特征,细化分割边界。以分割识别精度、参数量以及运行效率为评价指标,通过设置不同宽度系数以及输入图像分辨率选出最优模型。试验结果表明:本文模型的平均交并比达到87.58%,平均像素准确率为99.19%,帧频可达42.064帧/s,参数量仅为525 763,具有较高分割识别精度和较好实时性。该方法有效实现了甜菜与杂草的精确实时识别,可为后续机器人精确除草提供理论参考。

关 键 词:作物  图像分割  卷积神经网络  深度学习  甜菜  杂草  实时
收稿时间:2018/10/21 0:00:00
修稿时间:2019/3/6 0:00:00

Real-time recognition of sugar beet and weeds in complex backgrounds using multi-channel depth-wise separable convolution model
Sun Jun,Tan Wenjun,Wu Xiaohong,Shen Jifeng,Lu Bing and Dai Chunxia.Real-time recognition of sugar beet and weeds in complex backgrounds using multi-channel depth-wise separable convolution model[J].Transactions of the Chinese Society of Agricultural Engineering,2019,35(12):184-190.
Authors:Sun Jun  Tan Wenjun  Wu Xiaohong  Shen Jifeng  Lu Bing and Dai Chunxia
Institution:School of Electrical and Information Engineering, Jiangsu University, Zhenjiang 212013, China,School of Electrical and Information Engineering, Jiangsu University, Zhenjiang 212013, China,School of Electrical and Information Engineering, Jiangsu University, Zhenjiang 212013, China,School of Electrical and Information Engineering, Jiangsu University, Zhenjiang 212013, China,School of Electrical and Information Engineering, Jiangsu University, Zhenjiang 212013, China and School of Electrical and Information Engineering, Jiangsu University, Zhenjiang 212013, China
Abstract:Abstract: Mechanical weeding can reduce the use of pesticides and is of great significance to ensure high yield of crops. Real-time and accurate identification of crops is a key technical problem needs to be solved in mechanical weeding equipment. Because of the subjectivity of feature extraction process in weed recognition, the accuracy of traditional methods in actual field environment is low. In recent years, the method of weed identification based on convolution neural network has been widely studied. Although the accuracy is obviously improved, there are still problems such as large parameters and poor real-time performance. In order to solve the above problems, a four-channel input image is constructed by collecting near infrared and visible images of sugar beet in the field, and a lightweight convolution neural network based on codec structure is proposed. In this paper, Sugarbeet and weed images collected from a farm in Bonn, Germany, in 2016 were used as data sets, which covered images of different growth stages of sugar beet, and 226 pictures of which were randomly selected as training sets, and the remaining 57 pictures were used as test sets. The experimental data set was composed of three channels of visible light image and one channel of near infrared image, which are merged into a four-channel image by pixel level superposition, and the depth-wise separable convolution was used in the deep model. Firstly, the input feature image was convoluted in 2 dimensions convolution kernel and the number of channels was expanded. Then, the 1×1 convolution kernel was used to make the 3 dimensions convolution which combined channel feature and compressed the channels to enhance the nonlinear mapping ability of the model. In order to avoid the problem of the gradient disappearing, the residual block was used to connect the input and output of the depth-wise separate convolution. Finally, the coding and decoder structure was designed and the shallow features were combined with deep features to refine the segmentation effect. Due to the imbalance of pixel proportions of soil, crops and weeds, the weighted loss function was used to optimize the model. The segmentation accuracy, parameters and operating efficiency of models at different input resolutions and different width factor were introduced to evalute the model. When the width factor was 1, the segmentation accuracy of the model increased with the increase of the input image resolution, the model accuracy of four channel input was higher than that of the model based on original visible image input, which showed that the near-infrared image features can compensate the defects of ordinary RGB images to some extent, and make the model more suitable for the dark environment. Under the same input image resolution, the model with a width factor of 2 or 4 performs better than the model with a width factor of 1. With the increases of width factor, the parameters of the model increase greatly. The amount of calculation is related to the size of the input image, so the frame rate gradually decreases with the increase of the size of input image. The experimental results show that the optimal model in this paper is a four channel input model with a width coefficient of 2, and the average intersection union ration is 87.58%, the average pixel accuracy is 99.19%, the parameters are 525 763 and the frame rate is 42.064 frames/s. The model has high segmentation and recognition accuracy and good real-time performance, and can provide theoretical basis for the development of intelligent mechanization weeding equipment.
Keywords:image segmentation  crops  convolutional neural network  deep learning  sugar beet  weed  real-time
本文献已被 CNKI 等数据库收录!
点击此处可从《农业工程学报》浏览原始摘要信息
点击此处可从《农业工程学报》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号