首页 | 本学科首页   官方微博 | 高级检索  
     检索      

大视场下荔枝采摘机器人的视觉预定位方法
引用本文:陈燕,王佳盛,曾泽钦,邹湘军,陈明猷.大视场下荔枝采摘机器人的视觉预定位方法[J].农业工程学报,2019,35(23):48-54.
作者姓名:陈燕  王佳盛  曾泽钦  邹湘军  陈明猷
作者单位:1. 华南农业大学工程学院,广州 510642; 2. 华南农业大学南方农业机械与装备关键技术教育部重点实验室,广州 510642,1. 华南农业大学工程学院,广州 510642; 2. 华南农业大学南方农业机械与装备关键技术教育部重点实验室,广州 510642,1. 华南农业大学工程学院,广州 510642; 2. 华南农业大学南方农业机械与装备关键技术教育部重点实验室,广州 510642,1. 华南农业大学工程学院,广州 510642; 2. 华南农业大学南方农业机械与装备关键技术教育部重点实验室,广州 510642,1. 华南农业大学工程学院,广州 510642; 2. 华南农业大学南方农业机械与装备关键技术教育部重点实验室,广州 510642
基金项目:国家自然科学基金资助项目(31571568);广东省自然科学基金项目(2018A030307067)
摘    要:机器人采摘荔枝时需要获取多个目标荔枝串的空间位置信息,以指导机器人获得最佳运动轨迹,提高效率。该文研究了大视场下荔枝采摘机器人的视觉预定位方法。首先使用双目相机采集荔枝图像;然后改进原始的YOLOv3网络,设计YOLOv3-DenseNet34荔枝串检测网络;提出同行顺序一致性约束的荔枝串配对方法;最后基于双目立体视觉的三角测量原理计算荔枝串空间坐标。试验结果表明,YOLOv3-DenseNet34网络提高了荔枝串的检测精度与检测速度;平均精度均值(mean average precision,m AP)达到0.943,平均检测速度达到22.11帧/s。基于双目立体视觉的荔枝串预定位方法在3 m的检测距离下预定位的最大绝对误差为36.602 mm,平均绝对误差为23.007 mm,平均相对误差为0.836%,满足大视场下采摘机器人的视觉预定位要求,可为其他果蔬在大视场下采摘的视觉预定位提供参考。

关 键 词:机器人  图像处理  目标检测  荔枝采摘  大视场  卷积神经网络  立体视觉
收稿时间:2019/6/30 0:00:00
修稿时间:2019/11/11 0:00:00

Research on vision pre-positioning for litchi picking robot under large field of view
Chen Yan,Wang Jiasheng,Zeng Zeqin,Zou Xiangjun and Chen Mingyou.Research on vision pre-positioning for litchi picking robot under large field of view[J].Transactions of the Chinese Society of Agricultural Engineering,2019,35(23):48-54.
Authors:Chen Yan  Wang Jiasheng  Zeng Zeqin  Zou Xiangjun and Chen Mingyou
Institution:1. College of Engineering, South China Agricultural University, Guangzhou 510642, China; 2. Key Lab of Key Technology on Agriculture Machine and Equipment, Ministry of Education, South China Agricultural University, Guangzhou 510642, China,1. College of Engineering, South China Agricultural University, Guangzhou 510642, China; 2. Key Lab of Key Technology on Agriculture Machine and Equipment, Ministry of Education, South China Agricultural University, Guangzhou 510642, China,1. College of Engineering, South China Agricultural University, Guangzhou 510642, China; 2. Key Lab of Key Technology on Agriculture Machine and Equipment, Ministry of Education, South China Agricultural University, Guangzhou 510642, China,1. College of Engineering, South China Agricultural University, Guangzhou 510642, China; 2. Key Lab of Key Technology on Agriculture Machine and Equipment, Ministry of Education, South China Agricultural University, Guangzhou 510642, China and 1. College of Engineering, South China Agricultural University, Guangzhou 510642, China; 2. Key Lab of Key Technology on Agriculture Machine and Equipment, Ministry of Education, South China Agricultural University, Guangzhou 510642, China
Abstract:Litchi picking robot is an important tool for improving the automation of litchi picking operation. The spatial position information of litchi cluster needs to be acquired when the robot picks litchi normally. In order to guide the robot moving to the picking position and improve the picking efficiency, the vision pre-positioning method of litchi picking robot under large field of view is proposed in this paper studied. Firstly, using the binocular stereo vision system composed of two industrial cameras that have been calibrated, 250 pairs of litchi cluster images under large field of view was taken in the litchi orchard in Guangzhou, the spatial positions of key litchi clusters were recorded by using a laser range finder, and the results were compared with those tested in the paper. In order to expand the sample size, the original image and the polar line correction image were randomly cropped and scaled in a small range, and the final image data set was 4 000 sheets. After that, by using labeling, the data set of the target detection network was created. Secondly, by using the YOLOv3 network and the DenseNet classification network, combined with the characteristics of single target and single scene of litchi cluster detection task (only for orchard environment), the network structure was optimized, a Dense Module with a depth of 34 layers and a litchi cluster detection network YOLOv3-DenseNet34 based on the Dense Module was designed. Thirdly, Because of the the complexity of the background image under large field of view, the dense stereo matching degree of the whole image is low and the effect is poor, at the same time, some litchi clusters can not appear in the public view of the image at the same time, therefore, a method for calculating sub-pixel parallax was designed. By solving the quadratic curve composed of parallax and similarity, the parallax under sub-pixel was used to calculate the spatial positions of the litchi cluster. Through the comparison with the original network of YOLOv3, the test network performance of the paper was tested, and found that the YOLOv3-DenseNet34 network improved the detection accuracy and detection speed of the litchi cluster, the mAP (mean average precision) value was 0.943, the average detection speed was 22.11 frame/s and the model size was 9.3 MB, which was 1/26 of the original network of YOLOv3. Then, the detection results of the method were compared with the results of the laser range finder. The max absolute error of the pre-positioning at the detection distance of 3 m was 36.602 mm, the mean absolute error was 23.007 mm, and the average relative error was 0.836%. Test results showed that the vision pre-positioning method studied in this paper can basically meet the requirements of vision pre-positioning under large field of view in precision and speed. And this method can provide reference for other vision pre-positioning methods under large field of view of fruits and vegetables picking.
Keywords:robs  image processing  object detection  litchi picking  large field of view  convolutional neural network  stereo vision
本文献已被 CNKI 等数据库收录!
点击此处可从《农业工程学报》浏览原始摘要信息
点击此处可从《农业工程学报》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号