首页 | 本学科首页   官方微博 | 高级检索  
     检索      

基于YOLOv5m和CBAM-CPN的单分蘖水稻植株表型参数提取
引用本文:陈慧颖,宋青峰,常天根,郑立华,朱新广,张漫,王敏娟.基于YOLOv5m和CBAM-CPN的单分蘖水稻植株表型参数提取[J].农业工程学报,2024,40(2):307-314.
作者姓名:陈慧颖  宋青峰  常天根  郑立华  朱新广  张漫  王敏娟
作者单位:中国农业大学智慧农业系统集成研究教育部重点实验室,北京 100083;中国科学院分子植物科学卓越创新中心,植物分子遗传国家重点实验室,上海 200032;中国农业大学智慧农业系统集成研究教育部重点实验室,北京 100083;中国农业大学烟台研究院,烟台 264670
基金项目:国家自然科学基金项目(32201654);国家重点研发计划(2022YFD1900701)
摘    要:为快速获取单分蘖水稻植株的形态结构和表型参数,该研究提出了一种基于目标检测和关键点检测模型相结合的骨架提取和表型参数获取方法。该方法基于目标检测模型生成穗、茎秆、叶片的边界框和类别,将所得数据分别输入到关键点检测模型检测各部位关键点,按照语义信息依次连接关键点形成植株骨架,依据关键点坐标计算穗长度、茎秆长度、叶片长度、叶片-茎秆夹角4种表型参数。首先,构建单分蘖水稻的关键点检测和目标检测数据集;其次,训练Faster R-CNN、YOLOv3、YOLOv5s、YOLOv5m目标检测模型,经过对比,YOLOv5m的检测效果最好,平均精度均值(mean average precision,mAP)达到91.17%;然后,应用人体姿态估计的级联金字塔网络(cascaded pyramid network,CPN)提取植株骨架,并引入注意力机制CBAM(convolutional block attention module)进行改进,与沙漏网络(hourglass networks,HN)、堆叠沙漏网络模型(stacked hourglass networks,SHN)和CPN模型相比,CBAM-CPN模型的预测准确率分别提高了9.68、8.83和0.5个百分点,达到94.75%,4种表型参数的均方根误差分别为1.06、0.81、1.25 cm和2.94°。最后,结合YOLOv5m和CBAM-CPN进行预测,4种表型参数的均方根误差分别为1.48 、1.05 、1.74 cm和2.39°,与SHN模型相比,误差分别减小1.65、3.43、2.65 cm和4.75°,生成的骨架基本能够拟合单分蘖水稻植株的形态结构。所提方法可以提高单分蘖水稻植株的关键点检测准确率,更准确地获取植株骨架和表型参数,有助于加快水稻的育种和改良。

关 键 词:目标检测|注意力机制|水稻|关键点检测|骨架提取|表型参数|单分蘖植株
收稿时间:2023/4/17 0:00:00
修稿时间:2023/10/26 0:00:00

Extraction of the single-tiller rice phenotypic parameters based onYOLOv5m and CBAM-CPN
CHEN Huiying,SONG Qingfeng,CHANG Tiangen,ZHENG Lihu,ZHU Xinguang,ZHANG Man,WANG Minjuan.Extraction of the single-tiller rice phenotypic parameters based onYOLOv5m and CBAM-CPN[J].Transactions of the Chinese Society of Agricultural Engineering,2024,40(2):307-314.
Authors:CHEN Huiying  SONG Qingfeng  CHANG Tiangen  ZHENG Lihu  ZHU Xinguang  ZHANG Man  WANG Minjuan
Institution:Key Laboratory of Smart Agriculture Systems, Ministry of Education, China Agricultural University, Beijing 100083, China;National Key Laboratory of Plant Molecular Genetics, CAS Center for Excellence in Molecular Plant Sciences, Chinese Academy of Sciences, Shanghai 200032, China;Key Laboratory of Smart Agriculture Systems, Ministry of Education, China Agricultural University, Beijing 100083, China;Yantai Institute of China Agricultural University, Yantai 264670, China
Abstract:Rice is one of the most essential grain crops in China, providing an important guarantee for the food supply. The taste and nutrition of rice are ever-increasing with the development of society and the improvement of living standards. Therefore, it is necessary to accelerate the breeding and improvement to ensure the quantity and quality of rice. Among them, skeleton and phenotypic parameters can be used to represent the growth and health status of rice for better breeding and improvement. In this study, the object and key points detection models were used to extract the skeleton for the phenotypic parameters. The images of single-tiller rice were taken as the research object. The bounding box of spikes, stems and leaves was also detected by the object detection model. The predicted key points were connected to form the rice skeleton, according to the semantic information. Four phenotypic parameters were calculated, including spike length, stem length, leaf length, and leaf-stem angle, according to the key point coordinates. Firstly, 1 081 RGB images of single-tiller rice were collected in total. The datasets of single-tiller rice were created for object detection and key points detection. Secondly, four current mainstream object detection models were trained, namely Faster R-CNN, YOLOv3, YOLOv5s, and YOLOv5m. The best detection was achieved in YOLOv5m, with mean average precision (mAP) reaching 91.17%, compared with the rest, the mAP of YOLOv5m was improved by 49.55, 36.38, and 2.69 percentage points, respectively, compared with Faster R-CNN, YOLOv3 and YOLOv5s. The predicted bounding box and category were drawn on the original picture to observe the prediction of the model. The visualization results showed that YOLOv5m was basically detected in the bounding box and category of spikes, stems and leaves. Then, the cascaded pyramid network (CPN) model was used for human pose estimation and then applied to plant skeleton extraction. The attention mechanism squeeze and excitation networks (SENet) and convolutional block attention module (CBAM) were integrated into the backbone to improve the feature extraction ability of the model. By contrast, the key points prediction accuracies of SE-CPN and CBAM-CPN were higher than that of CPN. Furthermore, CBAM-CPN shared the highest prediction accuracy of key points, with accuracy of 95.24%, 95.74%, and 93.27% for spike, stem and leaf, respectively. The average accuracy reached 94.75%. The prediction accuracy of the CBAM-CPN model was improved by 9.68, 8.83, and 0.5 percentage points, respectively, compared with hourglass networks (HN), stacked hourglass networks (SHN) and CPN models. The root mean square errors (RMSE) of the phenotypic parameters were 1.06 cm, 0.81 cm, 1.25 cm, and 2.94° respectively. Lastly, the RMSE of four phenotypic parameters were 1.48 cm, 1.05 cm, 1.74 cm and 2.39°, compared with YOLOv5m and CBAM-CPN. The errors were reduced by 1.65 cm, 3.43 cm, 2.65 cm and 4.75°, respectively, compared with SHN. The better prediction was achieved in the improved model. Moreover, the formed skeleton can be expected to better fit the morphological structure of single-tiller rice. The feasibility of the improved model was further verified to combine the object and key points detection model, in order to extract the skeleton and phenotypic parameters of the single-tiller rice. In conclusion, higher detection accuracy was achieved in the key points of single-tiller rice plants. The skeleton and phenotypic parameters were extracted more efficiently and accurately. The findings can provide a strong reference to accelerate the breeding and improvement of rice.
Keywords:object detection|attention mechanism|rice|key points detection|skeleton extraction|phenotypic parameters|single-tiller plant
点击此处可从《农业工程学报》浏览原始摘要信息
点击此处可从《农业工程学报》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号