High-efficiency tea shoot detection method via a compressed deep learning model
Keywords:
deep learning, tea shoot detection, model compression, high-efficiencyAbstract
Achieving high-efficiency and accurate detection of tea shoots in fields is essential for tea robotic plucking. A real-time tea shoot detection method using the channel and layer pruned YOLOv3-SPP deep learning algorithm was proposed in this study. First, tea shoot images were collected and data augmentation was performed to increase sample diversity, and then a spatial pyramid pooling module was added to the YOLOv3 model to detect tea shoots. To simplify the tea shoot detection model and improve the detection speed, the channel pruning algorithm and the layer pruning algorithm were used to compress the model. Finally, the model was fine-tuned to restore its accuracy, and achieve the fast and accurate detection of tea shoots. The test results demonstrated that the number of parameters, model size, and inference time of the tea shoot detection model after compression reduced by 96.82%, 96.81%, and 59.62%, respectively, whereas the mean average precision of the model was only 0.40% lower than that of the original model. In the field test, the compressed model was deployed on a Jetson Xavier NX to conduct the detection of tea shoots. The experimental results demonstrated that the detection speed of the compressed model was 15.9 fps, which was 3.18 times that of the original model. All the results indicate that the proposed method could be deployed on tea harvesting robots with low computing power to achieve high efficiency and accurate detection. Keywords: deep learning, tea shoot detection, model compression, high-efficiency DOI: 10.25165/j.ijabe.20221503.6896 Citation: Li Y T, He L Y, Jia J M, Chen J N, Lyu J, Wu C Y. High-efficiency tea shoot detection method via a compressed deep learning model. Int J Agric & Biol Eng, 2022; 15(3): 159–166.References
[1] Food and Agriculture Organization of the United Nations (FAO). Current market situation and medium term outlook: FAO. Available: http://www.fao.org/3/BU642en/bu642en.pdf. Accessed on [2020-11-23].
[2] Han Y, Xiao R H, Song Y Z, Ding Q W. Design and evaluation of tea-plucking machine for improving quality of tea. Applied Engineering in Agriculture, 2019; 35(6): 979–986.
[3] Yang H, Chen L, Ma Z, Chen M, Zhong Y, Deng F, et al. Computer vision-based high-quality tea automatic plucking robot using Delta parallel manipulator. Computers and Electronics in Agriculture, 2021; 181: 105946. doi: 10.1016/j.compag.2020.105946.
[4] Motokura K, Takahashi M, Ewerton M, Peters J. Plucking motions for tea harvesting robots using probabilistic movement primitives. IEEE Robotics and Automation Letters, 2020; 5(2): 3275–3282.
[5] Yuan J. Research process analysis of robotics selective harvesting technologies. Transactions of the CSAM, 2020; 51(9): 1–17. (in Chinese)
[6] Tang Y, Han W, Hu A, Wang W. Design and experiment of intelligentized tea-plucking machine for human riding based on machine vision. Transactions of the CSAM, 2016; 47(7): 15–20. (in Chinese)
[7] Zhang L, Zhang H, Chen Y, Dai S, Li X, Kenji I, et al. Real-time monitoring of optimum timing for harvesting fresh tea leaves based on machine vision. Int J Agric & Biol Eng, 2019; 12(1): 6–9.
[8] Karunasena G, Priyankara H. Tea bud leaf identification by using machine learning and image processing techniques. International Journal of Scientific & Engineering Research, 2020; 11(8): 624–628.
[9] Zhang L, Zou L, Wu C, Jia J, Chen J. Method of famous tea sprout identification and segmentation based on improved watershed algorithm. Computers and Electronics in Agriculture, 2021; 184: 106108. doi: 10.1016/j.compag.2021.106108.
[10] Li Y, He L, Jia J, Lyu J, Chen J, Qiao X, et al. In-field tea shoot detection and 3D localization using an RGB-D camera. Computers and Electronics in Agriculture, 2021; 185: 106149. doi: 10.1016/j.compag.2021.106149.
[11] Kamilaris A, Prenafeta-Boldu F X. Deep learning in agriculture: A survey. Computers and Electronics in Agriculture, 2018; 147(1): 70–90.
[12] Yang H, Chen L, Chen M, Ma Z, Deng F, Li M, et al. Tender tea shoots recognition and positioning for picking robot using improved YOLO-V3 Model. IEEE Access, 2019; 7: 180998-181011.
[13] Chen Y-T, Chen S-F. Localizing plucking points of tea leaves using deep convolutional neural networks. Computers and Electronics in Agriculture, 2020; 171: 105298. doi: 10.1016/j.compag.220.105298.
[14] LeCun Y, Bengio Y, Hinton G. Deep learning. Nature, 2015; 521(7553): 436–444.
[15] Liu L, Ouyang W, Wang X, Fieguth P, Chen J, Liu X, et al. Deep learning for generic object detection: A survey. International Journal of Computer Vision, 2020; 128(2): 261–318.
[16] Szegedy C, Wei L, Yangqing J, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convolutions. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, USA: IEEE, 2015; pp.1–9. doi: 10.1109/CVPR.2015.7298594.
[17] Tan M X, Pang R, Le Q V. EfficientDet: Scalable and efficient object detection. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, USA: IEEE, 2020; pp.10778–10787. doi: 10.1109/CVPR42600.2020.01079.
[18] Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C. Learning efficient convolutional networks through network slimming. In: 2017 IEEE International Conference on Computer Vision, Venice, Italy: IEEE, 2017; pp.2755–2763. doi: 10.1109/ICCV.2017.298.
[19] Molchanov P, Mallya A, Tyree S, Frosio I, Kautz J. Importance estimation for neural network pruning. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, USA: IEEE, 2019; pp.11256–11264. doi: 10.1109/CVPR.2019.01152.
[20] Han S, Pool J, Tran J, Dally W J. Learning both weights and connections for efficient neural networks. In: Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, Canada: MIT Press, 2015; pp.1135–1143.
[21] Zhang P Y, Zhong Y X, Li X Q. SlimYOLOv3: Narrower, faster and better for real-time UAV applications. In: 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul: IEEE, 2019; pp.37–45. doi: 10.1109/ICCVW.2019.00011.
[22] Wu D, Lyu S, Jiang M, Song H. Using channel pruning-based YOLO v4 deep learning algorithm for the real-time and accurate detection of apple flowers in natural environments. Computers and Electronics in Agriculture, 2020; 178: 105742. doi: 10.1016/j.compag.2020.105742.
[23] Wang D, He D. Channel pruned YOLO V5s-based deep learning approach for rapid and accurate apple fruitlet detection before fruit thinning. Biosystems Engineering, 2021; 210: 271–281.
[24] Xu P, Cao J, Shang F, Sun W, Li P. Layer pruning via fusible residual convolutional block for deep neural networks. arXiv. 2020; arXiv: 2011.14356v1. doi: 10.48550/arXiv.2011.14356.
[25] Redmon J, Farhadi A. Yolov3: An incremental improvement. arXiv, 2018; arXiv:1804.02767. doi: 10.48550/arXiv.1804.02767.
[26] He K M, Zhang X Y, Ren S Q, Sun J. Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, USA: IEEE, 2016; pp.770–778. doi: 10.1109/CVPR.2016.90.
[27] He Y, Zhang X, Sun J. Channel pruning for accelerating very deep neural networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy: IEEE, 2017; pp.1398–1406. doi: 10.1109/ ICCV.2017.155.
[2] Han Y, Xiao R H, Song Y Z, Ding Q W. Design and evaluation of tea-plucking machine for improving quality of tea. Applied Engineering in Agriculture, 2019; 35(6): 979–986.
[3] Yang H, Chen L, Ma Z, Chen M, Zhong Y, Deng F, et al. Computer vision-based high-quality tea automatic plucking robot using Delta parallel manipulator. Computers and Electronics in Agriculture, 2021; 181: 105946. doi: 10.1016/j.compag.2020.105946.
[4] Motokura K, Takahashi M, Ewerton M, Peters J. Plucking motions for tea harvesting robots using probabilistic movement primitives. IEEE Robotics and Automation Letters, 2020; 5(2): 3275–3282.
[5] Yuan J. Research process analysis of robotics selective harvesting technologies. Transactions of the CSAM, 2020; 51(9): 1–17. (in Chinese)
[6] Tang Y, Han W, Hu A, Wang W. Design and experiment of intelligentized tea-plucking machine for human riding based on machine vision. Transactions of the CSAM, 2016; 47(7): 15–20. (in Chinese)
[7] Zhang L, Zhang H, Chen Y, Dai S, Li X, Kenji I, et al. Real-time monitoring of optimum timing for harvesting fresh tea leaves based on machine vision. Int J Agric & Biol Eng, 2019; 12(1): 6–9.
[8] Karunasena G, Priyankara H. Tea bud leaf identification by using machine learning and image processing techniques. International Journal of Scientific & Engineering Research, 2020; 11(8): 624–628.
[9] Zhang L, Zou L, Wu C, Jia J, Chen J. Method of famous tea sprout identification and segmentation based on improved watershed algorithm. Computers and Electronics in Agriculture, 2021; 184: 106108. doi: 10.1016/j.compag.2021.106108.
[10] Li Y, He L, Jia J, Lyu J, Chen J, Qiao X, et al. In-field tea shoot detection and 3D localization using an RGB-D camera. Computers and Electronics in Agriculture, 2021; 185: 106149. doi: 10.1016/j.compag.2021.106149.
[11] Kamilaris A, Prenafeta-Boldu F X. Deep learning in agriculture: A survey. Computers and Electronics in Agriculture, 2018; 147(1): 70–90.
[12] Yang H, Chen L, Chen M, Ma Z, Deng F, Li M, et al. Tender tea shoots recognition and positioning for picking robot using improved YOLO-V3 Model. IEEE Access, 2019; 7: 180998-181011.
[13] Chen Y-T, Chen S-F. Localizing plucking points of tea leaves using deep convolutional neural networks. Computers and Electronics in Agriculture, 2020; 171: 105298. doi: 10.1016/j.compag.220.105298.
[14] LeCun Y, Bengio Y, Hinton G. Deep learning. Nature, 2015; 521(7553): 436–444.
[15] Liu L, Ouyang W, Wang X, Fieguth P, Chen J, Liu X, et al. Deep learning for generic object detection: A survey. International Journal of Computer Vision, 2020; 128(2): 261–318.
[16] Szegedy C, Wei L, Yangqing J, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convolutions. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, USA: IEEE, 2015; pp.1–9. doi: 10.1109/CVPR.2015.7298594.
[17] Tan M X, Pang R, Le Q V. EfficientDet: Scalable and efficient object detection. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, USA: IEEE, 2020; pp.10778–10787. doi: 10.1109/CVPR42600.2020.01079.
[18] Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C. Learning efficient convolutional networks through network slimming. In: 2017 IEEE International Conference on Computer Vision, Venice, Italy: IEEE, 2017; pp.2755–2763. doi: 10.1109/ICCV.2017.298.
[19] Molchanov P, Mallya A, Tyree S, Frosio I, Kautz J. Importance estimation for neural network pruning. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, USA: IEEE, 2019; pp.11256–11264. doi: 10.1109/CVPR.2019.01152.
[20] Han S, Pool J, Tran J, Dally W J. Learning both weights and connections for efficient neural networks. In: Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, Canada: MIT Press, 2015; pp.1135–1143.
[21] Zhang P Y, Zhong Y X, Li X Q. SlimYOLOv3: Narrower, faster and better for real-time UAV applications. In: 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul: IEEE, 2019; pp.37–45. doi: 10.1109/ICCVW.2019.00011.
[22] Wu D, Lyu S, Jiang M, Song H. Using channel pruning-based YOLO v4 deep learning algorithm for the real-time and accurate detection of apple flowers in natural environments. Computers and Electronics in Agriculture, 2020; 178: 105742. doi: 10.1016/j.compag.2020.105742.
[23] Wang D, He D. Channel pruned YOLO V5s-based deep learning approach for rapid and accurate apple fruitlet detection before fruit thinning. Biosystems Engineering, 2021; 210: 271–281.
[24] Xu P, Cao J, Shang F, Sun W, Li P. Layer pruning via fusible residual convolutional block for deep neural networks. arXiv. 2020; arXiv: 2011.14356v1. doi: 10.48550/arXiv.2011.14356.
[25] Redmon J, Farhadi A. Yolov3: An incremental improvement. arXiv, 2018; arXiv:1804.02767. doi: 10.48550/arXiv.1804.02767.
[26] He K M, Zhang X Y, Ren S Q, Sun J. Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, USA: IEEE, 2016; pp.770–778. doi: 10.1109/CVPR.2016.90.
[27] He Y, Zhang X, Sun J. Channel pruning for accelerating very deep neural networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy: IEEE, 2017; pp.1398–1406. doi: 10.1109/ ICCV.2017.155.
Downloads
Published
2022-06-30
How to Cite
Li, Y., He, L., jia, J., Chen, J., Lyu, J., & Wu, C. (2022). High-efficiency tea shoot detection method via a compressed deep learning model. International Journal of Agricultural and Biological Engineering, 15(3), 159–166. Retrieved from https://ijabe.migration.pkpps03.publicknowledgeproject.org/index.php/ijabe/article/view/6896
Issue
Section
Information Technology, Sensors and Control Systems
License
IJABE is an international peer reviewed open access journal, adopting Creative Commons Copyright Notices as follows.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).