Efficient detection model of green target fruit based on optimized Transformer network

被引:0
|
作者
Jia W. [1 ,2 ]
Meng H. [1 ]
Ma X. [1 ]
Zhao Y. [1 ]
Ji Z. [3 ]
Zheng Y. [1 ]
机构
[1] School of Information Science and Engineering, Shandong Normal University, Jinan
[2] Key Laboratory of Facility Agriculture Measurement and Control Technology and Equipment of Machinery Industry, Zhenjiang
[3] School of Engineering, Cardiff University, Cardiff
来源
Zheng, Yuanjie (yjzheng@sdnu.edu.cn) | 1600年 / Chinese Society of Agricultural Engineering卷 / 37期
关键词
Bootstrapping; Green fruit; Object detection; Transfer learning; Transformer networks;
D O I
10.11975/j.issn.1002-6819.2021.14.018
中图分类号
学科分类号
摘要
The posture of target fruit is ever-changing in the complex orchard environment. Some target fruits are homochromatic with background, and the limited number of samples have brought great challenges to accurately detect the target, due mainly to the difficulty of collecting some environmental data. Therefore, the detection needs to meet the high requirements of intelligent yield measurement and automatic harvesting, in terms of both accuracy and efficiency. In this study, an efficient detection model was proposed for the green target fruits suitable for small samples under different light conditions and fruit postures. The optimized transformer network was also employed in this model. Firstly, the convolutional neural network (CNN) was used to extract image features. The transformer encoder was input after feature dimension reduction and positional encoding. Multi-head attention and feed-forward network (FFN) were then selected to obtain the encoder output. Secondly, the transformer decoder processed the input using multi-head attention and feed-forward network. The positional encoding was then added to each link. After that, the outputs were generated with different data sizes. The bounding boxes were much larger than the actual objects after prediction, indicating a low missing rate of green target fruit after decoder settlement. Finally, the feed-forward network (FFN) was utilized to predict the detection. The training of detection model was mostly used sufficient samples to avoid overfitting in the training process for higher generalization of the model. Bootstrapping was also introduced to repeatedly mapping the original data for several times. As such, the expanding dataset was utilized to meet the high requirement of larger samples for the higher accuracy of detection mode in the training process. Transfer learning was selected to significantly improve the training efficiency of the model, while, accelerate the convergence of the network. The apple and persimmon datasets were made separately for the model training. The experimental results show that the training efficiency of the model was greatly improved by more than 13% after migration learning. An excellent illustration of features transferability increased the speed and efficiency of detection, as the difference decreased between the pre-training and target task. Transfer learning was adopted to improve the efficiency of the model, where the model converged faster and was better suitable for the complex orchard environment. The new model can widely be expected to effectively realize the detection of green target fruit in the complex orchard environment with multiple postures, illumination, and scenes, indicating better generalization ability and robustness. The accuracies of detection were 93.27% and 91.35%, respectively, when testing green persimmons and green apples. Consequently, the new optimized model presented the best performance, compared with the conventional. The finding can also provide a sound theoretical reference for the target detection of green fruits and vegetables in the intelligent yield measurement and automated harvesting in orchards. © 2021, Editorial Department of the Transactions of the Chinese Society of Agricultural Engineering. All right reserved.
引用
收藏
页码:163 / 170
页数:7
相关论文
共 32 条
  • [1] Li Yinian, Du Shiwei, Yao Min, Et al., Method for wheatear counting and yield predicting based on image of wheatear population in field, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 34, 21, pp. 185-194, (2018)
  • [2] van Klompenburg T, Kassahun A, Catal C., Crop yield prediction using machine learning: A systematic literature review, Computers and Electronics in Agriculture, 177, (2020)
  • [3] Wang Dandan, Song Huaibo, He Dongjian, Research advance on vision system of apple picking robot, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 33, 10, pp. 59-69, (2017)
  • [4] Jia W K, Zhang Y, Lian J, Et al., Apple harvesting robot under information technology: A review, International Journal of Advanced Robotic Systems, 17, 3, (2020)
  • [5] Ebrahimi M, Khoshtaghaza M, Minaei S, Et al., Vision-based pest detection based on SVM classification method, Computers and Electronics in Agriculture, 137, pp. 52-58, (2017)
  • [6] Ngugi L C, Abelwahab M, Abo-Zahhad M., Recent advances in image processing techniques for automated leaf pest and disease recognition: A review, Information Processing in Agriculture, 8, 1, pp. 27-51, (2021)
  • [7] Zhao Dean, Wu Rendi, Liu Xiaoyang, Et al., Apple positioning based on YOLO deep convolutional neural network for picking robot in complex background, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 35, 3, pp. 164-173, (2019)
  • [8] Guo Cailing, Liu Gang, Reconstruction method of apple tree canopy point-leaf model based on 3D point clouds, Transactions of the Chinese Society for Agricultural Machinery, 51, 4, pp. 173-180, (2020)
  • [9] Li Guoli, Ji Changying, Gu Baoxing, Recognition and location of oscillating fruit based on monocular vision and ultrasonic testing, Transactions of the Chinese Society for Agricultural Machinery, 46, 11, pp. 1-8, (2015)
  • [10] Xu Linyun, Liu Guanhua, Xuan Yan, Et al., Mathematical model of fruit's aerial movement in vibration response, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 35, 16, pp. 206-213, (2019)