Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods

被引:15
|
作者
Coll-Ribes, Gabriel [1 ]
Torres-Rodriguez, Ivan J. [1 ]
Grau, Antoni [2 ]
Guerra, Edmundo [2 ]
Sanfeliu, Alberto [1 ,2 ]
机构
[1] UPC, CSIC, Inst Robot & Informat Ind, Llorens & Artiga 4-6, Barcelona 08028, Spain
[2] Univ Politecn Cataluna, Pau Gargallo 5, Barcelona 08028, Spain
基金
欧盟地平线“2020”;
关键词
Image segmentation; Monocular depth; Grape bunch and peduncle detection; Grape bunch and peduncle depth estimation; Robot harvesting;
D O I
10.1016/j.compag.2023.108362
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape harvesting with robots. The bunch and peduncle detection methods presented in this work rely on a combination of instance segmentation and monocular depth estimation using Convolutional Neural Networks (CNN). Regarding depth estimation, we propose a combination of different depth techniques that allow precise localization of the peduncle using traditional stereo cameras, even with the particular complexity of grape peduncles. The methods proposed in this work have been tested on the WGISD (Embrapa Wine Grape Instance Segmentation) dataset, improving the results of state-of-the-art techniques. Furthermore, within the context of the EU project CANOPIES, the methods have also been tested on a dataset of 1,326 RGB-D images of table grapes, recorded at the Corsira Agricultural Cooperative Society (Aprilia, Italy), using a Realsense D435i camera located at the arm of a CANOPIES two-manipulator robot developed in the project. The detection results on the WGISD dataset show that the use of RGB-D information (mAP = 0.949) leads to superior performance compared to the use of RGB data alone (mAP = 0.891). This trend is also evident in the CANOPIES Grape Bunch and Peduncle dataset, where the mAP for RGB-D images (mAP = 0.767) outperforms that of RGB data (mAP = 0.725). Regarding depth estimation, our method achieves a mean squared error of 2.66 cm within a distance of 1 m in the CANOPIES dataset.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] Attention-Based Grasp Detection With Monocular Depth Estimation
    Xuan Tan, Phan
    Hoang, Dinh-Cuong
    Nguyen, Anh-Nhat
    Nguyen, Van-Thiep
    Vu, Van-Duc
    Nguyen, Thu-Uyen
    Hoang, Ngoc-Anh
    Phan, Khanh-Toan
    Tran, Duc-Thanh
    Vu, Duy-Quang
    Ngo, Phuc-Quan
    Duong, Quang-Tri
    Ho, Ngoc-Trung
    Tran, Cong-Trinh
    Duong, Van-Hiep
    Mai, Anh-Truong
    IEEE ACCESS, 2024, 12 : 65041 - 65057
  • [22] Multi-scale Deep CNN Network for Unsupervised Monocular Depth Estimation
    Wan Yingcai
    Fang Lijing
    Zhao Qiankun
    2019 9TH IEEE ANNUAL INTERNATIONAL CONFERENCE ON CYBER TECHNOLOGY IN AUTOMATION, CONTROL, AND INTELLIGENT SYSTEMS (IEEE-CYBER 2019), 2019, : 475 - 479
  • [23] Realtime depth estimation and obstacle detection from monocular video
    Wedel, Andreas
    Franke, Uwe
    Klappstein, Jens
    Brox, Thomas
    Cremers, Daniel
    PATTERN RECOGNITION, PROCEEDINGS, 2006, 4174 : 475 - 484
  • [24] Computer vision methods for robot tasks: Motion detection, depth estimation and tracking
    Martinez-Martin, E.
    AI COMMUNICATIONS, 2012, 25 (04) : 373 - 375
  • [25] Bayesian cue integration of structure from motion and CNN-based monocular depth estimation for autonomous robot navigation
    Mumuni, Fuseini
    Mumuni, Alhassan
    INTERNATIONAL JOURNAL OF INTELLIGENT ROBOTICS AND APPLICATIONS, 2022, 6 (02) : 191 - 206
  • [26] Bayesian cue integration of structure from motion and CNN-based monocular depth estimation for autonomous robot navigation
    Fuseini Mumuni
    Alhassan Mumuni
    International Journal of Intelligent Robotics and Applications, 2022, 6 : 191 - 206
  • [27] Obstacles detection and depth estimation from monocular vision for inspection robot of high voltage transmission line
    Li Cheng
    Wu, Gongping
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (02): : S2611 - S2627
  • [28] Obstacles detection and depth estimation from monocular vision for inspection robot of high voltage transmission line
    Li Cheng
    Gongping Wu
    Cluster Computing, 2019, 22 : 2611 - 2627
  • [29] Depth Estimation Matters Most: Improving Per-Object Depth Estimation for Monocular 3D Detection and Tracking
    Jing, Longlong
    Yu, Ruichi
    Kretzschmar, Henrik
    Li, Kang
    Qi, Charles R.
    Zhao, Hang
    Ayvaci, Alper
    Chen, Xu
    Cower, Dillon
    Li, Yingwei
    You, Yurong
    Deng, Han
    Li, Congcong
    Anguelov, Dragomir
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022,
  • [30] Lightweight Self-Supervised Monocular Depth Estimation Through CNN and Transformer Integration
    Wang, Zhe
    Zou, Yongjia
    Lv, Jin
    Cao, Yang
    Yu, Hongfei
    IEEE ACCESS, 2024, 12 : 167934 - 167943