Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods

被引:15
|
作者
Coll-Ribes, Gabriel [1 ]
Torres-Rodriguez, Ivan J. [1 ]
Grau, Antoni [2 ]
Guerra, Edmundo [2 ]
Sanfeliu, Alberto [1 ,2 ]
机构
[1] UPC, CSIC, Inst Robot & Informat Ind, Llorens & Artiga 4-6, Barcelona 08028, Spain
[2] Univ Politecn Cataluna, Pau Gargallo 5, Barcelona 08028, Spain
基金
欧盟地平线“2020”;
关键词
Image segmentation; Monocular depth; Grape bunch and peduncle detection; Grape bunch and peduncle depth estimation; Robot harvesting;
D O I
10.1016/j.compag.2023.108362
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape harvesting with robots. The bunch and peduncle detection methods presented in this work rely on a combination of instance segmentation and monocular depth estimation using Convolutional Neural Networks (CNN). Regarding depth estimation, we propose a combination of different depth techniques that allow precise localization of the peduncle using traditional stereo cameras, even with the particular complexity of grape peduncles. The methods proposed in this work have been tested on the WGISD (Embrapa Wine Grape Instance Segmentation) dataset, improving the results of state-of-the-art techniques. Furthermore, within the context of the EU project CANOPIES, the methods have also been tested on a dataset of 1,326 RGB-D images of table grapes, recorded at the Corsira Agricultural Cooperative Society (Aprilia, Italy), using a Realsense D435i camera located at the arm of a CANOPIES two-manipulator robot developed in the project. The detection results on the WGISD dataset show that the use of RGB-D information (mAP = 0.949) leads to superior performance compared to the use of RGB data alone (mAP = 0.891). This trend is also evident in the CANOPIES Grape Bunch and Peduncle dataset, where the mAP for RGB-D images (mAP = 0.767) outperforms that of RGB data (mAP = 0.725). Regarding depth estimation, our method achieves a mean squared error of 2.66 cm within a distance of 1 m in the CANOPIES dataset.
引用
收藏
页数:17
相关论文
共 50 条
  • [11] FastMDE: A Fast CNN Architecture for Monocular Depth Estimation at High Resolution
    Thien-Thanh Dao
    Quoc-Viet Pham
    Hwang, Won-Joo
    IEEE ACCESS, 2022, 10 : 16111 - 16122
  • [12] LightDepthNet: Lightweight CNN Architecture for Monocular Depth Estimation on Edge Devices
    Liu, Qingliang
    Zhou, Shuai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (04) : 2389 - 2393
  • [13] Car depth estimation within a monocular image using a light CNN
    Tighkhorshid, Amirhossein
    Tousi, Seyed Mohamad Ali
    Nikoofard, Amirhossein
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (16): : 17944 - 17961
  • [14] Car depth estimation within a monocular image using a light CNN
    Amirhossein Tighkhorshid
    Seyed Mohamad Ali Tousi
    Amirhossein Nikoofard
    The Journal of Supercomputing, 2023, 79 : 17944 - 17961
  • [15] Depth Estimation by Combining Binocular Stereo and Monocular Structured-Light
    Xu, Yuhua
    Yang, Xiaoli
    Yu, Yushan
    Jia, Wei
    Chu, Zhaobi
    Guo, Yulan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 1736 - 1745
  • [16] Accurate unsupervised monocular depth estimation for ill-posed region
    Wang, Xiaofeng
    Sun, Jiameng
    Qin, Hao
    Yuan, Yuxing
    Yu, Jun
    Su, Yingying
    Sun, Zhiheng
    FRONTIERS IN PHYSICS, 2023, 10
  • [17] Knowledge Distillation for Fast and Accurate Monocular Depth Estimation on Mobile Devices
    Wang, Yiran
    Li, Xingyi
    Shi, Min
    Xian, Ke
    Cao, Zhiguo
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 2457 - 2465
  • [18] SharpNet: Fast and Accurate Recovery of Occluding Contours in Monocular Depth Estimation
    Ramamonjisoa, Michael
    Lepetit, Vincent
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 2109 - 2118
  • [19] Self-supervised monocular depth estimation with direct methods
    Wang, Haixia
    Sun, Yehao
    Wu, Q. M. Jonathan
    Lu, Xiao
    Wang, Xiuling
    Zhang, Zhiguo
    NEUROCOMPUTING, 2021, 421 : 340 - 348
  • [20] Self-supervised monocular depth estimation with direct methods
    Wang H.
    Sun Y.
    Wu Q.M.J.
    Lu X.
    Wang X.
    Zhang Z.
    Neurocomputing, 2021, 421 : 340 - 348