Physically Grounded Spatio-temporal Object Affordances

被引:0
|
作者
Koppula, Hema S. [1 ]
Saxena, Ashutosh [1 ]
机构
[1] Cornell Univ, Dept Comp Sci, Ithaca, NY 14853 USA
来源
关键词
Object Affordances; 3D Object Models; Functional Representation of Environment; Generative Graphical Model; Trajectory Modeling; Human Activity Detection; RGBD Videos;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Objects in human environments support various functionalities which govern how people interact with their environments in order to perform tasks. In this work, we discuss how to represent and learn a functional understanding of an environment in terms of object affordances. Such an understanding is useful for many applications such as activity detection and assistive robotics. Starting with a semantic notion of affordances, we present a generative model that takes a given environment and human intention into account, and grounds the affordances in the form of spatial locations on the object and temporal trajectories in the 3D environment. The probabilistic model also allows uncertainties and variations in the grounded affordances. We apply our approach on RGB-D videos from Cornell Activity Dataset, where we first show that we can successfully ground the affordances, and we then show that learning such affordances improves performance in the labeling tasks.
引用
收藏
页码:831 / 847
页数:17
相关论文
共 50 条
  • [1] Recognizing Object Affordances in Terms of Spatio-Temporal Object-Object Relationships
    Pieropan, Alessandro
    Ek, Carl Henrik
    Kjellstrom, Hedvig
    [J]. 2014 14TH IEEE-RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2014, : 52 - 58
  • [2] Spatio-Temporal Object Recognition
    De Geest, Roeland
    Deboeverie, Francis
    Philips, Wilfried
    Tuytelaars, Tinne
    [J]. ADVANCED CONCEPTS FOR INTELLIGENT VISION SYSTEMS, ACIVS 2015, 2015, 9386 : 681 - 692
  • [3] Action Prediction Based on Physically Grounded Object Affordances in Human-Object Interactions
    Dutta, Vibekananda
    Zielinska, Teresa
    [J]. 2017 11TH INTERNATIONAL WORKSHOP ON ROBOT MOTION AND CONTROL (ROMOCO), 2017, : 41 - 46
  • [4] Probabilistic Spatio-Temporal Fusion of Affordances for Grasping and Manipulation
    Pohl, Christoph
    Asfour, Tamim
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3226 - 3233
  • [5] Object extraction by spatio-temporal assembling
    Qin, Xiaoke
    Tang, Liang
    Zhou, Be
    [J]. 2007 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOLS 1-7, 2007, : 2405 - +
  • [6] Spatio-temporal Object Detection Proposals
    Oneata, Dan
    Revaud, Jerome
    Verbeek, Jakob
    Schmid, Cordelia
    [J]. COMPUTER VISION - ECCV 2014, PT III, 2014, 8691 : 737 - 752
  • [7] Spatio-Temporal Attention Models for Grounded Video Captioning
    Zanfir, Mihai
    Marinoiu, Elisabeta
    Sminchisescu, Cristian
    [J]. COMPUTER VISION - ACCV 2016, PT IV, 2017, 10114 : 104 - 119
  • [8] Spatio-temporal relationships and video object extraction
    Deng, YN
    Manjunath, BS
    [J]. CONFERENCE RECORD OF THE THIRTY-SECOND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, VOLS 1 AND 2, 1998, : 895 - 899
  • [9] A query calculus for spatio-temporal object databases
    Griffiths, T
    Fernandes, AAA
    Djafri, N
    Paton, NW
    [J]. EIGHTH INTERNATIONAL SYMPOSIUM ON TEMPORAL REPRESENTATION AND REASONING, PROCEEDINGS, 2001, : 101 - 110
  • [10] Spatio-temporal detection of video moving object
    Ren, Ming-Yi
    Li, Xiao-Feng
    Li, Zai-Ming
    [J]. Guangdianzi Jiguang/Journal of Optoelectronics Laser, 2009, 20 (07): : 911 - 915