Joint learning of object and action detectors

被引:24
|
作者
Kalogeiton, Vicky [1 ,2 ]
Weinzaepfel, Philippe [3 ]
Ferrari, Vittorio [2 ]
Schmid, Cordelia [1 ]
机构
[1] Univ Grenoble Alpes, INRIA, CNRS, Grenoble INP,LJK, F-38000 Grenoble, France
[2] Univ Edinburgh, Edinburgh, Midlothian, Scotland
[3] Naver Labs Europe, Meylan, France
关键词
D O I
10.1109/ICCV.2017.219
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While most existing approaches for detection in videos focus on objects or human actions separately, we aim at jointly detecting objects performing actions, such as cat eating or dog jumping. We introduce an end-to-end multi-task objective that jointly learns object-action relationships. We compare it with different training objectives, validate its effectiveness for detecting objects-actions in videos, and show that both tasks of object and action detection benefit from this joint learning. Moreover, the proposed architecture can be used for zero-shot learning of actions: our multitask objective leverages the commonalities of an action performed by different objects, e.g. dog and cat jumping, enabling to detect actions of an object without training with these object-actions pairs. In experiments on the A2D dataset [50], we obtain state-of-the-art results on segmentation of object-action pairs. We finally apply our multitask architecture to detect visual relationships between objects in images of the VRD dataset [24].
引用
收藏
页码:2001 / 2010
页数:10
相关论文
共 50 条
  • [41] Joint Learning with Group Relation and Individual Action
    Nakatani, Chihiro
    Kawashima, Hiroaki
    Ukita, Norimichi
    2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,
  • [42] From Fly Detectors to Action Control: Representations in Reinforcement Learning
    Rusanen, Anna-Mari
    Lappi, Otto
    Pekkanen, Jami
    Kuokkanen, Jesse
    PHILOSOPHY OF SCIENCE, 2021, 88 (05) : 1045 - 1054
  • [43] Learning the semantics of object-action relations by observation
    Aksoy, Eren Erdal
    Abramov, Alexey
    Doerr, Johannes
    Ning, Kejun
    Dellen, Babette
    Woergoetter, Florentin
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2011, 30 (10): : 1229 - 1249
  • [44] Me or we? Action-outcome learning in synchronous joint action
    Marschner, Maximilian
    Dignath, David
    Knoblich, Guenther
    COGNITION, 2024, 247
  • [45] Joint Action Understanding improves Robot-to-Human Object Handover
    Grigore, Elena Corina
    Eder, Kerstin
    Pipe, Anthony G.
    Melhuish, Chris
    Leonards, Ute
    2013 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2013, : 4622 - 4629
  • [46] Higher-order action planning for individual and joint object manipulations
    Marlene Meyer
    Robrecht P. R. D. van der Wel
    Sabine Hunnius
    Experimental Brain Research, 2013, 225 : 579 - 588
  • [47] Higher-order action planning for individual and joint object manipulations
    Meyer, Marlene
    van der Wel, Robrecht P. R. D.
    Hunnius, Sabine
    EXPERIMENTAL BRAIN RESEARCH, 2013, 225 (04) : 579 - 588
  • [48] SegFlow: Joint Learning for Video Object Segmentation and Optical Flow
    Cheng, Jingchun
    Tsai, Yi-Hsuan
    Wang, Shengjin
    Yang, Ming-Hsuan
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 686 - 695
  • [49] Joint learning of visual attributes, object classes and visual saliency
    Wang, Gang
    Forsyth, David
    2009 IEEE 12TH INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2009, : 537 - 544
  • [50] Joint learning of foreground, background and edge for salient object detection
    Wu, Qin
    Zhu, Pengcheng
    Chai, Zhilei
    Guo, Guodong
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 240