Recognizing Personal Locations From Egocentric Videos

被引:28
|
作者
Furnari, Antonino [1 ]
Farinella, Giovanni Maria [1 ]
Battiato, Sebastiano [1 ]
机构
[1] Univ Catania, Dept Math & Comp Sci, I-95124 Catania, Italy
关键词
Context-aware computing; egocentric dataset; egocentric vision; first person vision; personal location recognition; CONTEXT; CLASSIFICATION; RECOGNITION; SCENE; SHAPE;
D O I
10.1109/THMS.2016.2612002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Contextual awareness in wearable computing allows for construction of intelligent systems, which are able to interact with the user in a more natural way. In this paper, we study how personal locations arising from the user's daily activities can be recognized from egocentric videos. We assume that few training samples are available for learning purposes. Considering the diversity of the devices available on the market, we introduce a benchmark dataset containing egocentric videos of eight personal locations acquired by a user with four different wearable cameras. To make our analysis useful in real-world scenarios, we propose a method to reject negative locations, i.e., those not belonging to any of the categories of interest for the end-user. We assess the performances of the main state-of-the-art representations for scene and object classification on the considered task, as well as the influence of device-specific factors such as the field of view and the wearing modality. Concerning the different device-specific factors, experiments revealed that the best results are obtained using a head-mounted wide-angular device. Our analysis shows the effectiveness of using representations based on convolutional neural networks, employing basic transfer learning techniques and an entropy-based rejection algorithm.
引用
收藏
页码:6 / 18
页数:13
相关论文
共 50 条
  • [21] Recognizing persons in images by learning from videos
    Hoerster, Eva
    Lux, Jochen
    Lienhart, Rainer
    MULTIMEDIA CONTENT ACCESS: ALGORITHMS AND SYSTEMS, 2007, 6506
  • [22] Recognizing Realistic Actions from Videos "in the Wild"
    Liu, Jingen
    Luo, Jiebo
    Shah, Mubarak
    CVPR: 2009 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-4, 2009, : 1996 - +
  • [23] RECOGNIZING OFFENSIVE STRATEGIES FROM FOOTBALL VIDEOS
    Li, Ruonan
    Chellappa, Rama
    2010 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 2010, : 4585 - 4588
  • [24] Recognizing Actions in Videos from Unseen Viewpoints
    Piergiovanni, A. J.
    Ryoo, Michael S.
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4122 - 4130
  • [25] Self-Supervised Object Detection from Egocentric Videos
    Akiva, Peri
    Huang, Jing
    Liang, Kevin J.
    Kovvuri, Rama
    Chen, Xingyu
    Feiszli, Matt
    Dana, Kristin
    Hassner, Tal
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5202 - 5214
  • [26] Action Recognition from Egocentric Videos Using Random Walks
    Sahu, Abhimanyu
    Bhattacharya, Rajit
    Bhura, Pallabh
    Chowdhury, Ananda S.
    PROCEEDINGS OF 3RD INTERNATIONAL CONFERENCE ON COMPUTER VISION AND IMAGE PROCESSING, CVIP 2018, VOL 2, 2020, 1024 : 389 - 402
  • [27] Next-active-object prediction from egocentric videos
    Furnari, Antonino
    Battiato, Sebastiano
    Grauman, Kristen
    Farinella, Giovanni Maria
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 49 : 401 - 411
  • [28] Summarization of Egocentric Videos: A Comprehensive Survey
    del Molino, Ana Garcia
    Tan, Cheston
    Lim, Joo-Hwee
    Tan, Ah-Hwee
    IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2017, 47 (01) : 65 - 76
  • [29] Salient object detection in egocentric videos
    Zhang, Hao
    Liang, Haoran
    Zhao, Xing
    Liu, Jian
    Liang, Ronghua
    IET IMAGE PROCESSING, 2024, 18 (08) : 2028 - 2037
  • [30] Text Synopsis Generation for Egocentric Videos
    Sharghi, Aidean
    Lobo, Niels da Vitoria
    Shah, Mubarak
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 4252 - 4259