Semantic scene understanding on mobile device with illumination invariance for the visually impaired

被引:0
|
作者
Xu, Chengyou [1 ]
Wang, Kaiwei [1 ]
Yang, Kailun [1 ]
Cheng, Ruiqi [1 ]
Bai, Jian [1 ]
机构
[1] Zhejiang Univ, State Key Lab Modern Opt Instrumentat, Hangzhou 310027, Peoples R China
关键词
Semantic segmentation; robustness; illumination invariance; scene understanding; mobile device;
D O I
10.1117/12.2532550
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For Visually Impaired People (VIP), it's very difficult to perceive their surroundings. To address this problem, we propose a scene understanding system to aid VIP in indoor and outdoor environments. Semantic segmentation performance is generally sensitive to the environment and illumination changes, including the change between indoor and outdoor environments and the change across different weather conditions. Meanwhile, most existing methods have paid more attention on either the accuracy or the efficiency, instead of the balance between both of them. In the proposed system, the training dataset is preprocessed by using an illumination-invariant transformation to weaken the impact of illumination changes and improve the robustness of the semantic segmentation network. Regarding the structure of semantic segmentation network, the lightweight networks such as MobileNetV2 and ShuffleNet V2 are employed as the backbone of DeepLabv3+ to improve the accuracy with little increasing of computation, which is suitable for mobile assistance device. We evaluate the robustness of the segmentation model across different environments on the Gardens Point Walking dataset, and demonstrate the extremely positive effect of the illumination-invariant pre-transformation in challenging real-world domain. The network trained on computer achieves a relatively high accuracy on ADE20K relabeled into 20 classes. The frame rate of the proposed system is up to 83 FPS on a 1080Ti GPU.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Semantic Foggy Scene Understanding with Synthetic Data
    Sakaridis, Christos
    Dai, Dengxin
    Van Gool, Luc
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2018, 126 (09) : 973 - 992
  • [42] Adopting Abstract Images for Semantic Scene Understanding
    Zitnick, C. Lawrence
    Vedantam, Ramakrishna
    Parikh, Devi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (04) : 627 - 638
  • [43] VPER - A Visual HelPER Device for the Visually Impaired Users
    Chen, Christine P.
    Huang, Leila Z.
    Tang, Tiffany Y.
    Wang, Relic Y.
    Winoto, Pinata
    2015 SECOND INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND MACHINE INTELLIGENCE (ISCMI), 2015, : 88 - 92
  • [44] RaMonic: Obstacle warning device for Visually Impaired People
    Warunsin, Kulwarun
    Atthayuwat, Kiattichai
    LifeTech 2022 - 2022 IEEE 4th Global Conference on Life Sciences and Technologies, 2022, : 130 - 132
  • [45] Assistive device to aid the visually impaired to play basketball
    Murnyack, Roberta
    Viktoriany, Khristi
    Sediqe, Ajmal
    Chong, Waiseng
    American Society of Mechanical Engineers, Bioengineering Division (Publication) BED, 1999, 43 : 11 - 12
  • [46] Exploiting context for semantic scene content understanding
    Luo, Jiebo
    ICIS '06: INTERNATIONAL CONGRESS OF IMAGING SCIENCE, FINAL PROGRAM AND PROCEEDINGS: LINKING THE EXPLOSION OF IMAGING APPLICATIONS WITH THE SCIENCE AND TECHNOLOGY OF IMAGING, 2006, : 479 - 479
  • [47] MAVI: An Embedded Device to Assist Mobility of Visually Impaired
    Kedia, Rajesh
    Yoosuf, K. K.
    Dedeepya, Pappireddy
    Fazal, Munib
    Arora, Chetan
    Balakrishnan, M.
    2017 30TH INTERNATIONAL CONFERENCE ON VLSI DESIGN AND 2017 16TH INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (VLSID 2017), 2017, : 213 - 218
  • [48] Assessment of a simple obstacle detection device for the visually impaired
    Lee, Cheng-Lung
    Chen, Chih-Yung
    Sung, Peng-Cheng
    Lu, Shih-Yi
    APPLIED ERGONOMICS, 2014, 45 (04) : 817 - 824
  • [49] An IoT based Smart Assistive Device for the Visually Impaired
    Talukdar, Parbin S.
    Hazarika, Bishmita
    Sharma, Jayshree
    Matam, Rakesh
    2020 IEEE REGION 10 SYMPOSIUM (TENSYMP) - TECHNOLOGY FOR IMPACTFUL SUSTAINABLE DEVELOPMENT, 2020, : 291 - 294
  • [50] Semantic Foggy Scene Understanding with Synthetic Data
    Christos Sakaridis
    Dengxin Dai
    Luc Van Gool
    International Journal of Computer Vision, 2018, 126 : 973 - 992