Attention-based video streaming

被引:8
|
作者
Dikici, Cagatay [1 ]
Bozma, H. Isil [1 ]
机构
[1] Bogazici Univ, Elect Elect Engn Dept, Intelligent Syst Lab, Istanbul, Turkey
关键词
Biologically motivated attentive vision; Foveation; Spatio-temporal pre-processing; Face tracking; Neural networks; Video streaming; FOVEATION; REGION;
D O I
10.1016/j.image.2010.08.002
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper considers the problem of video streaming in low bandwidth networks and presents a complete framework that is inspired by the fovea-periphery distinction of biological vision systems. First, an application specific attention function that serves to find the important small regions in a given frame is constructed a priori using a back-propagation neural network that is optimized combinatorially. Given a specific application, the respective attention function partitions each frame into foveal and periphery regions and then a spatial-temporal pre-processing algorithm encodes the foveal regions with high spatial resolution while the periphery regions are encoded with lower spatial and temporal resolution. Finally, the pre-processed video sequence is streamed using a standard streaming server. As an application, we consider the transmission of human face videos. Our experimental results indicate that even with limited amount of training, the constructed attention function is able to determine the foveal regions which have improved transmission quality while the peripheral regions have an acceptable degradation. Crown Copyright (C) 2010 Published by Elsevier B.V. All rights reserved.
引用
收藏
页码:745 / 760
页数:16
相关论文
共 50 条
  • [1] PRIOR: Deep Reinforced Adaptive Video Streaming with Attention-Based Throughput Prediction
    Yuan, Danfu
    Zhang, Yuanhong
    Zhang, Weizhan
    Liu, Xuncheng
    Du, Haipeng
    Zheng, Qinghua
    [J]. PROCEEDINGS OF THE 32ND WORKSHOP ON NETWORK AND OPERATING SYSTEMS SUPPORT FOR DIGITAL AUDIO AND VIDEO, NOSSDAV 2022, 2022, : 36 - 42
  • [2] Intelligent Video Streaming at Network Edge: An Attention-Based Multiagent Reinforcement Learning Solution
    Tang, Xiangdong
    Chen, Fei
    He, Yunlong
    [J]. FUTURE INTERNET, 2023, 15 (07)
  • [3] LiveAE: Attention-based and Edge-assisted Viewport Prediction for Live 360° Video Streaming
    Pan, Zipeng
    Zhang, Yuan
    Lin, Tao
    Yan, Jinyao
    [J]. PROCEEDINGS OF THE 2023 WORKSHOP ON EMERGING MULTIMEDIA SYSTEMS, EMS 2023, 2023, : 28 - 33
  • [4] Attention-Based Convolutional LSTM for Describing Video
    Liu, Zhongyu
    Chen, Tian
    Ding, Enjie
    Liu, Yafeng
    Yu, Wanli
    [J]. IEEE ACCESS, 2020, 8 : 133713 - 133724
  • [5] Multimodal attention-based transformer for video captioning
    Hemalatha Munusamy
    Chandra Sekhar C
    [J]. Applied Intelligence, 2023, 53 : 23349 - 23368
  • [6] An Attention-based Activity Recognition for Egocentric Video
    Matsuo, Kenji
    Yamada, Kentaro
    Ueno, Satoshi
    Naito, Sei
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2014, : 565 - +
  • [7] Residual Attention-based Fusion for Video Classification
    Pouyanfar, Samira
    Wang, Tianyi
    Chen, Shu-Ching
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 478 - 480
  • [8] Multimodal attention-based transformer for video captioning
    Munusamy, Hemalatha
    Sekhar, C. Chandra
    [J]. APPLIED INTELLIGENCE, 2023, 53 (20) : 23349 - 23368
  • [9] Attention-Based Multimodal Fusion for Video Description
    Hori, Chiori
    Hori, Takaaki
    Lee, Teng-Yok
    Zhang, Ziming
    Harsham, Bret
    Hershey, John R.
    Marks, Tim K.
    Sumi, Kazuhiko
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 4203 - 4212
  • [10] Attention-based video object segmentation algorithm
    Cao, Ying
    Sun, Lijuan
    Han, Chong
    Guo, Jian
    [J]. IET IMAGE PROCESSING, 2021, 15 (08) : 1668 - 1678