Recurrent Partial Kernel Network for Efficient Optical Flow Estimation

被引:0
|
作者
Morimitsu, Henrique [1 ]
Zhu, Xiaobin [1 ]
Ji, Xiangyang [2 ]
Yin, Xu-Cheng [1 ]
机构
[1] Univ Sci & Technol Beijing, Sch Comp & Commun Engn, Beijing, Peoples R China
[2] Tsinghua Univ, Dept Automat, Beijing, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optical flow estimation is a challenging task consisting of predicting per-pixel motion vectors between images. Recent methods have employed larger and more complex models to improve the estimation accuracy. However, this impacts the widespread adoption of optical flow methods and makes it harder to train more general models since the optical flow data is hard to obtain. This paper proposes a small and efficient model for optical flow estimation. We design a new spatial recurrent encoder that extracts discriminative features at a significantly reduced size. Unlike standard recurrent units, we utilize Partial Kernel Convolution (PKConv) layers to produce variable multi-scale features with a single shared block. We also design efficient Separable Large Kernels (SLK) to capture large context information with low computational cost. Experiments on public benchmarks show that we achieve state-of-the-art generalization performance while requiring significantly fewer parameters and memory than competing methods. Our model ranks first in the Spring benchmark without finetuning, improving the results by over 10% while requiring an order of magnitude fewer FLOPs and over four times less memory than the following published method without finetuning. The code is available at github. com/hmorimitsu/ptlflow/tree/main/ptlflow/models/rpknet.
引用
收藏
页码:4278 / 4286
页数:9
相关论文
共 50 条
  • [41] DTEA: Optical Flow Estimation with Deep Taylor Expansion Approximation Network
    Zhu, Zifan
    Huang, Chen
    Wang, Zhicheng
    Xu, Wenduo
    Huang, Zhenghua
    FOURTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING, ICGIP 2022, 2022, 12705
  • [42] Motion estimation based on optical flow and an artificial neural network (ANN)
    Jiafeng Zhang
    Feifei Zhang
    Masanori Ito
    Artificial Life and Robotics, 2009, 14 (4) : 502 - 505
  • [43] Nonparametric kernel estimation of low flow quantiles
    Guo, S.L.
    Kachroo, R.K.
    Mngodo, R.J.
    1996, Elsevier Science B.V., Amsterdam, Netherlands (185) : 1 - 4
  • [44] Nonparametric kernel estimation of low flow quantiles
    Guo, SL
    Kachroo, RK
    Mngodo, RJ
    JOURNAL OF HYDROLOGY, 1996, 185 (1-4) : 335 - 348
  • [45] Spatio-Temporal Recurrent Networks for Event-Based Optical Flow Estimation
    Ding, Ziluo
    Zhao, Rui
    Zhang, Jiyuan
    Gao, Tianxiao
    Xiong, Ruiqin
    Yu, Zhaofei
    Huang, Tiejun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 525 - 533
  • [46] Generalization of deep recurrent optical flow estimation for particle-image velocimetry data
    Lagemann, Christian
    Lagemann, Kai
    Mukherjee, Sach
    Schroder, Wolfgang
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2022, 33 (09)
  • [47] STaRFlow: A SpatioTemporal Recurrent Cell for Lightweight Multi-Frame Optical Flow Estimation
    Godet, Pierre
    Boulch, Alexandre
    Plyer, Aurelien
    Le Besnerais, Guy
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2462 - 2469
  • [48] Partial efficient estimation of SUR models
    Qian, Hailong
    Bednarek, Heather L.
    ECONOMICS BULLETIN, 2015, 35 (01): : 338 - +
  • [49] Recurrent neural network equalization for partial response shaping of magneto-optical readback signals
    Ozgunes, I
    Hacioglu, K
    Kumar, BVKV
    OPTICAL DATA STORAGE '98, 1998, 3401 : 159 - 167
  • [50] Efficient Segmentation-Based PatchMatch for Large Displacement Optical Flow Estimation
    Chen, Jun
    Cai, Zemin
    Lai, Jianhuang
    Xie, Xiaohua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (12) : 3595 - 3607