An end-to-end dynamic point cloud geometry compression in latent space

被引:1
|
作者
Jiang, Zhaoyi [1 ,4 ]
Wang, Guoliang [1 ,4 ]
Tam, Gary K. L. [2 ,4 ]
Song, Chao [1 ,4 ]
Li, Frederick W. B. [3 ,4 ]
Yang, Bailin [1 ,4 ]
机构
[1] Zhejiang Gongshang Univ, Sch Comp Sci & Technol, Hangzhou 310018, Peoples R China
[2] Swansea Univ, Dept Comp Sci, Skewen, Wales
[3] Univ Durham, Dept Comp Sci, Durham, England
[4] Zhejiang Gongshang Univ, Sch Stat & Math, Hangzhou 310018, Peoples R China
基金
中国国家自然科学基金;
关键词
Dynamic point clouds compression; Geometry encoding; Latent scene flow; Deep entropy model;
D O I
10.1016/j.displa.2023.102528
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Dynamic point clouds are widely used for 3D data representation in various applications such as immersive and mixed reality, robotics and autonomous driving. However, their irregularity and large scale make efficient compression and transmission a challenge. Existing methods require high bitrates to encode point clouds since temporal correlation is not well considered. This paper proposes an end-to-end dynamic point cloud compression network that operates in latent space, resulting in more accurate motion estimation and more effective motion compensation. Specifically, a multi-scale motion estimation network is introduced to obtain accurate motion vectors. Motion information computed at a coarser level is upsampled and warped to the finer level based on cost volume analysis for motion compensation. Additionally, a residual compression network is designed to mitigate the effects of noise and inaccurate predictions by encoding latent residuals, resulting in smaller conditional entropy and better results. The proposed method achieves an average 12.09% and 14.76% (D2) BD-Rate gain over state-of-the-art Deep Dynamic Point Cloud Compression (D-DPCC) in experimental results. Compared to V-PCC, our framework showed an average improvement of 81.29% (D1) and 77.57% (D2).
引用
收藏
页数:11
相关论文
共 50 条
  • [31] DeepPCO: End-to-End Point Cloud Odometry through Deep Parallel Neural Network
    Wang, Wei
    Saputra, Muhamad Risqi U.
    Zhao, Peijun
    Gusmao, Pedro
    Yang, Bo
    Chen, Changhao
    Markham, Andrew
    Trigoni, Niki
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 3248 - 3254
  • [32] End-to-End 3D Object Detection using LiDAR Point Cloud
    Raut, Gaurav
    Patole, Advait
    2024 IEEE 3RD INTERNATIONAL CONFERENCE ON COMPUTING AND MACHINE INTELLIGENCE, ICMI 2024, 2024,
  • [33] End-to-End Optimized ROI Image Compression
    Cai, Chunlei
    Chen, Li
    Zhang, Xiaoyun
    Gao, Zhiyong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3442 - 3457
  • [34] Efficient end-to-end multispectral image compression
    Depoian, Arthur C., II
    Bailey, Colleen P.
    Guturu, Parthasarathy
    BIG DATA VI: LEARNING, ANALYTICS, AND APPLICATIONS, 2024, 13036
  • [35] End-to-End Deep ROI Image Compression
    Akutsu, Hiroaki
    Naruko, Takahiro
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2020, E103D (05): : 1031 - 1038
  • [36] End-to-End Quality Controllable Image Compression
    Wang, Luge
    Mao, Xionghui
    Zhang, Saiping
    Yang, Fuzheng
    2022 PICTURE CODING SYMPOSIUM (PCS), 2022, : 229 - 233
  • [37] An End-to-End Learning Framework for Video Compression
    Lu, Guo
    Zhang, Xiaoyun
    Ouyang, Wanli
    Chen, Li
    Gao, Zhiyong
    Xu, Dong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (10) : 3292 - 3308
  • [38] Comprehensive Review of End-to-End Video Compression
    Shi, Liangfan
    Lu, Huimin
    20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024, 2024, : 43 - 48
  • [39] End-to-End Optimized 360° Image Compression
    Li, Mu
    Li, Jinxing
    Gu, Shuhang
    Wu, Feng
    Zhang, David
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 6267 - 6281
  • [40] AN EFFICIENT END-TO-END IMAGE COMPRESSION TRANSFORMER
    Jeny, Afsana Ahsan
    Junayed, Masum Shah
    Islam, Md Baharul
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1786 - 1790