JOURNAL ARTICLE

BMT-PCGC: Point Cloud Geometry Compression with Bidirectional Mask Transformer Entropy Model

Abstract

Point cloud compression plays a critical role in efficiently handling large 3D datasets, enabling their practical usage. An accurate entropy model is essential for learned codecs to achieve good compression performance. Although autoregressive entropy models can explore dependencies in large contexts, they are inefficient in terms of complexity and may miss information from the inverse direction. To improve coding efficiency while guaranteeing fast decoding, we propose a bidirectional mask Transformer entropy model (BMTEM) for point cloud geometry compression, by leveraging bidirectional self-attention in the mask Transformer. The pre-defined mask schedules facilitate group-wise autoregression, which improves parallel computation for faster inference compared to a fully autoregressive approach. Comparative evaluations against point cloud coding standards G-PCC and V-PCC reveal that BMTEM achieves significant bitrate savings of 53.22% and 47.18% in terms of D1-PSNR, respectively. When compared to other deep learning-based methods like PCGCv2 and ANFPCGC, our approach demonstrates average bitrate reductions of 20.01% and 17.67% in terms of D1-PSNR, respectively.

Keywords:
Point cloud Transformer Entropy (arrow of time) Computer science Geometry Physics Electrical engineering Mathematics Artificial intelligence Engineering Voltage

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
21
Refs
0.09
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Optical measurement and interference techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
3D Shape Modeling and Analysis
Physical Sciences →  Engineering →  Computational Mechanics
Advanced Measurement and Metrology Techniques
Physical Sciences →  Engineering →  Mechanical Engineering
© 2026 ScienceGate Book Chapters — All rights reserved.