22email: [email protected], [email protected], [email protected]
Fast Point Cloud Geometry Compression
with Context-based Residual Coding
and INR-based Refinement
Abstract
Compressing a set of unordered points is far more challenging than compressing images/videos of regular sample grids, because of the difficulties in characterizing neighboring relations in an irregular layout of points. Many researchers resort to voxelization to introduce regularity, but this approach suffers from quantization loss. In this research, we use the KNN method to determine the neighborhoods of raw surface points. This gives us a means to determine the spatial context in which the latent features of 3D points are compressed by arithmetic coding. As such, the conditional probability model is adaptive to local geometry, leading to significant rate reduction. Additionally, we propose a dual-layer architecture where a non-learning base layer reconstructs the main structures of the point cloud at low complexity, while a learned refinement layer focuses on preserving fine details. This design leads to reductions in model complexity and coding latency by two orders of magnitude compared to SOTA methods. Moreover, we incorporate an implicit neural representation (INR) into the refinement layer, allowing the decoder to sample points on the underlying surface at arbitrary densities. This work is the first to effectively exploit content-aware local contexts for compressing irregular raw point clouds, achieving high rate-distortion performance, low complexity, and the ability to function as an arbitrary-scale upsampling network simultaneously.
Keywords:
Point cloud geometry compression Implicit neural representation Non-linear transform coding1 Introduction
Point cloud is a common 3D representation in many applications, such as virtual reality, robotics and autonomous driving. Compared with 2D images, 3D point clouds generate much greater volume of data, and thus need to be compressed for efficient storage and transmission. Point cloud compression poses a technical challenge in coding 3D points in unordered and irregular configurations. Although methods employing regular structures like octrees [24, 45, 14, 49] or voxel grids [53, 52, 51, 37, 62] are available, they have relatively low coding efficiency in compressing high-precision coordinates. Other methods directly compress raw point clouds without the voxelization. They typically adopt an autoencoder architecture and entropy code point-wise latent features [56, 25, 59, 55, 60, 21, 26]. Unlike in the entropy coding of image/video, it is difficult to characterize neighboring relations in context-based arithmetic coding. Two straightforward methods are proposed to circumvent the issue: global maximum pooling [56, 25] and downsampling [59, 55, 60, 21, 26]. These methods suffer from three common drawbacks: 1. inability to form local coding contexts due to irregularity of point clouds, reducing coding efficiency; 2. non-scalability with a fixed cardinality of the reconstructed point set, reducing flexibility for downstream tasks; 3. inability to balance the trade-off between rate-distortion performance and computational complexities, reducing practicability in real applications.
To overcome the limitations of the existing methods, we propose a novel method of Context-based Residual Coding and Implicit neural representation (INR) based Refinement, denoted by CRCIR. It consists of two compression layers: a non-learning base layer for main structure reconstruction and a learned refinement layer for recovering finer details. First, the base layer builds a coarse geometry representation with very low complexity. Then in the refinement layer, we use the K-Nearest Neighbors (KNN) to characterize neighboring relations for exploiting correlations between latent features in a locality. The CRCIR method allows us to construct a content-adaptive conditional entropy model that removes statistical redundancies among neighboring features, leading to superior rate-distortion performance. It is worth noting that our method achieves both lower complexity and better rate-distortion performance than the existing methods.
Furthermore, unlike in existing methods the decoder directly transforms latent features into a 3D point set of a fixed cardinality, our decoder allows flexible point sampling on a learned implicit 3D surface at arbitrary densities. This is achieved by integrating INR into the decoder of the refinement layer. In addition, the CRCIR decoder is conditioned on latent features extracted by the CRCIR encoder and maps the coordinates of any 3D point to the offset between it and the underlying surface. Additionally, incorporating INR into the refinement layer eliminates the need to retrain INR on a per-object basis [11, 12, 50, 41], streamlining the compression process. To our knowledge, this paper is the first to combine the advantages of both INR and non-linear transform coding (NTC) in point cloud geometry compression. The above design strategy may be generalized to utilize INR in data compression of other modalities.
The contributions of our paper are summarized as follows:
-
•
We design a novel CRCIR method for point cloud geometry compression. It achieves not only superior rate-distortion performance but also the reductions of model complexity and coding latency. Both reductions are by two orders of magnitude compared to the SOTA method 3QNet [26].
-
•
We propose a novel conditional entropy model for the compression of latent features, being the first to effectively form local contexts of adaptive arithmetic coding, against the irregularity of raw 3D point clouds.
-
•
We incorporate INR into the refinement layer of the CRCIR compression system, allowing the decoder to sample points on the learned implicit surface at arbitray densities.
2 Related Work
2.1 Deep compression of point cloud geometry
Recently, deep learning has significantly influenced the research on point cloud compression. One facet of this research endeavor involves utilizing deep conditional entropy models to losslessly encode the octree-structured point cloud geometry into a compact bitstream [24, 45, 14, 49]. Moreover, parallel to the advancements in deep lossy image compression [1, 2, 35, 8, 18, 63, 19, 28, 64, 32], some researchers have shifted to using a combination of autoencoder architecture and deep entropy models to optimize the rate-distortion trade-off in an end-to-end manner. The techniques can be further categorized into voxel-based [53, 52, 51, 37, 62] and point-based methods [56, 25, 55, 21, 59, 60, 26]. The former needs to voxelize the input before feeding it into stacked 3D convolutional layers, and the voxelization process results in a critical artifact of missing points in the reconstructed point clouds. In contrast, the latter preserves the original density and details through direct processing of raw point clouds.
Point-based methods struggle to efficiently reduce redundancy due to the irregular and unordered nature of raw point clouds. To reduce the bitrate, two straightforward methods are proposed: global maximum pooling [56, 25] and downsampling [59, 55, 60, 21, 26]. Global maximum pooling, although simple to implement, is only suitable for sparse point clouds, as it faces challenges in reconstructing a dense point cloud from a single feature vector. On the other hand, the second strategy is more effective for dense point clouds because a feature vector is only responsible for reconstructing a specific patch of points. Among them, 3QNet [26] utilizes a learned codebook for vector quantization of latent features, while others [59, 60, 21] use scalar quantization.
2.2 Point cloud upsampling
Point cloud upsampling aims to transform a sparse point cloud into a dense and uniformly distributed set of points, similar to how image super-resolution increases image resolution [10, 48, 23, 47, 17, 33, 34]. Previous methods typically expand the number of points by converting expanded features to coordinates or offsets [61, 58, 30, 43, 31], necessitating a model trained for a specific upsampling rate. Recently, magnification-flexible point cloud upsampling methods have become more popular due to the increased flexibility [57, 44, 13, 65, 22]. Among them, INR has been leveraged for its ability to naturally support arbitrary-scale upsampling [13, 65, 22]. However, such methods suffer from computational inefficiency. Specifically, Grad-PU [22] must iteratively refine initialized points, requiring both forward and compute-intensive backward propagation at each iteration to determine the refinement step and direction respectively.

3 Method
In this section, we will introduce the proposed CRCIR point cloud geometry compression system and its two key technical contributions: (1) an INR-based refinement layer for flexible point sampling on a learned implicit 3D surface at arbitrary densities; (2) a novel graph-based, content-adaptive conditional entropy model for the compression of unordered and irregular point clouds;
3.1 Overview
The overall architecture of the proposed CRCIR compression system is presented in Fig. 1. It consists of a low complexity base layer for reconstructing main structures and an INR-based refinement layer for further recovering finer details. Given a dense 3D point cloud to be compressed, the compression workflow involves the following steps:
-
1.
is first downsampled to using the farthest point sampling (FPS) method [42] and then compressed by an efficient traditional coordinate codec Google Draco [15]. The decompressed point cloud is denoted as and then fed into a non-learning upsampling module to obtain a dense yet coarse point cloud as the base layer point cloud.
-
2.
The residuals between and are transformed and quantized to latent features and then compressed using a learned NTC method as the refinement layer. Specifically, to improve the coding efficiency, we learn a conditional entropy model based on the downsampled point cloud for compressing the residuals .
-
3.
In the decoder, the decompressed residuals are predicted based on the compressed features and the base layer point cloud . After that, the base layer point cloud is refined by adding the decompressed residuals , resulting in a finer reconstruction with higher fidelity.
It’s noteworthy that an INR is incorporated into the decoder of the refinement layer. This INR-based refinement layer can take as input the coordinates of arbitrary density and refine the input points by predicting the residuals between the input and the underlying surface represented by a dense point set. That is to say, by adjusting the decoder-end upsampling rate to an arbitrary value (same with or different from the encoder-end upsampling rate ), we can control the density of the reconstructed point cloud to any desired level, not necessarily the same as the original point cloud. This design enables our compression system to function as an arbitrary-scale point cloud reconstruction network.
3.2 Low complexity base layer
In point clouds, most points belong to the smooth regions, also known as low-frequency components. These areas make up the main structures of point cloud and generally are easier to reconstruct. On the other hand, sharp regions, characterized by a minority of points and known as high-frequency components, present challenges in achieving accurate reconstruction. Existing methods for compressing point clouds often rely on complex neural networks to handle all points, including those in both smooth and sharp regions. However, this approach leads to inefficient use of computational resources, especially when reconstructing smooth regions. To address this issue, we design a base layer dedicated to reconstructing main structures with very low complexity. Specifically, the downsampling and compression module in the base layer both adopts the efficient off-the-shelf algorithms FPS [42] and Google Draco [15], respectively. To obtain the base layer point cloud , we design a simple but effective upsampling method to predict a dense point cloud from the downsampled and decompressed point cloud , which is defined as:
(1) |
Here represents the -th point in , the set comprises the nearest neighbors of in , and denotes the associated interpolation weights. The selection of the interpolation weights are provided in the supplementary material.

3.3 INR-based refinement layer
Point cloud reconstruction can be regarded as sampling points on the underlying surface. However, existing point-based compression methods [56, 25, 55, 21, 59, 26] often neglect the modeling of the underlying surface and simply learn a mapping from latent features to the coordinates at a predefined density. The absence of surface modeling prevents these methods from flexibly controlling the density of the reconstructed point cloud.
To address this issue, we propose an INR-based decoder for the refinement layer. This INR-based decoder take the coordinates of any points as input and predicts the residuals between these points and the underlying surface. In this context, the set of points whose residuals are implicitly represent the underlying surface. Sampling points on this learned implicit surface involves three steps: 1. sampling initialized points within the bounding box; 2. predicting residuals for the initialized points; 3. adding the predicted residuals to the initialized points. Specifically, the step 1 is achieved by the base layer and the sampling density is determined by the decoder-end base layer upsampling rate . By setting to a desired value, it becomes possible to sample a set of points on the underlying surface with the intended target cardinality. Besides, can exceed the encoder-end base layer upsampling rate . In this case, a denser point cloud can be reconstructed compared to the original one.
Furthermore, to eliminate the need for retraining INR on a per-object basis in decoder-only architecture, the proposed method pairs the INR-based decoder with an encoder . The decoder employs a similar architecture to the decoder in [40] and the architecture of the encoder is shown in Fig. 2. Instance-specific information is captured by the encoder and encoded into learned latent features, which then condition the shared decoder . This design allows the proposed CRCIR method to streamline the compression process, significantly reducing coding latency.
The overall refinement layer operates by fusing the residuals of the base layer point cloud into point-wise latent features with the encoder and predicting residuals with the decoder based on these features. To permit parallel computation, we partition into several clusters and let the encoder and the decoder work on each cluster independently. A cluster contains points in that is upsampled from the same point in decompressed sparse point cloud and the cluster centroid is set to this corresponding point in . The workflow of the refinement layer involves the following two stages:
-
1.
In the compression phase, the relative positions between points within a cluster and the cluster centroid, along with the corresponding residuals of these points, are concatenated and then fed into the encoder . For each cluster, the input is first mapped to features with two stacked ResNet blocks [20], then aggregated with an attention-based weighting block [6] and finally re-encoded into a latent feature vector. Subsequently, the latent features are quantized to and entropy coded.
-
2.
In the decompression phase, the base layer reconstructs with the upsampling rate that is either aligned with the downsampling rate in the compression phase or flexibly configured with a desired number. Next, the INR-based decoder predicts residuals for each point in to yield a finer reconstruction . To achieve this, for each query point in , its associated latent feature is determined by using the row of corresponding to the cluster centroid. Subsequently, the relative position between the query point and its corresponding cluster centroid, along with the associated latent feature, are fed into the decoder to estimate the residual. By adding the predicted residuals, points in are projected onto the learned implicit surface.
3.4 Graph-based conditional entropy model
Unlike the common 1-D sequences with natural order such as audio or text, the latent features generated by the encoder are arranged in an irregular and unordered layout. The lack of regularity and order poses challenges in forming local contexts, a crucial factor for coding efficiency. To our best knowledge, none of the existing point-based compression methods have effectively exploited local contexts for point cloud coding. The SOTA methods such as DPCC [21] often rely on an unconditional fully factorized entropy model [1], resulting in sub-optimal coding performance.
In contrast to the existing methods, we propose a novel graph-based content-adaptive conditional entropy model that removes statistical redundancies among neighboring features against the irregularity of raw 3D point clouds. Specifically, given the unordered latent features (extracted from residuals) to be compressed, we first use KNN to characterize the neighboring relations of based on the positions provided by the sparse point cloud . As such, latent features can be interpreted as a graph signal supported on a KNN graph with nodes where each node corresponds to a point in . By employing graph structure, the CRCIR method can exploit the statistical correlation between latent features residing in neighboring nodes. This statistical dependency in a locality is captured by introducing a paired hyperencoder and hyperdecoder to learn a hyperprior [2] on the latent representation . Specifically, the probability models of is assumed to be Gaussian distributions whose means and scales are conditioned on local content. The hyperencoder learns hyperprior from latent features and . The hyperdecoder predicts the means and scales of based on the learned hyperprior .
The detailed architectures of the proposed entropy model are as follows:
-
1.
As illustrated in Fig. 3, the hyperencoder progressively downsamples the graph-organized latent features to using stacked three graph coarsening blocks. At the same time, is also downsampled to with the same length of . Each feature of corresponds to a distinct point in , while the features in are aggregated and then quantized to followed by being entropy coded with an unconditional factorized entropy model. Specifically, each row of corresponds to a point in and exhibits strong correlation with a set of neighboring features in .
-
2.
As depicted in Fig. 4, in the hyperdecoder , is re-encoded into and the positional information of is derived by using three deterministic FPS processes to replicate where the downsampling rates match the values used in the compression phase. Next, for each point in , its nearest neighbors in are characterized and the distances are recorded as and features in are gathered. Subsequently, these features are combined using a weighted sum, where the weights are determined by . Finally, the means and scales are predicted from with a ResNet block.


4 Experiments
4.1 Experimental setup
4.1.1 Training details
We implement the CRCIR method using PyTorch [39] and CompressAI [3], training it with three NVIDIA 1080 Ti GPUs. We use the Adam optimizer [29] by setting and . The loss function is
(2) |
where is used to balance the rate-distortion trade-off, and are the rates of latent features and side information respectively. The training process involves two stages. Initially, only the distortion term in Eq. 2 is minimized, with a learning rate of . In the subsequent stage, both the rate and distortion terms are jointly optimized, with a learning rate of . The first training stage comprises 34,410 steps, followed by the second stage of 10,519 steps, each with a batch size of 24.
4.1.2 Baselines
We compare our method with two representative point-based compression methods: DPCC [21] and 3QNet [26]. Besides, two SOTA non-learning-based methods are also included: MPEG G-PCC [16] and Google Draco [15]. Moreover, we place in the supplementrary material a detailed comparison with two sparse-tensor-based methods: PCGCv2 [52] and SparsePCGC [51].
4.1.3 Datasets
Our main experiments are carried out on the ShapeNet dataset [7]. We follow [21] to split training/testing/validation sets and sample points from meshes. The coordinates of all point clouds are normalized to . For each shape, we densely sample its mesh representation at k points to generate the point cloud. To evaluate the generalizability to unseen data, we also use the real-scanned point clouds in the scene-level Redwood indoor dataset [38]. Furthermore, we provide experiments carried out on more diverse datasets [4, 5, 36] in the supplementary material.
4.1.4 Metrics
Following [21], we adopt point-to-point Chamfer distance (CD) and point-to-plane PSNR for geometry distortion and bits per point (bpp) for compression rate. When the ground truth mesh is available, we also use point-to-mesh distance (P2M) to evaluate the proximity to the ground truth surface. The above three distortion metrics are calculated using PyTorch3D [46] and we utilize the same method to calculate PSNR as used in [26]. When calculating these three metrics, we denormalize each point cloud to its original scale.



4.2 Comparison with SOTA
4.2.1 Rate-distortion performance
In Fig. 5, we present the RD-curves for three distortion metrics tested on the ShapeNet dataset. The CRCIR method achieves slight rate-distortion performance gain over 3QNet [26] in terms of the CD and P2M metrics, while we attain an improvement of approximately 2dB in the PSNR metric at the same bitrate (greater than 0.6 bpp). Additionally, the CRCIR method significantly outperforms the other three baseline methods on all three metrics across the entire bitrate spectrum, with particularly remarkable performance gains at low bitrates. The inferiority of MPEG G-PCC [16] and Google Draco [15] at low bitrates mainly arises from their inability to handle raw point clouds directly. These two methods recude bitrate by increasing the quantization step and the resulting quantization error makes them perform poorly. Though DPCC [21] achieves direct processing of raw point clouds like ours, its failure to exploit local contexts for feature coding limits its ability to effectively leverage statistical dependencies among neighboring features. Consequently, it requires a higher bit allocation to accurately model local details, leading to poor performance at low bitrates. In comparison to all competing methods, the CRCIR method employ a graph-based conditional entropy model to compress features in the refinement layer. This entropy model adjusts to local geometry and exploits the correlation among neighboring features, leading to significant bitrate reduction.
4.2.2 Generalizability
In Fig. 6, we present the performance on point clouds tested on the scene-level Redwood indoor dataset [38]. Despite being trained on the ShapeNet dataset comprising man-made CAD objects only, the CRCIR method still demonstrates superiority in terms of rate-distortion performance on the real-scanned living rooms within this dataset. The experimental results highlight the generalizability of the CRCIR method to large-scale real-world data.


4.2.3 Error visualization
For a better perceptual understanding of the reconstruction qualities of the competing point-based methods, we visualize the reconstruction errors in Fig. 7. Here we use the distance of a point to its nearest neighbors in the ground truth as the error metric, which is calculated with Open3D [66]. One can appreciate the overall superb visual quality of the CRCIR method because of its ability to reconstruct point clouds with fewer errors, especially around areas containing finer detail like the bicycle spokes and the head stalk of the guitar.
4.2.4 Complexity and latency
Tab. 1 presents several key metrics related to both the memory and time complexities of the model, including the number of model parameters, memory footprint of executable files or checkpoints, GPU vRAM consumption, and per-object coding latency. The CRCIR method possesses the lowest model parameters and memory footprints. Regarding encoding and decoding runtime, the CRCIR method is only surpassed by Draco [15], but the increase in coding latency is negligible. In comparison to 3QNet [26], which represents the SOTA approach, the CRCIR method achieves significant reductions in the number of model parameters, encoding latency, and decoding latency by factors of , and respectively. Furthermore, the CRCIR method requires less GPU vRAM, making it compatible with most consumer-level GPU cards. These advantages are primarily attributed to the proposed dual-layer architecture. In this design, the non-learning base layer reconstructs main structures with very low complexity, while the refinement layer focuses on recovering details, enabling even small networks to perform effectively.
Methods | Param. | GPU vRAM | Memory | Enc. Time | Dec. Time |
---|---|---|---|---|---|
M | MB | MB | s | s | |
GPCC [16] | - | - | 3.66 | 0.158 | 0.191 |
Draco [15] | - | - | 2.72 | 0.034 | 0.014 |
DPCC [21] | 0.088 | 858 | 0.44 | 1.271 | 0.395 |
3QNet [26] | 21.420 | 5050 | 81.71 | 6.867 | 1.447 |
Ours | 0.056 | 854 | 0.26 | 0.043 | 0.020 |

Methods | DPCC [21]+Grad-PU [22] | 3QNet [26]+Grad-PU [22] | Ours | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Up rates | CD | P2M | Time | CD | P2M | Time | CD | P2M | Time | |||
s | s | s | ||||||||||
27.639 | 15.216 | 2.309 | 26.622 | 19.033 | 2.415 | 10.922 | 5.238 | 0.172 | ||||
24.391 | 13.975 | 4.474 | 23.441 | 18.016 | 4.589 | 7.699 | 4.756 | 0.173 | ||||
22.087 | 13.101 | 7.577 | 21.572 | 17.450 | 7.818 | 6.262 | 4.614 | 0.171 | ||||
20.291 | 12.444 | 12.095 | 20.354 | 17.086 | 12.195 | 5.678 | 4.632 | 0.170 | ||||
18.869 | 11.936 | 17.297 | 19.439 | 16.783 | 17.484 | 5.396 | 4.695 | 0.173 | ||||
17.757 | 11.528 | 24.068 | 18.715 | 16.501 | 24.295 | 5.243 | 4.780 | 0.173 | ||||
16.846 | 11.208 | 31.179 | 18.164 | 16.276 | 31.230 | 5.171 | 4.873 | 0.170 |
4.3 Upsampling after decompression
One intriguing feature of the proposed CRCIR method is its ability to function as an arbitrary-scale upsampling network. To evaluate the upsampling performance, we compare it with two baselines: DPCC [21]+Grad-PU [22] and 3QNet [26]+Grad-PU [22], where Grad-PU is the SOTA arbitrary-scale upsampling method. We conduct comparative evaluations on the PU-GAN dataset [30] using point clouds progressively downsampled from k points to k points. Specifically, each downsampling process removes k points from the dense point cloud. The point cloud with k points is compressed and the decompressed point cloud is fed into the upsampling network, which performs upsampling at rates ranging from to . The remaining point clouds serve as the ground truth for corresponding upsampling rate. Tab. 2 presents the distortion and the sum of decompression and upsampling runtime under different upsampling rates. The CRCIR method notably and consistently surpasses the competing baselines in both distortion and efficiency metrics. Besides, the runtime of our methods remains unchanged despite increasing the upsampling rate, whereas competing baselines suffer from a significant increase in runtime. This discrepancy in upsampling efficiency stems from the technical difference between Grad-PU and ours in the method to refine the initialized point cloud. Grad-PU requires iterative refinement and compute-intensive backward propagation, whereas the CRCIR method circumvent them by directly estimating the residual between the initialized point and the underlying surface. This substantially alleviate the increased workload associated with refining more initialized points.
4.4 Ablation study
In Fig. 8, we present the effectiveness and necessity of exploiting content-aware conditional entropy model, using experimental results from the ShapeNet dataset. RD-curves of DPCC [21] and 3QNet [26] are included for reference. Utilizing the proposed conditional entropy model drastically shifts R-D curves towards lower bitrates and enables our method to surpass 3QNet [26] whose complexity is two orders of magnitude higher. This implies that employing a conditional entropy model rather than a simple unconditional fully factorized entropy model is a key factor in achieving outstanding results.



4.5 Limitation discussion
By incorporating INR, we enable our decoder to sample points on the underlying surface at arbitrary densities. This allows our method to function as both a compression and arbitrary-scale upsampling network for 3D point clouds. However, when dealing with highly sparse input point clouds like LiDAR data, our method remains effective for compression but less so for arbitrary-scale upsampling. This is primarily because of the challenges involved in inferring plausible underlying surface from sparse and unevenly distributed LiDAR point clouds. To our knowledge, no existing INR-based upsampling methods exhibits remarkable capacity in effectively upsampling LiDAR point clouds at arbitrary scales.
5 Conclusion
We introduce a novel CRCIR method for compressing irregular point cloud geometry. It comprises a base layer tailored for reconstructing the main structures at low complexity and a refinement layer for recovering fine details. By employing the coarse geometry from base layer, we form local context in which the latent features in the refinement layer are compressed with a content-adaptive conditional entropy model. This allows the CRCIR method to remove the redundancies among neighboring features, leading to significant bitrate saving. By incorporating INR into the refinement layer, the CRCIR method possesses the respective advantages of NTC and INR in streamlining the compression process and sampling points on the learned implicit surface at arbitrary densities. Experimental results demonstrate the superiority of our method over existing methods, achieving outstanding rate-distortion performance, the remarkable ability to work as an arbitrary-scale upampling network and significant reductions in model complexity and coding latency.
Acknowledgements
This work was supported by the Natural Sciences and Engineering Research Council of Canada.
References
- [1] Ballé, J., Laparra, V., Simoncelli, E.P.: End-to-end optimized image compression. arXiv preprint arXiv:1611.01704 (2016)
- [2] Ballé, J., Minnen, D., Singh, S., Hwang, S.J., Johnston, N.: Variational image compression with a scale hyperprior. arXiv preprint arXiv:1802.01436 (2018)
- [3] Bégaint, J., Racapé, F., Feltman, S., Pushparaja, A.: Compressai: a pytorch library and evaluation platform for end-to-end compression research. arXiv preprint arXiv:2011.03029 (2020)
- [4] Berretti, S., Bimbo, A.D., Pala, P.: Superfaces: A super-resolution model for 3d faces. In: ECCV Workshops (1). pp. 73–82 (2012)
- [5] Bogo, F., Romero, J., Pons-Moll, G., Black, M.J.: Dynamic FAUST: Registering human bodies in motion. In: IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) (Jul 2017)
- [6] Boulch, A., Marlet, R.: Poco: Point convolution for surface reconstruction. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 6302–6314 (June 2022)
- [7] Chang, A.X., Funkhouser, T., Guibas, L., Hanrahan, P., Huang, Q., Li, Z., Savarese, S., Savva, M., Song, S., Su, H., et al.: Shapenet: An information-rich 3d model repository. arXiv preprint arXiv:1512.03012 (2015)
- [8] Cheng, Z., Sun, H., Takeuchi, M., Katto, J.: Learned image compression with discretized gaussian mixture likelihoods and attention modules. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 7939–7948 (2020)
- [9] Choy, C., Gwak, J., Savarese, S.: 4d spatio-temporal convnets: Minkowski convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 3075–3084 (2019)
- [10] Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convolutional networks. IEEE transactions on pattern analysis and machine intelligence 38(2), 295–307 (2015)
- [11] Dupont, E., Goliński, A., Alizadeh, M., Teh, Y.W., Doucet, A.: Coin: Compression with implicit neural representations. arXiv preprint arXiv:2103.03123 (2021)
- [12] Dupont, E., Loya, H., Alizadeh, M., Goliński, A., Teh, Y.W., Doucet, A.: Coin++: Neural compression across modalities. arXiv preprint arXiv:2201.12904 (2022)
- [13] Feng, W., Li, J., Cai, H., Luo, X., Zhang, J.: Neural points: Point cloud representation with neural fields for arbitrary upsampling. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 18633–18642 (2022)
- [14] Fu, C., Li, G., Song, R., Gao, W., Liu, S.: Octattention: Octree-based large-scale contexts model for point cloud compression. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 36, pp. 625–633 (2022)
- [15] Galligan, F., Hemmer, M., Stava, O., Zhang, F., Brettle, J.: Google/draco: a library for compressing and decompressing 3d geometric meshes and point clouds (2018)
- [16] Graziosi, D., Nakagami, O., Kuma, S., Zaghetto, A., Suzuki, T., Tabatabai, A.: An overview of ongoing point cloud compression standardization activities: Video-based (v-pcc) and geometry-based (g-pcc). APSIPA Transactions on Signal and Information Processing 9, e13 (2020)
- [17] Guo, Y., Wu, X., Shu, X.: Data acquisition and preparation for dual-reference deep learning of image super-resolution. IEEE Transactions on Image Processing 31, 4393–4404 (2022)
- [18] Guo, Y., Zhang, X., Wu, X.: Deep multi-modality soft-decoding of very low bit-rate face videos. In: Proceedings of the 28th ACM International Conference on Multimedia. pp. 3947–3955 (2020)
- [19] He, D., Yang, Z., Peng, W., Ma, R., Qin, H., Wang, Y.: Elic: Efficient learned image compression with unevenly grouped space-channel contextual adaptive coding. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5718–5727 (2022)
- [20] He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 770–778 (2016)
- [21] He, Y., Ren, X., Tang, D., Zhang, Y., Xue, X., Fu, Y.: Density-preserving deep point cloud compression. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2333–2342 (2022)
- [22] He, Y., Tang, D., Zhang, Y., Xue, X., Fu, Y.: Grad-pu: Arbitrary-scale point cloud upsampling via gradient descent with learned distance functions. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5354–5363 (2023)
- [23] Hu, X., Mu, H., Zhang, X., Wang, Z., Tan, T., Sun, J.: Meta-sr: A magnification-arbitrary network for super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 1575–1584 (2019)
- [24] Huang, L., Wang, S., Wong, K., Liu, J., Urtasun, R.: Octsqueeze: Octree-structured entropy model for lidar compression. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 1313–1323 (2020)
- [25] Huang, T., Liu, Y.: 3d point cloud geometry compression on deep learning. In: Proceedings of the 27th ACM international conference on multimedia. pp. 890–898 (2019)
- [26] Huang, T., Zhang, J., Chen, J., Ding, Z., Tai, Y., Zhang, Z., Wang, C., Liu, Y.: 3qnet: 3d point cloud geometry quantization compression network. ACM Transactions on Graphics (TOG) 41(6), 1–13 (2022)
- [27] Ioffe, S., Szegedy, C.: Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning. pp. 448–456. pmlr (2015)
- [28] Kim, J.H., Heo, B., Lee, J.S.: Joint global and local hierarchical priors for learned image compression. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5992–6001 (2022)
- [29] Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
- [30] Li, R., Li, X., Fu, C.W., Cohen-Or, D., Heng, P.A.: Pu-gan: a point cloud upsampling adversarial network. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 7203–7212 (2019)
- [31] Li, R., Li, X., Heng, P.A., Fu, C.W.: Point cloud upsampling via disentangled refinement. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 344–353 (2021)
- [32] Liu, J., Sun, H., Katto, J.: Learned image compression with mixed transformer-cnn architectures. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 14388–14397 (2023)
- [33] Luo, F., Wu, X., Guo, Y.: Functional neural networks for parametric image restoration problems. Advances in Neural Information Processing Systems 34, 6762–6775 (2021)
- [34] Luo, F., Wu, X., Guo, Y.: And: Adversarial neural degradation for learning blind image super-resolution. Advances in Neural Information Processing Systems 36 (2024)
- [35] Minnen, D., Ballé, J., Toderici, G.D.: Joint autoregressive and hierarchical priors for learned image compression. Advances in neural information processing systems 31 (2018)
- [36] Pandey, G., McBride, J.R., Eustice, R.M.: Ford campus vision and lidar data set. The International Journal of Robotics Research 30(13), 1543–1552 (2011)
- [37] Pang, J., Lodhi, M.A., Tian, D.: Grasp-net: Geometric residual analysis and synthesis for point cloud compression. In: Proceedings of the 1st International Workshop on Advances in Point Cloud Compression, Processing and Analysis. pp. 11–19 (2022)
- [38] Park, J., Zhou, Q.Y., Koltun, V.: Colored point cloud registration revisited. In: ICCV (2017)
- [39] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems 32 (2019)
- [40] Peng, S., Niemeyer, M., Mescheder, L., Pollefeys, M., Geiger, A.: Convolutional occupancy networks. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part III 16. pp. 523–540. Springer (2020)
- [41] Postels, J., Strümpler, Y., Reichard, K., Van Gool, L., Tombari, F.: 3d compression using neural fields. arXiv preprint arXiv:2311.13009 (2023)
- [42] Qi, C.R., Yi, L., Su, H., Guibas, L.J.: Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in neural information processing systems 30 (2017)
- [43] Qian, G., Abualshour, A., Li, G., Thabet, A., Ghanem, B.: Pu-gcn: Point cloud upsampling using graph convolutional networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 11683–11692 (2021)
- [44] Qian, Y., Hou, J., Kwong, S., He, Y.: Deep magnification-flexible upsampling over 3d point clouds. IEEE Transactions on Image Processing 30, 8354–8367 (2021)
- [45] Que, Z., Lu, G., Xu, D.: Voxelcontext-net: An octree based framework for point cloud compression. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 6042–6051 (2021)
- [46] Ravi, N., Reizenstein, J., Novotny, D., Gordon, T., Lo, W.Y., Johnson, J., Gkioxari, G.: Accelerating 3d deep learning with pytorch3d. arXiv:2007.08501 (2020)
- [47] Saharia, C., Ho, J., Chan, W., Salimans, T., Fleet, D.J., Norouzi, M.: Image super-resolution via iterative refinement. IEEE transactions on pattern analysis and machine intelligence 45(4), 4713–4726 (2022)
- [48] Shi, W., Caballero, J., Huszár, F., Totz, J., Aitken, A.P., Bishop, R., Rueckert, D., Wang, Z.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 1874–1883 (2016)
- [49] Song, R., Fu, C., Liu, S., Li, G.: Efficient hierarchical entropy model for learned point cloud compression. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 14368–14377 (2023)
- [50] Strümpler, Y., Postels, J., Yang, R., Gool, L.V., Tombari, F.: Implicit neural representations for image compression. In: European Conference on Computer Vision. pp. 74–91. Springer (2022)
- [51] Wang, J., Ding, D., Li, Z., Feng, X., Cao, C., Ma, Z.: Sparse tensor-based multiscale representation for point cloud geometry compression. IEEE Transactions on Pattern Analysis and Machine Intelligence (2022)
- [52] Wang, J., Ding, D., Li, Z., Ma, Z.: Multiscale point cloud geometry compression. In: 2021 Data Compression Conference (DCC). pp. 73–82. IEEE (2021)
- [53] Wang, J., Zhu, H., Liu, H., Ma, Z.: Lossy point cloud geometry compression via end-to-end learning. IEEE Transactions on Circuits and Systems for Video Technology 31(12), 4909–4923 (2021)
- [54] Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph cnn for learning on point clouds. ACM Transactions on Graphics (tog) 38(5), 1–12 (2019)
- [55] Wiesmann, L., Milioto, A., Chen, X., Stachniss, C., Behley, J.: Deep compression for dense point cloud maps. IEEE Robotics and Automation Letters 6(2), 2060–2067 (2021)
- [56] Yan, W., Liu, S., Li, T.H., Li, Z., Li, G., et al.: Deep autoencoder-based lossy geometry compression for point clouds. arXiv preprint arXiv:1905.03691 (2019)
- [57] Ye, S., Chen, D., Han, S., Wan, Z., Liao, J.: Meta-pu: An arbitrary-scale upsampling network for point cloud. IEEE transactions on visualization and computer graphics 28(9), 3206–3218 (2021)
- [58] Yifan, W., Wu, S., Huang, H., Cohen-Or, D., Sorkine-Hornung, O.: Patch-based progressive 3d point set upsampling. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5958–5967 (2019)
- [59] You, K., Gao, P.: Patch-based deep autoencoder for point cloud geometry compression. In: ACM Multimedia Asia, pp. 1–7 (2021)
- [60] You, K., Gao, P., Li, Q.: Ipdae: Improved patch-based deep autoencoder for lossy point cloud geometry compression. In: Proceedings of the 1st International Workshop on Advances in Point Cloud Compression, Processing and Analysis. pp. 1–10 (2022)
- [61] Yu, L., Li, X., Fu, C.W., Cohen-Or, D., Heng, P.A.: Pu-net: Point cloud upsampling network. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 2790–2799 (2018)
- [62] Zhang, J., Chen, T., Ding, D., Ma, Z.: Yoga: Yet another geometry-based point cloud compressor. In: Proceedings of the 31st ACM International Conference on Multimedia. pp. 9070–9081 (2023)
- [63] Zhang, X., Wu, X.: Attention-guided image compression by deep reconstruction of compressive sensed saliency skeleton. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 13354–13364 (2021)
- [64] Zhang, X., Wu, X.: Lvqac: Lattice vector quantization coupled with spatially adaptive companding for efficient learned image compression. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 10239–10248 (2023)
- [65] Zhao, W., Liu, X., Zhong, Z., Jiang, J., Gao, W., Li, G., Ji, X.: Self-supervised arbitrary-scale point clouds upsampling via implicit neural representation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1999–2007 (2022)
- [66] Zhou, Q.Y., Park, J., Koltun, V.: Open3D: A modern library for 3D data processing. arXiv:1801.09847 (2018)
Supplementary Material of ’Fast Point Cloud Geometry Compression with Context-based Residual Coding and INR-based Refinement’
Hao Xu Xi Zhang Xiaolin WuCorresponding author.
6 Additional implementation details
The base layer employs Google Draco [15] to compress the downsampled point cloud . For this compression, the quantization parameter is set to 9, and the compression level is set to 10.
The decoder in our compression system has a similar architecture to the decoder in [40]; the differences are in the number and dimension of hidden layers (3 and 64 respectively in our case). The detailed hyperparameters of other network components are tabulated in Tab. 3.
Following 3QNet [26], we utilize a single model to achieve variable bitrate compression by adjusting the size of the downsampled point cloud . This model is trained with the rate-distortion trade-off parameter, , set to . The selected downsampling rate ranges from to , specifically including
Note that these downsampling rates are examples; the actual sizes for can be adjusted according to user needs and are not strictly limited to the values listed. During training, the size of is set to points, corresponding to a downsampling rate of . Experimental results show that the model, trained at this specific downsampling rate, generalizes effectively to various other downsampling rates.
Component | Layer | i | o | k | Normalization | Activation |
---|---|---|---|---|---|---|
Encoder | ResNet block #1 | 6 | 64 | 1 | BN | ReLU |
ResNet block #2 | 64 | 64 | 1 | BN | ReLU | |
Conv (upper) | 64 | 64 | 1 | - | - | |
Conv (lower) | 64 | 1 | 1 | - | - | |
ResNet block #3 | 64 | 8 | 1 | BN | ReLU | |
Hyperencoder | Graph coarsening block #1 | 8 | 32 | 1 | - | ReLU |
Graph coarsening block #2 | 32 | 32 | 1 | - | ReLU | |
Graph coarsening block #3 | 32 | 8 | 1 | - | ReLU | |
Hyperdecoder | ResNet block #1 | 8 | 32 | 1 | - | ReLU |
ResNet block #2 | 32 | 16 | 1 | - | ReLU |
7 The selection of interpolation weights
Our non-learning base layer employs a simple but effective upsampling method to predict a dense point cloud from the downsampled and decompressed point cloud , which is defined as:
(3) |
Here represents the -th point in , the set comprises the nearest neighbors of in , and denotes the associated interpolation weights. When interpolating a point between and its -th nearest neighbor , the objective is to determine the weight that minimizes the offset between this interpolated point and the underlying surface, ideally approaching . The golden-section search method is a fundamental tool for such optimization problems. Inspired by this, we set the interpolation weights using the golden ratio. Specifically, we set to . In this context, this approach can be seen as using the points predicted by a one-step golden section search as the upsampled point cloud. To mitigate significant deviations of interpolated points from , we penalize the weights for those distant neighbors. Hence, for , the weight is set to . Fine-tuning these hyperparameters meticulously is unnecessary, as the interpolation process aims solely to provide a satisfactory initialization for the subsequent learned refinement layer. This refinement layer will then learn how to improve the quality of the upsampled point cloud. We will address learning the optimal interpolation weights in future work.
8 Comparison with sparse-tensor-based methods
The sparse-tensor-based methods such as PCGCv2 [52] and SparsePCGC [51] use a memory-efficient data structure called sparse tensor [9] to store the voxelized geometry. By introducing regularity with such regular structure, the compression system can form local contexts for conditional entropy coding. However, due to quantization, these methods often suffer from severe artifacts, such as missing points in their reconstructions. In Fig. 9, we demonstrate this failure of the sparse-tensor-based approach. In contrast, our method can well preserve the original density information through direct processing of raw point clouds.
In Fig. 10, we compare our method with PCGCv2 and SparsePCGC and showcase the R-D curves evaluated on the ShapeNet [7] dataset. We utilize the official checkpoint of PCGCv2 to evaluate its performance. As the implementation of SparsePCGC is unavailable, we provide the results of PCGCv2 with a 40% bitrate gain under the same distortion as an estimate of the performance of SparsePCGC. While our method trails SparsePCGC in PSNR at low bitrates (below 0.4 bpp), the slope of our curves suggests the potential for achieving higher PSNR scores at higher bitrates. Additionally, our method achieves superb rate-distortion performance in terms of the other two metrics.




9 Comparisons on additional datasets
We evaluate the rate-distortion performance on the Florence Superface [4] dataset and the MPI Dynamic FAUST [5] dataset. The former contains 20 3D faces, while the latter consists of 200 3D human bodies, both of which are semantically different from our training data. For each shape in these two datasets, we densely sample its mesh representation at k points to generate the point cloud. We provide R-D curves tested on them in Fig. 11 and Fig. 12 respectively. Additionally, in Fig. 13, we present the performance evaluated on real-scanned LiDAR point clouds from a subset of the scene-level Ford dataset [36]. Despite our training data only contains man-made CAD models, our method still maintains superiority in compressing those unseen point clouds.
We present additional perceptual results on point cloud geometry compression in Fig. 14, Fig. 15 and Fig. 16. The error colormap demonstrates clearly superior performance of the proposed method to other competing methods. The more complex the 3D geometry, the more advantageous our method.










10 Additional ablation study
In Tab. 4, we showcase the efficacy of the proposed compression system in working as an arbitrary-scale upsampling network. This experiment is conducted using the PU-GAN dataset [30], employing the same experimental setup as outlined in the main text. For comparison, we employ Grad-PU [22] to upsample the sparse point cloud compressed by our method, serving as a competing baseline. Despite our compression system not being trained explicitly for upsampling, its upsampling performance closely rivals that of a dedicated upsampling network, while also achieving a significant reduction in upsampling runtime. This highlights the effectiveness of our compression method in serving as an arbitrary-scale upsampling network.
Methods | Ours (Comp)+Grad-PU [22] | Ours (Comp)+Ours (Up) | ||||||
---|---|---|---|---|---|---|---|---|
Up rates | CD | P2M | Time | CD | P2M | Time | ||
s | s | |||||||
8.894 | 4.592 | 2.053 | 10.922 | 5.238 | 0.172 | |||
6.594 | 4.294 | 4.226 | 7.699 | 4.756 | 0.173 | |||
5.448 | 4.156 | 7.336 | 6.262 | 4.614 | 0.171 | |||
4.861 | 4.096 | 11.877 | 5.678 | 4.632 | 0.170 | |||
4.511 | 4.069 | 17.118 | 5.396 | 4.695 | 0.173 | |||
4.302 | 4.076 | 23.935 | 5.243 | 4.780 | 0.173 | |||
4.192 | 4.109 | 30.894 | 5.171 | 4.873 | 0.170 |
11 Visualizing the reconstruction error of upsampled point clouds
In Section 4.3 of the main text, we discussed the objective performance of upsampling sparse decompressed point clouds. These sparse point clouds were compressed at near bitrates using various compression methods and subsequently upsampled at the upsampling rates ranging from to . Here, we present some perceptual results of upsampling sparse decompressed point clouds in Fig. 17 and Fig. 18. For each presented object, we also provide the bitrate to compress the sparse point cloud, along with the Chamfer distance (CD) between the upsampled point cloud and ground truth. The unit of CD is .

