PointDP: Diffusion-driven Purification against Adversarial Attacks on 3D Point Cloud Recognition
Abstract
3D Point cloud is becoming a critical data representation in many real-world applications like autonomous driving, robotics, and medical imaging. Although the success of deep learning further accelerates the adoption of 3D point cloud in the physical world, deep learning is notorious for its vulnerability to adversarial attacks. In this work, we first identify that the state-of-the-art empirical defense, adversarial training, has a major limitation in applying to 3D point cloud models due to gradient obfuscation. We further propose PointDP, a purification strategy that leverages diffusion models to defend against 3D adversarial attacks. We extensively evaluate PointDP on six representative 3D point cloud architectures, and leverage 10+ strong and adaptive attacks to demonstrate its lower-bound robustness. Our evaluation shows that PointDP achieves significantly better robustness than state-of-the-art purification methods under strong attacks. Results of certified defenses on randomized smoothing combined with PointDP will be included in the near future.
1 Introduction
Point cloud data is emerging as one of the most broadly used representations in 3D computer vision. It is a versatile data format available from various sensors like LiDAR and stereo cameras and computer-aided design (CAD) models, which depicts physical objects by a number of coordinates in the 3D space. Many deep learning-based 3D perception models have been proposed wang2015voting ; maturana2015voxnet ; Riegler2017OctNet ; wang2017cnn ; qi2017pointnet ; choy20194d and thus realized several safety-critical applications (e.g., autonomous driving) yin2021center ; shi2019pointrcnn ; shi2020pv . Although deep learning models qi2017pointnet ; qi2017pointnet++ have exhibited performance boost on many challenging tasks, extensive studies show that they are notoriously vulnerable to adversarial attacks cao2019adversarial ; sun2020lidar ; xiang2019generating , where attackers manipulate the input in an imperceptible manner, which will lead to incorrect predictions of the target model. Because of the broad applications of 3D point clouds in safety-critical fields, it is imperative to study the adversarial robustness of point cloud recognition models.
The manipulation space for 2D adversarial attacks is to change pixel-level numeric values of the input images. Unlike adversarial examples in 2D applications, the flexible representation of 3D point clouds results in an arguably larger attack surface. For example, adversaries could shift and drop existing points zheng2019pointcloud , add new points into the pristine point cloud sun2021adversarially , or even generate new point clouds zhou2020lg to launch attacks. To make attacks less perceptible, different strategies like limiting the number of altered points and constraining the maximal magnitude of shifted points sun2021adversarially . The flexibility of 3D point cloud data formats enables diverse attacks, thus hindering a practical and universal defense design.
Considering the safety criticalness involved in 3D point cloud applications, various studies have been devoted to advancing the robustness of 3D point cloud recognition models. DUP-Net zhou2019dup and GvG-PointNet++ dong2020self pioneered to add statistical outlier removal (SOR) modules as a pre-processing and in-network block, respectively, as mitigation strategies. More lately, Sun et al. sun2020adversarial broke the robustness of DUP-Net and GvG-PointNet++ by specific adaptive attacks. Adversarial training has been acknowledged as the most powerful defense to deliver empirical robustness on PointNet, DGCNN, and PCT sun2021adversarially . Meanwhile, advanced purification strategies like IF-Defense wu2020if and LPC li2022robust leverage more complex module to clean the adversarial point clouds. However, we for the first time demonstrate that standard adversarial training suffers from gradient obfuscation in the point cloud recognition models. We also extensively evaluate IF-Defense and LPC to show that their purification strategies are both vulnerable to stronger attacks (§ 4.3).
In this work, we further propose PointDP, an adversarial purification method that leverages a diffusion model as a cleanser module to defend against 3D adversaries. Lately, diffusion models have been emerging as dominant generative models ho2020denoising ; nichol2021improved ; dhariwal2021diffusion , which extend to the 3D space as well luo2021diffusion . Diffusion models have been proven to be effective in defending against attacks in the 2D space nie2022diffusion . Diffusion models take two steps to (i) diffuse the input data to noise gradually and (ii) reverse the noised data to its origin step by step (§ 2.1). Besides the high quality of generation, diffusion models add randomness in every step of its process, which could help preventing adaptive adversaries from launching attacks. We rigorously evaluate PointDP with six representative point cloud models and sixteen attacks. PointDP on average achieves 75.9% robust accuracy while maintaining similar clean accuracy to the original models.
In a nutshell, our contributions are summarized as two-fold:
-
•
We for the first time demonstrate that standard adversarial training madry2017towards ; sun2021adversarially , the most longstanding defense in the 2D space, has a major limitation of application in 3D point cloud models due to architecture designs. We leverage black-box attacks to demonstrate our claim that drop adversarially trained models’ robust accuracy to 10%.
-
•
We propose PointDP that leverage diffusion models to purify adversarial examples. We conduct extensive and rigorous evaluation on six representative models with numerous attacks to comprehensively understand the robustness of PointDP. Our evaluation shows that PointDP outperforms state-of-the-arts purification methods, IF-Defense wu2020if and LPC li2022robust by 12.6% and 40.3% on average, respectively.
2 PointDP: Diffusion-driven Purification against 3D Adversaries
We first introduce the preliminaries of diffusion models and then propose PointDP that first introduces noise to the adversarial 3D point clouds, followed by the forward process of diffusion models to get diffused point clouds. Purified point clouds are recovered through the reverse process (§-2.2). Next, we follow nie2022diffusion to apply the adjoint method to backward propagate through SDE for efficient gradient evaluation with strong adaptive attacks (§ 2.3).
2.1 Preliminaries
In this section, we briefly review the background of diffusion models in 3D vision tasks. As mentioned in § 1, diffusion models involve the forward and reverse processes.
Given a clean point cloud sampled from the unknown data distribution , the forward process of the diffusion model leverages a fixed Markov chain to gradually add Gaussian noise to the clean point cloud over a pre-defined time steps, resulting in a number of noisy point clouds , , …, . Mathematically, the forward process is defined as
(1) |
where is a scheduling function of the added Gaussian noise (e.g., ).
The reverse process, in contrast, is trained to recover the diffused point cloud in an iterative manner. 3D Point clouds have less semantics than 2D images due to the lack of texture information. Therefore, point cloud diffusion models leverage a separate encoder to as a latent feature as a condition to help recover the clean point cloud.
(2) |
where denotes the approximated mean value parameterized by a neural network, and . The training objective is to learn the variational bound of the negative log-likelihood luo2021diffusion . In practice, we jointly train the encoder with . Essentially, the sampling process is similar to the DDPM model dhariwal2021diffusion :
(3) |
where . Point cloud diffusion models have recently achieved SOTA performance on generating and autoencoding 3D point clouds, which provides us with opportunities for adversarial point cloud purification.

2.2 Design of PointDP
Figure 1 illustrates the pipeline of PointDP. Nie et al. nie2022diffusion have shown that diffusion-driven purification is able to remove the adversarial effect for 2D images. As mentioned in § 2.1, conditional diffusion models were proposed in the 3D point cloud space. Specifically, we use the design in luo2021diffusion as the base model for the purification process in our study. Note that we do not aiming at designing new point cloud diffusion models, but instead propose a novel purification pipeline along with rigorous evaluation as our main contributions.
Let be an adversarial example w.r.t. the pristine classifier , we initialize the input of the forward diffusion process as , i.e., . The forward diffusion process can be solved by Equation 4 from to
(4) |
where , . We leverage Equation 3 to recover the clean point clouds. Equivalently, the reverse can be also solved by the SDE solver in nie2022diffusion , noted as: :
(5) |
where the six inputs are initial value, drift coefficient, diffusion coefficient, Wiener process, initial time, and end time nie2022diffusion .
(6) |
where score function is derived from ho2020denoising .
Besides, the hyper-parameter and trades off the denoising performance and efficiency. We empirically choose and in our study, which has shown satisfactory results in our evaluation (§ 4).
2.3 Adpative Attacks on PointDP
PointDP is a pre-processing module that purifies the adversarial perturbations. athalye2018obfuscated have shown that input transformation-based methods can be broken by specifically designed attacks. Therefore, it is essential to model the adaptive attacks on PointDP to demonstrate its lower-bound adversarial robustness. We thus formulate two types of adaptive attacks on PointDP.
Attack on Latent Feature. As PointDP utilizes conditional diffusion models for adversarial purification, the latent feature is a good candidate for adversaries to launch attacks. Concretely, adversaries can set the goal to maximize some distance metric between the latent feature of the optimized adversarial examples and the oracle latent feature of clean inputs . Without loss of generality, the adaptive attacks can be formulated as:
(7) |
where denotes the adversarial examples from the -th step, is the function to project the adversarial examples to the pre-defined space , and is the attack step size. We choose two distance metrics in our study, where the first one is the KL divergence goldberger2003efficient and the other is the the norm distance. In our evaluation (§ 4), we report the lowest accuracy achieved under attacks with two distance metrics.
Attack Using BPDA. We follow nie2022diffusion to formulate the adaptive attack as an augmented SDE process. We re-state the attack formulation as below. For the SDE in Equation 5, the augmented SDE that computes the gradient of backward propagating through it is given by:
(8) |
where is the gradient of the objective w.r.t. the output of the SDE in Equatrion 5), and
where and denote the -dimensional vectors of all ones and all zeros, respectively. Nie et al. nie2022diffusion have demonstrated that such approximation align well with the true gradient value. Therefore, we leverage this adaptive attack formulation for our evaluation.
3 Related Work
In this section, we review the current progress of deep learning, adversarial attacks, and defenses for 3D point cloud recognition tasks.
3.1 Deep Learning on 3D Point Cloud Recognition
2D computer vision has achieved stellar progress on architectural designs of convolutional neural networks he2016deep , followed by vision transformers dosovitskiy2020image . However, there is currently no consensus on the architecture of 3D perception models since there is no standard data format for 3D perception sun2022benchmarking . As raw data from both 3D scanners and triangular meshes can be efficiently transformed into point clouds, they are becoming the most often utilized data format in 3D perception. 3D networks at the early stage use dense voxel grids for perception wang2015voting ; maturana2015voxnet ; DeepSlidingShapes ; tchapmi_segcloud_3dv17 , which discretize a point cloud to voxel cells for classification, segmentation, and object detection. PointNet pioneered to leverage global pooling help achieve memory-efficient permutation invariance in an end-to-end manner. PointNet++ qi2017pointnet++ and DGCNN wang2019dynamic followed up to add sophisticated local clustering operations to advance the performance. Sparse tensors are the other direction in 3D network designs SubmanifoldSparseConvNet ; choy20194d to use 3D convolutions to improve 3D perception performance. PointCNN and RSCNN reformed the classic pyramid CNN to improve the local feature generation li2018pointcnn ; liu2019relation . PointConv and KPConv designed new convolution operation for point cloud learning wu2019pointconv ; thomas2019kpconv . PointTransformer and PCT advanced self-attention blocks in the 3D space and achieved good performance zhao2021point ; guo2020pct . Various novel local clustering operations xiang2021walk ; ma2022rethinking also show enhancements on the clean performance. In this work, we focus on PointNet, PointNet++, DGCNN, PCT, CurveNet, and PointMLP as our evaluation backbones since they are representative and widely used and achieve state-of-the-art results in point cloud recognition mn40 .
3.2 Adversarial Attacks and Defenses
Adversarial attacks have become the main obstacle that hinder deep learning models from real-world deployments, especially in safety-critical applications eykholt2018robust ; sun2020lidar ; cao2019adversarial ; zhang2021emp ; Zhang_2022_CVPR . There are a lot of adversarial attacks proposed in the 2D space to break the various vision models carlini2017towards ; xiao2018generating ; yang2020patchattack ; xie2017adversarial ; huang2019universal ; huang2020universal ; xiao2018spatially ; sun2021certified . To fill this gap between standard and robust accuracies, many mitigation solutions have been studied and presented to improve the robustness against adversarial attacks yang2019me ; xu2017feature ; bafna2018thwarting ; papernot2016distillation ; meng2017magnet ; zhang2019towards ; xiao2018characterizing ; zhang2020robust ; xiao2019advit and 3D domains dong2020self ; zhou2019dup ; sun2020adversarial . However, most of them including adding randomization liu2019extending ; dhillon2018stochastic ; dong2020self , model distillation papernot2016distillation , adversarial detection meng2017magnet , and input transformation yang2019me ; xu2017feature ; papernot2017extending ; bafna2018thwarting ; zhou2019dup have been compromised by adaptive attacks sun2020adversarial ; tramer2020adaptive ; athalye2018obfuscated . Adversarial training (AT) madry2017towards ; goodfellow2014explaining ; wong2020fast ; shafahi2019adversarial , in contrast, delivered a more longstanding mitigation strategy xie2020smooth ; Xie2020Intriguing ; zhang2019theoretically . However, the robust accuracy achieved by AT is still not satisfactory enough to be used in practice. Most recently, Nie et al. proposed DiffPure nie2022diffusion that leverages diffusion models to defend against adversarial attacks, and following-up studies to extend it to certified defenses carlini2022certified .
Adversarial attacks and defenses also extend to 3D point clouds. Xiang et al. xiao2018generating first demonstrated that point cloud recognition models are vulnerable to adversarial attacks. They also introduced different threat models like point shifting and point adding attacks. Wen et al. wen2019geometry enhanced the loss function in C&W attack to achieve attacks with smaller perturbations and Hamdi et al. hamdi2020advpc presented transferable black-box attacks on point cloud recognition. Zhou et al. zhou2019dup and Donget al. dong2020self proposed to purify the adversarial point clouds by input transformation and adversarial detection. However, these methods have been successfully by sun2020adversarial through adaptive attacks. Moreover, Liu et al. liu2019extending made a preliminary investigation on extending countermeasures in the 2D space to defend against simple attacks like FGSM goodfellow2014explaining on point cloud data. Sun et al. sun2021adversarially conducted a more thorough study on the application of self-supervised learning in adversarial training for 3D point clodu recognition. Besides adversarial training, advanced purification methods IF-Defense wu2020if and LPC li2022robust were proposed to transform the adversarial examples to the clean manifold. In this work, we present PointDP, that utilizes 3D diffusion models to purify adversarial point clouds. We also demonstrate that standard adversarial training suffer from strong black-box attacks and SOTA purification methods (i.e., IF-Defense and LPC) are vulnerable to PGD-styled adversaries (§ 4.3).
4 Experiments and Results
In this section, we first introduce our experimental setups (§ 4.1). We then present the standard robustness evaluation of PointDP(§ 4.2). We next show that how the SOTA adversarial training and adversarial purification methods fail under various strong attacks (§ 4.3). We finally conduct stress test on PointDP to show its actual robustness under various stronger adaptive attacks (§ 4.4).
4.1 Experimental Setups
Datasets and Network Architectures. We conduct all the experiments on the widely used ModelNet40 point cloud classification benchmark wu20153d , consisting of 12,311 CAD models from 40 artificial object categories. We adopt the official split with 9,843 samples for training and 2,468 for testing. We also uniformly sample 1024 points from the surface of each object and normalize them into an edge-length-2 cube, following most of the prior arts qi2017pointnet . As mentioned before, there are various backbones for 3D point cloud recognition in the literature. To demonstrate the universality of PointDP, we select six representative model architectures including PointNet qi2017pointnet , PointNet++ qi2017pointnet++ , DGCNN wang2019dynamic , PCT guo2020pct , CurveNet xiang2021walk , and PointMLP ma2022rethinking . These backbones either have representative designs (e.g., Transformer and MLP) or achieve SOTA performance on the ModelNet40 benchmark.
Adversarial Attacks. As briefly described in § 3.2, adversarial attacks could be roughly categorized into C&W- and PGD-styled attacks. C&W attacks involves the perturbation magnitude into the objective term of the optimization procedure, while PGD attacks set the perturbation magnitude as a firm constraint in the optimization procedure. Moreover, adversarial attacks by norm as the distance metric for the perturbation. Although a number of attacks measure Chamfer and Handoff “distances” in 3D point cloud xiang2019generating , they are not formal distance metrics as they do not satisfy the triangular inequality. Therefore, we still leverage and , following most defense studies in both 2D and 3D vision tasks carlini2017towards ; sun2021adversarially . We also have designed adaptive attacks on our proposed method § 2.3. Besides naive C&W and PGD attacks, we leverage specific attacks designed to break the robustness of point cloud recognition such as NN tsai2020robust and AdvPC hamdi2020advpc . We also apply strong adaptive AutoAttack croce2020reliable (i.e., APGD) in our evaluation. Moreover, we use SPSA uesato2018adversarial and Nattack li2019nattack as black-box adversaries, followed by the suggestion of Carlini et al. carlini2019evaluating . We also leverage EOT-AutoAttack. Point adding (PA) and dropping/detaching (PD) attacks are also evaluated in our study, followed by the setups in sun2021adversarially . We set the attack steps to 200 to maximize the adversarial capability and follow the settings in sun2021adversarially for other attack parameters by default.
Evaluation Metrics. We leverage two main metrics to evaluate the performance of our defense proposal, which are standard and robust accuracy. The standard accuracy measures the performance of the defense method on clean data, which is evaluated on the whole test set from ModelNet40. The robust accuracy measures the performance on adversarial examples generated by different attacks. Because of the high computational cost of applying adaptive and black-box attacks to our method, we evaluate robust accuracy for our defense on a fixed subset of 128 point clouds randomly sampled from the test set. Notably, robust accuracies of most baselines do not change much on the sampled subset, compared to the whole test set. We evaluate the robust accuracy on the whole test set for other adversarial attacks with acceptable overhead (e.g., C&W and PGD attacks).
PointNet | PointNet++ | DGCNN | PCT | CurveNet | PointMLP | |
None | 90.1 | 92.8 | 92.5 | 92.8 | 93.2 | 93.5 |
PA | 44.1 | 19.9 | 35.1 | 20.8 | 48.9 | 7.2 |
PD | 33.3 | 69.8 | 64.5 | 53.0 | 72.6 | 71.1 |
Baseline. Without any defense applied to the original recognition models, the robust accuracy is mostly 0 for all models under and based attacks. DGCNN exceptionally achieves 64% on -based PGD, AutoAttack, respectively, due to its dynamic clustering design, which adaptively discards outlier points. PA and PD are two weaker attacks and Table 1 presents the robust accuracy against these two attacks.
4.2 Experiment Results of PointDP
In this section, we first present the evaluation results of PointDPunder attacks on the plain models. We train the diffusion and 3D point cloud recognition models in a sequential order. Table 2 presents the detailed results of PointDP against attacks on six models. We find that PointDP overall achieves satisfactory results across all models and attacks. The average robust accuracy against adversarial attacks is above 75%. We observe a drop on the clean accuracy for the chosen models, which is expected. As mentioned before, diffusion models for 3D point cloud is a more difficult task than 2D image diffusion, which may lead to partial semantic loss. The average drop of standard accuracy is 4.9%. We find that DGCNN still achieves the best robustness combined with PointDP, which has a 79.9% of robust accuracy. We further compare the performance of PointDPwith adversarial training, IF-Defense, and LPC in the next section.
PointNet | PointNet++ | DGCNN | PCT | CurveNet | PointMLP | ||
None | 86.8 | 87.9 | 86.9 | 87.0 | 88.0 | 88.2 | |
C&W | 77.9 | 78.6 | 78.9 | 76.8 | 73.1 | 76.2 | |
PGD | 78.1 | 80.6 | 80.3 | 77.2 | 74.8 | 79.8 | |
AdvPC | 69.7 | 76.6 | 79.1 | 79.4 | 72.6 | 75.2 | |
PA | 82.1 | 85.1 | 84.8 | 85.5 | 86.3 | 85.8 | |
C&W | 82.4 | 82.9 | 81.9 | 80.9 | 81.5 | 82.6 | |
PGD | 80.1 | 75.0 | 74.6 | 72.0 | 71.7 | 76.4 | |
AdvPC | 69.1 | 76.3 | 79.0 | 74.2 | 74.1 | 75.6 | |
NN | 83.5 | 82.9 | 83.3 | 82.3 | 81.5 | 83.1 | |
PD | 68.9 | 74.1 | 77.3 | 76.3 | 76.8 | 77.4 |
4.3 Failure of State-of-the-Art Defenses

In this section, we demonstrate how lately proposed defense solutions fail when encountered with stronger (adaptive) adversarial attacks on 3D point cloud recognition models.
PointNet | DGCNN | PCT | |
None | 87.8 | 90.6 | 89.7 |
PGD | 52.1 | 67.4 | 51.3 |
AutoAttack | 40.5 | 56.4 | 47.2 |
SPSA | 56.7 | 7.8 | 11.4 |
Nattack | 55.1 | 5.4 | 6.5 |
Adversarial training (AT) has been applied to PointNet, DGCNN, and PCT with the help of self-supervised learning sun2021adversarially that achieves satisfactory robustness. Such observations are consistent with the performance of AT for 2D perception models. However, we find that AT is, in fact, a weak defense solution in 3D perception models. First, as acknowledged by sun2021adversarially , point cloud models (e.g., PointNet++ and CurveNet) often leverage different sampling strategies to select anchor points, like furthest point sampling (FPS). Such sampling involves high randomness. AT either cannot converge with different random seeds in each iteration or overfits to a single random seed. Therefore, AT cannot fit these models. Moreover, we discover that the NN layers will cause severe gradient obfuscation in point cloud models as well. Different from 2D models that are almost fully differentiable, except for the max pooling layer. As shown in Figure 3, NN essentially applies top-. Therefore, gradient backward propagation through NN layers is indexing, which is not non-smooth. The heavy usage of NN layers in DGCNN and PCT will drastically hinder the gradient flow. As mentioned in § 4.1, we exploit black-box SPSA and Nattack to validate our findings. Table 3 presents the results of AT. SPSA and Nattack can greatly lower the average robust accuracy (7.8%) than white-box attacks (55.6%) on DGCNN and PCT, which confirms the effect of gradient obfuscation. PointNet, however, achieves better robustness under black-box attacks because it only has one max pooling layer and does not employ NN layers.
PointNet | PointNet++ | DGCNN | PCT | CurveNet | PointMLP | ||
ONet | None | 90.0 | 92.8 | 92.4 | 92.8 | 93.1 | 93.5 |
PGD | 69.9 | 74.0 | 61.0 | 54.1 | 51.9 | 61.6 | |
AdvPC | 69.4 | 72.8 | 61.6 | 53.9 | 53.6 | 62.5 | |
PGD | 74.2 | 77.5 | 70.5 | 67.2 | 68.7 | 70.5 | |
AdvPC | 69.0 | 72.9 | 63.0 | 64.5 | 55.4 | 67.9 | |
ConvONet | None | 90.1 | 92.8 | 92.5 | 92.8 | 93.2 | 93.5 |
PGD | 66.4 | 73.2 | 52.9 | 46.8 | 45.3 | 55.7 | |
AdvPC | 63.7 | 71.2 | 55.5 | 47.2 | 46.7 | 55.0 | |
PGD | 72.2 | 76.7 | 69.8 | 65.6 | 62.7 | 71.4 | |
AdvPC | 63.4 | 74.3 | 56.6 | 59.8 | 47.2 | 71.0 |
Existing purification-based defenses against 3D adversarial point clouds mainly leverage C&W-styled attacks in their evaluation. C&W attacks utilize the method of Lagrange multipliers to find tractable adversarial examples while minimizing the magnitudes of the perturbation. From the perspective of adversary, such attacks are desirable due to their stealthiness, while this does not hold from a defensive view. Defense methods should be evaluated against strong adaptive attacks carlini2019evaluating . IF-Defense and LPC are the SOTA adversarial purification methods for 3D point cloud models. We leverage PGD and AdvPC attacks, which assign constant adversarial budget in the adversarial optimization stage. We follow the original setups of IF-Defense and LPC in our study. Such evaluation is stronger than C&W attacks, while we note that they are not strict adaptive attacks since the adversarial target is still the classifier itself. Similar to PointDP , IF-Defense can be pre-pended to any point cloud classifier, but LPC uses a specific backbone. Table 4 presents the detailed evaluation results of IF-Defense under various settings and attacks. We find that PointDP achieves much better robustness than IF-Defense, which is on average an 12.6% improvements. However, IF-Defense achieves slightly higher clean accuracy (4.9%). This is because IF-Defense leverages SOR to smooth the point cloud zhou2019dup . However, such an operation has been demonstrated to be vulnerable sun2020adversarial . With specific adaptive attacks, there will be a even larger drop of robust accuracy for IF-Defense.
Figure 2 shows the comparison among PointDP and existing methods. PointDP overall achieves the best performance than prior arts, which are 12.6% and 40.3% improvements than IF-Defense and LPC, respectively. We find that even without adaptive attacks, adversaries with constant budgets can already hurt the robust accuracy by a significant margin. This suggests that IF-Defense and LPC fail to deliver strong robustness to 3D point cloud recognition models. Especially, LPC appears in the proceedings of CVPR 2022, but actually achieves trivial robustness, indicating that a rigorous benchmarking is highly required in this community.
4.4 Defense against Adaptive Threats
We have so far illustrated that state-of-the-art defenses can be easily broken by (adaptive) adversarial attacks and PointDP consistently achieves the best robustness. In this section, we further extensively evaluate the robustness of PointDP on even stronger adaptive attacks to demonstrate the actual robustness realized by PointDP. As mentions in § 4.1, we leverage two types of adaptive attacks in our study, and Table 5 presents their results. We also leverage black-box SPSA and Nattack to validate our results. We find that BPDA-PGD the strongest adaptive attacks, which align well with previous study on 2D diffusion-driven purification nie2022diffusion . Even though with strong adaptive attacks, PointDP still achieves much better robustness. Besides, black-box attacks are much less effective. Although we admit that PointDP still relies on gradient obfuscation, the extremely high randomness will hinder the black-box adversaries finding correct gradients.
PointNet | PointNet++ | DGCNN | PCT | CurveNet | PointMLP | ||
None | 86.8 | 87.9 | 86.9 | 87.0 | 88.0 | 88.2 | |
BPDA-PGD | 77.1 | 78.6 | 79.2 | 76.1 | 73.9 | 77.7 | |
EOT-AutoAttack | 78.0 | 79.9 | 79.1 | 76.5 | 75.9 | 78.9 | |
PGD | 80.8 | 80.7 | 82.9 | 82.5 | 80.8 | 79.9 | |
AdvPC | 69.9 | 76.8 | 79.4 | 79.8 | 72.9 | 75.4 | |
SPSA | 76.6 | 78.9 | 74.9 | 78.5 | 76.4 | 80.9 | |
Nattack | 75.2 | 77.9 | 74.4 | 78.0 | 76.1 | 78.9 | |
PA | 81.7 | 84.7 | 84.1 | 84.5 | 84.8 | 85.2 | |
BPDA-PGD | 78.9 | 73.3 | 73.3 | 71.2 | 70.7 | 75.1 | |
EOT-AutoAttack | 79.6 | 74.4 | 74.2 | 71.3 | 71.3 | 75.9 | |
PGD | 86.1 | 87.5 | 82.5 | 86.3 | 87.7 | 87.8 | |
AdvPC | 69.1 | 76.9 | 79.2 | 74.5 | 74.3 | 76.1 | |
SPSA | 76.1 | 77.0 | 74.4 | 74.5 | 77.0 | 78.9 | |
Nattack | 74.9 | 76.5 | 73.9 | 74.0 | 76.3 | 77.2 | |
PD | 61.3 | 72.1 | 73.5 | 75.9 | 74.1 | 74.4 |
5 Discussion
Adversarial robustness has been well-established in 2D vision tasks, where Carlini et al. carlini2019evaluating and many other researchers have devoted significant efforts to set up a rigorous evaluation protocol. In this study, we also emphasize that this evaluation protocol should be followed in 3D point cloud robustness study. Counter-intuitively, we have demonstrated that standard adversarial training is not a good candidate to deliver robustness against strong black-box adversaries because of gradient obfuscation. We propose PointDP as an adversarial purification strategy to mitigate the robustness loss in the 3D space. We would like to clarify that almost all purification methods (including PointDP) still depend on gradient obfuscation. However, we argue that proper usage of gradient obfuscation could still serve as a good defense, as long as the obfuscation is sophisticated enough. The multi-step purification in diffusion models adds extremely high-level randomness that EOT pmlr-v80-athalye18b and BPDA athalye2018obfuscated attacks are hard to model. Therefore, we believe our extensive evaluation reveals the actual robustness of PointDP.
Broader Impacts and Limitations. Mitigation solutions to adversarial attacks are critical and essential for modern machine learning systems. Given that 3D point cloud is heavily adopted in safety-critical applications, we believe our study is valuable in demonstrating the vulnerabilities of existing SOTA defenses. PointDP also. On the other hand, diffusion models needs multiple steps in the reverse process to recover the point cloud and hinder adaptive attacks, which will incur additional computational overhead. PointDP also limits itself to empirical robustness without theoretical guarantees. Therefore, we cannot exclude possibilities that PointDP could be broken by future stronger attacks. We plan to include certified defense (i.e., randomized smoothing cohen2019certified ) into our framework in the near future.
6 Conclusion
In this paper, we propose PointDP, an adversarial purification method against attacks on 3D point cloud recognition. We have first demonstrated that adversarial training and prior purification methods are actually vulnerable to strong attacks. We further leverage extensive evaluation to validate that PointDP outperforms existing SOTA methods by a significant margin in robust accuracy.
References
- [1] 3D Point Cloud Classification Benchmark on ModelNet40. https://paperswithcode.com/sota/3d-point-cloud-classification-on-modelnet40, 2021.
- [2] A. Athalye, N. Carlini, and D. Wagner. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. In International Conference on Machine Learning, pages 274–283. PMLR, 2018.
- [3] A. Athalye, L. Engstrom, A. Ilyas, and K. Kwok. Synthesizing robust adversarial examples. In J. Dy and A. Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 284–293. PMLR, 10–15 Jul 2018.
- [4] M. Bafna, J. Murtagh, and N. Vyas. Thwarting adversarial examples: An -robustsparse fourier transform. arXiv preprint arXiv:1812.05013, 2018.
- [5] Y. Cao, C. Xiao, B. Cyr, Y. Zhou, W. Park, S. Rampazzi, Q. A. Chen, K. Fu, and Z. M. Mao. Adversarial sensor attack on lidar-based perception in autonomous driving. In Proceedings of the 2019 ACM SIGSAC conference on computer and communications security, pages 2267–2281, 2019.
- [6] N. Carlini, A. Athalye, N. Papernot, W. Brendel, J. Rauber, D. Tsipras, I. Goodfellow, A. Madry, and A. Kurakin. On evaluating adversarial robustness. arXiv preprint arXiv:1902.06705, 2019.
- [7] N. Carlini, F. Tramer, J. Z. Kolter, et al. (certified!!) adversarial robustness for free! arXiv preprint arXiv:2206.10550, 2022.
- [8] N. Carlini and D. Wagner. Towards evaluating the robustness of neural networks. In 2017 ieee symposium on security and privacy (sp), pages 39–57. IEEE, 2017.
- [9] C. Choy, J. Gwak, and S. Savarese. 4d spatio-temporal convnets: Minkowski convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3075–3084, 2019.
- [10] J. Cohen, E. Rosenfeld, and Z. Kolter. Certified adversarial robustness via randomized smoothing. In International Conference on Machine Learning, pages 1310–1320. PMLR, 2019.
- [11] F. Croce and M. Hein. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In International Conference on Machine Learning, pages 2206–2216. PMLR, 2020.
- [12] P. Dhariwal and A. Nichol. Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems, 34:8780–8794, 2021.
- [13] G. S. Dhillon, K. Azizzadenesheli, Z. C. Lipton, J. Bernstein, J. Kossaifi, A. Khanna, and A. Anandkumar. Stochastic activation pruning for robust adversarial defense. arXiv preprint arXiv:1803.01442, 2018.
- [14] X. Dong, D. Chen, H. Zhou, G. Hua, W. Zhang, and N. Yu. Self-robust 3d point recognition via gather-vector guidance. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 11513–11521. IEEE, 2020.
- [15] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- [16] K. Eykholt, I. Evtimov, E. Fernandes, B. Li, A. Rahmati, C. Xiao, A. Prakash, T. Kohno, and D. Song. Robust physical-world attacks on deep learning visual classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1625–1634, 2018.
- [17] J. Goldberger, S. Gordon, H. Greenspan, et al. An efficient image similarity measure based on approximations of kl-divergence between two gaussian mixtures. In ICCV, volume 3, pages 487–493, 2003.
- [18] I. J. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572, 2014.
- [19] B. Graham and L. van der Maaten. Submanifold sparse convolutional networks. arXiv preprint arXiv:1706.01307, 2017.
- [20] M.-H. Guo, J.-X. Cai, Z.-N. Liu, T.-J. Mu, R. R. Martin, and S.-M. Hu. Pct: Point cloud transformer. arXiv preprint arXiv:2012.09688, 2020.
- [21] A. Hamdi, S. Rojas, A. Thabet, and B. Ghanem. Advpc: Transferable adversarial perturbations on 3d point clouds. In European Conference on Computer Vision, pages 241–257. Springer, 2020.
- [22] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- [23] J. Ho, A. Jain, and P. Abbeel. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020.
- [24] L. Huang, C. Gao, Y. Zhou, C. Xie, A. Yuille, C. Zou, and N. Liu. Universal physical camouflage attacks on object detectors, 2019.
- [25] L. Huang, C. Gao, Y. Zhou, C. Xie, A. L. Yuille, C. Zou, and N. Liu. Universal physical camouflage attacks on object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 720–729, 2020.
- [26] K. Li, Z. Zhang, C. Zhong, and G. Wang. Robust structured declarative classifiers for 3d point clouds: Defending adversarial attacks with implicit gradients. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15294–15304, 2022.
- [27] Y. Li, R. Bu, M. Sun, W. Wu, X. Di, and B. Chen. Pointcnn: Convolution on x-transformed points. Advances in neural information processing systems, 31:820–830, 2018.
- [28] Y. Li, L. Li, L. Wang, T. Zhang, and B. Gong. Nattack: Learning the distributions of adversarial examples for an improved black-box attack on deep neural networks. In International Conference on Machine Learning, pages 3866–3876. PMLR, 2019.
- [29] D. Liu, R. Yu, and H. Su. Extending adversarial attacks and defenses to deep 3d point cloud classifiers. In 2019 IEEE International Conference on Image Processing (ICIP), pages 2279–2283. IEEE, 2019.
- [30] Y. Liu, B. Fan, S. Xiang, and C. Pan. Relation-shape convolutional neural network for point cloud analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8895–8904, 2019.
- [31] S. Luo and W. Hu. Diffusion probabilistic models for 3d point cloud generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2837–2845, 2021.
- [32] X. Ma, C. Qin, H. You, H. Ran, and Y. Fu. Rethinking network design and local geometry in point cloud: A simple residual mlp framework. arXiv preprint arXiv:2202.07123, 2022.
- [33] A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu. Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083, 2017.
- [34] D. Maturana and S. Scherer. Voxnet: A 3d convolutional neural network for real-time object recognition. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 922–928. IEEE, 2015.
- [35] D. Meng and H. Chen. Magnet: a two-pronged defense against adversarial examples. In Proceedings of the 2017 ACM SIGSAC conference on computer and communications security, pages 135–147, 2017.
- [36] A. Q. Nichol and P. Dhariwal. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pages 8162–8171. PMLR, 2021.
- [37] W. Nie, B. Guo, Y. Huang, C. Xiao, A. Vahdat, and A. Anandkumar. Diffusion models for adversarial purification. arXiv preprint arXiv:2205.07460, 2022.
- [38] N. Papernot and P. McDaniel. Extending defensive distillation. arXiv preprint arXiv:1705.05264, 2017.
- [39] N. Papernot, P. McDaniel, X. Wu, S. Jha, and A. Swami. Distillation as a defense to adversarial perturbations against deep neural networks. In 2016 IEEE symposium on security and privacy (SP), pages 582–597. IEEE, 2016.
- [40] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- [41] C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 652–660, 2017.
- [42] C. R. Qi, L. Yi, H. Su, and L. J. Guibas. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. arXiv preprint arXiv:1706.02413, 2017.
- [43] G. Riegler, A. O. Ulusoy, and A. Geiger. Octnet: Learning deep 3d representations at high resolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
- [44] A. Shafahi, M. Najibi, A. Ghiasi, Z. Xu, J. Dickerson, C. Studer, L. S. Davis, G. Taylor, and T. Goldstein. Adversarial training for free! arXiv preprint arXiv:1904.12843, 2019.
- [45] S. Shi, C. Guo, L. Jiang, Z. Wang, J. Shi, X. Wang, and H. Li. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10529–10538, 2020.
- [46] S. Shi, X. Wang, and H. Li. Pointrcnn: 3d object proposal generation and detection from point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 770–779, 2019.
- [47] S. Song and J. Xiao. Deep Sliding Shapes for amodal 3D object detection in RGB-D images. In CVPR, 2016.
- [48] J. Sun, Y. Cao, Q. A. Chen, and Z. M. Mao. Towards robust lidar-based perception in autonomous driving: General black-box adversarial sensor attack and countermeasures. In 29th USENIX Security Symposium (USENIX Security 20), pages 877–894. USENIX Association, Aug. 2020.
- [49] J. Sun, Y. Cao, C. B. Choy, Z. Yu, A. Anandkumar, Z. M. Mao, and C. Xiao. Adversarially robust 3d point cloud recognition using self-supervisions. Advances in Neural Information Processing Systems, 34:15498–15512, 2021.
- [50] J. Sun, K. Koenig, Y. Cao, Q. A. Chen, and Z. M. Mao. On the adversarial robustness of 3d point cloud classification, 2020.
- [51] J. Sun, A. Mehra, B. Kailkhura, P.-Y. Chen, D. Hendrycks, J. Hamm, and Z. M. Mao. Certified adversarial defenses meet out-of-distribution corruptions: Benchmarking robustness and simple baselines. arXiv preprint arXiv:2112.00659, 2021.
- [52] J. Sun, Q. Zhang, B. Kailkhura, Z. Yu, C. Xiao, and Z. M. Mao. Benchmarking robustness of 3d point cloud recognition against common corruptions. arXiv preprint arXiv:2201.12296, 2022.
- [53] L. P. Tchapmi, C. B. Choy, I. Armeni, J. Gwak, and S. Savarese. Segcloud: Semantic segmentation of 3d point clouds. In International Conference on 3D Vision (3DV), 2017.
- [54] H. Thomas, C. R. Qi, J.-E. Deschaud, B. Marcotegui, F. Goulette, and L. J. Guibas. Kpconv: Flexible and deformable convolution for point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6411–6420, 2019.
- [55] F. Tramer, N. Carlini, W. Brendel, and A. Madry. On adaptive attacks to adversarial example defenses. arXiv preprint arXiv:2002.08347, 2020.
- [56] T. Tsai, K. Yang, T.-Y. Ho, and Y. Jin. Robust adversarial objects against deep learning models. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 954–962, 2020.
- [57] J. Uesato, B. O’donoghue, P. Kohli, and A. Oord. Adversarial risk and the dangers of evaluating against weak attacks. In International Conference on Machine Learning, pages 5025–5034. PMLR, 2018.
- [58] D. Z. Wang and I. Posner. Voting for voting in online point cloud object detection. In Robotics: Science and Systems, volume 1, pages 10–15607. Rome, Italy, 2015.
- [59] P.-S. Wang, Y. Liu, Y.-X. Guo, C.-Y. Sun, and X. Tong. O-cnn: Octree-based convolutional neural networks for 3d shape analysis. ACM Transactions on Graphics (TOG), 36(4):1–11, 2017.
- [60] Y. Wang, Y. Sun, Z. Liu, S. E. Sarma, M. M. Bronstein, and J. M. Solomon. Dynamic graph cnn for learning on point clouds. Acm Transactions On Graphics (tog), 38(5):1–12, 2019.
- [61] Y. Wen, J. Lin, K. Chen, and K. Jia. Geometry-aware generation of adversarial and cooperative point clouds. 2019.
- [62] E. Wong, L. Rice, and J. Z. Kolter. Fast is better than free: Revisiting adversarial training. arXiv preprint arXiv:2001.03994, 2020.
- [63] W. Wu, Z. Qi, and L. Fuxin. Pointconv: Deep convolutional networks on 3d point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9621–9630, 2019.
- [64] Z. Wu, Y. Duan, H. Wang, Q. Fan, and L. J. Guibas. If-defense: 3d adversarial point cloud defense via implicit function based restoration. arXiv preprint arXiv:2010.05272, 2020.
- [65] Z. Wu, S. Song, A. Khosla, F. Yu, L. Zhang, X. Tang, and J. Xiao. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1912–1920, 2015.
- [66] C. Xiang, C. R. Qi, and B. Li. Generating 3d adversarial point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9136–9144, 2019.
- [67] T. Xiang, C. Zhang, Y. Song, J. Yu, and W. Cai. Walk in the cloud: Learning curves for point clouds shape analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 915–924, 2021.
- [68] C. Xiao, R. Deng, B. Li, T. Lee, B. Edwards, J. Yi, D. Song, M. Liu, and I. Molloy. Advit: Adversarial frames identifier based on temporal consistency in videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3968–3977, 2019.
- [69] C. Xiao, R. Deng, B. Li, F. Yu, M. Liu, and D. Song. Characterizing adversarial examples based on spatial consistency information for semantic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 217–234, 2018.
- [70] C. Xiao, B. Li, J.-Y. Zhu, W. He, M. Liu, and D. Song. Generating adversarial examples with adversarial networks. arXiv preprint arXiv:1801.02610, 2018.
- [71] C. Xiao, J.-Y. Zhu, B. Li, W. He, M. Liu, and D. Song. Spatially transformed adversarial examples. arXiv preprint arXiv:1801.02612, 2018.
- [72] C. Xie, M. Tan, B. Gong, A. Yuille, and Q. V. Le. Smooth adversarial training. arXiv preprint arXiv:2006.14536, 2020.
- [73] C. Xie, J. Wang, Z. Zhang, Y. Zhou, L. Xie, and A. Yuille. Adversarial examples for semantic segmentation and object detection. In International Conference on Computer Vision. IEEE, 2017.
- [74] C. Xie and A. Yuille. Intriguing properties of adversarial training at scale. In International Conference on Learning Representations, 2020.
- [75] W. Xu, D. Evans, and Y. Qi. Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155, 2017.
- [76] C. Yang, A. Kortylewski, C. Xie, Y. Cao, and A. Yuille. Patchattack: A black-box texture-based attack with reinforcement learning. In European Conference on Computer Vision, pages 681–698. Springer, 2020.
- [77] Y. Yang, G. Zhang, D. Katabi, and Z. Xu. Me-net: Towards effective adversarial robustness with matrix estimation. arXiv preprint arXiv:1905.11971, 2019.
- [78] T. Yin, X. Zhou, and P. Krähenbühl. Center-based 3d object detection and tracking. CVPR, 2021.
- [79] H. Zhang, H. Chen, C. Xiao, S. Gowal, R. Stanforth, B. Li, D. Boning, and C.-J. Hsieh. Towards stable and efficient training of verifiably robust neural networks. arXiv preprint arXiv:1906.06316, 2019.
- [80] H. Zhang, H. Chen, C. Xiao, B. Li, D. S. Boning, and C.-J. Hsieh. Robust deep reinforcement learning against adversarial perturbations on observations. 2020.
- [81] H. Zhang, Y. Yu, J. Jiao, E. Xing, L. El Ghaoui, and M. Jordan. Theoretically principled trade-off between robustness and accuracy. In International Conference on Machine Learning, pages 7472–7482. PMLR, 2019.
- [82] Q. Zhang, S. Hu, J. Sun, Q. A. Chen, and Z. M. Mao. On adversarial robustness of trajectory prediction for autonomous vehicles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 15159–15168, June 2022.
- [83] X. Zhang, A. Zhang, J. Sun, X. Zhu, Y. E. Guo, F. Qian, and Z. M. Mao. Emp: Edge-assisted multi-vehicle perception. In Proceedings of the 27th Annual International Conference on Mobile Computing and Networking, pages 545–558, 2021.
- [84] H. Zhao, L. Jiang, J. Jia, P. H. Torr, and V. Koltun. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 16259–16268, 2021.
- [85] T. Zheng, C. Chen, J. Yuan, B. Li, and K. Ren. Pointcloud saliency maps. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1598–1606, 2019.
- [86] H. Zhou, D. Chen, J. Liao, K. Chen, X. Dong, K. Liu, W. Zhang, G. Hua, and N. Yu. Lg-gan: Label guided adversarial network for flexible targeted attack of point cloud based deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10356–10365, 2020.
- [87] H. Zhou, K. Chen, W. Zhang, H. Fang, W. Zhou, and N. Yu. Dup-net: Denoiser and upsampler network for 3d adversarial point clouds defense. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1961–1970, 2019.