Cross-modal Adversarial Reprogramming
Abstract
With the abundance of large-scale deep learning models, it has become possible to repurpose pre-trained networks for new tasks. Recent works on adversarial reprogramming have shown that it is possible to repurpose neural networks for alternate tasks without modifying the network architecture or parameters. However these works only consider original and target tasks within the same data domain. In this work, we broaden the scope of adversarial reprogramming beyond the data modality of the original task. We analyze the feasibility of adversarially repurposing image classification neural networks for Natural Language Processing (NLP) and other sequence classification tasks. We design an efficient adversarial program that maps a sequence of discrete tokens into an image which can be classified to the desired class by an image classification model. We demonstrate that by using highly efficient adversarial programs, we can reprogram image classifiers to achieve competitive performance on a variety of text and sequence classification benchmarks without retraining the network.
1 Introduction
Transfer learning [28] and adversarial reprogramming [4] are two closely related techniques used for repurposing well-trained neural network models for new tasks. Neural networks when trained on a large dataset for a particular task, learn features that can be useful across multiple related tasks. Transfer learning aims at exploiting this learned representation for adapting a pre-trained neural network for an alternate task. Typically, the last few layers of a neural network are modified to map to a new output space, followed by fine-tuning the network parameters on the dataset of the target task. Such techniques are especially useful when there is a limited amount of training data available for the target task.
Adversarial reprogramming shares the same objective as transfer learning with an additional constraint: the network architecture or parameters cannot be modified. Instead, the adversary can only adapt the input and output interfaces of the network to perform the new adversarial task. This more constrained problem setting of adversarial reprogramming poses a security challenge to neural networks. An adversary can potentially re-purpose cloud-hosted machine learning (ML) models for new tasks thereby leading to theft of computational resources. Additionally, the attacker may reprogram models for tasks that violate the code of ethics of the service provider. For example, an adversary can repurpose a cloud-hosted ML API for solving captchas to create spam accounts.
Prior works on adversarial reprogramming [4, 21, 14, 32] have demonstrated success in repurposing Deep Neural Networks (DNNs) for new tasks using computationally inexpensive input and label transformation functions. One interesting finding of [4] is that neural networks can be reprogrammed even if the training data for the new task has no resemblance to the original data. The authors empirically demonstrate this by repurposing ImageNet [1] classifiers on MNIST [16] digits with shuffled pixels showing that transfer learning does not fully explain the success of adversarial reprogramming. These results suggest that neural circuits hold properties that can be useful across multiple tasks which are not necessarily related. Hence neural network reprogramming not only poses a security threat, but also holds the promise of more reusable and efficient ML systems by enabling shared compute of the neural network backbone during inference time.
In existing work on adversarial reprogramming, the target adversarial task has the same data domain as the original task. Recent work has shown that network architectures based on the transformer model can achieve state-of-the-art results on language [34], audio [29] and vision [2] benchmarks suggesting that transformer networks serve as good inductive biases in various domains. Given this commonality between the neural architectures in different domains, an interesting question that arises is whether we can perform cross-modal adversarial reprogramming: For example, Can we repurpose a vision transformer model for a language task?
Cross-modal adversarial reprogramming increases the scope of target tasks for which a neural network can be repurposed. In this work, we develop techniques to adversarially reprogram image classification networks for discrete sequence classification tasks. We propose a simple and computationally inexpensive adversarial program that embeds a sequence of discrete tokens into an image and propose techniques to train this adversarial program subject to a label remapping defined between the labels of the original and new task. We demonstrate that we can reprogram a number of image classification neural networks based on both Convolutional Neural Network (CNN) [15] and Vision Transformer [2] architectures to achieve competitive performance on a number of sequence classification benchmarks. Additionally, we show that it is possible to conceal the adversarial program as a perturbation in a real-world image thereby posing a stronger security threat. The technical contributions of this paper are summarized below:
-
•
We propose Cross-modal Adversarial Reprogramming, a novel approach to repurpose ML models originally trained for image classification to perform sequence classification tasks. To the best of our knowledge, this is the first work that expands adversarial reprogramming beyond the data domain of the original task.
-
•
We demonstrate the feasibility of our method by re-purposing four image classification networks for six different sequence classification benchmarks covering sentiment, topic, and DNA sequence classification. Our results show that a computationally-inexpensive adversarial program can leverage the learned neural circuits of the victim model and outperform word-frequency based classifiers trained from scratch on several tasks studied in our work.
-
•
We demonstrate for the first time the threat imposed by adversarial reprogramming to the transformer model architecture by repurposing the Vision Transformer model for six different sequence classification tasks. The reprogrammed transformer model outperforms alternate architectures on five out of six tasks studied in our work.
2 Background and Related Work

2.1 Adversarial Reprogramming
Neural networks have been shown to be vulnerable to adversarial examples [5, 26, 25, 20, 35, 33, 22, 10, 11, 9] which are slightly perturbed inputs that cause victim models to make a mistake. Adversarial Reprogramming was introduced by [4] as a new form of adversarial threat that allows an adversary to repurpose neural networks to perform new tasks, which are different from the tasks they were originally trained for. The proposed technique trains a single adversarial perturbation that can be added to all inputs in order to re-purpose the target model for an attacker’s chosen task. The adversary achieves this by first defining a hard-coded one-to-one label remapping function that maps the output labels of the adversarial task to the label space of the classifier; and learning a corresponding adversarial reprogramming function that transforms an input from the input space of the new task to the input space of the classifier. The authors demonstrated the feasibility of their attack algorithm by reprogramming ImageNet classification models for classifying MNIST and CIFAR-10 data in a white-box setting, where the attacker has access to the victim model parameters.
While the above attack does not require any changes to the victim model parameters or architecture, the adversarial program proposed [4] is only applicable to tasks where the input space of the the original and adversarial task is continuous. To understand the feasibility of attack in a discrete data domain, [21] proposed methods to repurpose text classification neural networks for alternate tasks, which operate on sequences from a discrete input space. The attack algorithm used a context-based vocabulary remapping method that performs a computationally inexpensive input transformation to reprogram a victim classification model for a new set of sequences. This work was also the first in designing algorithms for training such an input transformation function in both white-box and black-box settings—where the adversary may or may not have access to the victim model’s architecture and parameters. They demonstrated the success of their proposed reprogramming functions by adversarially re-purposing various text-classification models including Long Short Term Memory networks (LSTM) [8], bi-directional LSTMs [6] and CNNs [36] for alternate text classification tasks.
Recent works [14, 32] have argued that reprogramming techniques can be viewed as an efficient training method and can be a superior alternative to transfer learning. Particularly [32] argue that one of the major limitations of current transfer learning techniques is the requirement of large amounts of target domain data, which is needed to fine-tune pre-trained neural networks. They demonstrated the advantage of instead using reprogramming techniques to repurpose existing ML models for alternate tasks, which can be done even when training data is scarce. The authors designed a black-box adversarial reprogramming method, that can be trained iteratively from input-output model responses, and demonstrated its success in repurposing ImageNet models for medical imaging tasks such as classification of autism spectrum disorders, melanoma detection, etc.
All of these existing reprogramming techniques are only able to reprogram ML models when the data domain of the target adversarial task and the original task are the same. We address this limitation in our work by designing adversarial input transformation functions that allow image classification models to be reprogrammed for sequence classification tasks such as natural language and protein sequence classification.
2.2 Transformers and Image Classifiers
While Convolutional Neural Networks (CNNs) have long achieved state-of-the-art performance on vision benchmarks, the recently proposed Vision Transformers (ViTs) [2] have been shown to outperform CNNs on several image classification tasks. Transformers [34] are known for achieving state-of-the-art performance in natural language processing (NLP). In order to train transformers for image classification tasks, the authors [2] divided an image into patches and provide the sequence of linear embeddings of these patches as an input to a transformer. Image patches are treated the same way as tokens (words) in an NLP application and the model is trained on image classification in a supervised manner. The authors report that when ViTs are trained on large-scale image datasets, they are competitive and also outperform state-of-the-art models on multiple image recognition benchmarks.
Since transformers can model both language and vision data in a similar manner, that is, as a sequence of embeddings, we are curious to investigate whether a vision transformer can be reprogrammed for a text classification task. In the process, we find that CNN network architectures can also be reprogrammed to achieve competitive performance on discrete sequence classification tasks. In the next section, we discuss our cross-modal adversarial reprogramming approach.
3 Methodology
3.1 Problem Definition
Consider a victim image classifier trained for mapping images to a label . That is,
An adversary wishes to repurpose this victim image classifier for an alternate text classification task of mapping sequences to a label . That is,
To achieve this goal, the adversary needs to learn appropriate mapping functions between the input and output spaces of the original and the new task. We solve this by first defining a label remapping that maps label spaces of the two tasks: ; and then learning a corresponding adversarial program that maps a sequence to an image i.e., such that acts as the target classifier .
We assume a white-box adversarial reprogramming setting where the adversary has complete knowledge about architecture and model parameters of the victim image classifier. In the next few sections we describe the adversarial program , the label remapping function and the training procedure to learn the adversarial program.
3.2 Adversarial Program
The goal of our adversarial program is to map a sequence of discrete tokens to an image . Without loss of generalizability, we assume to be the scaled input space of the image classifier where are the height and width of the input image and is the number of channels. The tokens in the sequence belong to some vocabulary list . We can represent the sequence as where is the vocabulary index of the token in sequence in the vocabulary list .
When designing the adversarial program it is important to consider the computational cost of the reprogramming function . This is because if a classification model that performs equally well can be trained from scratch for the classification task and is computationally cheaper than the reprogramming function, it would defeat the purpose of adversarial reprogramming.
Keeping the above in mind, we design a reprogramming function that looks up embeddings of the tokens and arranges them as contiguous patches of size in an image that is fed as input to the classifier . Mathematically, the reprogramming function is parameterized by a learnable embedding tensor and performs the transformation as per Algorithm 1.
The patch size and image dimensions determine the maximum length of the sequence that can be encoded into the image. We pad all the input sequences all the way up to the maximum allowed sequence length with a padding token to fill up the reprogrammed image and clip any sequences longer than the maximum allowed length from the end. More details about the hyper-parameters can be found in our experiments section.
Concealing the adversarial perturbation: Most past works on adversarial reprogramming have considered an unconstrained attack setting, where the reprogrammed image does not necessarily need to resemble a real-world image. However, as noted by [4], it is possible to conceal the reprogrammed image in a real-world image by constraining the output of the reprogramming function. We can conceal the reprogrammed image as an additive perturbation to some real-world base image by defining an alternate reprogramming function as follows:
(1) |
Since the output of the original reprogramming function is bounded between , we can control the norm of the added perturbation using the parameter .
Computational Complexity: As depicted in Figure 1, during inference, the adversarial program only looks up embeddings of the tokens in the sequence and arranges them in an image tensor which can optionally be added onto a base image. Asymptotically, the time complexity of this adversarial program is linear in terms of the length of the sequence . Since there are no matrix-vector multiplications involved in the adversarial program, it is computationally equivalent to just the embedding layer of a sequence-based neural classifier. Therefore the inference cost of the adversarial program is significantly less than that of a sequence-based neural classifier. Table 1 in our supplementary material compares the wall-clock inference time for a sequence of length for our adversarial program and various sequence-based neural classifiers used in our experiments.
3.3 Label Remapping and Optimization Objective
Past works [4, 21, 32] on adversarial reprogramming assume that the number of labels in the target task are less than than the number of labels in the original task. In our work, we relax this constraint and propose label remapping functions for both of the following scenarios:
1. Target task has fewer labels than the original task: Initial works on adversarial reprogramming defined a one-to-one mapping between the labels of the original and new task [4, 21]. However, recent work [32] found that mapping multiple source labels to one target label helps improve the performance over one-to-one mapping. Our preliminary experiments on cross-modal reprogramming confirm this finding, however, we differ in the way the final score of a target label is aggregated—[32] obtained the final score for a target label as the mean of the scores of the mapped original labels. We found that aggregating the score by taking the maximum rather than the mean over the mapped original labels leads to faster training. Another advantage of using max reduction is that during inference, we can directly map the original predicted label to our target label without requiring access to probability scores of any other label.
Consider a target task label , mapped to a subset of labels of the original task under the many-to-one label remapping function . We obtain the score for this target task label as the maximum of the scores of each label by classifier . That is,
(2) |
where and represent the score (before softmax) assigned to some label by classifier and respectively.
To define the label remapping , instead of randomly assigning source labels to a target label, we first obtain the model predictions on the base image (or a zero image in the case of an unbounded attack) and sort the labels by the obtained scores; We then assign the the highest scored source labels to each target label using a round-robin strategy until we have assigned source labels to each target label.
Note that while we need access to individual class scores during training (where we assume a white-box attack setting), during inference we can simply map the highest predicted label to the target label using the label remapping function without having to know the actual scores assigned to different labels.
2. Original task has fewer labels than the target task: In this scenario, we map the probability distribution over the original labels to a distribution over target labels to class scores for the target label space using a learnable linear transformation. That is,
(3) |
Here is a vector representing class scores (logits) for the target label space. are the learnable parameters of the linear transformation that are optimized along with the parameters of the reprogramming function . Note that unlike the previous scenario, in this setting, we assume that we have access to the probability scores of the original labels during both training and inference.
Optimization Objective: To train the parameters of our adversarial program, we use a cross-entropy loss between the target label and the model score predictions obtained as per Equation 2 or Equation 3. We also incorporate an regularization loss for better generalization on the test set and to encourage more imperceptible perturbation in the case of our bounded attack. Therefore our final optimization objective is the following:
Here is the regularization hyper-parameter and is the predicted class probability of the correct label for sequence . We use mini-batch gradient descent using an Adam optimizer [13] to solve the above optimization problem on the dataset of the target task.
4 Experiments
4.1 Victim Image Classifiers
To demonstrate cross-modal adversarial reprogramming, we perform experiments on four neural architectures trained on the ImageNet dataset. We choose both CNNs and the recently proposed Vision Transformers (ViT) [2] as our victim image classifiers. While CNNs have long achieved state-of-the-art performance on computer-vision benchmarks, the recently proposed ViTs have been shown to outperform CNNs on several image classification tasks. We choose the ViT-Base [2], ResNet-50 [7], InceptionNet-V3 [30] and EfficientNet-B7 [31] architectures. The details of these architectures are listed in Table 1. We perform experiments on both pre-trained and randomly initialized networks.
Accuracy (%) | |||||
---|---|---|---|---|---|
Model | Abbr. | Type | # Params | Top-1 | Top-5 |
ViT-Base | ViT | Transformer | 86.9M | 84.2 | 97.2 |
ResNet-50 | RN-50 | CNN | 25.6M | 79.0 | 94.4 |
InceptionNet-V3 | IN-V3 | CNN | 23.8M | 77.5 | 93.5 |
EfficientNet-B4 | EN-B4 | CNN | 19.3M | 83.0 | 96.3 |
Dataset Statistics | Accuracy (%) | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Avg | Neural Methods | TF-IDF | ||||||||
Dataset | Task Type | # Classes | # Train | # Test | Token | Length | Bi-LSTM | 1D-CNN | unigram | n-gram |
Yelp | Sentiment | 2 | 560,000 | 38,000 | word | 135.6 | 95.94 | 95.18 | 92.50 | 92.93 |
IMDB | Sentiment | 2 | 25,000 | 25,000 | word | 246.8 | 89.43 | 90.02 | 88.52 | 88.43 |
AG | Topic | 4 | 120,000 | 7,600 | word | 57.0 | 91.45 | 92.09 | 90.92 | 90.69 |
DBPedia | Topic | 14 | 560,000 | 70,000 | word | 47.1 | 97.78 | 98.09 | 97.12 | 97.16 |
Splice | DNA | 3 | 2,700 | 490 | neucleobase | 60.0 | 93.26 | 83.87 | 51.42 | 72.24 |
H3 | DNA | 2 | 13,468 | 1,497 | neucleobase | 500.0 | 86.84 | 85.43 | 75.68 | 78.89 |
4.2 Datasets and Reprogramming Tasks
In this work, we repurpose the aforementioned image classifiers for several discrete sequence classification tasks. We wish to analyze the performance of cross-modal adversarial reprogramming for different applications such as understanding language and analyzing sequential biomedical data. Biomedical datasets e.g. splice-junction detection in genes, often have fewer training samples than language based datasets and we aim to understand whether such limitations can adversely affect our proposed reprogramming technique.
Sentiment analysis and topic classification are popular NLP tasks. However, analyzing the underlying semantics of the sequence is often not necessary for solving these tasks since word-frequency based statistics can serve as strong discriminatory features. In contrast, tasks like DNA-sequence classification requires analyzing the sequential semantics of the input and simple frequency analysis of the unigrams or n-grams does not achieve competitive performance on these tasks. To evaluate the effectiveness of adversarial reprogramming in both of these scenarios, we consider the following tasks and datasets in our experiments:
4.2.1 Sentiment Classification
1. Yelp Polarity Dataset (Yelp) [36]: This is a dataset consisting of reviews from Yelp for the task of sentiment classification, categorized into binary classes of positive and negative sentiment.
2. Large Movie Review Dataset (IMDB) [18]: This is a dataset for binary sentiment classification of positive and negative sentiment from highly polar IMDB movie reviews.
4.2.2 Topic Classification
1. AG’s News Dataset (AG) [36]: is a collection of more than 1 million news articles. News articles have been gathered from more than 2000 news sources and contains 4 classes: World, Sports, Business, Sci/Tech.
2. DBPedia Ontology Dataset (DBPedia) [36]: consists of 14 non-overlapping categories from DBpedia 2014. The samples consist of the category and abstract of each Wikipedia article.
Unbounded | Bounded () | ||||||||||||
Pre-trained | Randomly Initialized | Pre-Trained | |||||||||||
Task | ViT | RN-50 | IN-V3 | EN-B4 | ViT | RN-50 | IN-V3 | EN-B4 | ViT | RN-50 | IN-V3 | EN-B4 | |
Yelp | 16 | 92.82 | 93.29 | 89.19 | 93.47 | 92.73 | 68.50 | 65.56 | 52.97 | 88.57 | 81.32 | 81.33 | 81.23 |
IMDB | 16 | 86.76 | 85.60 | 80.67 | 87.26 | 88.38 | 81.08 | 52.87 | 50.26 | 82.07 | 72.28 | 71.22 | 81.42 |
AG | 16 | 91.59 | 89.88 | 89.78 | 90.46 | 91.45 | 82.37 | 50.43 | 24.87 | 86.49 | 83.26 | 78.93 | 84.03 |
DBPedia | 32 | 97.62 | 96.31 | 95.70 | 96.77 | 97.56 | 30.12 | 52.87 | 19.61 | 92.79 | 80.64 | 81.46 | 79.53 |
Splice | 48 | 95.31 | 94.48 | 95.10 | 92.04 | 54.13 | 48.57 | 91.22 | 50.20 | 95.10 | 94.27 | 94.89 | 91.55 |
H3 | 16 | 82.57 | 78.16 | 80.29 | 80.16 | 77.02 | 73.00 | 64.20 | 51.17 | 76.62 | 72.01 | 75.55 | 75.42 |
4.2.3 DNA Sequence Classification
1. Splice-junction Gene Sequences (Splice): This dataset [24, 3] was curated for training ML models to detect splice junctions in DNA sequences. In DNA, there are two kinds of splice junction regions: Exon-Intron (EI) junction and Intron-Exon (IE) junction. This dataset contains sample DNA sequences of 60 base pair length categorized into 3 classes: “EI” which contains exon-intron junction, “IE” which contains intron-exon junction, and “N” which contain neither EI or IE regions.
2. Histone Protein Occupancy in DNA (H3): This dataset from [27, 23] indicates whether certain DNA sequences wrap around H3 histone proteins. Each sample is a sequence with a length of 500 neucleobases. Positive samples contain DNA regions wrapping around histone proteins while negative samples do not contain such DNA regions.
The statistics of these datasets are included in Table 2. To benchmark the performance that can be achieved on these tasks, we train various classifiers from scratch on the datasets for each task. We consider both neural network based classification models and frequency-based statistical models (such as TF-IDF) as our benchmarks. We use word-level tokens for sentiment and topic classification tasks and neucleobase level tokens for DNA sequence classification tasks.
The TF-IDF methods can work on either unigrams or n-grams for creating the feature vectors from the input data. For the n-gram model, we consider n-grams up to length 3 and choose the value of that achieves the highest classification accuracy on the hold-out set. We train a Stochastic Gradient Descent (SGD) classifier to classify the feature vector as one of the target classes. Additionally, we train DNN based text-classifiers: Bidirectional Long Short Term Memory networks (Bi-LSTM) [6, 8] and 1D CNN [12] models from scratch on the above tasks. We use randomly initialized token embeddings for all classification models, which are trained along with the network parameters. For Bi-LSTMs, we combine the outputs of the first and last time step for prediction. For the Convolutional Neural Network we follow the same architecture as [12]. The hyper-parameter details of these classifiers and architecture have been included in Table 2 of the supplementary material.
We report the accuracies on the test set of the above mentioned classifiers in Table 2. We find that while both neural and frequency based TF-IDF methods work well on sentiment and topic classification tasks, neural networks significantly outperform frequency based methods on DNA sequence classification tasks. This is presumably because the latter require structural analysis of the sequence rather than relying on keywords.
4.3 Experimental Details
Input image size and patch size: The ViT-Base model utilized in our work is trained on images of size and works on image patches of size . For all our experiments, we fix the input image size to be . When we use a patch of size for encoding a single token in our sequence, it allows for a maximum of tokens to be encoded into a single image. In our initial experiments we found that using larger patch sizes for smaller sequences leads to higher performance on the target task, since it encodes a sequence in a spatially larger area of the image. Therefore, we choose our patch size as the largest possible multiple of that can encode the longest sequence in our target task dataset. We list the patch size used for different tasks in Table 3.
Training hyper-parameters: We train each adversarial program on a single Titan 1080i GPU using a batch size of . We set the learning rate as for the unbounded attacks and for our bounded attacks (Equation 1). We set the regularization hyper-parameter for all our experiments and train the adversarial program for a maximum 100k mini-batch iterations in the unbouned attack setting and for 200k mini-batch iterations in the bounded attack setting. We map 10 original labels to each target label in the scenario when there are fewer labels for the target task than for the original task. We point the readers to our codebase for precise implementation.111https://github.com/paarthneekhara/multimodal_rerprogramming
5 Results
5.1 Pre-trained vs untrained victim models
Experimental results of our proposed cross-modal reprogramming method are reported in Table 3. In these experiments, the original task has more labels than the target task so we use the label remapping function given by Equation 2. We first consider the unbounded attack setting, where the output of the adversarial program does not need to be concealed in a real-world image. For these experiments, we use the reprogramming function described in Algorithm 1. We also note that the primary evaluation of past reprogramming works [4, 21, 32] is done in an unbounded attack setting.
When attacking pre-trained image classifiers, we achieve competitive performance (as compared to benchmark classifiers trained from scratch, reported in Table 2) across several tasks for all victim image classification models. To assess the importance of pre-training the victim model on the original dataset, we also experiment with reprogramming untrained randomly initialized networks.
Randomly initialized neural networks can potentially have rich structure which the reprogramming functions can exploit. Prior works [19, 17] have shown that wide neural networks can behave as Gaussian processes, where training specific weights in the intermediate layers is not necessary to perform many different tasks. However, in our experiments, we find that for CNN-based image classifiers, reprogramming pre-trained neural networks performs significantly better than reprogramming randomly initialized networks for all tasks. This is consistent with the findings of prior reprogramming work [4] which reports that adversarial reprogramming in the image domain is more effective when it targets pre-trained CNNs. For the ViT model, we find that we are able to obtain competitive performance on sentiment and topic classification tasks when reprogramming either randomly initialized or pre-trained models. Particularly, we find that reprogramming untrained vision transformers provides the highest accuracy on the IMDB classification task. However, for DNA sequence classification tasks (Splice and H3) that require structural analysis of the sequence rather than token-frequency statistics, we find that reprogramming pre-trained vision transformer model performs significantly better than a randomly initialized transformer model.
The ViT model outperforms other architectures on 5 out of 6 tasks in the unbounded attack setting. In particular, for the task of splice-junction detection in gene sequences, reprogramming a pre-trained ViT model outperforms both TF-IDF and neural classifiers trained from scratch. For sentiment analysis and topic classification tasks, which primarily require keyword detection, some reprogramming methods achieve competitive performance as the benchmark methods reported in Table 2.
Additionally, to assess the importance of the victim classifier for solving the target task, we study the extent to which the task can be solved without the victim classifier and using only the adversarial reprogramming function with a linear classification head. We present the results and details of this experiment in Table 3 of our supplementary material.

Concealing the adversarial perturbation: To conceal the output of the adversarial program in a real-world image, we follow the adversarial reprogramming function defined in Equation 1. We randomly select an image from the ImageNet dataset (shown in Figure 2) as the base image and train adversarial programs targeting different image classifiers for the same base image. We present the results at (on a 0 to 1 pixel value scale) distortion between the reprogrammed image and the base image on the right side of Table 3. It can be seen that for some drop in performance, it is possible to perform adversarial reprogramming such that the input sequence is concealed in a real-world image. Figure 1 in our supplementary material shows the accuracy on three target tasks for different magnitudes of allowed perturbation, while reprogramming a pre-trained ViT model.
5.2 Target task has more labels than original task
In a practical attack scenario, the adversary may only have access to a victim image classifier with fewer labels than the target task labels. To evaluate adversarial reprogramming in this scenario, we constrain the adversary’s access to the class-probability scores of just labels of the ImageNet classifier. We choose the most frequent ImageNet labels as the original labels, that can be accessed by the adversary; and perform our experiments on two tasks from our datasets, which have the highest number of labels—AG News (4 labels) and DBPedia (14 labels). We use the label remapping function given by Equation 3, and learn a linear transformation to map the predicted probability distribution over the original labels to the target task label scores.
We demonstrate that we are able to perform adversarial reprogramming even in this more constrained setting. We achieve similar performance as compared to our many-to-one label remapping scenario reported in Table 3 when is close to the number of labels in the target task. This is because we learn an additional mapping function for the output interface, which can potentially lead to better optimization. However as a downside, this setting requires access to all class probability scores for predicting the adversarial label, while in the previous many-to-one label remapping scenario, we only need to know the highest-scored original label for mapping it to one of the adversarial labels.
Accuracy (%) | ||||||
---|---|---|---|---|---|---|
Dataset | # Labels | ViT | RN-50 | IN-V3 | EN-B4 | |
AG | 4 | 3 | 89.42 | 87.18 | 86.66 | 89.18 |
DBPedia | 14 | 3 | 96.34 | 83.16 | 84.17 | 92.95 |
DBPedia | 14 | 10 | 98.01 | 96.84 | 94.88 | 97.16 |
6 Conclusion
We propose Cross-modal Adversarial Reprogramming, which for the first time demonstrates the possibility of repurposing pre-trained image classification models for sequence classification tasks. We demonstrate that computationally inexpensive adversarial programs can repurpose neural circuits to non-trivially solve tasks that require structural analysis of sequences. Our results suggest the potential of training more flexible neural models that can be reprogrammed for tasks across different data modalities and data structures. More importantly, this work reveals a broader security threat to public ML APIs that warrants the need for rethinking existing security primitives.
7 Acknowledgements
This work was supported by ARO under award number W911NF1910317, SRC under Task ID: 2899.001 and DoD UCR W911NF2020267 (MCA S-001364).
References
- [1] J. Deng, W. Dong, R. Socher, L. Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
- [2] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021.
- [3] Dheeru Dua and Casey Graff. UCI machine learning repository, 2017.
- [4] Gamaleldin F Elsayed, Ian Goodfellow, and Jascha Sohl-Dickstein. Adversarial reprogramming of neural networks. In ICLR, 2019.
- [5] Ian Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. In ICLR, 2015.
- [6] Alex Graves, Santiago Fernández, and Jürgen Schmidhuber. Bidirectional lstm networks for improved phoneme classification and recognition. In International Conference on Artificial Neural Networks: Formal Models and Their Applications, 2005.
- [7] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016.
- [8] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural Comput., 1997.
- [9] Shehzeen Hussain, Paarth Neekhara, Brian Dolhansky, Joanna Bitton, Cristian Canton Ferrer, Julian McAuley, and Farinaz Koushanfar. Exposing vulnerabilities of deepfake detection systems with robust attacks. ACM Journal of Digital Threats: Research and Practice.
- [10] Shehzeen Hussain, Paarth Neekhara, Shlomo Dubnov, Julian McAuley, and Farinaz Koushanfar. Waveguard: Understanding and mitigating audio adversarial examples. In 30th USENIX Security Symposium (USENIX Security 21). USENIX Association, 2021.
- [11] Shehzeen Hussain, Paarth Neekhara, Malhar Jere, Farinaz Koushanfar, and Julian McAuley. Adversarial deepfakes: Evaluating vulnerability of deepfake detectors to adversarial examples. In WACV, 2021.
- [12] Yoon Kim. Convolutional neural networks for sentence classification. In EMNLP, 2014.
- [13] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv:1412.6980, 2014.
- [14] Eliska Kloberdanz. Reprogramming of neural networks: A new and improved machine learning technique. Masters Thesis, 2020.
- [15] Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 1998.
- [16] Yann LeCun and Corinna Cortes. MNIST handwritten digit database. 2010.
- [17] Jaehoon Lee, Jascha Sohl-dickstein, Jeffrey Pennington, Roman Novak, Sam Schoenholz, and Yasaman Bahri. Deep neural networks as gaussian processes. In ICLR, 2018.
- [18] Andrew L. Maas, Raymond E. Daly, Peter T. Pham, Dan Huang, Andrew Y. Ng, and Christopher Potts. Learning word vectors for sentiment analysis. In Human Language Technologies, 2011.
- [19] Alexander Matthews, Jiri Hron, Mark Rowland, Richard E. Turner, and Zoubin Ghahramani. Gaussian process behaviour in wide deep neural networks. In ICLR, 2018.
- [20] S. Moosavi-Dezfooli, A. Fawzi, O. Fawzi, and P. Frossard. Universal adversarial perturbations. In CVPR, 2017.
- [21] Paarth Neekhara, Shehzeen Hussain, Shlomo Dubnov, and Farinaz Koushanfar. Adversarial reprogramming of text classification neural networks. In EMNLP, 2019.
- [22] Paarth Neekhara, Shehzeen Hussain, Prakhar Pandey, Shlomo Dubnov, Julian McAuley, and Farinaz Koushanfar. Universal Adversarial Perturbations for Speech Recognition Systems. In Proc. Interspeech, 2019.
- [23] Nguyen Ngoc Giang, Vu Tran, Duc Ngo, Dau Phan, Favorisen Lumbanraja, M Reza Faisal, Bahriddin Abapihi, Mamoru Kubo, and Kenji Satou. Dna sequence classification by convolutional neural network. Journal of Biomedical Science and Engineering, 2016.
- [24] Michiel O. Noordewier, Geoffrey G. Towell, and Jude W. Shavlik. Training knowledge-based neural networks to recognize genes in dna sequences. In NIPS, 1990.
- [25] Nicolas Papernot, Patrick D. McDaniel, Ian J. Goodfellow, Somesh Jha, Z. Berkay Celik, and Ananthram Swami. Practical black-box attacks against deep learning systems using adversarial examples. arXiv, abs/1602.02697, 2016.
- [26] Nicolas Papernot, Patrick D. McDaniel, Somesh Jha, Matt Fredrikson, Z. Berkay Celik, and Ananthram Swami. The limitations of deep learning in adversarial settings. arXiv, abs/1511.07528, 2015.
- [27] Dmitry Pokholok, Christopher Harbison, Stuart Levine, Megan Cole, Nancy Hannett, Tong Lee, George Bell, Kimberly Walker, P Rolfe, Elizabeth Herbolsheimer, Julia Zeitlinger, Fran Lewitter, David Gifford, and Richard Young. Genome-wide map of nucleosome acetylation and methylation in yeast. Cell, 2005.
- [28] Rajat Raina, Alexis Battle, Honglak Lee, Benjamin Packer, and Andrew Y. Ng. Self-taught learning: Transfer learning from unlabeled data. In ICML, 2007.
- [29] Yi Ren, Yangjun Ruan, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, and Tie-Yan Liu. Fastspeech: Fast, robust and controllable text to speech. In Neurips, 2019.
- [30] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In CVPR, 2016.
- [31] Mingxing Tan and Quoc Le. Efficientnet: Rethinking model scaling for convolutional neural networks. In ICML, 2019.
- [32] Yun-Yun Tsai, Pin-Yu Chen, and Tsung-Yi Ho. Transfer learning without knowing: Reprogramming black-box machine learning models with scarce data and limited resources. In ICML, 2020.
- [33] James Tu, Mengye Ren, Sivabalan Manivasagam, Ming Liang, Bin Yang, Richard Du, Frank Cheng, and Raquel Urtasun. Physically realizable adversarial examples for lidar object detection. In CVPR, 2020.
- [34] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In NIPS, 2017.
- [35] Puyudi Yang, Jianbo Chen, Cho-Jui Hsieh, Jane-Ling Wang, and Michael I. Jordan. Greedy attack and gumbel attack: Generating adversarial examples for discrete data. arXiv, 2018.
- [36] Xiang Zhang, Junbo Jake Zhao, and Yann LeCun. Character-level convolutional networks for text classification. In NIPS, 2015.