Modulating Bottom-Up and Top-Down Visual Processing via Language-Conditional Filters
Abstract
How to best integrate linguistic and perceptual processing in multi-modal tasks that involve language and vision is an important open problem. In this work, we argue that the common practice of using language in a top-down manner, to direct visual attention over high-level visual features, may not be optimal. We hypothesize that the use of language to also condition the bottom-up processing from pixels to high-level features can provide benefits to the overall performance. To support our claim, we propose a U-Net-based model and perform experiments on two language-vision dense-prediction tasks: referring expression segmentation and language-guided image colorization. We compare results where either one or both of the top-down and bottom-up visual branches are conditioned on language. Our experiments reveal that using language to control the filters for bottom-up visual processing in addition to top-down attention leads to better results on both tasks and achieves competitive performance. Our linguistic analysis suggests that bottom-up conditioning improves segmentation of objects especially when input text refers to low-level visual concepts. Code is available at https://github.com/ilkerkesen/bvpr.
1 Introduction
As human beings, we can easily perceive our surroundings with our visual system and interact with each other using language. Since the work of Winograd [78], developing a system that understands human language in a situated environment has been one of the long-standing goals of artificial intelligence. Recent successes of deep learning studies in both language and vision domains have increased the interest in tasks that combine language and vision [3, 80, 43, 72, 2, 34]. However, how to best integrate linguistic and perceptual processing is still an important open problem. Towards this end, we investigate whether language should be used for conditioning bottom-up visual processing as well as top-down attention.
In the human visual system, attention is driven by both top-down cognitive processes (e.g. focusing on a given shape) and bottom-up salient, behaviourally relevant stimuli (e.g. fast moving objects and contrasting colors) [14, 13, 74]. Studies on embodied language explore the link between linguistic and perceptual representations [66, 76, 22] and often assume that language has a high-level effect on perception and drives the top-down visual attention [5, 37, 18]. However, recent studies from cognitive science point out that language comprehension also affects low-level visual processing [61, 6, 54]. Motivated by this, we propose a model that can modulate either or both of bottom-up and top-down visual processing with language and compare different designs for language modulation.
Current deep learning systems for language-vision tasks typically start with low-level image processing, then connect the language representation with high-level visual features to control the visual focus. To integrate both modalities, concatenation [56], element-wise multiplication [51, 41], attention from language to vision [79, 86, 50, 1, 91] and transformers [73, 17, 20] are commonly used. These studies typically do not condition low-level visual features on language. Some methods [15, 65] do the opposite by conditioning only the bottom-up visual processing on language.
To evaluate language-modulation on the bottom-up and top-down visual branches independently, we develop an architecture that clearly separates these two branches (based on U-Net [68]) and allows us to experiment with modulating one or both branches with language. The bottom-up branch starts from low-level visual features and applies a sequence of contracting filters that result in successively higher level feature maps with lower spatial resolution. Following this, a top-down branch takes the final low resolution feature map and applies a sequence of expanding filters that eventually result in a map with the original image resolution. Information flows between branches through skip connections between contracting and expanding filters at the same level. Our proposed architecture is task-agnostic and it can be used for various vision-language tasks involving dense prediction. We evaluate our model with different language-modulation settings on two different tasks: referring expression segmentation and language-guided image colorization.
In the referring expression segmentation (RES) task, given an image and a natural language description, the aim is to obtain the segmentation mask that marks the object(s) described. We can contrast this with pure image based object detection [25, 67] and semantic segmentation [49, 9] tasks which are limited to predefined semantic classes. The language input may contain various visual attributes (e.g. shape), spatial information (e.g. in front of), actions (e.g. running) and interactions/relations between different objects (e.g. arm of the chair that the cat is sitting on).
In language-guided image colorization (LIC) task, given a grayscale image and a description, the aim is to predict pixel color values. The absence of color information in the input images makes this problem interesting to experiment with because color words do not help in conditioning the bottom-up branch when the input image is grayscale.
We find that conditioning both branches leads to better results, achieving competitive performance on both tasks. Our experiments suggest that conditioning the bottom-up branch on language is important to ground low-level visual information. On RES, we find that modulating only the bottom-up branch performs significantly better than modulating only the top-down branch especially when color-dependent language is present in the input. Our findings on LIC show that when color information absent in input images, the bottom-up baseline naturally fails to predict and manipulate colors of target objects specified by input language. That said, conditioning the bottom-up branch still improves the colorization quality by helping our model to accurately segment and colorize the target objects as a whole.
2 Related Work
Referring Expression Comprehension (REC). In this problem, the goal is to locate a bounding box for the object(s) described in the input language. The proposed solutions can be divided into two categories: two-stage and one-stage methods. Two-stage methods [58, 30, 63, 89, 77, 12, 30, 89, 48, 81] rely on a pre-trained object detector [67, 27] to generate object proposals in the first stage. In the second stage, they assign scores to the object proposals depending on how much they match with input language. One-stage methods [53, 82, 45, 85, 84, 17] directly localize the referred objects in one step. Most of these methods condition only the top-down visual processing on language, while some fuse language with multi-level visual representations.
Referring Expression Segmentation (RES). In this task, the aim is to generate a segmentation mask for the object(s) referred in the input language [31]. To accomplish this, multi-modal LSTMs [47, 60], ConvLSTMs [70, 47, 7, 87], word-level attention [69, 35, 7, 87], cross-modal attention [88, 32, 33, 53, 52, 38], and transformers [75, 20] have been used. Each one of these methods modulates only the top-down branch with language. As one exception, EFN [21] conditions the bottom-up branch on language, but does not modulate the top-down branch with language.
Language-guided Image Colorization (LIC). In this task, the aim is to predict colors for all the pixels of a given input grayscale image based on input descriptive text. Specifically, [57] inserts extra Feature-wise Linear Modulation (FiLM) layers [65] into a pre-trained ResNet to predict color values in LAB color space. Multi-modal LSTMs [47, 94] and generative adversarial networks [26, 4, 8] are also used in this context to colorize sketches. Similar to us, Tag2Pix [40] extends U-Net to perform colorization on line art data, but it modulates only the top-down visual processing with symbolic input using concatenation.
Language-conditional Parameters. Here we review methods that use input-text-dependent dynamic parameters to process visual features. To control a visual model with language, MODERN and FiLM [15, 65] used conditional batch normalization layers with language-conditioned coefficients rather than customized filters. Numerous methods [44, 23, 24, 60, 11, 62] generate language-conditional dynamic filters to convolve visual features. Some RES models [60, 11] also incorporate language-conditional filters into the their top-down visual processing. To map instructions to actions in virtual environments, LingUNet [62] extends U-Net by adding language-conditional filters to the top-down visual processing only. Each one of these methods conditions either top-down or bottom-up branch only.
Comparison. To support our main research question, our architecture clearly separates bottom-up and top-down visual processing. This allows us to experiment with modulating either one branch or both branches with language and evaluate their individual contributions. The majority of related work conditions only the top-down visual processing on language. Other U-Net-based methods [62, 40] and most transformer models [20, 17] which implement cross-modal attention between textual and visual representations in top-down visual processing fall into this category. A few exceptions [15, 65, 21] do the opposite by conditioning only the bottom-up branch. Some methods [53, 85, 33] fuse language with a multi-level visual representation, which leads to good results, but this kind of fusion does not allow the evaluation of language conditioning on top-down vs bottom-up visual processing. Our architecture allows language to control either or both of top-down and bottom-up branches. We show that (i) the bottom-up conditioning is important to ground language to low-level visual features, and (ii) conditioning both branches on language leads to best results.

3 The Model
Here, we describe our proposed model in detail. Figure 1 shows an overview of our proposed architecture. First, the model extracts a tensor of low-level features using a pre-trained convolutional neural network and encodes the given natural language expression to a vector representation using an LSTM [29]. Starting with the visual feature tensor, the model generates feature maps through a contracting and expanding path where the output head takes the final map and performs dense prediction, similar to U-Net [68]. Our proposed architecture modulates both of the contracting and expanding paths with language using convolutional filters generated from the given expression. It is important to emphasize that the previous works either have a language-guided top-down or a language-conditional bottom-up visual processing branch. As will be discussed in the next section, our experiments show that modulating both of these paths improves the performance dramatically.
3.1 Low-level Image Features
Given an input image , we extract visual features by using a pre-trained convolutional network. In the RES task, we use DeepLab-v3+ backbones [10], and in the LIC task, we use ResNet-101 pre-trained on ImageNet [28, 16]. On the task of referring expression segmentation, we also concatenate 8-D location features to this feature map following the previous work [47, 88].
3.2 Language Representation
Consider a language input where each word is represented with a -dimensional GloVe embedding [64]. We map the language input to hidden states using an LSTM as . We use the final hidden state of the LSTM as the language representation, . Later on, we split this language representation into pieces to generate language-conditional filters.
3.3 Multi-modal Encoder
After generating image and language representations, our model generates a multi-modal feature map representing the input image and the given natural language expression. Our multi-modal encoder module extends U-Net by conditioning both contracting and expanding branches on language using language-conditional filters.
In the bottom-up branch, our model applies convolutional modules to the image representation . Each module, , takes the concatenation of the previously generated feature map () and its convolved version with language-conditional filters and produces an output feature map (). Each has a 2D convolution layer followed by batch normalization [36] and ReLU activation function [55]. The convolution layers have filters with and halving the spatial resolution, and they all have the same number of output channels.
Similar to [62], we split the textual representation to equal parts (), and then use each to generate a language-conditional filter for th bottom-up layer ():
(1) |
Each is an affine transformation followed by normalizing and reshaping to convert the output to convolutional filters. After obtaining the filters, we convolve them over the feature map obtained from the previous module () to relate expressions to image features:
(2) |
Then, the concatenation of these text-modulated features () for th bottom-up layer and the previously generated features () is fed into module for the next step.
In the top-down branch, we generate feature maps starting from the final output of the contracting branch as:
(3) | ||||
(4) | ||||
(5) |
Similar to the bottom-up branch, is the modulated feature map with language-conditional filters defined as:
(6) |
where is again an affine transformation followed by normalizing and reshaping for th layer of the top-down branch. Here, we convolve the filter () over the feature maps from the contracting branch (). Each upsampling module gets the concatenation () of the text related features and the feature map () generated from the previous module. Only the first module operates on just convolved features. Each consists of a 2D deconvolution layer followed by a batch normalization and ReLU activation function. Final output becomes our joint feature map representing the input image / language pair. The deconvolution layers have filters with and doubling the spatial resolution, and they all have the same number of output channels.
3.4 Output Heads
As mentioned earlier, we develop our model as a generic solution which can be used to solve language-vision problems involving dense prediction. In this direction, we adapt our model to two different tasks by varying the output head: referring expression segmentation and language-guided image colorization.
Segmentation. In the referring expression segmentation problem, the goal is to generate segmentation mask for a given input image and language pair. After generating the joint feature map , we apply a stack of layers (, , …, ) to map to the exact image size. Similar to upsampling modules, each is a 2D deconvolution layer followed by batch normalization and the ReLU activation. Each preserves the number of channels except for the last one which maps the features to a single channel for the mask prediction. We omit the batch normalization and the ReLU activation for the final module, instead we apply a sigmoid function to turn the final features into probabilities. Given these probabilities and ground-truth mask, we train our network by using binary cross entropy loss.
Colorization. In the language-guided image colorization task, the goal is to predict pixel color values for given input image with the guidance of language input. A convolutional layer by with 33 filters generates class scores for each spatial location of . We apply bilinear upsampling to these predicted scores to match input image size. Given predicted scores and ground-truth color classes, we train our model by using a weighted cross entropy loss. To create compound LAB color classes and their weights, we follow the exactly same process with [57, 92].
Method | UNC | UNC+ | G-Ref | ReferIt | ||||||
val | testA | testB | val | testA | testB | val (G) | val (U) | test (U) | test | |
CMSA [88] | 58.32 | 60.61 | 55.09 | 43.76 | 47.60 | 37.89 | 39.98 | - | - | 63.80 |
STEP [7] | 60.04 | 63.46 | 57.97 | 48.19 | 52.33 | 40.41 | 46.40 | - | - | 64.13 |
BRINet [32] | 61.35 | 63.37 | 59.57 | 48.57 | 52.87 | 42.13 | 48.04 | - | - | 63.46 |
CMPC [33] | 61.36 | 64.53 | 59.64 | 49.56 | 53.44 | 43.23 | 49.05 | - | - | 65.53 |
LSCM [35] | 61.47 | 64.99 | 59.55 | 49.34 | 53.12 | 43.50 | 48.05 | - | - | 66.57 |
EFN [21] | 62.76 | 65.69 | 59.67 | 51.50 | 55.24 | 43.01 | 51.93 | - | - | 66.70 |
BUSNet [83] | 63.27 | 66.41 | 61.39 | 51.76 | 56.87 | 44.13 | 50.56 | - | - | - |
Our model | 64.63 | 67.76 | 61.03 | 51.76 | 56.77 | 43.80 | 50.88 | 52.12 | 52.94 | 66.01 |
MCN† [53] | 62.44 | 64.20 | 59.71 | 50.62 | 54.99 | 44.69 | - | 49.22 | 49.40 | - |
CGAN† [52] | 64.86 | 68.04 | 62.07 | 51.03 | 55.51 | 44.06 | 46.54 | 51.01 | 51.69 | - |
LTS† [38] | 65.43 | 67.76 | 63.08 | 54.21 | 58.32 | 48.02 | - | 54.40 | 54.25 | - |
VLT† [20] | 65.65 | 68.29 | 62.73 | 55.50 | 59.20 | 49.36 | 49.76 | 52.99 | 56.65 | - |
Our model† | 67.01 | 69.63 | 63.45 | 55.34 | 60.72 | 47.11 | 53.51 | 55.09 | 55.31 | 57.09 |
4 Experimental Analysis
This section contains the details of our experiments on referring expression segmentation (Section 4.1) and language-guided image colorization (Section 4.2) tasks111We provide the implementation details, and present the complete ablation experiments in the supplementary material..
4.1 Referring Expression Segmentation
Datasets. We evaluate our model on ReferIt (130.5k expressions, 19.9k images) [39], UNC (142k expressions, 20k images), UNC+ (141.5k expressions, 20k images) [90] and Google-Ref (G-Ref) (104.5k expressions, 26.7k images) [58] datasets. Unlike UNC, location-specific expressions are excluded in UNC+ through enforcing annotators to describe objects by their appearance. ReferIt, UNC, UNC+ datasets are collected through a two-player game and have short expressions (avg. 4 words). G-Ref have longer and richer expressions, its expressions are collected from Amazon Mechanical Turk instead of a two-player game. G-Ref does not contain a test split, and Nagaraja et al. [63] extends it by having separate splits for validation and test, which are denoted as val (U) and test (U).
Evaluation Metrics. Following previous work, we use intersection-over-union (IoU) and as evaluation metrics. Given the predicted segmentation mask and the ground truth, the IoU metric is the ratio between the intersection and the union of the two. There are two different ways to calculate IoU: the overall IoU calculates the total intersection over total union score throughout the entire dataset and the mean IoU calculates the mean of IoU scores of each individual example. For a fair comparison, we use both IoU metrics. The second metric, , calculates the percentage of the examples that have IoU score higher than the threshold .
Quantitative Results. Table 1 shows the comparison of our model with previous methods. Bold faces highlight the highest achieved scores. We evaluate our model by using both IoU metrics for a fair evaluation. Table 2 presents the comparison of our model with the state-of-the-art in terms of . The difference between our model and the state-of-the-art increases as the threshold increases. This indicates that our model is better at segmenting the referred objects including smaller ones.
Method | [email protected] | [email protected] | [email protected] | [email protected] | [email protected] |
---|---|---|---|---|---|
CMSA | 66.44 | 59.70 | 50.77 | 35.52 | 10.96 |
STEP | 70.15 | 63.37 | 53.15 | 36.53 | 10.45 |
BRINet | 71.83 | 65.05 | 55.64 | 39.36 | 11.21 |
LSCM | 70.84 | 63.82 | 53.67 | 38.69 | 12.06 |
EFN | 73.95 | 69.58 | 62.59 | 49.61 | 20.63 |
MCN | 76.60 | 70.33 | 58.39 | 33.68 | 5.26 |
LTS | 75.16 | 69.51 | 60.74 | 45.17 | 14.41 |
Our Model | 76.67 | 71.77 | 64.76 | 51.69 | 22.73 |
Qualitative Results. We visualize some of the segmentation predictions of our model to gain better insights about the trained model. Figure 2 shows some of the cases that our model segments correctly. These examples demonstrate that our model can learn a variety of language and visual reasoning patterns. For example, the first two examples of the fourth column show that our model learns to relate superlative adjectives (e.g., taller) with visual comparison. Examples including spatial prepositions (e.g., near to) demonstrate the spatial reasoning ability of the model. Our model can also learn a domain-specific nomenclature (e.g. catcher) that is present in the dataset. Lastly, we can see that the model can also detect certain actions (e.g., sitting).

Figure 3 shows failure cases from our model on the UNC test split. Examples in (a) show that our model tends to fail in case of typos. Our model segments the correct objects for these two examples when the typos are fixed (e.g. pink instead of pick). Examples in (b) show that some of the expressions are ambiguous, where there are multiple objects that could be referred to by the expression. In this case, the model seems to segment the most salient object. Some of the annotations contain incorrect or incomplete ground-truth mask (c). Finally, some of the examples (d) are hard to segment completely due to the lack of light or objects that mask the referred objects.

Method | Backbone | IoU |
---|---|---|
Top-down Baseline | ResNet-50 | 58.06 |
Bottom-up Baseline | ResNet-50 | 60.74 |
Our Model | ResNet-50 | 63.59 |
Our Model | ResNet-101 | 64.63 |
Ablation Study. We implemented 3 different models, the top-down baseline, the bottom-up baseline and our model, to show the effect of modulating language in expanding and contracting visual branches. While the bottom-up baseline modulates language in bottom-up branch only, the top-down baseline modulates language in top-down branch only. Our model conditions language on both branches. Table 3 shows us the results. The bottom-up baseline outperforms the top-down one with 2.7 IoU improvement. Modulating language in both branches yields the best results by improving the bottom-up baseline with 2.85 IoU score.

Figure 4 visualizes the predictions of the different models on the same examples. The bottom-up baseline performs better when the description has color information as we show in the first three examples. The top-down-only baseline also fails to detect object categories in some cases, and segments additional unwanted objects with similar category or appearance (e.g. banana vs. orange). Overall, our model which conditions both visual branches on language gives the best results.
Method | +C | -C | +IN | +IN,-C | +JJ* | +JJ*,-C |
---|---|---|---|---|---|---|
Top-down Baseline | 52.59 | 59.59 | 54.84 | 55.66 | 57.66 | 61.16 |
Bottom-up Baseline | 60.40 | 60.02 | 56.05 | 55.49 | 61.18 | 61.86 |
Our Model | 62.98 | 63.57 | 59.60 | 59.27 | 64.45 | 65.56 |
Language-oriented Analysis. To analyze the effect of language on model performance, we divided UNC test splits into subsets depending on the different types of words (e.g. colors) and phrases (e.g. noun phrases with multiple adjectives) included in input expressions. Table 4 shows us the results of different models on these subsets. The first column stands for models, and the rest stand for different input expression categories. We exclude the categories which do not contribute to our analysis. We use DeepLab-v3+ ResNet-50 as visual backbone in each method. The notation of the categories are similar to Part of Speech (POS) tags [59], where we denote prepositions with IN, examples with adjectives with JJ*, and colors with C. Preceding plus and minus signs stand for inclusion and exclusion. For instance, +IN,-C column stands for the subset where each expression contains at least one preposition without any color words. Color words (e.g. red, darker) has the most impact on the performance in comparison to other types of words and phrases. Our model and the bottom-up baseline performs significantly better than top-down baseline on the subset that includes colors. In the opposite case, where input expressions with colors are excluded, the top-down baseline has performance similar to the bottom-up baseline, and our final model outperforms both single branch models. Since colors can be seen as low-level sensory information, low performance in the absence of the bottom-up branch can be expected. This demonstrates the importance of conditioning the bottom-up visual branch on language to capture low-level visual concepts.
4.2 Language-guided Image Colorization
Datasets. Following the prior work [57], we use a modified version of COCO dataset [46] where the descriptions that do not contain color words are excluded. In this modified version, the training split has 66165, and the validation split has 32900 image / description pairs, and all images have a resolution of pixels.
Evaluation Metrics. Following the previous work, we use pixel-level top-1 (acc@1) and top-5 accuracies (acc@5) in LAB color space, and additionally PSNR and LPIPS [93] in RGB for evaluation. A lower score is better for LPIPS, and a higher score is better for the rest.
Method | acc@1 | acc@5 | PSNR | LPIPS |
---|---|---|---|---|
FiLMed ResNet | 23.70 | 60.50 | - | - |
FiLMed ResNet (ours) | 20.22 | 49.57 | 20.89 | 0.1280 |
Top-down Baseline | 22.83 | 51.85 | 21.29 | 0.1226 |
Bottom-up Baseline | 21.85 | 51.34 | 20.98 | 0.1448 |
Our Model | 23.38 | 54.27 | 21.42 | 0.1262 |
Our Model w/o balancing | 33.74 | 67.83 | 22.75 | 0.1250 |
Quantitative Results. We present the quantitative performance of our model in Table 5, and compare it with different design choices and previous work. FiLMed ResNet [57] uses FiLM [65] to perform language-conditional colorization. FiLMed ResNet (ours) denotes the results reproduced by the implementation provided by the authors. To show the effect of language modulation on different branches, we train 3 different models again: the top-down baseline, the bottom-up baseline and our model. We also re-train our model without class rebalancing and denote it as Our Model w/o balancing.
Contrary to the segmentation experiments, the top-down baseline performs better than the bottom-up baseline on the colorization task in all measures. Since, color information is absent in input images, bottom-up branch cannot encode low-level image features by modulating color-dependent language.
When we disable class rebalancing in the training phase, we observe a large improvement in acc@1 and acc@5 due to the imbalanced color distribution, where the model predicts the frequent colors exist in the backgrounds.
Qualitative Results. We visualize some of the colorization outputs of the trained models to analyze them in more detail in Figure 5. FiLMed ResNet (ours) can understand all colorization hints, and it can manipulate object colors with some incorrectly predicted areas. The top-down baseline also performs similar to FiLMed ResNet (ours), where both models condition only the top-down branch on language.
In this task, since the models are blind to color, the bottom-up baseline loses its effectiveness to some degree, and starts to predict the most probable colors. This can be seen on the second and the last example, where the bottom-up baseline predicts red for the stop sign and blue for the sky. Although, the bottom-up baseline performs worse in this task, modulating the bottom-up branch with language still contributes to our final model to localize and recognize the objects present in the scene. This can be seen on the last two examples, where the top-down baseline mixes colors up in some object parts (e.g. the red parts in the motorcycle). Our model w/o balancing tends to predict more grayish colors (e.g. dog, sky).
Figure 6 highlights some of the failure cases we observed throughout the dataset. In the first two examples, our model is able to localize and recognize the target objects, but it fails to colorize them successfully by colorizing not only the targeted parts but also other parts. Models generally fail to colorize small objects since the data is imbalanced and it contains vast backgrounds and big objects frequently. The last two examples show that models fail to colorize reflective or transparent objects like glasses or water, these were also difficult in the language based segmentation task (see Figure 3 (d)).


5 Conclusion
In this work, we suggested that conditioning both top-down and bottom-up visual processing on language is beneficial for grounding language to vision. To support this claim, we proposed a generic architecture with explicit bottom-up and top-down visual branches for vision-language problems involving dense prediction. Our experiments on two different tasks demonstrated that conditioning both visual branches on language gives the best results. Our experiments on the referring expression segmentation task revealed that conditioning the bottom-up branch on language plays a vital role to process color-dependent input language. The language-guided image colorization experiments demonstrated similar conclusions, the bottom-up baseline failed to colorize the target objects since the color information is absent in the input images.
Limitations. We share common failure cases in Figure 3 and Figure 6. The performance of our model on both tasks decreases in the presence of transparent and/or reflective objects. Our model also fails to colorize small objects, mostly due to having an imbalanced color distribution. Finally, our current model is limited to integrated vision and language tasks involving dense prediction, and we did not perform experiments on other vision and language problems.
Acknowledgements. This work was supported in part by an AI Fellowship to I. Kesen provided by the KUIS AI Center, GEBIP 2018 Award of the Turkish Academy of Sciences to E. Erdem, and BAGEP 2021 Award of the Science Academy to A. Erdem.
References
- [1] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In CVPR, pages 6077–6086, 2018.
- [2] Peter Anderson, Qi Wu, Damien Teney, Jake Bruce, Mark Johnson, Niko Sünderhauf, Ian Reid, Stephen Gould, and Anton van den Hengel. Vision-and-Language Navigation: Interpreting Visually-Grounded Navigation Instructions in Real Environments. In CVPR, June 2018.
- [3] Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C. Lawrence Zitnick, and Devi Parikh. VQA: Visual Question Answering. In ICCV, Dec. 2015.
- [4] Hyojin Bahng, Seungjoo Yoo, Wonwoong Cho, David Keetae Park, Ziming Wu, Xiaojuan Ma, and Jaegul Choo. Coloring with words: Guiding image colorization through text-based palette generation. In ECCV, pages 431–447, 2018.
- [5] Paul Bloom. How children learn the meanings of words. MIT press, 2002.
- [6] Bastien Boutonnet and Gary Lupyan. Words jump-start vision: A label advantage in object recognition. Journal of Neuroscience, 35(25):9329–9335, 2015.
- [7] Ding-Jie Chen, Songhao Jia, Yi-Chen Lo, Hwann-Tzong Chen, and Tyng-Luh Liu. See-through-text grouping for referring image segmentation. In ICCV, pages 7454–7463, 2019.
- [8] Jianbo Chen, Yelong Shen, Jianfeng Gao, Jingjing Liu, and Xiaodong Liu. Language-based image editing with recurrent attentive models. In CVPR, pages 8721–8729, 2018.
- [9] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. TPAMI, 40(4):834–848, 2017.
- [10] Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig Adam. Encoder-decoder with atrous separable convolution for semantic image segmentation. In ECCV, pages 801–818, 2018.
- [11] Y.-W. Chen, Y.-H. Tsai, T. Wang, Y.-Y. Lin, and M.-H. Yang. Referring Expression Object Segmentation with Caption-Aware Consistency. In BMVC, 2019.
- [12] Volkan Cirik, Taylor Berg-Kirkpatrick, and Louis-Philippe Morency. Using syntax to ground referring expressions in natural images. In AAAI, 2018.
- [13] Charles E Connor, Howard E Egeth, and Steven Yantis. Visual attention: bottom-up versus top-down. Current biology, 14(19):R850–R852, 2004.
- [14] Maurizio Corbetta and Gordon L Shulman. Control of goal-directed and stimulus-driven attention in the brain. Nature reviews neuroscience, 3(3):201–215, 2002.
- [15] Harm De Vries, Florian Strub, Jérémie Mary, Hugo Larochelle, Olivier Pietquin, and Aaron C Courville. Modulating early visual processing by language. In NeurIPS, pages 6594–6604, 2017.
- [16] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In CVPR, pages 248–255, 2009.
- [17] Jiajun Deng, Zhengyuan Yang, Tianlang Chen, Wengang Zhou, and Houqiang Li. Transvg: End-to-end visual grounding with transformers. In ICCV, pages 1769–1779, October 2021.
- [18] Banchiamlack Dessalegn and Barbara Landau. More than meets the eye: The role of language in binding and maintaining feature conjunctions. Psychological science, 19(2):189–195, 2008.
- [19] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In ACL, pages 4171–4186, June 2019.
- [20] Henghui Ding, Chang Liu, Suchen Wang, and Xudong Jiang. Vision-language transformer and query generation for referring segmentation. In ICCV, 2021.
- [21] Guang Feng, Zhiwei Hu, Lihe Zhang, and Huchuan Lu. Encoder fusion network with co-attention embedding for referring image segmentation. In CVPR, pages 15506–15515, 2021.
- [22] Vittorio Gallese and George Lakoff. The brain’s concepts: The role of the sensory-motor system in conceptual knowledge. Cognitive neuropsychology, 22(3-4):455–479, 2005.
- [23] Peng Gao, Hongsheng Li, Shuang Li, Pan Lu, Yikang Li, Steven CH Hoi, and Xiaogang Wang. Question-guided hybrid convolution for visual question answering. In ECCV, pages 469–485, 2018.
- [24] Kirill Gavrilyuk, Amir Ghodrati, Zhenyang Li, and Cees GM Snoek. Actor and action video segmentation from a sentence. In CVPR, pages 5958–5966, 2018.
- [25] Ross Girshick. Fast R-CNN. ICCV, Dec. 2015.
- [26] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Q. Weinberger, editors, NeurIPS, volume 27, 2014.
- [27] Kaiming He, Georgia Gkioxari, Piotr Dollar, and Ross Girshick. Mask R-CNN. ICCV, Oct. 2017.
- [28] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, pages 770–778, 2016.
- [29] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
- [30] Ronghang Hu, Marcus Rohrbach, Jacob Andreas, Trevor Darrell, and Kate Saenko. Modeling Relationships in Referential Expressions with Compositional Modular Networks. CVPR, July 2017.
- [31] Ronghang Hu, Marcus Rohrbach, and Trevor Darrell. Segmentation from Natural Language Expressions. Lecture Notes in Computer Science, pages 108–124, 2016.
- [32] Zhiwei Hu, Guang Feng, Jiayu Sun, Lihe Zhang, and Huchuan Lu. Bi-Directional Relationship Inferring Network for Referring Image Segmentation. In CVPR, pages 4424–4433, 2020.
- [33] Shaofei Huang, Tianrui Hui, Si Liu, Guanbin Li, Yunchao Wei, Jizhong Han, Luoqi Liu, and Bo Li. Referring image segmentation via cross-modal progressive comprehension. In CVPR, pages 10488–10497, 2020.
- [34] Drew A Hudson and Christopher D Manning. GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering. CVPR, 2019.
- [35] Tianrui Hui, Si Liu, Shaofei Huang, Guanbin Li, Sansi Yu, Faxi Zhang, and Jizhong Han. Linguistic structure guided context modeling for referring image segmentation. In ECCV, pages 59–75. Springer, 2020.
- [36] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In ICML, pages 448–456. PMLR, 2015.
- [37] Ray Jackendoff and Ray S Jackendoff. Foundations of language: Brain, meaning, grammar, evolution. Oxford University Press, USA, 2002.
- [38] Ya Jing, Tao Kong, Wei Wang, Liang Wang, Lei Li, and Tieniu Tan. Locate then segment: A strong pipeline for referring image segmentation. In CVPR, pages 9858–9867, 2021.
- [39] Sahar Kazemzadeh, Vicente Ordonez, Mark Matten, and Tamara Berg. Referitgame: Referring to objects in photographs of natural scenes. In EMNLP, pages 787–798, 2014.
- [40] Hyunsu Kim, Ho Young Jhoo, Eunhyeok Park, and Sungjoo Yoo. Tag2pix: Line art colorization using text tag with secat and changing loss. In ICCV, pages 9056–9065, 2019.
- [41] Jin-Hwa Kim, Sang-Woo Lee, Donghyun Kwak, Min-Oh Heo, Jeonghee Kim, Jung-Woo Ha, and Byoung-Tak Zhang. Multimodal residual learning for visual qa. In NeurIPS, pages 361–369, 2016.
- [42] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- [43] Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li-Jia Li, David A Shamma, et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. IJCV, 123(1):32–73, 2017.
- [44] Zhenyang Li, Ran Tao, Efstratios Gavves, Cees GM Snoek, and Arnold WM Smeulders. Tracking by natural language specification. In CVPR, pages 6495–6503, 2017.
- [45] Yue Liao, Si Liu, Guanbin Li, Fei Wang, Yanjie Chen, Chen Qian, and Bo Li. A real-time cross-modality correlation filtering method for referring expression comprehension. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
- [46] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, pages 740–755, 2014.
- [47] Chenxi Liu, Zhe L. Lin, Xiaohui Shen, Jimei Yang, Xin Lu, and Alan L. Yuille. Recurrent Multimodal Interaction for Referring Image Segmentation. ICCV, pages 1280–1289, 2017.
- [48] Daqing Liu, Hanwang Zhang, Feng Wu, and Zheng-Jun Zha. Learning to Assemble Neural Module Tree Networks for Visual Grounding. In ICCV, Oct. 2019.
- [49] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In CVPR, pages 3431–3440, 2015.
- [50] Jiasen Lu, Caiming Xiong, Devi Parikh, and Richard Socher. Knowing when to look: Adaptive attention via a visual sentinel for image captioning. In CVPR, pages 375–383, 2017.
- [51] Jiasen Lu, Jianwei Yang, Dhruv Batra, and Devi Parikh. Hierarchical question-image co-attention for visual question answering. In NeurIPS, pages 289–297, 2016.
- [52] Gen Luo, Yiyi Zhou, Rongrong Ji, Xiaoshuai Sun, Jinsong Su, Chia-Wen Lin, and Qi Tian. Cascade grouped attention network for referring expression segmentation. In Proceedings of the 28th ACM International Conference on Multimedia, pages 1274–1282, 2020.
- [53] Gen Luo, Yiyi Zhou, Xiaoshuai Sun, Liujuan Cao, Chenglin Wu, Cheng Deng, and Rongrong Ji. Multi-task collaborative network for joint referring expression comprehension and segmentation. In CVPR, pages 10034–10043, 2020.
- [54] Gary Lupyan and Andy Clark. Words and the world: Predictive coding and the language-perception-cognition interface. Current Directions in Psychological Science, 24(4):279–284, 2015. Publisher: Sage Publications Sage CA: Los Angeles, CA.
- [55] Andrew L Maas, Awni Y Hannun, and Andrew Y Ng. Rectifier nonlinearities improve neural network acoustic models. In ICML, volume 30, page 3, 2013.
- [56] Mateusz Malinowski, Marcus Rohrbach, and Mario Fritz. Ask your neurons: A neural-based approach to answering questions about images. In ICCV, pages 1–9, 2015.
- [57] Varun Manjunatha, Mohit Iyyer, Jordan Boyd-Graber, and Larry Davis. Learning to color from language. In NAACL-HLT, vol. 2, pages 764–769, June 2018.
- [58] Junhua Mao, Jonathan Huang, Alexander Toshev, Oana Camburu, Alan Yuille, and Kevin Murphy. Generation and Comprehension of Unambiguous Object Descriptions. CVPR, June 2016.
- [59] Mitchell P Marcus, Mary Ann Marcinkiewicz, and Beatrice Santorini. Building a large annotated corpus of english: the penn treebank. Computational Linguistics, 19(2):313–330, 1993.
- [60] Edgar Margffoy-Tuay, Juan C Pérez, Emilio Botero, and Pablo Arbeláez. Dynamic multimodal instance segmentation guided by natural language queries. In ECCV, pages 630–645, 2018.
- [61] Lotte Meteyard, Bahador Bahrami, and Gabriella Vigliocco. Motion detection and motion verbs: Language affects low-level visual perception. Psychological Science, 18(11):1007–1013, 2007.
- [62] Dipendra Misra, Andrew Bennett, Valts Blukis, Eyvind Niklasson, Max Shatkhin, and Yoav Artzi. Mapping instructions to actions in 3D environments with visual goal prediction. In EMNLP, pages 2667–2678, Oct.-Nov. 2018.
- [63] Varun K. Nagaraja, Vlad I. Morariu, and Larry S. Davis. Modeling Context Between Objects for Referring Expression Understanding. Lecture Notes in Computer Science, pages 792–807, 2016.
- [64] Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In EMNLP, pages 1532–1543, 2014.
- [65] Ethan Perez, Florian Strub, Harm de Vries, Vincent Dumoulin, and Aaron C. Courville. FiLM: Visual Reasoning with a General Conditioning Layer. In AAAI, 2018.
- [66] Friedemann Pulvermüller. Words in the brain’s language. Behavioral and brain sciences, 22(2):253–279, 1999.
- [67] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. TPAMI, 39(6):1137–1149, June 2017.
- [68] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In MICCAI, pages 234–241, 2015.
- [69] Hengcan Shi, Hongliang Li, Fanman Meng, and Qingbo Wu. Key-Word-Aware Network for Referring Expression Image Segmentation. In ECCV, Sept. 2018.
- [70] Xingjian Shi, Zhourong Chen, Hao Wang, Dit-Yan Yeung, Wai-kin Wong, and Wang-chun WOO. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In NeurIPS, pages 802–810. 2015.
- [71] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. JMLR, 15(1):1929–1958, 2014.
- [72] Alane Suhr, Mike Lewis, James Yeh, and Yoav Artzi. A Corpus of Natural Language for Visual Reasoning. In ACL vol. 2, pages 217–223, Vancouver, Canada, July 2017. Association for Computational Linguistics.
- [73] Hao Tan and Mohit Bansal. Lxmert: Learning cross-modality encoder representations from transformers. In EMNLP-IJCNLP, pages 5100–5111, 2019.
- [74] Jan Theeuwes. Top–down and bottom–up control of visual selection. Acta psychologica, 135(2):77–99, 2010.
- [75] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \textbackslashLukasz Kaiser, and Illia Polosukhin. Attention is all you need. In NeurIPS, pages 5998–6008, 2017.
- [76] Gabriella Vigliocco, David P Vinson, William Lewis, and Merrill F Garrett. Representing the meanings of object and action words: The featural and unitary semantic space hypothesis. Cognitive psychology, 48(4):422–488, 2004.
- [77] Peng Wang, Qi Wu, Jiewei Cao, Chunhua Shen, Lianli Gao, and Anton van den Hengel. Neighbourhood Watch: Referring Expression Comprehension via Language-Guided Graph Attention Networks. CVPR, June 2019.
- [78] Terry Winograd. Understanding natural language. Cognitive psychology, 3(1):1–191, 1972.
- [79] Huijuan Xu and Kate Saenko. Ask, attend and answer: Exploring question-guided spatial attention for visual question answering. In ECCV, pages 451–466. Springer, 2016.
- [80] Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. Show, attend and tell: Neural image caption generation with visual attention. In ICML, pages 2048–2057, 2015.
- [81] Sibei Yang, Guanbin Li, and Yizhou Yu. Graph-structured referring expressions reasoning in the wild. 2020.
- [82] Sibei Yang, Guanbin Li, and Yizhou Yu. Propagating over phrase relations for one-stage visual grounding. In ECCV, pages 589–605. Springer, 2020.
- [83] Sibei Yang, Meng Xia, Guanbin Li, Hong-Yu Zhou, and Yizhou Yu. Bottom-up shift and reasoning for referring image segmentation. In CVPR, pages 11266–11275, 2021.
- [84] Zhengyuan Yang, Tianlang Chen, Liwei Wang, and Jiebo Luo. Improving one-stage visual grounding by recursive sub-query construction. In ECCV, pages 387–404. Springer, 2020.
- [85] Zhengyuan Yang, Boqing Gong, Liwei Wang, Wenbing Huang, Dong Yu, and Jiebo Luo. A fast and accurate one-stage approach to visual grounding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4683–4693, 2019.
- [86] Zichao Yang, Xiaodong He, Jianfeng Gao, Li Deng, and Alex Smola. Stacked Attention Networks for Image Question Answering. In CVPR, June 2016.
- [87] Linwei Ye, Zhi Liu, and Yang Wang. Dual convolutional lstm network for referring image segmentation. TMM, 22(12):3224–3235, 2020.
- [88] Linwei Ye, Mrigank Rochan, Zhi Liu, and Yang Wang. Cross-Modal Self-Attention Network for Referring Image Segmentation. CVPR, June 2019.
- [89] Licheng Yu, Zhe Lin, Xiaohui Shen, Jimei Yang, Xin Lu, Mohit Bansal, and Tamara L. Berg. MAttNet: Modular Attention Network for Referring Expression Comprehension. CVPR, June 2018.
- [90] Licheng Yu, Patrick Poirson, Shan Yang, Alexander C. Berg, and Tamara L. Berg. Modeling Context in Referring Expressions. Lecture Notes in Computer Science, pages 69–85, 2016.
- [91] Rowan Zellers, Yonatan Bisk, Ali Farhadi, and Yejin Choi. From Recognition to Cognition: Visual Commonsense Reasoning. In CVPR, June 2019.
- [92] Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In ECCV, pages 649–666, 2016.
- [93] Richard Zhang, Phillip Isola, Alexei A Efros, Eli Shechtman, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric. In CVPR, 2018.
- [94] Changqing Zou, Haoran Mo, Chengying Gao, Ruofei Du, and Hongbo Fu. Language-based colorization of scene sketches. TOG), 38(6):1–16, 2019. Publisher: ACM New York, NY, USA.
Appendix A Supplementary Material
This supplementary material contains the implementation details (Section A.1) and the complete ablation studies (Section A.2) of our work.
A.1 Implementation Details
Referring Expression Segmentation. Following previous work [47, 60, 88, 7], we limit the maximum length of expressions to . We set input image size to and for training and inference phase respectively. We use the first four layers of DeepLabv3+ with ResNet-101 backbone, pre-trained on COCO dataset by excluding images appear on the validation and the test sets of UNC, UNC+ and G-Ref datasets similar to previous work [53, 20, 89]. Thus, our low-level visual feature map has the size of in training, and in inference phase, both including 8-D location features. In all convolutional layers, we set the filter size, stride, and number of filters as , , and , respectively. The depth is in the multimodal encoder part of the network. For the bottom-up-only baseline, we used grouped convolution in the bottom-up branch to prevent linguistic information leakage to the top-down visual branch. We apply dropout regularization [71] to language representation with probability. We use Adam optimizer [42] with default parameters. We freeze the DeepLab-v3+ ResNet-101 weights. There are examples in each minibatch. We train our model for epochs on a Tesla V100 GPU with mixed precision and each epoch takes at most two hours depending on the dataset.
Language-guided Image Colorization. Unless otherwise speficied, we follow the same design choices applied for the referring expression segmentation task. We set the number of language-conditional filters as , replace the LSTM encoder with a BiLSTM encoder, and we use the first two layers of ResNet-101 trained on ImageNet as image encoder to have a similar model capacity and make a fair comparison with the previous work [57]. We set input image width and height to in both training and validation. Thus, the low-level visual feature map has the size of , and we don’t use location features. Additionally, in our experimental analysis, we consider the same design choices with previous work [57, 92]. Specifically, we use LAB color space, and our model predicts color values for all the pixels of the input image. We perform the class re-balancing procedure to obtain class weights for weighted cross entropy objective. We use classes present in ImageNet dataset, and encode color values to classes by assigning them to their nearest neighbors. We use input images with a size of , and output target images with a size of which is same with the previous work.
A.2 Ablation Studies
# | Top-down | Bottom-up | Depth | Layer | Visual | Textual | [email protected] | [email protected] | [email protected] | [email protected] | [email protected] | IoU |
---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | ✓ | 4 | Conv | ResNet-50 | LSTM | 66.40 | 58.59 | 49.35 | 36.01 | 13.42 | 58.06 | |
✓ | 4 | Conv | ResNet-50 | LSTM | 71.40 | 65.14 | 57.36 | 45.11 | 19.04 | 60.74 | ||
✓ | ✓ | 4 | Conv | ResNet-50 | LSTM | 75.12 | 70.08 | 63.32 | 50.50 | 22.29 | 63.59 | |
2 | ✓ | ✓ | 3 | Conv | ResNet-50 | LSTM | 69.96 | 63.13 | 55.04 | 41.33 | 15.98 | 60.23 |
✓ | ✓ | 4 | Conv | ResNet-50 | LSTM | 75.12 | 70.08 | 63.32 | 50.50 | 22.29 | 63.59 | |
✓ | ✓ | 5 | Conv | ResNet-50 | LSTM | 75.56 | 70.59 | 63.82 | 51.68 | 22.84 | 63.52 | |
3 | ✓ | ✓ | 4 | Conv | ResNet-50 | LSTM | 75.12 | 70.08 | 63.32 | 50.50 | 22.29 | 63.59 |
✓ | ✓ | 4 | FiLM | ResNet-50 | LSTM | 71.18 | 65.14 | 57.32 | 44.66 | 18.75 | 61.12 | |
4 | ✓ | ✓ | 4 | Conv | ResNet-50 | LSTM | 75.12 | 70.08 | 63.32 | 50.50 | 22.29 | 63.59 |
✓ | ✓ | 4 | Conv | ResNet-50 | BERT | 75.60 | 70.39 | 63.05 | 49.93 | 21.16 | 63.57 | |
5 | ✓ | ✓ | 4 | Conv | ResNet-50 | LSTM | 75.12 | 70.08 | 63.32 | 50.50 | 22.29 | 63.59 |
✓ | ✓ | 4 | Conv | ResNet-101 | LSTM | 76.67 | 71.77 | 64.76 | 51.69 | 22.73 | 64.63 |
We performed additional ablation experiments on referring expression segmentation task in order to understand the contributions of the remaining components of our model. We share results in Table A1. Each row stands for a different architectural setup. Horizontal lines separate the different ablation studies we performed, and first column denotes the ablation study group. Columns on the left determine these architectural setups. ✓on the Top-down column indicates that the corresponding setup modulates top-down visual branch with language, and similarly ✓on the Bottom-up column indicates that the corresponding setup modulates bottom-up visual branch with language. Depth indicates how many layers the multi-modal encoder has. Layer indicates the type of language-conditional layer used. Visual and Textual indicates which visual encoder and textual encoder used for the corresponding setup, respectively. The remaining columns stand for results.
Network Depth (2). We performed experiments by varying the depth size of the multi-modal encoder. We originally started with the depth size of 4. Increasing the depth size slightly increased the scores for some metrics, but more importantly, decreasing the depth size caused the model to perform worse than the bottom-up baseline. This happens because decreasing the depth size shrinks the receptive field of the network, and the model becomes less capable of drawing conclusions for the scenes that requires to be seen as a whole in order to fully understand.
FiLM vs. Language-conditional Filters (3). Another method for modulating language is using conditional batch normalization [15] or its successor, FiLM layers. When we replaced language-conditional filters with FiLM layers in our model, we observed 2.5 IoU decrease. This is natural, since the FiLM layer can be thought as grouped convolution with language-conditional filters, where the number of groups is equal to number of channels / filters.
LSTM vs. BERT as language encoder (4). We also experimented with BERT [19] as input language encoder in addition to LSTM network. We update BERT weights simultaneously with the rest of our model, where we use a smaller learning rate for BERT (). We use the CLS output embedding as our language representation , than split this embedding into pieces to create language-conditional filters. Our model achieved similar quantitative results using BERT as language encoder. This points out a language encoder pre-trained on solely textual data might be sub-optimal for integrating vision and language.
The impact of the visual backbone (5). We first start training our model with DeepLabv3+ ResNet-50 backbone pre-trained on Pascal VOC dataset. Then, we pre-trained a DeepLabv3+ with ResNet-101 backbone on COCO dataset by excluding the images appear on the validation and the test splits of all benchmarks similar to the previous work [89, 53, 20]. We only used 20 object categories present in Pascal VOC. Thus, using a more sophisticated visual backbone resulted with improvement on the IoU score.