Curriculum Learning for Vision-and-Language Navigation
Abstract
Vision-and-Language Navigation (VLN) is a task where an agent navigates in an embodied indoor environment under human instructions. Previous works ignore the distribution of sample difficulty and we argue that this potentially degrade their agent performance. To tackle this issue, we propose a novel curriculum-based training paradigm for VLN tasks that can balance human prior knowledge and agent learning progress about training samples. We develop the principle of curriculum design and re-arrange the benchmark Room-to-Room (R2R) dataset to make it suitable for curriculum training. Experiments show that our method is model-agnostic and can significantly improve the performance, the generalizability, and the training efficiency of current state-of-the-art navigation agents without increasing model complexity.
1 Introduction
Vision-and-Language (VLN) navigation task proposed recently by (Anderson et al., 2018) is a step towards building smart robots. It requires the agent to perceive the environment, understand human language instructions and finally unify the multi-modal information to make actions. Many state-of-the-art methods have been proposed. Some of them focus on the alignment between visual and textual inputs by improving model structure (Ma et al., 2019) or proposing novel auxiliary losses (Zhu et al., 2020) whereas others put their attention on the data augmentation (Fried et al., 2018; Tan et al., 2019; Hong et al., 2020; Ku et al., 2020). Large scale pre-training has also been employed to better generalize the downstream VLN tasks (Li et al., 2019; Majumdar et al., 2020; Hao et al., 2020; Huo et al., 2021).
Despite great progress previous works have made, very few of them care about how much the agent learns from the dataset, i.e. is the agent a good student? In computer vision, (Hlynsson et al., 2019) tries to answer this question by measuring the data efficiency — performance as a function of training set size — of deep learning methods. In vision-and-language navigation, (Huang et al., 2019) develops a discriminator that can filter the low quality instruction-path pairs to boost the learning efficiency. In this work, we focus on another aspect: can a VLN agent be further educated without model structure change and data modification? We observe that many works neglect the internal distribution of the sample difficulty. For example, a navigation task within a single room should be considered easier than one that needs to travel two or more rooms. Current training methods simply flush in the data and do not distinguish difficulty levels among the training samples. As shown in Figure 1(a), navigation agents trained by such learning process do not perform well on those "easy" tasks even on the previously seen environments. We monitor the first error an agent makes during the navigation and demonstrate the ratio of different types of these errors in Figure 1(b). We find that when the navigation agent fails, about 50% errors are caused by the agent wrongly predicts the next in-room direction. The ratio of this kind of error decreases as the navigation task spans more rooms, but still remains a relatively high level. These phenomenon indicates that the navigation agent is limited by its ability to navigate inside one room and cross two rooms.


The poor performance of agents on those easy cases inspires us to borrow the idea from curriculum learning (Bengio et al., 2009) and propose a curriculum-based training paradigm. The basic idea of curriculum learning is to start small, learn easier aspects of the task and then gradually increase the difficulty level. The definition of "difficulty" is therefore worthwhile. We hypothesise that the difficulty of a navigation task is closely related to the rooms the agent needs to pass by towards the destination. On top of this, we introduce the curriculum design for VLN and re-arrange the benchmark Room-to-Room (R2R) dataset (Anderson et al., 2018) to make it suitable for curriculum learning (Section 3). We incorporate human prior knowledge about training samples into the training process of a navigation agent via self-paced curriculum learning (SPCL) (Jiang et al., 2015). We adapt the traditional SPCL algorithm to efficiently train deep learning models (Section 4). Experiments show that our method can consistently improve both the navigation performance and the training efficiency of navigation agents (Section 5).
In summary, our main contributions are:
-
•
We propose to explicitly incorporate human prior knowledge about training samples into navigation agent training process.
-
•
We design the curriculum for VLN and put forward the training paradigm of a navigation agent by curriculum learning without increasing the complexity of the model.
-
•
We empirically validate that the role of curriculum learning is to smooth loss landscape and hence find a better local optima.
2 Related Work
Vision-and-Language Navigation.
Vision-and-language Navigation (VLN) is a task where an agent navigates to a goal location in a photo-realistic 3D environment under human instructions. (Anderson et al., 2018) formalized this task, proposed the benchmark Room-to-Room (R2R) dataset and set up an attention-based sequence-to-sequence baseline model. Other related VLN datasets include Touchdown dataset (Chen et al., 2019), which is the first large-scale outdoor VLN dataset, and CVDN dataset (Thomason et al., 2019) that emphasizes on the robot-human dialogues during navigation.
Embodied navigation tasks suffer from a limited size of training data, which results in several different research focuses. For data augmentation, (Fried et al., 2018) developed a speaker-follower model where the speaker model is usually used as a tool for instruction generation, and (Tan et al., 2019) proposed to apply a effective environmental dropout layer to mimic unseen environments during training. For better generalization, (Wang et al., 2018) introduced a hybrid model that integrated the model-based and model-free reinforcement learning whereas (Wang et al., 2019) proposed a Reinforced Cross-Modal (RCM) agent and a self-supervised imitation learning method to explore the unseen environment. Besides, (Li et al., 2019; Huo et al., 2021; Hao et al., 2020) applied pre-training skills towards better instruction understanding and generalization.
Another resolution to the paucity of training data is rather straightforward: generating more training data and annotations. For example, (Hong et al., 2020) split the instructions in R2R dataset into sub-instructions and annotated the corresponding sub-paths. (Ku et al., 2020) proposed Room Across Room (RxR) dataset, a multilingual VLN dataset with dense spatiotemporal grounding. Furthermore, like (Wang et al., 2020) one can use multitask learning framework to grasp common knowledge from other homologous datasets so as to better the agent’s performance on the current task.
These works are worthwhile but we observe that most of them ignore how much the model can learn from the dataset. In this paper, we focus to improve training methods to better exploit data. A better training paradigm can benefit both previous and future works on VLN tasks.
Curriculum Learning.
Curriculum Learning (Bengio et al., 2009) is a learning paradigm that mimic the learning principle underlying the cognitive process of humans and animals, where a model is learned by gradually including from easy to complex samples in training. (Jiang et al., 2015) bridged the gap between curriculum learning and self-paced learning by proposing a novel self-paced curriculum learning method. (Graves et al., 2017) proposed an automated curriculum learning method that can automatically select the curriculum syllabus. Later, (Matiisen et al., 2017) introduced a teacher-student framework for automatic curriculum learning, where the student tries to learn a complex task and the teacher automatically chooses the task for student to train on.
Curriculum learning has been applied to several natural language processing tasks, such as question answering (Sachan and Xing, 2016; Liu et al., 2018), machine translation (Platanios et al., 2019) and so on. An empirical work done by (Hacohen and Weinshall, 2019) conclude that curriculum learning can effectively modify the optimization landscape and under mild conditions it does not change the corresponding global minimum of the optimization function. These works encourage us to apply the curriculum learning into the VLN tasks as it does not need any modification towards the navigation agent and is capable of improving the agent performance. To the best of our knowledge, this is the first work that introduces curriculum learning purely as a training paradigm for VLN tasks.
Babywalk (Zhu et al., 2020) is another work that applies the curriculum learning idea. It aims to enhance the transfer ability of navigation agents. Therefore, Babywalk uses dynamic programming to decompose the long instructions into shorter ones and then applies these sub-instructions for curriculum-based reinforcement learning. It does not pay much attention to how an agent performs within a dataset. Babywalk is a carefully designed VLN agent with complex model structure whereas our method is a easy, extendable model-agnostic training method. Our method will change neither the difficulty of training data nor the model complexity.

3 Curriculum Design for VLN
We observe that different samples in the dataset have different navigation difficulties. Our intuition is that, for human beings it is easy to find an object or a place within a small range. After exploration, it is natural for us to exploit the knowledge about the environment and complete the harder cross-room tasks. Hence, we hypothese that the number of rooms a path could cover (namely room length, ) dominates the difficulty level of a navigation task. We therefore propose to re-arrange the R2R dataset based on . Two examples with various are shown in Figure 2.
Train Set | Paths | Instructions | Average Trajectory Length (m) | Average Instruction Length (words) | Start Coverage (%) | Room Coverage (%) |
R2R | 4675 | 14039 | 9.91 | 29.41 | 89.5 | 96.1 |
CLR2R | ||||||
- Round 1 | 345 | 1037 | 8.86 | 24.31 | 12.0 | 12.8 |
- Round 2 | 471 | 1415 | 8.25 | 24.78 | 20.2 | 38.3 |
- Round 3 | 1632 | 4897 | 9.19 | 28.08 | 59.5 | 79.1 |
- Round 4 | 1530 | 4593 | 10.42 | 31.16 | 59.5 | 83.0 |
- Round 5 | 697 | 2097 | 12.12 | 34.33 | 33.5 | 62.2 |
We firstly investigate the distributions of rooms that can be covered by a R2R path. Approximately 17% of the paths only pass through less than two rooms whereas about 15% of the paths span 5 rooms or more. We then propose to split the train set of R2R dataset into 5 mutually exclusive subsets and each subset contains paths whose are constrained. We believe that the learning from simple to difficult is very similar to play arcade games, so the subsets are named round 1 to 5 according to its difficulty. To be specific, round () subset contains samples whose ground truth path covers rooms. Round 5 subset contains the left samples.
The basic statistics of the re-arranged R2R dataset is summarized in Table 1. As the construction of such dataset incorporate human prior knowledge and such dataset will be futher used for curriculum learning, we call it Room-to-Room for Curriculum Learning (CLR2R) dataset.
It should be noted that the definition of "difficulty" is open and task-specific, we throw a possibility here and encourage other attempts. Similar standards are easy to find in other datasets, for example, the number of dialogue rounds and the richness of dialogue phenomenon can both be taken as difficulty indicators on CVDN dataset(Thomason et al., 2019). Finally, we leave validation and test set of R2R dataset unchanged so as to have comparable experimental results.
4 Methods
There are multiple settings of curriculum learning. When a series of tasks is presented, we can apply automated curriculum learning (Graves et al., 2017) where agents are trained by tasks adapted to their capacities. While for single task setting, self-paced curriculum learning (Jiang et al., 2015) is more appropriate if the task dataset has samples with various difficulty levels. Our observations better fit the latter mode, hence we adopt self-paced curriculum learning as the training method.
4.1 General Setup
Self-paced curriculum learning (SPCL) is an “instructor-student collaborative” learning method as it considers both prior knowledge known before training and information learning during training in a unified framework. To be specific, the objective loss function of SPCL is defined as
(1) | ||||
where denotes the navigation agent, is the weight variable reflecting the sample importance. is called the self-paced function which controls the learning scheme, and is a hyper-parameter that limits the learning pace. is a feasible region that encodes the information of a predetermined curriculum, defined below.
Definition 4.1
For training samples , given a curriculum defined on it, the feasible region, defined by,
is a curriculum region of if it holds: 1) is nonempty; 2) for all for all
where vector that parameterizes a linear space is a function of curriculum . It is not hard to see that, prior human knowledge influence the feasible region which constrains the weight parameter . Hence, the update of is subject to the learning progress , self-paced function and feasible region . Based on our curriculum design, samples within each round should be given the same curriculum rank. Therefore, 5 scalars is enough to define a good vector .
Compared with self-paced learning (SPL) (Kumar et al., 2010), SPCL is more general by introducing self-paced function as a flexible regularization term. Self-paced functions are defined to be convex, which ensures that we can find good solutions of within the linear curriculum region. Following (Jiang et al., 2014)’s work, (Jiang et al., 2015) further discussed several examples of self-paced functions. In this paper, we focus on the simplest two: (1) Binary scheme and (2) Linear scheme. As summarized in Table 2, the convex optimization problem
(2) |
have the closed form solution if . Therefore, to solve the linear constrained convex optimization problem, we can simply apply a projection gradient descent method to obtain the optimal weight .
Function Form | Closed Form Optimal Solution | |
Binary Scheme | = | |
Linear Scheme | = |
4.2 Algorithm
(Jiang et al., 2015) proposed an alternative convex search (ACS) (M. Bazaraa and Shetty, 1993) algorithm to solve Equation 1. The main problem of the original algorithm is at step 4, where the optimal model parameter is learned with most recent weight vector fixed. In the training of navigation agents, it is impossible to compute the exact optimum of due to both the time consumption and the lack of global convergence guarantees. We propose not to compute the exact minimum but replace step 4 with several gradient descent update steps. Doing so benefits the algorithm as the training is much faster then before.
According to (Gorski et al., 2007), the original SPCL algorithm proposed by (Jiang et al., 2015) is guaranteed to converge globally as the objective function is monotonically decreasing and bounded below. Our algorithm does not have the monotonically decreasing property as the stochastic gradient descent (SGD) update can not assure the continuous decrease of the function value. Alternatively, our algorithm can be viewed as a naive version of mini-batch randomized block coordinate descent (MRBCD) method. (Zhao et al., 2014) claimed that MRBCD using semi-stochastic optimization scheme can attain linear rates of convergence, while the convergence of naive MRBCD is still left to be discussed. Further experiments show that our algorithm is empirically converged, which encourages the theoretical analysis in this direction.
Algorithm 1 takes the input of a predetermined curriculum , an instantiated self-paced function , a stepsize parameter and an update interval ; it outputs an optimal model parameter .
5 Experiment Results and Analysis
5.1 Experiment Setup
Navigation agents.
We experiment with three state-of-the-art VLN agents using different training paradigms and compare their performance on the original R2R validation set. We choose several state-of-the-art navigation agents including the Speaker-Follower agent (Fried et al., 2018) which applies the panoramic action space, the Self-Monitoring agent (Ma et al., 2019) which enforces cross-model alignment and EnvDrop with Back Translation (Tan et al., 2019) that applies reinforcement learning. We reproduce these agents in a unified code framework and replicate the experimental results presented in original papers. We do not use any data augmentation tricks and all three agents are trained within the R2R dataset range.
Implementation details.
We experiment with three training paradigms. They include the traditional machine learning (ML) strategy that training by uniformly sample mini-batches from R2R dataset, a naive curriculum learning (NCL) strategy that the agent is firstly trained on CLR2R round 1 split, then round 1~2 splits, and gradually trained on the whole CLR2R train set, and previously introduced self-paced curriculum learning (SPCL) strategy. As discussed in Section 4, we set for samples in CLR2R round split. The constant is chosen within range . For self-paced functions, we choose the binary and linear scheme. For EnvDrop agent (Tan et al., 2019) which is trained by a mixed loss of imitation learning and reinforcement learning, we only use the ground truth trajectory-based loss to update the weight variable. More information is available in Appendix A.1.
Evaluation metrics.
We follow the standard metrics that (Fried et al., 2018) employed for evaluating the agent’s performance, including average Navigation Error (NE) which measures the distance between the end location and the target location, Success Rate (SR) which is the percentage of predicted end location within 3m of the target location, and Oracle Success Rate (OSR) which is the percentage of trajectory the shortest distance between which and the target location is within 3m. We also adopt success weighted by path length (SPL) as recommended by (Anderson et al., 2018). Furthermore, we consider the coverage weighted by length score (CLS) (Jain et al., 2019) which measures the overall correspondence between the predicted and ground truth trajectories. Actually, nDTW and SDTW proposed by (Magalhães et al., 2019) are both useful metrics that capture path fidelity, hence we supplement a full-metric version of main results in Appendix A.2.
5.2 Overall Performance
Model | Validation Seen | Validation Unseen | ||||||||
NE ↓ (m) | SR (%) | OSR (%) | SPL (%) | CLS | NE ↓ (m) | SR (%) | OSR (%) | SPL (%) | CLS | |
Follower | 4.85 | 52.3 | 65.3 | 44.3 | 58.2 | 7.12 | 28.6 | 40.9 | 20.3 | 35.0 |
+ Naïve CL | 5.03 | 48.6 | 62.0 | 40.4 | 55.9 | 7.13 | 31.1 | 42.8 | 21.2 | 34.3 |
+ Self-Paced CL | 4.23 | 58.7 | 69.2 | 51.1 | 63.3 | 6.69 | 32.2 | 44.2 | 24.5 | 38.9 |
Self-Monitoring | 4.27 | 58.4 | 67.0 | 51.9 | 64.1 | 6.42 | 38.4 | 48.1 | 28.3 | 41.5 |
+ Naïve CL | 4.08 | 61.0 | 69.8 | 54.6 | 64.9 | 6.30 | 40.0 | 51.7 | 28.6 | 41.1 |
+ Self-Paced CL | 4.19 | 58.8 | 68.2 | 53.3 | 65.4 | 5.98 | 41.0 | 52.4 | 30.8 | 43.9 |
EnvDrop | 4.55 | 57.7 | 65.6 | 54.4 | 67.2 | 5.92 | 45.7 | 54.2 | 41.8 | 57.0 |
+ Naïve CL | 4.49 | 57.8 | 63.1 | 54.8 | 67.2 | 5.93 | 44.3 | 50.5 | 41.3 | 57.6 |
+ Self-Paced CL | 4.42 | 58.1 | 65.6 | 54.8 | 67.4 | 5.48 | 47.6 | 54.3 | 44.1 | 59.1 |
We first compare the performance of the different training paradigms on the CLR2R(R2R) validation set. Table 3 shows overall results. Experiments show that naïve curriculum learning is surprisingly effective. It improves the navigation outcomes of Follower and Self-Monitoring agent on validation unseen split. In contrast to machine learning where the whole training set is available, naïve curriculum learning exposes the training samples sequentially from easy to hard. Hence, follower agent trained by NCL is underperformed on the validation seen split due to its limited learning ability. For self-paced curriculum learning, we initialize the algorithm to pay more attention on round 1 and 2 splits, then allow it actively assigns higher weight for examples which the agent learns better. During training, the weight for each sample can finally converge nearly to one. Experiments show that agents trained by SPCL consistently achieve good performance on both validation seen and unseen split.

Furthermore, we compare the learning speed of agents trained by different training paradigms. As shown in Figure 3, there is obviously a gap between agents trained by machine learning and self-paced curriculum learning. Agents trained by curriculum learning can learn fast and achieve better performance after the same number of iterations. This indicates that self-paced curriculum learning can not only improve the performance, but also optimize the training efficiency of the agents.

5.3 Further Analysis
We also investigate the hyperparameter robustness of self-paced curriculum learning. Moreover, we present the transfer learning outcomes of curriculum learning method on CLR2R and RxR dataset.
Hyperparameter robustness.
To understand how the weight initialization and stepsize choice influence the training paradigm, we grid search these two hyperparameters and the result is shown in Figure 4. Following the core idea of curriculum learning by starting small, we enforce the initial weight for round 1 and 2 samples as one. Then represents the initial weight for samples in round 3~5 splits. As shown in Figure 4, SPCL favors a smaller . This validates the principle of curriculum learning. Besides, SPCL is rather robust to weight initialization and the choice of stepsize as in most cases, the results in validation unseen split are better than the machine learning baseline. This also indicates a better generalizability of agents trained by SPCL. For validation seen split, the performance is usually no worse than the baseline. The overall outcomes prove that curriculum learning can improve the data exploitation efficiency and enhance the generalization ability.

Loss landscape.
Following (Santurkar et al., 2018), we investigate the loss landscape during training by computing the distance between maximum and minimum loss. As shown in Figure 5, SPCL narrows the loss gap for Follower and Self-Monitoring agent. For EnvDrop agent, SPCL differs little from ML. We attribute this to the mixed learning strategy of EnvDrop, whose loss is a weighted sum of imitation learning(0.2) and reinforcement learning(0.8). Hence, the power of SPCL is not as significant as in other agents. In general, our experimental results correspond to the theoretical result (Hacohen and Weinshall, 2019) that curriculum learning can effectively smooth the optimization landscape.
Transfer learning.
To explore whether curriculum learning helps the agent to obtain better transferring ability, we conduct the following experiment as summarized in Table 5. As mentioned in Section 2, RxR dataset (Ku et al., 2020) is a large-scale multilingual VLN dataset built upon Matterport3D simulator (Chang et al., 2017). We select the RxR dataset as the target dataset of transfer learning since it shares the same environment with R2R dataset but is much harder in many aspects (see Table 4). As RxR dataset is multilingual, we employ its english subset, RxR-en, to avoid the language bias. For fairly comparison, the number of training iterations is fixed at 80,000 for line (3)~(5) in table. Specifically, for line (3) the agent is trained on a integrated R2R+RxR-en dataset for 80,000 iterations; for line (4), the agent is firstly trained on R2R trainset for 40,000 iterations and then transferred to RxR-en dataset for another 40,000 iterations; for line (5), the agent is firstly trained on R2R trainset and then trained on the integrated R2R+RxR-en dataset for another 40,000 iterations.
Line (1) and (2) indicates that transfer learning (R2RRxR and vice-versa) performance is much weaker than the in-domain results. For line (3), the agent trained on an integrated dataset performs better on RxR-en validation split, but its performance on R2R validation split it not as good as line (1). The difficulty gap between samples from two data source confuses the agent. Line (3) and (4) indicates that starting from simple improves the agents performance on RxR dataset. Compared with line (4), curriculum learning in line (5) helps the agent to preserve its navigation ability on R2R dataset. Actually, line (4) can also be taken as a curriculum learning experiment whose target distribution is RxR dataset instead of R2R and RxR dataset.
This exploration experiment suggests that to transfer from an easier dataset towards a harder one, curriculum learning paradigm helps the agent to improve the performance on the harder dataset without degrading the performance on the easier dataset.
Dataset | Average Path Length (m) | Average Edges | Average Instruction Length (words) |
R2R (Anderson et al., 2018) | 9.4 | 5 | 29 |
RxR (Ku et al., 2020) | 14.9 | 8 | 78 |
Train Set | Validation Unseen | |||||||||||
NE(m) | SR(%) | OSR(%) | SPL(%) | nDTW | ||||||||
R2R | RxR | R2R | RxR | R2R | RxR | R2R | RxR | R2R | RxR | R2R | RxR | |
(1) | ✓ | 6.42 | 10.39 | 38.4 | 16.0 | 48.1 | 29.9 | 28.3 | 9.1 | 43.7 | 28.0 | |
(2) | ✓ | 7.72 | 10.12 | 16.7 | 18.7 | 30.4 | 32.8 | 5.9 | 6.9 | 18.6 | 18.2 | |
(3) | ✓ | ✓ | 6.66 | 10.17 | 32.7 | 20.5 | 45.8 | 33.7 | 22.4 | 10.9 | 36.4 | 23.9 |
(4) | F | S | 6.96 | 9.91 | 26.4 | 22.1 | 38.6 | 34.8 | 18.0 | 14.3 | 36.8 | 29.8 |
(5) | Naïve CL | 6.46 | 9.90 | 36.3 | 21.6 | 44.4 | 33.4 | 27.4 | 13.5 | 43.0 | 29.3 |
Combination with Pre-training.
Pre-training methods are proved strong in many areas. We believe that navigation agents trained by our proposed curriculum training paradigm can also benefit from pre-training skills. That is, curriculum learning does not conflict with pre-training. To validate this, we combine agents trained by curriculum based methods with VLN-BERT (Majumdar et al., 2020), a visiolinguistic transformer-based model. VLN-BERT is a scoring function that evaluates the compatibility between path-instruction pairs, hence it can be easily applied to assist our navigation agents. Since VLN-BERT has to be used under beam search settings , we implement beam search only here. In (Majumdar et al., 2020), authors use a beam size as 30 to include as much path candidates as possible. While in this experiment, we aim to evaluate the usefulnees of combining curriculum and pre-training. Therefore, we restrict the beam size as 5 and purely use VLN-BERT model to scoring and selecting the path-instruction pairs. We experiment with Follower agent under three different test modes, i.e. single-run, beam search with size 1, and vln-bert. Results are shown in Figure 6. Beam search and VLN-BERT scorer both can improve the agent performance. Navigation agents trained by curriculum-based methods obtain more improvements.

6 Conclusion and Future Works
We propose to apply curriculum learning as a useful training paradigm of VLN agents. We adapt the traditional self-paced curriculum learning and design the first curriculum for VLN task based on R2R dataset. Experiments illustrate that our method is model-agnostic and can improve both the performance and learning efficiency of the navigation agents. We verify the power of curriculum learning comes from its ability to smooth the loss landscape. Our further exploration indicates that curriculum learning is suitable for transfer learning and combination with pre-training methods.
In the future, we would like to explore the VLN task in two directions. Along the path of this paper, we would like to explore curriculum learning for vision-language navigation in a different scenario of city traveling where the agent needs to perform multi-scale navigation, i.e. city-level and building-level. Alternatively, inspired by the ampleness of VLN tasks and datasets, i.e. R2R (Anderson et al., 2018), CVDN (Thomason et al., 2019), HANNA (Nguyen and Daumé, 2019), TOUCHDOWN (Chen et al., 2019), REVERIE (Qi et al., 2020) and so on, it is possible to apply meta-learning on top of these datasets and get an adaptable navigation agent.
Acknowledgements
This work is partially supported by Natural Science Foundation of China (No.71991471, No.6217020551), Science and Technology Commission of Shanghai Municipality Grant (No.20dz1200600, 21QA1400600) and Zhejiang Lab (No. 2019KD0AD01).
References
- Anderson et al. (2018) P. Anderson, Q. Wu, D. Teney, J. Bruce, M. Johnson, N. Sünderhauf, I. Reid, S. Gould, A. V. D. Hengel, Vision-and-Language Navigation: Interpreting Visually-Grounded Navigation Instructions in Real Environments, in: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, pp. 3674–3683.
- Ma et al. (2019) C.-Y. Ma, J. Lu, Z. Wu, G. Al-Regib, Z. Kira, R. Socher, C. Xiong, Self-monitoring navigation agent via auxiliary progress estimation, ArXiv abs/1901.03035 (2019).
- Zhu et al. (2020) F. Zhu, Y. Zhu, X. Chang, X. Liang, Vision-language navigation with self-supervised auxiliary reasoning tasks, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020) 10009–10019.
- Fried et al. (2018) D. Fried, R. Hu, V. Cirik, A. Rohrbach, J. Andreas, L.-P. Morency, T. Berg-Kirkpatrick, K. Saenko, D. Klein, T. Darrell, Speaker-follower models for vision-and-language navigation, in: NeurIPS, 2018.
- Tan et al. (2019) H. Tan, L. Yu, M. Bansal, Learning to navigate unseen environments: Back translation with environmental dropout, ArXiv abs/1904.04195 (2019).
- Hong et al. (2020) Y. Hong, C. Rodriguez-Opazo, Q. Wu, S. Gould, Sub-instruction aware vision-and-language navigation, in: EMNLP, 2020.
- Ku et al. (2020) A. Ku, P. Anderson, R. Patel, E. Ie, J. Baldridge, Room-across-room: Multilingual vision-and-language navigation with dense spatiotemporal grounding, in: EMNLP, 2020.
- Li et al. (2019) X. Li, C. Li, Q. Xia, Y. Bisk, A. Çelikyilmaz, J. Gao, N. A. Smith, Y. Choi, Robust navigation with language pretraining and stochastic sampling, in: EMNLP, 2019.
- Majumdar et al. (2020) A. Majumdar, A. Shrivastava, S. Lee, P. Anderson, D. Parikh, D. Batra, Improving vision-and-language navigation with image-text pairs from the web, ArXiv abs/2004.14973 (2020).
- Hao et al. (2020) W. Hao, C. Li, X. Li, L. Carin, J. Gao, Towards learning a generic agent for vision-and-language navigation via pre-training, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020) 13134–13143.
- Huo et al. (2021) Y. Huo, M. Zhang, G. Liu, H. Lu, Y. Gao, G. xing Yang, J. Wen, H. Zhang, B. Xu, W. Zheng, Z. Xi, Y. Yang, A. Hu, J. Zhao, R. Li, Y. Zhao, L. Zhang, Y. Song, X. Hong, W. Cui, D. Hou, Y. Li, J. Li, P. Liu, Z. Gong, C. Jin, Y. Sun, S. Chen, Z. Lu, Z. Dou, Q. Jin, Y. Lan, W. X. Zhao, R. Song, J. Wen, Wenlan: Bridging vision and language by large-scale multi-modal pre-training, ArXiv abs/2103.06561 (2021).
- Hlynsson et al. (2019) H. D. Hlynsson, A. N. Escalante, L. Wiskott, Measuring the data efficiency of deep learning methods, in: ICPRAM, 2019.
- Huang et al. (2019) H. Huang, V. Jain, H. Mehta, J. Baldridge, E. Ie, Multi-modal discriminative model for vision-and-language navigation, ArXiv abs/1905.13358 (2019).
- Bengio et al. (2009) Y. Bengio, J. Louradour, R. Collobert, J. Weston, Curriculum learning, in: ICML ’09, 2009.
- Jiang et al. (2015) L. Jiang, D. Meng, Q. Zhao, S. Shan, A. Hauptmann, Self-paced curriculum learning, in: AAAI, 2015.
- Chen et al. (2019) H. Chen, A. Suhr, D. K. Misra, N. Snavely, Y. Artzi, Touchdown: Natural language navigation and spatial reasoning in visual street environments, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019) 12530–12539.
- Thomason et al. (2019) J. Thomason, M. Murray, M. Cakmak, L. Zettlemoyer, Vision-and-dialog navigation, in: CoRL, 2019.
- Wang et al. (2018) X. E. Wang, W. Xiong, H. Wang, W. Y. Wang, Look before you leap: Bridging model-free and model-based reinforcement learning for planned-ahead vision-and-language navigation, in: ECCV, 2018.
- Wang et al. (2019) X. E. Wang, Q. Huang, A. Çelikyilmaz, J. Gao, D. Shen, Y. Wang, W. Y. Wang, L. Zhang, Reinforced cross-modal matching and self-supervised imitation learning for vision-language navigation, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019) 6622–6631.
- Wang et al. (2020) X. Wang, V. Jain, E. Ie, W. Y. Wang, Z. Kozareva, S. Ravi, Environment-agnostic multitask learning for natural language grounded navigation, ArXiv abs/2003.00443 (2020).
- Graves et al. (2017) A. Graves, M. G. Bellemare, J. Menick, R. Munos, K. Kavukcuoglu, Automated curriculum learning for neural networks, ArXiv abs/1704.03003 (2017).
- Matiisen et al. (2017) T. Matiisen, A. Oliver, T. Cohen, J. Schulman, Teacher-student curriculum learning, CoRR abs/1707.00183 (2017).
- Sachan and Xing (2016) M. Sachan, E. Xing, Easy questions first? a case study on curriculum learning for question answering, in: ACL, 2016.
- Liu et al. (2018) C. Liu, S. He, K. Liu, J. Zhao, Curriculum learning for natural answer generation, in: IJCAI, 2018.
- Platanios et al. (2019) E. A. Platanios, O. Stretcu, G. Neubig, B. Poczos, T. M. Mitchell, Competence-based curriculum learning for neural machine translation, arXiv preprint arXiv:1903.09848 (2019).
- Hacohen and Weinshall (2019) G. Hacohen, D. Weinshall, On the power of curriculum learning in training deep networks, ArXiv abs/1904.03626 (2019).
- Zhu et al. (2020) W. Zhu, H. Hu, J. Chen, Z. Deng, V. Jain, E. Ie, F. Sha, Babywalk: Going farther in vision-and-language navigation by taking baby steps, in: ACL, 2020.
- Kumar et al. (2010) M. Kumar, B. Packer, D. Koller, Self-paced learning for latent variable models, in: NIPS, 2010.
- Jiang et al. (2014) L. Jiang, D. Meng, T. Mitamura, A. Hauptmann, Easy samples first: Self-paced reranking for zero-example multimedia search, Proceedings of the 22nd ACM international conference on Multimedia (2014).
- M. Bazaraa and Shetty (1993) H. S. M. Bazaraa, C. Shetty, Nonlinear programming: Theory and algorithms, in: John Wiley and Sons, Inc., 1993.
- Gorski et al. (2007) J. Gorski, F. Pfeuffer, K. Klamroth, Biconvex sets and optimization with biconvex functions: a survey and extensions, Mathematical Methods of Operations Research 66 (2007) 373–407.
- Zhao et al. (2014) T. Zhao, M. Yu, Y. Wang, R. Arora, H. Liu, Accelerated mini-batch randomized block coordinate descent method, Advances in neural information processing systems 27 (2014) 5614.
- Anderson et al. (2018) P. Anderson, A. X. Chang, D. S. Chaplot, A. Dosovitskiy, S. Gupta, V. Koltun, J. Kosecka, J. Malik, R. Mottaghi, M. Savva, A. Zamir, On evaluation of embodied navigation agents, ArXiv abs/1807.06757 (2018).
- Jain et al. (2019) V. Jain, G. Magalhães, A. Ku, A. Vaswani, E. Ie, J. Baldridge, Stay on the path: Instruction fidelity in vision-and-language navigation, in: ACL, 2019.
- Magalhães et al. (2019) G. Magalhães, V. Jain, A. Ku, E. Ie, J. Baldridge, Effective and general evaluation for instruction conditioned navigation using dynamic time warping, ArXiv (2019).
- Santurkar et al. (2018) S. Santurkar, D. Tsipras, A. Ilyas, A. Madry, How does batch normalization help optimization?, in: NeurIPS, 2018.
- Chang et al. (2017) A. X. Chang, A. Dai, T. Funkhouser, M. Halber, M. Nießner, M. Savva, S. Song, A. Zeng, Y. Zhang, Matterport3d: Learning from rgb-d data in indoor environments, 2017 International Conference on 3D Vision (3DV) (2017) 667–676.
- Nguyen and Daumé (2019) K. Nguyen, H. Daumé, Help, anna! visual navigation with natural multimodal assistance via retrospective curiosity-encouraging imitation learning, in: ViGIL@NeurIPS, 2019.
- Qi et al. (2020) Y. Qi, Q. Wu, P. Anderson, X. Wang, W. Wang, C. Shen, A. V. Hengel, Reverie: Remote embodied visual referring expression in real indoor environments, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020) 9979–9988.
Appendix A Appendix
A.1 Training Details
We train all agents on a single NVIDIA GeForce RTX2080 GPU, using the same model hyperparameters as the officially released codes, except the language encoder of Follower agent (Fried et al., 2018) where we enforce a bidirectional LSTM module and we do not use GloVe embeddings as initialization. For fairly comparison, basic training hyperparameters are fixed, i.e. the maximum training epoch is 200. We sample mini-batches with size 64 for 200 iterations per epoch. The learning rate is a constant and is fixed at .
With regard to self-paced curriculum learning, we choose linear scheme with for Follower agent, binary scheme with for Self-Monitoring agent, and linear scheme with for EnvDrop agent. Value is initialized as 2 for all agents. It is updated by if it is smaller than currently maximum item loss and is updated by half of otherwise.
Model | Validation Seen | ||||||
NE ↓ | SR ↑ | OSR ↑ | SPL ↑ | nDTW ↑ | SDTW ↑ | CLS ↑ | |
Follower | 4.85 | 52.3 | 65.3 | 44.3 | 59.4 | 44.1 | 58.2 |
+ Naïve CL | 5.03 | 48.6 | 62.0 | 40.4 | 57.4 | 40.5 | 55.9 |
+ Self-Paced CL | 4.23 | 58.7 | 69.2 | 51.1 | 64.5 | 50.5 | 63.3 |
Self – Monitoring | 4.27 | 58.4 | 67.0 | 51.9 | 65.4 | 51.5 | 64.1 |
+ Naïve CL | 4.08 | 61.0 | 69.8 | 54.6 | 66.2 | 53.9 | 64.9 |
+ Self-Paced CL | 4.19 | 58.8 | 68.2 | 53.3 | 66.3 | 52.2 | 65.4 |
EnvDrop | 4.55 | 57.7 | 65.6 | 54.4 | 67.2 | 51.4 | 67.2 |
+ Naïve CL | 4.49 | 57.8 | 63.1 | 54.8 | 67.5 | 51.5 | 67.2 |
+ Self-Paced CL | 4.42 | 58.1 | 65.6 | 54.8 | 67.7 | 51.6 | 67.4 |
Model | Validation Unseen | ||||||
NE ↓ | SR ↑ | OSR ↑ | SPL ↑ | nDTW ↑ | SDTW ↑ | CLS ↑ | |
Follower | 7.12 | 28.6 | 40.9 | 20.3 | 37.0 | 20.8 | 35.0 |
+ Naïve CL | 7.13 | 31.1 | 42.8 | 21.2 | 36.9 | 22.1 | 34.3 |
+ Self-Paced CL | 6.69 | 32.2 | 44.2 | 24.5 | 40.9 | 24.5 | 38.9 |
Self – Monitoring | 6.42 | 38.4 | 48.1 | 28.3 | 43.7 | 28.8 | 41.5 |
+ Naïve CL | 6.30 | 40.0 | 51.7 | 28.6 | 43.4 | 29.4 | 41.1 |
+ Self-Paced CL | 5.98 | 41.0 | 52.4 | 30.8 | 46.0 | 31.0 | 43.9 |
EnvDrop | 5.92 | 45.7 | 54.2 | 41.8 | 56.7 | 39.3 | 57.0 |
+ Naïve CL | 5.93 | 44.3 | 50.5 | 41.3 | 57.3 | 38.3 | 57.6 |
+ Self-Paced CL | 5.48 | 47.6 | 54.3 | 44.1 | 59.3 | 41.2 | 59.1 |
Model | Test | |||
NE ↓ | SR ↑ | OSR ↑ | SPL ↑ | |
Follower | 7.05 | 29.0 | 41.3 | 20.7 |
+ Naïve CL | 7.16 | 28.3 | 40.9 | 19.6 |
+ Self-Paced CL | 6.95 | 30.9 | 42.3 | 24.3 |
Self – Monitoring | 6.29 | 40.5 | 50.2 | 30.9 |
+ Naïve CL | 6.29 | 40.8 | 53.0 | 30.6 |
+ Self-Paced CL | 6.29 | 39.3 | 49.9 | 30.8 |
EnvDrop | 5.71 | 46.5 | 54.2 | 43.5 |
+ Naïve CL | 5.90 | 44.8 | 50.0 | 42.5 |
+ Self-Paced CL | 5.41 | 48.4 | 53.9 | 45.5 |
A.2 Full Metric Results
As coverage weighted by length score (CLS) (Jain et al., 2019) which measures the overall correspondence between the predicted and ground truth trajectories, other metrics like normalized dynamic timewarping (nDTW) and success weighted by normalized dynamic time warping (SDTW) can also capture the path fidelity. Therefore, to give readers a general picture, we supplement a full-metric version here on validation set (see Table 6).
Besides, we also supplement the results on test set. However, due to the limit of EVAL platform, results on test set only have 4 evaluation metrics and hence is not presented in paper.
A.3 Exploration Experiments
Randomness Check
Except for the main results reported in paper, we repeat the experiments for 4 more times and every time we use the same training settings except the random seed. Table 7 summarizes our results on validation unseen split. The agent’s performance is consistent with paper.
Model | Validation Unseen | ||||
NE ↓ | SR ↑ | OSR ↑ | SPL ↑ | nDTW ↑ | |
Follower | 6.98 (0.14) | 29.71 (1.68) | 40.75 (1.45) | 20.66 (1.46) | 37.10 (1.15) |
+ Naïve CL | 7.03 (0.07) | 29.95 (0.84) | 42.92 (1.34) | 20.01 (0.99) | 36.52 (0.81) |
+ Self-Paced CL | 6.75 (0.11) | 32.08 (0.77) | 40.39 (1.16) | 23.83 (0.86) | 40.39 (1.10) |
Self – Monitoring | 6.35 (0.07) | 38.22 (1.32) | 48.65 (2.01) | 29.15 (1.73) | 44.62 (1.42) |
+ Naïve CL | 6.35 (0.06) | 39.83 (0.43) | 50.44 (1.60) | 29.77 (1.30) | 44.44 (1.53) |
+ Self-Paced CL | 6.18 (0.14) | 40.22 (1.33) | 50.05 (1.96) | 32.12 (1.16) | 47.62 (1.07) |
EnvDrop | 5.79 (0.09) | 45.44 (0.28) | 54.11 (0.89) | 41.84 (0.30) | 57.69 (0.61) |
+ Naïve CL | 5.91 (0.09) | 44.66 (0.84) | 51.91 (0.97) | 41.46 (0.72) | 57.28 (0.42) |
+ Self-Paced CL | 5.71 (0.15) | 46.11 (0.89) | 54.13 (1.21) | 42.52 (0.98) | 58.18 (0.91) |
Order Check
From the results in Table 3, naive curriculum method has negligible gains on EnvDrop model, who uses a mixed learning strategy, when compared with other two models. To further ensure the success on Follower and Self-Monitoring models does not come from side effect of simple ordering of samples, we conduct an experiment where samples are organized by number of rooms in the path and the feeding order are randomly selected from 1 to 5. In Table 8, the success rate of different agents presents a consistent decreasing trend when the input order is randomly selected.
Model | Validation Seen | Validation Unseen | ||||||||
NE ↓ | SR | OSR | SPL | nDTW | NE ↓ | SR | OSR | SPL | nDTW | |
Follower | 4.85 | 52.3 | 65.3 | 44.3 | 59.4 | 7.12 | 28.6 | 40.9 | 20.3 | 37.0 |
+ Random | 5.27 | 47.6 | 61.8 | 37.2 | 52.2 | 7.00 | 28.1 | 40.8 | 18.8 | 35.9 |
+ Reverse CL | 4.82 | 51.2 | 67.5 | 42.4 | 58.2 | 7.04 | 28.9 | 41.6 | 19.1 | 35.6 |
+ Naïve CL | 5.03 | 48.6 | 62.0 | 40.4 | 57.4 | 7.13 | 31.1 | 42.8 | 21.2 | 36.9 |
Self – Monitoring | 4.27 | 58.4 | 67.0 | 51.9 | 65.4 | 6.42 | 38.4 | 48.1 | 28.3 | 43.7 |
+ Random | 5.08 | 53.5 | 64.1 | 45.6 | 59.5 | 6.79 | 36.4 | 46.4 | 26.3 | 40.3 |
+ Reverse CL | 4.32 | 58.5 | 71.5 | 52.1 | 65.6 | 6.49 | 38.8 | 52.3 | 27.8 | 43.1 |
+ Naïve CL | 4.08 | 61.0 | 69.8 | 54.6 | 66.2 | 6.30 | 40.0 | 51.7 | 28.6 | 43.4 |
Furthermore, we reverse the input order, that is, the feeding order is monotonically decreasing, i.e. the agent is gradually trained on round 5 split, round 5~4 splits, round 5~3 splits and finally the whole CLR2R dataset. For follower agent, results show that the best success rate for seen and unseen split are 51.2% and 28.9% respectively, which is very close to the normally trained agent. For self-monitoring agent, the reverse experiment gives 58.5% and 38.8% success rate respectively and such performance is also close to the normally trained agent (but lower than the CL trained agent). Hence, it is the curriculum training paradigm that contributes to the good performance.
Extension to FGR2R
Sub-instruction and sub-paths can be considered simpler navigation tasks and hence they are naturally suitable to be included in a CL framework. We combine both R2R (Anderson et al., 2018) and FGR2R dataset (Hong et al., 2020), then traine an agent using naive CL method on that. Specifically, we split the training samples from R2R and FGR2R datasets into 3 splits using the same room-coverage heuristic. These splits contains instruction-path pairs for 1 room, 2 rooms, and 3 rooms respectively. As shown in Table 9, with more information given, curriculum learning paradigm can further improve the agent’s performance.
Validation Seen | Validation Unseen | |||||||||
NE ↓ | SR | OSR | SPL | nDTW | NE ↓ | SR | OSR | SPL | nDTW | |
Normally w/ R2R | 4.85 | 52.3 | 65.3 | 44.3 | 59.4 | 7.12 | 28.6 | 40.9 | 20.3 | 37.0 |
Naive CL w/ CLR2R | 5.03 | 48.6 | 62.0 | 40.4 | 57.4 | 7.13 | 31.1 | 42.8 | 21.2 | 36.9 |
Naive CL w/ FGR2R | 4.41 | 57.5 | 70.1 | 49.8 | 62.6 | 6.79 | 32.4 | 43.8 | 24.0 | 40.3 |