This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

APPLI: Adaptive Planner Parameter Learning From Interventions

Zizhao Wang∗1, Xuesu Xiao∗2, Bo Liu∗2, Garrett Warnell3, and Peter Stone2 Equal Contribution1Department of of Electrical and Computer Engineering, University of Texas at Austin, Austin, Texas 78712 [email protected]2Department of of Computer Science, University of Texas at Austin, Austin, Texas 78712 {xiao, bliu, pstone}@cs.utexa.edu3Computational and Information Sciences Directorate, Army Research Laboratory, Adelphi, MD 20783 [email protected]
Abstract

While classical autonomous navigation systems can typically move robots from one point to another safely and in a collision-free manner, these systems may fail or produce suboptimal behavior in certain scenarios. The current practice in such scenarios is to manually re-tune the system’s parameters, e.g. max speed, sampling rate, inflation radius, to optimize performance. This practice requires expert knowledge and may jeopardize performance in the originally good scenarios. Meanwhile, it is relatively easy for a human to identify those failure or suboptimal cases and provide a teleoperated intervention to correct the failure or suboptimal behavior. In this work, we seek to learn from those human interventions to improve navigation performance. In particular, we propose Adaptive Planner Parameter Learning from Interventions (appli), in which multiple sets of navigation parameters are learned during training and applied based on a confidence measure to the underlying navigation system during deployment. In our physical experiments, the robot achieves better performance compared to the planner with static default parameters, and even dynamic parameters learned from a full human demonstration. We also show appli’s generalizability in another unseen physical test course, and a suite of 300 simulated navigation environments.

I INTRODUCTION

Decades of research has been devoted to developing mobile robot navigation systems that are capable of moving a robot safely from one point to another in obstacle-occupied spaces without collisions. Classical navigation systems, such as Elastic-Bands [quinlan1993elastic] or Dynamic Window Approach (DWA[fox1997dynamic], have been robustly deployed on mobile robots with verifiable guarantees of safety and explainability and are able to achieve optimal navigation in most cases.

However, in some situations, those classical navigation systems fail or suffer from suboptimal behaviors (Fig. 1). For example, the robot may not be able to find a feasible action in highly-constrained spaces [xiao2020toward, liu2020lifelong], or may drive unnecessarily slowly in open spaces [xiao2020agile]. The current solution to these problems is to manually re-tune the parameters of the underlying navigation system (e.g. max speed, sampling rate, inflation radius) to correct the failure cases or suboptimal behaviors in those places. This re-tuning process not only requires expert knowledge onsite during deployment, but also runs the risk of the re-tuned parameters targeted at the failed or suboptimal comprising performance in the originally good scenarios.

Meanwhile, even a non-expert user (i.e., one who is not familiar with the inner workings of the underlying navigation system) can easily identify the situations where the robot fails or performs suboptimally by watching, and then can intervene by teleoperating the vehicle. In this work, we utilize those human interventions to improve future autonomous navigation in those troublesome places, while maintaining good performance in others.

In particular, we introduce Adaptive Planner Parameter Learning from Interventions (appli). With a set of teleoperated human interventions, appli learns a set of navigation parameters, which are selected dynamically to eliminate failures or suboptimal behaviors during deployment. To assure the learned parameters will not jeopardize navigation performance in other places, the robot only uses the learned parameters when it is confident that they will benefit the current navigation. In our experiments, appli learns from interventions in a real-world navigation task. We test appli in the same training and another unseen physical environment. More than twenty thousands simulation trials are conducted in unseen environments to test appli’s generalizability. Our results show appli can improve upon default navigation performance and can generalize well to unseen environments, indicating that interventions are a uniquely valuable form of human interaction for building navigation systems.

Refer to caption
Figure 1: While classical navigation systems perform well in most places (green), they may fail (red) or suffer from suboptimal behavior (yellow) in others. appli utilizes human interventions in these two scenarios (we name them Type A and Type B interventions, respectively) to learn adaptive planner parameters and, based on a confidence measure, uses them during deployment.

II RELATED WORK

In this section, we review existing work on adaptive planner parameters, learning from intervention, and uncertainty measurement in deep learning.

II-A Adaptive Parameters for Classical Navigation

Classical navigation methods enjoy safety, explainability, and stable generalization to new environments. However, when facing new environments, they still need a great deal of tuning, which often requires expert robotics knowledge [zheng2017ros, xiao2017uav]. Prior work has considered automated parameter tuning, e.g., finding trajectory optimization weights [teso2019predictive] for the dwa planner [fox1997dynamic], or designing novel systems that can leverage gradient descent to match expert demonstrations [bhardwaj2019differentiable]. Specifically, Xiao et al [xiao2020appld] adopted black-box optimization to automatically map a robot’s local observation to the optimal planner parameters via learning from human demonstration. While this technique can be applied to any parameter-based planner, it is not expected to generalize well in environments not seen in the demonstration. In contrast, appli only requires a few, short, local interventions when classical navigation does not perform well, instead of a demonstration of the full trajectory. appli also includes confidence estimation over candidate planner parameters during deployment in unseen environments. Notably, in the worst case our method reduces to the planner with default parameters, rather than a poorly chosen parameter set, and therefore enjoys better generalization.

II-B Learning from Intervention

Due to the cost of providing full demonstrations, human intervention is a popular approach to providing minimal guidance for learning. It has been widely used in reinforcement learning [saunders2017trial, prakash2019improving] and imitation learning [goecks2019efficiently, kelly2019hg, zhang2016query][goecks2019efficiently, kelly2019hg, zhang2016query, spencerlearning, kahn2021land]. Learning from intervention essentially focuses the agent on learning from its mistakes, thus improving the data efficiency and reducing the demonstration cost. In this work, we leverage the benefits of learning from intervention to enable robust robot navigation, and further categorize interventions based on the expert’s estimation of the necessity of such interventions.

II-C Measuring Uncertainty in Deep Learning

Recent advances in deep learning have provided a family of tools for measuring the uncertainty in a deep model’s prediction. There are mainly three types of approaches. (1) Bayesian Neural Networks (BNN) represent distributions over network weights and the prediction uncertainty is indirectly inferred via weight uncertainty [kononenko1989bayesian]; (2) Deep Ensemble (DE) uses the outputs from multiple networks, each trained with partial data, as a Monte-Carlo estimator for uncertainty [lee2019ensemble]. A specific example is Dropout learning[gal2016dropout]; (3) Other methods train a single network with stationary weights but directly model the predictions in terms of a distribution. Evidential Deep Learning (EDL)[sensoy2018evidential] is one particular method that models a discrete class of predictions with a Dirichlet distribution. We incorporate EDL into appli due to its simplicity and efficiency in terms of both time and space complexity, compared to methods from the other two approaches, which is essential for robot learning.

III APPROACH

In this section, we introduce our method, appli, which has two novel features: (1) compared with Learning from Demonstration that requires demonstration of the whole task, appli only needs a few interventions in challenging scenarios where the default navigation system does not work well; (2) with a confidence measure on candidate parameters learned from interventions, our method knows when to switch back to the default parameters. This confidence measure enables appli to generalize well to unseen environments.

III-A Problem Definition

We denote a classical parameterized navigation system as G:𝒳×Θ𝒜G:\mathcal{X}\times\Theta\rightarrow\mathcal{A}, where 𝒳\mathcal{X} is the state space of the robot (e.g. goal, sensor observations), Θ\Theta is the parameter space for GG (e.g. max speed, sampling rate, inflation radius), and 𝒜\mathcal{A} is the action space (e.g. linear and angular velocities). During deployment, the navigation system repeatedly estimates state xx and takes action aa calculated as a=G(x;θ¯)a=G(x;\bar{\theta}). Typically, the default parameter set θ¯\bar{\theta} is tuned by a human designer trying to achieve good performance in most environments. However, being good at everything often means being great at nothing: θ¯\bar{\theta} usually exhibits suboptimal performance in some situations and may even fail (is unable to find feasible motions, or crashes into obstacles) in particularly challenging ones.

To mitigate this problem, a human can supervise the navigation system’s performance at state xx by observing its action aa and judging whether (s)he should intervene. Here, we consider two types of interventions. A type A intervention is one in which the system performs so poorly that the human must intervene (e.g. imminent collision or a signal for help). A type B intervention is one in which a human might intervene in order to improve otherwise suboptimal performance (e.g., driving too slowly in an open space). For the ithi^{\text{th}} intervention, we assume that the human resets the robot to the position where the failure or suboptimal behavior first occurred and then gives a short teleoperated intervention Ii={xt,at}t=1TiI_{i}=\{x_{t},a_{t}\}_{t=1}^{T_{i}} of length TiT_{i}, where x1:Tix_{1:T_{i}} is the trajectory starting from the reset state induced by intervention actions a1:Tia_{1:T_{i}}. As this short demonstration shows a cohesive navigation behavior in a specific segment of the environment (open space, narrow corridor, etc), we refer to the segment as a context cic_{i} and denote the space of contexts as 𝒞\mathcal{C}. Given NN interventions I1:NI_{1:N}, appli finds (1) a mapping M:𝒞ΘM:\mathcal{C}\rightarrow\Theta that determines the parameter set θi\theta_{i} for each intervention context cic_{i}, and (2) a parameterized predictor Bϕ:𝒳𝒞B_{\phi}:\mathcal{X}\rightarrow\mathcal{C} that determines to which context (if any) the current state xx belongs.

III-B Parameter Learning

After collecting a set of NN interventions I1:NI_{1:N}, for each IiI_{i}, we learn a set of navigation parameters θi\theta_{i} that can best imitate the demonstrated correction. To find such parameters, we use the same training procedure as in the approach by Xiao et al [xiao2020appld], i.e., we use Behavior Cloning to minimize the difference between the actions from the human and those generated by the navigation system with new parameters θi\theta_{i}. To be specific,

θi=argminθ(x,a)IiaG(x;θ)λ,\theta_{i}=\operatorname*{arg\,min}_{\theta}\sum_{(x,a)\in I_{i}}\|a-G(x;\theta)\|_{\lambda}, (1)

where dλ=λidi2\|d\|_{\lambda}=\sum\lambda_{i}d_{i}^{2} is the norm of the action difference with λ\lambda weighting the different action dimensions (in our case, linear and angular velocity, vv and ω\omega). The loss in Eqn. (1) is minimized with a black-box optimizer, such as CMA-ES [hansen2003reducing]. After identifying parameters in each context, the mapping MM is simply M(i)=θiM(i)=\theta_{i}.

III-C Confidence-Based Context Prediction

So far, we have described how to learn multiple parameter sets θ1:N\theta_{1:N} from human interventions I1:NI_{1:N} in contexts c1:Nc_{1:N}. In order to select the correct parameters at deployment time, we must also determine if the current state xtx_{t} falls into any one of the collected intervention contexts cic_{i}. If such a determination can be made, then we direct the robot to use the parameter set θi\theta_{i} to avoid making the same mistake as before. If it cannot be determined that xtx_{t} belongs to a particular intervention context, then we direct the robot to use the default parameters θ¯\bar{\theta}, as they are optimized for most cases and are expected to generalize better than any parameter set learned for a specific scenario. In our system, the determination above is made using a predictor, BϕB_{\phi}. To train this predictor, we first use the collected interventions to build a dataset, {{xt,ci}t=1Ti}i=1N\{\{x_{t},c_{i}\}_{t=1}^{T_{i}}\}_{i=1}^{N}, and train an intermediate classifier fϕ(x)f_{\phi}(x) with parameter set ϕ\phi using the Evidential Deep Learning method (EDL) [sensoy2018evidential]. A feature of EDL is that it supplies both a predicted label prediction and a confidence in that prediction ui(0,1]u_{i}\in(0,1], i.e.,

fϕ(xi)=(ci,ui).f_{\phi}(x_{i})=(c_{i},u_{i}). (2)

After training fϕf_{\phi} and during deployment, we can build a confidence-based classifier gϕg_{\phi} as

gϕ(xi)=ci𝟙(uiϵu),g_{\phi}(x_{i})=c_{i}\mathbbm{1}(u_{i}\geq\epsilon_{u}), (3)

where ϵu\epsilon_{u} is the threshold on confidence and 𝟙\mathbbm{1} is the indicator function. For state xix_{i}, gϕg_{\phi} determine its context from N+1N+1 contexts (NN intervention contexts and 1 default context). If uiϵuu_{i}\geq\epsilon_{u}, it suggests the classifier fϕf_{\phi} is confident and gϕg_{\phi} predicts cic_{i}. Otherwise, when fϕf_{\phi} is unsure about its prediction, ci𝟙{uiϵu}=0c_{i}\mathbbm{1}\{u_{i}\geq\epsilon_{u}\}=0. In this case, gϕg_{\phi} believes the current state xix_{i} is not similar to any intervention context and instead classifies xix_{i} as the default context. For this default context labeled as ci=0c_{i}=0, navigation utilizes the default navigation parameters θ¯\bar{\theta} (i.e., we set M(0)=θ¯M(0)=\bar{\theta}).

Then we define our context predictor BϕB_{\phi} as:

Bϕ(xt)=mode({gϕ(xi)}i=tw+1t).B_{\phi}(x_{t})=\text{mode}(\left\{g_{\phi}(x_{i})\right\}_{i=t-w+1}^{t}). (4)

To avoid a context estimation ctc_{t} that changes frequently (e.g. caused by gϕg_{\phi}’s wrong classifications), BϕB_{\phi} acts as a mode filter with window length ww and chooses the context ctc_{t} that the majority of classifications agree with over the past ww time steps.

III-D appli

Putting together all the components presented above, the entire appli pipeline is summarized in Alg. 1. In the training stage, it collects NN interventions from a human supervisor (line 1), and then learns corresponding navigation parameters θ1:N\theta_{1:N} i.e., the mapping MM (lines 2-3) and a context predictor BϕB_{\phi} (lines 5-6). During deployment, we use M(Bϕ(xt))M(B_{\phi}(x_{t})) to select the parameters for the navigation system at time tt (lines 8-10).

1 Training
Input: human interventions I1:N={{xt,at}t=1Ti}i=1NI_{1:N}=\{\{x_{t},a_{t}\}_{t=1}^{T_{i}}\}_{i=1}^{N}, navigation system GG, parameter space Θ\Theta.
2 for i=1,,Ni=1,\dots,N do
3      find parameter θi\theta_{i} for context ii using Eqn. (1).
4      end for
5     train the context classifier fϕf_{\phi} on {{xt,ci}t=1Ti}i=1N\{\{x_{t},c_{i}\}_{t=1}^{T_{i}}\}_{i=1}^{N}.
6      build mapping M(i)=θiM(i)=\theta_{i} and context predictor Bϕ(x)B_{\phi}(x).
7     Deployment
      Input: navigation system GG, parameter mapping MM, context predictor Bϕ(x)B_{\phi}(x), confidence threshold ϵu\epsilon_{u}, fallback parameters θ¯\bar{\theta}.
8      for t=1,t=1,\dots do
9           identify the current context ct=Bϕ(xt)c_{t}=B_{\phi}(x_{t}) with confidence threshold ϵu\epsilon_{u}.
10           navigate with G(xt,M(ct))G(x_{t},M(c_{t})).
11           end for
Algorithm 1 appli

IV EXPERIMENTS

In our experiments, we aim to show that appli can improve navigation performance by learning from only a few interventions and, with the confidence measurement, that the overall system can generalize well to unseen environments. We apply appli on a ClearPath Jackal ground robot in a physical obstacle course. Navigation performance learned through appli is then tested both in the same training environment, and also in another unseen physical test course. Furthermore, to investigate generalizability, we test the learned systems on a benchmark suite of 300 unseen simulated navigation environments.

IV-A appli Implementation

Our Jackal is a differential-drive robot equipped with a Velodyne LiDAR that we use to compute a 720-dimensional planar laser scan with a 270 field of view. The robot uses the Robot Operating System move_base navigation stack with Dijkstra’s global planner and the default dwa local planner, which works in most situations, but fails or behaves suboptimally in others (see Fig. 1).

During data collection, one of the authors (the intervener) follows the robot through the test course and intervenes when necessary, reporting if the intervention is to drive the robot out of a failure case (Type A) or to correct a suboptimal behavior (Type B). The four interventions are shown in Fig. 1: before the two Type A interventions (shown in red), the default system (dwa with θ¯\bar{\theta}) fails to plan feasible motions and starts recovery behaviors (rotates in place and moves backward); before the two Type B interventions (shown in yellow), the robot drives unnecessarily slowly in a relatively open space and enters the narrow corridor with unsmooth motions. For every intervention, the intervener stops the robot, drives it back to where they deem the failure or suboptimal behavior to have begun, and then provides recorded teleoperation II that avoids the problematic behavior. To compare the performance learned from interventions and learned from a full demonstration, we also collect extra demonstrations for those places where the default planner already works very well (denoted in green in Fig. 1).

This set of interventions comprises the input I1:N={{xt,at}t=1T}i=1NI_{1:N}=\{\{x_{t},a_{t}\}_{t=1}^{T}\}_{i=1}^{N} to Alg. 1, where xtx_{t} is all the sensory data fed into the move_base stack, GG, and ata_{t} is the linear and angular velocity (vv and ω\omega) from teleoperation. The default and learned parameters are shown in Tab. I, including those learned from Type A and B interventions (A1, A2 and B1, B2), and the extra demonstrations (D1, D2).

TABLE I: Default and Learned Planner Parameters:
max_vel_x (v), max_vel_theta (w), vx_samples (s), vtheta_samples (t), occdist_scale (o), pdist_scale (p), gdist_scale (g), inflation_radius (i)
v w s t o p g i
def. 0.50 1.57 6 20 0.10 0.75 1.00 0.30
A1 0.26 2.00 13 44 0.57 0.76 0.94 0.02
A2 0.22 0.87 13 31 0.30 0.36 0.71 0.30
B1 1.91 1.70 10 47 0.08 0.71 0.35 0.23
B1 0.72 0.73 19 59 0.62 1.00 0.32 0.24
D1 0.37 1.33 9 6 0.95 0.83 0.93 0.01
D2 0.31 1.05 17 20 0.45 0.61 0.22 0.23

IV-B Physical Experiments

After the training in Alg. 1 with the collected interventions, we deploy the learned mapping MM and context predictor BϕB_{\phi} on the move_base navigation stack GG. We use a confidence threshold ϵu=0.8\epsilon_{u}=0.8.

We first deploy appli in the same training physical environment (Fig. 1). We compare the performance of the dwa planner with default parameters, appli learned only with Type A interventions, appli learned with Type A and Type B interventions, and appli learned with a full demonstration (which is basically the appld framework [xiao2020appld] enhanced by the confidence measure). The motivation for the variation of appli learned only with Type A interventions is to study the effect of an unfocused or inexperienced human intervener. In this case, the human would still conduct all Type A interventions, as those mistakes are severe and easy to identify—some robots may even actively ask for help (e.g. by starting recovery behaviors). However, the human may fail to conduct Type B interventions as (s)he is not paying attention, or isn’t equipped with the knowledge to identify suboptimal behaviors. For all methods, we run five trials each and report the mean and standard deviation of the traversal time in Tab. II. If the robot gets stuck, we introduce a penalty value of 200 seconds. We also deploy the same sets of variants in an unseen physical environments (Fig. 2) and report the results in Tab. III.

TABLE II: Traversal Time in Training Environment
Default Type A Type A+B Full Demo
134.0±\pm60.6s 77.4±\pm2.8s 70.6±\pm3.2s 78.0±\pm2.7s
Refer to caption
Figure 2: appli Running in an Unseen Physical Environment
TABLE III: Traversal Time in Unseen Environment
Default Type A Type A+B Full Demo
109.2±\pm50.8s 71.0±\pm0.7s 59.0±\pm0.7s 62.0±\pm2.0s

For both training and unseen environments, Type A interventions alone significantly improve upon the default parameters, by correcting all recovery behaviors such as rotating in place or driving backwards, and eliminating all failure cases. Adding Type B interventions further reduces traversal time, since the robot learns to speed up in relatively open spaces and to execute smooth motion when the tightness of the surrounding obstacles changes. All the interventions are able to improve navigation in both training and unseen environments, suggesting appli’s generalizability. Surprisingly, in both environments, appli learned from only Type A and Type B interventions can even outperform appli learned from an entire demonstration. One possible reason for this better performance from fewer human interactions is the additional human demonstrations may be suboptimal, especially since they are collected in places where the default navigation system was already deemed to have performed well. For example, in the full demonstration, we find the human intervener is more conservative than the default navigation system and drives slowly in some places. Hence, learning from these suboptimal behaviors introduces suboptimal parameters and consequently worse performance in contexts similar to that intervention.

IV-C Simulated Experiments

TABLE IV: Percentage of Simulation Environments that Method 1 is Significantly Worse than Method 2 in Terms of Traversal Time
(Methods are listed in order of increasing performances. Results mentioned in experiment analysis are bold for better identification)
Method 2
appli (A) dwa appli (A+c) appli (A+B+D+c) appli (A+B+D) appli (A+B+c) appli (A+B)
Method 1 appli (A) 0 50 53 62 63 68 66
dwa 10 0 6 33 40 44 47
appli (A+c) 6 4 0 31 37 45 45
appli (A+B+D+c) 5 7 11 0 25 31 33
appli (A+B+D) 5 7 7 10 0 21 21
appli (A+B+c) 3 3 4 3 5 0 9
appli (A+B) 2 5 5 6 4 6 0

To further test appli’s generalizability to unseen environments, we test our method and compare it with two baselines on the Benchmark for Autonomous Robot Navigation (BARN) dataset [perille2020benchmarking]. The benchmark dataset consists of 300 simulated navigation environments generated using Cellular Automata, ranging from easy ones with a lot of open spaces to challenging ones where the robot needs to get through dense obstacles. Navigation trials in three example environments with low, medium, and high difficulty levels are shown in Fig. 3. Using the same training data collected from the physical environment shown in Fig. 1, we test the following seven variants:

Refer to caption
Figure 3: Navigation Trials in Example Environments with Low, Medium, and High Difficulty Levels
Refer to caption
Figure 4: Average Performance in 300 Simulation Environments under 12 Different Runs
  • appli learned from Type A and B inventions with confidence measure, denoted as appli (A+B+c).

  • appli learned from Type A and B inventions without confidence measure, i.e., appli (A+B).

  • appli learned from only Type A interventions with confidence measure, i.e., appli (A+c).

  • appli learned from only Type A interventions without confidence measure, i.e., appli (A).

  • appli learned from full demonstration with confidence measure, i.e., appli (A+B+D+c).

  • appli learned from full demonstration without confidence measure, i.e., appli (A+B+D).

  • the dwa planner with default parameters.

Testing these variations aims at studying the effect of learning from different modes of interventions caused by different degrees of human attention and experience levels, i.e. imperative interventions (A), optional interventions (A + B), and a full demonstration (A + B + D). They also provide an ablation study for the confidence measure in the EDL context classifier fϕf_{\phi}: when deployed without the confidence measure, the robot has to choose among the parameters learned from interventions and never uses the default parameters.

For each method in each simulation environment, we measure the traversal time for 12 different runs (the run is terminated after 50s if the robot gets stuck), resulting in 25200 total navigation trials. The average traversal time for each method in all simulation environments are shown in order of increasing performance in Fig. 4. We then conduct a pair-wise t-test for all methods in order to compute the percentage of environments in which one method (denoted as Method 1) is significantly worse (p<0.05p<0.05) than another (denoted as Method 2). For better illustration, we also reorder the method by their performance and show the pairwise comparison in Tab. IV.

appli (A+B+c) and appli (A+B+D+c) outperform dwa: they are significantly better in 33% and 44% of environments respectively and significantly worse in only 7% and 3% of environments than dwa. However, appli (A+c) is only significantly better than dwa 6% of the time, which suggests that even though type B inventions only correct suboptimal performances, they are crucial for performance improvement. In detail, as the robot learns to go through narrow passages from Type A interventions, the first Type B intervention further teaches the robot to drive fast in safe open spaces, significantly reducing the traversal time for simulation environments whose beginnings and ends are relatively open. Meanwhile, from the second Type B intervention, the robot learns how to take sharp turn despite constrained surroundings.

In terms of the effect of confidence, appli (A) only selects parameters learned from 2 Type A inventions and never uses the default parameters even when they are more appropriate. Removing confidence greatly harms its performance, making it significantly worse than appli (A+c) in 53% of environments. However, appli (A+B+c) and appli (A+B+D+c), which use more interventions or even the full demonstration to train the parameter mapping MM and context predictor BϕB_{\phi} are more confident about their predictions most of the time. As a result, removing confidence in the context predictor doesn’t result in a significant difference.

Lastly, a conterintuitive, but similar result as in the physical experiments is that compared with appli (A+B+D+c) which uses the full demonstration, appli (A+B+c) learned from only Type A and B interventions achieves superior performance by being significantly better than appli (A+B+D+c) and appli (A+B+D) in 31% and 21% of the environments, respectively. However, similar to the discussions about physical experiments, unnecessary human demonstrations are most likely suboptimal. In this sense, appli not only reduces the required human interactions from a full demonstration to only a few interventions, but also reduces the chance of performance degradation caused by suboptimal demonstrations.

V CONCLUSIONS

In this work, we introduce appli, Adaptive Planner Parameter Learning from Interventions. In contrast to most existing end-to-end machine learning for navigation approaches, appli utilizes existing classical navigation systems and inherits all their benefits, such as safety and explainability. Furthermore, instead of requiring a full expert demonstration or random exploration based on trial-and-error, appli only needs a few interventions, where the default underlying navigation system fails or exhibit poor behavior. It also introduces a confidence measure to assure generalizability in unseen environments. We show appli’s improved performance in training and unseen physical environments. We further test appli’s generalizability with 25200 simulated navigation trials in 300 unseen environments. While we allow the intervener to start interventions by rewinding the robot navigation to a state before the failure or suboptimal behavior occurs, an interesting direction for future work is to further investigate interventions without “rewinding”, i.e. where the intervener takes over from where the robot fails and drives it forward to a good state.

ACKNOWLEDGMENT

This work has taken place in the Learning Agents Research Group (LARG) at the Artificial Intelligence Laboratory, The University of Texas at Austin. LARG research is supported in part by grants from the National Science Foundation (CPS-1739964, IIS-1724157, NRI-1925082), the Office of Naval Research (N00014-18-2243), Future of Life Institute (RFP2-000), Army Research Office (W911NF-19-2-0333), DARPA, Lockheed Martin, General Motors, and Bosch. The views and conclusions contained in this document are those of the authors alone. Peter Stone serves as the Executive Director of Sony AI America and receives financial compensation for this work. The terms of this arrangement have been reviewed and approved by the University of Texas at Austin in accordance with its policy on objectivity in research.