Improved active output selection strategy for noisy environments
Abstract
The test bench time needed for model-based calibration can be reduced with active learning methods for test design. This paper presents an improved strategy for active output selection. This is the task of learning multiple models in the same input dimensions and suits the needs of calibration tasks. Compared to an existing strategy, we take into account the noise estimate, which is inherent to Gaussian processes. The method is validated on three different toy examples. The performance compared to the existing best strategy is the same or better in each example. In a best case scenario, the new strategy needs at least 10% less measurements compared to all other active or passive strategies. Further efforts will evaluate the strategy on a real-world application. Moreover, the implementation of more sophisticated active-learning strategies for the query placement will be realized.
keywords:
Gaussian Processes, Active Learning, Regression, Active Output Selection, Drivability Calibration, Experiment Designadieresis=ä,germandbls=ß
1 Introduction
Active learning – also known as online design of experiments or optimal experimental design – reduces the number of necessary measurements to achieve sufficient model accuracy. This is especially relevant for tasks with high labeling or measurement costs. Additionally, it facilitates the test design procedure since the placement of measurement points is conducted during test execution. The scientific community mostly focuses on optimally learning regression models with one output. Our field of research and application, the drivability calibration, requires models for multiple outputs with the same input dimensions. There is an importance of assuring a balanced model accuracy for all outputs, which originates in their purpose: The identified models are mostly used for optimization. In this case, the model with the lowest accuracy determines the quality of the following optimization. For a test engineer, this raises a question: In which order should the models actively learn to have the lowest number of measurements?
For these boundary conditions, we introduced a new active learning task, which is called active output selection, in a previous study, see Prochaska et al. (2020). We compared already existing heuristic strategies to a new strategy which chooses the leading model based on the cross-validation error. This paper advances the presented strategy to overcome the drawbacks. Multiple toy examples are used to illustrate the effectiveness of the new strategy.
Section 2 of this paper introduces previous works in context of active learning in general and in particular for regression tasks. A special focus lies on applications in the automotive calibration domain. Section 3 focuses on the task of active output selection and the specialties of active learning in the context of drivability calibration. Section 4 describes the analyzed approaches. Furthermore, an improved strategy for active output selection is presented. The approaches are evaluated using different toy examples. Experimental details and a discussion of results are shown in section 5. At the end, section 6 concludes the results and presents fields of possible future works.
2 Previous works
The field of active learning is a growing field in machine learning. In statistics literature, it is also called optimal experimental design, see Cohn (1996). Since there is a necessity of labeled data, it is a subdomain of supervised learning. Settles (2009) illustrates a broad overview of the current state of the art in active learning and gives an outlook to multiple possible future work fields. Recent advances in the scientific community mostly focused on classification problems. Those methods are often applied to speech recognition and text information extraction tasks, see Settles (2009).
Regression tasks in the context of active learning have not been as popular. However, the methodological advances are relevant as well. Sugiyama and Rubens (2008) showed an approach which actively learns an ensemble of models for the same task and selects the best one to query new points. Cai et al. (2013) introduced expected model change maximization (EMCM) for improving active learning for gradient boosted decision trees. Cai et al. (2017) later extended EMCM to choose a set of informative queries and to Gaussian process regression models (GPs). Park and Kim (2020) evolved EMCM for active learning even further into a learning algorithm, which handles outliers more robustly than before. The active learning strategies presented in those publications focus on improvements for single-output regression models. Zhang et al. (2016) introduced a learning algorithm for multiple-output Gaussian processes (MOGP). The algorithm improves the prediction accuracy of one target output model by taking into account several correlated auxiliary output models. It outperforms multiple single-output Gaussian processes (SOGP) and additionally indicates that a global consideration of multiple outputs is beneficial.
In automotive calibration tasks, the identification of multiple process outputs in the same experiment is more relevant to the application. There were also advances in active learning in this field. Hartmann and Nelles (2013) presented HiLoMoT-DoE, a method for design of experiments with hierarchical local model trees. Klein et al. (2013) applied this method successfully to an engine calibration task. Each of their two application examples had two outputs as well as five an seven inputs, respectively. A sequential strategy was chosen to model the two outputs in the application example. This strategy defines one output model as leading and identifies it completely before moving to another output.
Reichart et al. (2008) introduced a round-robin strategy for combining two linguistic annotation tasks111The round-robin strategy is called alternating selection by Reichart et al. (2008).. It switches the leading model after each iteration and performs better than separate annotation.
Dursun et al. (2015) adapted HiLoMoT-DoE to drivability calibration examples. Their algorithm identified multiple static regression tasks with identical input dimensions during the test run. Furthermore, they applied the round-robin strategy of Reichart et al. (2008) to their regression task and compared it to other existing strategies. In this single real-world experiment, the round-robin strategy performs better than offline methods and the online sequential strategy. This indicates, that round-robin is preferably used in general.
In an earlier publication, we thoroughly analyzed those strategies in a reproducible, noisy environment, see Prochaska et al. (2020). We introduced a new strategy, which selects the leading model based on the evaluation of the cross-validation error. The results can be described as mixed. Depending on the setup, it outperforms all other strategies by up to 20%. However, the strategy has difficulties when dealing with different signal-to-noise-ratios on different outputs. This leads to the conclusion, that the strategy’s general idea is advantageous, but its robustness to noise has to be increased.
3 Problem definition
In general, active learning is applied to tasks with high labeling costs. As stated in section 2, most advances in active learning improve the learning strategies for one model output. This paper is a contribution to a subdomain of active learning which is called active output selection (AOS). For these tasks, an algorithm identifies models. Their outputs are possibly uncorrelated, but their input dimensions are identical. In case of drivability calibration, measurements determine the values of the modeled outputs. The measurements always contain noise to some extent. A measure for the noise of output is defined by the signal-to-noise-ratio , where is the vector of all measurements of output and the standard deviation of normally distributed noise. For drivability criteria, the SNR lies approximately in a range of and can be different for each output. The measurements are conducted on powertrain test benches. On the one hand, one measurement is cost-expensive, hence the number of measurements ought to be reduced by active learning. Furthermore, the costs of measuring one output are identical to measuring all outputs, which then results in always measuring all outputs for one query. On the other hand, the duration of one measurement is timely more expensive than the evaluation of code. Consequently, a specific complexity analysis of the presented algorithms is lacking in this paper, but advantages and disadvantages are discussed in section 4. After identification, the models are used for optimizations of control unit parameters. Therefore, there is no primary or target output, but an adequate model quality for all models is crucial.
4 Active output selection strategies
Each of the process outputs is modeled using GPs, since they are suited for drivability applications and handle noise robustly, see Tietze (2015). In each iteration of the active learning setup, the leading process output defines the query for the next measurement. This paper analyzes different strategies to determine the leading process output. The queries are placed by the leading output using a maximum variance strategy, which was presented by MacKay (1992) in a general form. The next query of output , which is inside the input space , maximizes the same model’s output variance .
(1) |
The implementation is straightforward for GPs since the output variance is directly evaluated at each individual input point. Equation 2 and eq. 3 show the calculations of the predicted mean and output variance of a GP. is the vector of covariances between the measured training points and a single test point . are the covariances of and y contains the observations under noise with variance , see Rasmussen and Williams (2008).
(2) | ||||
(3) |
The active learning strategy defines the next query if the leading model is already selected. However, an AOS strategy has to define the leading model of the learning process beforehand. In the following, all analyzed AOS strategies are described. These descriptions include a space-filling, a round-robin and a strategy. Furthermore, this paper presents a normalized strategy, which is an improvement of of the strategy. To the author’s knowledge, the latter has never been presented before and therefore is entirely new.
4.0.1 sequential space-filling strategy (passive, SF)
This is the baseline strategy. We do not assume any prior knowledge about the modeled process outputs. In drivability calibration, the usage of space-filling designs with a high number of points is common. An offline space-filling design can only be applied if either the number of measurements is fixed before the experiments. Contrary to that, the sequential space-filling strategy is passive but can be compared with other active methods, since the number of points is sequentially increased. As all AOS strategies, SF starts with a number of random points. Every succeeding point maximizes the minimum Mahalanobis-distance between the already measured points and a huge set of candidate points .222For uncorrelated data in a range between 0 and 1, the Mahalanobis distance and the Euclidean distance are the same. This strategy does not use active learning at all, since the models do not have an impact on the query placement. In contrast to AOS strategies, SF is a passive strategy.
4.0.2 round-robin strategy (RR)
This is the second baseline strategy. RR changes the leading model after each iteration. When a model reaches the desired model quality, it is no leading model until the end of the experiment. RR does not compare model qualities of the process outputs with each other. This leads RR to not operating optimally because models that cannot be further improved but do not reach the desired model quality or models that do not benefit from being in the lead at the current status of experiment are still active every iterations. Due to the characteristics of this heuristic strategy, RR is robust under noisy conditions and outperforms SF in general.
4.0.3 strategy (CVH)
This strategy is presented and analyzed in Prochaska et al. (2020). It assumes that the model with the lowest model accuracy benefits the most from leading the learning process. The algorithm uses the normalized root mean squared -fold cross-validation-error with . is filtered with a digital moving average filter for stability reasons, see Prochaska et al. (2020). In each iteration, the CVH calculates for all outputs and selects the one with the highest as leading. Each of the models has to be identified for times to calculate , which results in higher computational costs compared to other strategies.
Prochaska et al. (2020) showed that CVH is beneficial in most setups. It handles similar and different complexities in model outputs well and outperforms baseline as well as other AOS strategies. However, it is not able to distinguish between bad model quality and high noise. Setups with an uncomplex but noisy output, on the one hand, and a complex but less noisy output, on the other hand, lead to poor performance. Even the SF baseline outperforms CVH in such scenarios. This lead to the conclusion that CVH has to be improved into handling such scenarios better.
4.0.4 normalized strategy (CVHn)
The drawbacks described in the previous paragraph show that an improvement of the existing AOS strategy is necessary. CVHn picks up the idea of selecting a model based on its . However, it includes a noise estimate in the decision criterion. Therefore, CVHn chooses the leading model based on the normalized . This strategy uses a characteristic of the GPs. As shown in equation 3, a standard deviation estimate of the process noise is directly determined during the model identification. Since the real noise of a process output limits the minimum possible model error, is normalized with the noise estimate . We assume that the model with a high normalized cross-validation score benefits most from leading the learning process.
Algorithm 1 shows the pseudo-code of CVHn. First, CVHn measures an initial set of points. Afterwards, the algorithm identifies all models in each iteration and calculates . The algorithm defines the model that has the highest as leading. CVHn stops if the maximum number of points or the desired model quality is reached.
As CVH, CVHn has to identify all models for times. This leads to higher computational costs compared to the other strategies. The execution time of those algorithms is not crucial for the drivability calibration applications on test benches, since the time needed to conduct one measurement is much bigger.
5 Experiments
As described previously, all AOS strategies have different strengths and disadvantages. The following section analyzes the effectiveness of those strategies on toy examples. The AOS strategies are tested for three different toy examples. These three examples are specially designed to show their characteristics. 15 runs are conducted for each AOS strategy and for each toy example. Any random seeds are reset for each run of one strategy but changed after every round. Due to the same prerequisites, one run of each strategy is comparable to another. The model accuracy of each process output is evaluated with the (see eq. 4).
(4) |
To compare multiple strategies with each other, one single criterion is favored over different ones. Therefore, , the squared sum of the accuracy criteria of all process outputs, is used.
(5) |
To test the algorithms, three different setups of toy examples that reflect distinct possible real world cases are used. These setups are described in Prochaska et al. (2020) and only mentioned in this paper. Each setup has three different outputs that are non-correlated. Each of those outputs has an underlying analytical model, which is created by a random function generator, which was introduced by Belz and Nelles (2015). The output values are overlaid with normally distributed noise; the standard deviation varies from setup to setup. Each output has two input dimensions in the same range. Because this toy example is derived from test bench applications, querying one input from the setup hands values for each output over to the learning algorithm (see section 3). The three setups are distinguished by the complexity of the underlying model as well as the signal-to-noise-ratio SNR of each output:
-
•
Setup 1 has similar model complexities and SNRs for all outputs.
-
•
Setup 2 has similar SNRs for all outputs but one output with considerably higher complexity. The complexity of said output is due to step functions, which represent a strong local non-linearity.
-
•
Setup 3 also has one output with higher complexity. Additionally, a different output is characterized by a lower SNR.
The number of measurements is limited to 100 for each strategy and run. For validation purposes, 121 points on an equidistant 11-by-11-grid are considered. The advantage of the toy examples is that the true validation point without the influence of any output noise are known.
Figure 1 shows the mean and standard deviation of the validation error over the number of measurements for setup 1. The active learning strategies outperform the passive SF. RR and CVH reduce the number of points to reach SF’s end value by 10% and 15%, respectively. CVHn further reduces the number of measurement points and reaches the end value of SF after 80%. For , CVHn performs significantly better than SF. In this setup, CVHn outperforms the other active output selection strategies after . The passive and all active learning strategies strike with low variance over the runs.

Figure 2 shows the performances of those strategies on setup 2. The benefits of considering for determining the leading output is clear at this point. CVHn performs similarly well as CVH; both reduce the number of points by 35% and 40% compared to SF and by 30% each compared to RR. This is accomplished with a lower standard deviation over the runs. CVH has a slightly lower mean value than CVHn over all runs. However, this difference is not statistically significant. This shows that the newly presented strategy CVHn performs as well as the already shown CVH in a setup were the latter showed to be very effective in the earlier publication, see Prochaska et al. (2020).

In the same paper, setup 3 showed the limits of CVH. The strategy often puts the noisy output in the lead because of its high . Figure 3 shows the evolution that takes place from CVH to CVHn. While CVH is not even able to outperform the passive SF in the end, the robust heuristic RR reaches the end value of SF after 85% of the measurements. The consideration of the models’ estimated noise for normalization of reduces the drawbacks of CVH. After , CVHn already outperforms CVH. Even in this setup, CVHn reaches the end value of SF and RR after 70% and 80% of the measurements.

The results convincingly show the advantages of the new CVHn strategy. Furthermore, the limitations of CVH in the presence of noise compared to CVHn are very apparent.
6 Conclusion
This paper introduces an improvement for the CVH active output selection strategy, which was presented by Prochaska et al. (2020). Active output selection strategies identify multiple process outputs with the same input dimensions. Depending on the strategy, different outputs are selected as leading and therefore actively querying measurements. CVH uses the 10-fold cross-validation error to determine the leading model. This strategy showed drawbacks in certain scenarios. The new strategy CVHn normalizes the 10-fold cross-validation error to increase the informative content and takes into account a noise estimate of each process output.
The paper analyzes CVHn and compares it to CVH as well as two other baseline strategies. Three toy examples are used to test said strategies in a reproducible but noisy environment. Each of those toy examples addresses different difficulties in an active output selection task. The results show that active learning strategies outperform the passive space-filling baseline strategy in most cases. In one of the presented examples, CVHn performs similar to the existing CVH, but there is no statistical significance. In any other example, CVHn outperforms all other strategies. In a best case scenario, the number of points is reduced by 10% at least and 60% at best. Specific weaknesses of CVH concerning the robustness to noise could be reduced with almost no compromise.
These results confirm the effectiveness of CVHn. However, this paper is lacking results of a benchmark data set or real-world applications examples. Therefore, further research must focus on testing the strategy in a real-world environment and creating a benchmark data set fitting to the specific application. Moreover, this paper focuses on a very simple active learning strategy for each model. Implementation of more advanced active learning methods could further decrease the needed number of points. The AOS strategy, which was presented here, is very likely to benefit from such an implementation.
References
- Belz and Nelles (2015) Belz, J. and Nelles, O. (2015). Proposal for a function generator and extrapolation analysis. In 2015 International Symposium on Innovations in Intelligent SysTems and Applications (INISTA). IEEE.
- Cai et al. (2017) Cai, W., Zhang, M., and Zhang, Y. (2017). Batch Mode Active Learning for Regression With Expected Model Change. IEEE Transactions on Neural Networks and Learning Systems, 28.
- Cai et al. (2013) Cai, W., Zhang, Y., and Zhou, J. (2013). Maximizing Expected Model Change for Active Learning in Regression. In 2013 IEEE 13th International Conference on Data Mining. IEEE, Dallas, TX, USA.
- Cohn (1996) Cohn, D.A. (1996). Neural Network Exploration Using Optimal Experiment Design. Neural Networks, 9.
- Dursun et al. (2015) Dursun, Y., Kirschbaum, F., Jakobi, R., Gebhardt, A., Goos, J.C., and Rinderknecht, S. (2015). Ansatz zur adaptiven Versuchsplanung für die Längsdynamikapplikation von Fahrzeugen auf Prüfständen. In 6. Internationales Symposium für Entwicklungsmethodik.
- Hartmann and Nelles (2013) Hartmann, B. and Nelles, O. (2013). Adaptive Test Planning for the Calibration of Combustion Engines – Methodology. Design of Experiments (DoE) in Engine Development.
- Klein et al. (2013) Klein, P., Kirschbaum, F., Hartmann, B., Bogachik, Y., and Nelles, O. (2013). Adaptive Test Planning for the Calibration of Combustion Engines – Application. Design of Experiments (DoE) in Engine Development.
- MacKay (1992) MacKay, D.J.C. (1992). Information-Based Objective Functions for Active Data Selection. Neural Computation, 4.
- Park and Kim (2020) Park, S.H. and Kim, S.B. (2020). Robust expected model change for active learning in regression. Applied Intelligence, 50.
- Prochaska et al. (2020) Prochaska, A., Pillas, J., and Bäker, B. (2020). Active Output Selection Strategies for Multiple Learning Regression Models. arXiv:2011.14307 [cs].
- Rasmussen and Williams (2008) Rasmussen, C.E. and Williams, C.K.I. (2008). Gaussian processes for machine learning. MIT Press, Cambridge, Mass., 3. print edition.
- Reichart et al. (2008) Reichart, R., Tomanek, K., Hahn, U., and Rappoport, A. (2008). Multi-task active learning for linguistic annotations. In Proceedings of ACL-08: HLT.
- Settles (2009) Settles, B. (2009). Active Learning Literature Survey. Technical report, University of Wisconsin–Madison.
- Sugiyama and Rubens (2008) Sugiyama, M. and Rubens, N. (2008). Active Learning with Model Selection in Linear Regression. In Proceedings of the 2008 SIAM International Conference on Data Mining. Society for Industrial and Applied Mathematics.
- Tietze (2015) Tietze, N. (2015). Model-based Calibration of Engine Control Units Using Gaussian Process Regression. Ph.D. thesis, Technische Universität Darmstadt, Darmstadt.
- Zhang et al. (2016) Zhang, Y., Hoang, T.N., Low, K.H., and Kankanhalli, M. (2016). Near-optimal active learning of multi-output Gaussian processes. In Thirtieth AAAI Conference on Artificial Intelligence.