Active Learning-Based Multistage Sequential Decision-Making Model with Application on Common Bile Duct Stone Evaluation
Abstract
Multistage sequential decision-making scenarios are commonly seen in the healthcare diagnosis process. In this paper, an active learning-based method is developed to actively collect only the necessary patient data in a sequential manner. There are two novelties in the proposed method. First, unlike the existing ordinal logistic regression model which only models a single stage, we estimate the parameters for all stages together. Second, it is assumed that the coefficients for common features in different stages are kept consistent. The effectiveness of the proposed method is validated in both a simulation study and a real case study. Compared with the baseline method where the data is modeled individually and independently, the proposed method improves the estimation efficiency by 62%-1838%. For both simulation and testing cohorts, the proposed method is more effective, stable, interpretable, and computationally efficient on parameter estimation. The proposed method can be easily extended to a variety of scenarios where decision-making can be done sequentially with only necessary information.
keywords:
Active Learning; sequential decision-making; common bile duct stone; common coefficients assumption; incomplete data1 Introduction
Multistage sequential decision-making is commonly seen in many real-world applications for the sake of accuracy, efficiency, and cost-effectiveness, such as heathcare audit [1], optimal investment [2], energy [3], optimization [6], treatment [5], and design of experiments [4]. Meanwhile, active learning is a special case of machine learning in which a learning algorithm can interactively query a user or some other information source to label new data points with the desired outputs [7], and has been integrated with decision-making in recent years in various contexts such as imbalanced data [8, 9] and Bayesian decision-making [10]. In this paper, we aim at developing an active learning-based multistage sequential decision-making model to assist doctors making reliable diagnostic decisions and treatment recommendations in a cost-effective and convenient manner.
Our research is motivated from the evaluation of gallstone. Gallstone are solid particles that can form from cholesterol, bilirubin, and other substances within the gallbladder. These densities are often benign when localized to the gallbladder, but can cause pain, infection, and liver damage when they become stuck in the common bile duct (CBD) and impede the flow of bile into the digestive tract. A stone that becomes impacted in the CBD can be difficult to detect definitively, but its presence requires a procedural intervention, endoscopic retrograde cholangiopancreatography (ERCP). As with most procedures, ERCP requires anesthesia and carries the risk of complications, including pancreatitis, infection, and bleeding. As a result, it is imperative to ensure that ERCP is only performed when there is a stone definitively obstructing the duct. This is particularly true in children [11]. Thus, one important objective is how to efficiently predict the presence of a CBD stone in a child with high accuracy and specificity.
From the machine learning or statistical point of view, this might be straightforward: one would like to collect as many informative features as possible for each patient, which will lead to higher prediction accuracy. However, from the medical point of view, this is highly non-trivial, since some informative features could be expensive and hard to obtain in many real-world scenarios. Obtaining additional clinical data often comes with a price, including delayed time to intervention, cost, and complications. This price can be particularly high in pediatric care. For instance, while magnetic resonance imaging (MRI) is the most accurate way to diagnose a CBD stone, only 5% of pediatric patients in our real dataset have such diagnostic, as pediatric MRI often involves sedation, long scheduling time, etc. Indeed a less cost-effective approach such as laboratory tests or ultrasound (US) often might be sufficient. Thus a common clinical challenge is how to reduce the cost of data collection for diagnosis decision-making, yet avoid computational issues associated with limited or incomplete data.
In this article, a multistage sequential decision-making model is developed to assist doctors making reliable diagnostic decisions and treatment recommendations with high accuracy in a cost-effective and convenient manner. In our approach, some patients are allowed to take fewer examinations if their symptoms are apparent according to early-stage features, and some patients will be recommended to take more examinations if their symptoms are ambiguous and doctors have difficulty making reliable treatment recommendations.
There are several benefits to the proposed model. This novel platform allows the healthcare provider to actively collect only the necessary data sequentially rather than collecting all the data at inception. As a result, the diagnostic process can be faster, more convenient, and cost effective, while retaining significant accuracy. In the research setting, our model offers a novel solution to address challenges raised by incomplete data through offering sequential decision-making in an active learning way. The active learning-based model allows some missing data yet guarantees reliability and prediction accuracy, which makes the incomplete data missing “at control” rather than missing at random.
In the proposed framework, one important thing is to identify a proper predictive model for each stage as the building block. General categorization is required to utilize the model. Patients must be stratified to: certainly healthy (“0”), indeterminate (“0.5”), and certainly sick (“1”). After comparing the performance of different machine learning models, including neural networks, random forest, logistic regression, we selected the logistic regression model, which provides stability and accuracy with easy interpretability, see our previous work [11]. In the sections to follow, a generalized logistic regression model will be utilized as a building block for each stage.
Multiclass logistic regression, also called softmax/multinoulli/multinomial regression, is a classification method that generalizes logistic regression to multiclass problems (). The idea behind this classification model is one-vs-all or one-vs-rest, which means binary classifier models will be trained to ensure every class has a binary classifier to identify whether a sample belongs to this specific class or not. This model enables multiple outcomes but does not involve the ordering of the categorical dependent variable. The ordinal logistic regression model, also called the ordered logit model or proportional odds model, overcomes this limitation by calculating the log of the odds of cumulative events [14, 15, 16]. Unlike simple logistic regression, ordinal logistic models consider the probability of an event and all the events that are below the focal event in the ordered hierarchy. The existing ordinal logistic regression model satisfies our needs in single stages but failed to deal with multistage scenarios simultaneously. Therefore, in this research, we’ll develop our own multistage model with the existing ordinal logistic regression model as a building block for each stage, but estimate the parameters in all stages together.
The contribution of our work is twofold. In terms of statistical analysis, an active learning model is proposed in a multistage decision-making process. This learning algorithm can interactively query a patient collect additional clinical data for collecting necessary information [17]. Also, the proposed model introduces a novel way of dealing with incomplete data by allowing some data to be absent and modeling with incomplete but necessary data directly, rather than guessing the incomplete data with uncertainty. In terms of healthcare application, a general strategy is developed to help the doctor making diagnostic decisions and treatment recommendations sequentially when some laboratory data are more expensive, invasive, or time-consuming to collect than other data. By recommending requisite testing to those patients in whom symptoms do not support a definitive diagnosis, the proposed process is fast, cost-effective, yet reliable. Also, the methodology applies to a large variety of disease diagnose that involve a sequence of indirect examinations and some unpleasant or inaccurate examinations.
Note that our proposed model/method is closely related to incomplete data analysis in the statistics literature, where many widely-used methods have been developed, e.g., simple deletion on feature or record, mean/median/mode/zeros substitution, regression imputation, last observation carried forward, EM, etc. [13, 12]. However, these existing methods were developed from two primary approaches. One is deleting the records or features with missing values, which will lead to a waste of available data and some intrinsic problems associated with limited data size. The other approach is “guessing” or “learning” the missing data from the observed data, which can be dangerous for clinical modeling since the synthetic data brings extra uncertainties. Our approach is different from these existing methods in the sense that we actively decide which kinds of data to be observed and which kinds of data to be missing.
In Section 2, the proposed model is described. Section 3 discusses the novel approach to parameter estimation for the proposed model. In Section 4, the performance of the proposed methodology on synthetic data is reported. Finally, in section 5, the analytical result on real clinical data is demonstrated.
2 Proposed Active Learning Multistage Sequential Decision-Making Model
In this section, we first present a high-level framework of our proposed active learning-based multistage sequential decision-making model, and then introduce necessary notation to represent observed data in our framework. Finally, our model is rigorously developed based on the existing ordinal logistic regression model.
2.1 Model Framework

The proposed multistage sequential decision-making model is shown in Fig 1. To achieve the goal of optimizing the diagnostic process, under our proposed model, the features of a patient’s clinical data are classified into several different stages according to the relevant cost, invasiveness, time, patient’s willingness to participate, etc. The earlier stages involve features that are patient-friendly, inexpensive, though potentially less informative. The latter stages could involve some more time-consuming, expensive, unpleasant, but potentially more informative examinations. As in group sequential clinical trials, if we can make a definitive yes or no decision at a given stage, then we will stop taking features/observations and make the corresponding decision; but if we are not sure, then we will move to the next stages by taking additional features/observations from the patient.
To be more specific, for the -th stage, based on the cumulative features, an ordinal logistic regression model is trained as the classifier to estimate the probability of sickness for each patient (). For every classifier, the corresponding optimal cutoff points, , or for the last stage, will be augment clinical decision making: if , patient will be labeled as healthy (“0”) and will not undergo further intervention; if , patient will be labeled as sick (“1”), they will proceed to the recommended intervention, such as ERCP or surgery; otherwise, the patient will be labeled as indeterminate (“0.5”) and undergo further diagnostic workup. Following the last stage, as no further diagnostic examinations are available, the provider must make a treatment recommendations based on a single threshold, , and the patient will be classified as either healthy or sick. By quantifying the probability of being sick via predictive models based on currently available features, the diagnosis could be more objective and stable compared with an empirical estimation.
In the evaluation of a newly presenting patient, this framework with trained parameters can be utilized to arrive at a diagnosis and recommend treatments accordingly. In such a way, the necessary features will be collected sequentially and actively.
In the proposed multistage model, ordinal logistic regression is adopted as the basic predictive model in each stage. A novel feature of this model is, the patients’ groups in the different stages have a nested structure rather than independent. Further, the features in the current stage include all those of the former stages and newly added features in the given category . This allows us to consider the predictive models in different stages together rather than in an isolated fashion. The details will be discussed further in the parameter estimation discussion in Section 3.
2.2 Data Organization for 2-stage Model as Example
To illustrate the proposed active learning style multistage sequential decision-making model, the 2-stage decision-making model will be elaborated as an illustrative example. As discussed earlier, all the capturable features are first divided into several categories, each corresponds to a certain stage in our multistage model. The criteria for feature grouping are flexible and can be adjusted according to real scenarios case-by-case: 1) For the features inside a category, the cost of collection and the information contained therein should be similar. For example, all indices can be obtained from one examination such as a blood test. 2) Each category should contain some useful information, and excess categories should be avoided. For example, there is no need to treat each variable as a separate category, especially less informative variables.
To illustrate the data organization for the proposed model, without loss of generality, we can take the 2-stage sequential decision-making model as an example as follows. In the observed (training) data, denote by and the number of patients in stage 1 and stage 2, respectively, and denote by and the number of features collected in the two stages. In the first stage, the observed features or explanatory variables are denoted by and the corresponding decision variable. , where implies that the more features need to be collected on patient in stage 2 for diagnosis. In stage 2, the new observed features or explanatory variables are denoted by and the corresponding decision variable Note that the feature set from stage 1 will be used in stage 2 cumulatively, and thus the number of features in stage 2 is actually .
Hence, for the 2-stage model, the (training) data can be summarized as:
-
•
Stage 1:
-
•
Stage 2:
For a new patient, after observing we need to predict When we predict this is the final decision. When we predict we will then take more observations from this patient in the stage 2, i..e, and make another prediction which will be our final decision.
2.3 Our Proposed Model
In our proposed model, the building block is an ordinal logistic regression model, which can serve as a filter in each stage to split patients into 3 groups based on the ordered probability of being sick. For better presentation, let us first provide a quick review of ordinal logistic regression, which can be seen a form of a generalized linear model (GLM) and an extension of binary logistic regression. Under ordinal logistic regression model, we assume that there is an underlying continuous latent variable
(1) |
where the error term has a logistic distribution with the cumulative distribution function (CDF) Of course, other distributions of ’s are possible such as Gaussian/normal distribution. Then the observable response variable will fall into several ordinal categories based on the values of latent variable and the cutoff points.
Now we are ready to present our proposed 2-stage sequential decision-making model. We assume that there are two latent variables, and for the two stages as defined in (1). Then the decision at each stage is given by
(2) |
Note that the cutoff points applying on the latent variable . This is equivalent to applying the cutoff points on the probability estimation in the Fig 1 after going through a link function transformation.
Recall that in our proposed multistage model, the output at the first stage represents healthy, indeterminate, and sick, respectively. At the last stage, indicates healthy and sick. It is useful to mention that the outputs are ordered, i.e.
A subtlety in our proposed multistage model is the relationship of the parameter in (1) across different stages. A computationally simple baseline approach is to assume that they can be completely different, where we can simply apply ordinal logistic regression models to estimate parameters each stage individually and independently. Unfortunately, this does not consider the nested structure and potential correlations between the two adjacent stages, and thus might cause mis-understanding and confusing interpretation in practice. In our paper, we will make a further assumption that and enjoys the nested structure in the sense of having the same values with the same features, i.e.,
Under this common coefficients assumption, the newly added features from additional stages will not affect the weights of existing ones, which makes the model more interpretable for doctors and patients.
3 Parameter Estimation
At the high-level, we propose to use the maximum likelihood estimation (MLE) method to estimate not only the unknown ’s values in different stages, but also the cutoff values . Note that the coefficients ’s can also be estimated by the penalized likelihood estimation framework such as LASSO for variable selection when there are a large of available features, which is beyond the scope of this article and will be presented elsewhere. Here we assume that we pre-screen all features, and only use those useful features in our model and medical decision making.
To better present our main ideas, let us begin with a single stage with a ordinal logistic regression model. To simplify the notation, we assume that the response results from the underlying latent variable by using cutoff points :
(3) |
where and are introduced for the technical reason to simplify the notation. Also note that at the last stage, we have , which will force one to make a decision or Under this notation, the conditional distribution of is given by
(4) | ||||
where are model coefficients and cutoff points to be estimated from the data, is the abbreviation of probability, and is the CDF of the error term which takes on the role of the inverse link function.
For a given stage or a given ordinal logistic regression model with subjects, the joint log-likelihood function becomes
It has been well investigated to maximize this log-likelihood objective function which possesses several nice properties. McCullagh presented a Fisher scoring algorithm for ML estimation, expressing the likelihood function using cumulative probabilities. McCullagh showed that a sufficiently large guarantees a unique maximum of the likelihood [14, 19]. Burridge and Pratt showed that the log-likelihood is concave for many cumulative link models, including the logit, probit, and complementary log-log [19]. Iterative algorithms usually converge rapidly to the ML estimates. The optimization computing algorithms are developed based on these useful properties. The existing approaches compute the MLE by iteratively reweighted least squares (IRLS). Actually, for packages in software, all GLM are fit using the IRLS estimation method for the log-likelihood.
Now let us go back to the parameter estimation of our proposed multistage model. For simplicity, denote as , and as . Under our notation, the joint log-likelihood function for all samples in all stages will be as follows:
where is the number of stages in the proposed model, is the number of patients in the stage, and is the number of categories in the stage for which we have
Given the joint likelihood function , the next step is to simultaneously estimate , all the parameters in the stages, that maximizes the likelihood function. Here we propose to adopt the gradient descent algorithm and a key step is to compute the gradient for each parameter, which can be written as:
where is the probability density function (PDF) of : .
We should also remark some technical details in computation. Note that the coefficient in stage , , contains and newly added coefficients for features in the stage. This makes the coefficients in different stages of different lengths. To address this issue, we propose finding the gradient in each stage individually first. Then, adding them together with the front ends aligned and pad the shorter ends with 0’s: . In addition, the indicator function can be replaced with a mask matrix with boolean values. Alternatively, the indicator function can be avoided by data preprocessing. For instance, we can split the patients into different groups to mark the identity, e.g., using to denote the group of patients who have label in stage , see Appendix A for more details.
To numerically solve our proposed joint-optimization problem, we follow the framework of the existing function in the package MASS in R software, and adopt the function as optimization solver. The is a function designed for general-purpose optimization based on Nelder–Mead, quasi-Newton and conjugate-gradient algorithms. In , the Broyden–Fletcher–Goldfarb–Shanno algorithm (BFGS) method, an iterative method for solving unconstrained nonlinear optimization problems [20], is adopted, since it works best with analytic gradients based on our extensive experiences.
4 Simulation Study
In this section, a simulation is conducted to verify the effectiveness and efficiency of the proposed methodology. First, a 2-stage synthetic dataset, is generated with prescribed coefficients and classification thresholds . Then, we conduct both a baseline approach and the proposed approach on the synthetic dataset to demonstrate the performance of the proposed method in the aspect of prediction accuracy, stability, interpretability, and relative efficiency.
Let denote the total number of synthetic patients at the outset. We design the features of patients as follows: , , , , , , . For stage 1, we design feature set as: . For stage 2, we add 3 new features: . The error term follows standard logistic regression for both stages: . Note that if the error term follows logistic regression with other parameters, the estimation of and will be scaled, see Eq. (2). The coefficients are designed as: which satisfys the common coefficients assumption. Next, we can calculate the latent variable according to Eq. (1). To cut the into ordered categories, we design the cutoff points as . The value and are chosen since and , where and are the designed cutoff points on probability . Via such design, we have . Given , the corresponding true label can be obtained according to Eq. (2).
In the baseline approach, the 2 stages proceed independently: in stage 1, a standard ordinal logistic regression is applied on via function in R; and in stage 2, a standard logistic regression is applied on via the function in R. While in the proposed method, all the parameters in 2 stages are estimated together via MLE. To provide reliable statistical inference, we adopt the Monte-Carlo method and run the data generation and parameter estimation 100 times.

The comparison of the mean coefficients estimated from the two methods is visualized in Fig. 2. Clearly, the mean coefficients estimated from the two methods are similar and both coincide well with ground truth, indicating that both methods have little or no bias on mean parameter estimation, proving the effectiveness of both methods on estimating. To quantify the subtle difference, we can utilize mean square error (MSE) to measure the distance between the estimated coefficients and the ground truth. For coefficients in stage 1, the MSE of the estimation from the baseline method and the proposed method are and , respectively. Meanwhile, for stage 2, the MSE of the baseline method and the proposed method are 0.0011 and 0.0003, respectively. This indicates a slight advantage of the proposed methods on mean coefficients estimation. See Appendix B for more comparison on mean prediction metrics.
Fig. 4 shows the standard deviation (std) of coefficients estimation from the two methods. It can be seen that the standard deviation of estimation from the proposed method is significantly smaller than those from baseline method in both stages. To quantify this advantage, the relative efficiency of the proposed method with respect to the baseline method, defined as , is plotted in Fig. 4, where and represent the estimators for in the proposed method and baseline method, respectively. It can be seen that the relative efficiency of the estimator in the proposed method for the coefficients in stage 1 is very big, as high as 19.38. And for the estimator for the coefficients in stage 2, the lowest relative efficiency is still . This indicates that the efficiency of the parameter estimation is improved in the proposed methodology. The reason for such improvement is straightforward: in the proposed methods, all available data is involved in the parameter estimation, while in the baseline method, only a small subset of the data is involved in later stages.


For the estimation of cutoff points, the baseline method and the proposed method give the mean estimations as and , respectively. And the standard of deviation of the estimations by the baseline and proposed methods are and , respectively. Compared with the ground truth, , both methods are functional, yet the proposed approach yields a lower mean deviation and standard deviation.
Overall, the proposed 2-stage model provides a more accurate and efficient estimation since it involves all data in the parameter estimation. Also, in implementation, the proposed algorithm enjoys an advantage with regard to computational convenience since we estimated all parameters in at once. Additionally, the common coefficients assumption in the proposed method reduces the number of parameters to be estimated from to since the coefficients for features in stage 1 are not repeatedly estimated in later stages, but instead assumed to be consistent.
5 Case Study
In this section, we analyze a real data set in [11] and demonstrate the usefulness of our proposed approach in the sense of being able to learn from data to effectively evaluate gallstone. In particular, the real clinical data is modeled as a 2-stage sequential decision-making model. We have patients in the first stage. Based on the missing profile, we filter the patients who have features in category 2 into the second stage and label them as . As a result, . Note that here we made the assumption that if a patient has available data in the next stage, the patient should be labeled as “0.5” at the current stage. The reason for such assumption is that we believe that the provider only recommends further diagnostic interventions to a patient who has indeterminate presentation at the current stage (“0.5”). In other words, the real data comes from the provider’s knowledge and experience, and the missing pattern is not missing at random but deliberately interfered with. Note that features in include hemolytic disease, gender, age, body mass index (BMI), total bilirubin, alanine transaminase (ALT), aspartate transaminase (AST), lipase, amalyse, gamma-glutamyl transferase (GGT), alkaline phosphatase. These features include demographic and laboratory tests. For features in , based on , it involves imaging data, including CBD diameter and the presence of CBDS by ultrasound (US) and magnetic resonance imaging (MRI) examinations, respectively. See Table 1 in Appendix C for descriptive statistics.

Fig. 5 visualizes the coefficients for features in category 1 estimated by baseline and proposed approach, respectively. It can be seen that with the baseline approach, the coefficients estimated by the two individual models are similar for most features, , but significantly different for . In the proposed 2-stage model, the coefficients estimated for feature seem to resemble a weighted average of the coefficients obtained from baseline individual models. Also, compared with the baseline method, the coefficients estimated by the proposed method have no dramatic spikes. Also, with the common coefficients assumption, newly added features will not affect the weights of existing ones. More importantly, in the proposed method, no matter how many stages a provider utilized, the calculation can be done at once efficiently.
Our model and method can be useful for providing the treatment recommendation for a new patient, and a diagnostic mobile App was developed in 2021 in Apple Store based on this research. To be concrete, when a new patient presents, given the trained proposed method, the provider (e.g., clinician or doctor), can first recommend the patient undergo diagnostic tests included in category 1, and calculate the accordingly. Based on the and , the patient will be labeled as healthy (“1”), sick (“0”), or indeterminate and proceed to the next stage (“0.5”). Then, only when the patient’s presentation is indeterminate, additional testing such as ultrasound will be undertaken. In other words, only the necessary data are actively collected sequentially for sequential decision-making. This approach simplifies and expedites the diagnostic workup. Additionally, the provider can adjust the multistage sequential decision-making model flexibly by designing the number of stages and the features in each stage, also see Section 2 for the details on the feature grouping recommendations.
6 Conclusion and Discussion
In this study, to augment healthcare clinical decision-making based on limited examinations, a multistage sequential decision-making model is developed to actively collect only minimal necessary diagnostic data.To illustrate the proposed method, we optimized a 2-stage model on synthetic data in the simulation study and applied the proposed algorithm to pediatric patients with suspicion for obstructive common bile duct stones in an effort to more efficiently and accurately arrive at a diagnosis that allowed expedited intervention with ERCP. In both the simulation study and real case study on collected pediatric data, the proposed method is more effective, stable, interpretable, and simpler in estimating the coefficients and cutoff points for all stages. It estimates all parameters simultaneously and fully considers the cumulative characteristics between these stages by making common coefficients assumption.
One limitation of the proposed methodology is that it involves all variables without selection. A potential future improvement to the model will be the addition of regularization term to the objective function to realize the variable selection.
Also, we can extend our proposed algorithm to satisfy more real-world cases with special requirements. For example, consideration of expense, delay to intervention, and need for invasive testing can be specified and minimized with modification to the objective function as , where the are the relevant cost per patient in stage 1 and stage 2, respectively. is the tuning parameter to adjust the weight between the negative likelihood and the relevant cost to the test. Note that here the can also be variables to be optimized over since the relevant cost depends on the number of patients in each stage.
In healthcare settings, the penalty for misdiagnosis can vary based on the patient and diagnosis. For example, false-positive detection may lead to unnecessary procedures or excess intervention, while false-negative detection may result in missed or delayed treatment. For such cases with weighted penalties on different types of misclassification, we can modify the objective function by specifying the penalties as , where the and are the tuning parameter for false-negative prediction and false-positive prediction, respectively.
The proposed algorithm is also applicable to a broad range of decision-making scenarios outside of healthcare as it utilizes active learning to drive the sequential collection of minimal requisite information, rather than all possible information simultaneously. This will be an area of future development.
References
- [1] T. Ekin and R.M. Musal, Integrated statistical and decision models for multi-stage health care audit sampling, J. Appl. Stat.(2021), pp. 1–19.
- [2] G. Sirbiladze, I. Khutsishvili, and B. Ghvaberidze, Multistage decision-making fuzzy methodology for optimal investments based on experts’ evaluations, Eur. J. Oper. Res. 232.1(2014), pp. 169–177.
- [3] H. Gerking, Modeling of multi-stage decision-making processes in multi-period energy-models, Eur. J. Oper. Res. 32.2 (1987), pp. 191–204.
- [4] A.R.G Mukkula, M. Mateáš, M. Fikar, and R. Paulen, Robust multi-stage model-based design of optimal experiments for nonlinear estimation, Comput. Chem. Eng. (2021).
- [5] Y. Tao, and L. Wang, Adaptive contrast weighted learning for multi-stage multi-treatment decision-making, Biometrics 73.1 (2017), pp. 145–155.
- [6] B. Han, C. Shang, and D. Huang, Multiple kernel learning-aided robust optimization: Learning algorithm, computational tractability, and usage in multi-stage decision-making, Eur. J. Oper. Res. 292.3 (2021), pp. 1004–1018.
- [7] B. Settles, Active learning literature survey, (2009).
- [8] I Sundin, P. Schulam, E. Siivola, A. Vehtari, S. Saria, and S. Kaski, Active learning for decision-making from imbalanced observational data, ICML (2019), pp. 6046–6055.
- [9] J. Lee, Y. Wu, and H. Kim, Unbalanced data classification using support vector machines with active learning on scleroderma lung disease patterns J. Appl. Stat. 42.3 (2015), pp. 676–689.
- [10] L. Filstroff, I. Sundin, P. Mikkola, A. Tiulpin, J. Kylmäoja, and S. Kaski, Targeted Active Learning for Bayesian Decision-Making, arXiv preprint arXiv:2106.04193 (2021).
- [11] R.Z. Cohen, H. Tian, C.G. Sauer,F.F Willingham, M.T. Santore, Y. Mei, A.J. Freeman, Creation of a Pediatric Choledocholithiasis Prediction Model, J. Pediatr. Gastroenterol. Nutr. , (2021).
- [12] X. Ma, and Q. Zhong, Missing value imputation method for disaster decision-making using K nearest neighbor, J. Appl. Stat. 43.4(2016), pp. 767-781.
- [13] H. Kang, The prevention and handling of the missing data, Korean J. of Anesthesiol. 64.5 (2013), pp. 402.
- [14] P. McCullagh, Regression models for ordinal data, J. R. Stat. Soc. Series B42.2 (1980), pp. 109-127.
- [15] H.S. Kim, Topics in ordinal logistic regression and its applications, Texas A&M University (2004).
- [16] G.S. Watson, Generalized Linear Models (P. Mccullagh and JA Nelder), SIAM Review, 28.1 (1986), pp. 128-130.
- [17] F. Olsson F. A literature survey of active machine learning in the context of natural language processing, (2009).
- [18] A. Agresti, and C. Tarantola, Simple ways to interpret effects in modeling ordinal categorical data, Stat. Neerl. 72.3 (2018), pp. 210-223.
- [19] A. Agresti, Categorical data analysis, John Wiley & Sons (2003).
- [20] M. Avriel M, Nonlinear programming: analysis and methods, Courier Corporation (2003).
- [21] C.G. Broyden, J.E. Dennis Jr, J.J Moré, On the local and superlinear convergence of quasi-Newton methods. IMA J Appl. Math. 12.3 (1973), pp. 223-245.
Appendix
Appendix A – Alternative Notation for Likelihood and Gradient for 2-Stage Case
Take our 2-stage illustrating case as an example, the log-likelihood function can be written as:
(5) | ||||
Since here , to avoid any potential confusing notation, we rewrite them as:
Given the joint log-likelihood function, we can write the derivative for each coefficient in individually accordingly.
For the common coefficients, for , we have:
(6) | ||||
For the coefficients for newly added features in stage 2, for , we have:
(7) |
For the cutoff points in all 2 stages, we have:
(8) |
(9) |
(10) |
Appendix B – Mean Prediction Metrics Comparison of the Baseline Method and Proposed Method in Simulation Study
Fig. 6 shows the mean prediction metrics for prediction for 2 stages with coefficients estimated from the two methods respectively. The evaluation metrics include sensitivity (Sensi.), specificity (Speci.), positive predictive values (PPV), negative predictive values (NPV), and balanced accuracy, precision, recall, F1 value, prevalence, detection rate (Detec. Rate), detection prevalence (Detec. Prev.), and balanced accuracy (Bal. Accu). From the predictive performance comparison in Fig. 6, it can be seen that the proposed method has a significant advantage for the prediction for stage 2 with almost higher values in every index, especially the prevalence, and detection rate, and detection prevalence. And for the prediction for stage 1, there’s no significant difference between the proposed method and baseline method, as we can see that the blue and black lines are too close to each other to distinguish them from each other visually.

Appendix C – Table 1. Descriptive Statistics for Demographic, Laboratory, and Imaging Data in Case Study
CBD Stone Present | No CBD Stone Present | p-value | |
Total patients | 120 | 196 | |
IOC | 81 | 196 | |
ERCP | 39 | 0 | |
Hemolytic disease | |||
Female | |||
Age | |||
BMI | |||
Total bilirubin | |||
ALT (U/L) | |||
AST (U/L) | |||
Lipase (U/L) | |||
Amylase (U/L) | |||
GGT (U/L) | |||
Alkaline phosphatase (U/L) | |||
CBD diameter (mm, by US) | |||
Presence of CBD stone (US) | |||
CBD diameter (mm, by MRI) | |||
Presence of CBD stone (MRI) |