This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Conformal Depression Prediction

Yonghong Li and Xiuzhuang Zhou This work was funded in part by the National Natural Science Foundation of China under Grants 61972046 (Corresponding author: Xiuzhuang Zhou). Yonghong Li and Xiuzhuang Zhou are with the School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing 100876, China (Email: [email protected]).
Abstract

While existing depression prediction methods based on deep learning show promise, their practical application is hindered by the lack of trustworthiness, as these deep models are often deployed as black box models, leaving us uncertain on the confidence of their predictions. For high-risk clinical applications like depression prediction, uncertainty quantification is essential in decision-making. In this paper, we introduce conformal depression prediction (CDP), a depression prediction method with uncertainty quantification based on conformal prediction (CP), giving valid confidence intervals with theoretical coverage guarantees for the model predictions. CDP is a plug-and-play module that requires neither model retraining nor an assumption about the depression data distribution. As CDP provides only an average coverage guarantee across all inputs rather than per-input performance guarantee, we further propose CDP-ACC, an improved conformal prediction with approximate conditional coverage. CDP-ACC firstly estimates the prediction distribution through neighborhood relaxation, and then introduces a conformal score function by constructing nested sequences, so as to provide a tighter prediction interval adaptive to specific input. We empirically demonstrate the application of CDP in uncertainty-aware facial depression prediction, as well as the effectiveness and superiority of CDP-ACC on the AVEC 2013 and AVEC 2014 datasets. Our code is publicly available at https://github.com/PushineLee/CDP.

Index Terms:
facial depression prediction, uncertainty quantification, conformal prediction, approximate conditional coverage.

I Introduction

Depression is a prevalent mental disorder presented as persistent feelings of sadness, debilitation, and loss of interest in activities [1, 2, 3]. It increases the risk of suicide and accounts for a substantial psychological burden [4, 5]. The existing primary diagnostic methods used involve mental health reports, such as the Beck Depression Inventory (BDI-II) [6] (target variable focused on in this paper), the Hamilton Depression Rating Scale (HRSD) [7], and the Patient Health Questionnaire (PHQ-8) [8]. The diagnostic process depends on interviews and demands considerable time and effort from both psychiatrists and patients. Also, it heavily relies on the clinicians’ subjective experience, as well as the patients’ cognitive abilities and psychological states.

Over the past decade, continuous efforts have been made to explore effective biomarkers and methods for automated depression prediction. For depression biomarkers, initial attention was focused on hand-crafted features such as Local Binary Patterns (LBP) [9], Local Gabor Binary Patterns from Three Orthogonal Planes (LGBP-TOP) [10], and Local Binary Pattern from Three Orthogonal Planes (LBP-TOP) [11, 12] extracted from facial videos. Handcrafted features extracted from other behavioral signals have also been employed in this task, including speech, facial action units, facial landmarks, head poses, and gazes [13, 14]. The development of these hand-crafted features heavily relies on specific knowledge related to the depression prediction tasks. Acquiring such knowledge is time-consuming, subject to researchers’ subjective cognition and task specificity, and lacks good generalization. Moreover, these features struggle to identify certain implicit and difficult-to-distinguish patterns of depression [15]. Fortunately, with the advent of deep learning [16], a new pathway has been opened for depression prediction tasks. In this pipeline end-to-end deep neural networks can be trained using depression data including facial video [17, 18, 19, 20], speech [21, 22, 13] and other data sources. These deep models are expected to discover subtle, indistinguishable depression-related features and implicitly discriminate among them for improved prediction.

Refer to caption
Figure 1: Uncertainty quantification for facial depression prediction can provide an additional confidence interval to score prediction at any given confidence level, such as 90%111To protect personal privacy, we show facial landmarks instead of the original facial images..

While the rapid development of deep learning and computer vision has led to exciting progress in depression prediction, the reliability and trustworthiness of deep learning models-considered black boxes [23, 24, 25]-have not been well addressed. These models may exhibit overconfidence in predicting failures, potentially leading to catastrophic risks. For automated depression prediction, the accuracy performance should not be the only criterion determining the usefulness of a predictive model. The lack of statistically rigorous uncertainty quantification is a key factor undermining the reliability of depression prediction. Consider the application context of depression prediction. We are more willing to accept reliable predictions that provide insight into potential risks. Beyond the depression severity prediction, we need additional outcomes to represent their uncertainty. As shown in Fig. 1, our aim is to provide statistically rigorous confidence intervals in depression prediction. These intervals should cover the actual depression scores with any user-defined level of confidence.

In this work, we propose conformal depression prediction (CDP) to quantify the uncertainty of depression severity estimation based on conformal prediction (CP [26, 27]). CDP can provide confidence intervals to satisfy marginal coverage for the actual depression scores. We show that, CDP as a plug-and-play module, can validly quantify uncertainty for most depression prediction models, without retraining the model. Furthermore, inspired by the adaptive prediction sets (APS) [28] in uncertainty-aware classification tasks, which construct a prediction set per-input adaptive to its prediction distribution, we propose conformal depression prediction with approximate conditional coverage (CDP-ACC) to overcome the shortcomings of conformal prediction with marginal coverage, providing adaptive confidence intervals for depression prediction.

The main contributions are summarized as follows:

1. We propose an uncertainty quantification framework CDP for facial depression prediction. CDP provides statistically rigorous confidence intervals that satisfy marginal coverage for most existing facial depression prediction models.

2. We develop CDP-ACC, an improved CDP algorithm that satisfies approximate conditional coverage, yielding tighter confidence intervals while maintaining good coverage.

3. We conduct extensive experiments to demonstrate the effectiveness of our proposed method in uncertainty quantification of facial depression prediction.

II RELATED WORK

II-A Depression Prediction

Over the past decade, depression prediction based on deep learning has attracted significant interest and attention in the field, leading to many meaningful research advancements. Studies using labeled depression data from various modalities, including speech, text, video, and EEG signals, as well as the integration of these modalities, have taken a leading role.

For single-modality depression prediction, He et al. employed a Deep Convolutional Neural Network (DCNN) to extract depression-related features from speech data [21]. These learned features were subsequently integrated with additional hand-crafted features to predict depression scores. Zhao et al. introduced a hybrid network that integrates Self-Attention Networks (SAN) and DCNN to handle both low-level acoustic features and 3D Log-Mel spectrogram data [22]. Following this, feature fusion techniques along with support vector regression are employed for depression prediction. Melo et al. concentrated on efficiently extracting spatiotemporal features from facial videos associated with depression [17, 18]. They introduced a deep Multiscale Spatiotemporal Network (MSN) and a deep Maximization-Differentiation Network (MDN) to predict depression levels. Sharma et al. proposed CNN-LSTM hybrid networks to extract depression features from EEG signals [29]. Seal et al. developed the DeprNet, a DCNN designed for classification of EEG data from depressed and normal subjects [30].

For multimodal depression prediction, Niu et al. introduced a novel Spatio-Temporal Attention (STA) network and a Multimodal Attention Feature Fusion (MAFF) strategy to acquire a multimodal representation of depression cues from facial and speech spectrum [31]. Uddin et al. introduced a novel deep multi-modal framework that effectively integrates facial and verbal cues for automated depression prediction [32]. They extracted features separately from video and speech and employed a Multi-modal Factorized Bilinear pooling (MFB) strategy to efficiently fuse these multi-modal features. Niu et al. proposed a novel multi-layer perceptron (MLP) architecture named DepressionMLP, designed to automatically predict depression levels through facial keypoints and action units (FKRS and AURS) [33].

These studies, either using single-modality data or multimodal data, primarily focus on the prediction accuracy, while overlooking the analysis of predictive uncertainty. Interestingly, works such as those of Melo et al. introduced minimizing expected error to learn distribution of depression levels [34], while Zhou et al. proposed DJ-LDML to learn depression representation using a label distribution and metric learning [20]. These studies implicitly explore uncertainty in depression prediction, yet without providing a clear and intuitive uncertainty quantification. The method closely related to ours is the one proposed by Nouretdinov et al. [35], which leveraged conformal prediction for uncertainty quantification of MRI-based depression classification. Different from the work [35] that dedicated to the classification task, we focus on regression task in the context of deep learning, and we further develop an approximate conditional coverage (ACC) algorithm to overcome the shortcomings of conformal prediction with marginal coverage.

II-B Uncertainty Estimation in Affective Computing

Uncertainty has been widely considered for various tasks in affective computing, such as emotion recognition, facial expression recognition, and apparent personality recognition [36, 37, 38, 39, 40, 41, 42]. In emotion recognition, Harper et al. utilized a Bayesian framework to support emotional valence classification from the perspective of uncertainty [36]. Prabhu et al. proposed BNNs to model the label uncertainty based on subjectivity in emotion recognition [37]. Wu et al. employed deep evidence regression to jointly model aleatoric and epistemic uncertainties, aiming to enhance the performance of emotion attribute estimation [38]. Tellamekala et al. introduced calibrated and ordinal latent distribution to capture modality-wise uncertainty for multimodal fusion in emotion recognition [43]. Lo et al. utilized probabilistic uncertainty learning to extract information from low-resolution data and proposed the emotion wheel to learn label uncertainty, yielding robust facial expression recognition [39]. In facial expression recognition, She et al. adopted a pairwise uncertainty estimation approach, assigning lower confidence scores to more ambiguous samples to address annotation ambiguities [40]. Le et al. proposed a method based on uncertainty-aware label distribution to adaptively construct the training sample distribution [41]. Additionally, Tellamekala et al. utilized a neural latent variable model to model aleatoric and epistemic uncertainty and integrated both to enhance the performance of apparent personality recognition [42].

The uncertainty-inspired algorithms mentioned above may not be suitable for depression prediction, as the available datasets with annotated depression scores are limited. Re-stabilizing a well-performing, uncertainty-inspired depression prediction algorithm is not straightforward. Furthermore, while the above methods may achieve better accuracy performance in their respective tasks by introducing uncertainty-inspired heuristics, they do not establish a rigorously valid uncertainty quantification. In contrast to previous methods, our proposed CDP is a plug-and-play uncertainty quantification method for depression prediction, which does not require model retraining and does not assume the depression data distribution. Thanks to the conformal prediction [26, 44], our CDP can achieve valid confidence intervals with theoretical coverage guarantees for depression predictions.

Refer to caption
Figure 2: The pipeline of CDP. CDP takes as input a pre-trained depression prediction model f^\hat{f}, a calibration set 𝒟cal\mathcal{D}_{cal} and a given confidence level 1α1-\alpha. CDP sets the prediction biases (on the calibration set) as the conformal scores and employs the (1α)(1-\alpha)-th empirical quantile of the conformal scores to form confidence interval for the prediction f^(Xtest)\hat{f}(X_{test}).

On the other hand, current conformal prediction methods tend to produce unnecessarily conservative intervals, resulting in the confidence interval to be unnecessarily wide [45]. In depression prediction, overly wide intervals are meaningless. For instance, the prediction with BDI-II (ranging from 0 to 63) has a 100% confidence interval of [0, 63]. Clearly, such a predictive uncertainty lacks practicality. Additionally, marginal coverage guarantees only valid average intervals, which overcovers simple subgroups and undercovers difficult ones [46]. Differently, our proposed conformal prediction method with approximate conditional coverage (CDP-ACC) can produce valid and tighter confidence interval adaptive to the specific depression prediction.

III METHOD

III-A Preliminaries

Formally, let us assume that we have the facial depression-associated video data denoted by XX as input and the corresponding label denoted by yy (the BDI-II score [6], for example). We randomly split the data into disjoint training set 𝒟train\mathcal{D}_{train}, calibration set 𝒟cal={(Xi,yi)}i=1N\mathcal{D}_{cal}=\{(X_{i},y_{i})\}_{i=1}^{N}, and test set 𝒟test\mathcal{D}_{test}. We use 𝒟train\mathcal{D}_{train} to train a DNN model f^()\hat{f}(\cdot) for depression prediction. Here, y^\hat{y}\in\mathbb{R} denotes the point prediction for XX by f^()\hat{f}(\cdot). In the deep learning context, we can formulate the depression prediction as a deep regression task by minimizing the mean squared error (MSE) loss as the learning objective.

MSE=12|𝒟train|i=1|𝒟train|(y^iyi)2\begin{split}\mathcal{L}_{MSE}=\frac{1}{2|\mathcal{D}_{train}|}\sum_{i=1}^{|\mathcal{D}_{train}|}(\hat{y}_{i}-y_{i})^{2}\end{split} (1)

where |𝒟train||\mathcal{D}_{train}| denotes the number of data in 𝒟train\mathcal{D}_{train}.

In depression prediction based on such regression models, the point prediction y^\hat{y} is the expectation 𝔼[y|X]\mathbb{E}[y|X] derived from estimating the conditional probability distribution P(y|X)P(y|X). However, we do not have access to the oracle of P(y|X)P(y|X). To quantify the uncertainty of depression prediction y^\hat{y}, we expect to construct an interval function 𝒞α(X)\mathcal{C}_{\alpha}(X), which takes depression data XX and miscoverage rate α\alpha as input and outputs a confidence interval. The interval contains the depression target yy with any confidence level 1α1-\alpha, as shown in Fig. 1 (α=0.1\alpha=0.1). Clearly, wider interval |𝒞α(X)||\mathcal{C}_{\alpha}(X)| indicates greater uncertainty for the model’s prediction y^\hat{y}.

III-B Depression Prediction with UQ

The straightforward method for uncertainty quantification of depression prediction is to train a DNN to estimate the conditional probability distribution P(y|X=x)P(y|X=x) of depressed individuals. Given our lack of knowledge regarding the true P(y|X=x)P(y|X=x), one possible solution is to assume that the depression target yy follow a Gaussian distribution y𝒩(y^,σ2)y\sim\mathcal{N}(\hat{y},\sigma^{2}), and uses the negative log-likelihood (NLL) [47] as the loss function:

NLL=i=1|𝒟train|logσi22+(yiy^i)22σi2\mathcal{L}_{NLL}=\sum_{i=1}^{|\mathcal{D}_{train}|}\frac{\text{log}\sigma_{i}^{2}}{2}+\frac{(y_{i}-\hat{y}_{i})^{2}}{2\sigma_{i}^{2}} (2)

The learning objective allows us to train a DNN capable of predicting the distribution P(y|X=x)P(y|X=x) for any given individual xx. Furthermore, under the Gaussian assumption, the α\alpha-th quantile can be calculated by

y^α=y^+2σerf1(2α1)\hat{y}_{\alpha}=\hat{y}+\sqrt{2}\sigma\cdot\mathrm{erf}^{-1}(2\alpha-1) (3)

where erf1\mathrm{erf}^{-1} denotes the inverse Gaussian error function. In this way, we can obtain the confidence interval [y^α/2,y^1α/2][\hat{y}_{\alpha/2},\hat{y}_{1-\alpha/2}] that satisfies P(y[y^α/2,y^1α/2])1αP(y\in[\hat{y}_{\alpha/2},\hat{y}_{1-\alpha/2}])\geq 1-\alpha.

In practice, however, depression datasets (e.g., AVEC 2013 and AVEC 2014 [48, 49]) exhibit high data imbalance, and the assumption of the Gaussian distribution may not be appropriate in this context. We prefer a distribution-free method for uncertainty quantification. Indeed, by introducing the pinball loss (Eq. (4)), we can train a DNN as a conditional quantile estimator [50, 51], allowing us to obtain arbitrary quantile values for P(y|X=x)P(y|X=x):

pinball={q(yy^q),if yy^q(1q)(y^qy),if y<y^q\mathcal{L}_{pinball}=\begin{cases}q\cdot(y-\hat{y}_{q}),&\text{if }y\geq\hat{y}_{q}\\ (1-q)\cdot(\hat{y}_{q}-y),&\text{if }y<\hat{y}_{q}\end{cases} (4)

where q[0,1]q\in[0,1], y^q\hat{y}_{q} denotes the qq-th quantile of the prediction, and yy is the ground truth. Using the trained regressor, we can use α/2{\alpha/2} and 1α/2{1-\alpha/2} quantile values of the predictive distribution to derive the confidence interval [y^α/2,y^1α/2][\hat{y}_{\alpha/2},\hat{y}_{1-\alpha/2}], thereby achieving a distribution-free uncertainty quantification.

Input: 𝒟cal={Xi,yi}i=1N\mathcal{D}_{cal}=\{X_{i},y_{i}\}_{i=1}^{N}: calibration set;
f^()\hat{f}(\cdot): pre-trained depression prediction model;
XtestX_{test}: test data;
α(0,1)\alpha\in(0,1): miscoverage rate;
𝒬(,)\mathcal{Q}(\cdot,\cdot): quantile estimator
Output: Confidence interval 𝒞α(Xtest)\mathcal{C}_{\alpha}(X_{test})
1
// Calculate the conformal scores {si}i=1N\{s_{i}\}_{i=1}^{N} on 𝒟cal\mathcal{D}_{cal}
2 for i=1i=1 to NN do
3      y^i=f^(Xi)\hat{y}_{i}=\hat{f}(X_{i})
4       si=|y^iyi|s_{i}=|\hat{y}_{i}-y_{i}|
5 end for
6
7q=N+1N(1α)q=\lceil\frac{N+1}{N}\rceil(1-\alpha)
// Calculate the qq-th quantile of the conformal scores
8
9s^=𝒬({si}i=1N,q)\hat{s}=\mathcal{Q}(\{s_{i}\}_{i=1}^{N},q)
// Calculate the confidence interval
10 𝒞α(Xtest)=[f^(Xtest)s^,f^(Xtest)+s^]\mathcal{C}_{\alpha}(X_{test})=[\hat{f}(X_{test})-\hat{s},\hat{f}(X_{test})+\hat{s}]
return 𝒞α(Xtest)\mathcal{C}_{\alpha}(X_{test})
Algorithm 1 Conformal Depression Prediction

However, due to limited depression data and imbalanced data distributions, using Eq. (2) or Eq. (4) to train depression prediction models is challenging. In practice, it has been observed that those methods suffer from overconfidence, where P(y𝒞α(X))P(y\in\mathcal{C}_{\alpha}(X)) can be significantly lower than 1α1-\alpha. In this context, there is an urgent need for a post-processing method that is distribution-free to quantify the uncertainty of depression prediction. In particular, we seek for constructing the confidence interval 𝒞α()\mathcal{C}_{\alpha}(\cdot) to guarantee the marginal coverage so as to maintain an average coverage rate across the target, as expressed by

P(ytest𝒞α(Xtest))1α,(Xtest,ytest)𝒟testP(y_{test}\in\mathcal{C}_{\alpha}(X_{test}))\geq 1-\alpha,\ \ \ (X_{test},y_{test})\in\mathcal{D}_{test} (5)

To that end, we introduce conformal prediction (CP [26, 52]) for uncertainty quantification of facial depression prediction. Specifically, we develop conformal depression prediction (CDP) as shown in Fig. 2. As premise of CP is to have a calibration set that is exchangeable with the test sample (This is guaranteed if the i.i.d. assumption holds), we calculate the conformal score ss on the calibration set,

si=|yiy^i|s_{i}=|y_{i}-\hat{y}_{i}| (6)

which reflects the discrepancy between the depression prediction and the ground truth. Next, we calculate the qq-th quantile of ss and denote it as s^\hat{s}. For any test sample XtestX_{test}, we have P(stests^)=qP(s_{test}\leq\hat{s})=q. In this way, we can obtain its prediction intervals 𝒞α(Xtest)=[y^tests^,y^test+s^]\mathcal{C}_{\alpha}(X_{test})=[\hat{y}_{test}-\hat{s},\hat{y}_{test}+\hat{s}], which is valid with

1αP(ytest𝒞α(Xtest)1α+1N+11-\alpha\leq P(y_{test}\in\mathcal{C}_{\alpha}(X_{test})\leq 1-\alpha+\frac{1}{N+1} (7)

The implementation of CDP are summarized in Algorithm 1. With CDP, we can acquire valid, theoretical coverage guaranteed confidence interval, and the interval width relates to the user-defined confidence levels 1α1-\alpha. Note that this CP-based uncertainty quantification achieves marginal coverage for the target depression scores at any confidence levels, as proven in [26, 52].

III-C CDP with Approximate Conditional Coverage

There is a stronger notion than the marginal coverage, namely conditional coverage, as defined by

P(ytest𝒞α(y|Xtest))1αP(y_{test}\in\mathcal{C}_{\alpha}(y|X_{test}))\geq 1-\alpha (8)

For the model prediction y^test\hat{y}_{test} on any test sample XtestX_{test}, it aims to provide a confidence interval CαC_{\alpha} with 1α1-\alpha coverage. In contrast to marginal coverage, conditional coverage is more suitable for uncertainty quantification in individual depression prediction. In practical applications, each individual only cares about whether the prediction is trustworthy (or within which confidence interval the prediction falls). Unfortunately, achieving conditional coverage is practically challenging in a distribution-free setting [44]. Our work draws inspiration from ordinal adaptive prediction sets [53] for classification tasks. The motivation behind our CDP-ACC is to generate adaptive confidence intervals for pre-trained depression predictor, achieving approximate conditional coverage in a model-agnostic and distribution-free manner.

Refer to caption
Figure 3: The pipeline of CDP-ACC. The predictions on the calibration set are firstly partitioned into MM subintervals. For each subinterval, CDP-ACC estimates the approximate conditional distribution P(y|y^)P(y|\hat{y}), computes the conformal score based on the calibration samples in that interval, and finally constructs the adaptive confidence interval.

III-C1 Approximate Conditional Coverage

Given a pre-trained depression predictor with point prediction, y^=f^(X)\hat{y}=\hat{f}(X), we cannot directly access to the conditional distribution P(y|X)P(y|X). To accurately estimate P(y|X)P(y|X) in depression prediction, we rely on a key assumption [54]: similar XX (in feature embedding space) will have similar conditional distributions P(y|X)P(y|X) and similar conditional expectation. This assumption is reasonable in our practical observations. For example, different video clips of the same patient tend to have similar depression predictions.

Let Δ(Xi)\Delta(X_{i}) denote the neighborhood of XiX_{i} in the feature embedding space. Our aim is to approximate P(y|X=Xi)P(y|X=X_{i}) with P(y|XΔ(Xi))P(y|X\in\Delta(X_{i})). In the context of facial depression prediction, XiX_{i} denotes the facial video clips, and Δ(Xi)\Delta(X_{i}) can represent clips belonging to the same video, or clips with the same depression score. However, for unseen test video clips, we do not have oracle with access to Δ(Xi)\Delta(X_{i}). We need to further explore relaxation scheme to approximate Δ(Xi)\Delta(X_{i}).

We revisit Δ(Xi)\Delta(X_{i}) from the perspective of the model predictions (with pre-trained model f^\hat{f}): Δ(yi^)=f^(Δ(Xi))\Delta(\hat{y_{i}})=\hat{f}(\Delta(X_{i})). The model predicts similar depression scores for what it perceives as similar XiX_{i}. Thus, we can use Δ(yi^)\Delta(\hat{y_{i}}) as the condition instead of Δ(Xi)\Delta(X_{i}), that is, similar predictions y^\hat{y} can also have similar conditional distributions. With this approximation, we have P(y|X=Xi)P(y|y^Δ(y^))P(y|X=X_{i})\approx P(y|\hat{y}\in\Delta(\hat{y})), and we can estimate the conditional distribution by grouping neighboring points of the depression predictions. Moreover, by incorporating the approximation, we can introduce inductive bias, which implies that for samples falling in Δ(y^)\Delta(\hat{y}), the more concentrated the points, the higher the credibility.

We uniformly divide the range of y^\hat{y} into MM subintervals Bm,m=1,,MB_{m},\ m=1,...,M. Let ym,Ly_{m,L} and ym,Uy_{m,U} denote the lower and upper bounds of yy in each subinterval mm. Based on the approximation P(y|X=Xi)P(y|y^Bm)P(y|X=X_{i})\approx P(y|\hat{y}\in B_{m}), we can only use the depression label yy (corresponding to y^\hat{y} within BmB_{m}) to approximately estimate the conditional distribution P(y|X=Xi)P(y|X=X_{i}). Following [52, 53], we use histogram binning to estimate the conditional probability distribution Pm(y|y^Bm)P_{m}(y|\hat{y}\in B_{m}) for each subinterval mm (shown in Fig. 4).

Refer to caption
Figure 4: Histogram estimation of the conditional distribution P(y|y^Δ(y^))P(y|\hat{y}\in\Delta(\hat{y})). For samples falling in y^Δ(y^)\hat{y}\in\Delta(\hat{y}), the more concentrated the points, the higher the credibility.

III-C2 Confidence Interval

In conformal prediction, setting the conformal score is particularly important. The choice conformal score function essentially determines the effectiveness of the conformal prediction method. In CDP-ACC, setting the conformal score involves constructing nested sequence of approximate oracle intervals.

To maintain the continuity of prediction interval, we construct a nested sequence E()E(\cdot) so that the interval with higher confidence covers the interval with lower confidence, all while endeavoring to minimize the width of each interval. As shown in Fig. 5, for each yiy_{i} whose prediction falls into BmB_{m} we can construct the confidence interval Em(τ,yi)E_{m}(\tau,y_{i}) with the shortest width:

Em(τ,yi)=argmin[yl,yh][ym,L,ym,U],yi[yl,yh]\displaystyle E_{m}(\tau,y_{i})=\mathop{\arg\min}\limits_{\begin{subarray}{c}[y_{l},y_{h}]\subset[y_{m,L},y_{m,U}],\\ y_{i}\in[y_{l},y_{h}]\end{subarray}} {yhyl:\displaystyle\{y_{h}-y_{l}: (9)
Fm(yh|y^)Fm(yl|y^)τ}\displaystyle F_{m}(y_{h}|\hat{y})-F_{m}(y_{l}|\hat{y})\geq\tau\}

where τ(0,1)\tau\in(0,1) denotes the confidence level, m[1,M]m\in[1,M] and FmF_{m} denotes the cumulative probability distribution of the conditional probability distribution Pm(y|y^Bm)P_{m}(y|\hat{y}\in B_{m}). If the optimal solution is not unique, we can add a little noise ϵ\epsilon into the BmB_{m} to break the ties [55]. It should be noted that, given yiy_{i}, Em(τ,yi)E_{m}(\tau,y_{i}) is a nested sequence as τ\tau grows, for example, Em(0.1,yi)Em(0.2,yi)Em(0.3,yi)E_{m}(0.1,y_{i})\subseteq E_{m}(0.2,y_{i})\subseteq E_{m}(0.3,y_{i}). For any sample yiy_{i} whose prediction falls into BmB_{m}, we set the width of its interval EmE_{m} as the conformal score sis_{i}:

si(m,α)=|Em(1α,yi)|s_{i}(m,\alpha)=|E_{m}(1-\alpha,y_{i})| (10)

The conformal score typically reflects the prediction bias. Higher scores indicate poorer prediction quality while wider interval corresponds to worse predictions in our context.

Refer to caption
Figure 5: The setting of the conformal score for CDP-ACC. Given the approximate conditional distribution P(y|y^)P(y|\hat{y}), various prediction intervals can be acquired containing yiy_{i} with 1α1-\alpha coverage, of which the width of the shortest one is set as the conformal score for our CDP-ACC.

After calculating the conformal scores on the calibration set, we can estimate the confidence interval for any test samples based on the assumption that the calibration set and the test samples are exchangeable to generate valid confidence intervals. A good confidence interval should achieve the specified coverage rate at least. Narrower interval widths are desirable as they provide more precise uncertainty quantification. For Em(τ,yi)E_{m}(\tau,y_{i}), if the user sets the confidence level τ\tau to 100%, the confidence interval for any samples yiy_{i} will be unique (Em(1,yi)E_{m}(1,y_{i})). Otherwise, there will be multiple prediction intervals that cover yiy_{i} with at least 1α1-\alpha probability. According to our smoothness approximation, the concentration degree of samples within a subinterval can represent prediction uncertainty. We use interval width as the conformal score and identify the narrowest interval that covers yiy_{i} with 1α1-\alpha probability as its confidence prediction interval. Specifically, given all conformal scores calculated on the calibration set and a user-defined miscoverage rate α\alpha, CDP-ACC constructs confidence interval for any unseen test sample XtestX_{test} according to the following steps. First, for each subinterval mm, we calculate the (1α)(1-\alpha)-th quantile of the conformal scores {si(m,α)}\{s_{i}(m,\alpha)\}.

s^(m,α)=𝒬({si(m,α)},1α)\hat{s}(m,\alpha)=\mathcal{Q}(\{s_{i}(m,\alpha)\},1-\alpha) (11)

Then, for each subinterval mm we can obtain its confidence interval with shortest width according to

𝒜(m,α)=argmin[yl,yh][ym,L,ym,U],yi[yl,yh]{yhyl:si(m,α)s^(m,α)}\displaystyle\mathcal{A}(m,\alpha)=\mathop{\arg\min}\limits_{\begin{subarray}{c}[y_{l},y_{h}]\subset[y_{m,L},y_{m,U}],\\ y_{i}\in[y_{l},y_{h}]\end{subarray}}\{y_{h}-y_{l}:s_{i}(m,\alpha)\leq\hat{s}(m,\alpha)\} (12)

Finally, based on the interval into which the prediction of the test sample falls, we can retrieve the corresponding confidence interval. The proposed CDP-ACC is summarized in Algorithm 2.

Input: 𝒟cal={Xi,yi}i=1N\mathcal{D}_{cal}=\{X_{i},y_{i}\}_{i=1}^{N}: calibration set;
f^()\hat{f}(\cdot): pre-trained depression prediction model;
XtestX_{test}: test data;
α(0,1)\alpha\in(0,1): miscoverage rate;
L,UL,U: lower and upper bounds of the predictions on 𝒟cal\mathcal{D}_{cal};
()\mathcal{H}(\cdot): histogram estimator;
𝒬(,)\mathcal{Q}(\cdot,\cdot): quantile estimator;
MM: number of the subintervals
1
Output: Confidence interval 𝒞α(Xtest)\mathcal{C}_{\alpha}(X_{test})
2 Initialize subintervals {Bm}m=1M\{B_{m}\}_{m=1}^{M}\leftarrow\emptyset
// Subintervals dividing for y^\hat{y} on 𝒟cal\mathcal{D}_{cal}
3
4for i=1i=1 to NN do
5       y^i=f^(Xi)\hat{y}_{i}=\hat{f}(X_{i})
6      for m=1m=1 to MM do
7             if y^i[L+m1M(UL),L+mM(UL))\hat{y}_{i}\in[L+\frac{m-1}{M}(U-L),L+\frac{m}{M}(U-L)) then
8                   Bm{yi}B_{m}\cup\{y_{i}\}
9             end if
10            
11       end for
12      
13 end for
14
// Estimation of the conditional distribution
15 for m=1m=1 to MM do
16       Pm(y|y^)P_{m}(y|\hat{y}) = (Bm)\mathcal{H}(B_{m})
17       Fm(y|y^)F_{m}(y|\hat{y}) = ym,LyPm(t|y^)𝑑t\int_{y_{m,L}}^{y}P_{m}(t|\hat{y})dt
18 end for
19
// Calculate the confidence interval
20
21y^test=f^(Xtest)\hat{y}_{test}=\hat{f}(X_{test})
22 for m=1m=1 to MM do
23       Nm=|Bm|N_{m}=|B_{m}|
       // Calculate conformal score according to Eq. (9) and Eq. (10)
24       {si(m,α)}i=1Nm={|Em(1α,yi)|}i=1Nm\{s_{i}(m,\alpha)\}_{i=1}^{N_{m}}=\{|E_{m}(1-\alpha,y_{i})|\}_{i=1}^{N_{m}}
25      
      // Calculate the (1α1-\alpha)-th quantile of the conformal scores
26       s^(m,α)=𝒬({si(m,α)}i=1Nm,1α)\hat{s}(m,\alpha)=\mathcal{Q}(\{s_{i}(m,\alpha)\}_{i=1}^{N_{m}},1-\alpha)
       // Calculate the interval according to Eq. (12)
27       [ym,l,ym,h]=𝒜(m,α)[y_{m,l},y_{m,h}]=\mathcal{A}(m,\alpha)
28      
29 end for
30if y^testBm\hat{y}_{test}\in B_{m} then
31       𝒞α(Xtest)=[ym,l,ym,h]\mathcal{C}_{\alpha}(X_{test})=[y_{m,l},y_{m,h}]
32       return 𝒞α(Xtest)\mathcal{C}_{\alpha}(X_{test})
33 end if
Algorithm 2 Depression Prediction with CDP-ACC

IV EXPERIMENTS

IV-A Dataset

In order to verify the effectiveness of our uncertainty quantification methods in depression prediction, we conduct experiments on two commonly used facial depression datasets, AVEC 2013 [48] and AVEC 2014 [49]. AVEC 2013 is a subset of the AViD-Corpus, consisting of 150 facial videos. The videos range in length from 20 to 50 minutes, with a frame rate of 30 FPS and a resolution of 640 ×\times 480 pixels. The content of the videos includes facial recordings of subjects during human-computer interaction tasks, such as reading assigned content and improvising on given sentences. We divide the dataset into 70 videos for training, 30 videos for validation (calibration), and 50 videos for testing. The facial depression levels are annotated using BDI-II scores [6], with each video assigned a single value representing the depression level. The BDI-II questionnaire consists of 21 questions, each with multiple-choice answers scored from 0 to 3. The total BDI-II score ranges from 0 to 63: 0-13 indicates minimal depression, 14-19 indicates mild depression, 20-28 indicates moderate depression, and 29-63 indicates severe depression, as shown in Table IV.

AVEC 2014, a subset of AVEC 2013, consists of two tasks: Northwind and Freeform. It comprises 150 Northwind videos and 150 Freeform videos. The original videos of AVEC 2014 are similar to those used in the AVEC 2013 but include five pairs of previously unseen videos, replacing a few videos considered unsuitable. We use 200 videos for training, 50 videos for validation (calibration), and 50 videos for testing.

The videos in the AVEC 2014 dataset are relatively short, with an overlap of 8 frames between adjacent video clips, whereas there is no overlap in the AVEC 2013 dataset. During data processing, landmarks in all video frames are detected using OpenFace [56], and facial region alignment is performed using the eyes, nose, and mouth.

IV-B Experimental Setup

IV-B1 Prediction Model

This work focuses on methods for quantifying uncertainty in depression prediction rather than on the predictive performance of the models themselves. We use the classic C3D [57, 19] and the SlowFast [58] networks as baseline models for facial depression prediction. In particular, for vanilla regression, we employ MSE loss as the learning objective during training and modify the last FC layer from 4096 to 64 to enhance training stability and prevent overfitting. In the context of Quantile Regression (QR [50]), we use 99 quantiles evenly spaced between 0.01 and 0.99 for the pinball loss. The output of the last FC layer transforms from a single output in vanilla regression to 99 outputs in QR.

IV-B2 UQ for Depression Prediction

To quantify the uncertainty of depression prediction models, we perform NLL [47] and QR [50] that achieves UQ by retraining the model. We also implement several post-processing methods related to CP [26, 44] to construct valid confidence intervals for depression prediction. Specifically, CQR [45] is a post-processing method that applies conformal prediction to calibrate the interval of pre-trained QR. Its conformal score is set as max{f^(X;α/2)y,yf^(X;1α/2)}max\{\hat{f}(X;\alpha/2)-y,y-\hat{f}(X;1-\alpha/2)\}. CDP represents the method for uncertainty quantification in vanilla regression, as shown in Fig. 2. Its conformal score is set as |yy^||y-\hat{y}|, which achieves marginal coverage. CDP-ACC is a method that achieves approximate conditional coverage and provides valid and adaptive confidence intervals. We set the miscoverage rate α\alpha to 0.1. Given that the depression score ranges from 0 to 63, we use truncation operator to ensure the predictive interval [y^l,y^h][0,63][\hat{y}_{l},\hat{y}_{h}]\subseteq[0,63]. For CDP-ACC, we set the subintervals number MM to 14, lower bound LL to 0 and upper bound UU to 63. To validate the generalizability of our method, we also evaluate our UQ method on seven publicly available datasets for general regression task (for more information, please refer to the Appendix).

IV-B3 Metrics

In depression prediction, mean absolute error (MAE) and root mean squared error (RMSE) are commonly used as the evaluation metrics, aiming to quantify the accuracy performance. To further measure the inherent uncertainty in predictions, we adopt the prediction interval coverage probability (PICP) and mean prediction interval width (MPIW) [51]. PICP and MPIW are two seemingly contradictory metrics. For example, if the PICP worsens, the MPIW may actually improve. We aim to minimize MPIW while ensuring PICP coverage rate. If we cannot ensure that PICP is not less than (1α1-\alpha) ×\times 100%, merely shortening the interval width is meaningless. Similarly, constructing an interval with a larger width to ensure coverage is also meaningless.

PICP=1|𝒟test|i=1|𝒟test|𝟏(yi𝒞α(Xi))PICP=\frac{1}{|\mathcal{D}_{test}|}\sum_{i=1}^{|\mathcal{D}_{test}|}\mathbf{1}(y_{i}\in\mathcal{C}_{\alpha}(X_{i})) (13)
MPIW=1|𝒟test|i=1|𝒟test||y^i,hy^i,l|MPIW=\frac{1}{|\mathcal{D}_{test}|}\sum_{i=1}^{|\mathcal{D}_{test}|}|\hat{y}_{i,h}-\hat{y}_{i,l}| (14)

where Xi𝒟testX_{i}\in\mathcal{D}_{test} and [y^i,l,y^i,h]=𝒞α(Xi)[\hat{y}_{i,l},\hat{y}_{i,h}]=\mathcal{C}_{\alpha}(X_{i}).

TABLE I: Prediction errors of the regression models with different training losses and backbones
Method Backbone Dataset MAE RMSE
MSE C3D AVEC 2013 7.14 8.97
NLL [47] C3D AVEC 2013 7.02 8.92
QR [50] C3D AVEC 2013 6.82 8.64
MSE C3D AVEC 2014 6.54 8.32
NLL [47] C3D AVEC 2014 7.10 9.24
QR [50] C3D AVEC 2014 6.65 8.24
MSE SlowFast AVEC 2013 7.49 9.37
NLL [47] SlowFast AVEC 2013 7.30 9.21
QR [50] SlowFast AVEC 2013 7.22 9.04
MSE SlowFast AVEC 2014 7.00 8.70
NLL [47] SlowFast AVEC 2014 6.29 8.11
QR [50] SlowFast AVEC 2014 6.89 8.65
TABLE II: Uncertainty quantification performance of different methods
Method Backbone Dataset PICP MPIW
NLL [47] C3D AVEC 2013 74.33% 13.17
QR [50] C3D AVEC 2013 17.61% 4.73
CQR [45] C3D AVEC 2013 87.94% 23.32
CDP C3D AVEC 2013 91.78% 22.82
CDP-ACC C3D AVEC 2013 92.00% 21.48
NLL [47] C3D AVEC 2014 76.52% 14.15
QR [50] C3D AVEC 2014 21.84% 5.35
CQR [45] C3D AVEC 2014 87.14% 23.53
CDP C3D AVEC 2014 91.53% 27.47
CDP-ACC C3D AVEC 2014 93.72% 20.17
NLL [47] SlowFast AVEC 2013 66.94% 17.69
QR [50] SlowFast AVEC 2013 22.30% 5.38
CQR [45] SlowFast AVEC 2013 92.29% 27.78
CDP SlowFast AVEC 2013 95.25% 28.29
CDP-ACC SlowFast AVEC 2013 91.25% 24.13
NLL [47] SlowFast AVEC 2014 69.57% 19.59
QR [50] SlowFast AVEC 2014 70.49% 16.83
CQR [45] SlowFast AVEC 2014 93.06% 27.76
CDP SlowFast AVEC 2014 93.86% 26.28
CDP-ACC SlowFast AVEC 2014 92.18% 23.29
α=0.1\alpha=0.1

Furthermore, we utilize size-stratified coverage (SSC [59, 46]) to assess the extent to which different methods achieve conditional coverage for subsets of varying depression severity. We divide all possible test data into GG bins based on the severity of depression (shown in Table IV), denoted as 1,2,,G\mathcal{I}_{1},\mathcal{I}_{2},\ldots,\mathcal{I}_{G}.

SSC=ming{1,..,G}1|g|ig𝟏{yi𝒞α(Xi)}SSC=\underset{g\in\{1,..,G\}}{\min}\frac{1}{|\mathcal{I}_{g}|}\sum_{i\in\mathcal{I}_{g}}{\mathbf{1}\{y_{i}\in\mathcal{C}_{\alpha}(X_{i})\}} (15)

where g{1,2,,ntest}\mathcal{I}_{g}\subset\{1,2,\ldots,n_{\text{test}}\} represents the set of sample indices falling into bin gg, |g||\mathcal{I}_{g}| denotes the size of g\mathcal{I}_{g} and 𝟏()\mathbf{1}(\cdot) denotes the indicator function.

IV-C Results

In this section, we provide a comprehensive evaluation of uncertainty quantification methods for depression prediction. Firstly, we show the prediction errors for vanilla regression, NLL and QR, allowing for a comparison of their accuracy performance in depression prediction. Then, we compare the uncertainty quantification performance of different UQ methods in depression prediction. We also perform uncertainty analysis of CDP-ACC for some instances to gain insight into CDP-ACC. Additionally, in the Appendix, we experimentally evaluate the proposed method on several general regression datasets.

TABLE III: Conditional coverage performance of different methods on AVEC 2014
Method Backbone SSC
NLL [47] C3D 0.2426
QR [50] C3D 0.1715
CQR [45] C3D 0.8280
CDP [26] C3D 0.7035
CDP-ACC C3D 0.8980
NLL [47] SlowFast 0.5293
QR [50] SlowFast 0.4623
CQR [45] SlowFast 0.8343
CDP [26] SlowFast 0.8140
CDP-ACC SlowFast 0.8407
α=0.1\alpha=0.1

IV-C1 Results of the UQ

Table I shows the prediction errors of the C3D and SlowFast models under different datasets and training losses. We can find that comparing vanilla regression using MSE or NLL as the learning objective, the prediction error for quantile regression is slightly smaller in AVEC 2013. This indicates the QR with the 50% quantile has a smaller prediction error than the vanilla regression and NLL. However, in the experiments with the SlowFast model on AVEC 2014, the NLL yields smaller errors compared to the other two methods. This also indicates that if only the prediction accuracy is considered, the NLL method as well as the QR may present better performance.

Table II shows the results of several uncertainty quantification methods with different depression prediction models. Due to NLL and QR heavily relying on the training process, it’s intuitive to observe that the coverage of QR is the worst, followed by NLL. In the current context of depression prediction, unstable training process and imbalanced datasets make it difficult to obtain effective quantile values or fitting Gaussian distributions. For pre-trained QR, CQR is a valid method to calibrate coverage by adjusting the interval width. In this way, CQR can provide adaptive confidence intervals, but its adaptiveness primarily stems from QR, while the conformal prediction mainly provides theoretical guarantees for marginal coverage. In situations where QR exhibits overconfidence, CQR unavoidably relies on average marginal coverage. As for our CDP-ACC, it is an approximate condition coverage method for regression. It can reduce the interval width as much as possible by providing adaptive confidence intervals while maintaining coverage.

TABLE IV: BDI-II scores and depression severity levels
BDI-II Score Severity Level
0-13 minimal
14-19 mild
20-28 moderate
29-63 severe
Refer to caption
Figure 6: The prediction intervals for CDP and CDP-ACC on some sampled video clips with 90% confidence (α\alpha = 0.1). C3D is used as the depression predictor. To protect privacy, we show facial landmarks of the first frame to represent the facial video clips.

In Fig. 6, we present several facial video samples with varying levels of depression and their prediction uncertainties, where the predictor uses the C3D regression model. From the figure, we can see that although the predictor exhibits different prediction biases on samples with different levels of depression, CDP and CDP-ACC provide valid confidence intervals for the model’s predictions. A wider interval indicates potentially larger uncertainty in the model’s predictions. It can also be seen that compared to CDP, CDP-ACC provides tighter confidence intervals while maintaining coverage. On the other hand, we can observe that even though the predictor may have small prediction biases on some samples (e.g., the second example from the left in Fig. 6), its prediction uncertainty can still be large. This reminds us that in clinical practice, we cannot focus solely on prediction accuracy. From this perspective, our proposed uncertainty quantification scheme offers new insights and perspectives for assessing potential risks in depression prediction.

IV-C2 Evaluation of the Conditional Coverage

Table III shows the conditional coverage of different UQ methods (with α=0.1\alpha=0.1) on test instances with various levels of depression severity, which are categorized into four groups (G=4G=4, shown in Table IV). A SSC closer to 1α1-\alpha indicates better conditional coverage of the UQ method. We can clearly observe that NLL and QR perform the worst in terms of SSC due to unstable training and overconfidence. As an effective method for quantifying depression uncertainty, CDP can only guarantee marginal coverage, hence its results are not as good as CQR. However, the conditional coverage of CQR is limited by the pre-trained QR model. In contrast, CDP-ACC can directly approximate conditional coverage based on the point prediction-based vanilla regression models, which does not require retraining a QR model and achieves better conditional coverage.

Refer to caption
Figure 7: PICP of CDP-ACC with varying MM on AVEC 2014.
Refer to caption
Figure 8: MPIW of CDP-ACC with varying MM on AVEC 2014.

IV-C3 Parameter Analysis

The effectiveness of CDP-ACC largely depends on the accuracy of conditional distribution estimation. In our histogram estimation, the parameter MM affects this accuracy. A smaller MM implies a coarser histogram estimation, while a larger MM results in a finer estimation but reduces the number of sample points within each subinterval mm, which can also hinder the effectiveness of the approximation. In our experiments, half of the calibration set is used as the validation set for tuning the parameter MM. As shown in Figs. 7 and 8, the optimal value of the parameter (M=14M=14) can be obtained by tuning on the validation set.

V CONCLUSION

By introducing conformal prediction, we propose a plug-and-play uncertainty quantification method CDP to produce confidence intervals for facial depression prediction. The intervals predicted by CDP can achieve marginal coverage guarantee at any given miscoverage rate. Furthermore, we develop CDP-ACC for facial depression prediction. By approximate conditional coverage, it alleviates the issue of marginal coverage practices focusing solely on the average coverage while overlooking individuals with great uncertainty, and provides valid and adaptive confidence intervals for better uncertainty quantification in facial depression prediction. Although our proposed method is used for facial depression prediction, it is also applicable to various other data modalities. We expect that this work will attract more researchers to pay attention to the issue of uncertainty quantification for automated depression prediction in modern context.

To further validate the effectiveness of our CDP-ACC for general regression tasks, we conducted supplementary experiments on seven publicly available datasets, as also used in [45, 52]. These seven datasets are: Physicochemical Properties of Protein Tertiary Structure (bio) [60], BlogFeedback (blog) [61], Facebook Comment Volume Dataset [62], variants one (fb1) and two (fb2), from the UCI Machine Learning Repository [63], and Medical Expenditure Panel Survey (meps19 - meps21) [64]. Regarding the preprocessing of the datasets and the experimental setup, we followed the settings in [45, 52], with detailed information available on GitHub.

Datasets

The Physicochemical Properties of Protein Tertiary Structure (bio) dataset [60] falls under the subject heading of biology. It consists of data extracted from CASP 5-9 and contains 45,730 decoys ranging in size from 0 to 21 angstroms. This dataset is commonly used for multivariate regression analysis.

The BlogFeedback (blog) dataset [61] belongs to the field of social science. It comprises 60,021 instances, each consisting of features extracted from blog posts and a target variable representing the number of comments the post received within 24 hours. This task is a multivariate regression task commonly used to predict the number of comments a post will receive.

The Facebook Comment Volume Dataset (fb1, fb2) [62] is similar to the blog dataset in that it aims to predict the number of comments on a post. It comprises 40,949 instances, each with 53 features. The dataset includes 5 variants; however, we only utilized the first two variants (fb1 and fb2) in our experiment.

The Medical Expenditure Panel Survey (meps 19 - 21) [64] is a series of large-scale surveys conducted on households, individuals, healthcare providers, and employers across the United States. MEPS serves as the most comprehensive data source on healthcare expenses, utilization, and the extent of medical insurance coverage. In our experiments, we exclusively utilized data from panels 19 to 21.

Implementation Details

Our experimental setup follows [45, 52], where we train a conditional quantile estimator (implemented as a three-layer neural network) to handle the regression task described above. We employ the pinball loss (Eq. 4) as our loss function and ensure that the output quantiles are sorted to prevent crossing.

We train the neural network using the Adam optimizer with a learning rate of 0.02. The output quantiles serve as part of the post-processing calibration for conformal prediction. Unlike SCP [26, 44, 27] and CDP-ACC, which do not rely on quantiles, we typically treat the 50% quantile as the output for vanilla regression. We evaluate using Prediction Interval Coverage Probability (PICP) and Mean Prediction Interval Width (MPIW) as metrics, with a miscoverage rate α\alpha set to 0.1. For specific definitions of these evaluation metrics, please refer to our main paper.

Results

We present the uncertainty quantification results of several methods, including SCP [26, 44, 27], QR [50], CQR [45], CHR [52], and CDP-ACC, on the aforementioned general datasets. QR and CQR are introduced in the main paper. CHR is a histogram estimation method that calibrates the QR method by constructing nested intervals and integrating them with conformal prediction.

From Table V, it is clear that SCP generally achieves marginal coverage but tends to construct unnecessarily conservative intervals. In contrast, in general regression tasks, QR maintains normal coverage rates and effectively reduces the mean interval width (However, in depression prediction task, QR exhibits a trend of overconfidence, with quantile values overly concentrated around predictions, leading to lower coverage rates.)

CQR and CHR serve as calibration methods for QR, aiming to minimize interval width while maintaining adequate coverage rates. However, CQR is constrained by the simplicity of QR in constructing conformal scores. In contrast, CHR integrates histogram estimation with conformal prediction, enabling more refined optimization of prediction intervals.

For QR, CQR, and CHR, retraining the QR model is crucial. However, vanilla regression models with point prediction can only allow for obtaining prediction intervals through post-processing methods like SCP. Nevertheless, SCP has significant drawbacks such as fixed interval width and overly conservative intervals (unnecessarily wide). CDP-ACC, as an adaptive post-processing method, can be applied to both QR and vanilla regression models. It achieves a finer approximation of conditional distributions through binning, thereby further reducing interval width. From Table V, we observe that CDP-ACC achieves competitive uncertainty quantification results, demonstrating its effectiveness on general regression tasks..

TABLE V: Results of Uncertainty Quantification.
Methods Data PICP MPIW
SCP [26, 44, 27] bio 90.10% 14.87
QR [50] bio 90.05% 14.53
CQR [45] bio 90.15% 14.56
CHR [52] bio 90.35% 13.13
CDP-ACC bio 89.55% 12.26
SCP [26, 44, 27] blog 91.20% 16.10
QR [50] blog 92.35% 14.65
CQR [45] blog 91.20% 14.65
CHR [52] blog 89.85% 10.24
CDP-ACC blog 90.02% 9.56
SCP [26, 44, 27] fb1 88.25% 16.15
QR [50] fb1 92.00% 15.68
CQR [45] fb1 90.00% 15.67
CHR [52] fb1 89.60% 12.23
CDP-ACC fb1 89.15% 12.11
SCP [26, 44, 27] fb2 90.05% 18.51
QR [50] fb2 92.55% 15.44
CQR [45] fb2 89.95% 15.43
CHR [52] fb2 90.25% 12.14
CDP-ACC fb2 89.70% 10.27
SCP [26, 44, 27] meps19 90.05% 32.55
QR [50] meps19 89.80% 31.23
CQR [45] meps19 90.95% 31.37
CHR [52] meps19 90.04% 22.73
CDP-ACC meps19 89.90% 17.00
SCP [26, 44, 27] meps20 88.85% 23.74
QR [50] meps20 88.45% 26.47
CQR [45] meps20 89.10% 26.53
CHR [52] meps20 89.35% 16.56
CDP-ACC meps20 90.05% 16.00
SCP [26, 44, 27] meps21 90.55% 26.58
QR [50] meps21 88.80% 29.91
CQR [45] meps21 90.05% 29.99
CHR [52] meps21 90.05% 19.44
CDP-ACC meps21 91.15% 15.00
α=0.1{}^{*}\alpha=0.1

References

  • [1] W. H. Organization et al., “Depression and other common mental disorders: global health estimates,” World Health Organization, Tech. Rep., 2017.
  • [2] A. Aprilia and D. Aminatun, “Investigating memory loss: How depression affects students’memory endurance,” Journal of English Language Teaching and Learning, vol. 3, no. 1, pp. 1–11, 2022.
  • [3] S. Marwaha, E. Palmer, T. Suppes, E. Cons, A. H. Young, and R. Upthegrove, “Novel and emerging treatments for major depression,” The Lancet, vol. 401, no. 10371, pp. 141–153, 2023.
  • [4] S. Bachmann, “Epidemiology of suicide and the psychiatric perspective,” International journal of environmental research and public health, vol. 15, no. 7, p. 1425, 2018.
  • [5] B. Singh, T. Olds, R. Curtis, D. Dumuid, R. Virgara, A. Watson, K. Szeto, E. O’Connor, T. Ferguson, E. Eglitis et al., “Effectiveness of physical activity interventions for improving depression, anxiety and distress: an overview of systematic reviews,” British journal of sports medicine, vol. 57, no. 18, pp. 1203–1209, 2023.
  • [6] A. T. Beck, R. A. Steer, and G. Brown, “Beck depression inventory–ii,” Psychological assessment, 1996.
  • [7] M. Hamilton, “The hamilton rating scale for depression,” in Assessment of depression.   Springer, 1986, pp. 143–152.
  • [8] K. Kroenke, T. W. Strine, R. L. Spitzer, J. B. Williams, J. T. Berry, and A. H. Mokdad, “The phq-8 as a measure of current depression in the general population,” Journal of affective disorders, vol. 114, no. 1-3, pp. 163–173, 2009.
  • [9] C. Shan, S. Gong, and P. W. McOwan, “Facial expression recognition based on local binary patterns: A comprehensive study,” Image and vision Computing, vol. 27, no. 6, pp. 803–816, 2009.
  • [10] T. R. Almaev and M. F. Valstar, “Local gabor binary patterns from three orthogonal planes for automatic facial expression recognition,” in 2013 Humaine association conference on affective computing and intelligent interaction.   IEEE, 2013, pp. 356–361.
  • [11] A. Dhall and R. Goecke, “A temporally piece-wise fisher vector approach for depression analysis,” in 2015 International conference on affective computing and intelligent interaction (ACII).   IEEE, 2015, pp. 255–259.
  • [12] L. Wen, X. Li, G. Guo, and Y. Zhu, “Automated depression diagnosis based on facial dynamic analysis and sparse coding,” IEEE Transactions on Information Forensics and Security, vol. 10, no. 7, pp. 1432–1441, 2015.
  • [13] J. R. Williamson, T. F. Quatieri, B. S. Helfer, G. Ciccarelli, and D. D. Mehta, “Vocal and facial biomarkers of depression based on motor incoordination and timing,” in Proceedings of the 4th international workshop on audio/visual emotion challenge, 2014, pp. 65–72.
  • [14] Z. Du, W. Li, D. Huang, and Y. Wang, “Encoding visual behaviors with attentive temporal convolution for depression prediction,” in 2019 14th IEEE international conference on automatic face & gesture recognition (FG 2019).   IEEE, 2019, pp. 1–7.
  • [15] L. He, M. Niu, P. Tiwari, P. Marttinen, R. Su, J. Jiang, C. Guo, H. Wang, S. Ding, Z. Wang et al., “Deep learning for depression recognition with audiovisual cues: A review,” Information Fusion, vol. 80, pp. 56–86, 2022.
  • [16] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, 2012.
  • [17] W. C. de Melo, E. Granger, and A. Hadid, “A deep multiscale spatiotemporal network for assessing depression from facial dynamics,” IEEE transactions on affective computing, vol. 13, no. 3, pp. 1581–1592, 2020.
  • [18] W. C. de Melo, E. Granger, and M. B. Lopez, “Mdn: A deep maximization-differentiation network for spatio-temporal depression detection,” IEEE transactions on affective computing, 2021.
  • [19] Y. Zhu, Y. Shang, Z. Shao, and G. Guo, “Automated depression diagnosis based on deep networks to encode facial appearance and dynamics,” IEEE Transactions on Affective Computing, vol. 9, no. 4, pp. 578–584, 2017.
  • [20] X. Zhou, Z. Wei, M. Xu, S. Qu, and G. Guo, “Facial depression recognition by deep joint label distribution and metric learning,” IEEE Transactions on Affective Computing, vol. 13, no. 3, pp. 1605–1618, 2022.
  • [21] L. He and C. Cao, “Automated depression analysis using convolutional neural networks from speech,” Journal of biomedical informatics, vol. 83, pp. 103–111, 2018.
  • [22] Z. Zhao, Q. Li, N. Cummins, B. Liu, H. Wang, J. Tao, and B. Schuller, “Hybrid network feature extraction for depression assessment from speech,” in Proceedings INTERSPEECH 2020.   ISCA-INST SPEECH COMMUNICATION ASSOC, 2020, pp. 4956–4960.
  • [23] D. Castelvecchi, “Can we open the black box of ai?” Nature News, vol. 538, no. 7623, p. 20, 2016.
  • [24] V. Buhrmester, D. Münch, and M. Arens, “Analysis of explainers of black box deep neural networks for computer vision: A survey,” Machine Learning and Knowledge Extraction, vol. 3, no. 4, pp. 966–989, 2021.
  • [25] M. Xu, X. Zhang, and X. Zhou, “Confidence-calibrated face and kinship verification,” IEEE Transactions on Information Forensics and Security, 2023.
  • [26] V. Vovk, A. Gammerman, and C. Saunders, “Machine-learning applications of algorithmic randomness,” in Proceedings of the Sixteenth International Conference on Machine Learning, 1999, pp. 444–453.
  • [27] G. Shafer and V. Vovk, “A tutorial on conformal prediction.” Journal of Machine Learning Research, vol. 9, no. 3, 2008.
  • [28] Y. Romano, M. Sesia, and E. Candes, “Classification with valid and adaptive coverage,” Advances in Neural Information Processing Systems, vol. 33, pp. 3581–3591, 2020.
  • [29] G. Sharma, A. Parashar, and A. M. Joshi, “Dephnn: a novel hybrid neural network for electroencephalogram (eeg)-based screening of depression,” Biomedical signal processing and control, vol. 66, p. 102393, 2021.
  • [30] A. Seal, R. Bajpai, J. Agnihotri, A. Yazidi, E. Herrera-Viedma, and O. Krejcar, “Deprnet: A deep convolution neural network framework for detecting depression using eeg,” IEEE Transactions on Instrumentation and Measurement, vol. 70, pp. 1–13, 2021.
  • [31] M. Niu, J. Tao, B. Liu, J. Huang, and Z. Lian, “Multimodal spatiotemporal representation for automatic depression level detection,” IEEE transactions on affective computing, 2020.
  • [32] M. A. Uddin, J. B. Joolee, and K.-A. Sohn, “Deep multi-modal network based automated depression severity estimation,” IEEE transactions on affective computing, 2022.
  • [33] M. Niu, Y. Li, J. Tao, X. Zhou, and B. W. Schuller, “Depressionmlp: A multi-layer perceptron architecture for automatic depression level prediction via facial keypoints and action units,” IEEE Transactions on Circuits and Systems for Video Technology, 2024.
  • [34] W. C. De Melo, E. Granger, and A. Hadid, “Depression detection based on deep distribution learning,” in 2019 IEEE international conference on image processing (ICIP).   IEEE, 2019, pp. 4544–4548.
  • [35] I. Nouretdinov, S. G. Costafreda, A. Gammerman, A. Chervonenkis, V. Vovk, V. Vapnik, and C. H. Fu, “Machine learning classification with confidence: application of transductive conformal predictors to mri-based diagnostic and prognostic markers in depression,” Neuroimage, vol. 56, no. 2, pp. 809–813, 2011.
  • [36] R. Harper and J. Southern, “A bayesian deep learning framework for end-to-end prediction of emotion from heartbeat,” IEEE transactions on affective computing, vol. 13, no. 2, pp. 985–991, 2020.
  • [37] N. R. Prabhu, N. Lehmann-Willenbrock, and T. Gerkmann, “End-to-end label uncertainty modeling in speech emotion recognition using bayesian neural networks and label distribution learning,” IEEE Transactions on Affective Computing, 2023.
  • [38] W. Wu, C. Zhang, and P. Woodland, “Estimating the uncertainty in emotion attributes using deep evidential regression,” in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023, pp. 15 681–15 695.
  • [39] L. Lo, B.-K. Ruan, H.-H. Shuai, and W.-H. Cheng, “Modeling uncertainty for low-resolution facial expression recognition,” IEEE transactions on affective computing, 2023.
  • [40] J. She, Y. Hu, H. Shi, J. Wang, Q. Shen, and T. Mei, “Dive into ambiguity: Latent distribution mining and pairwise uncertainty estimation for facial expression recognition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 6248–6257.
  • [41] N. Le, K. Nguyen, Q. Tran, E. Tjiputra, B. Le, and A. Nguyen, “Uncertainty-aware label distribution learning for facial expression recognition,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 6088–6097.
  • [42] M. K. Tellamekala, T. Giesbrecht, and M. Valstar, “Dimensional affect uncertainty modelling for apparent personality recognition,” IEEE Transactions on Affective Computing, vol. 13, no. 4, pp. 2144–2155, 2022.
  • [43] M. K. Tellamekala, S. Amiriparian, B. W. Schuller, E. André, T. Giesbrecht, and M. Valstar, “Cold fusion: Calibrated and ordinal latent distribution fusion for uncertainty-aware multimodal emotion recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  • [44] V. Vovk, “Conditional validity of inductive conformal predictors,” in Asian conference on machine learning.   PMLR, 2012, pp. 475–490.
  • [45] Y. Romano, E. Patterson, and E. Candes, “Conformalized quantile regression,” Advances in neural information processing systems, vol. 32, 2019.
  • [46] A. N. Angelopoulos and S. Bates, “A gentle introduction to conformal prediction and distribution-free uncertainty quantification,” arXiv preprint arXiv:2107.07511, 2021.
  • [47] P. Cui, W. Hu, and J. Zhu, “Calibrated reliable regression using maximum mean discrepancy,” Advances in Neural Information Processing Systems, vol. 33, pp. 17 164–17 175, 2020.
  • [48] M. Valstar, B. Schuller, K. Smith, F. Eyben, B. Jiang, S. Bilakhia, S. Schnieder, R. Cowie, and M. Pantic, “Avec 2013: the continuous audio/visual emotion and depression recognition challenge,” in Proceedings of the 3rd ACM international workshop on Audio/visual emotion challenge, 2013, pp. 3–10.
  • [49] M. Valstar, B. Schuller, K. Smith, T. Almaev, F. Eyben, J. Krajewski, R. Cowie, and M. Pantic, “Avec 2014: 3d dimensional affect and depression recognition challenge,” in Proceedings of the 4th international workshop on audio/visual emotion challenge, 2014, pp. 3–10.
  • [50] J. W. Taylor, “A quantile regression neural network approach to estimating the conditional density of multiperiod returns,” Journal of forecasting, vol. 19, no. 4, pp. 299–311, 2000.
  • [51] N. Tagasovska and D. Lopez-Paz, “Single-model uncertainties for deep learning,” Advances in neural information processing systems, vol. 32, 2019.
  • [52] M. Sesia and Y. Romano, “Conformal prediction using conditional histograms,” Advances in Neural Information Processing Systems, vol. 34, pp. 6304–6315, 2021.
  • [53] C. Lu, A. N. Angelopoulos, and S. Pomerantz, “Improving trustworthiness of ai disease severity rating in medical imaging with ordinal conformal prediction sets,” in International Conference on Medical Image Computing and Computer-Assisted Intervention.   Springer, 2022, pp. 545–554.
  • [54] Y. Chung, W. Neiswanger, I. Char, and J. Schneider, “Beyond pinball loss: Quantile methods for calibrated uncertainty quantification,” Advances in Neural Information Processing Systems, vol. 34, pp. 10 971–10 984, 2021.
  • [55] R. F. Barber, E. J. Candes, A. Ramdas, and R. J. Tibshirani, “Predictive inference with the jackknife+,” 2021.
  • [56] T. Baltrušaitis, P. Robinson, and L.-P. Morency, “Openface: an open source facial behavior analysis toolkit,” in 2016 IEEE winter conference on applications of computer vision (WACV).   IEEE, 2016, pp. 1–10.
  • [57] D. Tran, L. Bourdev, R. Fergus, L. Torresani, and M. Paluri, “Learning spatiotemporal features with 3d convolutional networks,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 4489–4497.
  • [58] C. Feichtenhofer, H. Fan, J. Malik, and K. He, “Slowfast networks for video recognition,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 6202–6211.
  • [59] A. N. Angelopoulos, S. Bates, M. Jordan, and J. Malik, “Uncertainty sets for image classifiers using conformal prediction,” in International Conference on Learning Representations.
  • [60] P. Rana, “Physicochemical Properties of Protein Tertiary Structure,” UCI Machine Learning Repository, 2013, DOI: https://doi.org/10.24432/C5QW3H.
  • [61] K. Buza, “BlogFeedback,” UCI Machine Learning Repository, 2014, DOI: https://doi.org/10.24432/C58S3F.
  • [62] K. Singh, “Facebook Comment Volume Dataset,” UCI Machine Learning Repository, 2016, DOI: https://doi.org/10.24432/C5Q886.
  • [63] A. Asuncion and D. Newman, “Uci machine learning repository,” 2007.
  • [64] J. W. Cohen, S. B. Cohen, and J. S. Banthin, “The medical expenditure panel survey: a national information resource to support healthcare cost research and inform policy and practice,” Medical care, vol. 47, no. 7_Supplement_1, pp. S44–S50, 2009.
[Uncaptioned image] Yonghong Li received a Bachelor of Engineering degree from Chang’an University, Shannxi, China, in 2020. He is currently pursuing his PhD degree at Beijing University of Posts and Telecommunications in Beijing, China, with a focus on affective computing and deep learning.
[Uncaptioned image] Xiuzhuang Zhou (Member, IEEE) received the PhD degree from the School of Computer Science, Beijing Institute of Technology, China, in 2011. He is currently a full professor at the Beijing University of Posts and Telecommunications, Beijing, China. His research interests include computer vision, pattern recognition, multimedia computing, and machine learning. He has authored more than 80 scientific papers in peer reviewed journals and conferences including several top venues such as the IEEE Transactions on Pattern Analysis and Machine Intelligence, CVPR. He serves as an associate editor for the Neurocomputing.