Random Survival Forest for Censored Functional Data
Abstract
This paper introduces a Random Survival Forest (RSF) method for functional data. The focus is specifically on defining a new functional data structure, the Censored Functional Data (CFD), for dealing with temporal observations that are censored due to study limitations or incomplete data collection. This approach allows for precise modelling of functional survival trajectories, leading to improved interpretation and prediction of survival dynamics across different groups. A medical survival study on the benchmark SOFA data set is presented. Results show good performance of the proposed approach, particularly in ranking the importance of predicting variables, as captured through dynamic changes in SOFA scores and patient mortality rates.
Keywords: Functional Data Analysis, Survival Analysis, Random Survival Forest, Functional Principal Component Analysis, Functional Random Survival Forest.
1 Introduction
Survival analysis focuses on estimating and predicting the time until an event of interest, such as death or another unique occurrence, using historical data Kleinbaum and Klein (2012). This type of analysis can be challenging because the exact timing of the event may be unknown in certain instances. Survival trees (STs), a specific type of model within this field, allow for uncovering intricate nonlinear relationships in an intuitive, interpretable format. By iteratively dividing the population into subgroups and predicting a unique survival distribution for each terminal node, STs provide a powerful tool for understanding and forecasting time-to-event data. STs and their ensembles are widespread in non-parametric modelling. STs are particularly noteworthy for their ability to handle complex relationships and interactions among variables Ishwaran et al. (2008). They offer a clear picture of the decision-making procedure, making it easier to interpret survival patterns and recognise critical prognostic factors. As a result, STs have established a unique position in survival analysis by delivering a visually understandable framework for modelling and predicting survival probabilities. Well-known limitations of single STs are variance and overfitting. To address these concerns, Bagging (SB) and Random Survival Forest (RSF) Wang et al. (2019) are frequently used. These techniques enhance reliability by averaging multiple trees, reducing variance, preventing overfitting, and improving prediction accuracy and robustness. An extensive review can be found in Wang and Li (2017). SB trains multiple STs on bootstrapped samples generated from the original dataset. Then, the single predictions of STs are combined through averaging or voting. RSF expands the concept of traditional Random Forest (RF) to survival analysis Hemant et al. (2008), Biemann and Kearney (2010). In RSF, each tree is trained on a random subset of observations and a random subset of features at each split. Introducing randomness in the tree-building process, RSF diminishes the correlation between STs and improves the diversity of the ensemble, resulting in better generalisation performance. Further research also focused on combining survival analysis with various modelling approaches, discussed in more detail in Wey et al. (2015).
Studying the relationship between time-varying processes and time-to-event data is challenging in clinical research. A modern alternative solution involves using Functional Data Analysis (FDA) Ramsay and Silverman (2005). FDA transforms longitudinal data into curves, which are then integrated into survival analysis models. A first approach proposed by Lin et al. (2021) incorporates multiple longitudinal outcomes by extracting features using two methods: multivariate functional principal component analysis and multivariate fast covariance estimation for sparse functional data. These extracted features are subsequently used as covariates in a survival model, with FPCA aiding in the estimation process. Spreafico et al. (2023) Spreafico et al. (2023) proposed a novel approach for addressing time-varying covariates in survival analysis using FDA. The latter research aims to transform longitudinal data into functional data, facilitating their treatment and analysis within the survival framework. For longitudinal studies, patients often return for follow-up visits at regular intervals. Therefore, the true event time is only known to lie within an interval between visits, and the exact time is obscured. Multiple studies have demonstrated biased survival outcome estimates using the Cox model’s well-developed methodology for modelling interval-censored data Zhang and Sun (2010). In many studies, the focus has been on reconstructing curves over the entire follow-up period of study Delaigle and Hall (2013)Strzalkowska-Kominiak and Romo (2021) to manage the data at the same observation time. Starting from these latter ideas, our research strives to extend the concept of RSF to the functional data framework, exploiting the FDA’s potential in dimensionality reduction, predictive power, interpretability, and ability to reconstruct censored data.
The paper positions itself within the literature on combining FDA and statistical learning. The starting point of our proposal is to extend the Functional Random Forest (FRF) offered by Maturo and Verde (2023), Maturo and Verde (2022) to the context of irregular data in survival analysis, with a particular focus on the issues of censured data reconstruction and explainability of the RSF.
Implementing RSF in the functional context poses significant challenges, especially given the vast amount of available data and the incomplete nature of each statistical unit’s temporal sequences. This work considers a new type of functional data, defined as Censored Functional Data (CFD). Unlike previous studies, we concentrate on the actual observation period to reconstruct the trajectories associated with each unit rather than extending the curves over the entire temporal domain of the follow-up study. This approach aims to maximise the utilisation of available information during the study period without excessive extensions or interpolations. Essentially, the main goal is to accurately model the trajectory of curves, considering only the observation period associated with each unit, thus ensuring a precise representation of the survival dynamic within that specific period.
By integrating the CFD with FPCA, we aim to capture relationships and dynamic patterns in the data, leading to more accurate and interpretable survival predictions within STs. Section 2 introduces preliminaries on Functional Data Analysis and Survival Random Forest and describes our contribution. Section 3 assesses the performance of our proposed approach on the well-known "SOFA" data set. The paper ends with a discussion and conclusion on our proposal.
2 Material and Methods
2.1 Preliminaries on Functional Data Analysis (FDA)
The field of Functional Data Analysis (FDA) Ramsay et al. (2009); Ramsay and Silverman (2005); Ferraty (2011) deals with data represented in functional form rather than using traditional discrete observations. Typically, data are handled as vectors or matrices, with an assumption of independence among observations. However, in many situations, the observations are displayed on a discrete time scale or spatial domain, and the information is not available on the entire study domain. Indeed, FDA aims to capture the underlying structure and the variability present in discrete data by transforming it into a functional data object that assumes values within a functional space denoted as . In general, the functional data object can be represented as a random variable , where is a common domain and denotes the value that the function assumes at time or position . In this article, we will consider to be a space-time domain. This real-valued function can be viewed as the realization of a one-dimensional stochastic process, often assumed to be in a Hilbert space, such as defined on the compact domain . Here, a stochastic process is said to be an process if and only if it satisfies . The space is defined as a normed space of functions where the norm is given by . In general, considering a time domain , in the Hilbert space is defined the distance between two functions and as follows:
(1) |
FDA aims to convert discrete data into a functional form using techniques such as smoothing, interpolation, and regression Schimek (2013). We observe the discrete-time data given by pairs , where and are the values recorded at the time for statistic units. The -th recorded value corresponds to a realization of a generic function at time . In FDA, we can define the function by finite linear combination Ramsay and Silverman (2005) as follows:
(2) |
where is a set of basis functions used in the representation, and are the coefficients of each basis function. However, in real cases, the -th observation can be affected by the -th term’s error, , due to measurement inaccuracies or other factors introducing variability. Thus, we observe as . In the Equation (2), the functional data is approximated by a finite linear combination of basis functions that can be chosen based on the specific problem and the characteristics of the data trend. The coefficients can be calculated using the sum of squared errors (SSE), where the problem is given by:
(3) |
where is the number of data points observed, represents the -th instant time, and is the number of basis functions chosen for the functional representation. In the context of FDA, considering the data functions expressed as , we can introduce the standard estimators for the mean and covariance functions:
(4) |
and
(5) |
Moreover, the FDA provides many useful tools for dealing with functional data to interpret and make predictions e.g., the Functional Principal Component Analysis (FPCA) Ramsay and Silverman (2005); Ferraty and Vieu (2006), Functional Regression Analysis (FRA) Ramsay and Silverman (2005); Febrero-Bande and de la Fuente (2012a) and Functional Classification Trees (FCTs) Maturo and Verde (2023, 2022, 2024). FPCA extends classical Principal Component Analysis (PCA) in a functional context. The method allows us to reduce the high dimensionality of the data, preserving the maximum amount of information Ramsay and Silverman (2005); Aguilera and Aguilera-Morillo (2013); Febrero-Bande and de la Fuente (2012b). In this context, the approximated functional data is given by:
(6) |
where represents the total number of Functional Principal Components Scores (FPCs), where denotes the eigenfunction for the -th of the function . The explained variance of these curves is determined by , where represents the variance of the -th FPC. It’s important to note that the variance explained by each FPC decreases as the index increases, that is . We assume functions, for , be centered, i.e., , which implies that the overall mean function is zero. The -th FPC can be obtained as follows:
(7) |
where the eigenfunctions are obtained resolving this problem:
(8) |
s.t.
(9) |
and
(10) |
In statistics and machine learning, the data points can be classified into distinct categories using the features collected during the study, which can vary across different fields and applications. In many situations, utilizing functional data objects as features can be advantageous. Starting from a classic Classification Tree (CT), we can move to the functional approach, which involves adapting traditional techniques to accommodate functional data, which are functions defined over a continuous domain. The method is called Functional Classification Tree (FCT) and allows us to deal with functions defined in a functional space . The methodology takes the pairs , where is a predictor curve for -th unit defined within metric space , and is a scalar response observed for -th unit. Thus, the FCT algorithm predicts the vector-response considering as features the FPCs’ scores obtained from decomposition (6) applied for each to obtain the score-matrix as follows:
(11) |
where and correspond to FPC score of the -th curve relative to the -th eigenfuncion for and Maturo and Verde (2022, 2023). This approach is called the Functional Classification Tree with Principal Components (FCT-FPCs). The FCT-FPCs organize data into groups based on their characteristics. Starting with all data points, which are the functions, the algorithm involves stratifying and segmenting the predictor space into rectangular regions, called terminal nodes or lives, based on certain criteria that make the groups more homogeneous by using as features the score-vector for . The method selects the best features and the best threshold to define at each internal node of the tree two half-planes given by:
(12) |
Then, to determine the best feature and threshold to split on, the algorithm uses the splitting criterion such as the Gini index or the Shannon-Weiner index Chao and Shen (2003); Chao et al. (2014). Let be the splitting criterion at node , which divides it into two daughter nodes left-daughter, right-daughter. Let , be the splitting criteria for the node and its two daughter nodes. The impurity is computed as follows:
(13) |
Where is the number of samples in the splitting-node, and are the number of samples in the left and right daughter nodes. The feature and threshold that maximize are chosen for splitting at each node.
2.2 Contribute
This section focuses on managing survival time data, starting with the creation of censored functional data and ending with the creation of the Functional Random Survival Forest (FRSF) based on FPCs. We explore the complexity of managing survival data, address challenges posed by censoring, and use techniques such as Principal Components Analysis through Conditional Expectation (PACE) to extract meaningful features from irregular functional data.
2.2.1 Censored Functional Data
In a survival study, subjects are enrolled in a given period called a follow-up. In this period, different data are collected for each -th subject, with . In general, the times denote the survival time, and denote the censored time, i.e., an event did not occur for the subject. Let be defined for each subject as follows:
(14) |
During the follow-up study, values are collected for each subject until the event time, denoted by , occurs. This information defines the Survival Time Data (STD) given by , where refers to the observed values vector with and is the time vector for the -th unit.
In this context, the main challenge of employing the FDA approach and leveraging functional tools is to retrieve data for every subject, even when there are no observations during the entire observation period. For this reason, our approach involves the introduction of CFD, enabling the continuous reconstruction of information for each subject at any given moment. The main issue of this extension is how to replace these data by using different functions according to the number of recording values . Note that for each -th unit and no longitudinal measurements are available after . Thus, the observation of the function consist of pairs for and .
Constructing functional data requires a different approach in survival analysis with censored data. Hence, we aim to develop a methodology that handles censoring and addresses the unique aspects of functional data, such as temporal truncation, to ensure accurate modelling and interpretation of the underlying patterns over time. In our case, we have as observed data, which are -dimensional. For this reason, the first step is to convert the observed values into a functional form by considering different recording values. In this connection, the standard approach to estimating the functional form by starting with observed values is the basis approximation. The CFD can be defined as follows:
Definition 1
Let functional space where is a compact interval. The CFD is a functional data expressed as follows:
(15) |
where is the true event time and is the censored time.
As in (15), we can express these functions according to numbers of . For which, using a fixed basis representation and other functional representations, we obtain the CFD as follows:
Definition 2
Let functional space defined on the compact interval and pairs observed values , where for , the CFD is built as follows:
(16) |
where corresponds to the -th coefficient for -th unit and to -th basis function that can be chosen to approximate the full basis expansion.
Moreover, is determined individually for -th unit when the number of recorded values exceeds or equals . It is determined using leave-one-out cross-validation. The method minimizes the prediction error of the number of components choosing the best for the -th subject as follows:
(17) |
In addition, considering that in real-world applications the measurement of may be subject to errors, we consider the random noise with and , where are independent across and and the errors are considered homoschedastic with . Then, we can express the observed values as , with , max, where is the length of the study follow-up. Hence, for different numbers using the definition (16) we obtain the follow model:
(18) |
In order to evaluate the coefficients for the (18), we use the SSD criterion as follows:
(19) |
calculated for all subjects .
2.2.2 Principal component analysis for Censored Functional Data
FPCA is an extension of the classical PCA because the principle is to replace vectors with functions, matrices by linear operators, and, in particular, covariance by auto-covariance operators Ramsay and Silverman (2005). Moreover, scalar products in vector are replaced by scalar products in function space . Let’s suppose that is a generic trajectory defined on Hilbert space with its mean function and its covariance function , with . The covariance function can be expressed by the spectral decomposition as , where , correspond to a set non-increasing eigenvalues such that and correspond to the eigenfunctions. Thus, the trajectory admits the following Karhunen-Loevè expansion:
(20) |
where the coefficient corresponds to the -th Functional Principal Component (FPC) score of and satisfies the condition with if and otherwise.
For the case of sparse or irregular functional data, a fully non-parametric approach denoted as PACE Yao et al. (2005) is used. Let be the -hth observation obtained from CFD (16) evaluating the -th curve at time , where is the step size chosen for the time increment by considering for . Then, including additional measurement errors , uncorrelated among them with and , to a model with measurements calculated at time from (20) we obtain:
(21) |
where the expansion (21) can be approximated as follows:
(22) |
Moreover, the mean function, the covariance function and the eigenfunctions are calculated using the local linear smoothers Fan and Gijbels (1992). From (21), we note that the for which , where is a mean function obtained by applying a local linear smoother to the scatterplot . The estimated covariance surface, , is calculated considering the raw estimates and then applying two-dimensional smoothing to the scatterplot . Nevertheless, the calculation of scores is done according to the definition of integrals in , but in this case having that are available only at discrete random times we can obtain the estimates from (21) as . The goal is to obtain predicted trajectories for irregular data . In summary, by discretising continuous trajectories with step sizes , we can effectively analyse functional data even when observations are sparse or irregular. This approach allows for the practical implementation of FPCA, capturing essential features of the data while accounting for measurement errors and irregular observation patterns.
An alternative to the Riemann sum given by the previous formula is to assume that in (21), the and are jointly Gaussian. For this reason, defining , and under Gaussian assumptions, we obtain the best prediction for the -th FPC score for the -th subject as conditional expectation as follows:
(23) |
where , with and identity matrix. Thus, , where if and otherwise. Equation (23) can be obtained by using the estimates of , , and as follows:
(24) |
Finally, from (24) the prediction for the trajectory for th subject can be done considering the first eigenfunctions as follows:
(25) |
2.2.3 Functional Survival Tree (FST)
In supervised learning, Survival Trees (STs) extend the classical Classification Trees (CTs) grown on censored data. Unlike classical decision trees, STs are not directly applicable to censored data; a ST focuses on predicting the time until an event of interest occurs for a subject. These problems introduce a distinctive challenge when dealing with survival data, making certain aspects of implementing ST more intricate than decision trees for classification tasks. In survival studies, we have, for each subject, an observation vector that is a data collected in values until the event time occurs for the -th unit and the response vector that can have binary values given by or for . Thus, we observe, for the -th subject, the data given by , with censoring indicator.
In traditional CTs, the non-pruned trees lead to identifying homogeneous terminal nodes (leaves) using some splitting criteria based on decreasing an impurity metric, e.g. the Gini or Shannon-Weier indexes. However, in STs, the splitting criterion used in each node of the tree, denoted by , differs from classical methods when dealing with censored data. Indeed, iteratively, the procedure grows a tree until each node defines, at the very least, a singular distinct event by using survival splitting as a means for maximizing between-node survival differences given by curves defined through parametric (Cox regression) and non-parametric (Kaplan-Meier) approaches Shimokawa et al. (2015).
In a classical decision tree, the features vectors have the same size for each unit: on the contrary, in STs the challenge is how to grow a tree having, for each subject, data collected during follow-up study given by for values with for some with . For this reason, starting from these data types and defining the CFD, we deal with new features using the values estimated on CFD and applying the PACE decomposition on them. Therefore, extending Functional Classification Trees Maturo and Verde (2022) to the context of survival analysis, we propose the Functional Survival Classification Tree with Principal Components (FSCTs-FPCs). The latter classifier considers the new data , where corresponds to the -th score vector obtained from Equation 24. Hence, the features matrix is defined as follows:
(26) |
where and is the -th FPC score of the -th curve relative to the eigenfunction for and .
Typically, considering the node of the survival tree , given a score-predictor , the method finds a value and selects the best such that survival differences between the two inequalities, and , are maximized in order to predict the response . The aim is to define in node two regions and to identify times for the two groups, with . One of the survival criteria used within node , to select the threshold and the functional score-prediction , is the log-rank test as splitting method Ziegler et al. (2007). It can be used to maximize a splitting value for a predictor variable as follows:
(27) |
with number of events in daughter nodes at time (time of observation), where is the total number of events within node . Moreover, is the number of individuals who are at risk (alive) in daughter nodes at time , where is the total number of individuals alive within node . is the total number of times between two daughter nodes. The number corresponds to for which , while corresponds to for which within node . The Equation (27) allows finding the best and the best which for each and chosen randomly. This process is repeated randomly at every node until the terminal node is reached. The FSCTs-FPCs use different methods for constructing predictive models. Two common alternative approaches include log-rank splitting, which divides data based on survival differences, and the Nelson-Aalen estimator, used for estimating cumulative hazard. The last method helps us to identify significant predictors of survival outcomes in the terminal nodes. Once the survival tree is built, let be the terminal node, where , in which are defined the times and let and the number of events and individuals at risk at time , the survival predictor in is defined in terms of the predictor within each terminal node . For this, the Cumulative Hazard Function (CHF) and the Survival Function (SF) are obtained by using the Nelson-Aalen estimator and Kaplan-Meier estimator defined as follows:
(28) |
where and are the hazard and the survival estimate for a terminal node obtained considering all distinct event times . However, predicting survival outcomes after the survival tree construction is essential. This evaluation often involves comparing the predicted survival probabilities with the observed outcomes. One standard method for evaluating the importance of a survival tree is using the Concordance Index (C-index), also known as Harrell’s C or the C-statistics, which tells us how accurately the model ranks people’s survival times. A higher C-index indicates better predictive performance of the survival tree model. Other metrics, such as the Brier score, can be used for further evaluation.
2.2.4 Functional Random Survival Forest (FRSF)
The Functional Random Survival Forest (FRSF) extends the classic RSF. In our approach, the method is built for the CFD considering the pairs , where is a CFD curve for -th unit defined in the metric space , while is a scalar response. Starting from the dataset used for the realization of the model, where is the -th score calculated by PACE on CFD, is a binary response, , and is a censoring indicator at sample . The method generates bootstrap samples from by creating the Training data set (In-bag data) and Validation data set (Out-of-bag data) to construct multiple STs.
The following pseudo-code can summarise the FRSF:
The step of the algorithm can be explained from (28) as follows:
(29) |
where is the in-bag CHF, while is the out-of-bag CHF for -th survival tree.
The metrics derived from the ensemble CHF, such as the predictor error and the variable importance (VIMP), provide crucial insights into the model’s goodness-of-fit calculated using only the OOB data. This process evaluates the accuracy of predictors and determines their relative importance in predicting survival outcomes. In FRSF the Breiman-Cutler Variable Importance (VIMP), also known as permutation importance Breiman (2002) is used to quantify the importance of each predictor variable by evaluating its impact on prediction error. The VIMP method permutes the out-of-bag (OOB) values of within each tree. This permutation process entails randomising the values of , while keeping the values of the other predictor variables unchanged. Moreover, the modified data influences decision paths through the tree structure based on the altered values. The resulting out-of-bag error, indicative of prediction accuracy with the permuted is compared to the original error calculated with unaltered values. The discrepancy between these errors quantifies the importance of in predicting outcomes. Aggregating these individual importance scores across all trees in the forest yields the permutation importance for variable , providing insights into its contribution to the overall predictive performance of the model.
3 Application
We consider a dataset from the field of critical care medicine called the Sequential Organ Failure Assessment (SOFA) score, designed to provide a comprehensive evaluation of organ function in critically ill patients. The SOFA score dataset is available in the R package ’refund’. Many applications have been developed utilizing the SOFA score dataset, as evidenced by several studies Moreno et al. (2023). This dataset concerns specifically patients admitted to the Intensive Care Unit (ICU) with Acute Lung Injury. For each patient, daily measurements of SOFA scores, which range from to , have been recorded for each patient over ICU days as shown in Figure (1), indicating the severity of organ failure.

Other variables collected during hospitalization include ICU death indicators, ICU length of stay (LOS), patient age, gender, and the Charlson co-morbidity index, providing insights into baseline health status. Figure (2) shows descriptive statistics for the variables. Boxplots visually summarize the distributional characteristics of the numerical variables, while barplots compare categorical data. All data concerning the outcome variable indicate the survival status, corresponding to patients died and lived until the last recording time.

The first problem we address is converting the SOFA scores into a functional form to monitor recovery progress throughout hospitalization. We thus propose a functional approach based on the previously described definition in Equation 2.2.1, where the response variable corresponds to the SOFA scores, and the independent variable represents the at which the score was calculated for different recorded values. For each statistical unit, truncated curves will be defined based on the scores recorded during hospitalization as follows:
(30) | |||||||
(31) | |||||||
with | (32) |
where in the equation (32) we select the set of functions as B-splines of order . The truncated curves have been visualized respect to the corresponding outcome variable, as shown in the Figure (3). To assess the model quality of the FRSF, we propose to use different training datasets, comprising , , , and of the SOFA dataset.

Train Dataset | STD Model | CFD Model(h=0.5) | CFD Model(h=0.2) | |||
(OOB) CRPS | (OOB) RPE | (OOB) CRPS | (OOB) RPE | (OOB) CRPS | (OOB) RPE | |
50% | 0.11830311 | 0.14836430 | 0.11861220 | 0.14734700 | 0.11130006 | 0.13733469 |
60% | 0.11631665 | 0.14853905 | 0.11517514 | 0.14402103 | 0.11273948 | 0.14061401 |
70% | 0.09189498 | 0.14906593 | 0.09253855 | 0.14428571 | 0.08753912 | 0.14195055 |
80% | 0.10149814 | 0.14921755 | 0.10135369 | 0.14619833 | 0.09974069 | 0.14570930 |
The model is evaluated considering the four scenarios obtained from the previously defined partition. In particular, we compare our proposed method (FRSF for CFD) and the traditional FRSF for STD data. We assess the models’ performance across the different training dataset proportions using various metrics, such as the Continuous Ranked Probability Score (CRPS) and the Requested Performance Error (RPE) on the OOB data proportion. As seen in Table 1, the models show varying degrees of accuracy depending on the proportion of the training dataset.
A more detailed analysis of temporal evaluations with different temporal discretisation values, represented by the parameter , enables capturing information that may otherwise be lost during the follow-up period. By carefully selecting , we ensure a more detailed and accurate temporal assessment, filling in gaps where the information was not collected initially. Specifically, we have chosen two different bandwidth values, and , within the CFD model. Moreover, the model’s performance has been analyzed using the RPE scores based on the number of trees and the CRPS scores to observe the forecasted temporal trend. Graphs for the four different scenarios (4) illustrate the model’s predictive performance evolution.




It can be noted that there is a significant improvement in the RPE (left panel) and CRPS (right panel) in all cases.
The CRPS function shows better performance for the FRSF with CFD during intermediate daily periods when a substantial amount of data is available. For example, when examining the CRPS and RPE graphs for the Model CFD with , we observe a gradual decrease in both metrics over time, indicating improved model accuracy. Furthermore, the quality of the models has been shown through Variable Importance (VIMP) and Relative Importance (RI) for scenarios (50%, 60%, 70%, and 80% of the Training dataset) for the three models. In our specific analysis, we focus on the scenario using 80% of the training dataset (results for other scenarios are provided in the appendix). The variables considered include the first four FPCs scores, Age, Charlson Comorbidity Index, and Gender. By examining the variable importance values across different scenarios and models, we can determine each feature relative contribution to the predictive performance of the models (Table 2).
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4138 | 1.0000 |
PC2 | 0.1325 | 0.3201 |
PC4 | 0.0786 | 0.1900 |
PC3 | 0.0510 | 0.1231 |
Age | 0.0435 | 0.1052 |
Charlson | 0.0284 | 0.0686 |
Gender | -0.0004 | -0.0009 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4202 | 1.0000 |
PC2 | 0.1287 | 0.3062 |
PC4 | 0.0806 | 0.1919 |
PC3 | 0.0487 | 0.1159 |
Age | 0.0416 | 0.0989 |
Charlson | 0.0270 | 0.0644 |
Gender | -0.0008 | -0.0019 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4396 | 1.0000 |
PC2 | 0.1191 | 0.2709 |
PC4 | 0.0945 | 0.2149 |
PC3 | 0.0532 | 0.1211 |
Age | 0.0434 | 0.0987 |
Charlson | 0.0346 | 0.0786 |
Gender | -0.0004 | -0.0008 |
In general, PC1 consistently exhibits the highest importance across all cases and models, followed by PC2 and PC4. Age and Charlson Comorbidity Index also show moderate importance, while Gender has an insignificant impact on the model. These values underscore the varying degrees of influence that different variables have on the predictive accuracy of the models.
4 Discussion and Conclusions
This study presents a new approach to survival analysis by developing Functional Random Survival Forest (FRSF), specially designed for Censored Functional Data (CFD). This approach embeds Functional Data Analysis within the framework of Random Survival Forest, thereby utilizing the strength of FPCA in managing complex, high-dimensional, censored, and temporally correlated data.
This study’s critical contribution is introducing the FRSF model, a novel extension of traditional RSF incorporating functional data. This offers significant advantages compared to conventional methods, especially when the data are censored and/or high-dimensional. The implementation of CFD provides a robust framework for handling incomplete or irregularly spaced data by enabling the reconstruction of continuous trajectories from discrete observations and performing a more comprehensive analysis of survival dynamics. Following the work of Maturo and Verde on FRF Maturo and Verde (2023), the current study adapts and extends FRF to the context of survival analysis and mainly deals with the difficulties posed by censored data. Fusing FPCA into the FRSF framework allows for dimension reduction while retaining key features, which gives the model interpretability to the FST and explainability for the overall RSF model. That means the model remains flexible in accounting for differing observation times and censoring mechanisms, leading to robust inferences without some common assumptions made for the parametric survival models.
A crucial aspect of this study is the introduction of the parameter , which denotes the step size for time increments in the evaluation of functional data. The parameter is meaningful as it permits discretizing continuous trajectories, enabling effective functional data analysis even when observations are sparse or irregular. Different values of can capture various levels of detail in the data, with smaller values providing finer resolution and potentially revealing more fine patterns in the temporal evolution of the data. Our analysis used different values of to demonstrate how different discretization levels impact the model’s performance. The proposed FRSF model was empirically validated using the SOFA dataset, which consists of daily measurements of critically ill patients.
Despite its strengths, the FRSF model has certain limitations. One of the primary challenges is the computational complexity associated with the FPCA decomposition and the subsequent construction of survival trees. The choice of basis functions and FPCs can also impact the model’s performance. Future research should optimise these aspects to enhance the model’s efficiency and scalability and test different basis functions, such as wavelets, to capture more complex and localised patterns in the data. In addition, future research could explore integrating FRSF with other machine learning approaches, such as boosting or deep learning, which could further enhance its predictive capabilities and scalability.
In conclusion, the FRSF offers a powerful and versatile tool for survival analysis, particularly in the presence of censored and irregularly spaced functional data. Its integration of FDA and RSF techniques, along with empirical validation, demonstrates its potential for significant contributions to theoretical advancements and practical applications in survival analysis.
Funding and/or Conflicts of interests/Competing interests
All the authors declare that they did not receive support from any organisation for the submitted work. All authors certify that they have no affiliations with or involvement in any organization or entity with any financial or non-financial interest in the subject matter or materials discussed in this manuscript.
References
- Aguilera and Aguilera-Morillo [2013] A. Aguilera and M. Aguilera-Morillo. Penalized PCA approaches for b-spline expansions of smooth functional data. Applied Mathematics and Computation, 219(14):7805–7819, mar 2013. doi: 10.1016/j.amc.2013.02.009. URL https://doi.org/10.1016%2Fj.amc.2013.02.009.
- Biemann and Kearney [2010] T. Biemann and E. Kearney. Size does matter: How varying group sizes in a sample affect the most common measures of group diversity. Organizational Research Methods, 13(3):582–599, jul 2010. doi: 10.1177/1094428109338875. URL https://doi.org/10.1177%2F1094428109338875.
- Breiman [2002] L. Breiman. Manual on setting up, using, and understanding random forests. (1), 2002.
- Chao and Shen [2003] A. Chao and T.-J. Shen. Nonparametric estimation of shannon’s index of diversity when there are unseen species in sample. Environmental and Ecological Statistics, 10(4):429–443, 2003. doi: 10.1023/B:ENVR.0000043140.70017.ac.
- Chao et al. [2014] A. Chao, N. J. Gotelli, T. C. Hsieh, E. L. Sander, K. H. Ma, R. K. Colwell, and A. M. Ellison. Rarefaction and extrapolation with hill numbers: a framework for sampling and estimation in species diversity studies. Ecological Monographs, 84(1):45–67, feb 2014. doi: 10.1890/13-0133.1. URL https://doi.org/10.1890%2F13-0133.1.
- Delaigle and Hall [2013] A. Delaigle and P. Hall. Classification using censored functional data. Journal of the American Statistical Association, 108(504):1269–1283, 2013. doi: 10.1080/01621459.2013.824893. URL https://doi.org/10.1080/01621459.2013.824893.
- Fan and Gijbels [1992] J. Fan and I. Gijbels. Variable bandwidth and local linear regression smoothers. Annals of Statistics, 20(4):2008–2036, December 1992. doi: 10.1214/aos/1176348900.
- Febrero-Bande and de la Fuente [2012a] M. Febrero-Bande and M. de la Fuente. Statistical computing in functional data analysis: The r package fda.usc. Journal of Statistical Software, Articles, 51(4):1–28, 2012a. doi: 10.18637/jss.v051.i04. URL https://www.jstatsoft.org/v051/i04.
- Febrero-Bande and de la Fuente [2012b] M. Febrero-Bande and M. O. de la Fuente. Statistical computing in functional data analysis: The R package fda.usc. Journal of Statistical Software, 2012b. ISSN 15487660. doi: 10.18637/jss.v051.i04.
- Ferraty [2011] F. Ferraty. Recent Advances in Functional Data Analysis and Related Topics. Physica-Verlag HD, 2011. doi: 10.1007/978-3-7908-2736-1. URL https://doi.org/10.1007%2F978-3-7908-2736-1.
- Ferraty and Vieu [2006] F. Ferraty and P. Vieu. Nonparametric Functional Data Analysis. Springer New York, 2006. doi: 10.1007/0-387-36620-2. URL https://doi.org/10.1007%2F0-387-36620-2.
- Hemant et al. [2008] I. Hemant, U. Kogalur, E. Blackstone, and M. Lauer. Random survival forests. The Annals of Applied Statistics, 2:841–860, 2008.
- Ishwaran et al. [2008] H. Ishwaran, U. B. Kogalur, E. H. Blackstone, and M. S. Lauer. Random survival forests. The Annals of Applied Statistics, 2:841–860, 2008.
- Kleinbaum and Klein [2012] D. G. Kleinbaum and M. Klein. Survival Analysis: A Self-learning Text. Springer, third edition, 2012. ISBN 978-1441966452.
- Lin et al. [2021] J. Lin, K. Li, and S. Luo. Functional survival forests for multivariate longitudinal outcomes: Dynamic prediction of Alzheimer’s disease progression. Statistical Methods in Medical Research, 30(1):99–111, 2021. doi: 10.1177/0962280220941532. URL https://doi.org/10.1177/0962280220941532.
- Maturo and Verde [2022] F. Maturo and R. Verde. Pooling random forest and functional data analysis for biomedical signals supervised classification: Theory and application to electrocardiogram data. Statistics in Medicine, 41(12):2247–2275, 2022. doi: https://doi.org/10.1002/sim.9353. URL https://onlinelibrary.wiley.com/doi/abs/10.1002/sim.9353.
- Maturo and Verde [2023] F. Maturo and R. Verde. Supervised classification of curves via a combined use of functional data analysis and tree-based methods. Computational Statistics, 38:419–459, 2023. doi: 10.1007/s00180-022-01236-1. URL https://doi.org/10.1007/s00180-022-01236-1.
- Maturo and Verde [2024] F. Maturo and R. Verde. Combining unsupervised and supervised learning techniques for enhancing the performance of functional data classifiers. Computational Statistics, 39(1):239–270, 2024. doi: 10.1007/s00180-022-01259-8.
- Moreno et al. [2023] R. Moreno, A. Rhodes, L. Piquilloud, et al. The sequential organ failure assessment (sofa) score: has the time come for an update? Critical Care, 27(1):15, 2023.
- Ramsay and Silverman [2005] J. Ramsay and B. Silverman. Functional Data Analysis, 2nd edn. Springer, New York, 2005.
- Ramsay et al. [2009] J. Ramsay, G. Hooker, and S. Graves. Introduction to functional data analysis. In Functional Data Analysis with R and MATLAB, pages 1–19. Springer New York, 2009. doi: 10.1007/978-0-387-98185-7\_1. URL https://doi.org/10.1007%2F978-0-387-98185-7_1.
- Schimek [2013] M. G. Schimek, editor. Smoothing and Regression: Approaches, Computation, and Application. John Wiley & Sons, 2013.
- Shimokawa et al. [2015] A. Shimokawa, Y. Kawasaki, and E. Miyaoka. Comparison of splitting methods on survival tree. The International Journal of Biostatistics, 11:175 – 188, 2015. URL https://api.semanticscholar.org/CorpusID:11441090.
- Spreafico et al. [2023] M. Spreafico, F. Ieva, and M. Fiocco. Modelling time-varying covariates effect on survival via functional data analysis: application to the MRC BO06 trial in osteosarcoma. Statistical Methods & Applications, 32:271–298, 2023. doi: 10.1007/s10260-022-00647-0. URL https://doi.org/10.1007/s10260-022-00647-0.
- Strzalkowska-Kominiak and Romo [2021] E. Strzalkowska-Kominiak and J. Romo. Censored functional data for incomplete follow-up studies. Statistics in Medicine, 40:2821–2838, 2021. doi: 10.1002/sim.8930. URL https://doi.org/10.1002/sim.8930.
- Wang and Li [2017] H. Wang and G. Li. A selective review on random survival forests for high dimensional data. Quantitative Biology, 36(2):85–96, 2017. doi: 10.22283/qbs.2017.36.2.85.
- Wang et al. [2019] P. Wang, Y. Li, and C. K. Reddy. Machine learning for survival analysis: A survey. ACM Computing Surveys (CSUR), 51(6):1–36, 2019.
- Wey et al. [2015] A. Wey, J. Connett, and K. Rudser. Combining parametric, semi-parametric, and non-parametric survival models with stacked survival models. Biostatistics, 16(3):537–549, 02 2015. ISSN 1465-4644. doi: 10.1093/biostatistics/kxv001. URL https://doi.org/10.1093/biostatistics/kxv001.
- Yao et al. [2005] F. Yao, H.-G. Müller, and J.-L. Wang. Functional linear regression analysis for longitudinal data. The Annals of Statistics, pages 287 3–2903, 2005.
- Zhang and Sun [2010] Z. Zhang and J. Sun. Interval censoring. Statistical Methods in Medical Research, 19(1):53–70, Feb 2010. doi: 10.1177/0962280209105023. URL https://doi.org/10.1177/0962280209105023.
- Ziegler et al. [2007] A. Ziegler, S. Lange, and R. Bender. Survival analysis: log rank test. Dtsch Med Wochenschr, 132(Suppl 1):e39–e41, 2007.
Appendix
VIMP Tables
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4466 | 1.0000 |
PC2 | 0.1209 | 0.2708 |
PC4 | 0.0610 | 0.1366 |
Age | 0.0526 | 0.1177 |
PC3 | 0.0512 | 0.1147 |
Charlson | 0.0398 | 0.0891 |
Gender | -0.0004 | -0.0008 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4549 | 1.0000 |
PC2 | 0.1233 | 0.2710 |
PC4 | 0.0701 | 0.1541 |
Age | 0.0497 | 0.1093 |
PC3 | 0.0459 | 0.1010 |
Charlson | 0.0388 | 0.0854 |
Gender | 0.0000 | 0.0000 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4749 | 1.0000 |
PC2 | 0.1144 | 0.2409 |
PC4 | 0.0800 | 0.1685 |
Age | 0.0475 | 0.0999 |
PC3 | 0.0464 | 0.0977 |
Charlson | 0.0381 | 0.0801 |
Gender | -0.0001 | -0.0002 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4473 | 1.0000 |
PC2 | 0.1078 | 0.2409 |
PC4 | 0.0768 | 0.1716 |
Age | 0.0656 | 0.1467 |
PC3 | 0.0467 | 0.1043 |
Charlson | 0.0424 | 0.0947 |
Gender | -0.0007 | -0.0015 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4571 | 1.0000 |
PC2 | 0.1117 | 0.2445 |
PC4 | 0.0824 | 0.1803 |
Age | 0.0585 | 0.1280 |
PC3 | 0.0399 | 0.0873 |
Charlson | 0.0389 | 0.0852 |
Gender | -0.0006 | -0.0012 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4674 | 1.0000 |
PC2 | 0.1094 | 0.2340 |
PC4 | 0.0916 | 0.1959 |
Age | 0.0536 | 0.1147 |
PC3 | 0.0423 | 0.0904 |
Charlson | 0.0414 | 0.0886 |
Gender | -0.0003 | -0.0006 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4338 | 1.0000 |
PC2 | 0.1169 | 0.2694 |
PC4 | 0.1067 | 0.2459 |
Age | 0.0612 | 0.1411 |
PC3 | 0.0567 | 0.1308 |
Charlson | 0.0362 | 0.0836 |
Gender | -0.0001 | -0.0003 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4457 | 1.0000 |
PC2 | 0.1104 | 0.2477 |
PC4 | 0.1062 | 0.2382 |
Age | 0.0557 | 0.1249 |
PC3 | 0.0515 | 0.1156 |
Charlson | 0.0432 | 0.0969 |
Gender | -0.0001 | -0.0001 |
Variable | Importance | Relative Importance |
---|---|---|
PC1 | 0.4465 | 1.0000 |
PC4 | 0.1045 | 0.2340 |
PC2 | 0.1024 | 0.2293 |
Age | 0.0524 | 0.1174 |
PC3 | 0.0517 | 0.1157 |
Charlson | 0.0415 | 0.0928 |
Gender | -0.0003 | -0.0008 |