Optimization of model independent gravitational wave search using machine learning
Abstract
The Coherent WaveBurst (cWB) search algorithm identifies generic gravitational wave (GW) signals in the LIGO-Virgo strain data. We propose a machine learning (ML) method to optimize the pipeline sensitivity to the special class of GW signals known as binary black hole (BBH) mergers. Here, we test the ML-enhanced cWB search on strain data from the first and second observing runs of Advanced LIGO and successfully recover all BBH events previously reported by cWB, with higher significance. For simulated events found with a false alarm rate less than , we demonstrate the improvement in the detection efficiency of for stellar-mass BBH mergers and for intermediate mass black hole binary mergers. To demonstrate the robustness of the ML-enhanced search for the detection of generic BBH signals, we show that it has the increased sensitivity to the spin precessing or eccentric BBH events, even when trained on simulated quasi-circular BBH events with aligned spins.
I Introduction
The detection of the first gravitational wave (GW) signal GW150914 [1] commenced the age of the GW astronomy. Since then, the Advanced LIGO [2] and Advanced Virgo [3] network of detectors have identified 11 GW candidates during the first two observing runs (O1 and O2) [4], 40 GW candidates in the first half of the third observing run (O3a) [5], and provided 20 GW public alerts to EM astronomers during the second half of the third observing run (O3b) [6]. With the improving sensitivity of the GW detector network, it is essential to refine the search algorithms used to detect GW signals.
Coherent WaveBurst (cWB) is an algorithm that searches for excess power in the time-frequency domain to identify GW signals in the LIGO-Virgo strain data [7, 8]. Unlike other analysis pipelines which search for binary black hole (BBH) mergers, cWB does not use template waveform models. Instead, the cWB algorithm is model independent, which makes it a valuable tool in the search for poorly modeled or unexpected GW sources. cWB played an integral role in the discovery of the first BBH merger GW150914 [1] and, more recently, in the first direct detection of an intermediate mass black hole (IMBH) GW190521 [9, 10]. Also, cWB has contributed to the detection of 22 BBH events in the O1, O2, and O3a observing runs [4, 5].
The cWB pipeline generates summary statistics for every identified event. These summary statistics describe generic properties of reconstructed events such as the characteristic frequency, duration, cross-correlation between different detectors, and other parameters described in Appendix A. In the standard cWB search framework, we use summary statistics to construct vetoes designed to reject noise events from the analysis and increase the significance of detected GW events. Although this veto procedure generally works well, it risks the removal of GW signal events that lie near the border of the predefined veto thresholds. In addition, designing vetoes is challenging since they need to be redefined for each detector network and are dependent on the run conditions.
Machine learning (ML) offers a novel approach to solving complex problems. Accordingly, the interest of ML techniques in GW astronomy has grown in recent years [11] as ML has been applied to categorize noise artifacts in the GW detector strain data [12, 13, 14], classify GW signals [15, 16], and estimate GW source parameters [17, 18]. ML has already been used in combination with cWB for various other studies [19, 20, 21, 22].
In this paper, we propose to use a decision tree based ensemble learning algorithm called eXtreme-Gradient Boost (XGBoost) [23] to automate the signal-noise classification in cWB and optimize the pipeline sensitivity to BBH mergers. To preserve the waveform independent analysis, we do not attempt to train the ML model directly on the GW strain data. Instead, we utilize cWB to reconstruct events and generate their summary statistics, and then we carefully select a subset of summary statistics used for the construction of the ML model. The end result is the ML-enhanced search pipeline which is resistant to overfitting and provides the robust recovery of GW events with the waveform parameters that could be outside of the training set. We test the cWB pipeline enhanced with XGBoost on publicly available LIGO Hanford and LIGO Livingston strain data from O1 and O2 [24].
The paper is organized as follows. In Section II, we introduce the cWB search pipeline. In Section III, we describe the data used to train and test our ML algorithm. In Section IV, we demonstrate the implementation of ML into the detection procedure and define our new cWB reduced detection statistic. In Section V, we compare the sensitivity of the ML-enhanced cWB search against the sensitivity of the standard cWB search. We report the updated significance of BBH events detected during the O1 and O2 observing runs. Finally, in Section VI, we state the conclusions of our study.
II cWB search algorithm
The cWB search algorithm is designed to detect GW signals with minimal assumptions on the signal model [25, 8]. The detector strain data is mapped to the time-frequency domain using the Wilson Daubechies Meyer (WDM) wavelet transform [26] where the data is normalized by the rms of the detector noise. The algorithm then identifies WDM wavelets with excess power above the average fluctuations of the detector noise. The selected nearby wavelets are grouped into clusters. The pipeline generates an event for each selected cluster and reconstructs the signal waveform using the constrained maximum likelihood method [8].
For each event, various summary statistics are estimated by the pipeline which describes the time-frequency structure, signal strength, and coherence across the detector network. The main detection statistic for the cWB generic GW search is the signal-to-noise ratio (SNR) defined for the LIGO detector network as:
(1) |
Here, denotes the coherent energy estimated by cross-correlating the reconstructed signal waveforms across different detectors, and where is the estimated residual noise energy and is the the number of independent wavelet amplitudes describing the event. The correction in Equation 1, which is close to unity for genuine GW events, reduces the non-Gaussian noise contribution. For the cWB searches which target BBH events, the detection statistic is modified to favor events which frequency is increasing with time:
(2) |
where is the event energy fraction and is the event ellipticity defined in Ref. [27]. Both parameters and are close to unity for BBH events and penalize events which time-frequency evolution is significantly different from the chirping BBH signal.
GW detector data is hindered by noise artifacts known as glitches, and consequently, some noise events are reconstructed by the pipeline and leak into the analysis. In the standard cWB analysis, we apply a series of vetoes to target and remove these glitches. This approach, henceforth known as the veto method, improves the significance of candidate GW events by reducing excess background. The veto method consists of applying a priori defined veto thresholds on a set of the cWB summary statistics. This procedure discretely classifies generated events into one of the two categories: signal-like events and noise-like events. Events that fall into the noise-like category are removed from the analysis. This process could inevitably result in discarding borderline GW events which do not pass the veto thresholds and at the same time makes the pipeline vulnerable to the high SNR glitches which do pass the vetoes. Designing vetoes in the multidimensional space of the summary statistics is challenging, and furthermore, requires re-tuning of the veto thresholds for each detector network configuration and each observing run.
In the standard cWB setup, the veto method is tuned separately to improve the search sensitivity to stellar-mass BBH mergers () and IMBH binary mergers (). While the GW waveforms of these two classes are conceptually similar, the corresponding GW signals observed in the LIGO frequency band are quite different. A GW signal originating from the stellar-mass BBH merger usually exhibits the full inspiral-merger-ringdown waveform, whereas GW signals from IMBH binary mergers are short in duration and contain mostly the merger-ringdown waveform, with the inspiral signal buried inside the low-frequency seismic noise. As a result, we utilize two configurations of the cWB search tuned for these systems: the BBH configuration which targets stellar-mass BBH mergers, and the IMBH configuration which targets IMBH binary mergers [10]. IMBH binaries are expected to merge at lower frequencies compared to the stellar-mass BBH mergers, and so the corresponding cWB search configurations apply different vetoes to account for the difference in the signal morphology.
III Data
We analyze publicly available strain data from Advanced LIGO’s first two observational runs [24]. Here, we only examine data from the LIGO Hanford and LIGO Livingston detectors, with the inclusion of Virgo data left for future work. To train and test our ML model, we require a representative set of noise events and signal events.
Noise events (background) are generated by systematically time-shifting the data from one detector with respect to other detectors in the detector network. Each time shift is chosen to be greater than the time of flight between detectors to exclude true astrophysical signals. This process is repeated multiple times for various time lags, and we count the number of background events generated over the total accumulated background time. For the O1 run, we accumulated approximately 16,000 years and 4,000 years of the background time for the BBH and IMBH searches, respectively. For the O2 run, we accumulated approximately 11,000 years of background for each search.
To generate a representative set of the signal events, we add (inject) simulated GW signals to the detector data and reconstruct them with cWB. In this work, we investigate four sets of simulated signals: (i) a quasi-circular spin-aligned stellar-mass BBH set, (ii) a quasi-circular IMBH binary set, (iii) an eccentric BBH set, and (iv) a quasi-circular precessing BBH set. Only the first two simulation sets of the quasi-circular signals were used for ML training, whereas the remaining two sets are used to test the robustness of the ML implementation. In all four cases, the binary orientation parameters (sky location, inclination angle) for every simulated waveform are randomly drawn from uniform distributions. The redshift is drawn from a uniform distribution in co-moving volume, assuming Planck 2015 cosmology [28].
To simulate the stellar-mass BBH set, we use the SEOBNRv3 [29] and SEOBNRv4 [30] waveform approximants. These waveform approximants include only the dominant () harmonic mode. The source frame total mass for these simulations ranges from approximately to , and the mass ratio ranges from approximately 1/4 to 1. Component black hole spins are aligned with the orbital plane.
For the IMBH binary set, we use numerical relativity waveforms which include higher-order harmonics. We consider 17 massbins, as used in Ref. [31], which range in source frame total mass from to , with mass ratios ranging from 1 to 1/10.
For the high mass, eccentric BBH set, we also use numerical relativity waveforms [32, 33]. We consider 28 massbins which range in total mass from to , mass ratio equal to 1, with eccentricities ranging from 0.66 to 0.99.
For the precessing stellar-mass BBH set, we use the SEOBNRv4PHM [34] waveform approximant, which includes precession and higher-order harmonic modes. The source frame total mass ranges from to , with mass ratios ranging from 1 to 1/20. Component black hole spins are isotropically distributed.
IV Machine Learning implementation
The veto method used in the standard cWB search categorizes events into two discrete bins: signal events and noise events. Here, the veto method effectively acts as a decision tree with only two leaves, where the summary statistics for a given event are compared against various rules at each decision node until the event is classified as either a signal event or a noise event. Since this method produces a discrete outcome, it unavoidably removes signal events which do not pass all veto thresholds.
We propose using ML, which produces continuous ranking criteria for all reconstructed events, to replace the veto method. Binary classification is a standard problem in the ML literature. Moreover, many prominent ML algorithms are based on the decision tree structure, which we expect is well suited for the cWB classification problem.
We use the boosted decision tree based ML algorithm called XGBoost [23]. In XGBoost, instead of using a single decision tree to classify events, an ensemble of decision trees is generated. A decision tree is used as the base learner, and subsequent learners (trees) are formed based on the residual errors obtained after each iteration (boosting). This process is expected to be more accurate and more robust than using a single decision tree used by the veto method. A continuous score is calculated by taking the weighted average of output, obtained from each decision tree in the ensemble. The final output is computed by taking the sigmoid of the continuous score, where a value close to zero denotes a noise-like event and a value close to one denotes a signal-like event.
To construct the ML model, we select a subset of 14 summary statistics estimated by cWB as input features for the ML algorithm. Selected summary statistics describe the signal strength and the correlation across the detectors (, , ), the quality of the likelihood fit (, ), the time-frequency evolution of the event (, , , , ) and the two different estimators for the number of cycles in the reconstructed waveform (, ). The detailed list of selected summary statistics and their definitions can be found in Appendix A.
IV.1 Tuning XGBoost hyper-parameters
XGBoost has a number of free hyper-parameters which control various properties of the learning process to prevent overfitting. These hyper-parameters need to be tuned for each specific application. We use a small data set consisting of yr of background data (which equates to approximately 500,000 noise events) and 2,000 simulated IMBH binary events to tune the XGBoost hyper-parameters. These background and simulation events are drawn from the O2 run.
We perform a grid search over a range of six standard XGBoost hyper-parameters, listed in Table 1. We find the most optimal set by evaluating each configuration of XGBoost hyper-parameters with respect to the precision-recall area under the curve (AUC PR) metric [35] over 10-fold cross-validation. The optimal configuration of hyper-parameters, according to this criteria, is shown in bold in Table 1. We use a method known as early stopping to optimize the total number of trees generated. In early stopping, a small fraction of the training data set is set aside for validation. When the validation AUC PR score stops improving, the training ends to prevent the XGBoost from overfitting.
Overall, we found that the performance of the model was not highly dependent on the chosen hyper-parameter values. As a result, we keep this hyper-parameter configuration for all models presented in this paper.
XGBoost hyper-parameter | entry |
---|---|
objective | binary:logistic |
tree_method | hist |
grow_policy | lossguide |
n_estimators | 20,000 |
max_depth | 7, 9, 11, 13 |
learning_rate | 0.1, 0.03 |
min_child_weight | 5.0, 10.0 |
colsample_bytree | 0.6, 0.8, 1.0 |
subsample | 0.4, 0.6, 0.8 |
gamma | 2.0, 5.0, 10.0 |
IV.2 XGBoost model training
We train a separate ML model for each search configuration and each observing run. In this paper, we have two search configurations (BBH, IMBH) and analyze two observing runs (O1, O2), a total of four models. The estimated central frequency of a GW signal is expected to be inversely proportional to the red-shifted total mass of the binary system. As such, we select events with to train the BBH search models and events with to train the IMBH search models. Each trained model uses the same XGBoost hyper-parameters discussed in Section IV.1, and consists of around total trees for O1 run and total trees for O2 run, with an average of 20 leaves per tree.
For training, we select of background data (approximately 250,000 noise events) per data chunk. For every of background data we select approximately simulation events. We use simulated quasi-circular stellar-mass BBH mergers (simulation set i) to train the BBH search model, and we use simulated quasi-circular IMBH binary mergers (simulation set ii) for the IMBH search model. The remaining background and simulation data are used for testing. Generally, ML classifiers are expected to be more accurate when the same number of events per class are used for training. However, in our case, using balanced classes is not feasible. It is difficult to arbitrarily increase the number of simulated events due to the computational cost. Moreover, down-sampling the noise event set is not prudent since we could lose valuable information related to the high SNR tail of the background distribution. Instead, we apply a weight to every sample noise event to reduce the class imbalance. This weight is dependent on and gives less importance to the low SNR glitches. The weighting procedure is described in more detail in Appendix B.
IV.3 cWB+ML detection statistic
We incorporate the predictions made by the ML model directly into the detection statistic to improve noise rejection. We define the reduced detection statistic used for the ML-enhanced cWB search as:
(3) |
where is the XGBoost penalty factor. To compute , we first apply a correction to the XGBoost output, defined in Appendix C.1. This correction is designed to suppress numerous noise events which have less than one cycle in the time domain waveform, which is typical for the known family of glitches found in the GW detector data. Next, we apply a monotonic transformation, defined in Appendix C.2, to obtain the penalty factor . This transformation accentuates the ranking of events with the values of very close to unity.
Although itself could be used as a detection statistic, we find that it is susceptible to assigning high significance to low SNR noise events. Instead, we use it as a penalty factor to the estimated effective correlated SNR . The end result is a detection statistic which is enhanced by the ML classification and resistant to overfitting the low SNR noise events.
V Results
In this section, we present the results of the ML-enhanced cWB search and compare its sensitivity to the standard cWB search. The significance of a given candidate event is estimated by its false alarm rate (FAR). It is computed by counting the number of background events with the equal or higher value of the detection statistic than for the candidate event, divided by the total background time.


For the ML-enhanced cWB search, FAR is calculated by using the reduced detection statistic (Equation 3). Whereas for the standard cWB BBH search, we first remove vetoed events and then calculate FAR by using the detection statistic (Equation 2). We compare the performance of the different cWB searches by comparing their detection efficiency defined as the number of detected simulated events with FAR equal to or less than a given threshold, divided by the total number of recovered simulation events.
V.1 Re-analysis of O1 and O2 data
First, we examine the background noise distributions for the ML-enhanced search. Figure 1 shows the rate of the noise events for each search (BBH, IMBH) and each observing run (O1, O2) as a function of the reduced detection statistic . In all cases, we do not observe significant tails of the high events, which indicates that the ML-enhanced detection statistic efficiently suppresses the high SNR outliers.
Next, we examine the sensitivities of each search for various simulation data sets. The top panel of Figure 2 shows the detection efficiency as a function of FAR for the various searches over the O1 (left) and O2 (right) data. The ML-enhanced search (shown in red) is more sensitive than the standard search with vetoes (black) in the wide range of the FAR thresholds. For events detected with FAR , we estimate the 26% improvement for the BBH configuration and the 16% improvement for the IMBH configuration, averaged over the two observing runs. For high significance detection (FAR ), we estimate the 22% improvement for the BBH configuration and the 13% improvement for the IMBH configuration.
The bottom panel of Figure 2 shows the detection efficiency for events detected with FAR as a function of the central frequency . Here, we see that for most frequency bins, the ML-enhanced search (red) is more sensitive than the standard search (black). This indicates that the ML-enhanced search is not overly tuned to any specific frequency bins but shows consistent improvement over the entire frequency range considered by the cWB search configurations. Since the central frequency , is expected to be inversely proportional to the detector frame mass, we can infer that the ML-enhanced search is more sensitive over the entire BBH mass range accessible by LIGO.
Table 2 reports the BBH candidates identified by the ML-enhanced cWB search in the O1-O2 observing runs. We recover 7 BBH candidates previously reported by the standard cWB search [4], all identified with higher significance. Additionally, the ML-enhanced cWB search detected GW170809 with a FAR of , which was previously vetoed in the standard cWB search. No other candidate events were identified with FAR less than .
Event | Standard cWB | ML-enhanced cWB |
---|---|---|
( + vetoes) | () | |
FAR [yr-1] | FAR [yr-1] | |
GW150914 | ||
GW151226 | ||
GW170104 | ||
GW170608 | ||
GW170729 | ||
GW170809 | ||
GW170814 | ||
GW170823 |
V.2 Test of model robustness
As a final test, we analyze the performance of the ML-enhanced search on simulated waveforms outside of the training set. First, we investigate the sensitivity of the ML-enhanced IMBH search, which is trained on quasi-circular binaries, to high mass BBH systems with highly eccentric orbits (simulation set iii). In Figure 3, we show the sensitivity to the eccentric IMBH mergers of the ML-enhanced IMBH search compared to the standard IMBH search. The ML-enhanced search is more sensitive to the eccentric BBH mergers despite the fact that we trained the ML model only on the quasi-circular IMBH waveforms.
Next, we test the ML-enhanced BBH search on precessing BBH systems (simulation set iv). The training set consists of simulated waveforms with the (2,2) harmonic mode, aligned spins, and low mass ratio (1 - 1/4). The testing set consists of simulated waveforms with higher-order modes, precessing spins, and a higher mass ratio (1 - 1/20). In this case, the model is also trained on O2 data, whereas the testing set consists of O3a simulation and background, which has very different GW detector sensitivity. Figure 4 shows improved detection efficiency of the ML-enhanced BBH search compared to the standard BBH search.


These results demonstrate that the ML-enhanced search is agnostic to the details of the BBH dynamical evolution including eccentricity, spin effects, and higher-order modes. It retains the robustness of the standard search, but with increased sensitivity.
VI Conclusion
In conclusion, we introduce a novel method to automate the noise rejection in cWB with ML, and we demonstrate that the ML-enhanced cWB search has improved sensitivity to the BBH mergers. Unlike the standard cWB search, which decimates noise events with the veto method, the ML-enhanced search uses a continuous ranking statistic and does not remove any events. Instead, it penalizes noise events with the newly designed detection statistic , which combines the correlated network SNR with the penalty factor based on the signal-noise classification provided by the XGBoost model. This new detection statistic is resistant to loud noise glitches and allows cWB identification of BBH events with higher significance.
For the stellar-mass BBH mergers, the detection efficiency of the ML-enhanced cWB search is improved by 26% compared to the standard cWB search (at FAR less than ). Similarly, for the simulation set of IMBH binary mergers, the detection efficiency is improved by 16%. While we do not claim any new BBH detections in the O1 and O2 data, we do improve the detection confidence of the previously identified GW candidates and recovered the GW170809 event missed by the standard cWB search.
The ML-enhanced cWB search is capable of detecting BBH signals well outside of the training set. We demonstrate the improved pipeline sensitivity to the highly eccentric BBH mergers despite only being trained on the quasi-circular IMBH signals. We also found the search to be agnostic to other binary waveform properties including precession, high mass ratio, and higher-order harmonic modes. Waveforms with similar time-frequency morphology, but not predicted by general relativity should also be detectable.
The ML-enhanced BBH search is a promising addition to the cWB pipeline for future planned observing runs, where we expect numerous BBH detections. While in this study, we use only the LIGO Hanford and LIGO Livingston detector network, in future work we will expand the ML-enhanced search to the other detector networks which include Virgo and Kagra detectors. This will further improve the cWB sensitivity to the BBH mergers. The ML-enhanced pipeline could be also used for the low latency cWB searches in the future observing runs.
Acknowledgements.
This research has made use of data, software, and/or web tools obtained from the Gravitational Wave Open Science Center, a service of LIGO Laboratory, the LIGO Scientific Collaboration, and the Virgo Collaboration. This work was supported by the NSF Grant No. PHY 1806165. We gratefully acknowledge the support of LIGO and Virgo for the provision of computational resources. I. B. acknowledges support by the NSF Grant No. PHY 1911796, the Alfred P. Sloan Foundation and by the University of Florida. We thank Ik Siong Heng, Erik Katsavounidis, Peter Shawhan, and Michele Zanolin for their continued participation and effort in reviewing cWB analyses. We acknowledge the use of open source Python packages including NumPy [36], Pandas [37], Matplotlib [38], and scikit-learn [39].Appendix A cWB estimated summary statistics
The ML algorithm is tuned and trained on the data from a selected subset of cWB summary statistics for each event. We start by truncating this subset by selecting the summary statistics that have a low correlation with each other. We also aggregate a few summary statistics together to prune the list of summary statistics that are used as input features for the ML algorithm. We select 14 cWB summary statistics in total that are used by the ML algorithm as the input list of features. The summary statistics are listed below:
-
•
— Main cWB detection statistic for the generic GW search. For the ML study, we cap the value at 8 (any event with higher is assigned a value of 8) so the algorithm is not affected by the high SNR events in the background distribution, which is a steep function of .
-
•
— Coherent energy divided by the sum of coherent energy and null energy, defined in Ref. [25].
-
•
— Effective number of time-frequency resolutions used for event detection and waveform reconstruction.
-
•
— Ratio of the coherent energy to the network likelihood.
-
•
— Energy weighted signal duration.
-
•
— Energy weighted signal bandwidth.
-
•
— Energy weighted signal central frequency.
-
•
— Chirp mass parameter estimated in the time-frequency domain, defined in Ref. [27].
-
•
— Chirp mass goodness of the fit metric, presented in Ref. [27].
- •
-
•
— An estimation of the effective number of cycles in a cWB event.
-
•
— for the loudest pixel, the ratio between the pixel energy and the total energy of an event [42].
-
•
— quality of the event reconstruction, where is the residual noise energy estimated and is the number of independent wavelet amplitudes describing an event.
-
•
— Data chunk number. LIGO-Virgo data is divided into time segments known as chunks, which typically contain a few days of strain data. Including the data chunk number allows the ML algorithm to respond to changes in detector sensitivity across separate observing runs and chunks.
Appendix B Noise event sample weight
In the initial testing phase with the trained XGBoost model, we found a tail of high SNR background events, most of which were consistent with blip glitches. This tail is caused by the suboptimal ML model due to the high class imbalance between the high SNR noise events and signal events. To correct for this tail, we applied an dependent sample weight to the noise events while training the XGBoost models. This weight provides us with a weighted background distribution as shown in Figure 5, which is similar for any observing run.

The sample weight for the simulation events is set to 1. For the noise events, we divide into 231 bins with values ranging from 5 to 8 and apply the sample weight for each bin as follows:
(4) |
where is the number of signal events and is the number of noise events, in a given bin. We found that the sample weight works reasonably well for the following choices of values: , , and . The number of simulation events for 8 were re-sampled to match the number of background events with the same range of . This application of sample weight enables the algorithm to successfully remove almost all the high SNR outliers while keeping the weighted events class imbalance () to around 20 for any given observing run. Without the application of the sample weight, the class imbalance ranges from 50 to 600 or even higher for the given training setup depending on the observing run and the search configuration.

Appendix C XGBoost penalty factor
C.1 cWB correction to in the IMBH configuration
Further investigation revealed that in the IMBH configuration, high SNR background outliers were found in a specific parameter space represented by the - plot in Figure 6. In the standard veto method, the application of the and vetoes would have removed all events below the predefined thresholds at the cost of losing a small fraction of simulated events. In the ML method, we apply a correction to the ranking criteria such that we suppress the background outliers in the affected parameter space below the magenta-colored curve in Figure 6.
The correction to the continuous ranking criteria output from XGBoost is applied only to the IMBH configuration (shorter signals with fewer cycles in the time domain), given by the following equations:
(5) |
This correction enhances detection of signal and suppresses glitches in the desired part of the - parameter space explicitly without any changes to the XGBoost training and testing procedure. The correction comes at the cost of losing a small fraction of IMBH signals with small values of .
C.2 Monotonic transformation
The monotonic transformation applied on is defined as follows,
(6) |
which counterbalances the steep sigmoid function used by XGBoost while producing . The output by XGBoost for any event has a precision of up to 5 decimal places. As a result, the effect of using directly as a penalty factor is not very significant. The transformation in Equation 6 allows us to zoom in on the high precision, for values close to unity. This helps us differentiate between high SNR background events and simulation events that typically end up with very high values of .
References
- Abbott et al. [2016] B. P. Abbott et al. (LIGO Scientific Collaboration and Virgo Collaboration), Phys. Rev. Lett. 116, 061102 (2016).
- Aasi et al. [2015] J. Aasi et al. (LIGO Scientific Collaboration), Class. Quant. Grav. 32, 074001 (2015), arXiv:1411.4547 [gr-qc] .
- Acernese et al. [2015] F. Acernese et al. (Virgo Collaboration), Class. Quant. Grav. 32, 024001 (2015), arXiv:1408.3978 [gr-qc] .
- Abbott et al. [2019a] B. P. Abbott et al. (LIGO Scientific Collaboration, Virgo Collaboration), Phys. Rev. X9, 031040 (2019a), arXiv:1811.12907 [astro-ph.HE] .
- Abbott et al. [2020a] R. Abbott et al. (LIGO Scientific, Virgo), (2020a), arXiv:2010.14527 [gr-qc] .
- Gra [2020] Gravitational-Wave Candidate Event Database, LIGO/Virgo Public Alerts (2020).
- Klimenko et al. [2008a] S. Klimenko et al., Class. Quant. Grav. 25, 114029 (2008a).
- Klimenko et al. [2016] S. Klimenko et al., Phys. Rev. D 93, 042004 (2016), arXiv:1511.05999 [gr-qc] .
- Abbott et al. [2020b] B. P. Abbott et al. (LIGO Scientific Collaboration and Virgo Collaboration), Phys. Rev. Lett. 125, 101102 (2020b).
- Szczepanczyk et al. [2020] M. Szczepanczyk, S. Klimenko, B. O’Brien, I. Bartos, V. Gayathri, G. Mitselmakher, G. Prodi, G. Vedovato, C. Lazzaro, E. Milotti, F. Salemi, M. Drago, and S. Tiwari, (2020), arXiv:2009.11336 [astro-ph.HE] .
- Cuoco et al. [2020] E. Cuoco et al., Machine Learning: Science and Technology 2, 011002 (2020).
- Colgan et al. [2020] R. E. Colgan, K. R. Corley, Y. Lau, I. Bartos, J. N. Wright, Z. Márka, and S. Márka, Phys. Rev. D 101, 102003 (2020).
- Zevin et al. [2017] M. Zevin, S. Coughlin, S. Bahaadini, E. Besler, N. Rohani, S. Allen, M. Cabero, K. Crowston, A. K. Katsaggelos, S. L. Larson, and et al., Classical and Quantum Gravity 34, 064003 (2017).
- Cuoco et al. [2018] E. Cuoco, M. Razzano, and A. Utina, in 2018 26th European Signal Processing Conference (EUSIPCO) (2018) pp. 2648–2652.
- Gebhard et al. [2019] T. D. Gebhard, N. Kilbertus, I. Harry, and B. Schölkopf, Phys. Rev. D 100, 063015 (2019).
- Iess et al. [2020] A. Iess, E. Cuoco, F. Morawski, and J. Powell, Machine Learning: Science and Technology 1, 025014 (2020).
- George and Huerta [2018] D. George and E. Huerta, Physics Letters B 778, 64 (2018).
- Schmidt et al. [2021] S. Schmidt, M. Breschi, R. Gamba, G. Pagano, P. Rettegno, G. Riemenschneider, S. Bernuzzi, A. Nagar, and W. Del Pozzo, Physical Review D 103, 10.1103/physrevd.103.043020 (2021).
- Vinciguerra et al. [2017] S. Vinciguerra, M. Drago, G. A. Prodi, S. Klimenko, C. Lazzaro, V. Necula, F. Salemi, V. Tiwari, M. C. Tringali, and G. Vedovato, Classical and Quantum Gravity 34, 094003 (2017).
- Cavaglià et al. [2020] M. Cavaglià, S. Gaudio, T. Hansen, K. Staats, M. Szczepańczyk, and M. Zanolin, Machine Learning: Science and Technology 1, 015005 (2020).
- Gayathri et al. [2020a] V. Gayathri, D. Lopez, P. R. S., I. S. Heng, A. Pai, and C. Messenger, Phys. Rev. D 102, 104023 (2020a).
- O’Brien et al. [2021] B. O’Brien, M. Szczepańczyk, V. Gayathri, I. Bartos, G. Vedovato, G. Prodi, G. Mitselmakher, and S. Klimenko, In preparation, LIGO-DCC-P2100128 (2021).
- Chen and Guestrin [2016] T. Chen and C. Guestrin, in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16 (Association for Computing Machinery, New York, NY, USA, 2016) p. 785–794.
- Abbott et al. [2021] R. Abbott et al., SoftwareX 13, 100658 (2021).
- Klimenko et al. [2008b] S. Klimenko, I. Yakushin, A. Mercer, and G. Mitselmakher, Classical and Quantum Gravity 25, 114029 (2008b).
- Necula et al. [2012] V. Necula, S. Klimenko, and G. Mitselmakher, Gravitational waves. Numerical relativity - data analysis. Proceedings, 9th Edoardo Amaldi Conference, Amaldi 9, and meeting, NRDA 2011, Cardiff, UK, July 10-15, 2011, J. Phys. Conf. Ser. 363, 012032 (2012).
- Tiwari et al. [2015] V. Tiwari, S. Klimenko, V. Necula, and G. Mitselmakher, Classical and Quantum Gravity 33, 01LT01 (2015).
- Ade et al. [2016] P. A. R. Ade et al., Astronomy & Astrophysics 594, A13 (2016).
- Babak et al. [2017] S. Babak, A. Taracchini, and A. Buonanno, Phys. Rev. D95, 024010 (2017), arXiv:1607.05661 [gr-qc] .
- Bohé et al. [2017] A. Bohé et al., Physical Review D 95, 10.1103/physrevd.95.044028 (2017).
- Abbott et al. [2019b] B. P. Abbott et al. (LIGO Scientific Collaboration, Virgo Collaboration), Phys. Rev. D100, 064064 (2019b), arXiv:1906.08000 [gr-qc] .
- Healy et al. [2017] J. Healy, C. O. Lousto, Y. Zlochower, and M. Campanelli, Classical and Quantum Gravity 34, 224001 (2017).
- Gayathri et al. [2020b] V. Gayathri, J. Healy, J. Lange, B. O’Brien, M. Szczepanczyk, I. Bartos, M. Campanelli, S. Klimenko, C. Lousto, and R. O’Shaughnessy, (2020b), arXiv:2009.05461 [astro-ph.HE] .
- Ossokine et al. [2020] S. Ossokine et al., Phys. Rev. D 102, 044055 (2020), arXiv:2004.09442 [gr-qc] .
- Davis and Goadrich [2006] J. Davis and M. Goadrich, in Proceedings of the 23rd international conference on Machine learning - ICML '06 (ACM Press, 2006).
- Harris et al. [2020] C. R. Harris et al., Nature 585, 357 (2020).
- pan [2020] pandas-dev/pandas: Pandas (2020).
- Hunter [2007] J. D. Hunter, Computing in Science & Engineering 9, 90 (2007).
- Pedregosa et al. [2011] F. Pedregosa et al., Journal of Machine Learning Research 12, 2825 (2011).
- Vedovato [2018a] G. Vedovato, The Qveto algorithm (2018a).
- McIver [2012] J. McIver, Classical and Quantum Gravity 29, 124010 (2012).
- Vedovato [2018b] G. Vedovato, The Lveto algorithm (2018b).