This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bayesian Insights into post-Glitch Dynamics: Model comparison and parameter constraint from decades long observation data of the Crab pulsar

Chun Huang Physics Department, Central China Normal University, Luoyu Road, 430030, Wuhan, China Physics Department and McDonnell Center for the Space Sciences, Washington University in St. Louis; MO, 63130, USA Xiao-ping Zheng Physics Department, Central China Normal University, Luoyu Road, 430030, Wuhan, China [email protected] (XPZ)
Abstract

The Crab Pulsar has exhibited numerous glitches accompanied by persistent shifts in its spin-down rate. The explanation of the observed persistent shifts remain a challenge. We perform a detailed Bayesian analysis to compare four data-fitting models, ranging from a simple linear model to more complex power-law and logarithmic models, using a dataset of observed glitches and persistent shifts. Our results show the large observed events are difficult to explain by the usually assumed linear model due to starquakes. A particularly notable finding is that the logarithmic model provides the best fit to the observation data but the two power-law models show a close tie to it. Detail differences of these models may be further clarified by the understanding of internal physics of neutron stars.

Stars: Neutron(1108) — Bayesian statistics (1900)

1 Introduction

The Crab Pulsar has been observed for over half a century, particularly through radio band observations( eg. Lyne et al. (1992); Espinoza et al. (2011a); Fuentes et al. (2017); Serafin Nadeau et al. (2024)), resulting in a substantial accumulation of data on its rotational evolution, with significant contributions from the Jordrell Bank Observatory Espinoza et al. (2011b); Lyne et al. (2015); Basu et al. (2022). Nearly 30 instances of a sudden increase in the pulsar’s spin frequency, known as glitches, have been recorded. These glitches are typically accompanied by a permanent shift in the spin-down rate of the pulsar(see Alpar et al. (1985a, 1996); Akbal et al. (2017); Wang & Zheng (2019); Ge et al. (2020)). Despite extensive observations, the underlying mechanisms driving these glitches remain unclear. One hypothesis suggests that glitches may result from starquakes Baym & Pines (1971); Pines et al. (1972); Alpar & Baykal (1994); Link et al. (1998); Larson & Link (2002); Keer & Jones (2015); Akbal & Alpar (2018); Lu et al. (2023); Sourie & Chamel (2020) on the neutron star, while another model attributes them to the superfluidity of the neutron star’s matte Anderson & Itoh (1975); Ruderman (1976); Alpar (1977); Anderson et al. (1981); Alpar et al. (1984, 1985b); Link et al. (1993); Link & Cutler (2002); Glampedakis & Andersson (2009); Link (2014); Wlazłowski et al. (2016), which may also account for the abrupt increase in spin frequency.

The Crab Pulsar is particularly active in exhibiting glitches, likely due to its relative youth. Detailed modeling of glitch data from this pulsar is therefore crucial for understanding the phenomenon of pulsar glitches Alpar et al. (1985b, 1996); Lyne et al. (2000); Crawford & Demiański (2003); Wang & Zheng (2019); Sourie & Chamel (2020). In this study, we focus primarily on the persistent post-glitch shifts in spin-down rate following glitches, with a particular emphasis on the relationship between these effects and glitch sizes. This relationship may provide a useful probe for distinguishing among various glitch models, as the recovery behavior following a glitch can be modeled in greater detail due to its extended timescale, thereby offering rich information about the glitch mechanisms.

To analyze the glitch data, we will employ modern Bayesian techniques, which are well-suited for extracting information from observational datasets and for conducting model comparisons Kass & Raftery (1995). Bayesian inference allows for the systematic comparison of different models on a consistent basis, and also facilitates the constraints of model parameters within various frameworks. The objective of this study is to bridge the gap between models and observational data through systematic statistical modeling of post-glitch activity in the Crab Pulsar.

This paper is organized as follows: In Section 2, we will discuss the datasets used and the Bayesian framework implemented, along with the modeling motivations for the different data-fitting models. Section 3 will present the posterior results of the Bayesian inference, comparing the Bayesian evidence across different models, and explore the inter-model numerical conversion relations, to emphasis the deep connection between each two models. In section 4, the discussion of recent observation from 2017 event will be presented

       Glitch size Δν\Delta\nu (μHz)(\mu\mathrm{Hz})       Persistent shift Δν˙p\Delta\dot{\nu}_{p} (1015s2)\left(10^{-15}\mathrm{~{}s}^{-2}\right)
      1.08(1)       -112(2)
      2.43(1)       -150(5)
      1.133(3)       -116(5)
      0.20(1)       -25(3)
      0.855(3)       -53(3)
      0.94(3)       -70(10)
      0.151(3)       -8(2)
      6.595(20)       250(20)
      0.65(1)       - 30(5)
      1.46(1)       -132(5)
Table 1: Summary for the data-set of observed glitch size and the persistent shift. See Table 3 in Lyne et al. (2015)

2 Method description

The Bayesian technique requires several ingredients: observation data-set, the theoretical models, prior setting of all the model parameters, likelihood set-up and nested sampling algorithm. With accumulative observations of several decades, the observation data of the Crab pulsar already very significant, In Lyne et al. (2015), people summarized the 45 years of glitch observation on the Crab pulsar, with detailed record of the glitch sizes and the persistent shifts of post-glitch spin-down rates. (See Table 3 of Lyne et al. (2015)). Our primary focus is the relation between persistent shift Δν˙p\Delta\dot{\nu}_{p} and glitch size Δν\Delta\nu. The data is listed in the table 1. Larger glitch size normally corresponds to a larger step, this general correlation, in Lyne et al. (2015) was fitted by a linear model. While the detail treatment of this correlation is crucial for distinguishing different models, and only the linear model will prohibit a very bad explanation of the observation data. To explore this correlation systematically by modern statistical approach motivates this study.

2.1 Fitting models

In this current study, four different models will be examined by the observation data, regardless of realistic physics. (a) The linear fitting model Δν˙p=k0Δν+b0\Delta\dot{\nu}_{p}=k_{0}\Delta\nu+b_{0} (denoted as M1), where k0k_{0} and b0b_{0} are free parameters that need to explored by Bayesian technique. This model was presented by Lyne et al. (2015).(b) The power-law fitting model Δν˙p=k1Δνγ1\Delta\dot{\nu}_{p}=k_{1}\Delta\nu^{\gamma_{1}} (denoted as M2), where k1k_{1} is the proportional constant, γ1\gamma_{1} is the power-law index, these are the free parameters in this model. This model release the polynomial index of linear model from 1 to a free parameter to embed possible nonlinear correlation between glitch size and persistent shift. (c) The modified power-law fitting model Δν˙p=Δνγ1(α1+ξ1Δνγ2)1\Delta\dot{\nu}_{p}=\Delta\nu^{\gamma_{1}}(\alpha_{1}+\xi_{1}\Delta\nu^{\gamma_{2}})^{-1}(denoted as M3), where the modified term on the simple power law ξ1Δνγ2\xi_{1}\Delta\nu^{\gamma_{2}} is a small quantity compared to the coefficient α1\alpha_{1}. Here there are four free parameters γ1\gamma_{1}, γ2\gamma_{2}, α1\alpha_{1} and ξ1\xi_{1}. This model naturally seems to include some higher-order effects of possible physical process. (4) The Log fitting model, Δν˙p=k2ln(α2Δν+ξ2)\Delta\dot{\nu}_{p}=k_{2}\ln(\alpha_{2}\Delta\nu+\xi_{2}) (denoted as M4) , where the k2k_{2}, α2\alpha_{2} and ξ2\xi_{2} are the free parameters. This model is currently lack of appropriate motivation, however, based on the observation of the overall trend of the data-set, this is a reasonable pick mathematically. In the following study, Bayesian analysis is performed to constrain the model parameters and the inter-connection among these models will be discussed in detail.

2.2 Bayesian inference set-up

Prior setting of these models reflects our preknowledge about the fitting free parameters, this is the essential components of Bayesian analysis. For linear model, the prior for k0k_{0} and b0b_{0} are uniform probablity distribution from 0 to 300, and -300 to 300, which denoted as 𝒰(0,300)\mathcal{U}(0,300) and 𝒰(300,300)\mathcal{U}(-300,300). The power-law model priors defined as the prior of k1k_{1} is 𝒰(0,300)\mathcal{U}(0,300), for power-law index γ1\gamma_{1} is 𝒰(0,3)\mathcal{U}(0,3). In modified power-law model, we set 𝒰(0,3)\mathcal{U}(0,3) for γ1\gamma_{1} and γ2\gamma_{2}, the α2\alpha_{2} and ξ2\xi_{2} prior are both 𝒰(0,0.1)\mathcal{U}(0,0.1).

Refer to caption
Figure 1: The heatmap of different Bayes factors given different models: (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M3: Log model

Regarding the likelihood set-up, our likelihood should reflect the observation data and their uncertainties, and as the constraint condition for reshaping the prior space of each models. Here, each individual observation of the glitch event with their accompanied persistent shift will associate with a likelihood function then compose all of them to a single global likelihood. As the uncertainty of glitch size measurement and persistent step width given, we view each individual observation as a Gaussian likelihood function with their 1-sigma width is determined by the uncertainty number. Every likelihood computation will be carried as following: Sampling the glitch size of the correlated persistent step observation via the given uncertainty-defined Gaussian distribution function. Then take the sampled glitch size into our prior-defined fitting model get a corresponded persistent shift value. Put these values inside of the Gaussian likelihood defined by Persistent shift observation, compute the likelihood from this observation then draw the posterior from this process by multiplying the likelihood with prior which defined the theoretical model utilized in this computation. This process could be expressed as a formula:

p(𝜽𝒅,)p(𝜽)\displaystyle p(\boldsymbol{\theta}\mid\boldsymbol{d},\mathcal{M})\propto p(\boldsymbol{\theta}\mid\mathcal{M}) (1)
×jp(Δν˙p,jdradio,j)p(Δνjdradio,j)\displaystyle\times\prod_{j}p\left(\Delta\dot{\nu}_{p,j}\mid d_{\mathrm{radio},\mathrm{j}}\right)p\left(\Delta\nu_{j}\mid d_{\mathrm{radio},\mathrm{j}}\right)

The meaning of this euqation is, for given model and data-set the posterior probability distribution of the model parameters are defined, according to Bayes’ theorem, as the product of the prior distribution of these parameters for a given model and the prior probability distribution of glitch size Δν\Delta\nu and persistent shift Δν˙p\Delta\dot{\nu}_{p} giving the radio observation data set dradiod_{\mathrm{radio}} as the likelihood.

All analysis in this study were performed using the CompactObject package, created by the authors. This open-source tool is designed for applying Bayesian constraints to the interiors of neutron stars, based on observational data. This research marks the first time the package has been applied to another physics problem other than equation of state inference. Previous studies using this package include those referenced in Huang et al. (2024). The sampling process employed in this research utilizes UltraNest, which uses a highly efficient step-sampler for nested sampling Buchner (2021). We configured the model comparisons with 50,000 live points for each model, a setting that dictates the extent of Bayesian exploration and guarantees a fair comparison among models. The average sampling duration for these models on a standard personal laptop is about 600 seconds, demonstrating the computational viability of this method given the manageable size of the observational data.

       Models        Baysian evidence (log(Z)\log(Z))
      M1       -280.4 ±\pm 0.032
      M2       -204.4 ±\pm 0.023
      M3       -167.7 ±\pm 0.046
      M4       -156.9 ±\pm 0.018
Table 2: Summary for the Bayesian evidence for different models: (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M4: Log model
       Models        Parameters posteriors (log(Z)\log(Z))
      M1       k0=68.941.61+1.61k_{0}=68.94^{+1.61}_{-1.61}, b0=11.591.60+1.60b_{0}=11.59^{+1.60}_{-1.60}
      M2       k1=87.311.25+1.23k_{1}=87.31^{+1.23}_{-1.25}, γ1=0.740.02+0.02\gamma_{1}=0.74^{+0.02}_{-0.02}
      M3       γ1=0.440.05+0.04\gamma_{1}=0.44^{+0.04}_{-0.05}, γ2=0.0960.044+0.032\gamma_{2}=-0.096^{+0.032}_{-0.044}, ξ1=0.00350.0006+0.0004\xi_{1}=0.0035^{+0.0004}_{-0.0006}, α1=0.00440.0010+0.0010\alpha_{1}=0.0044^{+0.0010}_{-0.0010}
      M4       k2=120.5912.21+13.80k_{2}=120.59^{+13.80}_{-12.21}, α2=0.890.03+0.03\alpha_{2}=0.89^{+0.03}_{-0.03}, ξ2=1.260.20+0.23\xi_{2}=1.26^{+0.23}_{-0.20}
Table 3: Summary for the Bayesian evidence for different models: (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M3: Log model
Refer to caption
Figure 2: The projection of posterior samples in ΔνΔν˙p\Delta\nu-\Delta\dot{\nu}_{p} given different models: (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M3: Log model. WhereΔν\Delta\nu is in μHz\mu\mathrm{Hz}, Δν˙p\Delta\dot{\nu}_{p} is in 1015s210^{-15}\mathrm{~{}s}^{-2}. The contour levels are 68% and 84% separately from inner to outer.
Refer to caption
Figure 3: The projection of residual of each pair of models in ΔνΔν˙p\Delta\nu-\Delta\dot{\nu}_{p} given different models pair: (1) M1 and M2 (blue) (2) M2 and M3 (green)(3) M3 and M4 (red). WhereΔν\Delta\nu is in μHz\mu\mathrm{Hz}, Δν˙p\Delta\dot{\nu}_{p} is in 1015s210^{-15}\mathrm{~{}s}^{-2}. The contour levels are 68% and 84% separately from inner to outer.

3 Inference result

Different model choices result in varying Bayesian evidence, reflecting how well each model explains the data. By analyzing this Bayesian evidence, we can perform model comparisons and draw data-driven conclusions about which models are favored by the observational data. Table 2 lists the logarithms of the Bayesian evidence for each models.

Model M1, the linear model that probably matches the starquake hypothesis, is the least favored by the observation data, with a very low value of -280.4. In contrast, the non-linear model, M2, shows a significant improvement in Bayesian evidence, rooted in the fact that the power-law function fits the large glitch-accompanied persistent shift better. Introducing a higher-order correction to Model M2, that is Model M3, further improves the Bayesian evidence significantly.

The final model, M4, particularly noteworthy because its functional form is quite another, yet it demonstrates the highest explanatory power for the observational data. This logarithmic-like dependence might represent some missed correction to the power-law models or a different mechanism altogether. In the following section, we will statistically show the convergence of each model, aiding in understanding the potential origin of this purely mathematical model.

However, it is important to note that, across all four models, the Bayesian evidence remains quite low, with the maximum value being -156.9. Considering this is the log(Z)\log(Z), the exponential of this value is extremely small. This may arise from the fact that the observational data for small glitch sizes accompanied by persistent shifts are well-observed, resulting in very small uncertainties, yet these phenomena seem challenging to explain smoothly using any elementary functional form. This raises questions about the reliability of the observations or suggests that the small glitch-related persistent shifts may require a more sophisticated model for adequate explanation.

In Figure 1, the heatmap of Bayes’ factors for different model pairs is presented. The Bayes factor is a standard method for model selection (see Kass & Raftery (1995)). A Bayes factor greater than 1 but less than 3.2 for model A versus model B suggests that the observations moderately favor model A, though this is not definitive. If the Bayes factor exceeds 3.2, it provides strong evidence that the data prefer model A. In this heatmap, the natural logarithm of the Bayes’ factors is shown.

This map indicates that from M1 to M4, each successive model represents a significant improvement over the previous one. The conclusion is robust, as the Bayes’ factors are exceptionally large, demonstrating that the measurements are sufficiently precise to distinguish among different theoretical models in this context. This provides overwhelmingly strong evidence from Bayesian inference to rule out the validity of the linear model. The large Bayes’ factor suggests that the linear model fails to capture some dominant feature of the system.

When comparing M3 to M2, the large Bayes factor indicates that the higher-order correction to an idealized power-law model is crucial for explaining the data. Even though this correction is theoretically minor, the data are already significant enough to distinguish this model from the one without the correction. The improvement from M3 to M4 is relatively smaller but still extremely strong. This gives us additional motivation to explore the physical meaning behind this model, as the improvement is critical, even though the model is currently only mathematical in nature.

The detailed posterior contour plots are provided in the supplementary materials. While the parameter space itself may be less interesting in this study—since the parameters lack direct physical interpretation and are purely model parameters—they become crucial in future theoretical discussions. However, translating these posterior samples back to the ΔνΔν˙p\Delta\nu-\Delta\dot{\nu}_{p} space is of interest, as it allows us to see the distinct regions in which these models reside. By comparing these regions with the original observational data, we can gain insights into the performance of the various models and correlate these findings with the Bayesian evidence discussed earlier.

All the parameters posteriors are summarized in Table 3. In Figure 2, we illustrate the projection of these posterior distributions of the four different models onto the ΔνΔν˙p\Delta\nu-\Delta\dot{\nu}_{p} plane. Our analysis shows that for all theoretical models, when the glitch size is smaller than approximately 3 μ\muHz, their performance is comparable, as they tend to occupy similar regions. However, for larger glitches, particularly the last data point with a glitch size exceeding 6 μ\muHz, the observed persistent shift is much smaller than predicted by M1 and M2, with the observation falling outside their 84% credible interval.

Comparing the overall bandwidth of different models, we observe that the 84% credible interval bandwidth increases sequentially from M1 to M4. This reflects that, from M1 to M4, the models expand the available parameter space to better explain the observational data. This expansion is also reflected in the Bayesian evidence, which measures the volume of the posterior parameter space.

An interesting observation is the sequentially increasing trend in Bayesian evidence from M1 to M4, which aligns with the models’ progressively better performance in explaining the last large glitch observation. This suggests that the large glitch observation is the key factor in distinguishing among the models from a statistical perspective. Notably, this large data point cannot be well-reproduced without compromising the fit for the small glitch accompanied by persistent shift observations, highlighting the challenge in balancing model accuracy across different scales of glitch sizes.

This increasing trend of approaching the last observation also suggests an inter-model numerical convergence relationship. Each model performs similarly at lower glitch sizes, but in the larger glitch region, from M1 to M4, the models sequentially approach the largest data point. In Figure 3, we show the posterior projection residuals between each pair of models, defined as Δν˙p,MiΔν˙p,Mj\Delta\dot{\nu}_{p,M_{i}}-\Delta\dot{\nu}_{p,M_{j}}, where the i-th model’s predicted persistent shift is subtracted from the j-th model’s prediction. This allows us to clearly see that all model comparisons approach zero for smaller glitches. However, for larger glitch sizes, the correction from the i-th model to the j-th model predicts a lower Δν˙p\Delta\dot{\nu}_{p}.

Statistically, this indicates that the models with better performance and larger Bayesian evidence account for higher-order corrections that become significant only in the case of large glitches. For the M3-M2 pair, this higher-order correction is theoretically well understood—it stems from new physics beyond the power-law process. As we show, this correction is relatively small for small glitches but becomes dominant for large glitch sizes.

For the M4-M3 pair, this higher-order correction is still evident. This suggests that the reason M4 outperforms the other models in Bayesian inference is likely due to this higher-order correction, assuming both models share the same foundational framework. However, the physical meaning behind the higher-order correction that M4 accounts for remains unclear and will require further modeling work. Despite this, the statistical behavior of the correction is clear from our analysis.

       Models        Baysian evidence (log(Z)\log(Z))
      M1       -635.1 ±\pm 0.026
      M2       -250.8 ±\pm 0.028
      M3       -173.3 ±\pm 0.052
      M4       -173.0 ±\pm 0.089
Table 4: Summary for the Bayesian evidence for different models after implementing the 2017 glitch event observation: (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M3: Log model
Refer to caption
Figure 4: The heatmap of different Bayes factors given different models after implementing the 2017 glitch event observation: (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M3: Log model
Refer to caption
Figure 5: The distribution of the Bayesian evidence accross different models, comparing the result with and without 2017 observation for (1) M1: Linear model. (2) M2: Power-law model. (3) M3: Modified power-law model. (4) M3: Log model
Refer to caption
Refer to caption
Figure 6: The Residue comparison (1) left panel: between the M3 and M2, (2) right panel: between the M4 and M3, The unite of the residue is 101510^{-15}μ\muHz\cdots2s^{-2}

4 Test in ultra glitch

In 2017, a new glitch event was recorded, accompanied by a persistent shift in the spin-down rate of the star. This glitch was obscured by the occurrence of three much smaller glitches in later about 600 days Shaw et al. (2018); Basu et al. (2020, 2022); Shaw et al. (2021) and hence they were incorported into an effective event as 16.43 μ\muHz, with a persistent shift of 434(8) in units of (1015s2)\left(10^{-15}\mathrm{~{}s}^{-2}\right). Since this event is more than twice the glitch size of the previous largest event, it presents a significant test of the current models discussed in this study. As a result, a detailed Bayesian inference was conducted with this new observation added to the dataset.

Table 4 summarizes the updated Bayesian evidence for the four models inclusion of the 2017 event. The overall sequence of Bayesian evidence remains consistent: from M1 to M4, the Bayesian evidence improves, supporting the conclusion that higher-order corrections are incorporated into the models, making them more accurate. Notably, the Bayesian evidence for the linear model has dropped to an extremely small value, which allows us to confidently exclude this linear model as an explanation and provides strong motivation to propose new theoretical models to interpret this observation.

In Figure 4, we present an updated heatmap of the Bayes’ factors. All the Bayes’ factors in the lower-left corner have increased, reflecting the fact that the inclusion of this new data makes the distinction among models even stronger. The only exception is the M4/M3 comparison, where the logarithm of the Bayes’ factor decreases from 12.90 to 0.3 after incorporating the 2017 event. This is a particularly interesting observation, as it suggests that the 2017 event might be better understood by M3. This single observation also is significant enough to support the overall conclusion. The dataset as a whole still favors Model 4, as the Bayes’ factor remains larger than 1, providing potential evidence that M4 is still the best model to explain the entire dataset.

In Figure 5, we compare the Bayesian evidence of different models with and without the 2017 event included. The conclusion is that, with the inclusion of this new observation, all models more and less face challenges in explaining this large event, as the logarithms of Bayesian evidence have dropped, and the posterior space has become more constrained. Especially, M1 and M2 exhibit a significant drop in evidence, suggesting that these two models are strongly constrained and potentially excluded by the new observation.

For M3 and M4, the decrease in Bayesian evidence is not as pronounced, particularly when compared to M1 and M2. If we compute the Bayes’ factor for the same model with and without the 2017 event, the logarithm of the Bayes’ factor is relatively small. Interestingly, the Bayes’ factor for M4 is even smaller than for M3, consistent with our earlier analysis showing that only the M4/M3 Bayes’ factor decreased when comparing the scenarios with and without the 2017 event.

5 Inter-models numerical conversions

Most of the models considered in this study are primarily mathematical in nature, with the physical interpretation of their parameters remaining largely undefined. However, even when treating these models as purely mathematical constructs, we can still gain insights into their characteristics by performing numerical expansions to various orders. This approach allows us to explore the inter-model conversions based on these expansions. It is possible that truncating higher-order effects could reduce the more data-favored models to resemble the less-favored ones.

In this analysis, we focus exclusively on the best-fitting parameter sets for the different models, specifically selecting the optimal parameter combinations listed in Table 3. The methodology is as follows: for each model pair, we subtract the less-favored model from the more data-favored model, and then numerically integrate this difference from zero to the final glitch observation point. This integration provides a global residue that quantifies the discrepancy between the models. Our primary focus is on the M3-M2 and M4-M3 model pair conversions. The M2-M1 pair is straightforward, as the first-order expansion of M2 can easily revert to the linear M1 model.

For the M3-M2 pair, see the left panel of Figure 6. The x-axis represents the expansion orders of M3, where the expansion is based on a Taylor series. We observe that the residue is minimized at the fourth-order expansion; beyond this point, increasing the expansion order causes the residue to rise again. This suggests that the fourth-order expansion of M3 is the best numerical fit for M2. The superior performance of M3 can be attributed to its incorporation of higher-order modifications from unknown physical sources, thereby numerically demonstrating the correlation between different models.

The residue for the M4-M3 pair is shown in the right panel of Figure 6. This comparison is more complex because the usual Taylor expansion of the M4 logarithmic model does not monotonically increase with additional terms—each positive term is followed by a negative one, making the expansion less effective at reproducing the less-favored model. To address this, we employ a special expansion of the logarithmic function:

k2ln(ξ2+α2Δν)=k2m=0((ξ2+α2Δν)1)mm(ξ2+α2Δν)mk_{2}\ln\left(\xi_{2}+\alpha_{2}\Delta\nu\right)=k_{2}\sum_{m=0}^{\infty}\frac{\left(\left(\xi_{2}+\alpha_{2}\Delta\nu\right)-1\right)^{m}}{m\left(\xi_{2}+\alpha_{2}\Delta\nu\right)^{m}} (2)

This expansion holds true as long as the condition Re(ξ2+α2Δν)>1/2\operatorname{Re}\left(\xi_{2}+\alpha_{2}\Delta\nu\right)>1/2 is satisfied. This condition is automatically met across the entire Δν\Delta\nu range, given that ξ2\xi_{2} is set to 1.26 and the minimum value of Δν\Delta\nu is zero. Using this expansion, and as shown in Figure 6, the residue stabilizes after 14 terms. Thus, we can assert that the M3 modified power-law model is equivalent to the 14th-order expansion of the M4 logarithmic model. The better performance of M4 on explaining the observation data can be attributed to its inclusion of higher-order corrections absent in the previous model.

This finding is particularly intriguing because, even without an explicit theoretical framework to derive the logarithmic model, we can numerically establish the connection between existing models and identify the orders of the source of the corrections. Interpreting the physical significance of these higher-order terms will require further theoretical modeling.

6 Conclusion

In this study, we conducted a detailed Bayesian analysis of the glitch-induced persistent increases in the spin-down rate of the Crab pulsar. By systematically comparing four models, ranging from a simple linear model (M1) to more sophisticated power-law and modified power-law models (M2 and M3), as well as a logarithmic model (M4). We hope to understand how well these models explain the observational data.

Our results indicate that higher-order corrections are crucial for accurately explaining the persistent shift accompanying larger glitches. From M1 to M4, the Bayesian evidence increases, showing that the models progressively improve in capturing the post-glitch behavior, especially when considering larger glitches. The linear model (M1) is significantly disfavored by the data. This highlights the need for more complex models that incorporate effects like superfluidity and its associated corrections, as represented by M2 and M3.

The recent 2017 glitch is a extremely large event. Its inclusion in our analysis further constrained the models, particularly M1 and M2, which were strongly disfavored. However, M3 and M4 remain competitive, with M4 still slightly outperforming M3 statistically and M3 maybe better capturing the nuances of this 2017 largest glitch event property. Further modeling work is needed to fully understand success of model M4.

The post-glitch behavior of the Vela is a classical relaxation process, associated with a response to vortex creep. Unlike this standard case, the Crab pulsar cannot relax toward the pre-glitch state after a glitch and has a persistent torque increase. The effect was generally attributed to a external torque variation triggered by starquakes Link et al. (1992); Link & Epstein (1997); Epstein & Link (2000). According to “plate tectonic” proposed by Ruderman (1991), the persistent shift has linear dependence of glitch size found by Zheng et al. (2024). However our analysis evidently disfavors the linear model explanation of observed data, which means that Bayesian evidence rules out starquake hypothesis. This convinces us that possible physics inside neutron stars should be taken seriously. Our future work will focus on finding real post-glitch physics of the Crab pulsar by the understandings of these models.

7 Software and third party data repository citations

CompactObject package: An full-scope open-source Bayesian inference framework especially designed for Neutron star physics https://github.com/ChunHuangPhy/CompactOject, documentation: https://chunhuangphy.github.io/CompactOject/. UltraNest: Buchner (2021), https://github.com/JohannesBuchner/UltraNest.

This paper is supported by the National SKA Program of China (Grant No. 2020SKA0120300) and the National Natural Science Foundation of China (Grant Nos. 12033001,12473039). C.H. also acknowledges support from an Arts & Sciences Fellowship at Washington University in St. Louis and from NASA grant 80NSSC24K1095.

References

  • Akbal & Alpar (2018) Akbal, O., & Alpar, M. A. 2018, MNRAS, 473, 621, doi: 10.1093/mnras/stx2378
  • Akbal et al. (2017) Akbal, O., Alpar, M. A., Buchner, S., & Pines, D. 2017, MNRAS, 469, 4183, doi: 10.1093/mnras/stx1095
  • Alpar (1977) Alpar, M. A. 1977, ApJ, 213, 527, doi: 10.1086/155183
  • Alpar & Baykal (1994) Alpar, M. A., & Baykal, A. 1994, MNRAS, 269, 849, doi: 10.1093/mnras/269.4.849
  • Alpar et al. (1996) Alpar, M. A., Chau, H. F., Cheng, K. S., & Pines, D. 1996, ApJ, 459, 706, doi: 10.1086/176935
  • Alpar et al. (1985a) Alpar, M. A., Nandkumar, R., & Pines, D. 1985a, ApJ, 288, 191, doi: 10.1086/162780
  • Alpar et al. (1985b) —. 1985b, ApJ, 288, 191, doi: 10.1086/162780
  • Alpar et al. (1984) Alpar, M. A., Pines, D., Anderson, P. W., & Shaham, J. 1984, ApJ, 276, 325, doi: 10.1086/161616
  • Anderson et al. (1981) Anderson, P. W., Alpar, M. A., Pines, D., & Shaham, J. 1981, in IAU Symposium, Vol. 95, Pulsars: 13 Years of Research on Neutron Stars, ed. W. Sieber & R. Wielebinski, 299
  • Anderson & Itoh (1975) Anderson, P. W., & Itoh, N. 1975, Nature, 256, 25, doi: 10.1038/256025a0
  • Basu et al. (2020) Basu, A., Joshi, B. C., Krishnakumar, M. A., et al. 2020, MNRAS, 491, 3182, doi: 10.1093/mnras/stz3230
  • Basu et al. (2022) Basu, A., Shaw, B., Antonopoulou, D., et al. 2022, MNRAS, 510, 4049, doi: 10.1093/mnras/stab3336
  • Baym & Pines (1971) Baym, G., & Pines, D. 1971, Annals of Physics, 66, 816, doi: 10.1016/0003-4916(71)90084-4
  • Buchner (2021) Buchner, J. 2021, The Journal of Open Source Software, 6, 3001, doi: 10.21105/joss.03001
  • Crawford & Demiański (2003) Crawford, F., & Demiański, M. 2003, ApJ, 595, 1052, doi: 10.1086/377470
  • Epstein & Link (2000) Epstein, R. I., & Link, B. 2000, in Astrophysics and Space Science Library, Vol. 254, Astrophysics and Space Science Library, ed. K. S. Cheng, H. F. Chau, K. L. Chan, & K. C. Leung, 95, doi: 10.1007/978-94-010-0878-5_12
  • Espinoza et al. (2011a) Espinoza, C. M., Jordan, C., Bassa, C., et al. 2011a, The Astronomer’s Telegram, 3777, 1
  • Espinoza et al. (2011b) Espinoza, C. M., Lyne, A. G., Stappers, B. W., & Kramer, M. 2011b, MNRAS, 414, 1679, doi: 10.1111/j.1365-2966.2011.18503.x
  • Fuentes et al. (2017) Fuentes, J. R., Espinoza, C. M., Reisenegger, A., et al. 2017, A&A, 608, A131, doi: 10.1051/0004-6361/201731519
  • Ge et al. (2020) Ge, M. Y., Zhang, S. N., Lu, F. J., et al. 2020, ApJ, 896, 55, doi: 10.3847/1538-4357/ab8db6
  • Glampedakis & Andersson (2009) Glampedakis, K., & Andersson, N. 2009, Phys. Rev. Lett., 102, 141101, doi: 10.1103/PhysRevLett.102.141101
  • Huang et al. (2024) Huang, C., Raaijmakers, G., Watts, A. L., Tolos, L., & Providência, C. 2024, Mon. Not. Roy. Astron. Soc., 529, 4650, doi: 10.1093/mnras/stae844
  • Kass & Raftery (1995) Kass, R. E., & Raftery, A. E. 1995, Journal of the American Statistical Association, 90, 773. http://www.jstor.org/stable/2291091
  • Keer & Jones (2015) Keer, L., & Jones, D. I. 2015, MNRAS, 446, 865, doi: 10.1093/mnras/stu2123
  • Larson & Link (2002) Larson, M. B., & Link, B. 2002, MNRAS, 333, 613, doi: 10.1046/j.1365-8711.2002.05439.x
  • Link (2014) Link, B. 2014, ApJ, 789, 141, doi: 10.1088/0004-637X/789/2/141
  • Link & Cutler (2002) Link, B., & Cutler, C. 2002, MNRAS, 336, 211, doi: 10.1046/j.1365-8711.2002.05726.x
  • Link & Epstein (1997) Link, B., & Epstein, R. I. 1997, ApJ, 478, L91, doi: 10.1086/310549
  • Link et al. (1992) Link, B., Epstein, R. I., & Baym, G. 1992, ApJ, 390, L21, doi: 10.1086/186362
  • Link et al. (1993) —. 1993, ApJ, 403, 285, doi: 10.1086/172202
  • Link et al. (1998) Link, B., Franco, L. M., & Epstein, R. I. 1998, ApJ, 508, 838, doi: 10.1086/306457
  • Lu et al. (2023) Lu, R., Yue, H., Lai, X., et al. 2023, MNRAS, 520, 4289, doi: 10.1093/mnras/stad270
  • Lyne et al. (2015) Lyne, A. G., Jordan, C. A., Graham-Smith, F., et al. 2015, MNRAS, 446, 857, doi: 10.1093/mnras/stu2118
  • Lyne et al. (2000) Lyne, A. G., Shemar, S. L., & Smith, F. G. 2000, MNRAS, 315, 534, doi: 10.1046/j.1365-8711.2000.03415.x
  • Lyne et al. (1992) Lyne, A. G., Smith, F. G., & Pritchard, R. S. 1992, Nature, 359, 706, doi: 10.1038/359706a0
  • Pines et al. (1972) Pines, D., Shaham, J., & Ruderman, M. 1972, Nature Physical Science, 237, 83, doi: 10.1038/physci237083a0
  • Ruderman (1976) Ruderman, M. 1976, ApJ, 203, 213, doi: 10.1086/154069
  • Ruderman (1991) —. 1991, ApJ, 366, 261, doi: 10.1086/169558
  • Serafin Nadeau et al. (2024) Serafin Nadeau, T., van Kerkwijk, M. H., Bassa, C. G., et al. 2024, ApJ, 962, 57, doi: 10.3847/1538-4357/ad18c5
  • Shaw et al. (2021) Shaw, B., Keith, M. J., Lyne, A. G., et al. 2021, MNRAS, 505, L6, doi: 10.1093/mnrasl/slab038
  • Shaw et al. (2018) Shaw, B., Lyne, A. G., Stappers, B. W., et al. 2018, MNRAS, 478, 3832, doi: 10.1093/mnras/sty1294
  • Sourie & Chamel (2020) Sourie, A., & Chamel, N. 2020, MNRAS, 493, L98, doi: 10.1093/mnrasl/slaa015
  • Wang & Zheng (2019) Wang, W., & Zheng, X. 2019, arXiv e-prints, arXiv:1906.12060, doi: 10.48550/arXiv.1906.12060
  • Wlazłowski et al. (2016) Wlazłowski, G., Sekizawa, K., Magierski, P., Bulgac, A., & Forbes, M. M. 2016, Phys. Rev. Lett., 117, 232701, doi: 10.1103/PhysRevLett.117.232701
  • Zheng et al. (2024) Zheng, X.-P., Wang, W.-H., Huang, C., & Yuan, J.-P. 2024, ApJ in prep