Understanding Artificial Neural Network’s Behavior from Neuron Activation Perspective
Abstract
This paper explores the intricate behavior of deep neural networks (DNNs) through the lens of neuron activation dynamics. We propose a probabilistic framework that can analyze models’ neuron activation patterns as a stochastic process, uncovering theoretical insights into neural scaling laws, such as over-parameterization and the power-law decay of loss with respect to dataset size. By deriving key mathematical relationships, we present that the number of activated neurons increases in the form of , and the neuron activation should follows power-law distribution. Based on these two mathematical results, we demonstrate how DNNs maintain generalization capabilities even under over-parameterization, and we elucidate the phase transition phenomenon observed in loss curves as dataset size plotted in log-axis (i.e. the data magnitude increases linearly). Moreover, by combining the above two phenomenons and the power-law distribution of neuron activation, we derived the power-law decay of neural network’s loss function as the data size scale increases. Furthermore, our analysis bridges the gap between empirical observations and theoretical underpinnings, offering experimentally testable predictions regarding parameter efficiency and model compressibility. These findings provide a foundation for understanding neural network scaling and present new directions for optimizing DNN performance.
1 Introduction
Deep neural networks (DNNs) have achieved unprecedented success across a wide range of applications, including natural language processing, computer vision, and scientific discovery (?, ?, ?, ?). Despite these advancements, understanding the underlying mechanisms that govern the behavior of DNNs remains a fundamental challenge.
Recent empirical studies have revealed consistent scaling laws in neural networks, highlighting two key phenomena: over-parameterization (?, ?) and the power-law decay of loss with respect to dataset size (?, ?) . Over-parameterization refers to the observation that DNNs with parameter sizes growing faster than dataset sizes can achieve remarkable generalization performance (?). Concurrently, the loss function of DNNs decreases approximately as a power-law function of the dataset size (?). While these scaling laws have been validated empirically, theoretical explanations for their emergence remain limited and often lack generality or consistency with experimental observations.
From the generalization error perspective, classical statistical learning theory suggests that the generalization error decreases with the size of the dataset, often inversely proportional to or (?). Some researchers attempt to extend these ideas to high-dimensional spaces by correlating the power-law exponent with the intrinsic dimensionality of the data. While this offers a plausible explanation, it falls short in explaining why such scaling laws are predominantly observed in neural networks and not in other models, such as random forests, which share similar statistical properties.
On the other hand, the capability quanta perspective, inspired by concepts from quantum mechanics, proposes that learning involves acquiring discrete units of task-solving ability, termed ”capability quanta.” (?) This theory effectively explains the observed power-law relationships but relies on counterintuitive assumptions, such as the existence of discrete generalization status, which violate the classical understanding to continuous generalization error bound, and power-law distribution of capability quanta.
In contrast, our approach focuses on the activation patterns of neurons and models their evolution as a stochastic process. By doing so, we derive theoretical results that not only explain key aspects of neural scaling laws, such as over-parameterization and the power-law decay of loss, but also align with experimental findings that are seldom discussed or newly reported. Moreover, our framework bridges the gap between empirical observations and theoretical explanations, offering new avenues for understanding the behavior of neural networks. Specifically, we analyze the evolution of neuronal activation as a stochastic process and study its relationship with dataset expansion and parameter growth. Through this approach, we derive theoretical results that align with key experimental findings:
-
•
The neural networks in over-parameterization, where the parameter size grows faster than the dataset size , i.e. , can still generalize.
-
•
We can observe a phase transition in the generalization process of neural network if we plot the data size in log-axis plot. This finding is implied by the figure 2 in the first neural scaling law paper from OpenAI (?), but seldom discussed by previous researches.
-
•
The power-law decay of loss as a function of , expressed as , where .
Furthermore, we extend our framework to make experimentally testable predictions regarding the role of noise-to-signal parameter ratios and the compressibility of large models. These predictions not only offer new insights into neural network scaling but also provide a foundation for future experimental validation.
The remainder of this paper is structured as follows. Section 2 presents the notations and assumptions of our probabilistic framework. In Section 3, we propose our central hypothesis, whose form is simple and clear, which assumes that when the training data size is increasing, the neuron activation distribution is steady in linear axis. In Section 4, based on our central hypothesis, we pointed out that the neuron activation distribution is a power-law distribution and acquired the growth function of activated neurons. In Section 5, based on the derived growth function, we present why neural networks under over-parameterization can still generalize, and why there exist a clear phase transition if we plot the data size in log-axis. Then using the phase transition effect and the power-law distribution, we derived the power-law decay of loss as a function of . After that, in Section 6, we provides new predictions, mainly about the parameter efficiency and pruning for neural network, based on our derived conclusions and theory framework. In Section 7, we review related work on neural scaling laws and existing theoretical interpretations. Finally, Section 8 concludes with a discussion of our limitation and potential future research directions.
2 Notations and Preliminary
In this paper, we consider a single layer of neurons in a deep neural network. The notation used throughout are as follows:
-
•
Number of neurons: Let denote the total number of neurons in the layer, and the index of each neuron is denoted by .
-
•
Training dataset: The size of the training dataset is .
-
•
Activation Function: We consider a neural network that uses ReLU as activation function:
(1) Such a function satisfy two important properties: (1) only activated neurons contribute to the final loss of the model; and (2) only activated neurons can receive gradients in backward propagation, which means that a sample is learnt by a neuron only if the neuron is activated.
-
•
Activation of neurons: During training on dataset , not every neuron is activated by every sample (Otherwise, the neural network will lose non-linearity). Let represent the number of samples that activate neuron . Since a neuron can only learn (i.e., be back-propagated) when it is activated, the also represent the sample number that neuron has learnt.
-
•
Expect of Activated Neuron Number: Modern deep neural networks often include normalization layers (e.g., batch normalization, layer normalization, or group normalization (?, ?, ?)), which constrain the number of neurons activated by each sample to be approximately constant across different samples, denoted as . The should be only relevant to hyper-parameter setting and the neuron number , and intuitively it should only increase when increases.
3 Central Hypothesis
In this paper, we consider a steady process that introduce a new sample into an existing dataset, i.e., moving from . This new sample is assumed to be drawn independently and identically distributed (i.i.d.) with the existing samples. Based on the above defined notations, we provide our central hypothesis: Stable Activation Probability Hypothesis:
-
•
Based on the i.i.d property of the new sample, the probability of neuron being activated by the new sample is proportional to , where is the number that neuron is activated by samples in , and is a small positive constant ensuring that even unactivated neurons have a non-zero probability of being activated.
Based on the above assumption, the probability of neuron being activated by the new sample is given by:
(2) |
where is a constant ensuring that unactivated neurons have a non-zero activation probability, is the approximate number of neurons activated per sample, and is the total number of activations across all neurons.
4 The dynamic of neuron activation
In this section, we analyze how the distribution of evolves as increases. Neurons are divided into two categories: ”working neurons”, which have been activated by at least one sample, and ”free neurons”, which have not been activated. By analyzing the stochastic process of neuron activation, we present two conclusions:
-
•
The number of working neuron, denoted as , increase as a function of , in the form of .
-
•
The distribution of activation number in the working neuron, denoted as , approximately follows power-law distribution: , where
4.1 The dynamic of working neuron number
We consider the transition from free to working neurons. Since the transition from free to working neurons is unidirectional, let denote the number of working neurons. Then, is a monotonically increasing function of . Based on the above probability definition, we derive the recurrence relation:
By approximating the discrete difference as a continuous change, we rewrite the relation in its differential form:
Solving the above differential equation yields the following solution for the number of working neurons :
where and represent the initial conditions. Solutions can be found in the Appendix. This solution describes how the number of working neurons evolves with the growth of the dataset size . For large model on large dataset (i.e. ), approximately follows the form of .
4.2 The distribution of working neuron’s activation number
We now consider the distribution of activation counts at a total sample size , denoted as . When is sufficiently large, the distribution reaches a steady state, where .
At steady state, the clusters with have a probability proportional to of transitioning to , while the clusters with have a probability proportional to of transitioning back to . This leads to the equilibrium equation:
Expanding this recurrence relation back to , we find that is proportional to . This distribution approximates a power-law distribution ,where . This range is because , which implies decrease as increases but slower than . Therefore, the forllows:
(3) |
where is the maximum of . Since we know , we have:
(4) |
Thus, we have .
5 Understanding Deep Neural Network Behavior
In this section, we apply the above two conclusions to understand three key empirical phenomenons:
-
•
The generalization of over-parametrized neural network
-
•
The phase transition of neural network generalization
-
•
The power-law decay of neural network loss regarding the increase of data size scale.
5.1 The generalization of over-parametrized neural network
From the above results, we can understand why overparamterized neural network (i.e. the neural network with ) highly generalize. Given that every sample activates neurons, we consider the average sample number of every base learner (working neuron), which is defined as:
Suppose that N grows with D with linearity and , i.e. N=kD=O(D), then the will become:
which grows when D and N grow. Thus, it means that if N grows with D with linearity, then the average sample of each base learner will grow together with D, finally leading to underfitting in some settings. Thus, N should grow a little bit faster then D, which aligns with experimental observation that .
5.2 The phase transition of neural network generalization

In (?), particularly the figure 2, the authors reported the curves about how neural language model’s performance change with data size when the parameter size is given. They reported the curves of model with various parameter sizes. And it can be observed that the validation loss of all models shows a phase transition phenomenon. More specifically, at very beginning, the model’s loss is very steady, which looks approximately as a flat line. Then as data increases, the model will get into a transition phase, in which the model’s loss dramatically reduces. After that, the model’s loss again become a approximately steady flat line.
Such phenomenon is very counterintuitive for machine learning area, in which the generalization error bound of a model should be a continuous function that decrease as data size increase (?). However, in this section, we are going to explain why such phase transition process generally happens for different models.
Given a series of neurons, satisfies the following approximation:
We provide the theoretic analysis to the above approximation. To give more intuitive justification, we put a figure (Fig. 1) showing how change with when (i.e. 10B, the normal scale of large language models), (approximately the square root of ) and . From the figure, we can see that across multiple magnitudes that is lower than , the linear approximation fits very well. And when the magnitude crosses , it dramatically change to a constant.
When the , i.e. , we see , which implies . This means that the working neuron number much fewer than free neurons, implying the output of the whole network is dominant by a random noise generated by a free noise with initial parameters. Thus, the loss should be close to the loss of a randomly initialized neural network, denoted as . Meanwhile, when , i.e., , from the over-parameterization nature of deep neural network, we can know that the sample numbers is sufficiently enough for the network to learn parameters that are close to optimal. We denote this loss as . It is obviously that there exists a clear gap between and . Thus, we can observe a phase transition phenomenon when we apply axis
5.3 The neural scaling law
In this section, we will discuss how to approximately derive neural scaling law by analyzing the above dynamics and distributions that we acquired.
Since the activation probabilities of the ”working neurons” in our theory also follows power-law distribution, it is very natural to relate the ”working neuron” with the ’ability quanta’ in (?). Meanwhile, when (i.e., ), the probability that a free neuron is activated is close to 0:
(5) |
Thus, when is sufficiently large, the overall loss of the neural network is only contributed by the ”working neurons” that are activated. In other words, the overall loss can be represented by the expect of the loss contributed by the activated neuron:
(6) | ||||
where represent the neuron index and is the probability that the -th neuron is activated.
Since a neuron’s output is actually the output of the sub-network that consists of its predecessor neurons and itself, then each working neuron’s loss should also satisfy the phase transition phenomenon that we previously discussed, i.e. :
Due to the symmetry of neural network architecture, for two neuron’s in the same layer, their predecessor layers are the same, and thus share same configuration of . These same configuration also implies same random initialized loss and optimal loss of these sub-networks, which implies:
where means the transition phase loss. We denoted the shared configuration as . Therefore, we have:
(7) | ||||
where refers to the expect of loss from those neurons in transition phase.
Meanwhile, as we derived previously in Section 4, the distribution of follows :
(8) |
where is the shape parameter of ’s distribution.
Then we have:
(9) |
Since , we can do Taylor expansion to to acquire . Thus . When and , we have
Therefore, we have:
(10) | ||||
From the expression of , we know that monotonically increases and . Thus, when is sufficiently large, . Thus, have:
(11) |
As we discussed, . Thus, the is actually a power-law function of and
6 Predictions of Unknown Phenomena
In the previous sections, we demonstrated how the proposed model explains two key phenomena in scaling laws: over-parameterization and the power-law decay of loss with dataset size. In this section, we extend the analysis to predict some previously unexplored phenomena based on our model.
We focus on the noise-to-signal parameter ratio, which quantifies the relationship between the model’s total parameters and the scale of free neurons . Particularly, this ratio can help us understand the optimal or sub-optimal ratio of pruning and compressing a neural network (?, ?), which shows significance in many areas (?). Using the formula for , we can derive:
When grows faster than , this ratio approaches 1 as increases. This result implies that the proportion of ineffective parameters in a large model increases as both the model size . When N grows faster than linearity of , we also see that when datasize increases, the ineffective parameter ratio increases too. Based on this observation, we propose several predictions:
1. Larger models are inherently more compressible.
2. With the same architecture, models trained on larger datasets achieve a lower compression ratio compared to those trained on smaller datasets. The ratio should approximately decrease as a power-low function of in the form of
3. With the same dataset, larger models generally achieve higher compression ratios than smaller models. The ratio should approximately grows as a power-low function in the form of
4. Over time, naively pre-trained large models will exhibit lower capability density compared to smaller models trained with more data.
It is worth noting that the second conclusion has been empirically verified in recent work. Specifically, the concept of capability density, introduced in (?), highlights that the effective parameter size of a model relative to its actual parameter size is increasing over time. This aligns with our theoretical prediction that the density of useful parameters in larger, naively pre-trained models diminishes relative to smaller, efficiently trained models. The study shows that the capability density of open-source models has doubled approximately every three months, supporting the idea that smaller, well-trained models can achieve comparable or even superior performance with fewer parameters.
7 Related Work
This work builds upon and extends previous studies in neural scaling laws (?, ?), which examine the relationship between compute, parameter size, dataset size, and the loss of neural networks. Since compute can be expressed as a function of parameter size and dataset size, we primarily focus on the relationships between parameter size and dataset size , as well as the relationship between dataset size and neural network loss .
Here, we define:
-
•
Dataset size : The number of training samples used to optimize the neural network.
-
•
Parameter size : The total number of trainable parameters in the neural network.
-
•
Loss : The objective function value that the neural network minimizes during training, often representing the prediction error on the dataset.
7.1 Generalization under Overparameterization
One critical aspect of neural scaling laws is the phenomenon of overparameterization. As neural networks scale, the parameter size increases faster than the dataset size . Mathematically, this can be expressed as:
indicating that the parameter size grows superlinearly with respect to the dataset size. Empirically, (?), indicating that the parameter size scales sub-linearly with respect to the dataset size . This overparameterization has been shown to improve generalization, despite the apparent risk of overfitting. Recent studies have empirically and theoretically validated this relationship across various architectures and tasks, highlighting that sufficiently large models are capable of capturing complex patterns in data, even with finite datasets.
7.2 Power-Law Relationship between Loss and Dataset Size
Another key observation in neural scaling laws (?) is the power-law decay of loss with increasing dataset size . Specifically, the loss can be expressed as:
where is a positive constant. This power-law relationship has been consistently observed in both language and vision models, suggesting that increasing the dataset size leads to diminishing returns in reducing loss. The exponent depends on the model architecture, the quality of the data, and the task, but its existence reflects a universal trend in neural scaling.
The insights from these two aspects of neural scaling laws provide a foundation for the theoretical analysis in this paper, particularly the derivation of scaling behavior in loss and the role of overparameterization in improving model performance.
7.3 Interpretations of Neural Scaling Laws
Existing works attempting to interpret neural scaling laws generally follow two primary approaches.
1. Generalization Error Perspective. From the perspective of classical statistical learning theory, the generalization error is typically inversely proportional to either the dataset size or the square root of (?, ?). Based on this, some researchers have proposed that the power-law exponent observed in scaling laws could be explained as the intrinsic dimensionality of the data in high-dimensional spaces. However, this explanation encounters a critical limitation: theoretically, it should also apply to other machine learning models, such as random forests. In practice, however, the power-law relationships are predominantly observed in neural networks and not in other machine learning models. This discrepancy casts doubt on the universality of this approach.
2. Capability Quanta Perspective. Another approach, inspired by ideas from quantum mechanics, has been proposed by researchers with a background in physics (?). This approach hypothesizes the existence of ”capability quanta,” analogous to energy quanta in quantum mechanics. Each capability quantum represents a discrete unit of task-solving ability. When a model acquires a new capability quantum, it unlocks a specific ability. Under this framework, the distribution of capability quanta follows a power-law distribution, and the data and parameter requirements for learning each quantum also exhibit power-law growth.
While this theory successfully explains the observed power-law relationships between loss, dataset size, and parameter size, it has several significant limitations:
-
•
Counterintuitive Assumptions: The theory assumes the existence of capability quanta. These quanta has discrete status, i.e. ”learnt” and ’non-learnt’. And the losses contributed by quantas with different status have a clear gap. As admitted by the authors themselves in their discussion of paper limitation, this assumption is very counterintuitive, and violates the observation that neural networks are optimized gradually. In machine learning area, the generalization error bound of a model is usually characterized as a continuous function of sample number .
-
•
Specific Distribution Assumptions: It further assumes that the distribution of capability quanta follows a power-law distribution. Other distributions, such as exponential or Poisson distributions with appropriate parameters, are not considered, making the framework appear overly tailored to fit the power-law observations.
These limitations highlight the challenges in fully explaining neural scaling laws and suggest the need for more robust and empirically consistent theoretical frameworks.
In this paper, we propose a new perspective to understand the behavior of neural networks. By analyzing the activation patterns of neurons and modeling their evolution as a stochastic process, we study how neural networks scale with increasing dataset size and parameter size . Through this approach, we derive conclusions consistent with existing experimental results, including the benefits of over-parameterization and the power-law decay of loss with . Furthermore, within this theoretical framework, we provide experimentally testable predictions.
8 Conclusion, Limitation and Future Work
This paper provides a probabilistic framework for understanding the impact of individual neuron activation patterns on the behavior of deep neural networks. The derived results align with experimental observations and offer testable predictions that can guide future research.
Concurrently, this paper leaves huge space for future works to improve on, including but not limited to:
-
•
Our stable activation probability hypothesis implies a data sample clustering process that is similar as Dirichlet Process and Pitman-Yor process (?, ?, ?). A natural question is: what is the deep mechanism of such clustering process in neural network?
-
•
In our paper, we mainly discuss the power-law decreasing relationship between and loss. Deriving the relationship about needs to analyze the term. This term is not related to our data size, but the expressive power of a neural network with limited parameters.
-
•
Extending this analysis to multiple layers. Our current analysis did not specify the configuration. As a result, we do not need to specify which layer we are analyzing. Therefore, the form of power-law relationship that we acquired can be extended to every layers, explaining the dynamics of deep neural networks. However, since different layers have different hyper-parameter configurations, precisely calculate the aggregation of them to acquire the precise equation for the whole neural network is still non-trivial. We suggest such analysis may require background knowledge in fractal theory and recursive analysis.
-
•
Why is shape unimportant for Transformer? This phenomenon, which is also important in neural scaling law, is still not answered. Intuitively, understanding this phenomenon requires us to precisely calculate the aggregation of different layer’s configurations so that we can acquire the precise equation for the whole neural network.
References and Notes
- 1. Zeyuan Allen-Zhu, Yuanzhi Li, and Zhao Song. A convergence theory for deep learning via over-parameterization. In International conference on machine learning, pages 242–252. PMLR, 2019.
- 2. Yasaman Bahri, Ethan Dyer, Jared Kaplan, Jaehoon Lee, and Utkarsh Sharma. Explaining neural scaling laws. Proceedings of the National Academy of Sciences, 121(27):e2311878121, 2024.
- 3. Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- 4. Zixiang Chen, Yuan Cao, Difan Zou, and Quanquan Gu. How much over-parameterization is sufficient to learn deep relu networks? arXiv preprint arXiv:1911.12360, 2019.
- 5. Pascal Germain, Francis Bach, Alexandre Lacoste, and Simon Lacoste-Julien. Pac-bayesian theory meets bayesian inference. Advances in Neural Information Processing Systems, 29, 2016.
- 6. Babak Hassibi and David Stork. Second order derivatives for network pruning: Optimal brain surgeon. Advances in neural information processing systems, 5, 1992.
- 7. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- 8. Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, et al. Training compute-optimal large language models. arXiv preprint arXiv:2203.15556, 2022.
- 9. Sergey Ioffe. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167, 2015.
- 10. Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361, 2020.
- 11. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097–1105, 2012.
- 12. Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. ArXiv e-prints, pages arXiv–1607, 2016.
- 13. Hao Li, Asim Kadav, Igor Durdanovic, Hanan Samet, and Hans Peter Graf. Pruning filters for efficient convnets, 2017.
- 14. Xiuyu Li, Yijiang Liu, Long Lian, Huanrui Yang, Zhen Dong, Daniel Kang, Shanghang Zhang, and Kurt Keutzer. Q-diffusion: Quantizing diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 17535–17545, 2023.
- 15. Sanae Lotfi, Marc Finzi, Sanyam Kapoor, Andres Potapczynski, Micah Goldblum, and Andrew G Wilson. Pac-bayes compression bounds so tight that they can explain generalization. Advances in Neural Information Processing Systems, 35:31459–31473, 2022.
- 16. Eric Michaud, Ziming Liu, Uzay Girit, and Max Tegmark. The quantization model of neural scaling. Advances in Neural Information Processing Systems, 36, 2024.
- 17. Issei Sato and Hiroshi Nakagawa. Topic models with power-law using pitman-yor process. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 673–682, 2010.
- 18. Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 15(1):1929–1958, 2014.
- 19. Yee Teh and Dilan Gorur. Indian buffet processes with power-law behavior. Advances in neural information processing systems, 22, 2009.
- 20. Yee Whye Teh et al. Dirichlet process. Encyclopedia of machine learning, 1063:280–287, 2010.
- 21. Yuxin Wu and Kaiming He. Group normalization. In Proceedings of the European conference on computer vision (ECCV), pages 3–19, 2018.
- 22. Chaojun Xiao, Jie Cai, Weilin Zhao, Guoyang Zeng, Xu Han, Zhiyuan Liu, and Maosong Sun. Densing law of llms. arXiv preprint arXiv:2412.04315, 2024.
Appendix A Derivative Equation Solution
Step 1: Rewrite the Equation First, rewrite the equation in the standard form of a first-order linear differential equation:
Here:
-
•
is a function of .
-
•
, , and are constants.
Step 2: Determine the Integrating Factor
The standard form of a first-order linear differential equation is:
Where:
-
•
-
•
The integrating factor is calculated as:
Step 3: Apply the Integrating Factor
Multiply both sides of the differential equation by the integrating factor:
This simplifies to:
Substituting the known and :
Simplify the right-hand side:
Step 4: Integrate Both Sides
Integrate both sides with respect to :
Where is the constant of integration.
Step 5: Solve for
Divide both sides by :
General Solution
Therefore, the general solution to the differential equation is:
Where is an arbitrary constant determined by initial conditions.
Example
Suppose there is an initial condition , then:
Solving for :
Substituting back into the general solution gives the particular solution that satisfies the initial condition:
This provides the complete solution to the differential equation.
Appendix B Approximation of
Here, we justify the following approximation:
Let us define . Then we can rewrite as a function of :
(12) |
When is much smaller than (i.e., ), the term approaches 1. This simplification allows us to approximate the function and its derivative as follows:
Raising both sides to the power of , we get:
Substituting this approximation back into the original function :
Taking the derivative of the approximated function:
Alternatively, when is much larger than (i.e., ), we can approximate the function and its derivative as follows:
Similarly, we substitute this approximation back into the original function :
Since the magnitude of is larger than , we have . Thus, we have:
which is a constant.