Fusing Multiple Algorithms for Heterogeneous Online Learning
Abstract
This study addresses the challenge of online learning in contexts where agents accumulate disparate data, face resource constraints, and use different local algorithms. This paper introduces the Switched Online Learning Algorithm (SOLA), designed to solve the heterogeneous online learning problem by amalgamating updates from diverse agents through a dynamic switching mechanism contingent upon their respective performance and available resources. We theoretically analyze the design of the selecting mechanism to ensure that the regret of SOLA is bounded. Our findings show that the number of changes in selection needs to be bounded by a parameter dependent on the performance of the different local algorithms. Additionally, two test cases are presented to emphasize the effectiveness of SOLA, first on an online linear regression problem and then on an online classification problem with the MNIST dataset.
1 Introduction
Multi-agent learning frequently involves scenarios in which agents gather disparate data at varying rates, collectively seeking to address an online optimization problem. Some instances include collaborative localization, search-and-rescue operations, coverage control, etc. Compounding the complexity of these scenarios is the constraint of limited data processing and computational capabilities, and the need for time-sensitive decision-making. Conventionally, either heterogeneous data or gradients are pooled [Zhang et al.(2018)Zhang, Cisse, Dauphin, and Lopez-Paz, Esfandiari et al.(2021)Esfandiari, Tan, Jiang, Balu, Herron, Hegde, and Sarkar, Aketi et al.(2023)Aketi, Kodge, and Roy], or heterogeneity is ignored and a distributed learning algorithm is deployed [Le Bars et al.(2023)Le Bars, Bellet, Tommasi, Lavoie, and Kermarrec, Zhu et al.(2021)Zhu, Xu, Liu, and Jin]. The former approach raises privacy concerns, while the latter proves suboptimal due to the inherent heterogeneity of the data. Effectively addressing the heterogeneity of data and resource constraints in an online learning problem remains an unresolved challenge.
An alternative approach to mitigate the challenges posed by data heterogeneity and resource constraints is to employ distinct algorithms tailored to the specific characteristics of the data, computation, and communication resources. Nevertheless, adopting distinct algorithms introduces the risk of underutilizing the available data. In this study, we present a systematic method to integrate updates provided by distinct algorithms to solve the heterogeneous online learning problem.
1.1 Problem Statement
We seek to solve the following online minimization problem:
(1) |
Here, is a parameter that needs to be collectively estimated by agents . The data gathered by each agent up to time is denoted by . More precisely, let denote the set of time instances when agent collects new data. Then , where denotes the round of sampling by agent . We use the set to denote all the time instances when new data is available. That is, . We drop the superscript to denote a time instance when new data is acquired by any agent. The elements of satisfy , where . Let the samples collected by agent at the round be denoted by . The data collected by agent until time is
(2) |
Every agent is constrained by their computation power and communication capabilities. Hence, they need to employ a local algorithm with their local data to solve the online optimization problem. For instance, a single agent can employ a centralized algorithm such as gradient descent (GD), stochastic gradient descent (SGD), or batch-wise GD [Dixit et al.(2019)Dixit, Bedi, Tripathi, and Rajawat]. An agent who is composed of a system of smaller units can collectively employ distributed algorithms such as Decentralized SGD or Federated Learning depending on the availability of resources. As an example, consider the following situation: an e-commerce entity equipped with data centers strategically situated across varied geographical locations, adopts an asynchronous data acquisition methodology from online users for the purpose of targeted advertisement display. In this scenario, each center employs a localized algorithm to process its specific dataset, owing to limitations in both computational and communicative capabilities.
Agents may choose to delay their decision-making process in order to accumulate an ample amount of data or computational resources necessary for solving problem (1). However, this approach may be suboptimal, given that decisions are frequently time-sensitive. A practical example is that of naval vessels mapping the sea for adversarial entities. Vessels positioned at varying distances from the shoreline collect sonar data, yet their computational and communicative capabilities are constrained by their respective locations. Specifically, vessels in closer proximity to the coast benefit from superior computational resources, albeit with diminished data quality, as elucidated by [Ferla and Jensen(2002)]. In this scenario, waiting to gather sufficient resources can be extremely dangerous. Therefore, the updates from different algorithms need to be fused in an online fashion as soon as updates are available to solve the problem (1).
1.2 Related Work
Distributed online optimization has been extensively studied as detailed in [McMahan(2017), Hoi et al.(2021)Hoi, Sahoo, Lu, and Zhao, Li et al.(2023b)Li, Xie, and Li]. However, these studies do not consider cooperatively using different algorithms in a constrained time-sensitive setting. Popular algorithms such as decentralized SGD and Federated Learning in the presence of asynchronous agents were studied in [Jiang et al.(2021)Jiang, Zhang, Gu, and Zhu, Chen et al.(2020)Chen, Ning, Slawski, and Rangwala]. In [Jiang et al.(2021)Jiang, Zhang, Gu, and Zhu], decentralized SGD is proposed with asynchronous agents but it does not incorporate agents running different algorithms and it also requires extensive communication between agents. Asynchronous online Federated Learning [Chen et al.(2020)Chen, Ning, Slawski, and Rangwala] requires the presence of a coordinator, and still does not fuse different algorithms. Model fusion has received attention in supervised learning [Li et al.(2023a)Li, Peng, Zhang, Ding, Hu, and Shen]. Model fusion in online learning case has received significantly less interest due to its complexities [Foster et al.(2015)Foster, Rakhlin, and Sridharan, Hoang et al.(2019)Hoang, Hoang, Low, and How, Cutkosky(2019)]. These works typically consider selecting models from several algorithms at every time step. Particularly, [Cutkosky(2019)] shows that algorithms with bounded regrets can be fused simply by averaging the parameters and still maintain bounded regret. However, in our case, we seek to fuse the updates from different agents by employing only a single agent at any given time. Therefore, it is unclear how agents with different data and resources could cooperatively solve problem (1) by running their own local algorithm.
1.3 Contributions
We provide an algorithm called the Switched Online Learning Algorithm (SOLA) to solve Problem (1) by fusing the updates from agents running different algorithms. We solve the considered problem by switching between the agents and fusing their updates based on their performance. We provide a sufficient condition to guarantee a bound on the regret of SOLA based on the rate at which different algorithms are chosen. To this end, we model SOLA as a switched dynamical system and ensure its contractivity. We numerically analyze the performance of our algorithm for the online linear regression problem and also the online classification problem with the MNIST dataset111Code repository: https://github.com/Shivanshu-007/Heterogeneous-online-optimization .
2 Switched Online Learning Algorithm (SOLA)
In this section, we describe the proposed Switched Online Learning Algorithm (SOLA). The input to any local algorithm at time is the data , and the parameter . Note that the dimension of is dependent on time as each agent acquires new data over time. The number of samples is given by and the number of features is . Henceforth, we simply use the variable to denote the discrete instances , i.e. . The update provided by algorithm is given by the map . At any instance , the selecting signal selects an agent if the agent has new data. Agent uses its local algorithm to update the parameter . The update provided by is used to solve the problem (1) as
(3) |
We call as the fusing variable. We introduce it to smoothly incorporate new updates from the chosen local algorithms . The fusing variable depends on the performance of the algorithm of the local algorithms. Let us define a performance metric as . Common performance metrics for classification problems are precision, recall, true positive rate, etc. [Hand(2012)]. In regression problems, performance is often computed as the inverse of the norm of error, or the trace of the inverse of error covariance [Kay(1993)]. We are particularly interested in metrics that have higher values to signify better performance. Given the performance metric, the fusing variable is defined as:
(4) |
where . Note that is using the performance of to incorporate its update. If the performance is poor on the local data , the fusing variable is closer to , whereas is closer to 1. This implies that the former parameter has a higher influence on the updated parameter . Similarly, if the performance of is superior to that of , more weight is given to . We describe the proposed algorithm SOLA in detail in Algorithm 1.
Remark 2.1.
(Naive switching): The design of the fusing variable is crucial to SOLA because the updates from can be vastly different from . For instance, by setting , the update of one agent acts as the input to the subsequent agents. This is a naive method of switching between agents without any consideration for the performance of each agent. Switching naively between different agents may cause abrupt and large changes in the parameters, which may be undesirable. This is illustrated in the online regression problem represented in Figure 1. SOLA chooses between two agents running centralized GD and decentralized SGD with five sub-units. Figure 1 compares the naive case when , and the case when is dependent on the performance of the local algorithms. We see frequent jumps in the parameter resulting in a chattering behavior with an improper choice of . However, when , we see an improvement in the performance and the chattering behavior is absent. The detailed simulation setting is provided in Section 4.1.
Remark 2.2.
(Choice of performance metric): SOLA uses updates from different agents for a general online optimization problem. Therefore, the optimal choice of a performance metric is dependent on the cost function, data and the constraints of each agent. For instance, consider the online linear regression problem. The recursive least squares algorithm in [Kay(1993)] incorporates data arriving incrementally to solve an online linear regression problem. Although the algorithm operates with only a single agent, it incorporates incoming updates by weighting them with the inverse of the error covariance. It can be shown that the choice of using the inverse of the error covariance in the recursive least-squares algorithm produces the Best Linear Unbiased Estimator (BLUE).
3 Design of Selecting Signal
The selecting signal determines the choice of subsystem for the update, and in turn, determines the regret of SOLA. In this section, we analyze the effect of the switching signal on the regret of the Switched Learning Algorithm. Let us denote any algorithm employed to solve problem (1) as . The regret of algorithm is denoted by and is defined for as
(5) |
Here refers to the parameters provided by algorithm at time using data , and is the optimal parameter given all the data a priori. It is essential for an algorithm to have bounded regret as it compares the performance of the algorithm to an optimal choice. We now show that SOLA has bounded regret under the right conditions of the selecting signal .
We analyze the regret achieved by SOLA by viewing the algorithm as a dynamical system. Optimization algorithms have received extensive attention from the view of dynamical systems [Elisseeff et al.(2005)Elisseeff, Evgeniou, Pontil, and Kaelbing, Shikhman and Stein(2009), Ross and Bagnell(2011), Saha et al.(2012)Saha, Jain, and Tewari, Sontag(2022), Cisneros-Velarde and Bullo(2022), Kozachkov et al.(2022)Kozachkov, Wensing, and Slotine]. Particularly, algorithms that are stable have been shown to have bounded regret [Ross and Bagnell(2011), Saha et al.(2012)Saha, Jain, and Tewari]. A useful tool to study these optimization algorithms and stability is contraction theory [Cisneros-Velarde and Bullo(2022), Kozachkov et al.(2022)Kozachkov, Wensing, and Slotine]. A contracting algorithm is defined as follows.
Definition 3.1.
(Contracting algorithm, [Lohmiller and Slotine(1998)]) Let the updates provided by an algorithm be given by the dynamical system
(6) |
where and are the parameter and data at time , respectively. The differential dynamics of the algorithm is then given by
The associated distance for the differential dynamics is denoted by
Here is a symmetric positive-definite matrix function, and it is uniformly bounded as , for all . The algorithm is said to be contracting if
(7) |
where , is the rate of contraction.
We show that any contracting algorithm achieves bounded regret in Appendix A. Given that contracting algorithms achieve bounded regret, it is sufficient to ensure that SOLA is a contracting algorithm by designing the switching signal . We make the following assumptions for our regret analysis.
-
(A1)
is -convex in : .
-
(A2)
Every local algorithm is contracting with a rate . Further, for each pair of algorithms , there exists such that the distance of differential dynamics is bounded as .
Assumption (A1) is a commonly used in regret analysis for online problems [McMahan(2017)]. However, (A2) needs more careful attention as it comes from the perspective of dynamical systems. Several algorithms such as gradient descent [Cisneros-Velarde and Bullo(2022)], SGD and decentralized SGD algorithms have been proven to be contractive [Boffi and Slotine(2020)]. When these algorithms are used to solve a common online learning problem, assumption (A2) captures the relationship between the differential dynamics of the local algorithms. Particularly, it provides an upper bound on the relative difference between the performance of the various local algorithms. We introduce and which will be useful for our regret analysis.
The variables corresponds to the largest difference in differential dynamics and corresponds to the slowest contracting rate. The following theorem addresses the design of the selecting signal such that SOLA achieves bounded regret.
Theorem 3.2.
(SOLA achieves bounded regret) For a selecting signal , let the number of switches between different agents be over any horizon , where . That is,
(8) |
Then, if the number of switches satisfies
(9) |
where is a constant, and , then SOLA achieves bounded regret: , where is a decreasing function in .
We refer the reader to Appendix B for the proof of Theorem 3.2. In the literature related to switched systems [Liberzon(2003)], is typically referred to as the chatter bound and is referred to as the average dwell time.
Remark 3.3.
(Effect of cost function, data, and algorithm on the number of switches) It is important to note that captures the effect of the cost function, the data and the characteristics of the different local algorithms on the admissible switching signal in Theorem 3.2. For instance, poor data or a slow learning rate for the agent results in high values of . This leads to a higher value of , which in turn means that the number of switches needs to be small. Conversely, algorithms that learn fast allow for fast switching. Another aspect of is the similarity of the behavior of local algorithms captured through Assumption (A2). If the different local algorithms behave similarly, . This leads to a small value of . Conversely, if the algorithms behave very differently is larger which restricts the frequency of switching local algorithms.
4 Numerical results
In this section, we show the effectiveness of SOLA for online linear regression and online classification using the MNIST dataset.
4.1 Online linear regression
We conduct experiments for online linear regression with a synthetic dataset. The agents acquire data , where , , and . The online linear regression problem is
In this experiment, the number of agents . One agent acquires data sampled as , where . The other agent collects data , where . Here, Figure 2 compares the performance of SOLA for different settings. The selecting signal periodically chooses between the two agents every ten instances, i.e. . In Figure 2, we compare the performance of SOLA in two different settings, (i) Agent 1 has one sub-unit that uses centralized gradient descent, and agent 2 uses decentralized SGD with 5 sub-units, (ii) Agent 1 has 5 sub-units that perform FedAvg [McMahan et al.(2017)McMahan, Moore, Ramage, Hampson, and y Arcas] and agent 2 uses decentralized SGD. Lastly, we compare the performance of only agent 2 using decentralized SGD without SOLA. From Figure 2, it is evident SOLA with GD and decentralized SGD performs the best, whereas SOLA with FedAvg with decentralized SGD converges slower. Decentralized SGD by itself converges slower and has a higher error.
In Figure 2(b), we compare SOLA with . The three agents perform centralized GD, decentralized GD with five sub-units and FedAvg with five sub-units. The data used by the agents is the same as mentioned above. It is evident that a naive choice of the fusing variable not only causes more error but also leads to chattering. SOLA has a higher error when as compared to when . This is because fusing more agents requires a good choice of the fusing variable and data of good quality.
4.2 Online classification
In this experiment, we consider the online classification problem with and the cost is the cross-entropy loss. Here, one agent samples data from the MNIST dataset and employs centralized gradient descent. It uses a neural network that has one hidden layer with 128 neurons. The second agent performs decentralized SGD with sub-units having the same neural network architecture as agent 1. We compare the performance of SOLA for different numbers of sub-units for the second agent. In the first case, there are five sub-units and each sub-unit acquires only two labels. For example, the first sub-unit receives labels and , the second sub-unit receives labels and , and so on. In the second case, there are ten sub-units and each sub-unit receives only one distinct label. Further, every sub-unit is restricted to possess only images (reducing computation). The images sampled by the sub-units are from the MNIST dataset but marred with Gaussian noise of variance and zero mean. Figures 3(a) and (b) show the accuracy and entropy loss of SOLA on testing data over time. We denote the number of sub-units as in Figure 3. We observed that the performance metric gave the best performance for both accuracy and loss. The selecting signal periodically chooses between agents every five instances, i.e. . It can be seen that SOLA still achieves an accuracy close to eighty-two percent for both cases of five and ten sub-units. However, the naive fusing choice with five sub-units has lower accuracy and more chattering.
5 Conclusions and Future Work
In this work, we considered the scenario where agents with different data and resources use different local algorithms to solve an online learning problem. Our proposed algorithm, SOLA, provides a way to systematically fuse the updates from different algorithms and ensure that regret is bounded. We also numerically analyze the performance of SOLA for different online learning scenarios. Future directions include the case of dynamically changing data distributions, tighter regret bounds and the adversarial case where byzantine agents provide malicious updates.
Appendix A Contracting Optimizers Achieve Bounded Regret
The connection between the stability of learning algorithms and bounded regret for online learning problems has been studied in [Poggio et al.(2011)Poggio, Voinea, and Rosasco, Ross and Bagnell(2011), Saha et al.(2012)Saha, Jain, and Tewari]. Particularly, in [Poggio et al.(2011)Poggio, Voinea, and Rosasco], the notion of online stability is defined as follows:
Definition A.1.
(Online stability) An algorithm is said to be online stable if
(10) |
where as .
The notion of online stability captures the fact that for an online stable algorithm, the change in cost incurred between any time instances is bounded by a non-increasing function . Importantly, Theorem 18 in [Ross and Bagnell(2011)] shows that online-stable algorithms achieve bounded regret where . Also, [Poggio et al.(2011)Poggio, Voinea, and Rosasco] shows that iterative gradient-based methods such as GD, and SGD achieve online stability. We first model any iterative gradient descent algorithm as a perturbated gradient descent.
(11) |
The perturbation to the true gradient is and is a constant learning rate. Further, the covariance of the perturbation satisfies , where denotes the zero matrix. This model is commonly used in the analysis of algorithms such as SGD [Jastrzebski et al.(2017)Jastrzebski, Kenton, Arpit, Ballas, Fischer, Bengio, and Storkey, Mandt et al.(2017)Mandt, Hoffman, and Blei, Li and Orabona(2019)] and decentralized SGD [Boffi and Slotine(2020)]. The dynamics of any single sub-unit of decentralized SGD or FedAvg can be expressed with (11). Further, the average parameter of all the sub-units also follows the same dynamics, however, the noise characteristics of differ. Here, we show that perturbed iterative gradient-based algorithms are contracting as given by Definition 3.1, and are are online stable.
Theorem A.2.
Proof A.3.
The differential dynamics of the system (11) is given by
(12) | ||||
(13) |
Here is treated as an external signal. For the Euclidean metric , the distance of the differential dynamics is given by
(14) |
Note that due to the -convexity of . Therefore,
(15) |
Taking the expectation with respect to the perturbation ,
(16) |
If the learning rate ensures that
(17) |
then is contractive. The proof of online stability follows along the same lines as Theorem 2 of [Poggio et al.(2011)Poggio, Voinea, and Rosasco]. The cost can be expressed using the first two derivatives by Taylor expansion. The stability of the perturbed gradient can be then used to bound the change in cost as given by Definition A.1.
Appendix B Proof of Theorem 3.2
First, consider SOLA with the fusing variable . Then,
(18) |
By assumption (A2), we have that between any two instances and
To ensure that SOLA is a contraction, we need that
To admit at least one switch if , we introduce . Therefore, when the number of switches satisfies
(19) |
SOLA is a contracting optimizer that achieves online stability. Further, when , SOLA uses a convex combination of the update by the local algorithm and the parameter at the previous time instance . The convex combination of contracting systems also results in a contracting system as shown in [Lohmiller and Slotine(1998)]. Hence, the overall SOLA algorithm achieves online stability.
References
- [Aketi et al.(2023)Aketi, Kodge, and Roy] S. A. Aketi, S. Kodge, and K. Roy. Neighborhood gradient mean: An efficient decentralized learning method for non-iid data. Transactions on Machine Learning Research, 2023.
- [Boffi and Slotine(2020)] N. M. Boffi and J.-J. E. Slotine. A continuous-time analysis of distributed stochastic gradient. Neural computation, 32(1):36–96, 2020.
- [Chen et al.(2020)Chen, Ning, Slawski, and Rangwala] Y. Chen, Y. Ning, M. Slawski, and H. Rangwala. Asynchronous online federated learning for edge devices with non-iid data. In 2020 IEEE International Conference on Big Data (Big Data), pages 15–24. IEEE, 2020.
- [Cisneros-Velarde and Bullo(2022)] P. Cisneros-Velarde and F. Bullo. A contraction theory approach to optimization algorithms from acceleration flows. In International Conference on Artificial Intelligence and Statistics, pages 1321–1335. PMLR, 2022.
- [Cutkosky(2019)] A. Cutkosky. Combining online learning guarantees. In Conference on Learning Theory, pages 895–913. PMLR, 2019.
- [Dixit et al.(2019)Dixit, Bedi, Tripathi, and Rajawat] Rishabh Dixit, Amrit Singh Bedi, Ruchi Tripathi, and Ketan Rajawat. Online learning with inexact proximal online gradient descent algorithms. IEEE Transactions on Signal Processing, 67(5):1338–1352, 2019.
- [Elisseeff et al.(2005)Elisseeff, Evgeniou, Pontil, and Kaelbing] A. Elisseeff, T. Evgeniou, M. Pontil, and L. P. Kaelbing. Stability of randomized learning algorithms. Journal of Machine Learning Research, 6(1), 2005.
- [Esfandiari et al.(2021)Esfandiari, Tan, Jiang, Balu, Herron, Hegde, and Sarkar] Y. Esfandiari, S. Y. Tan, Z. Jiang, A. Balu, E. Herron, C. Hegde, and S. Sarkar. Cross-gradient aggregation for decentralized learning from non-iid data. In International Conference on Machine Learning, pages 3036–3046. PMLR, 2021.
- [Ferla and Jensen(2002)] C. M. Ferla and F. B. Jensen. Are current environmental databases adequate for sonar predictions in shallow water? In Impact of Littoral Environmental Variability of Acoustic Predictions and Sonar Performance, pages 555–562. Springer, 2002.
- [Foster et al.(2015)Foster, Rakhlin, and Sridharan] D. J. Foster, A. Rakhlin, and K. Sridharan. Adaptive online learning. Advances in Neural Information Processing Systems, 28, 2015.
- [Hand(2012)] D. J. Hand. Assessing the performance of classification methods. International Statistical Review, 80(3):400–414, 2012.
- [Hoang et al.(2019)Hoang, Hoang, Low, and How] T. N. Hoang, Q. M. Hoang, K. H. Low, and J. How. Collective online learning of gaussian processes in massive multi-agent systems. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 7850–7857, 2019.
- [Hoi et al.(2021)Hoi, Sahoo, Lu, and Zhao] S. C. H. Hoi, D. Sahoo, J. Lu, and P. Zhao. Online learning: A comprehensive survey. Neurocomputing, 459:249–289, 2021.
- [Jastrzebski et al.(2017)Jastrzebski, Kenton, Arpit, Ballas, Fischer, Bengio, and Storkey] S. Jastrzebski, Z. Kenton, D. Arpit, N. Ballas, A. Fischer, Y. Bengio, and A. Storkey. Three factors influencing minima in sgd. arXiv preprint arXiv:1711.04623, 2017.
- [Jiang et al.(2021)Jiang, Zhang, Gu, and Zhu] J. Jiang, W. Zhang, J. Gu, and W. Zhu. Asynchronous decentralized online learning. Advances in Neural Information Processing Systems, 34:20185–20196, 2021.
- [Kay(1993)] S. M. Kay. Fundamentals of statistical signal processing: estimation theory. Prentice-Hall, Inc., 1993.
- [Kozachkov et al.(2022)Kozachkov, Wensing, and Slotine] L. Kozachkov, P. Wensing, and J. J. E. Slotine. Generalization as dynamical robustness–the role of riemannian contraction in supervised learning. Transactions on Machine Learning Research, 2022.
- [Le Bars et al.(2023)Le Bars, Bellet, Tommasi, Lavoie, and Kermarrec] B. Le Bars, A. Bellet, M. Tommasi, E. Lavoie, and A.M. Kermarrec. Refined convergence and topology learning for decentralized sgd with heterogeneous data. In International Conference on Artificial Intelligence and Statistics, pages 1672–1702. PMLR, 2023.
- [Li et al.(2023a)Li, Peng, Zhang, Ding, Hu, and Shen] W. Li, Yong Peng, Miao Zhang, Liang Ding, Han Hu, and Li Shen. Deep model fusion: A survey. arXiv preprint arXiv:2309.15698, 2023a.
- [Li and Orabona(2019)] X. Li and F. Orabona. On the convergence of stochastic gradient descent with adaptive stepsizes. In The 22nd international conference on artificial intelligence and statistics, pages 983–992. PMLR, 2019.
- [Li et al.(2023b)Li, Xie, and Li] X. Li, L. Xie, and N. Li. A survey on distributed online optimization and online games. Annual Reviews in Control, 56:100904, 2023b.
- [Liberzon(2003)] D. Liberzon. Switching in systems and control, volume 190. Springer, 2003.
- [Lohmiller and Slotine(1998)] W. Lohmiller and J.-J. E. Slotine. On contraction analysis for non-linear systems. Automatica, 34(6):683–696, 1998.
- [Mandt et al.(2017)Mandt, Hoffman, and Blei] S. Mandt, M. D. Hoffman, and D. M. Blei. Stochastic gradient descent as approximate bayesian inference. arXiv preprint arXiv:1704.04289, 2017.
- [McMahan et al.(2017)McMahan, Moore, Ramage, Hampson, and y Arcas] B. McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pages 1273–1282. PMLR, 2017.
- [McMahan(2017)] H. B. McMahan. A survey of algorithms and analysis for adaptive online learning. The Journal of Machine Learning Research, 18(1):3117–3166, 2017.
- [Poggio et al.(2011)Poggio, Voinea, and Rosasco] T. Poggio, S. Voinea, and L. Rosasco. Online learning, stability, and stochastic gradient descent. arXiv preprint arXiv:1105.4701, 2011.
- [Ross and Bagnell(2011)] S. Ross and J. A. Bagnell. Stability conditions for online learnability. arXiv preprint arXiv:1108.3154, 2011.
- [Saha et al.(2012)Saha, Jain, and Tewari] A. Saha, P. Jain, and A. Tewari. The interplay between stability and regret in online learning. arXiv preprint arXiv:1211.6158, 2012.
- [Shikhman and Stein(2009)] V. Shikhman and O. Stein. Constrained optimization: projected gradient flows. Journal of optimization theory and applications, 140:117–130, 2009.
- [Sontag(2022)] E. D. Sontag. Remarks on input to state stability of perturbed gradient flows, motivated by model-free feedback control learning. Systems & Control Letters, 161:105138, 2022.
- [Zhang et al.(2018)Zhang, Cisse, Dauphin, and Lopez-Paz] H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz. mixup: Beyond empirical risk minimization. In International Conference on Learning Representations, 2018.
- [Zhu et al.(2021)Zhu, Xu, Liu, and Jin] H. Zhu, J. Xu, S. Liu, and Y. Jin. Federated learning on non-iid data: A survey. Neurocomputing, 465:371–390, 2021.