One-Pass Learning via Bridging Orthogonal Gradient Descent and Recursive Least-Squares
Abstract
While deep neural networks are capable of achieving state-of-the-art performance in various domains, their training typically requires iterating for many passes over the dataset. However, due to computational and memory constraints and potential privacy concerns, storing and accessing all the data is impractical in many real-world scenarios where the data arrives in a stream. In this paper, we investigate the problem of one-pass learning, in which a model is trained on sequentially arriving data without retraining on previous datapoints. Motivated by the increasing use of overparameterized models, we develop Orthogonal Recursive Fitting (ORFit), an algorithm for one-pass learning which seeks to perfectly fit every new datapoint while changing the parameters in a direction that causes the least change to the predictions on previous datapoints. By doing so, we bridge two seemingly distinct algorithms in adaptive filtering and machine learning, namely the recursive least-squares (RLS) algorithm and orthogonal gradient descent (OGD). Our algorithm uses the memory efficiently by exploiting the structure of the streaming data via an incremental principal component analysis (IPCA). Further, we show that, for overparameterized linear models, the parameter vector obtained by our algorithm is what stochastic gradient descent (SGD) would converge to in the standard multi-pass setting. Finally, we generalize the results to the nonlinear setting for highly overparameterized models, relevant for deep learning. Our experiments show the effectiveness of the proposed method compared to the baselines.
I Introduction
While deep neural networks have been successful in numerous domains, their training is computationally demanding and requires iterating over the entire dataset multiple times. This hinders their deployment in many real-world settings such as robot learning, autonomy, and online decision making, where new datapoints are collected over time or become available sequentially. In such settings, storing all the datapoints and retraining the model at every step on all the data is extremely costly and often not feasible. In addition, in certain applications, storing the data may be prohibited for privacy reasons.
Thus, it is very desirable to come up with algorithms that can learn incrementally or in an online fashion, rather than by iterating over the entire data many times. However, it is well-known that deep neural networks are prone to entirely forgetting past information while learning new data, which is an issue referred to as “catastrophic forgetting” [10]. This begs the question:
“Can we learn streaming data efficiently without forgetting or retraining on previous data?”
This is a setting often referred to as one-pass learning. More specifically, one-pass learning concerns the setting where the algorithm (i) makes an update for the current datapoint without direct access to previous data; (ii) the new updates do not significantly affect the predictions on the previous data; moreover, (iii) the computational and memory costs of each update must not grow with the iteration count.
There has been growing attention on one-pass learning and its variants, and several works have attempted to address them under various contexts. In particular, [8] studied learning the ImageNet dataset in a single pass by revisiting some “important” previous datapoints at each learning step, and [19] investigated learning incremental ‘batches’ on a large scale by correcting the classifier’s bias towards new data. However, both methods train on previous data and are not adequate for one-pass learning. On the other hand, [15] proposed an effective one-pass learning method for support vector machines. However, their method is tailored to the specific setting of support vector machines. Further, [16] proposed a one-pass deep learning algorithm, but it relies on a specific network architecture and is vulnerable to forgetting the previous data unless the data is consistently arriving from the same distribution.
One-pass learning is also closely related to a classical problem studied in the context of control/estimation theory. More specifically, a classical algorithm known as recursive least-squares (RLS) (see, e.g., [17]) tackles one-pass learning for linear models (as elaborated in Section II-B). However, there are two limitations of the standard RLS: (i) it suffers from high computational and memory costs, and (ii) it is not well-suited for the overparameterized setting where zero training loss is desired. The main focus of this work is to develop a method that overcomes these limitations. Related to the overparameterized setting, a few works have recently discussed utilizing RLS to train popular deep neural networks such as FNN, CNN, RNN, and LSTM [21, 20]. However, these works are empirical in nature and consider multi-pass learning with mini-batches. Moreover, the theoretical properties of RLS for one-pass learning are not studied in those works.
I-A Contributions
Our main contributions can be summarized as follows.
-
•
We develop Orthogonal Recursive Fitting (ORFit), an algorithm for one-pass learning in the overparameterized setting which fits new data on the fly while updating the parameters in a direction that causes the least change to the predictions on previous data. Our algorithm uses memory efficiently by exploiting the structure of the streaming data via incremental principal component analysis (IPCA) to extract the essential information for the update. (§III)
-
•
Through the proposed method, we establish an interesting connection between two different algorithms from adaptive filtering and machine learning, namely, the recursive least-squares (RLS) algorithm and the orthogonal gradient descent (OGD). We further theoretically characterize the behavior of the proposed method in the linear overparameterized setting. (§IV)
-
•
We demonstrate the practicality of our approach and corroborate our theoretical findings through various experiments. (§V)
-
•
We discuss further extensions to overparameterized nonlinear models, relevant for deep learning. (§VI)
I-B Connections to Related Notions
One-pass learning shares some similarities with other settings that deal with streaming data such as online learning and incremental learning. While these terms are often inconsistently defined in the literature, one may distinguish them based on whether we learn a single datapoint or a batch of data at a time [14]. Although both of these approaches learn from streaming data, they typically assume that data is arriving from the same distribution. Compared to these settings, one-pass learning requires explicit efforts to not alter the predictions on the previous data while learning new data. Thus, it aims to preserve the predictions even when the new data comes from a different distribution. Another related setting is continual learning where batches of data from different tasks are sequentially learned [6].
II Preliminaries
II-A One-Pass Learning
Let be a model that the agent is trying to fit, where is the input and is the parameter (weight) vector. Consider a sequentially arriving stream of data where and . In overparameterized models, we have (and often ). For a loss function , let and . Then, one-pass learning considers the setting where, given an initial parameter , it updates the parameter after the new data arrived without revisiting previous datapoints .
For concreteness, we first focus on a linear overparameterized model with . We then discuss the extension of the results to nonlinear models (e.g. overparameterized architectures in deep learning) in Section VI.
Before introducing our algorithm and the results, we briefly review two related algorithms capable of one-pass learning, proposed in two different literatures.
II-B Recursive Least-Squares
First, we briefly review recursive least-squares (RLS) from the control/estimation theory literature (refer to [17] for details). At every step , RLS aims to find a parameter vector that solves the following regularized least-squares problem:
(1) |
where for a positive-definite matrix and is an initial parameter estimate. Note that the system is underdetermined/overparameterized, and the regularization term in (1) is necessary for the solution to be uniquely defined. While there is a closed-form solution for (1) given by with and , computing the solution directly requires storing all the previous data as well as recomputing the inverse of the covariance matrix for every new datapoint. RLS bypasses this issue by computing the new solution of (1) recursively from and .
We elaborate the algorithm more formally for a general version of RLS called exponentially-weighted recursive least-squares (EW-RLS) [17]. Consider the following problem:
(2) |
with a forgetting factor . Note that this reduces to the problem of the vanilla RLS (1) when . The exact solution of it is recursively updated as follows:
(3) |
with and . Here, can be alternatively written as where .
II-C Orthogonal Gradient Descent
An algorithm called orthogonal gradient descent (OGD) [7] has been recently proposed in the context of machine learning for a different but related problem. More specifically, [7] considers a continual learning setting in which tasks arrive sequentially, and each task consists of a set of datapoints. Continual learning can be understood as a “batch” version of one-pass learning. At a high level, when the -th task arrives, OGD updates the parameter to fit new samples from in a way that causes minimal changes to the predictions for previous tasks . The gradient of the model on a datapoint with respect to the parameter, , is the direction in the parameter space that causes the most change to the prediction on that datapoint. Thus, moving orthogonal to this direction, locally, keeps the prediction unchanged, which is the main idea behind OGD. More formally, the update direction is computed via projecting the current gradient of the loss onto the subspace orthogonal to :
(4) |
where is an orthogonal basis for and . The orthogonal basis is incrementally updated through the Gram-Schmidt procedure.
III Orthogonal Recursive Fitting
In this section, we propose a one-pass learning algorithm called orthogonal recursive fitting (ORFit). The algorithm consists of three main components: (i) orthogonal update of the parameter motivated by OGD; (ii) interpolation (perfect fitting) of new data in a single step; and (iii) efficient use of memory via incremental summary. We describe the details of each component below, starting from (i) and (ii). See Fig. 1 for an illustration.
III-A Orthogonal Recursive Update
We start by considering OGD directly applied to the one-pass learning setting. By treating each task in the continual learning setting as consisting of a single datapoint, OGD will run multiple gradient descent steps on the single datapoint. While perfect fitting/interpolation is desired in often highly overparameterized models [5, 4], it will take many iterations for OGD to perfectly fit the datapoint. Instead, inspired by the recent trend in meta-learning, we consider a one-step learning scheme that only runs a single gradient step to interpolate the new datapoint. We first begin with the following result that serves as a building block for our algorithm design; see Appendix -A for a proof.
Lemma 1.
Consider a linear model , and let be the projection defined by (4) of any vector . Then, for any step size , the new parameter preserves the predictions on the previous datapoints, i.e., for all .
The main takeaway of Lemma 1 is that the predictions for the previous datapoints do not change when we update the model along the direction . Hence, we may choose so that the updated parameter can perfectly fit the new datapoint, say , as . Following this principle, a straightforward calculation yields the following update rule for :
(5) |
where , is the initial weight vector, and the optimal step size is chosen as
(6) |
For intuition, we note that the optimal step size (6) is typically small for highly overparameterized models; there are many parameter vectors in the vicinity of the current solution that perfectly fit the new datapoint [2, 13, 1].

Remark 1 (Computational overhead).
Remark 2 (Nonlinear models).
Another distinction between the update rule (5) and OGD lies in the update of the orthogonal basis : OGD utilizes “fresher” gradient at the updated parameter by . Although the two bases actually span the same subspace for linear models, it turns out that ORFit in (5) leads to a natural generalization for nonlinear models; see Section VI for details.
Although the update rule (5) does not access previous datapoints, it still requires storing the orthogonal basis , whose size grows linearly in the number of visited datapoints. This is not desirable in practice when one needs to train the model on a large dataset. We address this issue next.
III-B Incremental Summary of Memory
In this section, we overcome the aforementioned memory issue by utilizing the structure of the streaming dataset. The main idea is to summarize the orthogonal basis using an incremental principal component analysis (IPCA) algorithm, known as the sequential Karhunen–Loeve (SKL) algorithm proposed in [12]. IPCA is a memory-efficient variant of PCA that enables sequential update for streaming/large datasets. Let us formally describe how IPCA incrementally summarizes the orthogonal basis.
Consider the orthogonal basis obtained by (5). Let the singular value decomposition (SVD) of be . Here, the crucial information of the SVD is the left-singular vectors that forms an orthonormal basis for . This information can be used to come up with a rank- approximation of by using the components corresponding to the top singular values.
Now suppose that the orthogonal basis is augmented with a newly projected gradient as in (5). We want to efficiently update for the new basis. Letting , the new basis matrix can be represented as:
(7) | ||||
(8) |
where is the SVD of . Then, (8) is the SVD of the new basis matrix. Hence to update , one can directly use the information from the previous iteration, namely and . The important aspect here is that the update can be made without storing and without having to recompute the SVD of the new basis matrix. Finally, one can store only the top singular values in and their corresponding components in . By repeatedly applying this IPCA algorithm in addition to (5), we obtain orthogonal recursive fitting (ORFit). See Algorithm 1 for the detailed procedure.
The main advantage of ORFit is its computational/memory efficiency. By only storing the top components, we can reduce the memory size from to . Moverover, the additional computational overhead to perform IPCA as well as the total time complexity of ORFit at each step is . Hence, ORFit can reduce both the computation and the memory complexity by appropriately choosing .
IV Theoretical Results
In this section, we provide the theoretical properties of the proposed method. We begin by discussing a formal connection between the proposed method and RLS.
IV-A Connection to RLS
It turns out ORFit corresponds to an extreme case of the well-known RLS method, as formally described in the following result; see Appendix -B for a proof.
Proposition 2.
Consider a linear overparameterized () model. Let be the initialization and be the memory limit for ORFit. Then, at each iteration , the update rule of ORFit results in the same parameter vector as the EW-RLS update rule (3) does with , , and initialization . In this setting, in (3) is the projection matrix onto the subspace orthogonal to .
Remark 3.
Remark 4.
ORFit in (5) has time and memory complexities, compared to those of for the EW-RLS update rule, where typically in the overparameterized setting.
One notable aspect of ORFit is that it bridges the two seemingly distinct algorithms OGD and RLS through Proposition 2. The connection provides us new insights into understanding the behavior of our proposed method, as we discuss next.
IV-B Characterizing the Solution of ORFit
Before presenting our main result, we first provide some intuitions. To understand the behavior of ORFit, let us first recall that the EW-RLS update rule (3) is the solution to the optimization problem (2). In light of Proposition 2, one might be tempted to claim that ORFit in (5) solves (2) with and . However, (2) is not well-defined for .
Nevertheless, intuitively one can regard ORFit as solving (2) in the limit of . Then, for sufficiently small , the first term in the objective of (2) outweighs the second term, which suggests that, in the overparameterized case, the solution should enforce for all , while minimizing . In the following theorem, we formalize this intuition and characterize the solution of ORFit; see Appendix -C for a proof.
Theorem 3.
Consider a linear overparameterized () model. Let be the initialization and be the memory limit. Then, at each iteration , the parameter vector obtained by ORFit is the solution of the following optimization problem:
(9) | ||||
s.t. |
It is known that, for a linear overparameterized model, in the standard multi-pass learning setting over the dataset , as the number of iterations goes to infinity, the iterates of stochastic gradient descent (SGD) initialized at with a sufficiently small step size converge to the solution of problem (9) (see, e.g., Proposition 1 in [3]). Thus, ORFit with just an epoch of training finds the solution that SGD in the limit of infinite number of iterations converges to.
Corollary 4.
Consider a linear overparameterized () model. Let be the initialization and be the memory limit. The parameter vector obtained by ORFit at each iteration is equal to what SGD would converge to by iterating over the dataset with a sufficiently small step size in the limit of infinite number of iterations.
V Experiments
In this section, we demonstrate the effectiveness of our proposed methods in the one-pass learning setting and corroborate the theoretical results presented in Section IV. We performed experiments for linear models in the Rotated MNIST setup described in [18]. In this setup, the inputs are rotated MNIST images for digit ‘’, whose size is . Our goal is to estimate the rotated angles in . For the training dataset, the angles are uniformly sampled from , and to introduce distribution shift, we order the dataset so that an image rotated with a smaller angle arrives earlier.


V-A Learning with Restrictions on Memory Size
In this experiment, we demonstrate the effectiveness of our proposed method in the memory-restricted setting, in which 100 datapoints are sequentially learned while we are only allowed to store up to basis vectors. For comparison, we consider the following baselines: (1) Greedy scheme: outputs only the label of the most recently learned datapoint, regardless of the input, as an extreme case of forgetting. (2) One-Step SGD: employs the one-step learning scheme with the step size in (6) but with empty orthogonal basis. (3) ORFit-random: ORFit that keeps randomly chosen basis vectors after each iteration instead of performing IPCA. (4) ORFit-latest: ORFit that keeps the latest basis vectors after each iteration instead of performing IPCA.
We first compared the test errors of the proposed method and the baselines. The test errors are measured with the test dataset that consists of images of digit ‘’ in the MNIST test set of which each is rotated with a random angle in . As shown in Fig. 2(a), ORFit outperforms other baselines after a sufficient number of training steps. Notably, ORFit results in lower variances over the independent runs with different initialization.
Next, we compared the degrees of forgetting for the proposed method and the baselines by keeping track of the prediction errors of a training datapoint throughout the training. As shown in Fig. 2(b), ORFit successfully keeps the prediction error low throughout the training. This is in stark contrast with other methods for which the prediction error quickly increases as other datapoints are learned.
V-B Learning without Memory Restriction
In this experiment, we follow the setup in Section V-A except that this time, we do not impose any memory restrictions. For comparison, we run vanilla SGD with a fixed step size . Vanilla SGD makes multiple passes over the entire dataset for epochs.


Reported in Fig. 3(a) are the training and test errors of vanilla SGD, One-Step SGD, and ORFit. Note that both SGD and ORFit learn the training dataset perfectly with almost zero training error. Moreover, after the training is finished, SGD and ORFit achieve similar test errors, corroborating Theorem 3.
In addition, Fig. 3(b) shows the prediction error of the -th training datapoint throughout the training (analogous to Fig. 2(b)). Note that ORFit perfectly preserves the prediction error of the sample, while One-Step SGD quickly deteriorates it. This result demonstrates the effectiveness of the orthogonal update in ORFit for one-pass learning.
VI Extension to Deep Learning
In this section, we extend our discussion to nonlinear models under the neural tangent kernel (NTK) regime [9, 11]. The main idea behind the NTK regime is that when the width of the neural network is chosen large enough, the model is well-approximated by its first-order approximation around the initialization:
(10) |
In particular, Lee et al. [11] discussed sufficient conditions (in terms of the width of the network) for which this approximation is valid; see their Theorem 2.1 for details. Under the linearized regime, we consider expanding the model around the parameter learned at the previous step as
(11) |
Throughout, we denote by the RHS of (11).
A notable feature of ORFit is that it is applicable to any differentiable nonlinear model . This is in contrast to the EW-RLS algorithm (3), which is only applicable to linear models . Based on this observation, we employ the step size calculated in (6) to fit the new data for a nonlinear model under the NTK regime (11). More specifically, we consider an analog of the EW-RLS (2) for nonlinear models:
(12) |
given a forgetting factor and . Then similarly as in (3), one can write out the solution of (12) in a recursive manner, while treating as the streamed data at the -th step:
(13) |
with initialization and . We call this generalized update rule NTK-RLS. Following a similar argument as in Section IV, we obtain the following result; see Appendix -D for a proof.
Theorem 5.
Consider a (nonlinear) overparameterized model with . Let be the initialization and be the memory limit for ORFit. Then, at each iteration , the update rule of ORFit results in the same parameter vector as the NTK-RLS update rule (13) does with , , and initialization . Moreover, at each iteration , the parameter vector obtained by ORFit is the solution of the following optimization problem:
(14) | ||||
s.t. |
Note that, under the NTK regime, for an overparameterized model, we have , and the solution obtained by ORFit in one pass is the same as that of SGD in the standard multi-pass setting (as characterized in, e.g., [3]). Compared to NTK-RLS (13), ORFit in (5) greatly reduces both time and memory complexities from to . This is particularly important for , common to the overparameterized settings (for instance, in ResNet-18, commonly used for the CIFAR-10 dataset, consisting of 50K samples).
VII Conclusion
In this paper, we proposed an algorithm called Orthognoal Recursive Fitting (ORFit) to tackle one-pass learning. We discussed the connection between the proposed method and orthogonal gradient descent (OGD), a practical algorithm in continual learning, as well as the recursive least-squares (RLS), a well-known method from adaptive filtering. Through this connection, we explained the advantages of the proposed method and theoretically characterized its behavior. Our theoretical findings reveal that ORFit attains the same solution as SGD, with much lower time and memory complexities. We validated our method and its theoretical properties through several experiments and discussed its extensions to nonlinear settings, relevant for deep learning.
We conclude with several interesting future directions. First, although ORFit exhibits outstanding performance in the memory-limited setting, some forgetting is still happening. It is caused by the information loss from summarizing the orthogonal basis via IPCA. Theoretically characterizing how much ORFit forgets would be of great importance to understand the practicality of the algorithm. Moreover, one can also come up with other methods to summarize the memory such as matrix sketching and compare them with ORFit. Next, we remark that RLS is a special case of the Kalman filter applied on a static system. Based on this connection, another interesting avenue is to build on ORFit and devise efficient learning/estimation methods for dynamic systems. Lastly, exploring the practicality of ORFit in deep learning based on a more comprehensive set of experiments would be of great interest.
References
- [1] Z. Allen-Zhu, Y. Li, and Z. Song. A convergence theory for deep learning via over-parameterization. In International Conference on Machine Learning, pages 242–252. PMLR, 2019.
- [2] N. Azizan and B. Hassibi. Stochastic gradient/mirror descent: Minimax optimality and implicit regularization. In International Conference on Learning Representations, 2018.
- [3] N. Azizan, S. Lale, and B. Hassibi. Stochastic mirror descent on overparameterized nonlinear models. IEEE Transactions on Neural Networks and Learning Systems, 2021.
- [4] P. L. Bartlett, P. M. Long, G. Lugosi, and A. Tsigler. Benign overfitting in linear regression. Proceedings of the National Academy of Sciences, 117(48):30063–30070, 2020.
- [5] M. Belkin, D. Hsu, S. Ma, and S. Mandal. Reconciling modern machine-learning practice and the classical bias–variance trade-off. Proceedings of the National Academy of Sciences, 116(32):15849–15854, 2019.
- [6] M. Delange, R. Aljundi, M. Masana, S. Parisot, X. Jia, A. Leonardis, G. Slabaugh, and T. Tuytelaars. A continual learning survey: Defying forgetting in classification tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.
- [7] M. Farajtabar, N. Azizan, A. Mott, and A. Li. Orthogonal gradient descent for continual learning. In International Conference on Artificial Intelligence and Statistics, pages 3762–3773. PMLR, 2020.
- [8] H. Hu, A. Li, D. Calandriello, and D. Gorur. One pass ImageNet. arXiv preprint arXiv:2111.01956, 2021.
- [9] A. Jacot, F. Gabriel, and C. Hongler. Neural tangent kernel: Convergence and generalization in neural networks. Advances in neural information processing systems, 31, 2018.
- [10] R. Kemker, M. McClure, A. Abitino, T. Hayes, and C. Kanan. Measuring catastrophic forgetting in neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018.
- [11] J. Lee, L. Xiao, S. Schoenholz, Y. Bahri, R. Novak, J. Sohl-Dickstein, and J. Pennington. Wide neural networks of any depth evolve as linear models under gradient descent. Advances in neural information processing systems, 32, 2019.
- [12] A. Levey and M. Lindenbaum. Sequential Karhunen-Loeve basis extraction and its application to images. IEEE Transactions on Image Processing, 9(8):1371–1374, 2000.
- [13] Y. Li and Y. Liang. Learning overparameterized neural networks via stochastic gradient descent on structured data. Advances in Neural Information Processing Systems, 31, 2018.
- [14] D. Nallaperuma, R. Nawaratne, T. Bandaragoda, A. Adikari, S. Nguyen, T. Kempitiya, D. De Silva, D. Alahakoon, and D. Pothuhera. Online incremental machine learning platform for big data-driven smart traffic management. IEEE Transactions on Intelligent Transportation Systems, 20(12):4679–4690, 2019.
- [15] P. Rai, H. Daumé, and S. Venkatasubramanian. Streamed learning: one-pass svms. In Proceedings of the 21st International Joint Conference on Artificial Intelligence, pages 1211–1216, 2009.
- [16] D. Sahoo, Q. Pham, J. Lu, and S. C. Hoi. Online deep learning: learning deep neural networks on the fly. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, pages 2660–2666, 2018.
- [17] A. H. Sayed. Fundamentals of adaptive filtering. John Wiley & Sons, 2003.
- [18] A. Sharma, N. Azizan, and M. Pavone. Sketching curvature for efficient out-of-distribution detection for deep neural networks. In Uncertainty in Artificial Intelligence, pages 1958–1967. PMLR, 2021.
- [19] Y. Wu, Y. Chen, L. Wang, Y. Ye, Z. Liu, Y. Guo, and Y. Fu. Large scale incremental learning. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 374–382. IEEE, 2019.
- [20] T. Yu, C. Zhang, Y. Wang, M. Ma, and Q. Song. Recursive least squares for training and pruning convolutional neural networks. arXiv preprint arXiv:2201.04813, 2022.
- [21] C. Zhang, Q. Song, H. Zhou, Y. Ou, H. Deng, and L. T. Yang. Revisiting recursive least squares for training deep neural networks. arXiv preprint arXiv:2109.03220, 2021.
-A Proof of Lemma 1
Consider for any .
(15) |
Since the orthogonal basis spans , can be represented as . Then,
(16) | ||||
(17) | ||||
(18) | ||||
(19) |
as if . Thus, . ∎
-B Proof of Proposition 2
In the update rule of ORFit (5), the new basis vector can be represented as
(20) | ||||
(21) |
where . Similarly,
(22) |
where denotes the derivative of w.r.t. its second argument.
Putting (22) into the parameter update in (5),
(23) |
Also, note that is symmetric and satisfies , and it corresponds to the projection matrix to the subspace orthogonal to . With these properties, can be expressed in a recursive form as
(24) |
Then, (23) and (24) are equivalent to the EW-RLS update rule (3) with , , and initialization , while . ∎
-C Proof of Theorem 3
We prove (9) using KKT conditions. First, we transform the RHS into an equivalent convex problem:
(25) |
Let and . Then, the Lagrangian of (25) is then represented as
(26) |
Since (25) is a convex problem, any primal and dual variables of (26) satisfying KKT conditions are primal and dual optimal. The KKT conditions for a pair of primal and dual variables are
(27) |
Now, we show that from the ORFit update rule (5) satisfies (27) with some so that is the primal optimal solution of (25) which shows (9). From (5), . Then for each ,
(28) | ||||
(29) |
where . Such exists since . Then,
(30) |
By letting , satisfies the first KKT condition. Also, with the choice of (6) with Lemma 1, fits all the data , which implies the second KKT condition. Thus, is the solution of (25) so that (9) holds. ∎
-D Proof of Theorem 5
We first prove the connection between ORFit and NTK-RLS. As in the proof of Prop. 2, the new basis vector and the projected gradient of the loss in the update rule of ORFit (5) can be represented as
(31) | ||||
(32) |
where . Putting (32) into the parameter update in (5),
(33) |
Since is symmetric and satisfies , with (31), can be expressed in a recursive form as
(34) |
Then, (33) and (34) are equivalent to the NTK-RLS update rule (13) with , , and initialization , while .
We then prove (14) using KKT conditions as in Thm. 3. First, we transform the RHS into an equivalent convex problem:
(35) |
Let , , and . Then, the Lagrangian of (35) is then represented as
(36) |
Since (35) is a convex problem, any primal and dual variables of (36) satisfying KKT conditions are primal and dual optimal. The KKT conditions for a pair of primal and dual variables are
(37) |