Online Forgetting Process for Linear Regression Models - Proceedings of Machine Learning Research
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Online Forgetting Process for Linear Regression Models Yuantong Li Chi-Hua Wang Guang Cheng Department of Statistics Department of Statistics Department of Statistics Purdue University Purdue University Purdue University Abstract Privacy Act Right (CCPA) (State of California Depart- ment of Justice, 2018), which also stipulates users to require companies and organizations such as Google, Motivated by the EU’s "Right To Be Forgot- Facebook, Twitter, etc to forget and delete these per- ten" regulation, we initiate a study of sta- sonal data to protect their privacy. Besides, users also tistical data deletion problems where users’ have the right to request the platform to delete his/her data are accessible only for a limited period obsolete data at any time or only to authorize the of time. This setting is formulated as an on- platform to hold his/her personal information such as line supervised learning task with constant photos, emails, etc, only for a limited period. Unfor- memory limit. We propose a deletion-aware tunately, given these data are typically incrementally algorithm FIFD-OLS for the low dimensional collected online, it is a disaster for the machine learn- case, and witness a catastrophic rank swinging ing model to forget these data in chronological order. phenomenon due to the data deletion oper- Such a challenge opens the needs to design and analyze ation, which leads to statistical inefficiency. deletion-aware online machine learning method. As a remedy, we propose the FIFD-Adaptive Ridge algorithm with a novel online regular- In this paper, we propose and investigate a class of ization scheme, that effectively offsets the un- online learning procedure, termed online forgetting pro- certainty from deletion. In theory, we pro- cess, to adapt users’ requests to delete their data before vide the cumulative regret upper bound for a specific time bar. To proceed with the discussion, both online forgetting algorithms. In the ex- we consider a special deletion practice, termed First periment, we showed FIFD-Adaptive Ridge In First Delete (FIFD), to address the scenario that outperforms the ridge regression algorithm the users only authorize their data for a limited period. with fixed regularization level, and hopefully (See Figure 1 for an illustration of the online forgetting sheds some light on more complex statistical process with constant memory limit s.) In FIFD, the models. agent is required to delete the oldest data as soon as receiving the latest data, to meet a constant memory limit. The FIFD deletion practice is inspired by the 1 Introduction situation that, the system may only use data from the past three months to train their machine learning model to offer service for new customers (Jon Porter, Today many internet companies and organizations are 2019; Google, 2020). The proposed online forgetting facing the situation that certain individual’s data can process is an online extension of recent works that no longer be used to train their models by legal require- consider offline data deletion (Izzo et al., 2020; Ginart ments. Such circumstance forces companies and organi- et al., 2019; Bourtoule et al., 2019) or detect data been zations to delete their database on the demand of users’ forgotten or not (Liu and Tsaftaris, 2020). willing to be forgotten. On the ground of the ‘Right to be Forgotten’, regularization is established in many In such a ‘machine forgetting’ setting, we aim to deter- countries and states’ laws, including the EU’s General mine its difference with standard statistical machine Data Protection Regulation (GDPR)(Council of Euro- learning methods via an online regression framework. pean Union, 2016), and the recent California Consumer To accommodate such limited authority, we provide solutions on designing deletion-aware online linear re- Proceedings of the 24th International Conference on Artifi- gression algorithms, and discuss the harm due to the cial Intelligence and Statistics (AISTATS) 2021, San Diego, "constant memory limit" setting. Such a setting is California, USA. PMLR: Volume 130. Copyright 2021 by challenging for a general online statistical learning task the author(s).
Online Forgetting Process for Linear Regression Models ✓b[1,s] ✓b[t s,t 1] x1 y1 x2 y2 …… xs ys xs+1 ys+1 …… xt s yt s …… xt 1 yt 1 xt yt time ✓b[2,s+1] Figure 1: Online Forgetting Process with constant memory limit s. since the "sample size" never grows to infinity but denotes the maximum value of vi , where i is in the stays a constant size along the whole learning process. indexes of negative coordinate. As an evil consequence, statistical efficiency is never For a positive semi-definite matrix 2 Rd⇥d ⌫0 , min ( ) improving as the time step grows due to the constant denotes the minimum eigenvalue of . We denote memory limit. as the generalized inverse of if it satisfies the Our Contribution. We first investigate the online condition = . The weighted 2-norm of vec- forgetting process in the ordinary linear regression. We tor x 2 Rd with respect p to positive definite matrix find a new phenomenon defined as Rank Swinging Phe- is defined by x = xT x . The inner product nomenon, which exists in the online forgetting process. is denoted by h·, ·i and the weighted inner-product If the deleted data can be fully represented by the is denoted by xT y = hx, yi . For any sequence data memory, then it will not introduce any regret. {xt }1t=0 , we denote matrix x[a,b] = {xa , xa+1 , . . . , xb } Otherwise, it will introduce extra regret to make this and x[a,b] 1 = supi,j |x[a,b] |(i,j) . For any matrix x[a,b] , online forgetting process task not online learnable. The Pb notation [a,b] = t=a xt x> t represents a gram matrix rank swinging phenomenon plays such a role that it with constant memory limit s = b a + 1 and notation indirectly represents the dissimilarity between the dele- Pb t=a xt xt + Id⇥d represents a gram matrix > ,[a,b] = tion data and the new data to affect the instantaneous with ridge hyperparameter . regret. Besides, if the gram matrix does not have full rank, it will cause the adaptive constant ⇣ to be un- stable and then the confidence ellipsoid will become 2 Statistical Data Deletion Problem wider. Taking both of these effects into consideration, the order of the FIFD-OLS’s regret will become linear At each time step t 2 [T ], where T is a finite time in time horizon T . horizon, the learner receives a context-response pair zt = (xt , yt ), where xt 2 Rd is a d-dimensional context The rank swinging phenomenon affects the regret scale and yt 2 R is the response. The observed sequence and destabilizes the FIFD-OLS algorithm. Thus, to of context {xt }t 1 are drawn i.i.d from a distribution remedy this problem, we propose the FIFD-Adaptive of PX with a bounded support X ⇢ Rd and xt2 L. Ridge to offset this phenomenon because when we add Let D[t s:t 1] = {zi }ti=t1 s denote the data collected at the regularization parameter, the gram matrix will have [t s, t 1] following the FIFD scheme. full rank. Different from using the fixed regularization parameter in the standard online learning model, we We assume that for all t 2 [T ], the response yt is a use the martingale technique to adaptively select the linear combination of context xt ; formally, regularization parameter over time. yt = hxt , ✓? i + ✏t , (1) Notation. Throughout this paper, we denote [T ] as the set {1, 2, . . . , T }. |S| denotes the number of el- where ✓? 2 Rd is the target parameter that sum- ements for any collection S. We use xp to denote marizes the relation between the context xt and re- the p-norm of a vector x 2 Rd and x1 = supi |xi |. sponse yt . The noise ✏t ’s are drawn independently from For any vector v 2 Rd , notation P(v) ⌘ {i|vi > 0} subgaussian distribution. That is, for every ↵ 2 R, denotes the indexes of positive coordinate of v and it is satisfied that E[exp(↵✏t )] exp(↵2 2 /2). N (v) ⌘ {i|vi < 0} denotes the indexes of negative Under the proposed FIFD scheme with a constant coordinate of v. Pmin (v) ⌘ min{vi |i 2 P(v)} denotes memory limit s, the algorithm A at time t only can the minimum value of vi , where i is in the indexes of keep the information of historical data from time step positive coordinate and Nmax (v) ⌘ max{vi |i 2 N (v)} t s to time step t 1 and then to make the prediction,
Yuantong Li, Chi-Hua Wang, Guang Cheng and previous data before time step t s 1 are not 3.1 FIFD-OLS Algorithm allowed to be kept and needs to be deleted or forgotten. The algorithm A is required to make total T s number The FIFD-OLS algorithm uses the least square es- of predictions in the time interval [s + 1, T ]. timator based on the constant data memory from time window [t s, t 1], defined as ✓ˆ[t s,t 1] = To be more precise, the algorithm A first receives a con- 1 ⇥ Pt 1 ⇤ text xt at time step t, and make a prediction based only [t s,t 1] i=t s yi xi . Then an incremental update ˆ formula for ✓ from time window [t s, t 1] to [t s+1, t] on the information in previous s time steps D[t s:t 1] (|D[t s:t 1] | = s) and hence the agent forgets the infor- is showed as follows. mation up to the time step t s 1. The predicted OLS incremental update: At time step t, the es- value ŷt computed by the algorithm A based on infor- timator ✓ˆ[t s+1,t] is updated by previous estimator mation D[t s:t 1] and current the context xt , which is ✓ˆ[t s,t 1] and new data zt , denoted by ŷt = A(xt , D[t s:t 1] ). (2) ✓ˆ[t s+1,t] =f ( 1 [t s,t 1] , xt s , xt ) After prediction, the algorithm A receives the true (4) · g(✓ˆ[t s,t 1] , [t s,t 1] , zt s , zt ), response yt and suffers a pseudo regret, r(ŷt , yt ) = (hxt , ✓? i ŷt )2 . We define the cumulative (pseudo)- where f ( 1 is defined as [t s,t 1] , xt s , xt ) regret of the algorithm A up to time horizon T as T X ( 1 [t s,t 1] ) (x> t s ( 1 [t s,t 1] ) 1) 1 RT,s (A) ⌘ (hxt , ✓? i 2 ŷt ) . (3) ⇥ 1 > 1 ⇤ ( [t s,t 1] )xt s xt s ( [t s,t 1] ) t=s+1 (5) 1 1 with ( [t s,t 1] ) ⌘ [t s,t 1] Our theoretical goal is to explore the relationship be- 1 ⇥ 1 1 ⇤ tween constant memory limit s and the cumulative re- (x> t [t s,t 1] xt + 1) 1 [t s,t > 1] xt xt [t s,t 1] gret RT,s (A). We proposed two algorithms FIFD-OLS and g(✓ˆ[t s,t 1] , [t s,t 1] , zt s , zt ) is defined as and FIFD-Adaptive Ridge and studied their cumula- ˆ [t s,t 1] ✓[t s,t 1] + yt xt yt s xt s . The detailed tive regret, respectively. In particular, we discusses online incremental update is in Appendix F. the effect of constant memory limit s, dimension d, The algorithm has two steps. First, f -step is to update subguassian parameter , and confidence level 1 on the inverse of gram matrix from [t 1 s,t 1] to [t 1 s+1,t] the obtained cumulative regret RT,s (A) for these two algorithms. For example, what’s the effect of constant and the Woodbury formula Woodbury (1950) is applied memory s on the order of the regret RT,s (A) if other to reduce the time complexity to O(s2 d) compared with parameters keep constant. In other words, how many directly computing the inverse of gram matrix; Sec- data do we need to keep in order to achieve the satis- ondly, g-step is a simple algebra and uses the definition fied performance under the FIFD scheme. Besides, is of adding and forgetting data to update the least square there any amazing or unexpected phenomenon both in estimator. The detailed update procedure of FIFD-OLS the experiment and theory occurred when the agent is displayed in Algorithm 1. used the ordinary least square method under the FIFD scheme and how to improve it? 3.2 Confidence Ellipsoid for FIFD-OLS Remark. The FIFD scheme can also be generalized Our first contribution is to obtain a confidence ellipsoid to the standard online learning paradigm when we add for the OLS method based on the sample set collected two and delete one data or add more and delete one under the FIFD scheme, showed in Lemma 1. data. These settings will automatically transfer to the Lemma 1. (FIFD-OLS Confidence Ellipsoid) For any standard online learning model when T becomes large. > 0, if the event min ( [t s,t 1] /s) > 2[t s,t 1] > 0 We presented the simulation result to illustrate it. holds, with probability at least 1 , for all t s + 1, ✓? lies in the set 3 FIFD OLS ⇢ C[t s,t 1] = ✓ 2 Rd : ✓ˆ[t s,t 1] ✓ In this section, we present the FIFD - ordinary least [t s,t 1] p (6) square regression (FIFD-OLS) algorithm under “First In First Delete” scheme and the corresponding confi- q[t s,t 1] (2d/s) log(2d/ ) dence ellipsoid for the FIFD-OLS estimator and regret analysis. Besides, the rank swinging phenomenon will where q[t s,t 1] = x[t s,t 2 1] 1 / [t s,t 1] is the adaptive be discussed in section 3.4. constant and we denote [t s,t 1] as the RHS bound.
Online Forgetting Process for Linear Regression Models Algorithm 1: FIFD-OLS Remarks. We develop the FIFD-OLS algorithm and prove an upper bound of the cumulative regret in or- Given parameters: s, T , 1 der of O( ⇣[(d/s)(log 2d/ )] 2 T ). The agent using this for t 2 {s + 1, ..., T } do Observe xt algorithm cannot improve the performance of this al- T gorithm because it has a constant memory limit s to [t s,t 1] = x[t s,t 1] x[t s,t 1] update the estimated parameters. What’s more, the if t = s+1 then adaptive constant q[t s,t 1] is unstable caused by the ✓ˆ[1,s] = [1,s] 1 xT[1,s] y[1,s] forgetting process. The main factor causing the oscil- else lation of gram matrix [t s,t 1] is that the minimum ✓ˆ[t s,t 1] = f ( [t 1 s 1,t 2] , xt s 1 , xt ) · eigenvalue is close to zero. In other words, the gram g(✓ˆ[t s 1,t 2] , [t s 1,t 2] , matrix is full rank at some time. Therefore, the adap- zt s 1 , zt 1 ) tive constant q[t s,t 1] will go to infinity at such time end points. So the generalization’s ability of FIFD-OLS is poor. Predict ŷt = h✓ˆ[t s,t 1] , xt i and observe yt Compute loss rt = |ŷt h✓? , xt i| Following the derivation of the regret upper bound of Delete data zt s = (xt s , yt s ) FIFD-OLS in Theorem 1, we find an interesting phe- end nomenon, which we called Rank Swinging Phenomenon (defined below Definition 1). This phenomenon will un- stabilize the FIFD-OLS algorithm and introduce some Proof. The key step to get the confidence ellipsoid is to extreme bad events, which results in a larger value of use the martingale noise technique presented in Bastani the regret upper bound of the FIFD-OLS algorithm. and Bayati (2020) to obtain an adaptive bound. The detailed proof can be obtained in Appendix A. Besides, 3.4 Rank Swinging Phenomenon the confidence ellipsoid in (6) requires the information of minimum eigenvalue of gram matrix [t s,t 1] and Below we give the definition of the Rank Swinging infinity norm of our observations with constant data Phenomenon. memory s. Thus, it is an adaptive confidence ellipsoid Definition 1. (Rank Swinging Phenomenon) At time following the change of data. t, when we delete data xt s and add data xt , the gram matrix switching from [t s,t 1] to [t s+1,t] will 3.3 FIFD OLS Regret cause its rank increasing or decreasing by 1 or 0 if Rank( [t s,t 1] ) d, The following theorem provides the regret upper bounds 8 for the FIFD-OLS algorithm. Theorem 1. (Regret Upper Bound of The FIFD-OLS Rank( [t s+1,t] ) = Rank( [t s,t 1] ), Case 3 > : Algorithm). Assume that for all t 2 [s + 1, T s] and Rank( [t s,t 1] ) 1, Case 4 Xt is i.i.d random variables with distribution PX . With (8) probability at least 1 2 [0, 1], for all T > s, s d, where four examples are illustrated in Appendix D. The we have an upper bound on the cumulative regret at real example can be found in Figure 2. time T : The rank swinging phenomenon results in unstable RT,s (AOLS ) regret, which is measured by the term called ‘Forgetting p 2 ⇣ (d/s) log(2d/ )(T s) (d log(sL2 /d) + (T s)), Regret Term’ (FRT) caused by deletion at time t. (7) Definition 2. (Forgetting Regret Term) where the adaptive constant ⇣ = max q[t s,t 1] . s+1tT FRT[t s,t 1] = Proof. We provide a roadmap for the proof of The- (9) xt s 2 1 (xt 2 1 sin2 (✓, [t 1 s,t 1] ) + 1), orem 1. The proof is motivated by (Abbasi-Yadkori [t s,t 1] [t s,t 1] et al., 2011; Bastani and Bayati, 2020). We first prove where FRT[t s,t 1] 2 [0, 2]. Lemma 1 to obtain a confidence ellipsoid holding for the FIFD-OLS estimator. Then we use the confidence Let’s use sin2 (✓, [t 1 s,t 1] ) denote the dissimilarity be- ellipsoid, to sum up the regret over time in Lemma 2 tween xt s and xt under [t 1 s,t 1] as follows, and find a key term called FRT (defined below equation 9), which affects the derivation of regret upper bound a sin2 (✓, [t 1 s,t 1] ) = sin2 1 ✓ lot. The detailed proof of this theorem can be found in [t s,t 1] the appendix C. = 1 cos2 (✓, [t 1 s,t 1] )
Yuantong Li, Chi-Hua Wang, Guang Cheng inverse of [t s,t 1] . Lemma 2. The cumulative regret of FIFD-OLS is partially determined by FRT at each time step, and the cumulative representation term is T X T X xt 2 1 2⌘OLS + FRT[t s,t 1] (10) [t s,t 1] t=s+1 t=s+1 where ⌘OLS = log(det( [T s,T ] )) is a constant based on data time window [T s, T ]. Proof. Let’s first denote rt as the instantaneous regret at time t and decompose the instantaneous p regret, rt = Figure 2: Rank switching phenomenon with memory h✓ˆ[t s,t 1] , xt i h✓? , xt i ( )x t 1 , [t s,t 1] limit s = 20, 60, 100, 150, 500 and d = 100 from left [t s,t 1] where the inequality is from Lemma 1. Thus, with to right. Xt ⇠ Unif(e1 , e2 , ...e100 ). The rank of gram probability at least 1 , for all T > s, matrix can decrease due to deletion operation. v u T u X hxt , xt si 1 RT,s (AOLS ) t(T s) rt2 1 [t s,t 1] cos(✓, [t s,t 1] ) = t=s+1 xt 1 xt s 1 v [t s,t 1] [t s,t 1] u T u X t(T s) max [t s,t 1] ( ) xt 2 1 . Remark. If FRT[t s,t 1] = 0, then it won’t introduce s+1tT t=s+1 [t s,t 1] extra regret to avoid it being online learnable which means the deleting data can be fully represented by the rest of data and the deletion operation won’t sab- 4 FIFD-Adaptive Ridge otage the representation power of the algorithm. If FRT[t s,t 1] 6= 0, it will introduce extra regret during To remedy the rank switching phenomenon, we take this online forgetting process, which means that the advantage of the ridge regression method to avoid this deletion data can’t be fully represented by the rest of happening in the online forgetting process. In this the data and the deletion operation will sabotage the section, we present the FIFD - adaptive ridge regres- representation power of the algorithm. sion (FIFD-Adaptive Ridge) algorithm under “First In First Delete” scheme and the corresponding confi- FRT is determined by two terms the ‘deleted dence ellipsoid for the FIFD-Adaptive Ridge estimator term’ xt s 2 1 and the ‘dissimilarity term’ and regret analysis. [t s,t 1] 1 xt 2 1 sin2 (✓, [t s,t 1] ). If this deleted term [t s,t 1] 4.1 FIFD-Adaptive Ridge Algorithm xt s 2 1 is zero, then FRT won’t introduce any [t s,t 1] extra regret, no matter how large the weight term The FIFD-Adaptive Ridge algorithm use the ridge xt 2 1 sin2 (✓, [t 1 s,t 1] ) is. The larger the dissimi- estimator ✓ˆ ,[t s,t 1] to predict the response. The defi- [t s,t 1] nition of it for time window [t s, t 1] is showed as larity term, the more regret it will introduce if the follows. deleted term is not zero since the new data intro- duces new direction in the representation space. If Adaptive Ridge update: FRT[t s,t 1] = 0, 8s < t T , we will achieve order p t 1 O( T ) cumulative regret, which is online learnable as ⇥ X ⇤ ✓ˆ ,[t s,t 1] = 1 yi x i . expected. However, this hardly happens. Therefore, ,[t s,t 1] i=t s the cumulative regret under the FIFD scheme is usually O(T ), which is proved in Theorem 1 and Theorem 2. Then we display the adaptive choice of hyperparameter In the following Lemma 2, we show the importance for time window [t s, t 1]. of FRT in getting the upper bound of Theorem 1 Lemma 3. (Adaptive Ridge Parameter [t s,t 1] ) If and Theorem 2. Here for the sake of simplicity, if the event min ( ,[t s,t 1] /s) > 2 ,[t s,t 1] holds, with Rank( [t s,t 1] ) < d, we will not discriminate the probability 1 , for any notation of generalized inverse [t s,t 1] and inverse 1 1 p [t s,t 1] , and uniformly use [t s,t 1] to represent the ( ) > x[a,b] 1 (2d/s) log(2d/ )/ 2 ,[a,b] ,
Online Forgetting Process for Linear Regression Models we h have a control of Li2 estimation error that If Assumption 1 holds, a high probability confidence Pr ✓ˆ ,[t s,t 1] ✓? 2 ( ) 1 , and to satisfy ellipsoid can be obtained for FIFD-Adaptive ridge. this condition, we select the adaptive [t s,t 1] as fol- Lemma 4. (FIFD-Adaptive Ridge Confidence Ellip- lows for the limited time window [t s, t 1], soid) For any 2 [0, 1], with probability at least 1 , p for all t s + 1, with Assumption 1 and the event [t s,t 1] x[t 1,t s] 1 2s log(2d/ )/✓?1 . (11) min ( ,[t s,t 1] /s) > 2 ,[t s,t 1] > 0 holds, then ✓? lies in the set The detailed derivation can be found in Appendix E. ⇢ C ,[t s,t 1] = ✓ 2 Rd : ✓ˆ ,[t s,t 1] ✓ The detailed update procedure of FIFD-Adaptive ,[t s,t 1] Ridge is displayed in Algorithm 2. p ⌫q ,[t s,t 1] d/2s , Algorithm 2: FIFD-Adaptive Ridge (13) Given parameters: s, T , where q ,[t s,t 1] = x[t 1,t s] 1 / 2 ,[t s,t 1] , = q for t 2 {s + 1, ..., T } do Observe xt log2 (6|P(✓? )|/ )/ log(2d/ ), and ⌫ = ✓?1 /Pmin (✓? ). x[t 1,t s] 1 = max x(i,j) 1is,1jd Proof. Key steps. The same technique used in Lemma ˆ[t s,t 1]= SD(y[t s,t 1] ) 1 is applied here to obtain an adaptive confidence ellip- [t s,t 1] = p soid for ridge estimator. Here we assume Assumption p 2sˆ[t s,t 1] x[t s,t 1] 1 log 2d/ 1 to make the confidence ellipsoid (4) more simplified. ,[t s,t 1] = The detailed proof can be obtained in the Appendix B. T x[t s,t 1] x[t s,t 1] + [t s,t 1] I Remark. The order of constant memory limit s in the ˆ 1 confidence ellipsoid for FIFD-OLS and ✓[t s,t 1] = xT y ,[t s,t 1] [t s,t 1] [t s,t 1] p FIFD-Adaptive Predict ŷt = h✓ˆ[t s,t 1] , xt i and observe yt ridge’s confidence ellipsoids are O( s). The adaptive Compute loss rt = |ŷt h✓? , xt i| ridge confidence ellipsoid requires the information of Delete data zt s = (xt s , yt s ) minimum eigenvalue of gram matrix ,[t s,t 1] with end hyperparameter and infinity norm of data x[t 1,t s] 1 , which is similar as FIFD-OLS confidence ellipsoid re- quired. The benefits of introduction of avoids the 4.2 Confidence Ellipsoid for FIFD-Adaptive singularity of gram matrix , which will stabilize the Ridge algorithm FIFD-Adaptive ridge and make the confi- dence ellipsoid narrow in most times. Besides, it is an Before we moving to the confidence ellipsoid of the adaptive confidence ellipsoid. adaptive ridge estimator, we define some notions about the true parameter ✓? , where Pmin (✓? ) represents the 4.3 FIFD Adaptive Ridge Regret weakest positive signal and Nmax (✓? ) serves as the weakest negative signal. To make the adaptive ridge In the following theorem, we provide the regret upper confidence ellipsoid more simplified, without loss of bound for the FIFD-Adaptive Ridge method. generality, we assume the following assumption. Theorem 2. (Regret Upper Bound of the FIFD- Assumption 1. (Weakest Positive to Strongest Signal Adaptive Ridge algorithm) With Assumption 1 and Ratio) We assume positive coordinate of ✓? dominates with probability at least 1 , the cumulative regret the bad events happening, satisfies: q Pmin (✓? ) RT,s (ARidge ) ⌫⇣ (d/s)(T s)[⌘Ridge + (T s)] WPSSR = ✓?1 (14) p q (12) x C3 + C3 + s2 log 2d log 12|P(✓? )| where ⇣ = max [t2 1,t s] 1 is the maximum adap- s+1tT ,[t s,t 1] 2d , tive constant, ⌘Ridge = d log(sL2 /d + [T s,T 1] ) s log log C2 ( ) is a constant related to the last data memory, QT s where C3 = log 6d log 2d . The WPSSR is monotone C2 ( ) = t=s+1 (1 + 2 ,[t s+1,t] ) ,[t s+1,t] ,[t s+1,t] increasing in s and P(✓? ), and is monotone decreasing is a constant close to 1, and ,[t s+1,t] = [t s+1,t] in d. In most cases, the LHS is greater than one, [t s,t 1] represents the change of over time steps. such as s = 100, d = 110, = 0.05, |P(✓? )| = 30, then WPS Ratio needs to be less than 1.02, which is satisfied Proof. The key step is to use the same technical lemma this assumption. used in the OLS setting to show a confidence ellipsoid
Yuantong Li, Chi-Hua Wang, Guang Cheng holds for the adaptive ridge estimator. Lemma 2 is used to compute FRT and then sums up FTRs to get the cumulative regret upper bound. The detailed proof of this theorem can be found in Appendix E. Remarks. (Relationship between regret upper bound and parameters) We develop the FIFD-Adaptive Ridge algorithm and then provide a regret upper bound in 1 order O( ⇣ L[d/s] 2 T ). This order represents the re- lationship between the regret, noise , dimension d, constant memory limit s, signal level ⌫, and confidence level 1 . Since the agent always keeps the constant memory limit s data to update the estimated param- eters, thus the agent can’t improve the performance of the algorithm. Therefore, the regret is O(T ) with respect to the decision number. Besides, we find that using the ridge estimator can offset the rank swinging Figure 3: Comparison of cumulative regret between of phenomenon since Ridge’s gram matrix is always full the Adaptive Ridge method and Fixed Ridge method. rank. The error bars represent the standard error of the mean regret over 100 runs. The blue line represents the Adaptive Ridge. The green line represents the 5 Simulation Experiments Fixed Ridge method with = {1, 2, 3} for each row. The red line represents the Fixed-Ridge method with We compare the FIFD-Adaptive Ridge method where = {10, 20, 30} for each row. The red line represents is chosen under Lemma 3 with the baselines with pre- the Fixed-Ridge method with = {100, 200, 300} for tuned . For all experiments unless otherwise specified, each row. we choose T = 3000, = 0.05, d = 100, L = 1 for all simulation settings. All results are averaged over 100 runs. we find that the cumulative regret of the adaptive ridge (blue line) decreases. Since we know that as Simulation settings. The setting for constant mem- the sample size increases, the regret will decrease in ory limit s is 20, 40, 60, 80 and subgaussian parame- theory. For each column, when we fix the memory ter is 1, 2, 3. Context {xt }t2[T ] is generated from limit s, as the noise level increases, the regret of N(0d , Id⇥d ), and then we normalize it. The response FIFD-Adaptive Ridge and Fixed Ridge increase with {yt }t2[T ] is generated by yt = hx, ✓? i + ✏t , where different hyperparameters. ✓? ⇠ N(0d , Id⇥d ) and then normalize it with ✓?2 = 1. i.i.d Overall, we find that the FIFD-Adaptive Ridge method We set ✏t ⇠ -subguassian. The adaptive ridge’s is more robust to the change of noise level when noise in the simulation is estimated by ˆ[t s,t 1] = we view figure 3 by row. Fixed Ridge methods with sd(y[t s,t 1] ), t 2 [s + 1, T ]. different hyperparameters such as green line and red Hyperparameter settings for competing meth- line are sensitive to the change of noise and memory ods. The hyperparameter we select for = 1 setting limit s because Fixed Ridge doesn’t have any prior is {1, 10, 100}. Since by the relationship of hyperparam- to adaptive the data. Although it performs well in eter and noise level , we know they are in linear order. the top left, it doesn’t work well in other settings. Thus, for the = 2 setting, we set = {2, 20, 200} The yellow line has relative comparable performance and for the = 3 setting, we set = {3, 30, 300}. The compared with the Adaptive Ridge method since its adaptive ridge hyperparameter is automatically calcu- hyperparameter is determined by the knowledge from lated according to Lemma 3 and we assume ✓?1 = 1 Adaptive Ridge. The FIFD-Adaptive Ridge is always since ✓?2 = 1. the best (2nd row and 3rd row) or close to the best (1st row) choice of among all of these settings, which Results. In figure 3, we present the simulation re- means that the Adaptive Ridge method is robust to sult of relationship between the cumulative regret the large noise. RT,s (ARidge ), noise level , hyperparameter , and constant memory limit s. We find all of the cumulative Besides, the Adaptive Ridge method can save compu- regret are linear in time horizon T just with different tational cost compared with the fixed Ridge method, constant levels. which needs cross-validation to select the optimal . From three rows, as the memory limit s increases, In addition, more results about the choice of hyperpa-
Online Forgetting Process for Linear Regression Models Figure 4: Here ✏t ⇠ tdf with degree of freedom df = Figure 5: Switching from Ridge to OLS & (+k, 1): {5, 10, 15}. This setting investigate the robustness of Comparison of cumulative regret between of the Switch- the proposed online regularization scheme. These lines’ ing Adaptive Ridge method when s 2d and Fixed colors are the same as they represent in figure 3. Ridge method. The error bars represent the standard error of the mean regret over 100 runs. The cyan line represents the Switching Adaptive Ridge. Other lines’ rameter and L2 estimation error of the estimator can colors are the same as they presented in Figure 3. be found in Appendix G. Heavy Tail Noise Distribution. Moreover, we also consider when the agent accumulates many data such test the robustness of these methods concerning dif- as n > 2d, where n represents the number of data the ferent noise. Here we assume ✏t ⇠ tdf , 8t 2 [T ]. The agent has at some time point t, and then transfer the degree of freedom df is set to be 5, 10, 15. As we algorithm from the Adaptive ridge method to the OLS know when df is greater than 30, it behaves like the method. This method called Switching Adaptive Ridge normal distribution. When df is small, t-distribution represented as the cyan line in figure 5. We see that has much heavier tails. Thus, in figure 4, we find that when the noise level is relatively small, it performs well. when df increases, the cumulative regret is decreasing However, when is large, it works worse than the green since the error is more like the normal distribution, line (fixed Ridge with prior knowledge about ) and which can be well captured by our algorithm. However, the blue line (Adaptive Ridge). fixed Ridge methods are not robust to the change of noise distribution, especially green line and red line. 6 Discussion and Conclusion Results for (+k, 1) addition-deletion operation. Here we test the pattern when we add more than one In this paper, we have proposed two online forgetting data point at each time step, such as k = 2, 3, 4, and algorithms under the FIFD scheme and provide the delete one data, denoted as (+k, 1) addition-deletion theoretical regret upper bound. To our knowledge, this pattern. In figure 5, the 1st, 2nd, 3rd column are cases is the first, theoretically well-proved work in the online that delete one data after receiving 2, 3, 4 data, respec- forgetting process under the FIFD scheme. tively. Each row has different noise level = 1, 2, 3. From the figure, we notice that the cumulative regret Besides, we find the existence of rank swinging phe- in each subplot has an increasing sublinear pattern. nomenon in the online least square regression and tackle For the first row, the noise level is the smallest, which it using the online adaptive ridge regression. In the shows the sublinear pattern quickly. For other rows, all future, we hope we can provide the lower bound for subplots show increasing sublinear patterns sooner or these two algorithms and design other online forgetting later, which satisfied our anticipation that when adding algorithms under the FIFD scheme. k > 1 data points and delete one data point at each time step, it finally will convert the normal online learn regime. However, the noise level will affect the time it References moves to the sublinear pattern. Abbasi-Yadkori, Y., Pál, D., and Szepesvári, C. (2011). Switching from Ridge to OLS. Besides, we also Improved algorithms for linear stochastic bandits. In
Yuantong Li, Chi-Hua Wang, Guang Cheng Advances in Neural Information Processing Systems, Liu, X. and Tsaftaris, S. A. (2020). Have you forgotten? pages 2312–2320. a method to assess if machine learning models have Bastani, H. and Bayati, M. (2020). Online decision forgotten data. arXiv preprint arXiv:2004.10129. making with high-dimensional covariates. Operations Liu, Y., Gadepalli, K., Norouzi, M., Dahl, G. E., Research, 68(1):276–294. Kohlberger, T., Boyko, A., Venugopalan, S., Timo- Bourtoule, L., Chandrasekaran, V., Choquette-Choo, feev, A., Nelson, P. Q., Corrado, G. S., et al. (2017). C., Jia, H., Travers, A., Zhang, B., Lie, D., and Detecting cancer metastases on gigapixel pathology Papernot, N. (2019). Machine unlearning. arXiv images. arXiv preprint arXiv:1703.02442. preprint arXiv:1912.03817. State of California Department of Justice (2018). Cal- ifornia consumer privacy act (ccpa). https://oag. Chen, M. X., Lee, B. N., Bansal, G., Cao, Y., Zhang, ca.gov/privacy/ccpa. S., Lu, J., Tsay, J., Wang, Y., Dai, A. M., Chen, Z., et al. (2019). Gmail smart compose: Real-time Villaronga, E. F., Kieseberg, P., and Li, T. (2018). assisted writing. In Proceedings of the 25th ACM Humans forget, machines remember: Artificial intel- SIGKDD International Conference on Knowledge ligence and the right to be forgotten. Computer Law Discovery & Data Mining, pages 2287–2295. & Security Review, 34(2):304–313. Council of European Union (2012). Council reg- Wainwright, M. J. (2019). High-dimensional statistics: ulation (eu) no 2012/0011. https://eur-lex. A non-asymptotic viewpoint, volume 48. Cambridge europa.eu/legal-content/EN/TXT/?uri=CELEX: University Press. 52012PC0011. Woodbury, M. (1950). Inverting modified matrices Council of European Union (2016). Council regulation (memorandum rept., 42, statistical research group). (eu) no 2016/678. https://eur-lex.europa.eu/ Zahavy, T. and Mannor, S. (2019). Deep neural eli/reg/2016/679/oj. linear bandits: Overcoming catastrophic forget- Dekel, O., Shalev-Shwartz, S., and Singer, Y. (2006). ting through likelihood matching. arXiv preprint The forgetron: A kernel-based perceptron on a fixed arXiv:1901.08612. budget. In Advances in neural information processing systems, pages 259–266. Dragomir, S. S. (2015). Reverses of schwarz inequality in inner product spaces with applications. Mathema- tische Nachrichten, 288(7):730–742. Gaillard, P., Gerchinovitz, S., Huard, M., and Stoltz, G. (2018). Uniform regret bounds over Rd for the sequential linear regression problem with the square loss. arXiv preprint arXiv:1805.11386. Ginart, A., Guan, M., Valiant, G., and Zou, J. Y. (2019). Making ai forget you: Data deletion in ma- chine learning. In Advances in Neural Information Processing Systems, pages 3513–3526. Google (2020). Clear browsing data. https: //support.google.com/chrome/answer/2392709? co=GENIE.Platform. Hsu, D., Kakade, S., Zhang, T., et al. (2012). A tail inequality for quadratic forms of subgaussian random vectors. Electronic Communications in Probability, 17. Izzo, Z., Smart, M. A., Chaudhuri, K., and Zou, J. (2020). Approximate data deletion from machine learning models: Algorithms and evaluations. arXiv preprint arXiv:2002.10077. Jon Porter (2019). Google will let you delete your lo- cation tracking data. https://www.theverge.com/ 2019/5/1/18525384/.
You can also read