Abstract
In this paper we extend the classical Follow-The-Regularized-Leader (FTRL) algorithm to encompass time-varying constraints, through adaptive penalization. We establish sufficient conditions for the proposed Penalized FTRL algorithm to achieve \(\mathcal O(\sqrt{t})\) regret and violation with respect to a strong benchmark \(\hat{X}^{max}_t\). Lacking prior knowledge of the constraints, this is probably the largest benchmark set that we can reasonably hope for. Our sufficient conditions are necessary in the sense that when they are violated there exist examples where \(\mathcal O(\sqrt{t})\) regret and violation is not achieved. Compared to the best existing primal-dual algorithms, Penalized FTRL substantially extends the class of problems for which \(\mathcal O(\sqrt{t})\) regret and violation performance is achievable.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Note the subtle yet crucial difference w.r.t. non-Pen-FTRL update (1).
- 2.
Recall that \(c\sum _{\tau =0}^t \frac{1}{\tau ^{1-c}} \approx c\int _0^t \frac{1}{\tau ^{1-c}} d\tau = t^{c}\) for \(0\le c\le 1\). Hence, with this choice \(E[n_{2,t}]\approx 0.1 t^{c}\) and \(E[p_{2,t}]\approx 0.1t^{c-1}\)..
References
Anderson, D., Iosifidis, G., Leith, D.J.: Lazy lagrangians for optimistic learning With budget constraints. IEEE ACM Trans. Netw. (2023). IEEE
Chen, T., Ling, Q., Giannakis, G.B.: An online convex optimization approach to proactive network resource allocation. IEEE Trans. Signal Process. 65(24), 6350–6364 (2017)
Hazan, E.: Introduction to online convex optimization. Found. Trends Optim. 2, 157–325 (2016)
Jenatton, R., Huang, J.C., Archambeau, C.: Adaptive algorithms for online convex optimization with long-term constraints. In: Proceedings of ICML, pp. 402–411 (2016)
Liakopoulos, N., Destounis, A., Paschos, G., Spyropoulos, T., Mertikopoulos, P.: Cautious regret minimization: online optimization with long-term budget constraints. In: Proceedings of ICML, pp. 3944–3952 (2019)
Mahdavi, M., Jin, R., Yang, T.: Trading regret for efficiency: online convex optimization with long term constraints. J. Mach. Learn. Res. 13(81), 2503–2528 (2012)
Mannor, S., Tsitsiklis, J.N., Yu, J.Y.: Online learning with sample path constraints. J. Mach. Learn. Res. 10(20), 569–590 (2009)
McMahan, H.B.: A survey of algorithms and analysis for adaptive online learning. J. Mach. Learn. Res. 18, 1–50 (2017)
Shalev-Shwartz, S.: Online learning and online convex optimization. Found. Trends Optim. 4, 107–194 (2011)
Sun, W., Dey, D., Kapoor, A.: Safety-aware algorithms for adversarial contextual bandit. In: Proceedings of ICML, pp. 3280–3288 (2017)
Valls, V., Iosifidis, G., Leith, D., Tassiulas, L.: Online convex optimization with perturbed constraints: optimal rates against stronger benchmarks. In: Proceedings of AISTATS, pp. 2885–2895 (2020)
Zangwill, W.J.: Nonlinear programming via penalty functions. Manag. Sci. 13(5), 344–358 (1967)
Yi, X., Li, X., Xie, L., Johansson, K.H.: Distributed online convex optimization with time-varying coupled inequality constraints. IEEE Trans. Signal Process. 68, 731–746 (2020)
Yu, H., Nelly, M., Wei, X.: Online convex optimization with stochastic constraints. In: Proceedings of NIPS (2017)
Zinkevich, M.: Online convex programming and generalized infinitesimal gradient ascent. In: Proceedings of ICML (2003)
Acknowledgments
The authors acknowledge support from Science Foundation Ireland (SFI) under grant 16/IA/4610, and from the European Commission through Grant No. 101017109 (DAEMON).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix A: Proofs
Appendix A: Proofs
1.1 A.1 Proof of Lemma 1
Proof
Firstly note that for feasible points \(x\in X\) we have that \(g^{(j)}(x)\le 0\), \(j=1,\cdots ,m\) and so \(F(x)=f(x)\). By definition \(f(x) \ge f^*=\inf _{x\in X} f(x)\) and so the stated result holds trivially for such points. Now consider an infeasible point \(w\notin X\). Let z be an interior point satisfying \(g^{(j)}(z)<0\), \(j=1,\cdots ,m\); by assumption such a point exists. Let \({\gamma }_0 = \frac{f^*-f(z)-1}{G}\). It is sufficient to show that \(F(w)> f^*\) for \({\gamma }\ge {\gamma }_0\) and \(G=\max _{j\in \{1,\cdots ,m\}}\{g^{(j)}(z)\}\).
Let \(v=\beta z + (1-\beta ) w\) be a point on the chord between points w and z, with \(\beta \in (0,1)\) and v on the boundary of X (that is \(g^{(j)}(v)\le 0\) for all \(j=1,\cdots ,m\) and \(g^{(j)}(v)=0\) for at least one \(j\in \{1,\cdots ,m\}\)). Such a point v exists since z lies in the interior of X and \(w\notin X\). Let \(A:=\{j:j\in \{1,\cdots ,m\},g^{(j)}(v)=0\}\) and \(t(x):=f(x)+{\gamma }\sum _{j\in A} g^{(j)}(x)\). Then \(t(v)=f(v)\ge f^*\). Also, by the convexity of \(g^{(j)}(\cdot )\) we have that for \(j\in A\) that \(g^{(j)}(v) = 0 \le \beta g^{(j)}(z) + (1-\beta ) g^{(j)}(w)\). Since \(g^{(j)}(z)<0\), it follows that \(g^{(j)}(w)>0\). Hence, \(\sum _{j\in A}g^{(j)}(w) = \sum _{j\in A}\max \{0,g^{(j)}(w)\} \le \sum _{j=1}^m\max \{0,g^{(j)}(w)\}\) and so \(t(w) \le F(w,{\gamma })\). Now, observe that \(t(z)= f(z)+{\gamma }\sum _{j\in A} g^{(j)}(z) \le f(z)+{\gamma }_0\sum _{j\in A} g^{(j)}(z)\) since \(g^{(j)}(z)<0\) and \({\gamma }\ge {\gamma }_0\). Hence,
Selecting G such that \(\frac{\sum _{j\in A} g^{(j)}(z)}{G}\ge 1\) then \( t(z) \le f^*-1 \le t(v) -1 \). So we have established that \(f^*\le t(v)\), \(t(z)\le t(v)-1\) and \(t(w) \le F(w)\). Finally, by the convexity of \(t(\cdot )\), \(t(v) \le \beta t(z) + (1-\beta ) t(w)\). Since \(t(z)\le t(v)-1\) it follows that \(t(v) \le \beta (t(v)-1) + (1-\beta ) t(w)\) i.e. \(t(v) \le -\frac{\beta }{1-\beta }+t(w)\). Therefore \(f^* \le -\frac{\beta }{1-\beta } + F(w)<F(w)\) as claimed.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Leith, D.J., Iosifidis, G. (2023). Penalized FTRL with Time-Varying Constraints. In: Amini, MR., Canu, S., Fischer, A., Guns, T., Kralj Novak, P., Tsoumakas, G. (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2022. Lecture Notes in Computer Science(), vol 13717. Springer, Cham. https://doi.org/10.1007/978-3-031-26419-1_19
Download citation
DOI: https://doi.org/10.1007/978-3-031-26419-1_19
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-26418-4
Online ISBN: 978-3-031-26419-1
eBook Packages: Computer ScienceComputer Science (R0)