Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Penalized FTRL with Time-Varying Constraints

  • Conference paper
  • First Online:
Machine Learning and Knowledge Discovery in Databases (ECML PKDD 2022)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13717))

Abstract

In this paper we extend the classical Follow-The-Regularized-Leader (FTRL) algorithm to encompass time-varying constraints, through adaptive penalization. We establish sufficient conditions for the proposed Penalized FTRL algorithm to achieve \(\mathcal O(\sqrt{t})\) regret and violation with respect to a strong benchmark \(\hat{X}^{max}_t\). Lacking prior knowledge of the constraints, this is probably the largest benchmark set that we can reasonably hope for. Our sufficient conditions are necessary in the sense that when they are violated there exist examples where \(\mathcal O(\sqrt{t})\) regret and violation is not achieved. Compared to the best existing primal-dual algorithms, Penalized FTRL substantially extends the class of problems for which \(\mathcal O(\sqrt{t})\) regret and violation performance is achievable.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    Note the subtle yet crucial difference w.r.t. non-Pen-FTRL update (1).

  2. 2.

    Recall that \(c\sum _{\tau =0}^t \frac{1}{\tau ^{1-c}} \approx c\int _0^t \frac{1}{\tau ^{1-c}} d\tau = t^{c}\) for \(0\le c\le 1\). Hence, with this choice \(E[n_{2,t}]\approx 0.1 t^{c}\) and \(E[p_{2,t}]\approx 0.1t^{c-1}\)..

References

  1. Anderson, D., Iosifidis, G., Leith, D.J.: Lazy lagrangians for optimistic learning With budget constraints. IEEE ACM Trans. Netw. (2023). IEEE

    Google Scholar 

  2. Chen, T., Ling, Q., Giannakis, G.B.: An online convex optimization approach to proactive network resource allocation. IEEE Trans. Signal Process. 65(24), 6350–6364 (2017)

    Article  MathSciNet  MATH  Google Scholar 

  3. Hazan, E.: Introduction to online convex optimization. Found. Trends Optim. 2, 157–325 (2016)

    Article  Google Scholar 

  4. Jenatton, R., Huang, J.C., Archambeau, C.: Adaptive algorithms for online convex optimization with long-term constraints. In: Proceedings of ICML, pp. 402–411 (2016)

    Google Scholar 

  5. Liakopoulos, N., Destounis, A., Paschos, G., Spyropoulos, T., Mertikopoulos, P.: Cautious regret minimization: online optimization with long-term budget constraints. In: Proceedings of ICML, pp. 3944–3952 (2019)

    Google Scholar 

  6. Mahdavi, M., Jin, R., Yang, T.: Trading regret for efficiency: online convex optimization with long term constraints. J. Mach. Learn. Res. 13(81), 2503–2528 (2012)

    MathSciNet  MATH  Google Scholar 

  7. Mannor, S., Tsitsiklis, J.N., Yu, J.Y.: Online learning with sample path constraints. J. Mach. Learn. Res. 10(20), 569–590 (2009)

    MathSciNet  MATH  Google Scholar 

  8. McMahan, H.B.: A survey of algorithms and analysis for adaptive online learning. J. Mach. Learn. Res. 18, 1–50 (2017)

    MathSciNet  MATH  Google Scholar 

  9. Shalev-Shwartz, S.: Online learning and online convex optimization. Found. Trends Optim. 4, 107–194 (2011)

    MATH  Google Scholar 

  10. Sun, W., Dey, D., Kapoor, A.: Safety-aware algorithms for adversarial contextual bandit. In: Proceedings of ICML, pp. 3280–3288 (2017)

    Google Scholar 

  11. Valls, V., Iosifidis, G., Leith, D., Tassiulas, L.: Online convex optimization with perturbed constraints: optimal rates against stronger benchmarks. In: Proceedings of AISTATS, pp. 2885–2895 (2020)

    Google Scholar 

  12. Zangwill, W.J.: Nonlinear programming via penalty functions. Manag. Sci. 13(5), 344–358 (1967)

    Article  MATH  Google Scholar 

  13. Yi, X., Li, X., Xie, L., Johansson, K.H.: Distributed online convex optimization with time-varying coupled inequality constraints. IEEE Trans. Signal Process. 68, 731–746 (2020)

    Article  MathSciNet  MATH  Google Scholar 

  14. Yu, H., Nelly, M., Wei, X.: Online convex optimization with stochastic constraints. In: Proceedings of NIPS (2017)

    Google Scholar 

  15. Zinkevich, M.: Online convex programming and generalized infinitesimal gradient ascent. In: Proceedings of ICML (2003)

    Google Scholar 

Download references

Acknowledgments

The authors acknowledge support from Science Foundation Ireland (SFI) under grant 16/IA/4610, and from the European Commission through Grant No. 101017109 (DAEMON).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Douglas J. Leith .

Editor information

Editors and Affiliations

Appendix A: Proofs

Appendix A: Proofs

1.1 A.1 Proof of Lemma 1

Proof

Firstly note that for feasible points \(x\in X\) we have that \(g^{(j)}(x)\le 0\), \(j=1,\cdots ,m\) and so \(F(x)=f(x)\). By definition \(f(x) \ge f^*=\inf _{x\in X} f(x)\) and so the stated result holds trivially for such points. Now consider an infeasible point \(w\notin X\). Let z be an interior point satisfying \(g^{(j)}(z)<0\), \(j=1,\cdots ,m\); by assumption such a point exists. Let \({\gamma }_0 = \frac{f^*-f(z)-1}{G}\). It is sufficient to show that \(F(w)> f^*\) for \({\gamma }\ge {\gamma }_0\) and \(G=\max _{j\in \{1,\cdots ,m\}}\{g^{(j)}(z)\}\).

Let \(v=\beta z + (1-\beta ) w\) be a point on the chord between points w and z, with \(\beta \in (0,1)\) and v on the boundary of X (that is \(g^{(j)}(v)\le 0\) for all \(j=1,\cdots ,m\) and \(g^{(j)}(v)=0\) for at least one \(j\in \{1,\cdots ,m\}\)). Such a point v exists since z lies in the interior of X and \(w\notin X\). Let \(A:=\{j:j\in \{1,\cdots ,m\},g^{(j)}(v)=0\}\) and \(t(x):=f(x)+{\gamma }\sum _{j\in A} g^{(j)}(x)\). Then \(t(v)=f(v)\ge f^*\). Also, by the convexity of \(g^{(j)}(\cdot )\) we have that for \(j\in A\) that \(g^{(j)}(v) = 0 \le \beta g^{(j)}(z) + (1-\beta ) g^{(j)}(w)\). Since \(g^{(j)}(z)<0\), it follows that \(g^{(j)}(w)>0\). Hence, \(\sum _{j\in A}g^{(j)}(w) = \sum _{j\in A}\max \{0,g^{(j)}(w)\} \le \sum _{j=1}^m\max \{0,g^{(j)}(w)\}\) and so \(t(w) \le F(w,{\gamma })\). Now, observe that \(t(z)= f(z)+{\gamma }\sum _{j\in A} g^{(j)}(z) \le f(z)+{\gamma }_0\sum _{j\in A} g^{(j)}(z)\) since \(g^{(j)}(z)<0\) and \({\gamma }\ge {\gamma }_0\). Hence,

$$\begin{aligned} t(z)&\le f(z)+(f^*-f(z)-1)\frac{\sum _{j\in A} g^{(j)}(z)}{G} \end{aligned}$$
(8)

Selecting G such that \(\frac{\sum _{j\in A} g^{(j)}(z)}{G}\ge 1\) then \( t(z) \le f^*-1 \le t(v) -1 \). So we have established that \(f^*\le t(v)\), \(t(z)\le t(v)-1\) and \(t(w) \le F(w)\). Finally, by the convexity of \(t(\cdot )\), \(t(v) \le \beta t(z) + (1-\beta ) t(w)\). Since \(t(z)\le t(v)-1\) it follows that \(t(v) \le \beta (t(v)-1) + (1-\beta ) t(w)\) i.e. \(t(v) \le -\frac{\beta }{1-\beta }+t(w)\). Therefore \(f^* \le -\frac{\beta }{1-\beta } + F(w)<F(w)\) as claimed.

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Leith, D.J., Iosifidis, G. (2023). Penalized FTRL with Time-Varying Constraints. In: Amini, MR., Canu, S., Fischer, A., Guns, T., Kralj Novak, P., Tsoumakas, G. (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2022. Lecture Notes in Computer Science(), vol 13717. Springer, Cham. https://doi.org/10.1007/978-3-031-26419-1_19

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-26419-1_19

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-26418-4

  • Online ISBN: 978-3-031-26419-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics