Abstract
In this paper, for solving large-scale nonlinear equations, we propose a nonlinear sampling Kaczmarz–Motzkin (NSKM) method. Based on the local tangential cone condition and the Jensen’s inequality, we prove convergence of our method with two different assumptions. Then, for solving nonlinear equations with the convex constraints, we present two variants of the NSKM method: the projected sampling Kaczmarz–Motzkin (PSKM) method and the accelerated projected sampling Kaczmarz–Motzkin (APSKM) method. With the use of the nonexpansive property of the projection and the convergence of the NSKM method, the convergence analysis is obtained. Numerical results show that the NSKM method with the sample of the suitable size outperforms the nonlinear randomized Kaczmarz method in terms of calculation times. The APSKM and PSKM methods are practical and promising for the constrained nonlinear problem.
Similar content being viewed by others
References
Bauschke HH, Borwein JM (1996) On projection algorithms for solving convex feasibility problems. SIAM Rev 38(3):367–426. https://doi.org/10.1137/S0036144593251710
Brinkhuis J (2020) Convex analysis for optimization: a unified approach, 1st edn. Springer Cham, Berlin. https://doi.org/10.1007/978-3-030-41804-5
Davis TA, Hu Y (2011) The University of Florida sparse matrix collection. ACM Trans Math Softw (TOMS) 38(1):1–25
De Loera JA, Haddock J, Needell D (2017) A sampling Kaczmarz–Motzkin algorithm for linear feasibility. SIAM J Sci Comput 39(5):S66–S87. https://doi.org/10.1137/16M1073807
Dennis JEJ, Moré JJ (1977) Quasi-Newton methods, motivation and theory. SIAM Rev 19(1):46–89. https://doi.org/10.1137/1019005
Haltmeier M, Leitao A, Scherzer O (2007) Kaczmarz methods for regularizing nonlinear ill-posed equations I: convergence analysis. Inverse Probl Imaging 1(2):289–298. https://doi.org/10.3934/ipi.2007.1.289
Jin B, Zhou Z, Zou J (2020) On the convergence of stochastic gradient descent for nonlinear ill-posed problems. SIAM J Optim 30(2):1421–1450. https://doi.org/10.48550/arXiv.1907.03132
Kelley CT (1995) Iterative Methods for Linear and Nonlinear Equations. Society for Industrial and Applied Mathematics, Philadelphia. https://doi.org/10.1002/cnm.1286
Li D, Fukushima M (1999) A globally and superlinearly convergent Gauss-Newton based BFGS method for symmetric nonlinear equations. SIAM J Numer Anal 37(1):152–172. https://doi.org/10.1137/S0036142998335704
Lukšan L, Matonoha C, Vlcek J (2018) Problems for nonlinear least squares and nonlinear equations. Tech. Rep. 1259, Institute of Computer Science, Academy of Sciences of the Czech Republic. https://doi.org/10.13140/RG.2.2.16112.10248
Qin T, Etesami SR (2020) A randomized algorithm for generalized accelerated projection method. IEEE Control Syst. Lett. 5(1):85–90. https://doi.org/10.1109/LCSYS.2020.3000714
Wang M, Bertsekas DP (2013) Incremental constraint projection-proximal methods for nonsmooth convex optimization. SIAM J Optimiz (to appear)
Wang Q, Li W, Bao W et al (2022) Nonlinear Kaczmarz algorithms and their convergence. J Comput Appl Math 399:113–720. https://doi.org/10.1016/j.cam.2021.113720
Yamashita N, Fukushima M (2001) On the rate of convergence of the Levenberg–Marquardt method. Topics in numerical analysis. Springer, Berlin, pp 239–249
Acknowledgements
This research was supported by the Fundamental Research Funds for the Central Universities (Grant No. 18CX02041A), the Shandong Provincial Natural Science Foundation (Grant No. ZR2020MD060), and the National Natural Science Foundation of China (Grant Nos. 42176011, 62231028).
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Andreas Fischer.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix
A Proof of Lemma 2
Proof
When \(k=0\), \(x_0\in \mathscr {B}_\rho (x_0)\) and \(|f_i(x_0)-f_i(x^*)-\nabla f_i(x_0)(x_0-x^*)|\le \eta _i|f_i(x_0)-f_i(x^*)|\) \((i=1,2,\ldots ,m) \), then we have
Since \(x^*\in \mathscr {B}_{\rho /2}(x_0)\) and (11), we have
Thus, \(x_1\in \mathscr {B}_\rho (x_0)\).
We assume that when \(k\le n\) \((n\in \mathbb {N})\), \(x_k\in \mathscr {B}_\rho (x_0)\) and (4) holds, then, for \(k=n+1\), similar to the derivation of \(k=0\), we have \(x_{n+1}\in \mathscr {B}_\rho (x_0)\) and (4) holds. \(\square \)
B Proof of Lemma 4
Proof
Since \(f: \mathscr {D}(f)\rightarrow \mathbb {R}\) is a convex function, we have that
By Taylor formula, it holds
Combining (12) and (13), we obtain that
Let \(\alpha \rightarrow 0\)
This completes the proof. \(\square \)
C Proof of Lemma 8
Proof
There are two cases to consider the following:
Case 1. \(x_{k+\frac{2}{4}}= x_{k+\frac{3}{4}}\). In this case, we have
Case 2. \(x_{k+\frac{2}{4}}\ne x_{k+\frac{3}{4}}\). By Lemma 7, we obtain that
Because \(x_{k+\frac{1}{4}}\ne x_{k+\frac{2}{4}}\), we get that
Besides, from \(x_{k+\frac{2}{4}}\ne x_{k+\frac{3}{4}}\), it can also be obtained that
Therefore, \(\Vert x_{k+\frac{3}{4}}-x^*\Vert _2^2<\Vert x_{k+\frac{2}{4}}-x^*\Vert _2^2 <\Vert x_{k+\frac{1}{4}}-x^*\Vert _2^2,\) which implies that \(x_{k+\frac{1}{4}}\ne x_{k+\frac{3}{4}}\).
This completes the proof. \(\square \)
D Proof of Lemma 9
Proof
We first show that \(\lambda _k\ge 1\). Observe that
where the last inequality follows from Lemma 7.
Hence \(\left\langle x_{k-\frac{3}{5}}-x_{k-\frac{1}{5}},x_{k-\frac{3}{5}}-x_{k-\frac{2}{5}}\right\rangle \le \Vert x_{k-\frac{2}{5}}-x_{k-\frac{3}{5}}\Vert ^2\), which implies that
Next, we will prove that \(x_{k}-x_{k-\frac{2}{5}}\) and \(x_{k-\frac{3}{5}}-x_{k-\frac{2}{5}}\) are orthogonal
Finally, we utilize (14) and (15) to prove (10).
For every \(x\in C_{\alpha _1^k} \cap C_{\alpha _2^k}\), we have
By writing \(\langle x_k-x_{k-\frac{1}{5}}, x_{k-\frac{1}{5}}-x \rangle =\langle x_k-x_{k-\frac{1}{5}}, x_{k-\frac{1}{5}}-x_k \rangle +\langle x_k-x_{k-\frac{1}{5}}, x_{k}-x \rangle \), we find that
By the definition of \(x_{k}\), we obtain that
For the first inner product of the above formula, we have
where the inequality comes from Lemma 7 and (15).
For the second inner product, we can obtain
where the first inequality follows Lemma 7, the second equality comes from the definition of \(x_k\), and the second inequality is from Lemma 7 and (14).
Thus
Since the iteration points \(x_i\) \((i=k-\frac{3}{5},k-\frac{2}{5},k-\frac{1}{5})\) are obtained by projecting on the closed convex sets, by Lemma 7, it results in
Thus
From (16) and (17), we get that
\(\square \)
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zhang, F., Bao, W., Li, W. et al. On sampling Kaczmarz–Motzkin methods for solving large-scale nonlinear systems. Comp. Appl. Math. 42, 126 (2023). https://doi.org/10.1007/s40314-023-02265-2
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s40314-023-02265-2
Keywords
- Large-scale nonlinear equations
- Finite convex constraints
- Sampling Kaczmarz–Motzkin method
- Projection method
- Randomized accelerated projection method