Numerische Mathematik: Efficient Algorithms For Solving The - Laplacian in Polynomial Time
Numerische Mathematik: Efficient Algorithms For Solving The - Laplacian in Polynomial Time
Numerische Mathematik: Efficient Algorithms For Solving The - Laplacian in Polynomial Time
https://doi.org/10.1007/s00211-020-01141-z Mathematik
Sébastien Loisel1
Received: 15 August 2018 / Revised: 19 March 2020 / Published online: 24 August 2020
© The Author(s) 2020
Abstract
The p-Laplacian is a nonlinear partial differential equation, parametrized by p ∈
[1, ∞]. We provide new numerical algorithms,
√ based on the barrier method, for solv-
ing the p-Laplacian numerically in O( n log n) Newton iterations for all p ∈ [1, ∞],
where n is the number of grid points. We confirm our estimates with numerical exper-
iments.
1 Introduction
p−2
∇ · (∇v2 ∇v) = f in and v = g on ∂, (1)
1/2
d
where w2 = j=1 |w j |
2 is the usual 2−norm on Rd . Prolonging g from ∂
to the interior and setting u = v − g, the variational form is
1, p 1 p
Find u ∈ W0 () such that J (u) = ∇(u + g)2 − f u is minimized.
p
(2)
A similar definition can be made in the case p = ∞ and will be discussed in Sect. 3.1.
For p = 1, the p-Laplacian is also known as Mean Curvature, and a solution with
f = 0 is known as a minimal surface [31]. The 1-Laplacian is related to a certain
“pusher-chooser” game [19] and compressed sensing [7]. The general p-Laplacian is
B Sébastien Loisel
S.Loisel@hw.ac.uk
123
370 S. Loisel
used for nonlinear Darcy flow [11], modelling sandpiles [2] and image processing [8].
We also mention the standard text of Heinonen et al. [16]; as well as the lecture notes
of Lindqvist [21].
One may discretize the variational form (2) using finite elements; we briefly outline
this procedure in Sect. 2.1 and refer to Barrett and Liu [3] for details.
One chooses
piecewise linear basis functions {φ j (x)} on and we let u h (x) = j u j φ j (x).
The energy J (u h ) can be approximated by quadrature; the quadrature is exact if the
elements are piecewise linear. This leads to a finite-dimensional energy functional
123
Efficient algorithms for solving the p-Laplacian 371
p p
g X p = ∇g2 d x.
Let > 0 be a tolerance. In exact arithmetic, the barrier method of Sect. 2.3, with barrier
(25), to minimize J (u) over u ∈ Vh , starting from (u, s) = (0, ŝ) given by (31), converges
in at most N ∗ iterations, where
N ∗ ≤ 14.4 ||h −d d! log h −1−17d R 5 (1 + g X p ) −1 + K ∗ . (4)
Here, the constant K ∗ = K ∗ (, ρ) depends on the domain and on the quasi-uniformity
parameter ρ of Th . At convergence, u satisfies
Computational complexity: When we vary 0 < √ h < 1 while freezing all other parame-
ters, the three estimates (4), (5) and (6) are O( n log n), where n is the number of grid
points in Th .
√
The iteration count O( n log n) also holds if is not frozen, provided that −1 grows
at most polynomially in n.
123
372 S. Loisel
123
Efficient algorithms for solving the p-Laplacian 373
cost of O(n 3.5 log n) FLOPS for our entire algorithms. This estimate can be improved
by using an O(n 2.373 ) FLOPS fast matrix inverse [20], making our overall algorithms
O(n 2.873 log n) FLOPS; we mention that this matrix inversion algorithm mostly of
theoretical interest since it is not practical for any reasonable value of n. We have
taken special care to preserve the sparsity
√ of this problem so that, if one assumes a
bandwidth of b (e.g. typically b = O( n) for d = 2 and b = O(n 2/3 ) for d = 3), one
obtains an O(b2 n) sparse matrix solve algorithm, resulting in O(n 2.5 log n) (d = 2)
or O(n 2.84 log n) (d = 3) FLOPS for our overall algorithms. In addition, we mention
many preconditioning opportunities [4,10,12,14,15,22–26,35]. Although solution by
preconditioning is possible, it is difficult to estimate the number of iterations a priori
since the diffusion coefficient of the stiffness matrix is difficult to estimate a priori;
in the best case (“optimal preconditioning”) where the elliptic solve at each Newton
iteration can be done in O(n) FLOPS, our algorithms are then O(n 1.5 log n) FLOPS.
Our paper is organized as follows. In Sect. 2, we give some preparatory material
on the p-Laplacian and the barrier method. In Sect. 3, we prove our main theorem for
1 ≤ p < ∞ and a separate theorem for the case p = ∞. In Sect. 4, we validate our
algorithms with numerical experiments. We end with some conclusions.
2 Preparatory material
Note that (7) is not a limit as p → ∞ of (2), e.g. because (2) uses the pth power of
· X p in its definition.
where we have used the triangle inequality for · 2 at (∗) and the convexity of
φ(z) = z p at (∗∗).
123
374 S. Loisel
We now prove strict convexity for the 1 < p < ∞ case. If we have equality at
(∗) then q(x) + s(x) and r (x) + s(x) are non-negative multiples of one another, i.e.
q + s = aw and r + s = bw where a(x), b(x) ≥ 0 and w(x) is vector-valued. Then
p
(∗∗) becomes ((ta + (1 − t)b)w2 ) p ≤ (ta p + (1 − t)b p )w2 . Note that
(ta + (1 − t)b) p < ta p + (1 − t)b p unless a = b so the inequality (**) is strict unless
1, p
∇u = ∇v almost everywhere. Since u, v ∈ W0 can be identified by their gradients,
we have proven strict convexity.
max 0, 1p − q1
From the norm equivalence u p ≤ d uq for x ∈ Rd , one obtains
− max 0, 1p − 21 max 0, 21 − 1p
d |u|W 1, p ≤ u X p () ≤ d |u|W 1, p . (8)
123
Efficient algorithms for solving the p-Laplacian 375
1
J (u) ≥ (u X p − g X p ) p − f L q u L p
p
1 1
≥ u X p − g X p − f L q L p −1/ p u X p .
p p
p p
1
1 p 2 p L q ( p/2) 1− p q
J (u) − J (0) ≥ u X p − g X p − f L q . (15)
2p p q
p p 1 q
Hence, if u X p > 4g X p + 2L q ( p/2) 1− p ( p − 1) f L q , then J (u) − J (0) > 0
and hence a minimizing sequence must satisfy (14).
The p = 1 case is as follows:
if u X 1 > g X 1 /(1 − f L ∞ L). The p = ∞ case is done in a similar fashion.
The a priori estimate above can also be used to show the existence of a minimizer
of J (u).
1, p
Lemma 4 Let 1 < p < ∞. There is a unique u ∈ V ⊂ W0 () that minimizes J (u).
Proof Let α = inf v J (v). We now show how to produce a minimizing sequence {u k } ⊂
1, p 1, p
W0 (). For k = 1, 2, . . ., let Bk = {u ∈ W0 () | J (u) < α + 1/k and u X p <
p q
4g X p + 4L ( p − 1) f L q + 1}, see (14). Note that each Bk is open and nonempty
2
123
376 S. Loisel
squares or d-cubes, then each P (k) is of the form diag(±h, . . . , ±h), and ~P (k) ~2 =
~[P (k) ]−1 ~−1 2 = h. In general, if Th is not necessarily a uniform lattice, we say that
the family of triangulations Th , parametrized by 0 < h < 1, is quasi-uniform with
parameter ρ < ∞ if h ≤ ~P (k) ~2 , ~[P (k) ]−1 ~−1 2 ≤ ρh. Note that on the reference
simplex, the basis functions are φ̂i (x̂) = x̂i for i = 1, . . . , d and φ̂0 (x̂) = 1 − i x̂i .
√ √
As a result, ∇ φ̂2 ≤ d and, from the chain rule, ∇φi (x)2 ≤ h −1 d.
Let span{φk (x) | k = 1, . . . , n} ⊂ W1, p () be the finite m element(i)space of
piecewise linear elements over Th and let w(x) d x ≈ i=1 ωi w(x ) be the
midpoint quadrature rule, which is exact for piecewise linear or piecewise constant
functions. We can construct a “discrete derivative” matrix D ( j) whose (i, k) entry is
( j)
Di,k = ∂φ (i)
∂ x j (x ). Then,
k
⎛ ⎞p
m d 2 2
1 ωi ⎝ (D ( j) (u + g))i ⎠ ;
p
∇(u + g)2 =
p p
i=1 j=1
note that the quadrature is exact provided that g is also piecewise linear. For the
midpoint rule, ωi is the volume of the simplex K i ; if the triangulation Th is quasi-
uniform then we find that
hd ρd hd
≤ ωi ≤ ; (16)
d! d!
123
Efficient algorithms for solving the p-Laplacian 377
123
378 S. Loisel
where T V (u) denotes the usual total variation of u. Any monotonically nondecreasing
function u(x) with u(0) = 0 and u(1) = 1 will minimize J (u) and satisfy the boundary
conditions.
A minimizing sequence for J (u) is the piecewise linear functions u n (x) =
min(1, max(0, 0.5 + n(x − 0.5))). This sequence converges to the indicating function
of [0.5, 1), which is not in W 1,1 (0, 1). This is because W 1,1 (0, 1) is not reflexive
and hence its unit ball is not weakly compact. Instead, the limit of u n is in BV , the
double-dual of W 1,1 (0, 1).
We now briefly show why the minimization of J (u) for u ∈ Vh is numerically
challenging.
Example 2 Consider J (x) = |x| p where x ∈ R and 1 ≤ p < ∞; this corresponds to
a 1-dimensional discrete p-Laplacian with a single grid point. The Newton iteration
xk+1 = xk − J (xk )/J (xk ) is
Hence, the Newton iteration converges linearly for p ∈ (1.5, 2) ∪ (2, ∞) and diverges
for 1 < p ≤ 1.5. The Newton iteration is undefined for p = 1 since J = 0.
The p-Laplacian for p = 1 is particularly hard; we now show two types of difficulties.
First, the Hessian may be singular, and regularizing the Hessian leads to gradient
descent.
The Hessian matrix J (x) is singular which makes the Newton iteration undefined. To
make matters worse, the kernel of J is spanned by J and hence any “regularization”
J + I leads to a simple gradient descent.
Yet another difficulty is that the 1-Laplacian may have nonunique solutions or no
solutions when the forcing is nonzero.
Example 4 Let c ∈ R and J (x) = |x| + cx; this corresponds to a 1-dimensional
1-Laplacian with a nonzero forcing term, discretized with a single grid point. Then,
123
Efficient algorithms for solving the p-Laplacian 379
J (x) is convex for all c ∈ R. However, J (x) has a unique minimum x = 0 if and only
if |c| < 1. When |c| = 1, J (x) has infinitely many minima. When |c| > 1, J (x) is
unbounded below and there is no minimum.
As a result, the energy J (u) of the 1-Laplacian may not be bounded below when
the forcing f = 0; see also Lemma 3.
In this section, we briefly review the theory and algorithms of convex optimization and
refer to Nesterov [27, Section 4.2] for details, including the notion of self-concordant
barriers.
Let Q ⊂ Rn be a bounded closed convex set that is the closure of its interior, c ∈ Rn
be a vector and consider the convex optimization problem
The barrier method (or interior point method) for solving (17) is to minimize tc T x +
F(x) for increasing values of t → ∞, where the barrier function F(x) tends to ∞
when x → ∂ Q. The minimizer x ∗ (t), parametrized by t ≥ 0, of tc T x + F(x), is
called the central path, and x ∗ (t) forms a minimizing sequence1 for (17) as t → ∞.
Assume
we have a ν-self-concordant barrier F(x) for Q. Define the norm v∗x =
v [F (x)]−1 v. The main path-following scheme is
T
1. Set t0 = 0, β = 1/9 and γ = 5/36. Choose an accuracy > 0 and x (0) ∈ Q such
that F (x (0) )∗x (0) ≤ β.
2. The kth iteration (k ≥ 0) is
γ
tk+1 = tk + and x (k+1) = x (k) − [F (x (k) )]−1 (tk+1 c + F (x (k) )). (18)
c∗x (k)
√
(β+ ν)β
3. Stop if tk ≥ ν + 1−β −1 =: tol−1 .
The invariant of this algorithm is that, if tk c + F (x (k) )∗x (k) ≤ β then also
tk+1 c + F (x (k+1) )∗x (k+1) ≤ β. The stopping criterion guarantees that, at conver-
gence, c T x (k) − c∗ ≤ . Starting this iteration can be difficult, since it is not always
obvious how to find an initial point x (0) ∈ Q such that F (x (0) )∗x (0) ≤ β. Define the
analytic center x F∗ by F (x F∗ ) = 0. We use an auxiliary path-following scheme2 to
approximate the analytic center x F∗ of Q:
1. Choose x (0) ∈ Q and set t0 = 1 and G = −F (x (0) ).
2. For the kth iteration (k ≥ 0):
γ
tk+1 = tk − and (19)
G∗x (k)
123
380 S. Loisel
The invariant of the auxiliary scheme is that tk G + F (x (k) )∗x (k) ≤ β for every k. At
convergence, one can show that F (x̄)∗x̄ ≤ β. Let x̂ ∈ Q be some starting point for
the auxiliary path-following scheme. Combining the auxiliary path-following scheme
to find the approximate analytic center x̄ of Q, followed by the main path-following
scheme to solve the optimization problem (17) starting from x (0) = x̄, completes in
at most N iterations, where
√
N = 7.2 ν 2 log ν + log F (x̂)∗x ∗ + log x̂∗x ∗ + log(1/) . (21)
F F
The path-following schemes of Sect. 2.3 are so-called “short step”, meaning that the
barrier parameter t increases fairly slowly when ν is large. It is well-known that long-
step algorithms, where t increases more rapidly, often converge faster overall than
short-step algorithms, even√though the worst case estimate O(ν log ν) is worse than
the short-step estimate O( ν log ν), see Nesterov and Nemirovskii [28] for details.
The main path-following scheme can be made “long-step” as follows:
1. Assume x (0) ∈ Q such that F (x (0) )∗x (0) ≤ β and let t0 = 0.
2. Set
⎧ " #
⎨max κt , t + γ if tk c + F (x (k) )∗x (k) ≤ β,
tk+1 =
k k c∗ (k) (22)
x
⎩
tk otherwise;
x (k+1) = x (k) − rk [F (x (k) )]−1 (tk+1 c + F (x (k) )), (23)
where 0 < rk ≤ 1 is found by line search, see e.g. Boyd and Vandenberghe [5,
with α √
Algorithm 9.2 = 0.01 and β = 0.25].
3. Stop if tk ≥ ν + (β+1−β
ν)β
−1 = tol−1 .
The parameter κ ≥ 1 determines the step size of the scheme. In convex optimization,
step sizes κ = 10 or even κ = 100 are often used, but we will see in Sect. 4 that
shorter step sizes are better suited for the p-Laplacian.
The long-step variant of the auxiliary path-following scheme is implemented in a
similar fashion; the criterion for decreasing tk+1 is then tk G + F (x (k) ∗x (k) ≤ β.
123
Efficient algorithms for solving the p-Laplacian 381
else we say that x (k) was a slow step. Let κ0 be an initial step size (we will take
κ0 = 10.)
1. If x (k) is accepted after 2 or fewer slow steps, put κk+1 = min{κ0 , κk2 }.
√
2. If x (k) is accepted after 8 or more slow steps, put κk+1 = κk .
3. If x (k) is still not accepted after 15 slow steps, replace x (k+1) and tk+1 by the most
1/4
recently accepted step and put κk+1 = κk . We call this procedure a rejection.
4. Otherwise, put κk+1 = κk .
The quantity tk+1 is computed as in the long step algorithm (22), with κ = κk+1 . Note
that whenever tk+1 coincides with the short step (20), then the step is automatically
accepted. The rejection is “wasteful” in that it discards possibly useful information, but
we will see in the numerical experiments that this adaptive scheme is quite efficient in
practice. Furthermore, the rejection step is the key that unlocks a very simple analysis
of our algorithm.
Theorem 2 For given c, F, , let N S and N A be the number of Newton steps of the
short step and adaptive step algorithms, respectively. Then,
√
N A ≤ 160.76 + 0.73 log(1 + 9 ν)N S . (24)
κmin
$
%& '
5
tk+1 ≥ 1+ √ tk .
4 + 36 ν
Starting from κ = 10, after r rejections, the step size is κ = 10(1/4) . When κ ≤ κmin ,
r
the short step is automatically accepted and hence the maximum number of rejections
is r = r− , where
log(log(κmin )/ log(10))
10(1/4)
r−
= κmin ⇒ r− = − .
log 4
Hence,
√
r ≤ 0.76 + 0.73 log(1 + 9 ν).
Since all the adaptive steps are at least as large as the short steps and the stopping
criterion is purely based on the barrier parameter tk , and noting that each rejection
corresponds to 15 slow steps (plus the initial accepted step), we obtain the estimate for
the main phase. The estimate for the auxiliary phase is obtained in a similar fashion.
123
382 S. Loisel
Theorem 2 states that the adaptive algorithm cannot be much worse than the short
√
step algorithm, which means that the adaptive algorithm scales at worse like Õ( ν),
where we have neglected some logarithms. The reader may be surprised that the
estimate for the adaptive scheme is slightly worse than the estimate for the short step
scheme, but this is a well-known phenomenon in convex optimization. The long step
estimates are quite pessimistic and in practice, long step and adaptive schemes work
much better than the theoretical estimates. Our result is especially interesting because
it is well-known that estimates
√for long step algorithms scale like Õ(ν), whereas our
new algorithm scales like Õ( ν).
3 Proof of Theorem 1
The proof of Theorem 1 is rather technical, so we begin by outlining the plan of our
proof. The idea is to estimate all the quantities in the bound (21) for the number N of
Newton iterations. The barrier parameter ν is estimated in Lemma 6. Some “uniform”
or “box” bounds are given for the central path in Lemma 7; these are an intermediate
step in converting as many estimates as possible into functions of h. Because (21)
depends on the Hessian F of the barrier, the lowest eigenvalue λmin of F is estimated
in Lemma 8. This bound itself depends on extremal singular values of the discrete
derivative matrices, which are estimated in Lemma 9, and these bounds are rephrased
in terms of h in Lemma 10. In Lemma 11, we establish the connection between the
number m of simplices and the grid parameter h, which is used in Lemma 12 to
estimate the quantities x̂2 and F (x̂)2 , which can be converted to estimates for
x̂∗x ∗ and F (x̂)∗x ∗ in (21) by dividing by λmin ; here x̂ is a starting point for the
F F
barrier method. Finally, the quantities R appearing in Theorem 1 are obtained by
starting from the estimates of Lemma 3, adding 1, and doubling them. This ensures
that the central path will be well inside the ball of radius R.
In the present section, we treat in detail the case 1 ≤ p < ∞. The case p = 1,
which is considered especially difficult, poses no special difficulty in the present
section, provided that the hypotheses of Lemma 3 are satisfied. The case p = ∞ is
deferred to Sect. 3.1.
Let 1 ≤ p < ∞ and define the barrier
F(u, s) = F p (u, s) = − log z i − σ log si − log τi where (25)
i i i
y ( j)
d $ %& '
[( D ( j) u + D ( j) g)i ]2 ,
2/ p
z i = si − τi = R − ωi si and (26)
j=1
(
2 if 1 ≤ p < 2 and
σ = σ ( p) = (27)
1 if p ≥ 2.
123
Efficient algorithms for solving the p-Laplacian 383
Lemma 6 The function F(u, t) is an m(σ + 2)-self-concordant barrier for the set
p
Q = {(u, s) : si ≥ ∇(u + g)| K i 2 , si ≥ 0 and max ωi si ≤ R}, (28)
i
The problem of minimizing J (u) over u ∈ Vh subject to the additional constraint that
maxi ωi ∇(u + g)| K i 2 ≤ R is equivalent to:
−f
min c x subject to x ∈ Q where c =
T
. (29)
ω
Here, we have abused the notation and used the symbol f for the vector whose ith
component is f (x)φi (x) d x.
Proof The functions B p (x, s) = − log(s 2/ p − x T x) are σ + 1-self-concordant so
m ( j)
− i log τi + i=1 B p ([ k Di,k (u + g)k ]dj=1 , si ) is m(σ + 2) self-concordant, see
Nesterov and Nemirovskii [28]. The rest is proved by inspection.
From Lemma 3, it is tempting to use a bound such as u X p < R, i.e. i ωu si ≤ R,
but this leads to a dense Hessian Fss . Instead, we have used the “uniform” bound:
ωi si ≤ ωi si = s ≤ R.
i
With this “looser” bound, the Hessian Fss is sparse.3 Furthermore, by using the R
value from the a-priori estimate Lemma 3, one can ensure that Q is non-empty and
contains minimizing sequences for J (u). Thus, put:
⎛ ⎞p
2
d
R ≥ R = 2(1 + g X p ) = 2 + 2 ⎝ [(D g)i ] ⎠ .
∗ (
p j) 2
(30)
j=1
Set
⎛ ⎞p
2
d
R
ŝi = 1 + ⎝ ( j)
[(D g)i ] 2⎠
; hence ŝi ≤ . (31)
2
j=1
123
384 S. Loisel
2/ p
Proof From w T s ≥ 0 and (26), we find τ ≤ R. From (26), we find z i ≤ si and
from 0 ≤ τi = R − ωi s, we find ωi si ≤ R.
4 ( j) ( j) T −2 2/ p−1
d
Fus = − (Y D ) Z S , (37)
p
j=1
2 2 4
Fss = − − 1 Z −1 S 2/ p−2 + 2 Z −2 S 4/ p−2 + σ S −2 + W 2 Z −2 , (38)
p p p
S = diag(s), W = diag(ω), Y = diag(y), Z = diag(z). (39)
d (k) T (k) and assume 0 <
k=1 [D ] D
Lemma 8 Let dmin 2 be the smallest eigenvalue of
h < 1. Let ωmin = mini ωi , and similarly for (z ∗F )max , etc... The smallest eigenvalue
λmin of F (u ∗F , s F∗ ) is bounded below by
λmin ≥ min 2(z ∗F )−1 2 2 ∗ −2
max dmin , ωmin (z F )max . (40)
Proof We consider the “Rayleigh quotient” x T F x/xT x, the extremal values of which
v
are the extremal eigenvalues of F . We put x = so that
w
We use the Cauchy-Schwarz inequality together with Young’s inequality to find that
⎛ ⎞
4 d T ( j) ( j) T −1 −1 2/ p−1
2|v T Fus w| = ⎝ v (Y D ) Z ⎠ Z S w
p j=1
123
Efficient algorithms for solving the p-Laplacian 385
)
* d
8*
≤ + v T (Y ( j) D ( j) )T Z −2 (Y (r ) D (r ) )v w T S 2/ p−1 Z −2 S 2/ p−1 w
p
j,r =1
d
4 T 2/ p−1 −2 2/ p−1
≤4 v T (Y ( j) D ( j) )T Z −2 (Y (r ) D (r ) )v + w S Z S w.
p2
j=1,r
Hence we find:
d
( j) T −1 ( j) 2 2 −1 2/ p−2
x F x ≥2
T
v [DT
] Z D v+ −1 w − Z S
T
w
p p
j=1
+ w T σ S −2 w + w T W 2 Z −2 w.
d
2
x T F x ≥ 2 v T [D ( j) ]T Z −1 D ( j) v +
− 1 w T T −1 W S −1 w + w T W 2 Z −2 w
p
j=1
−1 2 −2
≥ x22 min 2z max dmin , ωmin
2
z max .
A domain is said to be of width L when ⊂ S, where S is a strip of width L. The
1, p
Friedrichs inequality states that, for domains of width L > 0 and for u ∈ W0 (),
u L 2 ≤ L|u| H 1 () .
Lemma 9 Let be a polytope of width L < ∞, and assume that the triangulation
Th , which depends on the grid parameter 0 < h < 1, is quasi-uniform. Then, there is
a constant c > 0, which depends on and
the(k)quasi-uniformity parameter ρ of Th ,
2 of
such that the smallest eigenvalue dmin k [D ]T D (k) satisfies
2
dmin ≥ c > 0. (41)
d (k) ]T W D (k)
Proof Consider the matrix A = k=1 [D and note that
(ρh)d ∗ (k) T (k)
∗ ∗ (k) T (k)
u Au ≤ ωmax u [D ] D u≤ u [D ] D u.
d!
k k
123
386 S. Loisel
λmin ≥ c R −4 h 6d . (42)
Note that R ≥ R ∗ ≥ 1 = ||−1 || ≥ ||−1 ωmin and hence R/ωmin ≥ ||−1 and
− 4p 2+ 4
λmin ≥ min{2c , 1} max{||, 1}2 R ωminp .
Since R ≥ 1 and ωmin ≤ 1 (because h ≤ 1), we can find a lower bound by putting
p = 1 in the exponents. Under the quasi-uniform hypothesis, all the quadrature weights
are bounded below by ωi ≥ h d /(d!), which yields (42).
Lemma 11 For 0 < h < 1, assume Th is a quasi-uniform triangulation of . The
number n of vertices of Th inside and the number m of simplices in Th satisfy
n
≤ m ≤ ||h −d d!, (43)
d +1
ρ of Th .
Proof From (31) we have
m 2
R
x̂22 = ŝi2 ≤ m . (45)
2ωmin
i=1
123
Efficient algorithms for solving the p-Laplacian 387
2 −1 2/ p−1 1
≤ [D ( j) ]T Ẑ −1 D ( j) g2 + ẑ min ŝ 2 + σ ŝ −1 2 + ω2 .
p τ̂min
j
(47)
−1
Here, δ 2j = ~D ( j) ~22 ≤ ωmin ρ([D ( j) ]T W D ( j) ), where ρ(·) is the spectral radius. We
estimate the spectral radius as follows:
w T [D ( j) ]T W D ( j) w = wx2 j d x ≤ |w|2H 1 ≤ C I2S h −2 w2L 2 ≤ C I2S [K
2 2d−2
] h w22 ,
where we have used the inverse Sobolev inequality |w| H 1 ≤ C I S h −1 w L 2 for w ∈ Vh
(see e.g. Toselli and Widlund [37, Lemma B.27]) and the norm equivalence w L 2 ≤
K h d w2 . Thus,
2
−1/2 √
δ2 ≤ ωmin dC I S [K ]h d−1 .
As a result,
−1/2 √ −1/2
[D ( j) ]T Ẑ −1 D ( j) g2 ≤ ωmin dC I S [K
]h d−1 ωmin m 1/2 g X p (53)
j
−1−0.5d
≤ C I S K h g X p . (54)
123
388 S. Loisel
Hence,
F (x̂)2 ≤ C I S K
−1−0.5d
h g X p
√ R √ 2√
+ m +2 m+ m(ρh)d /(d!).
2ωmin R
Also
−1
F (x̂)∗x ∗ ≤ λ−1
min F
( x̂)2 ≤ c
R −4 6d
h C ∗ −1−1.5d
h R(1 + g X p ) (59)
F
−1 ∗ −1−7.5d 5
= [c ] C h R (1 + g X p ). (60)
The estimates R p=1 , R1< p<∞ were obtained by starting from the estimates of
Lemma 3, adding 1, and doubling them. Substituting these into N ∗ produces N p=1
and N1< p<∞ .
123
Efficient algorithms for solving the p-Laplacian 389
We notice that this definition of Q coincides with the definition (28) with p = 1
subject to the additional restriction that s1 = . . . = sm and subsequently dropping the
index i from si . As a result, one can obtain a barrier for Q by taking the barrier (25)
with p = 1 on the subspace of constant valued s vectors, hence the barrier F∞ and its
derivatives are
F∞ (u, s) = F1 (u, se), F∞ (u, s) = E F1 (u, se), F∞
(u, s) = E F1 (u, se)E T , (66)
⎡ ⎤
1
⎢ ⎥
where e = ⎣ ... ⎦ , E =
I O
. (67)
O eT
1
The starting point for the optimization is (û, ŝ) with û = 0 and
⎛ ⎞1
2
d
ŝ = 1 + max ⎝ ( j)
[(D g)i ] 2⎠
. (68)
i
j=1
and assume p = ∞, L f L 1 < 1. The barrier method to solve (65) requires at most
N p=∞ Newton iterations, where
5
2g ∞ ()
≤ 14.4 ||h −d d! log h −1−6.5d 2 + −1 + K ∗ .
X
N p=∞
1 − L f L 1
(70)
123
390 S. Loisel
Proof The proof of Theorem 3 follows the same logic as that of Theorem 1, so we
merely sketch it here. First, (34) and (35) are replaced by:
( j)
Fu 2[D ( j) ]T y z
F = = j ωj , (71)
Fs −2s j z1j − mσ s + j τj
F Fus F F
F = uu = uu T
us
where (72)
Fsu Fss Fus Fss
d
d
Fuu = 2 [D ( j) ]T Z −1 D ( j) + 4 (Y ( j) D ( j) )T Z −2 (Y (r ) D (r ) ), (73)
j=1 j,r =1
d
Fus = −4 (Y ( j) D ( j) )T z −2 s, (74)
j=1
Fss = −2 z −1
j +4 z −2
j s + σ ms
2 −2
+ ω2j z −2
j . (75)
j j j
The proof of (40) holds (changing what must be changed), ending with
x T F x ≥ x22 min{2z max
−1 2
dmin , ωk2 z k−2 }, (76)
k
where we have used (16), 1 ≤ ŝ ≤ R/(2 min ωi ), τ̂i ≥ R/(2 min ωi ), and ẑ j ≥ 1, and
C is some constant that depends only on . Thus,
F (x̂)∗x ∗ ≤ λ−1
min F ( x̂)2 (78)
F
−1
≤ c R −4 h 6d −1−0.5d
C I S K h g X ∞ + C Rh −2d (79)
−1−3d g X ∞
≤ K h , (80)
1 + f X 1
123
Efficient algorithms for solving the p-Laplacian 391
√
N ∗ ≤ 7.2 4m 2 log(4m) (81)
−1−3d g X ∞
+ log K h (82)
1 + f X 1
−1 ∗ 5 −7.5d
+ log [c ] C R h (83)
+ log( −1 ) , (84)
as required.
4 Numerical experiments
123
392 S. Loisel
Fig. 1 Solving the p-Laplacian for p = 1, 2, ∞ with the same boundary conditions g and zero forcing
f = 0 on a 200 × 200 grid. Because of the zero forcing, the minimum and maximum principles hold, which
provides some protection against the near-discontinuities in the boundary data, e.g. when p = ∞
123
Table 1 Newton iteration counts for various problem sizes n, various step strategies κ and various values of p
p = 1.0
Short 985 2165 3163 4392 6086 10, 512 17, 192 — — — — —
κ=2 128 141 164 171 190 211 243 289 357 392 496 533
κ=3 88 98 136 118 148 165 194 237 343 535 825 —
κ=4 60 68 114 96 124 138 160 340 496 720 877 —
κ=7 68 67 136 99 252 255 384 464 833 1113 — —
Adaptive 70 76 181 106 163 169 216 270 275 315 405 449
p = 1.1
Short 954 1844 2677 3674 5094 8803 14, 510 — — — — —
Efficient algorithms for solving the p-Laplacian
κ=2 128 141 156 166 170 195 225 247 272 279 312 327
κ=3 89 96 103 113 116 130 157 185 211 229 244 274
κ=4 61 77 84 97 102 115 145 164 182 195 227 256
κ=7 68 69 74 91 88 104 143 172 190 237 314 391
Adaptive 60 81 68 97 80 103 190 186 243 210 231 274
p = 1.2
Short 938 1779 2626 3566 4991 8589 14, 099 — — — — —
κ=2 128 141 156 162 169 193 218 244 259 265 291 295
κ=3 87 102 107 110 116 126 138 162 176 189 199 210
κ=4 60 67 74 86 89 98 114 124 139 152 168 176
κ=7 60 71 72 80 83 97 110 118 129 136 140 170
Adaptive 54 57 64 70 74 84 105 113 127 198 147 204
123
393
Table 1 continued
394
123
p = 1.5
Short 922 1731 2579 3484 4896 8405 13, 754 — — — — —
κ=2 129 146 154 163 174 193 220 249 266 273 287 294
κ=3 86 102 106 111 116 132 142 161 175 186 194 201
κ=4 58 71 73 84 88 102 116 126 131 148 156 162
κ=7 59 64 77 80 83 98 108 117 120 128 126 139
Adaptive 52 58 66 69 72 85 96 111 110 120 123 130
p = 2.0
Short 829 1543 2305 3102 4356 7440 12, 123 22,082 — — — —
κ=2 135 151 158 165 174 190 220 256 274 287 296 309
κ=3 93 100 108 113 120 129 146 169 184 192 201 209
κ=4 51 64 70 71 84 89 99 109 119 129 139 142
κ=7 62 74 80 81 98 113 125 132 145 148 158 160
Adaptive 59 63 69 73 79 86 99 107 113 118 126 134
p = 3.0
Short 857 1624 2451 3329 4701 8127 13, 372 — — — — —
κ=2 142 161 172 184 195 208 249 299 315 331 348 363
κ=3 99 113 122 129 136 144 164 193 212 223 239 249
κ=4 56 65 74 79 87 102 110 128 136 157 163 163
κ=7 67 75 84 90 102 116 127 138 154 156 165 175
Adaptive 84 70 76 112 88 94 109 118 129 135 145 147
S. Loisel
Table 1 continued
p = 4.0
Short 907 1765 2697 3696 5260 9197 15, 265 — — — — —
κ=2 153 179 192 205 217 236 273 346 373 413 433 —
κ=3 105 122 132 144 153 167 189 224 243 289 — —
κ=4 57 75 82 84 97 115 127 142 158 219 186 212
κ=7 69 84 94 99 108 128 141 159 174 178 193 —
Adaptive 66 77 80 91 100 110 125 137 148 157 512 —
p = 5.0
Short 970 1927 2974 4102 5872 10, 352 17, 280 — — — — —
Efficient algorithms for solving the p-Laplacian
123
395
396 S. Loisel
Fig. 2 The number of Newton iterations for various grid sizes n and parameters p and step sizes κ
In Fig. 2, most of the black curves are approximately straight lines, indicating
O(n α ) scaling, but there are notable exceptions when p = 1 or p = ∞, especially
when κ is also large. By contrast, the adaptive step size algorithm (red lines), with
κ0 = 10, is seen to be the best algorithm in most cases, and these red lines are much
straighter than the black lines. We denote by N p (n) the number of iterations required
for a certain value of p and problem size n for the adaptive step size algorithm. We
have fitted straight lines to the red curves of Fig. 2 in the least-squares sense and
obtained the following approximations:
1
Thus, it seems like the adaptive scheme requires about O(n 4 ) Newton iterations,
regardless of the value of p.
Note that the case p = 2 is a linear Laplacian that can be computed by solving
a single linear problem. When we embed this linear problem into the machinery of
convex optimization, the overall algorithm is very inefficient since it may require
hundreds of linear solves. We are including this test case for completeness, not as a
recommendation.
123
Efficient algorithms for solving the p-Laplacian 397
Fig. 3 Solving the 1-Laplacian (top row) and ∞-Laplacian (bottom row) in 3d. The left column shows
the solutions on the whole volumetric domain with transparency, while the right column shows a slice
through of the same solutions with opaque colors
4.1 3d experiments
123
398 S. Loisel
function), an hence g X ∞ is very large and so is the solution u + g. The 1-Laplacian
is better able to tolerate the boundary data g with (near)-jumps because the trace of a
W 1,1 () function is merely L 1 (∂), thus allowing jumps.
The solution for the p = 1-Laplacian seems very close to what one would obtain
if one were to put f = 0 instead of f = 1. This is not surprising, because the 1-
Laplacian is a linear program and the solutions of linear programs change in discrete
steps when the forcing changes continuously. For example, the unique minimizer of
J˜(x) = |x| + f x (x ∈ R) is x = 0 whenever | f | < 1, and switches to “undefined”
(or ±∞) when | f | > 1 because then J˜ is unbounded below.
For the p = ∞-Laplacian, the solution u + g is a large positive bump because
f > 0 and there is a minimum principle stating that the minimum must of u + g be
on the boundary ∂. When one takes f < 0 instead, the solution u + g is a large
negative bump because in that scenario, u + g satisfies a maximum principle. In the
2d experiments, the ∞-Laplacian did not develop large bumps because the boundary
data was between 0 and 1 and the forcing was 0. This meant that u + g had to satisfy
both minimum and maximum principles, and u was constrained by 0 ≤ u + g ≤ 1,
preventing the formation of large bumps in the solution.
We have presented new algorithms for solving the p-Laplacian efficiently for any given
tolerance and for all 1 ≤ p ≤ ∞. We have proven that our √algorithms compute a solu-
tion to any given tolerance in polynomial time, using √ O( n log n) Newton iterations,
and an adaptive stepping variant converges in O( n log2 n) Newton iterations. We
have confirmed these scalings with numerical experiments. We have further shown by
numerical experiments that the adaptive step variant of the barrier method converges
much faster than the short-step variant for the p-Laplacian and also usually faster than
long-step barrier methods, thus achieving the practical speedup of long-step algo-
rithms while avoiding the O(n log n) worst-case behavior of long-step algorithms. We
1
have numerically estimated that the adaptive step algorithm requires O(n 4 ) Newton
iterations across all values of 1 ≤ p ≤ ∞. We have observed numerical difficulties
for p ≥ 5, which are expected since large powers exhaust the accuracy of double
precision floating point arithmetic; this difficulty is not specific to our algorithm but
is inherent to the p-Laplacian for large values of p. Our algorithms are particularly
attractive when p ≈ 1 and p = ∞, where there are no other algorithms that are
efficient at all tolerances.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which
permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give
appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence,
and indicate if changes were made. The images or other third party material in this article are included
in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If
material is not included in the article’s Creative Commons licence and your intended use is not permitted
by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the
copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
123
Efficient algorithms for solving the p-Laplacian 399
References
1. Andersen, K.D., Christiansen, E., Conn, A.R., Overton, M.L.: An efficient primal-dual interior-point
method for minimizing a sum of Euclidean norms. SIAM J. Sci. Comput. 22, 243–262 (1999)
2. Andreu, F., Mazón, J.M., Rossi, J.D., Toledo, J.: A nonlocal p-Laplacian evolution equation with
nonhomogeneous Dirichlet boundary conditions. SIAM J. Math. Anal. 40(5), 1815–1851 (2009)
3. Barrett, J.W., Liu, W.B.: Finite element approximation of the p-Laplacian. Math. Comput. 61(204),
523–537 (1993)
4. Berninger, H., Loisel, S., Sander, O.: The 2-Lagrange multiplier method applied to nonlinear transmis-
sion problems for the Richards equation in heterogeneous soil with cross points. SIAM J. Sci. Comput.
36(5), A2166–A2198 (2014)
5. Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, New York (2004)
6. Calamai, P.H., Conn, A.R.: A projected Newton method for l p norm location problems. Math. Program.
38(1), 75–109 (1987). https://doi.org/10.1007/BF02591853
7. Candès, E.J., Romberg, J., Tao, T.: Robust uncertainty principles: exact signal reconstruction from
highly incomplete frequency information. IEEE Trans. Inf. Theory 52(2), 489–509 (2006)
8. Catté, F., Lions, P.-L., Morel, J.-M., Coll, T.: Image selective smoothing and edge detection by nonlinear
diffusion. SIAM J. Numer. Anal. 29(1), 182–193 (1992)
9. Chen, R.-L., Zeng, J.-P.: A new subspace correction method for nonlinear unconstrained convex opti-
mization problems. Acta Mathematicae Applicatae Sinica 28(4), 745–756 (2012). https://doi.org/10.
1007/s10255-012-0185-z. (English Series)
10. Côté, J., Gander, M.J., Laayouni, L., Loisel, S.: Comparison of the Dirichlet-Neumann and optimal
Schwarz method on the sphere. In: Domain Decomposition Methods in Science and Engineering, pp.
235–242. Springer, New York (2005)
11. Drábek, P.: The p-Laplacian-mascot of nonlinear analysis. Acta Mathematica Universitatis Comeni-
anae 76(1), 85–98 (2007)
12. Drury, S.W., Loisel, S.: Sharp condition number estimates for the symmetric 2-Lagrange multiplier
method. In: Domain Decomposition Methods in Science and Engineering, pp. 255–261. Springer, New
York(2013)
13. Ekeland, I., Témam, R.: Convex Analysis and Variational Problems. SIAM, Philadelphia (1999)
14. Gander, M.J., Loisel, S., Szyld, D.B.: An optimal block iterative method and preconditioner for banded
matrices with applications to pdes on irregular domains. SIAM J. Matrix Anal. Appl. 33(2), 653–680
(2012)
15. Greer, N., Loisel, S.: The optimised Schwarz method and the two-Lagrange multiplier method for
heterogeneous problems in general domains with two general subdomains. Numer. Algorithms 69(4),
737–762 (2015)
16. Heinonen, J., Kilperläinen, T., Martio, O.: Nonlinear Potential Theory of Degenerate Elliptic Equations.
Dover Publications, Mineola (2006)
17. Hertog, D., Jarre, F., Roos, C., Terlaky, T.: A sufficient condition for self-concordance, with application
to some classes of structured convex programming problems. Math. Program. 69(1–3), 75–88 (1995).
https://doi.org/10.1007/BF01585553
18. Huang, Y.Q., Li, R., Liu, W.: Preconditioned descent algorithms for p-Laplacian. J. Sci. Comput. 32(2),
343–371 (2007). https://doi.org/10.1007/s10915-007-9134-z
19. Kohn, R., Serfaty, S.: A deterministic-control-based approach motion by curvature. Commun. Pure
Appl. Math. 59(3), 344–407 (2006). https://doi.org/10.1002/cpa.20101
20. Le Gall, F.: Powers of tensors and fast matrix multiplication. In: Proceedings of the 39th International
Symposium on Symbolic and Algebraic Computation, pp. 296–303. ACM, New York (2014)
21. Lindqvist, P.: Notes on the p-Laplace Equation. University of Jyväskylä, Finland (2017)
22. Loisel, S., Nguyen, H.: An optimal Schwarz preconditioner for a class of parallel adaptive finite
elements. J. Comput. Appl. Math. 321, 90–107 (2017)
23. Loisel, S., Szyld, D.B.: On the convergence of optimized Schwarz methods by way of matrix analysis.
In: Domain Decomposition Methods in Science and Engineering, vol. XVIII, pp. 363–370. Springer,
New York (2009a)
24. Loisel, S., Szyld, D.B.: A maximum principle for l 2 -trace norms with an application to optimized
Schwarz methods. In: Domain Decomposition Methods in Science and Engineering, vol. XVIII, pp.
193–200. Springer, New York (2009b)
123
400 S. Loisel
25. Loisel, S., Côté, J., Gander, M.J., Laayouni, L., Qaddouri, A.: Optimized domain decomposition
methods for the spherical Laplacian. SIAM J. Numer. Anal. 48(2), 524–551 (2010)
26. Loisel, S., Nguyen, H., Scheichl, R.: Optimized Schwarz and 2-Lagrange multiplier methods for
multiscale elliptic PDEs. SIAM J. Sci. Comput. 37(6), A2896–A2923 (2015)
27. Nesterov, Y.: Introductory Lectures on Convex Optimization: A Basic Course, vol. 87. Springer, New
York (2013)
28. Nesterov, Y., Nemirovskii, A.: Interior-point polynomial algorithms in convex programming. Soc. Ind.
Appl. Math. (1994). https://doi.org/10.1137/1.9781611970791
29. Nocedal, J., Wächter, A., Waltz, R.A.: Adaptive barrier update strategies for nonlinear interior methods.
SIAM J. Optim. 19(4), 1674–1693 (2009)
30. Oberman, A.M.: Finite difference methods for the infinity Laplace and p-Laplace equations. J. Comput.
Appl. Math. 254, 65–80 (2013)
31. Pinkall, U., Polthier, K.: Computing discrete minimal surfaces and their conjugates. Exp. Math. 2(1),
15–36 (1993). https://doi.org/10.1080/10586458.1993.10504266
32. Quarteroni, A., Valli, A.: Numerical Approximation of Partial Differential Equations, vol. 23. Springer,
New York (2008)
33. Rektorys, K.: Variational Methods in Mathematics, Science and Engineering. Springer, New York
(2012)
34. Ridgway Scott, L., Zhang, S.: Finite element interpolation of nonsmooth functions satisfying boundary
conditions. Math. Comput. 54(190), 483–493 (1990)
35. Subber, W., Loisel, S.: Schwarz preconditioners for stochastic elliptic PDES. Comput. Methods Appl.
Mech. Eng. 272, 34–57 (2014)
36. Tai, X.-C., Jinchao, X.: Global and uniform convergence of subspace correction methods for some
convex optimization problems. Math. Comput. 71, 105–124 (2001)
37. Toselli, A., Widlund, O.: Domain Decomposition Methods-Algorithms and Theory, vol. 34. Springer,
New York (2006)
38. Xue, G., Ye, Y.: An efficient algorithm for minimizing a sum of p-norms. SIAM J. Optim. 10, 551–579
(1997)
Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published maps
and institutional affiliations.
123