Convexity of Chance Constraints With Dependent Random Variables - The Use o Copulae - Henrion, Strugarek
Convexity of Chance Constraints With Dependent Random Variables - The Use o Copulae - Henrion, Strugarek
Convexity of Chance Constraints With Dependent Random Variables - The Use o Copulae - Henrion, Strugarek
Acknowledgement. The work of the first author was supported by the DFG Research
Center Matheon Mathematics for key technologies in Berlin
1 Introduction
Chance constraints or probabilistic constraints represent an important tool for mod-
eling restrictions on decision making in the presence of uncertainty. They take the
typical form
P(h(x, ξ) ≥ 0) ≥ p, (1)
where x ∈ Rn is a decision vector, ξ : Ω → Rm is an m-dimensional random vector
defined on some probability space (Ω, A, P), h : Rn × Rm → Rs is a vector-valued
mapping and p ∈ [0, 1] is some probability level. Accordingly, a decision vector x
(which is subject to optimization with respect to some given objective function) is
declared to be feasible, if the random inequality system h(x, ξ) ≥ 0 is satisfied at
least with probability p. A compilation of practical applications in which constraints
of the type (1) play a crucial role, may be found in the standard references [Pre95],
2 René Henrion and Cyrille Strugarek
[Pre03]. Not surprisingly, one of the most important theoretical questions related to
such constraints is that of convexity of the set of decisions x satisfying (1).
In this paper, we shall be interested in chance constraints with separated random
vectors, which come as a special case of (1) by putting h(x, ξ) = g(x) − ξ. More
precisely, we want to study convexity of a set of feasible decisions defined by
It is well-known from Prékopa’s classical work (see, e.g., [Pre95], Th. 10.2.1) that
this set is convex for all levels p provided that the law P ◦ ξ −1 of ξ is a log-concave
probability measure on Rm and that the components gi of g are concave. The latter
condition is satisfied, of course, in the important case of g being a linear mapping.
However, the concavity assumption may be too strong in many engineering appli-
cations, so the question arises, whether it can be relaxed to some weaker type of
(quasi-) concavity. To give an example, the function ex is not concave, while its log
is. It has been shown in an earlier paper that such relaxation is possible indeed,
but it comes at the price of a rather strong assumption of ξ having independent
components and of a slightly weaker result, in which convexity of M (p) cannot be
guaranteed for any level p but only for sufficiently large levels. We shall refer to
this property as eventual convexity. Note, however, that this is not a serious restric-
tion in chance constrained programming, because probability levels are chosen there
close to 1 anyway. To provide an example (see Example 4.1 in [HS08]), let the two
dimensional random vector ξ have a bivariate standard normal distribution with
independent components and assume that
Then, for p∗ ≈ 0.7 one has that M (p) is nonconvex for p < p∗ whereas it is convex
for p ≥ p∗ . Note, that though the multivariate normal distribution is log-concave
(see [Pre95]), the result by Prékopa mentioned above does not apply because the
gi are not concave. As a consequence, M (p) is not convex for all p as it would be
guaranteed by that result. Nonetheless, eventually convexity can be verified by the
tools developed in [HS08].
The aim of this paper is to go beyond the restrictive independence assumption
made in [HS08]. While to do this directly for arbitrary multivariate distributions of
ξ seems to be very difficult, we shall see that positive results can be obtained in case
that this distribution is modeled by means of a copula. Copulae allow to represent
dependencies in multivariate distributions in a more efficient way than correlation
does. They may provide a good approximation to the true underlying distribution
just on the basis of its one-dimensional margins. This offers a new perspective also
to modeling chance constraints not considered extensively so far to the best of our
knowledge. The paper is organized as follows: in a first section, some basics on
copulae are presented and the concepts of r-concave and r-decreasing functions
introduced. The following section contains our main result on eventual convexity
of chance constraints defined by copulae. In a further section, log-exp concavity of
copulae, a decisive property in the mentioned convexity result, is discussed. Finally,
Convexity of chance constraints: the use of copulae 3
a small numerical example illustrates the application of the convexity result in the
context of chance constrained programming.
2 Preliminaries
2.1 Basics on copulae
In this section, we compile some basic facts about copulae. We refer to the intro-
duction [Nel06] as a standard reference.
Definition 1. A copula is a distribution function C : [0, 1]m → [0, 1] of some ran-
dom vector whose marginals are uniformly distributed on [0, 1].
A theorem by Sklar states that for any distribution function F : Rm → [0, 1] with
marginals (Fi )1≤i≤m there exists a copula C for F satisfying
Moreover, if the marginals Fi are continuous, then the copula C is uniquely given
by
C(u) = F F1−1 (u1 ), . . . , Fm
−1
(um ) ; Fi−1 (t) := inf r.
(5)
Fi (r)≥t
Note that Sklar’s Theorem may be used in two directions: either in order to identify
the copula of a given distribution function or to define a distribution function via a
given copula (for instance a copula with desirable properties could be used to fit a
given distribution function). The following are first examples for copulae:
m
• Independent or Product Copula: C(u) = Πi=1 ui
• Maximum or Comonotone Copula: C(u) = min1≤i≤m ui
Gaussian or Normal Copula: C Σ (u) = ΦΣ Φ−1 (u1 ), . . . , Φ−1 (um )
•
The Gaussian copula has much practical importance (similar to the normal dis-
tribution function). It allows to approximate a multivariate distribution function F
having marginals Fi which are not necessarily normally distributed, by a composition
of a multivariate normal distribution function with a mapping having components
Φ−1 Fi defined via one-dimensional distribution functions ([KW97]).
A lot of other practically important copulae is collected in a whole family:
4 René Henrion and Cyrille Strugarek
ψ is called the generator of the Archimedean copula C. If limu→0 ψ(u) = +∞, then
C is called a strict Archimedean copula.
Conversely, we can start from a generator ψ to obtain a copula. The following
proposition provides sufficient conditions to get a copula from a generator:
Proposition 1. Let ψ : [0, 1] → R+ such that
1. ψ is strictly decreasing, ψ(1) = 0, limu→0 ψ(u) = +∞
2. ψ is convex
dk
3. (−1)k dt kψ
−1
(t) ≥ 0 ∀k = 0, 1, . . . , m, ∀t ∈ R+ .
Then, C(u) = ψ −1 ( m
P
i=1 ψ(ui )) is a copula.
Table 1. t∗r - values in the definition of r-decreasing densities for a set of common
distributions.
Law Density t∗r
√
2 µ+ µ2 +4rσ 2
normal √1
2πσ
exp − (t−µ)
2σ 2 2
r
exponential λ exp (−λt) (t > 0) λ
b−1 b+r−1 1/b
exp −atb
Weibull abt (t > 0) ab
a
Gamma b
Γ (a)
exp (−bt) ta−1 (t > 0) a+r−1
b
2 √
χ 1
2n/2−1 Γ (n/2)
tn−1 exp − t2 (t > 0) n + r − 1
χ2 1
tn/2−1 exp − 2t
2n/2 Γ (n/2)
(t > 0) n + 2r − 2
(log t−µ)2 2
1
log-normal √2πσt exp − 2σ2 (t > 0) eµ+(r−1)σ
2
t2
√
Maxwell √2t exp − 2σ 2 (t > 0) σ r+2
2πσ 3
2 q
2t
Rayleigh λ
exp − tλ (t > 0) r+1
2
λ
We shall need as an auxiliary result the following lemma whose proof is easy and
can be found in [HS08] (Lemma 3.1):
Lemma 1. Let F : R → [0, 1] be a distribution function with (r + 1)-decreasing
density f for some r > 0. Then, the function z 7→ F (z −1/r ) is concave on 0, (t∗ )−r ,
∗
where t refers to Definition 4. Moreover, F (t) < 1 for all t ∈ R.
3 Main Result
The purpose of our analysis is to investigate convexity of the set
6 René Henrion and Cyrille Strugarek
1. There exist ri > 0 such that the components gi are (−ri )-concave.
2. The marginal distribution functions Fi admit (ri + 1)-decreasing densities fi .
m
3. The copula C is log exp-concave on Πi=1 [Fi (t∗i ), 1), where the t∗i refer to Defi-
nition 4 in the context of fi being (ri + 1)-decreasing.
Then, M (p) in (9) is convex for all p > p∗ := max{Fi (t∗i )|1 ≤ i ≤ m}.
Proof. Let p > p∗ , λ ∈ [0, 1] and x, y ∈ M (p) be arbitrary. We have to show that
λx + (1 − λ)y ∈ M (p). We put
where the strict inequalities rely on the second statement of Lemma 1. In particular,
by (11), the inclusions x, y ∈ M (p) mean that
C(q1x , . . . , qm
x
) ≥ p, C(q1y , . . . , qm
y
) ≥ p. (12)
∗
Now, (11), (6), (12) and the definition of p entail that
Note that, for τ ∈ (0, 1), F̃i (τ ) is a real number. Having a density, by assumption
2., the Fi are continuous distribution functions. As a consequence, the quantile
functions F̃i (τ ) satisfy the implication
gi (x) ≥ F̃i (qix ) > t∗i > 0, gi (y) ≥ F̃i (qiy ) > t∗i > 0 (i = 1, . . . , m) (14)
t∗i
(note that > 0 by Definition 4). In particular, for all i = 1, . . . , m, it holds that
h i
min{F̃i−ri (qix ), F̃i−ri (qiy )}, max{F̃i−ri (qix ), F̃i−ri (qiy )} ⊆ 0, (t∗i )−ri .
(15)
(i = 1, . . . , m) .
Owing to assumption 2., Lemma 1 guarantees that the functions z 7→ Fi (z −1/ri )
∗ −ri
are concave on 0, (ti ) . In particular, these functions are log-concave on the
indicated interval, as this is a weaker property than concavity (see Section 2.2). By
virtue of (15) and (8), this allows to continue (17) as
h iλ h i1−λ
Fi (gi (λx + (1 − λ)y)) ≥ Fi F̃i (qix ) Fi F̃i (qiy ) (i = 1, . . . , m) .
Exploiting the fact that the Fi as continuous distribution functions satisfy the re-
lation Fi (F̃i (q)) = q for all q ∈ (0, 1), and recalling that qix , qiy ∈ (0, 1) by (13), we
may deduce that
Fi (gi (λx + (1 − λ)y)) ≥ [qix ]λ [qiy ]1−λ (i = 1, . . . , m) .
Since the copula C as a distribution function is increasing w.r.t. the partial order in
Rm , we obtain that
C (F1 (g1 (λx + (1 − λ)y)), . . . , Fm (gm (λx + (1 − λ)y))) ≥
C [q1x ]λ [q1y ]1−λ , . . . , [qm
x λ y 1−λ
] [qm ] (18)
Remark 1. The critical probability level p∗ beyond which convexity can be guaran-
teed in Theorem 1, is completely independent of the mapping g, it just depends on
the distribution functions Fi . In other words, for given distribution functions Fi , the
convexity of M (p) in (2) for p > p∗ can be guaranteed for a whole class of map-
pings g satisfying the first assumption of Theorem 1. Therefore, it should come at
no surprise that, for specific mappings g even smaller critical values p∗ may apply.
Proposition 2. The independent, the maximum and the Gumbel copulae are log
exp-concave on [q, 1)m for any q > 0.
log C (eu1 , . . . , eum ) = log min eui = min log eui = min ui .
1≤i≤m 1≤i≤m 1≤i≤m
m
Both functions are concave on (−∞, 0) , hence the assertion follows. For the Gumbel
θ
copula (see Example 1), the strict generator ψ (t) := (− log t) (for some θ ≥ 1)
implies that ψ −1 (s) = exp −s1/θ , whence, for ui ∈ (−∞, 0),
m
! m
!
u1 um −1
X ui −1
X θ
log C (e , . . . , e ) = log ψ ψ (e ) = log ψ (−ui )
i=1 i=1
!1/θ !1/θ
m
X m
X
θ θ
= log exp − (−ui ) =− |ui |
i=1 i=1
= − kukθ .
Since k·kθ is a norm for θ ≥ 1 and since a norm is convex, it follows that − k·kθ is
concave on (−∞, 0)m . Thus, the Gumbel copula too is log exp-concave on [q, 1)m
for any q > 0.
Contrary to the previous positive examples, we have the following negative case:
Qm
Example 2. The Clayton copula is not log exp-concave on any domain i=1 [qi , 1)
where q ∈ (0, 1)m . Indeed, taking the generator ψ (t) := θ−1 t−θ − 1 , we see that
Fig. 1. Plot of the function log C(ex , ey ) for the bivariate Gaussian copula C in case
of positive (left) and negative (right) correlation
log C (λu + (1 − λ) v) =
log C (exp (log (λu1 + (1 − λ) v1 )) , . . . , exp (log (λum + (1 − λ) vm ))) ≥
log C (exp (λ log u1 + (1 − λ) log v1 ) , . . . , exp (λ log um + (1 − λ) log vm )) ≥
λ log C (u1 , . . . , um ) + (1 − λ) log C (v1 , . . . , vm ) =
λ log C (u) + (1 − λ) log C (v) .
Qm
Hence, C is log concave on i=1 [qi , 1).
5 An example
A small numerical example shall illustrate the application of Theorem 1. Consider
a chance constraint of type (2):
−3/4 −1/4
P ξ1 ≤ x1 , ξ2 ≤ x2 ≥ p.
0.1 0.20
1.4
0.5
1.2
0.15
1.0
0.8
0.10
0.6
0.3
0.4 0.2 0.05 0.8
0.7
0.2 0.4 0.6
0.0 0.00 0.9
0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 0.0 0.2 0.4 0.6 0.8 1.0
Fig. 2. Illustration of the feasible set M (p) in (9) for different probability levels in
a numerical example. Convexity appears for levels of approximately 0.8 and larger
n √ o
p∗ = max F1 (7/3) , F2 7 .
Using easily accessible evaluation functions for the cumulative distribution functions
of the
√ exponential and Maxwell distribution, we calculate F1 (7/3) ≈ 0.903 and
F2 7 ≈ 0.928. Hence, M (p) in (9) is definitely convex for probability levels larger
than 0.928.
Figure 2 illustrates the resulting feasible set for different probability levels (fea-
sible points lie below the corresponding level lines). By visual analysis, convexity
holds true for levels p larger than approximately 0.8. Not surprisingly, our theoretical
result is more conservative. One reason for the gap is explained in Remark 1.
References
[HS08] Henrion, R., Strugarek, C.: Convexity of chance constraints with indepen-
dent random variables. Comput. Optim. Appl., 41, 263–276 (2008)
[KW97] Klaassen, C.A.J., Wellner, J.A.: Efficient Estimation in the Bivariate Nor-
mal Copula Model: Normal Margins are Least Favourable, Bernoulli, 3,
55–77 (1997)
[Nel06] Nelsen, R. B.: An Introduction to Copulas. Springer, New York (2006)
[Pre95] Prékopa, A.: Stochastic Programming. Kluwer, Dordrecht (1995)
[Pre03] Prékopa, A.: Probabilistic Programming. In: Ruszczynski, A. and Shapiro,
A. (eds.): Stochastic Programming. Hamdbooks in Operations Research
and Management Science, Vol. 10. Elsevier, Amsterdam (2003)