Abstract
Uniform designs are widely used for experiments with mixtures. The uniformity of the design points is usually evaluated with a discrepancy criterion. In this paper, we propose a new criterion to measure the deviation between the design point distribution and a Dirichlet distribution. The support of the Dirichlet distribution, is defined by the set of d-dimensional vectors whose entries are real numbers in the interval [0,1] such that the sum of the coordinates is equal to 1. This support is suitable for mixture experiments. Depending on its parameters, the Dirichlet distribution allows symmetric or asymmetric, uniform or more concentrated point distribution. The difference between the empirical and the target distributions is evaluated with the Kullback–Leibler divergence. We use two methods to estimate the divergence: the plug-in estimate and the nearest-neighbor estimate. The resulting two criteria are used to build space-filling designs for mixture experiments. In the particular case of the flat Dirichlet distribution, both criteria lead to uniform designs. They are compared to existing uniformity criteria. The advantage of the new criteria is that they allow other distributions than uniformity and they are fast to compute.
Similar content being viewed by others
Notes
Good Lattice Point sets.
References
Borkowski JJ, Piepel GF (2009) Uniform designs for highly constrained mixture experiments. J Qual Technol. https://doi.org/10.1080/00224065.2009.11917758
Chuang SC, Hung YC (2010) Uniform design over general input domains with applications to target region estimation in computer experiments. Comput Stat Data Anal. https://doi.org/10.1016/j.csda.2009.08.008
Cornell JA (1981) Experiments with mixtures, designs, models, and the analysis of mixture data. Wiley, New York
Fang KT, Wang Y (1994) Number-theoretic methods in statistics. Chapman & Hall, London. https://doi.org/10.1007/978-1-4899-3095-8
Fang KT, Li R, Sudjianto A (2005) Design modeling for computer experiments. Chapman & Hall, London. https://doi.org/10.1201/9781420034899
Hickernell FJ (1998) A generalized discrepancy and quadrature error bound. Math Comput. https://doi.org/10.1090/S0025-5718-98-00894-1
Jin R, Chen W, Sudjianto A (2005) An efficient algorithm for constructing optimal design of computer experiments. J Stat Plan Inference. https://doi.org/10.1016/j.jspi.2004.02.014
Joe H (1989) Estimation of entropy and other functional of multivariate density. Ann Inst Stat Math. https://doi.org/10.1007/BF00057735
Jourdan A, Franco J (2009) Plans d’expériences numériques d’information de Kullback-Leibler minimale. J Soc Fr Stat 150(2):52–64
Jourdan A, Franco J (2010) Optimal Latin hypercube designs for the Kullback-Leibler criterion. Adv Stat Anal. https://doi.org/10.1007/s10182-010-0145-y
Kiefer J (1961) Optimum designs for regression model, II. Ann Math Stat. https://doi.org/10.1214/aoms/1177705160
Leonenko N, Pronzato L, Savani V (2008) A class of Rényi information estimators for multidimensional densities. Ann Stat. https://doi.org/10.1214/07-AOS539
Liu Y, Liu M (2016) Construction of uniform designs for mixture experiments with complex constraints. Commun Stat. https://doi.org/10.1080/03610926.2013.875576
Ning JH, Zhou YD, Fang KT (2011) Discrepancy for uniform design of experiments with mixtures. J Stat Plan Inference. https://doi.org/10.1016/j.jspi.2010.10.015
Prescott P (2008) Nearly uniform designs for mixture experiments. Commun Stat. https://doi.org/10.1080/03610920701824257
Pronzato L (2017) Minimax and maximin space-filling designs: some properties and methods for construction. J Soc Fr Stat 158(1):7–36
Scheffé H (1958) Experiments with mixtures. J R Stat Soc Ser B. https://doi.org/10.1111/j.2517-6161.1958.tb00299.x
Scott DW (1992) Multivariate density estimation: theory, practice and visualization. Wiley, New York. https://doi.org/10.1002/9780470316849
Wang Y, Fang KT (1990) Number theoretic methods in applied statistics (II). Chin Ann Math. https://doi.org/10.1142/9789812701190_0039
Wang Q, Kulkarni R, Verdu S (2006) A nearest-neighbor approach to estimating divergence between continuous random vectors. In: 2006 IEEE international symposium on information theory. https://doi.org/10.1109/ISIT.2006.261842
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendices
1.1 Appendix A. The proof of theorem 1
We apply the Jensen’s inequality to the expected value \({I}_{f}\left(g\right)=E\left[\text{log}(g({\varvec{X}})\right]\).
Let denote the function \(\varphi \left({\varvec{x}}\right)=\text{log}\left(g\left({\varvec{x}}\right)\right),\)
\(\varphi \) is a concave function since the logarithmic function is concave and \(\left({\alpha }_{k}-1\right)\) is positive for \({\alpha }_{k}\ge 1\). The Jensen’s inequality implies that, \(E\left[{\varphi }({\varvec{X}})\right]\le \varphi \left(E[{\varvec{X}}]\right)\). Let \(E\left[{\varvec{X}}\right]=\left({\mu }_{1},\dots ,{\mu }_{d}\right)\), then
since \(0<{\mu }_{k}<1\) (\(supp\left({X}_{k}\right)=\left[{0,1}\right]\) and we exclude the special case of a constant random variable equals to 0).
1.2 Appendix B. The proof of theorem 2
The prof of Theorem 2 is a direct application of a result demonstrated by Joe (1989). We just have to verify the assumptions.
The choice of a Gaussian kernel satisfies the conditions.
-
\(K(-z)=K(z)\)
-
The kernel is of the form \(K\left(z\right)=K\left({z}_{1},\dots ,{z}_{d}\right)=\prod_{j=1}^{d}{K}_{0}({z}_{j})\) where \({K}_{0}\) is a symmetric univariate density satisfying \(\int {u}^{2}{K}_{0}\left(u\right)du=1.\)
The \(d\) components of \({\varvec{X}}\) have approximately the same scale in [0,1], the logarithmic function is thrice differentiable. Moreover, we suppose that \(\int f\left({\varvec{x}}\right)log\left(f\left({\varvec{x}}\right)\right)d{\varvec{x}}\) and \(\int f\left({\varvec{x}}\right) log^{2} \left(f\left({\varvec{x}}\right)\right)d{\varvec{x}}\) exists. Hence all conditions are satisfied to apply the results demonstrated by Joe (1989).
We have already noted that the existence hypothesis of \(\int f\left({\varvec{x}}\right)log\left(f\left({\varvec{x}}\right)\right)d{\varvec{x}}\) is feasible since \(f\) is close to \(g\), and we proved the existence of this integral in Theorem 1. However, we have not demonstrated the existence of \(\int f\left({\varvec{x}}\right) log^{2} \left(f\left({\varvec{x}}\right)\right)d{\varvec{x}}\) when \(f=g\). This is demonstrated below in the case of \(d=2\) to simplify notation. It remains true in the general case.
The line \({x}_{1}+{x}_{2}=1\) has the parametric representation,
where \(t\in \left[-{1,0}\right].\) We define the mapping function\(M :\left[-{1,0}\right]\to \left[{0,1}\right]\times \left[{0,1}\right]\),\(M\left(t\right)=(-t,1+t)\). Then,
If \(f=g\),
With
\({I}_{1}<+\infty \) since \({\alpha }_{i}\ge 1\).
\({I}_{2}\) is an improper integral in 0, but \({\left(-t\right)}^{{\alpha }_{1}-1}{\left(1+t\right)}^{{\alpha }_{2}-1} log^{2} \left(-t\right)\sim {\left(-t\right)}^{{\alpha }_{1}-1} log^{2} \left(-t\right)\) when \(t\) tends to 0 and \({\int }_{-1}^{0}{\left(-t\right)}^{{\alpha }_{1}-1} log^{2} \left(-t\right)dt\) is a convergent Bertrand’s integral.
\({I}_{3}\) is an improper integral in -1, but \({\left(-t\right)}^{{\alpha }_{1}-1}{\left(1+t\right)}^{{\alpha }_{2}-1} log^{2} \left(1+t\right)\sim {\left(1+t\right)}^{{\alpha }_{2}-1} log^{2} \left(1+t\right)\) when \(t\) tends to -1 and \({\int }_{-1}^{0}{\left(1+t\right)}^{{\alpha }_{2}-1} log^{2} \left(1+t\right)dt\) is a convergent Bertrand’s integral.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Jourdan, A. Space-filling designs with a Dirichlet distribution for mixture experiments. Stat Papers 65, 2667–2686 (2024). https://doi.org/10.1007/s00362-023-01493-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00362-023-01493-2