1. Introduction
Let , where is a real separable Hilbert space, be an isonormal Gaussian process defined on some probability space (see Definition 1). The authors in [1] discovered a celebrated central limit theorem, called the “fourth moment theorem”, for a sequence of random variables belonging to a fixed Wiener chaos associated with B (see Section 2 for the definition of Wiener chaos).
Let be a sequence of random variables belonging to the th Wiener chaos with for all . Then if and only if , where Z is a standard Gaussian random variable and the notation denotes the convergence in distribution.
After that, the authors in [2] obtained a quantitative bound of the distances between the laws of F and Z by developing the techniques based on the combination between Malliavin calculus (see, e.g., [3,4,5,6,7]) and Stein’s method for normal approximation (see, e.g., [8,9,10]). These distances can be defined in several ways. More precisely, the distance between the laws of F and Z is given by
(1)
where D and denote the Malliavin derivative and the pseudo-inverse of the Ornstein–Uhlhenbeck generator, respectively (see Definitions 2 and 5), and the constant in (1) only depends on the distance d considered. In the particular case where F is an element in the qth Wiener chaos of B with , the upper bound (1) for Kolmogorov distance () is given by(2)
where is the fourth cumulant of F.The application of the Stein’s method related to Malliavin calculus has been extended from the normal distribution to the cases of Gamma and Pearson distributions (see e.g., [11,12]). Furthermore, the authors in [13] extend the upper bound (1) to a more general class of probability distribution. For a differentiable random variable in the sense of the Malliavin calculus, they obtain the upper bound of distance between its law and a law of a random variable with a density that is continuous, bounded, and strictly positive in the interval () with finite variance. Their approach is based on the construction of an ergodic diffusion that has a density p as an invariant measure. The diffusion with the invariant density p has the form
(3)
where W is a standard Brownian motion. Then, they consider the generator of the diffusion process X and use the integration by parts (see Definition 3 for the integration by parts formula) to find an upper bound for the distance between the law of a differentiable random variable G and the law of a random variable F with density . This bound contains D and as in the bound (1). Precisely, for a suitable class of functions ,(4)
If a random variable G admits a density with respect to the Lebesgue measure, the Kolmogorov (i.e., ) and total variation distance () can be bounded by(5)
We note that Scheffe’s theorem implies that the pointwise convergence of densities is stronger than convergence in distribution. In this paper, we assume that the law of G admits a density with respect to the Lebesgue measure. This assumption on G is satisfied for all distributions considered throughout examples in the paper [13]. Using the bound of (4) and the diffusion coefficient in (3) given by
we derive a bound of general distances in the left-hand side of (4), being expressed in terms of the density functions of two random variables F and G as in the case of Kolmogorov and total variation distances. In addition, we deal with the computation of the conditional expectation in (4). When G is general, it is difficult to find an explicit computation of this expectation. The random variables in all examples covered in [13] are just functions of a Gaussian vector. In this case, it is possible to compute the explicit expectation. If the law of these random variables admits a density with respect to the Lebesgue measure, like all examples considered in [13], we can find the formula from which we can easily compute this expectation.The rest of the paper is organized as follows. Section 2 reviews some basic notations and the results of Malliavin calculus. In Section 3, we describe the construction of a diffusion process with an invariant density p and derive an upper bound between the laws of F and G in terms of densities. In Section 4, we introduce a method that can directly compute the conditional expectation in (4). Finally, as an application of our main results, in Section 5, we obtain an upper bound of an example considered in [13]. Throughout this paper, c (or C) stands for an absolute constant with possibly different values in different places.
2. Preliminaries
In this section, we briefly review some basic facts about Malliavin calculus for Gaussian processes. For a more detailed explanation, see [6,7]. Fix a real separable Hilbert space , with inner product .
We say that a stochastic process defined on is an isonormal Gaussian process if B is a centered Gaussian family of random variables such that for every .
For the rest of this paper, we assume that is the -field generated by X. To simplify the notation, we write instead of . For each , we write to denote the closed linear subspace of generated by the random variables , , , where the space is the qth Hermite polynomial. The space is called the qth Wiener chaos of B. Let denote the class of smooth and cylindrical random variables F of the form
(6)
where is a -function such that its partial derivatives have at most polynomial growth, and , . Then, the space is dense in for every .For a given integer and , the pth Malliavin derivative of F with respect to B is the element of , where the space denotes the symmetric tensor product of , defined by
(7)
For a fixed and an integer , we denote by the closure of its associated smooth random variable class of with respect to the norm
For a given integer , we denote by the adjoint of the operator , called the multiple divergence operator of order p. The domain of , denoted by , is the subset of composed of those elements u such thatIf , then is the element of defined by the duality relationship
(8)
The above formula (8) is called an integration by parts formula. For a given integer and , the qth multiple integral of f is defined by . Let with . Then, for any integer , we have . From this, the linear mapping by has an isometric property. It is well known that any square integrable random variable can be expanded into a series of multiple integrals:
where the series converges in , and the functions , , are uniquely determined by F. Moreover, if , then for all .For a given , we say that F belongs to if
where is the projection operator from into , that is, , . For such an F, the operator L is defined through the projection operator , , as .
It is not difficult to see that the operator L coincides with the infinitesimal generator of the Ornstein–Uhlhenbeck semigroup . The following gives a crucial relationship between the operator D, , and L: Let . Then, we have if and only if and . In this case, , that is, for , the statement is equivalent to .
For any , we define the operator , called the pseudo-inverse of L, as .
Note that is an operator with values in and for all .
3. Diffusion Process with Invariant Measures
In this section, we explain how a diffusion process is constructed to have an invariant measure that admits a density function, say p, with respect to the Lebesgue measure (see [13,14] for more information). Let be a probability measure on ( with a continuous, bounded, and strictly positive density function p. We take a function that is continuous such that exists for which for and for are satisfied. Moreover, the function is bounded on I and
(9)
For , let us set(10)
Then, the stochastic differential equation (sde)(11)
has a unique ergodic Markovian weak solution with the invariant measure .The authors prove in [15] that the convergence of the elements of a Markov chaos to a Pearson distribution can be still bounded with just the first four moments by using the new concept of a chaos grade. Pearson diffusions are examples of the Markov triple and Itô diffusion given by the sde
(12)
where m is the expectation of , and(13)
Let us define where F is a random variable having its law of . For , where , we define Then, satisfies that In [13], the authors derive the Stein’s bound between the probability measure and the law of an arbitrary random variable G. This bound extends the results in [2,12] in the case where is a standard Gaussian and Gamma distribution, respectively.(Kusuoka and Tudor (2012) [13]). Let F be a random variable having the target law μ with a probability distribution associated to the diffusion given by sde (11). Let G be an I-valued random variable in with . Then, for every such that and are bounded, the following holds:
(14)
and(15)
When the laws of F and G admit densities and (with respect to Lebesgue measure), respectively, we derive an upper bound (14) in terms of the densities of F and G by using Theorem 2.
Let F be a random variable having the law μ with the density associated to the diffusion given by sde (11). Let G be a random variable in with . Suppose that the law of G has the density with respect to the Lebesgue measure. Then, for every such that and are bounded, we find that
(16)
Let be a -function having a bounded derivative with a compact support. Using the integration by parts yields
(17)
The above equality (17) obviously shows that(18)
Using the relations (10) and (17), the first expectation in the right-hand side of (15) can be written as(19)
Since we have that This implies that (19) can be written as(20)
Combining (15) and (20) completes the proof of this theorem. □In Theorem 2 of [13], the authors prove that if a random variable has the invariant measure , then and
(21)
Furthermore, if admits the density , it is obvious from (19) that (21) holds.We think it would be interesting to give numerical examples from the computational validity in Theorem 3. In this respect, although not a numerical example, we give a simple example to deduce an upper bound for between the laws of two centered Gaussain random variables.
Let F and G be two centered Gaussian random variables with variances and . Then,
(22)
where is the class of functions to be chosen depending on the type of the distance d.Obviously, the random variable F has the law with the density
associated to the diffusion given by sde with and . Since , the second sum in (16) is vanished. Hence, from Theorem 3, it follows that(23)
Since the distance between two distributions F and G is given by the proof of this proposition is completed. □Depending on the choice of , several types of distances can be defined (see Section 5.2). Comparing the upper bound in Proposition 3.6.1 of [6] obtained from an elementary application of Stein’s method with the upper bound in (22) is very interesting. This shows that our study is differentiated from the existing ones
4. Computation of
When G is general, it is difficult to find an explicit computation of the right-hand side of (15). In particular, when is not measurable with respect to the -field generated by G, there are cases where it is impossible to compute the expectation. The next proposition in [4] contains an explicit example.
Let , where B is an isonormal Gaussian process and is a uniquely defined measurable function a.e. Then, we have
(24)
so that(25)
Here, B and are defined on the product space such that stands for an independent copy of B. and denote the expectation with respect to and , respectively.If , where is a -function with bounded derivative and is a d-dimensional Gaussian random variable with zero mean and covariance , , where stands for the canonical basis of . By using Proposition 2, the following useful formula can be proved:
(26)
In order to show the significance of the bound (15), the authors in [13] consider the several random variables G. Here, among these random variables, we consider random variables with the uniform and Laplace distribution. The random variable defined by where and are independent standard Gaussian random variables, has the uniform distribution . The authors in [13] compute the right-hand side of (26) to prove that(27)
Computing in this way is tedious and lengthy. To overcome this situation, we can use Equation (18) to prove that (27) holds. Since G has the uniform distribution , we have(28)
In the case where G has a Laplace distribution, the authors in [13] consider two random variables:(29)
(30)
where , , are orthonormal functions in . It can be easily seen that , , has the Laplace distribution with parameter 1. In the paper [13], the authors prove, using Theorem 2 in [13], that for ,(31)
The authors argue that these identities are difficult to be proven directly. Here, we introduce a method that can directly prove these identities (31) by using the formula given in (18). Since , , has a Laplace distribution with parameter 1, we find that for ,(32)
An elementary computation yields that for a.s,(33)
and for a.s.(34)
Combining (33) and (34) proves that the identity (31) holds.5. Example
In this section, we illustrate the upper bound of probabilistic distances in Theorem 3 through an example considered in [13]. We denote the Wiener integral of by . Let be a sequence of orthonormal bases of and a sequence of random variables defined by
(35)
Let F be a random variable having log normal distribution with mean and variance . Then, the density of F is given by(36)
Next, we compute the density of the random variable given by (35). We first compute the cumulative distribution function of . Let us set . Then, the random variable has a Gamma distribution with parameters and , that is,
(37)
Using (37), we find that for ,(38)
Differentiating Equation (38) proves that(39)
From (39), it follows that(40)
5.1. Scheffe’s Theorem
First, we prove that converges in distribution to F by using Scheffe’s theorem and then find a convergence rate of the Kolmogorov and total variation distance. The right-hand side of (40) can be written as
(41)
For any fixed , we have, from (36) and (41), that(42)
To estimate the first term in (42), we can use the following specific version of the Stirling formula of the function, incorporating upper and lower bounds (see [16]):Let . Then for all ,
(43)
The term in (42) can be written as
(44)
where Obviously,(45)
Hence, form (43) and (44),
(46)
Using the Taylor expansion of we write as(47)
Since we will have that , and hence, from (42), This convergence implies, from Scheffe’s theorem, that as , An upper bound for the Kolmogorov and total variation distance is given in (5). Hence, converges in distribution to F. Next, we find the rate of convergence for an upper bound for these distances by using the bound (5). By using the change of variables , we find, from (36) and (40), that(48)
Using the Taylor expansion of , the right-hand side of (48) can be represented as(49)
From (46), it follows that(50)
Obviously,(51)
From (50) and (51), we prove that the rate of convergence of the Kolmogorov and total variation distance between the laws of F and is of order .5.2. General Distance
In this section, we consider general distances between the laws of F and defined by
(52)
where is a class of functions defined on . Depending on the choice of , several types of distances can be defined. In addition to the Kolmogorov distance and total variation distance, the following distances can be obtained: for example, if , where denotes the Lipschitz seminorm defined by then the distance in (52) is called Wasserstein. If , the Fortet-Mourier will be obtained. The rate of convergence of this distance can be found by using the bound given in Theorem 3. The drift coefficient of the associated diffusion is given by(53)
where the function denotes the distribution function of the standard Gaussian distribution. Let us set . From (18) and (39), it follows that(54)
where m is the expectation of given by The right-hand side of (54) can be written as(55)
Using the change of variables , we express the right-hand side of (55) as
(56)
By using the expansion the right-hand side of (56) can be expressed as(57)
The change of variables shows that (57) is(58)
The Taylor expansion of , , is given by(59)
Applying this expansion (59) to a function , we have(60)
Substituting (60) into the integrand in (58) yields that(61)
From (36) and (53), the drift coefficient of diffusion is given by(62)
The use of the change of variables makes the right-hand side of (62) equal to(63)
From (61) and (63), we write , whereFor every , we have
(64)
where .We write , where . Hence,
(65)
Since we have(66)
Substituting (66) into (65) proves this lemma. □Next, we estimate , . The Cauchy–Schwartz inequality and Lemma 2 give the estimate
(67)
By Hölder inequality and Lemma 2, we have(68)
Similarly,(69)
Combining the bounds in (67), (68) and (69), we obtain(70)
Therefore, we find that the rate of convergence of the general distance is of order .
6. Conclusions and Future Works
When a random variable F follows the invariant measure that admits a density and a differentiable random variable G in the sense of Malliavin allows a density function, this paper derives an upper bound on several probabilistic distances (e.g., Kolmogorov distance, total variation distance, Wasserstein distance, and Forter–Mourier distance, etc.) between the laws of F and G in terms of two densities. Among these distances, it is well known that the upper bound of the Kolmogorov and total variation distance can be easily expressed in terms of densities. The significant feature of our works is to show that the bounds of distances other than the two distances mentioned above can be expressed in some form of two density functions. An insight into the main result of this study is that it is possible by applying our results to express an upper bound for the distance of two distributions in terms of two density functions even when it is difficult to express the distance as a density function of two distributions.
Future works will be carried out in two directions: (1) Using the results worked in this paper, we plan to conduct a study on the upper bound that is more rigorous than the results obtained in the papers [15,17]. (2) In the case when G is a random variable belonging to a fixed Wiener chaos, we will prove the fourth moment theorem by using the bound obtained in this paper.
Conceptualization, Y.-T.K. and H.-S.P.; Methodology, H.-S.P.; Writing—original draft, H.-S.P.; Writing—review and editing, H.-S.P.; Funding acquisition, H.-S.P. All authors have read and agreed to the published version of the manuscript.
Not applicable.
We are very grateful to the anonymous referees for their suggestions and valuable advice.
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
1. Nualart, D.; Peccati, G. Central limit theorems for sequences of multiple stochastic integrals. Annu. Probab.; 2005; 33, pp. 177-193. [DOI: https://dx.doi.org/10.1214/009117904000000621]
2. Nourdin, I.; Peccati, G. Stein’s method on Wiener Chaos. Probab. Theory Relat. Fields; 2009; 145, pp. 75-118. [DOI: https://dx.doi.org/10.1007/s00440-008-0162-x]
3. Kim, Y.T.; Park, H.S. An Edgeworth expansion for functionals of Gaussian fields and its applications. Stoch. Process. Appl.; 2018; 44, pp. 312-320. [DOI: https://dx.doi.org/10.1016/j.spa.2018.01.006]
4. Nourdin, I.; Viens, F.G. Density formula and concentration inequalities with Malliavin calculus. Electron. J. Probab.; 2009; 14, pp. 2287-2309. [DOI: https://dx.doi.org/10.1214/EJP.v14-707]
5. Nourdin, I.; Peccati, G. Stein’s method meets Malliavin calculus: A short survey with new estimates. Recent Development in Stochastic Dynamics and Stochasdtic Analysis; Interdisciplinary Mathematical Sciences World Scientific Publishing: Hackensack, NJ, USA, 2010; Volume 8, pp. 207-236.
6. Nourdin, I.; Peccati, G. Normal Approximations with Malliavin Calculus: From Stein’s Method to Universality; Cambridge Tracts in Mathematica Cambridge University Press: Cambridge, UK, 2012; Volume 192.
7. Nualart, D. The Malliavin Calculus and Related Topics; 2nd ed. Probability and Its Applications Springer: Berlin/Heidelberg, Germany, 2006.
8. Chen, L.H.Y.; Goldstein, L.; Shao, Q.-M. Normal Approximation by Stein’s Method; Probability and Its Applications Springer: Berlin/Heidelberg, Germany, 2011.
9. Stein, C. A bound for the error in the normal approximation to the distribution of a sum of dependent random variables. Proceedings of the Sixth Berkeley Symposium on Mathematical Statistics and Probability, Volume II: Probability Theory; Berkeley, CA, USA, 21 June–18 July 1970; University of California Press: Berkeley, CA, USA, 1972; pp. 583-602.
10. Stein, C. Approximate Computation of Expectations; IMS: Hayward, CA, USA, 1986.
11. Eden, R.; Viens, F. General upper and lower tail estimates using Malliavin calculus and Stein’s equations. arXiv; 2010; arXiv: 1007.0514
12. Nourdin, I.; Peccati, G. Non-central convergence of multiple integrals. Annu. Probab.; 2009; 37, pp. 1412-1426.
13. Kusuoka, S.; Tudor, C.A. Stein’s method for invariant measures of diffusions via Malliavin calculus. Stoch. Process. Appl.; 2012; 122, pp. 1627-1651. [DOI: https://dx.doi.org/10.1016/j.spa.2012.02.005]
14. Bibby, B.M.; Skovgaard, I.M.; Sorensen, M. Diffusion-type models with given marginals and auto-correlation function. Bernoulli; 2003; 11, pp. 191-220.
15. Bourguin, S.; Campese, S.; Leonenko, N.; Taqqu, M.S. Four moments theorem on Markov chaos. Annu. Probab.; 2019; 47, pp. 1417-1446. [DOI: https://dx.doi.org/10.1214/18-AOP1287]
16. Jameson, G.J.O. A simple proof of Stirling’s formula for the gamma function. Math. Gaz.; 2015; 99, pp. 68-74. [DOI: https://dx.doi.org/10.1017/mag.2014.9]
17. Azmoodeh, E.; Campese, S.; Poly, G. Fourth moment theorems for Markov diffusion generators. J. Funct. Anal.; 2014; 266, pp. 2341-2359. [DOI: https://dx.doi.org/10.1016/j.jfa.2013.10.014]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The Kolmogorov and total variation distance between the laws of random variables have upper bounds represented by the
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer