1 Introduction
Numerical integration of high dimensional functions is a very common and challenging problem. Under the right conditions, quasiMonte Carlo (QMC) sampling and randomized QMC (RQMC) sampling can be very effective. A good result can be expected from (R)QMC if the following conditions, described in more detail below, all hold:

the (R)QMC points have highly uniform low dimensional projections,

the integrand is nearly a sum of low dimensional parts, and

those parts are regular enough to benefit from (R)QMC.
The first condition is a usual property of (R)QMC points. In a series of papers, Griebel, Kuo and Sloan [9, 10, 11] address the third condition by showing that the low dimensional parts of (defined there via the ANOVA decomposition) are at least as smooth as the original integrand and are often much smoother. They include conditions under which lower order ANOVA terms of functions with discontinuities (jumps) or discontinuities in their first derivative (kinks) are smooth. An alternative form of regularity, instead of smoothness, is for the low dimensional parts to have QMCfriendly discontinuities as described in [35]. In this article we explore sufficient conditions for the remaining second condition to hold. We use the mean dimension [26] to quantify the extent to which low dimensional components dominate the integrand.
This article is focused on ridge functions defined over . Ridge functions take the form for an orthonormal projection matrix where , with being an important special case. Ridge functions are useful here because we can find their integrals via low dimensional integration or even closed form expressions. That lets us investigate the impact of some qualitative features of on the integration problem. Additionally, many functions in science and engineering are well approximated by ridge functions with small values of [3], so good performance on ridge functions could extend well to many functions in the natural sciences. As one more example, the value of a European option under geometric Brownian motion is a ridge function of the Brownian increments and this is what allows the formula of Black and Scholes to be applied [8].
Our main finding is that there is an enormous difference between functions with jumps and functions with kinks. This is perhaps surprising. Based on criteria for finite variation in the sense of Hardy and Krause, one might have thought that a jump in dimensions would be similar to a kink in . Instead, we find that for Lipschitz continuous , the mean dimension of is bounded as and that bound can be quite low. For with step discontinuities, we find that the mean dimension can easily grow proportionally to . These effects were seen empirically in [28] where ridge functions were used to illustrate a scrambled Halton algorithm. Preintegration [12] turns a ridge function over with a jump into one with a kink, and ridge functions of Gaussian variables containing a jump can even become infinitely differentiable. The resulting Lipschitz constant need not be small. For a linear step function we find that preintegration can either increase mean dimension or reduce it from to .
An outline of this paper is as follows. Section 2 provides notation and background concepts related to quasiMonte Carlo and mean dimension. Section 3 introduces ridge functions and establishes upper bounds on their mean dimension in terms of Hölder and Lipschitz conditions and some spatially varying relaxations of those conditions. Corollary 1 there shows that a ridge function with Lipschitz constant
and variance
cannot have a mean dimension larger than in any dimension for any projection . Section 4 considers ridge functions with jumps. They can have mean dimension growing proportionally to and sparsity of makes a big difference. Section 5 considers the effects of preintegration on ridge functions. The preintegrated functions are also ridge functions with a Hölder constant no worse than the original function had. Preintegration can either raise or lower mean dimension. We give an example step function where preintegration leaves the mean dimension asymptotically proportional to with an increased lead constant. In another example, preintegration can change the mean dimension from growing proportionally to to having a finite bound as . Section 6 computes some mean dimensions using Sobol’ indices. Section 7 has conclusions, and a discussion of how generally these results may apply. Section 8 is an appendix containing the longer proofs.2 Background and notation
We use
for the standard Gaussian probability density function and
for the corresponding cumulative distribution function. We consider integration with respect to a
dimensional spherical Gaussian measure,where the quantile function
is applied componentwise. The (R)QMC approximations to take the form for points and . The distribution of is denoted or simply if is understood from context.QMC and KoksmaHlawka
For QMC, the KoksmaHlawka inequality [14]
(1) 
bounds the error in terms of the star discrepancy of the points used and the total variation of in the sense of Hardy and Krause. Constructions with are known [21, 5, 31], proving that QMC can be asymptotically better than Monte Carlo (MC) sampling which has a root mean squared error of . That argument requires which requires at a minimum that be a bounded function on . Scrambled net RQMC has a root mean squared error that is for any without requiring bounded variation [24].
Kinks and jumps
A kink function is continuous with a discontinuity in its first derivative along some manifold. Griebel et al. [12] consider kink functions of the form where is smooth. The kink takes place within the set . A jump function has a step discontinuity along some manifold. Griewank et al. [12] consider jump functions of the form where
is also smooth. There can be jump discontinuities within the set
. When , the result is a kink function. In the rest of this paper, denotes a unit vector.ANOVA and mean dimension
The ANOVA decomposition applies to any measurable and square integrable function of independent random inputs. In our case, those inputs will be either or .
We use for , and for , we write for the cardinality of and for the complement . The point has components for . The point has the components for . We abbreviate to . For and points , the hybrid point has for and otherwise.
The ANOVA decomposition [15, 33, 6] of is where depends on only through . For these functions, the line integral whenever and from that it follows that when and then
for variance components for and .
The mean dimension of (in the superposition sense) is
If we choose with probability proportional to then is the average of . Effective dimension is commonly defined via a high quantile of that distribution such as the 99’th percentile [2]. Such an effective dimension could well be larger than the mean dimension but it is more difficult to ascertain.
The mean dimension and a few other quantities that we use are not well defined when . In such cases, is constant almost everywhere and we will not ordinarily be interested in integrating it. We assume below, without necessarily stating it every time, that .
Sobol’ indices are used to quantify the importance of a variable or more generally a subset of them. We will use the (unnormalized) Sobol’ total index for variable ,
More generally, for , we set . An easy identity from [19] gives . Sobol’ [34] shows that
when and are independent random vectors with the same product distribution on . As a result we find that
(2) 
The expectation in the numerator of is a dimensional integral over independent and . It is commonly evaluated by (R)QMC.
Low effective dimension
Applying (1) componentwise yields
The coordinate discrepancies are known to decay rapidly when is small [5]. If also is negligible when is not small then can be considered to have low effective dimension and an apparent error for QMC can be observed. Some other ways to decompose a function into a sum of functions, one for each subset of , are described in [17]. For a survey of effective dimension methods in information based complexity, see [36].
To avoid the dependence on finite variation and to control the logarithmic terms we will use a version of RQMC known as scrambled nets. Under scrambled net sampling [23] each , while collectively remain digital nets with probability one, retaining their low discrepancy. The mean squared error of scrambled net sampling decomposes as
(3) 
where expectation refers to randomness in the [24]. If , then
(4) 
for some gain coefficient [25]. If also then
(5) 
If large have negligible and small are smooth enough for (5) to hold then RQMC may attain nearly root mean squared error. The logarithmic factors in (5) cannot make the variance much larger than the MC rate because the bound in (4) applies for finite .
The ANOVA decomposition of on is essentially the same as that of on . Specifically, .
Discontinuities can lead to severe deterioration in the asymptotic behavior of RQMC. He and Wang [13] obtain MSE rates of for jump discontinuities of the form where the set has a boundary with dimensional Minkowski content. When is the Cartesian product of a hyperrectangle and a dimensional set with a boundary of dimensional Minkowski content, then takes the place of in their rate. The smaller is, the more ‘QMCfriendly’ the discontinuity is.
3 Ridge functions
We let , choose an orthonormal matrix , and define the ridge function
(6) 
where . We must always have because otherwise is impossible to attain. Our main interest is in . Ridge functions can also be defined for but then the domain of becomes a complicated polyhedron called a zonotope [3].
When , we write
(7) 
where . Then, because we find that
We can get the answer and the corresponding RMSE under MC by one dimensional integration. For some , one or both of these quantities are available in closed form. Note that and above are both independent of and even of . For more general we find that and are dimensional integrals that do not depend on or on . Apart from a few remarks, we focus mostly on the case with .
By symmetry we can take all . It is reasonable to expect that sparse vectors will make the problem of intrinsically lower dimension. Sparsity is typically defined via small values of . It is common to use instead a proxy measure , with smaller values representing greater sparsity, relaxing an quantity to an quantity. By this measure, the ‘least sparse’ unit vectors are of the form while sparsest are of the form where is the ’th standard Euclidean basis vector.
We will need some fractional absolute moments of the
distribution. For define(8) 
This is from formula (18) in an unpublished report of Winkelbauer [37]. It can be verified directly by change of variable to .
Theorem 1
Let be a ridge function described by (6) for , where satisfies a Hölder condition for , , and . Then the mean dimension of satisfies
(9) 
where does not depend on .
Proof
Let and be independent random vectors. For , let be the ’th row of as a row vector. Then . Next
(10) 
because . Summing over gives (10). Finally, depends on the distribution of for which is independent of .
If , then we recognize as where is a matrix norm [22]. For , we get and this is then not a norm. If
is an orthogonal matrix, then
. Now so we can replace in (9) by . For , we get , the squared Frobenius norm of , and the bound in (9) simplifies to reveal a proportional dependence on .Corollary 1
Let be a ridge function described by (6) where is Lipschitz continuous with constant and with , for . Then
where does not depend on .
Proof
Take in Theorem 1.
The bound in Theorem 1 and its corollaries is conservative. It allows for the possibility that for all pairs of points . If that would hold for and , then it would imply that is linear. To see why, note that any triangle with points , , and , for distinct would have one angle equal to . A linear function would then have mean dimension , the smallest possible value when . A less conservative bound is in Section 3.1 below. The next result show that the bound has a dimensional effect when .
Corollary 2
Let be a ridge function given by (7) with , where is Hölder continuous with constant and exponent and is a unit vector for . Then
Proof
From Theorem 1, . The largest value this can take arises for . Then and so as required.
3.1 Spatially varying Hölder and Lipschitz constants
A Lipschitz or Hölder inequality provides a bound on that holds for all . The numerator in is a weighted average of over points and indices , and for a ridge function that reduces to a weighted average of Applying a Lipschitz or Hölder inequality bounds an quantity by the square of an quantity.
We say that satisfies a spatially varying Hölder condition if for some there is a function such that
(11) 
holds for all and . If , then satisfies a spatially varying Lipschitz condition. The well known locally Lipschitz condition is different. It requires that every be within a neighborhood on which has a finite Lipschitz constant . Equation (11) is stronger because it also bounds for .
We will use a Hölder inequality via and satisfying to slightly modify the proof in Theorem 1. Under (11)
(12) 
Allowing would have made and then the supremum norm of would be infinite, leading to a useless bound. For , we interpret as recovering Theorem 1. The bound (12) simplifies for and for . Under both simplifications,
To get a finite bound for it suffices for to have a finite moment of order for some .
3.2 A kink function
As a prototypical kink function, consider given by (7) with for some threshold . This is Lipschitz continuous with . Using indefinite integrals and , the first two moments of are
Because , we get . For , we get and and then
for any and any unit vector .
3.3 The least sparse case
The least sparse unit vectors have all . Because is symmetric we may take . In this case, it is easy to compute using Sobol’ indices. By symmetry, equals a three dimensional integral
(13) 
for any . Furthermore, by comparing results for to those for we can see some impact from sparsity because the least sparse unit vector for dimension will give the same answer as a very sparse dimensional vector with zeros and the remaining components equal.
4 Jumps
While both kinks and jumps may have smooth low dimensional ANOVA components, jumps do not necessarily have the same low mean dimension. They are also sensitive to sparsity of .
4.1 Linear step functions
First we consider a step function . We get upper and lower bounds for the mean dimension of this function in terms of the nominal dimension and , our sparsity measure. Over the range from sparsest to least sparse .
Theorem 2
Let for a threshold and a unit vector . Then, for ,
Proof
See Section 8.1 of the Appendix.
The rate in Theorem 2 arises for . More generally we get
For instance, if has components equal to and the rest equal to zero, then the upper bound is . There can thus be a significant improvement due to sparsity of .
Theorem 3
Let for a threshold and a unit vector . Then, for ,
Proof
See Section 8.3 of the Appendix.
The proof of Theorem 3 requires a certain lower bound on a bivariate Gaussian probability. We did not find many such lower bounds in the literature, so this may be new and may be of independent interest.
Lemma 1
Let
with and choose . Then
Proof
See Section 8.2 of the Appendix.
Choosing in Theorem 3 provides an example of a set of jump functions with mean dimension bounded below by a positive multiple of . Here again sparsity plays a role in the bound.
The bounds in both Theorems 2 and 3 depend on . The upper bound argument in Theorem 2 uses a mean value approximation where could be replaced by a value just over , yielding for that
by a Mills’ ratio inequality as . As a result the upper bound is not as sensitive to large as the presence of in the denominator from Theorem 2 would suggest.
The case is simpler. We find
using a definite integral from Section 2.5.2 of [30]. After some algebra
(14) 
Now as . Therefore holds if holds as . Thus there is no asymptotic factor when and, from the details of our proof, we suspect it is not present for other .
4.2 More general indicator functions
It is reasonable to expect indicator functions to have such large mean dimension for more general sets than just half spaces in
under a spherical Gaussian distribution. Here we sketch a generalization. First, for an indicator function
of a measurable set we have(15) 
for . The numerator expectations are with respect to random , and (15) holds for any distribution on with independent components, including and . We work with the latter case in what follows.
As in [10, 11] we take and place conditions on . Let be strictly monotone in each coordinate . Without loss of generality, suppose that is strictly increasing in each . Suppose additionally that and for all and all .
For any , there is a unique value for which . We write and sometimes suppress its dependence on . We can make a linear approximation to the boundary of at via where both , the normalized gradient of , and depend on . By monotonicity of , each . Let
Now . In words, is what we would get by sampling , finding the boundary points corresponding to the component directions , summing the corresponding values, and averaging the results over all samples. Each point leads to consideration of points . This process produces an unequally weighted average over points of a sum of values determined by the tangent plane at .
For a linear , we get , and we find from Theorem 3 that is then bounded below by a multiple of which can be as large as . For more general , the boundary set is no longer an affine flat, the sparsity measure varies spatially over , and so does the length . A large mean dimension, comparable to , could arise if has a nonsparse gradient over an appreciable proportion of .
If the assumption that fails, or if fails, for some value , then we can no longer find the corresponding point . In that case, the given value of and contribute nothing to the numerator of . The mean dimension can still be large due to contributions from other values of and from other . A similar issue came up in [11] where existence of for every proved not to be satisfied by an integrand from computational finance, and also proved not to be necessary for the smoothing effect of ANOVA to hold.
4.3 Cusps of general order
For and , consider a cusp of order given by
(16) 
taking . Now for [27], for , and more generally for . The higher the dimension, the greater smoothness is required to have finite variation. The boundary is not parallel to any of the coordinate axes, so this integrand is not QMCfriendly in any way.
These functions are carefully constructed to be among the simplest with the prescribed level of smoothness. As a result, we may find their mean dimension analytically.
Theorem 4
The function defined above for has mean dimension
Proof
See Section 8.4 of the Appendix.
Comments
There are no comments yet.