1 Introduction
Isotonic regression refers to the problem of estimating a monotone sequence
based on a noisy observation vector
which is assumed to be an additive perturbation of ,where the components of
are assumed to have zero mean and unit variance. It is commonly assumed that
are independent and identically distributed (i.i.d.) but we work with the more general assumption of exchangeability in this paper. A natural estimator for in this setting is the isotonic Least Squares Estimator (LSE), defined aswhere denotes the usual Euclidean norm on and is the monotone cone of length nondecreasing sequences. As is a closed convex cone, as defined above exists uniquely; it can also be computed in time by the pool adjacent violators algorithm [4, 11].
The statistical properties of are typically studied in terms of the risk or the normalized mean squared error:
A key quantity in understanding is
where denotes the law of the noise vector . Indeed, it is clear that
When are not all equal, let be the finest partition of such that is constant on each . It has been shown [15, 9, 3] that
(1.1) 
where denotes the marginal distribution of and is the length of the block for all . We emphasize that (1.1) holds for arbitrarily dependent with zero mean and finite variance. It was also shown in [3] that also bounds the risk of the isotonic LSE in misspecified settings where does not lie in .
The quantity therefore crucially controls the risk of the isotonic LSE. The goal of this paper is to explicitly determine for every under the additional assumption that is exchangeable. Specifically, under the assumption of exchangeability, we show in Corollary 3.3 that, for all ,
(1.2) 
where is the harmonic number and is the pairwise correlation. Combined with (1.1), our result provides a sharp nonasymptotic bound on the risk of isotonic regression for any exchangeable noise vector. In the special case when are i.i.d. with zero mean and unit variance, and thus (1.2) gives:
(1.3) 
Here is the common distribution of the independent variables .
Previously, the formula (1.3) was known when is the standard Gaussian probability measure on . This was observed by Amelunxen et al. [2] who proved it by observing first that when and is the standard Gaussian measure, the formula
(1.4) 
holds for every closed convex cone where is the intrinsic volume of . When is the monotone cone, the right hand side in equation (1.4) can be shown to be equal to by using the fact that the generating function can be computed in closed form. Amelunxen et al. [2] used the theory of finite reflection groups [7] to obtain the exact expression for this generating function. However, the exact expression for can already be found in the classical literature on isotonic regression (see Theorem 2.4.2 in Roberston et al. [16] and references therein).
The above proof does not work for nonGaussian mainly because the expression (1.4) does not hold for general . In fact, the best available result on for nonGaussian is in equation (2.11) of Zhang [18], who proved the asymptotic result:
This bound gives the right behavior as the right hand side of equation (1.3) but only as . We improve this result by proving for every that is always equal to the harmonic number for every probability measure having mean and variance .
We prove (1.2) by developing a precise characterization of the marginal distribution of each individual component of . Specifically, as long as is exchangeable, we show in Theorem 2.2 that has the same distribution as , the order statistic of the running averages . We prove Theorem 2.2 in Section 2, using a characterization of the components of the isotonic LSE as the lefthand slopes of the greatest convex minorant of the random walk with increments . This result and its continuoustime analogue may be of independent interest outside the study of isotonic regression, so in Section 2 we also address consequences for the greatest convex minorant of a stochastic process with exchangeable increments. The order statistics of the running averages can be fairly complicated even when is Gaussian; however, Theorem 2.2 easily implies results such as (1.2). In Section 3, we detail some risk calculations for isotonic regression and its variants which all follow from Theorem 2.2.
2 Main Result
Let denote the partial sums for , started at . Identify the random walk with its cumulative sum diagram (CSD) , where for integers
and linearly interpolated between integers. Let
denote the greatest convex minorant (GCM) of , i.e. the greatest convex function that lies below . See Figure 1 for a depiction of the GCM of the CSD. With this notation, we now recall the graphical representation of the isotonic LSE as given in Theorem 1.2.1 of Roberston et al. [16].Lemma 2.1.
For any vector , the isotonic LSE is given by the lefthand slopes of the greatest convex minorant of the cumulative sum diagram. For all
For the remainder of this section let
(2.1) 
denote the lefthand slope of the GCM at , so is equal to by the lemma. In particular, when we have . When , we have , and if then . Our next result generalizes this observation, showing that the slope is equal in distribution to the smallest running average if is exchangeable.
Theorem 2.2.
Suppose is exchangeable. Let denote the running average for and let denote their order statistics. Then
(2.2) 
marginally for all .
Proof.
As before, let denote the partial sum. Let be the last argmin of the sequence , and let be the amount of time the walk is nonpositive . We will use Corollary 11.14 of Kallenberg [12], due to SparreAndersen, which says as long as is exchangeable.
Note that the slope of the GCM switches from nonpositive to positive at time , since the horizontal line with intercept minorizes the GCM and touches it at time . Hence, no matter the sequence of increments , there is the identity of events
(2.3) 
Also, for the time that the walk is nonpositive, since if and only if , there is the identity of events
The equality in distribution then implies
If the sequence is modified to for some fixed , the modified sequence is exchangeable, and the values of and for the modified sequence are just and . Applying the above identity to the modified sequence gives
So and
have the same cumulative distribution function, hence the same distribution. ∎
The proof of Theorem 2.2 has a straightforward generalization to the setting where is a continuoustime stochastic process. Knight [13] showed that the analogous distributional identity holds when has exchangeable increments and . Hence, by a similar proof, we find that the slope of the greatest convex minorant of at time has the same distribution as the percentile point of the occupation measure for the process . We record this result as the following corollary.
Corollary 2.3.
Let denote a realvalued càdlàg stochastic process on with exchangeable increments, such that . Define as the slope of the greatest convex minorant of at , and let denote the (random) cdf associated with the occupation measure of ,
(2.4) 
where denotes Lebesgue measure. Then
(2.5) 
marginally for all .
See Abramson et al. [1] for a general study of convex minorants of random walks and processes with exchangeable increments. In the special cases where is a standard Brownian motion or Brownian bridge on the unit interval, Carolan & Dykstra [6] derive the distribution of the slope , jointly with the process and its convex minorant at , for a fixed value . Given our corollary, their explicit formula for the slope provides the distribution of , giving new information about the occupation measure of for Brownian motion and Brownian bridge. The distribution of the percentile point of the occupation measure for has been obtained under the same generality as Corollary 2.3: see the introduction of Dassios [8] and references therein.
3 Consequences for Isotonic Regression
Since the identity of Theorem 2.2 holds marginally, it allows us to simplify expectations of functions that are additive in the components of . As long as is exchangeable,
(3.1) 
Taking , we obtain our first corollary.
Corollary 3.1.
Suppose is exchangeable. For ,
(3.2) 
provided .
Remark 3.2.
Viewed through its graphical representation, is the leftderivative of the GCM at , so when the power , equation (3.2) yields the discrete arclength formula
(3.3) 
Closely related to this formula is the identity of Spitzer & Widom [17], which takes
to be a sequence of i.i.d. random variables in
(or the complex plane ) with finite variance. If is the partial sum and is the length of the perimeter of the convex hull , then(3.4) 
These formulas connect the geometry of the convex hull of a random walk to the magnitudes of the running means.
Consider the case when . Since is exchangeable, every pair of components has the same correlation . If we further assume has zero mean and unit variance, the right hand side of equation (3.2) can be computed explicitly
Summing over yields our next result.
Corollary 3.3.
Suppose is an exchangeable random vector with zero mean, unit variance, and pairwise correlation . Then
This result should be contrasted with other distributionfree identities, namely
provided has i.i.d. components with zero mean and unit variance. In particular, suppose we observe where has i.i.d. components with zero mean and unit variance, but it turns out that is constant. If we know is constant, we can estimate it by a constant sequence and pay a constant price in total risk. If we know nothing about the structure of and use , the risk is quite large by comparison. The monotone sequence estimate resides in the middle, with a much smaller risk of and knowledge only about the relative order. We explained in Section 1 how risk calculations when generalize to MSE bounds that are sharp in the low noise limit for arbitrary . For example, when has constant pieces, then (1.1), Corollary 3.3 and the fact that for every imply that
whenever are i.i.d. with mean zero and unit variance. Also if is not necessarily in , then Corollary 3.3, together with the results of [3], implies that
where is the number of constant pieces of the vector . These formulae (with the leading constant of 1 in front of the term on the right hand side) were previously only known when the distribution of was standard Gaussian.
Define the risk of the isotonic LSE
so that . We can similarly employ Theorem 2.2 to explicitly calculate the risk of the isotonic LSE when is constant and is Gaussian:
Corollary 3.4.
Suppose . Then for any ,
where .
Proof.
Note and apply the theorem. ∎
Corollary 3.4 should similarly be contrasted with the following identities when :
respectively. In particular, when , the bound holds for all , which is to say is bounded when whereas grows without bound as grows.
When is constant and , the risk of isotonic regression is
(3.5) 
When , Theorem 2.3 of Zhang [18] shows an asymptotic result for the risk on constant that agrees with equation (3.5).
The continuoustime distributional identity in Corollary 2.3 applies to the asymptotic distribution of the isotonic least squares estimator. A standard model for studying the asymptotic behavior of isotonic regression is
where is nondecreasing. We observe , a noisy version of , and calculate by projecting onto the monotone cone. The function estimate is defined by and linearly interpolated between design points. Here, as before, the dependence on in is suppressed, but now we are interested in the behavior of isotonic least squares at a fixed point as .
Define the partial sum process by , linearly interpolated between design points. When the function is constant, the quantity
is given by the leftderivative of the greatest convex minorant of at . By the invariance principle, this converges in distribution to the leftderivative of the greatest convex minorant of standard Brownian motion at . This asymptotic result is well known and a similar result was noted for the Grenander estimator in Carolan & Dykstra [5], where Brownian motion is replaced with a Brownian bridge. Corollary 2.3 relates this asymptotic distribution to the percentile points of the occupation measure for .
Finally, Corollary 3.3 on the projection onto extends over to that of the set of nonnegative monotone sequences . Theorem 1 of Németh & Németh [14] observes that the projection of onto is given by , the elementwise positive part of the projection onto . Hence the distributional identity Theorem 2.2 yields a similar set of identities for nonnegative isotonic regression.
Corollary 3.5.
For any exchangeable noise vector ,
(3.6) 
Provided ,
(3.7) 
Furthermore, if is symmetric with unit variance, the generalized statistical dimension of the monotone cone is
(3.8) 
where is the pairwise correlation.
Proof.
Equation (3.8) is also shown in Amelunxen et al. [2] in the special case using the theory of finite reflection groups. The identity (3.7) allows us to show equation (3.8) for a much wider variety of noise vectors, and as before also allows us to obtain relations for the expected norms of the projection of the noise vector. All of our exact formulae follow from the distributional identity in Theorem 2.2, which exploits the geometric characterization of the isotonic LSE in Lemma 2.1. An interesting open question is whether similar characterizations— such as for convex regression [10]—may yield exact nonasymptotic risk calculations in other shapeconstrained estimation problems.
References
 [1] J. Abramson, J. Pitman, N. Ross, and G. U. Bravo. Convex minorants of random walks and lévy processes. Electronic Communications in Probability, 16:423–434, 2011.

[2]
D. Amelunxen, M. Lotz, M. B. McCoy, and J. A. Tropp.
Living on the edge: Phase transitions in convex programs with random data.
Information and Inference: A Journal of the IMA, 3(3):224–294, 2014.  [3] P. C. Bellec. Sharp oracle inequalities for least squares estimators in shape restricted regression. The Annals of Statistics, 46(2):745–780, 2018.
 [4] H. D. Brunk, R. E. Barlow, D. J. Bartholomew, and J. M. Bremner. Statistical inference under order restrictions; the theory and application of isotonic regression. Wiley, NewYork, 1972.
 [5] C. Carolan and R. Dykstra. Asymptotic behavior of the grenander estimator at density flat regions. Canadian Journal of Statistics, 27(3):557–566, 1999.
 [6] C. Carolan and R. Dykstra. Marginal densities of the least concave majorant of Brownian motion. Ann. Statist., 29(6):1732–1750, 2001.
 [7] H. S. M. Coxeter and W. O. J. Moser. Generators and relations for discrete groups, volume 14. Springer Science & Business Media, 2013.

[8]
A. Dassios.
On the quantiles of brownian motion and their hitting times.
Bernoulli, 11(1):29–36, 2005.  [9] B. Fang and A. Guntuboyina. On the risk of convexconstrained least squares estimators under misspecification. arXiv preprint arXiv:1706.04276, 2017.
 [10] P. Groeneboom, G. Jongbloed, and J. A. Wellner. Estimation of a convex function: characterizations and asymptotic theory. The Annals of Statistics, 29(6):1653–1698, 2001.
 [11] S. J. Grotzinger and C. Witzgall. Projections onto order simplexes. Applied mathematics and Optimization, 12(1):247–270, 1984.
 [12] O. Kallenberg. Foundations of modern probability. Springer Science & Business Media, 2006.
 [13] F. B. Knight. The uniform law for exchangeable and Lévy process bridges. Astérisque, (236):171–188, 1996.
 [14] A. B. Németh and S. Z. Németh. How to project onto the monotone nonnegative cone using pool adjacent violators type algorithms. arXiv preprint arXiv:1201.2343, 2012.
 [15] S. Oymak and B. Hassibi. Sharp mse bounds for proximal denoising. Foundations of Computational Mathematics, 16(4):965–1029, 2016.
 [16] T. Robertson, F. T. Wright, and R. L. Dysktra. Order restricted statistical inference. 1988.
 [17] F. Spitzer and H. Widom. The circumference of a convex polygon. Proceedings of the American Mathematical Society, 12(3):506–509, 1961.
 [18] C.H. Zhang. Risk bounds in isotonic regression. The Annals of Statistics, 30(2):528–555, 2002.