1 Introduction
NewcombBenford’s law—see (Newcomb, 1881; Benford, 1938)—states that the occurrence of the first digit of a sequence follows a logarithmic distribution
(1) 
with . One paramount condition is that the sequence should span over multiple orders of magnitude to be true. The law can be extended to next significant digits—see (Hill, 1995)
. Starting from the 4th digit, the probability is almost equidistributed to all 10 digits with
. Note that the frequency of the digits is not independent. Hence, we cannot test for multiple digits occurrence at the same time without decreasing the power of the tests. Instead, the law is still valid if we are looking at the joint distribution of the leading digits.
can be seen as either , or . Doing so increases the power, but marginally (5%)—see (Joenssen, 2013). The loss of information due to rounding explains this behaviour.The NewcombBenford’s law has seen many applications ranging from fraud detection, to image analysis. Hill gives an explanation in (Hill, 1995): if random samples are generated from distributions selected at random, then the significant digits of the sample converge to the logarithmic distribution. Then, one can naturally wonder:
Can NewcombBenford’s law be used to assess the randomness of samples?
A sample corresponds to a given set of input parameters with , and is the number of dimensions. The set of samples is noted as and also called a Design of Experiments (DoE). DoE have numerous uses from numerical integration to experimental designs (Sacks et al., 1989). As the dimension grows, the volume of the hypercube increases exponentially. It quickly becomes intractable to completely fill the space.
Different metrics are used to characterize the space filling of samples. Mainly, there are geometrical and uniformity criteria. They respectively measure a distance between all points and measure how the points position deviates from the uniform distribution
(Fang et al., 2006; Androulakis et al., 2016). The latter is referred to as the discrepancy. Hence, using lowdiscrepancy methods such as Sobol’ (Sobol’, 1967) is a common practice. There are various ways to calculate the discrepancy, but they all share a common issue: their numerical complexity. The well used discrepancy is . When the measure is used in an optimization loop, the numerical cost can become quickly intractable.This work proposes a new method to assess the spacefilling of a sample which alleviates the computational complexity incurred with classical discrepancy measures. Starting from the sample to characterize, the values are transformed to the logarithmic law. The set is then compared to NewcombBenford’s law which gives a metric. These operations are very simple to perform and scalable.
The paper is organized as follows. Section 2 describes how to construct a discrepancy metric using NewcombBenford’s law. Section 3 demonstrates the performance of the method with respect to classical method used to assess the quality of a DoE. Finally, conclusions and future work are drawn in Section 4.
2 Presentation of the NewcombBenford discrepancy
The basis of the method is to compute the deviation from NewcombBenford’s law. First, the sample is scaled from the unit hypercube to the logarithmic law. Each sample’s first significant digit is then counted which leads to individual probabilities of occurence per digit. Goodness of fit is then used to compare this data with the logarithmic law which gives a metric. Algorithm 1 gives an overview of the process. In the following, it is referred to as the NewcombBenford discrepancy (NBD).
Goodness of fit is computed using Cramérvon Mises . It was shown in (Lesperance et al., 2016) to be a robust method to check the conformance with NewcombBenford’s law. Its discrete form corresponds to the Root Mean Square Error (RMSE) as
(2) 
with the digit to consider. The metric uses a flattened array—of size —so that all dimensions are taken into account: it is called the NewcombBenford discrepancy (NBD). This metric has a computational complexity of . Which is, not only a great improvement over the commonly used discrepancy, but the operations are also arguably simpler than classical discrepancy calculations.
One clear drawback when considering all dimensions at once is that the metric is invariant to coordinate permutations. When evaluating LHS designs, this could be an issue as a common optimization scheme consists in permuting the coordinates. Figure 1 shows two different LHS designs which are constructed using a permutation of digits from in 2dimensions. The design of the left is clearly superior in terms of space coverage. The discrepancy is respectively and . Still, the computed NBD would be the same as the set of digits would strictly be identical.
To mitigate this issue, I propose to consider all 2dimensional subprojections of the space—without the diagonal combinations. Hence, another metric can be computed based on the 2dimensional joint distribution of the coordinates. Considering higher order joint distribution would be intractable as there are possible digits to calculate the probabilities from. Hence, considering 2dimensional projections, the computational complexity becomes . It is still an improvement over classical discrepancies as we generally have .
3 Analysis of NewcombBenford discrepancy
In the following, crude MonteCarlo (MC) sampling and the low discrepancy sequences of Sobol’ are used to demonstrate the performance of the proposed method. Throughout the literature, and noticeably in (Kucherenko et al., 2015), Sobol’ proved to be superior in every way to MC. Hence, it is expected to show a smaller NBD measure for a given number of samples. Furthermore, the operations have been replicated 99 times to have converged statistics.
Let’s first take a look at Fig. 2. It presents the conformance with NewcombBenford’s law of two sets of
. The subfigures at the top present raw values whereas subfigures at the bottom present a boxplot of the NBD. Sobol’ method clearly follows the best the logarithmic law even at a low number of sample—along all 10 dimensions. This is confirmed when looking at the error levels as there is an order of magnitude between the two. The boxplot shows something more interesting, there is a clear heteroscedasticity of the error in case of MC whereas Sobol’ shows an homoscedasticity of the error. This indicates two things:
(i) from one run to another, Sobol’ would produce more consistent samples in terms of quality; (ii) it does not introduce biases as to favour more digits over the others.Then, the convergence of NBD is assessed with respect to for in Fig. 3. As expected, NBD converges with and at a faster rate and with lower values for Sobol’ than MC. This is in accordance with classical convergence results found in (Kucherenko et al., 2015). The jointNBD error seems to saturate as increases.
Looking back at the case in Fig. 1, the flattened NBD metric gives, as expected, the same value of for both designs. Whereas the 2dimensional joint version gives and —resp. for the design of the left and the design of the right. This is in accordance with results from the discrepancy which gives, resp., and . The hierarchy between the design is correct using both methods. But as seen in Section 2, the numerical complexity of the jointNBD is higher. Hence, if it is known that the designs to compare would not consist of permutations of coordinates (which is not the case for common LHS optimization methods for instance), the most efficient way can be used. Also, with large it seems that the joint version would not be able to be used to discriminate correctly the designs.
Note that both analyses were repeated with various number of samples—ranging from to —, and dimensions—ranging from to . In all cases, the results agreed with each other.
4 Conclusion
This work proposes a new method to assess the uniformity of a design. This method is based on NewcombBenford’s law, referred to as NBdiscrepancy (NBD). The RMSE between the logarithmic law is compared with the empirical probabilities of a given sample which leads to a sensible metric. It has also been shown that this strategy can be applied to joint distributions of digits in order to assess the uniformity over the different dimensions of the design. NBD allows having a fast measure of uniformity with a numerical complexity of . In case the designs to compare would consist of permuted coordinates, such as during the optimization of LHS, the joint distribution of digits can be used in the same way. In both cases, the convergence properties have been shown with respect to the number of samples and dimensions.
Compared to classical discrepancy such as the discrepancy, it is noticeably easier to implement. In (Fang et al., 2006), update strategies are given to update a LHS design and it leads to the same numerical complexity of NBD. But the individual operations are far more complex/expensive than simply counting the leading digits. Also, their update strategy is specific to the case of optimizing a LHS by permuting coordinates. The proposed method, on the other hand, is completely independent of the sampling strategy.
Being able to characterize a design is paramount as it determines the quality of the task it is used for. The proposed method provides an alternative to classical discrepancy methods used to control such design. It can be used to compare quantitatively designs, it is simple to implement and scalable which paves the way towards highdimensional optimizations.
Acknowledgements
The author acknowledges Prof. Art B. Owen from Stanford University for helpful discussions.
References
 Measures of uniformity in experimental designs: A selective overview. Communications in Statistics  Theory and Methods 45 (13), pp. 3782–3806. External Links: Document Cited by: §1.
 The Law of Anomalous Numbers. In Proceedings of the American Philosophical Society, Vol. 78, pp. 551–572. Cited by: §1.
 Design and modeling for computer experiments. Chapman & Hall/CRC. Note: Wrap arround Discrepancy error Cited by: §1, §4.
 The Significant Digit Phenomenon. The American Mathematical Monthly 102 (4), pp. 322–327. External Links: Document Cited by: §1, §1.
 Two Digit Testing for Benford’s Law. In Proceedings of the 59th World Statistics Congress of the International Statistical, pp. 3881–3886. Cited by: §1.
 Exploring multidimensional spaces: a Comparison of Latin Hypercube and Quasi Monte Carlo Sampling Techniques. The 8th IMACS Seminar on Monte Carlo Methods, pp. 1–32. External Links: 1505.02350 Cited by: §3, §3.

Assessing conformance with Benford’s law: Goodnessoffit tests and simultaneous confidence intervals
. PLoS ONE 11 (3), pp. 1–20. External Links: Document, ISSN 19326203 Cited by: §2.  Note on the Frequency of Use of the Different Digits in Natural Numbers. American Journal ofMathematics 4 (1), pp. 39–40. External Links: ISSN 00375497 Cited by: §1.
 Design and Analysis of Computer Experiments. Statistical Science 4 (4), pp. 409–423. External Links: Document Cited by: §1.
 On the distribution of points in a cube and the approximate evaluation of integrals. USSR Computational Mathematics and Mathematical Physics 7 (4), pp. 86–112. External Links: Document Cited by: §1.
Comments
There are no comments yet.