1 Introduction
The estimation of parameters of a model with “uncertain prior information” on parameters of interest began with Bancroft (1944) in the classical front. But a breakthrough came when Stein (1956) and James and Stein (1961) proved that the sample mean in a multivariate normal model is not admissible under a quadratic loss, for dimension more than two. This very result gave birth to a class of shrinkage estimators of various form and setup. A partial document on preliminary test and Steintype estimators are given by Judge and Bock (1978). The Steintype estimators have been reformulated and expanded by Saleh (2006, Ch 4.4.3) which includes asymptotic and nonparametric methods. Due to the immense impact of Stein’s results, scores of technical papers appeared in the literature covering various areas of applications. Here is one with the popular LASSO estimator.
This paper is devoted to the study of the performance characteristics of several improved estimators of LASSO based on preliminary test and Stein’s principle, and the comparison of the LASSO estimator with the least square estimator (LSE), improved preliminary test estimator (IPT), shrinkage preliminary test estimator (SPTE), Steintype shrinkage LASSO estimator (SLE), and Steintype positiverule shrinkage LASSO estimator (PSLE). An important characteristic of LASSOtype estimators is that they provide simultaneous estimation and selection of coefficients in linear models and can be applied when dimension of the parameter space exceeds the dimension of the sample space. Our conclusions are made based on simulated meansquared errors (MSE) and relative efficiency tables and graphs. It is shown that the modified LASSO carries on with the same dominance characteristics as the usual preliminary test and Steintype estimators (Saleh, 2006).
Our contribution in this paper is unique. We have proposed a set of LASSObased shrinkage estimators that perform superior to the classical LASSO estimator. We studied the theoretical properties of the estimators in terms of asymptotic mean squared errors (AMSE). Analytical expressions for the asymptotic risk functions of the proposed estimators have been provided. We carried out Monte Carlo simulation experiments to study the riskbehavior of the proposed estimators and their comparisons with the LASSO estimator. Application of the proposed estimators have been demonstrated using three real life data examples where the proposed estimators performed superiorly to the classical LASSO estimator.
The organization of the paper is as follows. Section 2 contains the basic informations about the LASSO and preliminary test and Steintype estimators. Proposed improved estimators of LASSO are presented here. Risk properties and riskcomparisons of various estimators are presented in this section. In section 3 we outline our simulation setup and discuss simulation results. Details of the analysis of relative efficiencies of estimators are presented in this section. In section 4, applications of the proposed estimators have been demonstrated using three real life data sets. Finally, conclusions are provided in section 5.
2 Linear model and estimators
Consider the linear multiple regression model
(2.1) 
where is the design matrix and iid error vector. Further we assume that and .
It is wellknown that the LSE of is given by
(2.2) 
which we use for obtaining preliminary test and shrinkage estimators. LSE
is the “best linear unbiased estimator (BLUE)” of
. The solution (2.2) depends on the nonsingularity of the matrix . If it is singular, the uniqueness of is lost and we end up with multiple solutions with varying variances and some of them may be very large. To avoid these problems, a class of penalty estimators evolved in the class of restricted estimators. A simple example of “restricted estimator” when we want an estimator of which belongs to the subspace defined by , where is a matrix and is a vector or real numbers.Next, we consider the classical penalty estimator, called the LASSO (least absolute shrinkage and selection operator) belonging to a class of restricted estimators.
The estimators of is obtained by minimizing the LS criterion subject to . Explicitly we may write
The solution for this problem is the “restricted estimator” and the tuning parameter can be explicitly obtained giving
(2.3) 
where , the LSE.
For this, we consider as the restriction and minimize
(2.4) 
yielding the solution as LASSO introduced by Tibshirani (1996), and is given by
(2.5) 
When , the solution may be written as
(2.6)  
where
Actually Frank and Friedman (1993) defined the class of generalized version of LASSO, namely, the “bridge estimator” as
(2.7) 
If the solution reduces to the LASSO estimator.
LASSO proposed by Tibshirani (1996) simultaneously estimates and makes selection of variables with appropriate interpretation and its viral popularity in applications. For computational solution and methodology see Tibshirani (1996) and Efron et al. (2004). Later Efron et al. (2004) proposed Least Angle Regression (LAR) which is a stepwise regression, and Friedman et al. (2010) developed an efficient algorithm for the estimation of a GLM with convex penalty. During the course of development of penalty estimators, Fan and Li (2001) defined good penalty functions as the one which yield (i) nearly unbiased estimator when true parameter is large to avoid unnecessary modeling bias, (ii) an estimator which is a threshold rule that sets small estimated coefficients to zero to reduce model complexity, and (iii) the resulting estimator to be continuous in the data to avoid instability in the model prediction. In this paper, we present an improved version of LASSO.
2.1 PTE and Steintype estimators
For the linear multiple regression model, , if we suspect the full hypothesis to be (nullvector), then the restricted estimator (RE) and the test for , vs may be based on the statistic
(2.8) 
where
(2.9) 
Under the conditions

, and

where is the th row of , –central chisquare variable with degrees of freedom (df).
Let be an upper level critical value from this null distribution; then we may define the preliminary test estimator (PTE) of as
(2.10) 
The PTE is a discrete variable. As a result some optimality properties when we consider assessing its MSE comparison is lost. We may define a continuous version of PTE as the JamesSteintype estimator (JSE) given by
(2.11) 
Note that we have replaced by in the definition of PTE. However, has an inherent problem of changing its sign due to the factor which may be larger than 1 in absolute value. If that happens, from applied point of view, its interpretation becomes blurred. Thus, we define another estimator, namely, the positiverule Steintype estimator (PRSE) as
(2.12) 
Next, we define an improved preliminary test (IPT) estimator defined by
(2.13) 
Thus, to set the stage, we have defined six estimators, namely, LSE, RE, PTE, JSE, PRSE, and IPT here, and one penalty estimator LASSO. Next, we use the definitions above to propose new shrinkagetype LASSO estimators.
2.2 Proposed improved estimators of LASSO
Let us redefine the LASSO estimator (LE) (Tibshirani, 1996) as
(2.14) 
From now on, we consider as our unrestricted estimator (UE). Then, similar to the definition of PTE, we define the preliminary test LASSO estimator (PTLE) as
(2.15) 
The Steintype shrinkage LASSO estimator (SLE) based on may be defined as
(2.16) 
where was defined in (2.8). Now we define the Steintype positive rule shrinkage LASSO estimator (PSLE) using and as
(2.17) 
We note that follows a noncentral distribution with noncentrality parameter under local alternatives.
2.3 Orthogonal case,
In this section, we consider the model (2.1) and assume that the design matrix is centered and . Under this condition, the LASSO estimator is given by
(2.18) 
where for .
It is known from Donoho and Johnstone (1994) as , the quadratic risk bound is given by
where under the local alternative
For the sparse solution, one has to use as such equals since we have coefficients satisfy and remaining equal to zero.
Now, we consider the PTE, SE and PRSE as defined in section (2.1) where we take . Thus, one can find that the asymptotic riskbound of the PTE, SE and PRSE are then given by
Here,
is the cdf of a noncentral chisquare distribution with
df and noncentrality parameter and2.4 Analysis of asymptotic MSE of the estimators
First, we note that uniformly in . Next, we compare and by taking the riskdifference
whenever
Note
or, 
Accordingly, PTLE is better than LE in the range of . Otherwise, for
LE is better than PTLE. Next, we note that for all , the riskdifferences
and riskdifference
Hence,
3 Simulation for orthogonal case
For the nonorthogonal case, we conduct Monte Carlo simulation experiments to study the performance of the proposed positiverule shrinkageLASSO estimator (PSLE) along with preliminary test LASSO estimator (PTLE), and shrinkage LASSO estimator (SLE). In particular, we study relative efficiencies of the proposed estimators compared to the LASSO estimator (LE) by Tibshirani (1996). In the simulation studies, mean squared errors (MSE) were computed for each of the proposed estimators, and their relative efficiencies were calculated by taking the ratio of MSE of the proposed estimators to the MSE of LE. Raheem et al. (2012)
have conducted similar studies where a subhypothesis was tested and relative efficiencies of various shrinkage and penalty estimators were studied in a partially linear regression setup. In this study, we are concerned with full model hypothesis
against the alternative . Saleh and Raheem (2015) have studied the performance of various shrinkage and penalty estimators under a full model hypothesis using the same setup. Next, we discuss the simulation setup.First, We generate the design matrix
from a multivariate normal distribution with mean vector
and covariance matrix . The offdiagonal elements of the covariance matrix are considered to be equal to with . We consider sample size to be , and the number of parameters, , equal to 10, 20, and 50. In our setup, is a vector and a function of . In the simulation, the vector is defined such that aindicates a data set being generated under null hypothesis, whereas
indicates a data set generated under alternative hypothesis. We considered 23 different values for , which are 0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1, 1.5, 2, 3, 5, 10, 15, 20, 25, 30, 35, 40, and 50. Each realization was repeated 1000 times to obtain biassquared and variance of the estimated regression parameters. Subsequently, MSEs were calculated for the least squared estimator (LSE), improved preliminary test estimator (IPT) which is based on LSE, SLE, and PSLE (READ/CHECK THESE LINES AGAIN). The responses were simulated from the following model:where with two different values for .
Secondly, the data generation setup was further modified to accommodate the number of nonzero s in the model. In particular, we partitioned as where indicates number of nonzero s, and indicates zeros–a function of . To translate the above, when and , we would have to generate the response under null hypothesis. When is introduced, e.g., , we would have to generate the response under alternative hypothesis. Clearly, inclusion of acts as a degree of violation of the null hypothesis. As increases, so does the degree of violation of the null hypothesis. We study the performance of the proposed estimators under varying degree of violation of null hypothesis as measured by .
Finally, the relative efficiencies were calculated using the following formula.
(3.1) 
where is the LASSO estimator, and is one of the estimators whose relative efficiency is to be computed. As in equation (3.1), a relative efficiency greater than 1 would indicate superiority of the proposed estimator compared to the LASSO estimator. On the other hand, a relative efficiency of equal to or less than 1 would indicate that the efficiency of the estimator is at or below that of the LASSO estimator. We used R statistical software (R Core Team, 2014) to carry out the simulation. For obtaining LASSO estimate, glmnet() R package (Friedman et al., 2014) was used.
In the following, we discuss the results of our simulation studies.
3.1 Discussion of simulation results
In this study, data have been generated with correlation between the ’s, for , , and the error variance . The relative efficiencies of the proposed estimators are presented in Tables 7 through 24. To visually compare the results of various configurations, relative efficiency of the estimators are compared to LASSO estimator as shown in Figures 10 through 12. Since positiverule shrinkageLASSO estimator (PSLE) outperforms all other estimators for most of the configurations, we separately compared its performance for and at various correlation coefficient , and the results are displayed in Figures 7 through 9. In all of these figures, a horizontal line was drawn at 1 on the axis to facilitate the comparison among the estimators. For a given estimator, any point above this line indicates superiority of the estimator compared to the LASSO estimator in terms of relative efficiency.
The findings of simulation studies may be summarized as follows.

LASSO dominates over LSE uniformly.

At any given and , relative efficiency of the estimators is a decreasing function of both and . As and increases, the relative efficiency decrease.

Gain in relative efficiency of the estimators is a decreasing function of the number of s that are zero as indicated by . Figures 10, 11, and 12 show that the relative efficiency of SLE and PSLE are the highest when (where relative efficiency is above 9). For the relative efficiency is around 6. The relative efficiency is around 4 for 3 and 5.

PTLE dominates IPT uniformly. However, IPT dominates the least squares estimator, which is consistent with the results found in literature.

Neither PTLE nor Steintype LASSO estimator dominate each other.
4 Real data examples
In the following, we study three real life examples. We preprocess the data sets by centering the predictor variables. We then fit linear regression models to predict the variable of interest using the available regressors. Lasso estimator (LE), improved preliminary test (IPT), Steintype shrinkage LASSO (SLE), and Steintype positiverule shrinkage LASSO (PSLE) estimators are then obtained for the regression parameters.
We evaluate the performance of the estimators by computing average cross validation error using fold cross validation. In cross validation, the data set is randomly divided into subsets of roughly equal size. One subset is left aside, termed as test set, while the remaining subsets, called training set, are used to fit the model. The fitted model is then used to predict the responses for the test data set. Finally, prediction errors are obtained by taking the squared deviation of the observed and predicted values in the test set.
In cross validation, the estimated prediction error varies across runs. Therefore, we repeat the process 1000 times, and calculate the average and standard deviation of the prediction errors. We found 1000 to be large enough number of runs to stabilize the standard deviations as no noticeable changes were observed for larger values.
4.1 Galapagos data
Faraway (2002) analyzed the data about species diversity on the Galapagos islands. The Galapagos data contains 30 rows and seven variables. Each row represents an island, and the covariates represent various geographic measurements. The covariates are: the number of endemic species, the area of the island (km), the highest elevation of the island (m), the distance from the nearest island (km), the distance from Santa Cruz island (km), the area of the adjacent island (km
). The original data set contained missing values for some of the covariates, which have been imputed by
Faraway (2002)for convenience. The response variable is the number of species of tortoise found on the island.
The summary statistics in shown in Table 1. The visual correlation matrix for the centered covariates is shown in Figure 1.
Min  Q1  Median  Mean  Q3  Max  SD (Response) 
2.00  13.00  42.00  85.23  96.00  444.00  114.63 
Bias Corrected CVE  

Estimator  Average  SD 
LE  76350.90  15057.76 
PTLE  76350.90  15057.76 
SLE  70569.66  13152.35 
PSLE  70569.66  13152.35 
Figure 2 displays the prediction error of the estimators for 1000 crossvalidated runs. The average and standard deviation of the predictor errors are summarized in Table 2. It is noted here that the prediction errors are unusually large for this data set. The reason is due to the variability present in the original data set (standard deviation for the response variable is 114.63).
We find SLE and PSLE to be performing better than the LE and PTLE. Notably, SLE and PSLE have smaller standard deviations compared to LE and PTLE.
4.2 State data
Faraway (2002) illustrated variable selection methods using the state data set. There are 97 observations (cases) on 9 variables. The variables are: population estimate as of July 1, 1975; per capita income (1974); illiteracy (1970, percent of population); life expectancy in years (196971); murder and nonnegligent manslaughter rate per 100,000 population (1976); percent highschool graduates (1970); mean number of days with minimum temperature 32 degrees (19311960) in capital or large city; and land area in square miles. We consider life expectancy as the response.
Summary statistics for this data set is given in Table 3. Correlation coefficients between the predictors is displayed in Figure 3. We notice moderate to strong correlation present between some of the predictors. In the simulation studies, we have observed that the proposed estimators perform superiorly when the correlation between the predictors is large.
Min  Q1  Median  Mean  Q3  Max  SD (Response) 
67.96  70.12  70.68  70.88  71.89  73.60  1.34 
Bias Corrected CVE  

Estimator  Average  SD 
LE  5786.94  148.94 
PTLE  5530.89  50.70 
SLE  5216.48  195.24 
PSLE  5552.93  81.89 
Table 4 gives averages and standard deviations of the predictor errors for the estimators. For this data, SLE has the smallest average prediction error followed by PSLE. LE has the largest average prediction error. Figure 4 shows the average prediction errors for the estimators visually, which demonstrates smaller yet highly variable prediction errors for the SLE.
4.3 Longley data
Longley data set (Longley, 1967) is a popular macroeconomic data set and is widely used to demonstrate the application of collinear regression. The data has 16 observations on seven variables. The predictors are GNP implicit price deflator, gross national product, the number of people unemployed, number of people in the armed forces, noninstitutionalized population aged 14 and older, the year (time). The response variable is the number of people employed. Table 5 shows data summary while visual correlation matrix for the predictors is shown in Figure 5. Note that the predictors are highly correlated. Table 6 shows SLE to be the best estimator in terms of prediction errors followed by PTLE and PSLE. LE has the largest average prediction error. Although SLE has the smallest prediction error, it has the largest variability. Of the four estimators, PTLE has the smallest variability. The plot of 1000 prediction errors in Figure 5 demonstrates SLE as the best performing estimator.
Min  Q1  Median  Mean  Q3  Max  SD (Response) 
60.17  62.71  65.50  65.32  68.29  70.55  3.51 
Bias Corrected CVE  

Estimator  Average  SD 
LE  5182.43  99.16 
PTLE  4751.56  5.97 
SLE  3163.05  337.86 
PSLE  4759.06  24.25 
5 Discussion and conclusion
In this paper, we proposed improved estimation technique for the LASSO estimator based on Steinrule. In particular, we used LASSO estimator to obtain preliminary test LASSO estimator (PTLE), Steintype shrinkage LASSO estimator (SLE), and Steintype positiverule shrinkage LASSO estimator (PSLE). We studied the performance of the proposed estimators under fullmodel hypothesis when the parameter space is small relative to sample size. Simulation studies have been performed to compare the estimators for various configurations of parameter sizes (), correlation coefficient between the predictors (), and the error variance (). We used relativeMSE criterion to compare the proposed estimators with the classical LASSO estimator. We varied the number of nonzero s, and evaluated the performance of the estimators under varying degree of model misspecification as guided by . We have provided relative efficiencies of the proposed estimators compared to classical LASSO estimator in Tables 7 through 24 as well as graphically for selected configurations as displayed in Figures 7 through 12.
The simulation results demonstrate that the classical LASSO dominates the LSE uniformly while PSLE has the smallest MSE among the proposed estimators. In particular, PSLE uniformly dominates classical LASSO estimator when the error variance is large (, in our setup). Also, neither PTLE nor SLE dominates one another. Relative efficiency of the proposed estimators increases when there are more nearzero parameters present in the model. Performance of the estimators decrease as we deviate from the null model. These results are consistent with the properties of traditional preliminary test and Steintype estimators found in the literature.
We have presented three real life data examples to demonstrate the use of proposed estimators. Average and standard deviations of the prediction errors based on the LE, PTLE, SLE, and PSLE have been obtained and compared. The proposed estimators outperform LASSO estimator in both average prediction error and standard deviation criteria. While PSLE dominated the other proposed estimators in the simulation experiments, the dominance picture was not obvious for the real data examples. As such we conclude that neither of the PTLE, SLE or PSLE may outperform each other in all real life applications. We tried some other data sets where the correlation between the predictors are low to moderately strong, and found that LASSO as well as the proposed estimators perform equally in those cases. Therefore, we conclude that the improved LASSO estimators would find their applications for data sets with moderate to strong correlations among the predictors.
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.10  1.84  2.50  4.06  4.12  0.17  1.24  1.44  2.08  2.09  0.28  1.24  1.21  1.70  1.70  0.38  1.02  1.03  1.45  1.46 
0.1  0.10  2.82  3.04  3.41  3.46  0.16  1.36  1.68  2.10  2.12  0.26  1.06  1.16  1.53  1.53  0.37  1.00  1.04  1.41  1.42 
0.2  0.12  1.79  2.37  3.17  3.18  0.17  1.33  1.59  2.04  2.06  0.27  1.07  1.13  1.55  1.55  0.40  1.06  1.04  1.43  1.43 
0.3  0.11  1.67  1.62  2.56  2.56  0.18  1.35  1.36  1.97  1.98  0.31  1.10  1.09  1.57  1.58  0.39  1.02  1.03  1.41  1.42 
0.4  0.11  1.52  1.78  2.57  2.58  0.19  1.32  1.28  1.88  1.89  0.28  1.07  1.12  1.50  1.50  0.40  1.06  1.04  1.45  1.46 
0.5  0.14  1.50  1.53  2.24  2.26  0.18  1.21  1.34  1.78  1.80  0.31  1.10  1.12  1.52  1.53  0.43  1.06  1.03  1.44  1.44 
0.6  0.15  1.53  1.61  2.38  2.38  0.21  1.25  1.28  1.87  1.87  0.29  1.07  1.07  1.49  1.49  0.40  1.06  1.07  1.42  1.43 
0.7  0.16  1.31  1.45  2.29  2.30  0.19  1.27  1.29  1.77  1.78  0.29  1.01  1.09  1.52  1.52  0.43  1.03  1.01  1.45  1.45 
0.8  0.14  1.59  1.82  2.27  2.28  0.22  1.15  1.20  1.76  1.76  0.33  1.05  1.08  1.51  1.51  0.42  1.02  1.03  1.48  1.48 
0.9  0.16  1.50  1.44  2.04  2.04  0.20  1.14  1.23  1.66  1.67  0.34  1.06  1.10  1.60  1.60  0.43  1.04  1.01  1.46  1.47 
1.0  0.16  1.29  1.37  2.05  2.06  0.20  1.00  1.15  1.68  1.69  0.33  1.05  1.07  1.51  1.52  0.45  1.06  1.05  1.44  1.44 
1.5  0.18  1.24  1.27  1.77  1.78  0.24  1.19  1.25  1.71  1.71  0.37  1.08  1.04  1.46  1.46  0.47  1.06  1.03  1.43  1.43 
2.0  0.18  1.04  1.38  1.64  1.66  0.25  1.02  1.12  1.56  1.56  0.36  1.01  1.04  1.43  1.43  0.50  1.07  1.00  1.42  1.42 
3.0  0.23  1.00  1.14  1.58  1.59  0.28  0.97  1.09  1.51  1.52  0.38  0.93  0.99  1.43  1.43  0.50  1.00  0.99  1.40  1.40 
5.0  0.29  0.79  0.91  1.26  1.27  0.32  0.77  0.93  1.28  1.28  0.46  0.88  0.94  1.31  1.32  0.50  0.90  0.93  1.28  1.28 
10.0  0.36  0.60  0.78  1.12  1.13  0.34  0.55  0.74  1.04  1.04  0.48  0.74  0.87  1.14  1.14  0.59  0.83  0.88  1.19  1.19 
15.0  0.33  0.48  0.71  0.94  0.94  0.39  0.55  0.78  0.98  0.98  0.49  0.67  0.85  1.08  1.08  0.60  0.82  0.91  1.12  1.12 
20.0  0.32  0.44  0.79  0.85  0.85  0.40  0.56  0.89  0.94  0.94  0.50  0.69  0.92  1.01  1.01  0.61  0.81  0.96  1.05  1.05 
25.0  0.37  0.47  0.84  0.86  0.86  0.39  0.50  0.86  0.87  0.87  0.48  0.63  0.92  0.98  0.98  0.57  0.73  0.96  1.00  1.00 
30.0  0.37  0.46  0.90  0.82  0.82  0.39  0.48  0.89  0.82  0.82  0.53  0.65  0.92  0.96  0.96  0.59  0.75  0.96  0.94  0.94 
35.0  0.36  0.46  0.94  0.85  0.85  0.41  0.51  0.96  0.85  0.85  0.48  0.61  0.99  0.96  0.96  0.56  0.71  1.00  0.96  0.96 
40.0  0.36  0.45  1.00  0.79  0.79  0.41  0.52  1.00  0.87  0.87  0.47  0.57  0.98  0.89  0.89  0.59  0.69  0.95  0.94  0.94 
50.0  0.36  0.43  0.97  0.77  0.77  0.36  0.43  1.00  0.81  0.81  0.49  0.58  1.00  0.87  0.87  0.62  0.73  1.00  0.95  0.95 
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.09  1.43  1.89  3.17  3.18  0.11  1.36  1.69  2.66  2.67  0.14  1.39  1.90  2.45  2.46  0.17  1.43  1.40  1.88  1.88 
0.1  0.11  1.68  1.80  3.31  3.34  0.13  1.57  1.60  2.68  2.69  0.14  1.32  1.50  2.09  2.09  0.17  1.38  1.45  2.00  2.00 
0.2  0.12  1.90  1.95  3.17  3.18  0.11  1.70  1.85  2.62  2.64  0.13  1.41  1.64  2.18  2.19  0.19  1.35  1.32  1.90  1.90 
0.3  0.11  1.81  2.17  3.30  3.33  0.12  1.72  2.28  2.75  2.76  0.15  1.26  1.40  2.09  2.10  0.18  1.40  1.39  1.93  1.93 
0.4  0.09  2.39  2.50  3.69  3.72  0.14  1.31  1.44  2.52  2.52  0.17  1.35  1.66  2.18  2.19  0.18  1.21  1.31  1.90  1.91 
0.5  0.13  2.16  2.23  3.19  3.21  0.13  1.64  1.84  2.75  2.77  0.17  1.50  1.52  2.26  2.26  0.17  1.23  1.43  1.96  1.96 
0.6  0.13  2.42  2.22  3.17  3.20  0.10  1.44  2.11  2.61  2.64  0.13  1.32  1.60  2.04  2.04  0.17  1.25  1.22  1.81  1.82 
0.7  0.11  1.84  1.80  3.20  3.23  0.14  1.53  1.57  2.36  2.39  0.15  1.55  1.78  2.12  2.14  0.18  1.09  1.26  1.78  1.79 
0.8  0.13  2.00  2.35  3.11  3.12  0.14  1.63  1.67  2.53  2.55  0.14  1.42  1.67  2.12  2.13  0.18  1.07  1.26  1.82  1.82 
0.9  0.12  1.68  2.05  3.05  3.07  0.13  1.56  1.65  2.55  2.55  0.16  1.31  1.53  2.03  2.04  0.18  1.35  1.37  1.86  1.87 
1.0  0.15  1.96  1.82  2.74  2.76  0.14  1.49  1.71  2.38  2.38  0.16  1.40  1.45  2.04  2.06  0.21  1.29  1.28  1.87  1.88 
1.5  0.12  1.97  1.64  2.64  2.65  0.14  1.63  1.99  2.44  2.45  0.20  1.31  1.33  2.02  2.02  0.22  1.30  1.27  1.86  1.87 
2.0  0.12  1.88  2.07  2.52  2.53  0.12  1.27  1.59  2.16  2.17  0.15  1.28  1.40  1.80  1.81  0.19  1.21  1.29  1.75  1.76 
3.0  0.16  1.42  1.50  2.18  2.18  0.16  1.30  1.41  2.06  2.07  0.18  1.28  1.39  1.80  1.81  0.19  1.21  1.34  1.69  1.70 
5.0  0.17  1.37  1.39  1.87  1.88  0.18  1.08  1.26  1.83  1.84  0.20  1.17  1.34  1.72  1.73  0.25  1.18  1.17  1.62  1.63 
10.0  0.23  1.04  1.16  1.61  1.62  0.26  0.96  1.12  1.60  1.61  0.27  1.11  1.13  1.55  1.55  0.32  1.06  1.03  1.53  1.53 
15.0  0.28  0.95  1.03  1.51  1.52  0.28  0.90  0.98  1.42  1.43  0.31  0.97  0.99  1.43  1.44  0.35  0.97  0.99  1.47  1.47 
20.0  0.31  0.83  0.95  1.36  1.36  0.29  0.80  0.93  1.37  1.37  0.32  0.78  0.92  1.28  1.28  0.39  0.92  0.94  1.39  1.40 
25.0  0.31  0.77  0.87  1.26  1.26  0.31  0.71  0.84  1.18  1.18  0.34  0.73  0.84  1.23  1.23  0.40  0.79  0.85  1.25  1.25 
30.0  0.35  0.69  0.78  1.16  1.16  0.36  0.75  0.83  1.20  1.20  0.38  0.78  0.89  1.30  1.30  0.42  0.79  0.87  1.26  1.26 
35.0  0.32  0.60  0.78  1.12  1.13  0.32  0.57  0.73  1.04  1.05  0.39  0.66  0.79  1.13  1.14  0.42  0.74  0.84  1.18  1.18 
40.0  0.33  0.57  0.72  1.06  1.06  0.34  0.59  0.76  1.07  1.07  0.41  0.67  0.83  1.13  1.13  0.40  0.67  0.83  1.11  1.11 
50.0  0.38  0.58  0.74  1.04  1.04  0.39  0.55  0.70  0.99  0.99  0.38  0.57  0.75  1.01  1.01  0.43  0.62  0.79  1.05  1.05 
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.06  1.33  1.81  5.18  5.21  0.07  1.41  1.73  2.63  2.66  0.14  1.05  1.22  1.90  1.91  0.17  1.07  1.12  1.58  1.59 
0.1  0.05  1.40  1.96  4.58  4.64  0.08  1.37  1.70  2.65  2.68  0.13  0.95  1.18  1.80  1.80  0.16  0.96  1.10  1.49  1.49 
0.2  0.06  2.25  2.94  5.17  5.23  0.08  1.12  1.37  2.41  2.43  0.13  1.04  1.21  1.71  1.72  0.18  0.96  1.07  1.66  1.66 
0.3  0.07  1.16  1.57  3.17  3.22  0.08  1.23  1.37  2.35  2.36  0.15  1.09  1.14  1.84  1.85  0.19  1.04  1.14  1.61  1.61 
0.4  0.07  1.87  2.35  3.93  3.96  0.10  1.22  1.42  2.47  2.49  0.13  1.09  1.24  1.79  1.80  0.18  0.95  1.07  1.59  1.60 
0.5  0.07  1.46  1.56  3.25  3.29  0.10  1.10  1.36  2.39  2.40  0.15  1.08  1.18  1.75  1.76  0.19  1.00  1.08  1.60  1.61 
0.6  0.07  1.35  1.48  3.06  3.10  0.09  1.43  1.62  2.35  2.36  0.15  1.21  1.29  1.86  1.87  0.19  0.99  1.10  1.61  1.61 
0.7  0.07  1.12  1.46  2.83  2.84  0.09  1.32  1.55  2.09  2.11  0.15  1.13  1.25  1.72  1.73  0.18  0.98  1.10  1.55  1.56 
0.8  0.05  1.44  2.03  2.65  2.67  0.10  1.14  1.32  2.02  2.03  0.16  1.03  1.16  1.71  1.71  0.19  1.01  1.11  1.52  1.53 
0.9  0.08  1.16  1.47  2.73  2.76  0.08  1.13  1.35  1.88  1.89  0.16  1.16  1.24  1.74  1.74  0.19  1.07  1.10  1.53  1.53 
1.0  0.08  1.32  1.89  2.76  2.83  0.11  1.28  1.46  2.01  2.03  0.15  1.01  1.14  1.75  1.75  0.21  0.99  1.10  1.57  1.57 
1.5  0.08  1.03  1.29  2.04  2.06  0.11  1.14  1.24  1.93  1.94  0.15  0.98  1.14  1.57  1.58  0.20  0.96  1.08  1.48  1.48 
2.0  0.13  1.20  1.23  2.17  2.18  0.13  1.06  1.17  1.71  1.74  0.17  1.03  1.11  1.56  1.57  0.22  0.95  1.06  1.49  1.50 
3.0  0.12  1.02  1.17  1.76  1.77  0.14  1.02  1.13  1.61  1.62  0.19  0.98  1.13  1.55  1.56  0.22  0.94  1.03  1.40  1.40 
5.0  0.17  0.86  0.99  1.63  1.65  0.16  0.80  0.95  1.36  1.38  0.20  0.85  0.97  1.36  1.36  0.27  0.93  0.98  1.40  1.40 
10.0  0.18  0.58  0.76  1.07  1.08  0.20  0.61  0.79  1.09  1.09  0.24  0.66  0.84  1.14  1.15  0.29  0.75  0.89  1.14  1.14 
15.0  0.18  0.45  0.70  0.91  0.91  0.20  0.50  0.76  0.98  0.98  0.27  0.63  0.83  1.06  1.06  0.30  0.62  0.81  1.04  1.04 
20.0  0.18  0.39  0.66  0.78  0.78  0.21  0.42  0.71  0.86  0.86  0.27  0.52  0.79  0.92  0.92  0.29  0.57  0.86  0.97  0.97 
25.0  0.18  0.36  0.76  0.73  0.73  0.21  0.41  0.79  0.82  0.83  0.27  0.50  0.84  0.91  0.91  0.29  0.53  0.91  0.93  0.93 
30.0  0.18  0.33  0.78  0.70  0.70  0.21  0.37  0.84  0.74  0.74  0.26  0.45  0.86  0.84  0.84  0.29  0.51  0.95  0.84  0.84 
35.0  0.20  0.33  0.80  0.69  0.69  0.20  0.34  0.92  0.72  0.72  0.23  0.38  0.93  0.74  0.74  0.31  0.52  0.96  0.90  0.90 
40.0  0.17  0.29  0.92  0.68  0.68  0.22  0.36  0.94  0.75  0.75  0.24  0.38  0.93  0.76  0.76  0.29  0.46  0.98  0.84  0.84 
50.0  0.17  0.27  0.93  0.65  0.65  0.19  0.29  1.00  0.67  0.67  0.25  0.38  0.96  0.77  0.77  0.32  0.47  0.96  0.84  0.84 
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.07  1.75  2.13  5.17  5.26  0.06  1.09  1.86  4.01  4.23  0.06  1.40  1.56  2.69  2.73  0.08  1.05  1.42  2.36  2.38 
0.1  0.04  1.48  2.53  5.89  6.28  0.06  1.59  2.09  3.55  3.56  0.06  1.65  1.92  2.71  2.72  0.09  1.26  1.30  2.39  2.42 
0.2  0.06  1.57  2.04  4.99  5.03  0.05  1.39  2.26  4.20  4.27  0.07  1.72  1.83  2.80  2.84  0.08  1.06  1.35  2.35  2.37 
0.3  0.04  1.39  2.53  5.85  5.89  0.07  1.42  1.64  3.92  3.98  0.06  1.24  1.71  2.48  2.49  0.09  1.37  1.60  2.48  2.49 
0.4  0.05  1.51  2.63  5.83  6.06  0.07  1.52  1.87  3.60  3.61  0.07  1.17  1.67  2.88  2.91  0.08  1.12  1.57  2.45  2.46 
0.5  0.06  1.75  2.42  5.23  5.28  0.06  1.66  1.83  3.65  3.72  0.07  1.46  1.77  2.84  2.86  0.07  1.06  1.35  1.98  1.99 
0.6  0.05  1.47  2.21  4.55  4.68  0.06  1.08  1.54  3.56  3.60  0.08  1.59  1.61  2.69  2.72  0.09  1.30  1.45  2.26  2.28 
0.7  0.06  1.56  1.64  4.16  4.21  0.05  1.54  2.21  3.72  3.77  0.06  1.14  1.55  2.27  2.31  0.09  1.31  1.42  2.35  2.38 
0.8  0.06  1.65  2.16  4.90  4.96  0.05  1.37  1.87  3.47  3.49  0.07  1.33  1.83  2.70  2.74  0.09  1.29  1.59  2.38  2.40 
0.9  0.04  1.84  3.65  4.83  5.08  0.07  1.15  1.90  3.28  3.32  0.08  1.52  1.92  2.97  3.01  0.09  1.46  1.55  2.35  2.38 
1.0  0.06  1.42  1.73  3.71  3.76  0.06  1.23  1.86  3.16  3.20  0.07  1.28  1.61  2.41  2.42  0.09  1.12  1.40  2.25  2.26 
1.5  0.07  1.43  1.66  3.51  3.57  0.08  1.39  1.69  3.22  3.26  0.06  1.06  1.62  2.27  2.29  0.09  1.06  1.34  2.06  2.08 
2.0  0.07  1.51  1.54  3.05  3.07  0.07  1.31  1.68  2.93  2.97  0.09  1.27  1.40  2.35  2.37  0.09  1.09  1.27  1.94  1.96 
3.0  0.06  1.21  1.69  2.61  2.64  0.08  1.14  1.45  2.60  2.61  0.10  1.30  1.32  2.20  2.21  0.10  1.10  1.21  2.07  2.07 
5.0  0.09  1.48  1.59  2.51  2.53  0.10  1.10  1.32  2.23  2.25  0.10  1.09  1.29  1.95  1.96  0.11  1.13  1.37  1.96  1.98 
10.0  0.09  1.00  1.27  1.75  1.78  0.12  1.08  1.32  1.94  1.96  0.13  1.02  1.12  1.71  1.72  0.15  1.05  1.17  1.73  1.74 
15.0  0.13  0.88  1.10  1.58  1.61  0.13  0.90  1.08  1.52  1.52  0.14  0.95  1.08  1.51  1.53  0.17  1.00  1.05  1.56  1.57 
20.0  0.14  0.87  0.98  1.45  1.46  0.16  0.91  1.04  1.49  1.50  0.18  0.88  1.01  1.57  1.58  0.18  0.89  1.00  1.48  1.49 
25.0  0.14  0.76  0.92  1.33  1.33  0.15  0.72  0.95  1.35  1.36  0.16  0.75  0.94  1.26  1.27  0.21  0.88  0.98  1.42  1.44 
30.0  0.18  0.75  0.89  1.34  1.35  0.17  0.69  0.88  1.29  1.30  0.19  0.70  0.89  1.30  1.30  0.19  0.73  0.92  1.24  1.24 
35.0  0.18  0.67  0.82  1.15  1.16  0.17  0.63  0.81  1.10  1.10  0.19  0.69  0.84  1.15  1.16  0.19  0.66  0.85  1.16  1.17 
40.0  0.19  0.59  0.78  1.23  1.23  0.18  0.62  0.80  1.09  1.10  0.20  0.62  0.82  1.12  1.13  0.21  0.64  0.83  1.16  1.17 
50.0  0.17  0.47  0.67  0.88  0.88  0.18  0.51  0.72  0.99  0.99  0.20  0.54  0.75  1.04  1.04  0.21  0.57  0.79  1.01  1.01 
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.02  0.57  2.24  9.33  10.34  0.02  0.71  1.52  3.61  3.68  0.04  0.70  1.39  2.18  2.20  0.04  0.66  1.20  1.68  1.69 
0.1  0.02  0.62  1.94  7.52  8.26  0.02  0.58  1.44  3.04  3.07  0.04  0.68  1.27  2.13  2.15  0.05  0.71  1.23  1.86  1.87 
0.2  0.02  0.58  1.63  5.86  6.17  0.02  0.65  1.47  2.87  2.99  0.03  0.71  1.32  1.94  1.97  0.05  0.75  1.18  1.73  1.75 
0.3  0.02  0.57  1.75  4.63  4.76  0.03  0.57  1.44  2.93  2.97  0.03  0.66  1.29  1.86  1.88  0.05  0.64  1.16  1.71  1.72 
0.4  0.02  0.94  2.03  5.59  5.84  0.02  0.49  1.46  2.57  2.61  0.04  0.59  1.34  1.98  2.00  0.05  0.76  1.18  1.82  1.82 
0.5  0.02  0.78  2.19  4.93  5.05  0.03  0.92  1.49  3.23  3.28  0.03  0.66  1.24  1.82  1.84  0.05  0.76  1.20  1.77  1.79 
0.6  0.02  0.69  1.98  4.24  4.36  0.02  0.63  1.75  2.44  2.49  0.04  0.79  1.38  2.05  2.07  0.05  0.64  1.10  1.72  1.73 
0.7  0.02  0.43  1.47  3.58  3.67  0.02  0.68  1.36  2.28  2.30  0.03  0.75  1.36  1.84  1.87  0.05  0.63  1.20  1.68  1.68 
0.8  0.02  0.51  1.59  3.63  3.77  0.02  0.71  1.70  2.32  2.35  0.04  0.63  1.36  1.83  1.85  0.05  0.71  1.22  1.67  1.68 
0.9  0.02  0.79  2.14  3.61  3.69  0.03  0.63  1.37  2.31  2.35  0.04  0.64  1.27  1.86  1.87  0.05  0.64  1.15  1.62  1.63 
1.0  0.02  0.69  1.71  3.31  3.36  0.03  0.64  1.28  2.69  2.74  0.04  0.67  1.22  1.71  1.72  0.05  0.63  1.14  1.56  1.57 
1.5  0.02  0.54  1.56  2.49  2.51  0.03  0.74  1.45  2.29  2.31  0.04  0.73  1.28  1.69  1.70  0.05  0.74  1.13  1.59  1.60 
2.0  0.03  0.76  1.50  2.53  2.58  0.03  0.67  1.21  2.07  2.08  0.05  0.72  1.25  1.73  1.74  0.06  0.65  1.14  1.59  1.60 
3.0  0.03  0.70  1.31  2.00  2.04  0.04  0.64  1.27  1.96  1.98  0.05  0.66  1.14  1.69  1.70  0.06  0.70  1.16  1.47  1.48 
5.0  0.04  0.62  1.11  1.57  1.59  0.05  0.65  1.13  1.59  1.60  0.05  0.64  1.09  1.41  1.43  0.07  0.63  1.06  1.40  1.41 
10.0  0.06  0.56  0.88  1.22  1.25  0.05  0.55  0.93  1.12  1.15  0.07  0.53  0.90  1.20  1.22  0.08  0.52  0.91  1.19  1.20 
15.0  0.06  0.40  0.73  0.95  0.97  0.07  0.44  0.79  1.08  1.09  0.08  0.48  0.80  1.06  1.06  0.08  0.50  0.82  1.01  1.02 
20.0  0.06  0.30  0.65  0.80  0.81  0.06  0.32  0.67  0.74  0.76  0.07  0.35  0.68  0.82  0.83  0.08  0.39  0.77  0.92  0.92 
25.0  0.06  0.27  0.64  0.72  0.73  0.07  0.30  0.65  0.74  0.75  0.08  0.34  0.73  0.84  0.85  0.08  0.33  0.76  0.83  0.83 
30.0  0.05  0.21  0.62  0.61  0.62  0.06  0.22  0.66  0.65  0.66  0.07  0.28  0.72  0.71  0.72  0.09  0.31  0.73  0.78  0.79 
35.0  0.06  0.22  0.58  0.58  0.59  0.07  0.24  0.67  0.67  0.67  0.07  0.25  0.65  0.64  0.64  0.09  0.29  0.79  0.78  0.78 
40.0  0.06  0.17  0.60  0.54  0.55  0.06  0.18  0.63  0.56  0.56  0.08  0.24  0.72  0.66  0.66  0.08  0.25  0.78  0.70  0.70 
50.0  0.07  0.18  0.68  0.55  0.55  0.05  0.15  0.69  0.49  0.49  0.08  0.22  0.80  0.63  0.63  0.08  0.22  0.83  0.63  0.63 
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.02  0.50  1.56  6.11  6.32  0.01  0.61  1.79  5.22  5.42  0.02  0.55  1.50  3.68  3.74  0.02  0.67  1.42  2.67  2.75 
0.1  0.02  0.74  1.53  6.18  6.45  0.02  0.50  2.50  6.17  6.36  0.02  0.71  1.95  4.01  4.12  0.02  0.53  1.59  2.75  2.82 
0.2  0.02  0.74  2.09  9.49  9.72  0.02  0.65  2.73  6.12  6.26  0.02  0.47  1.61  3.25  3.29  0.03  0.88  1.84  3.04  3.11 
0.3  0.02  0.67  1.76  7.36  7.74  0.01  0.40  1.90  5.15  5.38  0.02  0.68  1.69  3.72  3.84  0.02  0.73  1.73  2.94  3.02 
0.4  0.02  0.46  1.45  5.86  5.97  0.02  0.58  1.88  4.64  4.76  0.01  0.49  1.73  3.10  3.18  0.03  0.87  1.49  3.38  3.43 
0.5  0.01  0.61  2.19  6.87  7.68  0.02  0.56  1.54  4.56  4.71  0.02  0.71  1.83  3.33  3.44  0.02  0.74  1.56  2.84  2.90 
0.6  0.02  0.59  1.68  5.96  6.21  0.02  0.75  1.89  4.26  4.48  0.03  0.62  1.76  3.83  3.94  0.02  0.50  1.51  2.82  2.84 
0.7  0.02  0.75  1.82  5.28  5.61  0.02  0.74  1.99  5.19  5.50  0.02  0.49  1.47  2.89  3.01  0.03  0.61  1.48  2.98  3.01 
0.8  0.02  0.59  2.27  6.12  6.60  0.02  0.69  2.37  5.38  5.53  0.02  0.69  1.65  3.56  3.71  0.02  0.60  1.51  2.67  2.68 
0.9  0.02  0.80  4.01  7.71  8.17  0.02  0.85  1.94  4.64  4.86  0.02  0.49  1.50  3.22  3.27  0.02  0.71  1.66  2.80  2.84 
1.0  0.02  0.49  1.56  5.71  5.87  0.02  0.58  1.64  4.45  4.60  0.02  0.55  1.54  3.17  3.26  0.03  0.68  1.44  2.79  2.81 
1.5  0.02  0.65  2.13  4.96  5.08  0.02  0.56  1.61  3.96  4.00  0.03  0.80  1.56  3.27  3.34  0.03  0.58  1.30  2.67  2.69 
2.0  0.02  0.72  1.43  4.07  4.14  0.02  0.51  1.61  3.33  3.42  0.02  0.63  1.45  3.02  3.05  0.03  0.65  1.52  2.69  2.72 
3.0  0.02  0.62  1.49  3.96  4.06  0.02  0.67  1.58  3.19  3.25  0.02  0.63  1.66  2.70  2.78  0.04  0.71  1.45  2.87  2.91 
5.0  0.02  0.68  1.44  2.55  2.61  0.02  0.63  1.70  2.37  2.41  0.03  0.68  1.43  2.60  2.63  0.03  0.72  1.39  2.33  2.37 
10.0  0.03  0.76  1.29  2.36  2.39  0.03  0.60  1.33  1.99  2.06  0.04  0.70  1.45  2.03  2.08  0.04  0.62  1.29  1.74  1.77 
15.0  0.03  0.55  1.20  1.71  1.73  0.04  0.64  1.15  1.82  1.84  0.04  0.53  1.16  1.69  1.71  0.04  0.66  1.15  1.60  1.63 
20.0  0.04  0.61  1.19  1.64  1.67  0.04  0.65  1.13  1.61  1.62  0.05  0.66  1.13  1.68  1.69  0.04  0.59  1.07  1.41  1.43 
25.0  0.04  0.60  1.11  1.45  1.49  0.05  0.66  1.07  1.49  1.53  0.05  0.59  1.01  1.45  1.47  0.05  0.54  1.04  1.52  1.53 
30.0  0.05  0.49  0.95  1.29  1.33  0.05  0.64  1.03  1.43  1.44  0.05  0.54  0.99  1.38  1.40  0.06  0.54  0.98  1.38  1.40 
35.0  0.05  0.53  0.94  1.22  1.26  0.05  0.47  0.91  1.16  1.18  0.05  0.47  0.92  1.19  1.21  0.05  0.59  0.95  1.16  1.19 
40.0  0.05  0.43  0.82  1.05  1.09  0.05  0.42  0.87  1.12  1.14  0.05  0.50  0.90  1.12  1.13  0.06  0.55  0.91  1.16  1.18 
50.0  0.06  0.50  0.84  1.18  1.19  0.06  0.42  0.83  1.05  1.07  0.07  0.47  0.86  1.20  1.22  0.06  0.44  0.79  0.96  0.98 
LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  LSE  IPT  PTLE  SLE  PSLE  

0.0  0.08  1.98  2.23  3.66  3.67  0.14  1.27  1.54  2.30  2.31  0.22  0.83  1.12  1.79  1.79  0.32  0.67  1.01  1.60  1.60 
0.1  0.09  1.49  1.86  2.94  2.99  0.14  1.20  1.43  2.26  2.27  0.24  0.83  1.11  1.81  1.81  0.33  0.64  1.01  1.55  1.55 
0.2  0.10  1.75  1.90  2.86  2.87  0.13  1.16  1.47  2.25  2.26  0.24  0.79  1.10  1.76  1.77  0.31  0.63  1.00  1.56  1.56 
0.3  0.10  1.73  2.24  3.02  3.03  0.15  1.21  1.46  2.20  2.21  0.25  0.84  1.10  1.78  1.78  0.34  0.64  0.99  1.54  1.54 
0.4  0.11  1.52  1.75  2.73  2.74  0.16  1.17  1.45  2.25  2.26  0.26  0.78  1.08  1.76  1.76  0.33  0.63  1.00  1.54  1.54 
0.5  0.11  1.36  1.67  2.65  2.66  0.15  1.07  1.35  2.06  2.07  0.25  0.77  1.10  1.76  1.76  0.34  0.65  1.01  1.56  1.56 
0.6  0.10  1.35  1.69  2.53  2.54  0.16  1.11  1.44  2.07  2.08  0.24  0.66  1.06  1.70  1.71  0.35  0.65  1.00  1.54  1.54 
0.7  0.14  1.38  1.58  2.50  2.51  0.15  1.01  1.35  2.00  2.01  0.27  0.75  1.07  1.71  1.71  0.35  0.63  1.01  1.52  1.52 
0.8  0.13  1.36  1.60  2.43  2.44  0.16  1.04  1.35  1.97  1.98  0.28  0.76  1.06  1.69  1.69  0.35  0.64  1.00  1.53  1.53 
0.9  0.12  1.28  1.46  2.25  2.26  0.18  1.08  1.31  2.06  2.07  0.26  0.71  1.06  1.69  1.69  0.36  0.63  1.00  1.50  1.50 
1.0  0.15  1.21  1.34  2.16  2.17  0.18  1.03  1.28  1.92  1.93  0.27  0.72  1.04  1.67  1.68  0.35  0.62  0.99  1.48  1.48 
1.5  0.16  1.20  1.36  2.14  2.14  0.20  0.99  1.20  1.88  1.88  0.27  0.69  1.04  1.62  1.62  0.39  0.67  0.99  1.47  1.48 
2.0  0.16  1.03  1.22  1.82  1.84  0.19  0.90  1.22  1.82  1.83  0.28  0.67  1.02  1.60  1.60  0.39  0.66  1.00  1.45  1.45 
3.0  0.18  0.95  1.21  1.75  1.76  0.23  0.83  1.08  1.66  1.66  0.31  0.68  0.99  1.54  1.54  0.41  0.66  1.00  1.41  1.41 
5.0  0.21  0.79  1.05  1.51  1.52  0.25  0.69  0.98  1.47  1.47  0.36  0.67  0.97  1.44  1.44  0.41  0.62  0.99  1.35  1.35 
10.0  0.26  0.58  0.86  1.23  1.23  0.30  0.57  0.87  1.25  1.25  0.38  0.60  0.98  1.30  1.30  0.43  0.59  1.00  1.24  1.24 
15.0  0.29  0.51  0.82  1.11  1.11  0.32  0.52  0.92  1.18  1.18  0.39  0.57  0.98  1.21  1.21  0.45  0.59  1.00  1.19  1.19 
20.0  0.29  0.46  0.85  1.03  1.03  0.32  0.49  0.91  1.09  1.09  0.37  0.50  0.99  1.14  1.14  0.45  0.57  1.00  1.16  1.16 
25.0  0.29  0.43  0.89  0.99  0.99  0.30  0.43  0.93  1.02  1.02  0.38  0.50  0.99  1.11  1.11  0.46  0.57  1.00  1.14  1.14 
30.0  0.30  0.43  0.94  0.98  0.98  0.34  0.46  0.97  1.03  1.03  0.40  0.51  1.00  1.09  1.09  0.45  0.54  1.00  1.11  1.11 
35.0  0.30  0.41  0.98  0.96  0.96  0.32  0.42  0.99  1.00  1.00  0.37  0.46  1.00  1.07  1.07  0.44  0.53  1.00  1.09  1.09 
40.0  0.29  0.38  0.97  0.93  0.93  0.31  0.39  0.99  0.96  0.96  0.39  0.48  1.00  1.06  1.06  0.44  0.52  1.00  1.08  1.08 
50.0  0.31  0.40  1.00  0.93  0.93  0.32  0.39  1.00  0.93  0.93  0.38  0.46  1.00  1.04  1.04  0.46  0.53  1.00  1.07  1.07 
Comments
There are no comments yet.