1 Introduction
Anderson acceleration (AA) is a method employed to accelerate the convergence of the fixed point (FP) method for solving systems of nonlinear equations [2, 24, 16]. The formulation of this method leading to the most efficient implementation requires the solution of an unconstrained minimization problem, generally done through solving a least squares problem (LSP) via factorization [24, 7]. Because the LSP is highly sensitive, it is imperative to employ a stable algorithm, such as modified GramSchmidt, to update the factorization at each iteration. The application of modified GramSchmidt generates the main bottleneck of the algorithm when performed in a distributed memory parallel environment, due to the associated high communication costs of performing multiple dot products in sequence [9].
A solution to the costly updates within AA is to apply an algorithm that requires fewer synchronization points per iteration, i.e. performs multiple dot products simultaneously or computes an equivalent update at a lower parallel cost. Recent research by Świrydowicz et al. [23] and Bielich et al. [4] introduced low synchronization algorithms for MGS, CGS2 and GMRES [22], and because AA is equivalent to GMRES when solving systems of linear equations [24], it is natural to consider leveraging these advances in AA.
We have implemented and tested three low synchronization algorithms within the SUNDIALS KINSOL [14] implementation of AA. These methods reduce the number of global reductions to a constant number per iteration independent of the size of the AA iteration space. While these methods can reduce the number of synchronizations to 1 within Krylov methods, a single reduction is not possible within AA due to the requirement for normalization before the LSP solve. Nevertheless, these new methods are able to reduce required global communications to 2 or 3 per iteration.
In addition to demonstrating how to extend these ideas to AA, this paper also includes a problem independent performance study that illustrates the improved strong scalability of the new kernels, particularly for large iteration spaces and processor counts. The relative cost of the new methods is dependent upon on particular problem. Thus, we present convergence and timing data for multiple examples to demonstrate the reliability of the algorithms for AA and improved performance at scale.
The rest of the paper is organized as follows. In Section 2, we describe the AA algorithm and relevant components in the factorization process. In Section 3, we present and compare the low synchronization factorization variants. In Sections 5 and 4 we give performance results for the different methods employed within the AA algorithm itself and from several test problems respectively. Finally, Section 6 provides conclusions based on our findings.
2 Background
2.1 Anderson Acceleration
AA (Algorithm 1) utilizes a linear combination of prior function evaluations to accelerate the convergence of the FP iteration . The weights minimize the FP residual norm in the linear case and are computed by solving a LSP (creftype 6 in Algorithm 1) via a factorization of where is a matrix with linearly independent, normalized columns, and is an upper triangular matrix.
Input: , , tol, and maxIters. Output: 1: and 2:for do 3: Set and 4: Set with 5: Set with 6: Determine that solvesSolving the LSP (Algorithm 2) requires updating the
decomposition to incorporate the vector
. This update is comprised of two subroutines: QRAdd appends a vector to the factorization while performing the necessary orthogonalization and QRDelete removes the oldest vector from the factorization. Input: , , , , , , , and . Output: , , and 1:if then 2: Set and 3:else 4: if then 5: QRDelete(, , ) 6: end if 7: QRAdd(, , , ) 8:end if 9:SolveA key factor in the performance and convergence of AA is the number of residual history vectors or depth, . The depth determines the maximum size of the factorization and thus the number of vectors that must be orthogonalized each iteration in QRAdd. Note in the startup phase (), the number of vectors increases by one each iteration until vectors are retained. Once the history is full, the subsequent iterations require orthogonalization against vectors. As discussed below, the number of synchronizations required by QRAdd depends on and the method for updating the factorization. Generally, QRDelete does not require any communication, as initially noted by Loffeld and Woodward in [17]. In practice the depth is often kept small () however, there are potential convergence benefits to be gained if it were economical, performancewise, to run with larger .
2.2 QR Factorization with GramSchmidt
There exist several methods for computing the factorization of a matrix. In this paper, we consider methods derived from the GramSchmidt procedure. Specifically, we focus on the process of updating a given factorization to include a vector . Here, columnoriented algorithms are presented as opposed to those that proceed rowwise. These are better suited to a lowsynchronization formulation and promote datalocality for cachememory access both for CPUs and GPUs.
The classical GramSchmidt (CGS) process updates a given factorization with the vector by applying the projection to then normalizing . While this process is attractive in a parallel computing environment, due to the fact that can be applied with a single reduction, this algorithm is highly unstable and exhibits a loss of orthogonality dependent on [11, 10] where is machine precision and is the square of the condition number of . For this reason alone, it is unsuitable as a updating scheme in AA, where the LSP solver convergence may be highly sensitive to the loss of orthogonality [19].
A common solution for the instability of CGS is to apply modified GramSchmidt (MGS) instead. This process applies the rank–1 elementary projection matrices , , one for each of the linearly independent columns of , to the vector then normalizes
. Performing the GramSchmidt process in this way effectively reduces the loss of orthogonality to
[5]. However, applying the projections sequentially requires dot products, one for each of the columns, resulting in communication bottlenecks when performed in a parallel distributed environment (discussed more thoroughly in Section 3.1).New variants of CGS and MGS have been introduced to mitigate instability and parallel performance issues caused by these methods. In most cases, they introduce a correction matrix , leading to a projection operator of the form . In the case of CGS, CGS2 (classical GramSchmidt with reorthogonalization) corrects the projection by reorthogonalizing the vectors of and thereby reduces the loss of orthogonality to [10, 20]. The form of the correction matrix for this algorithm was derived in [23] and is discussed further in Sections 3.4 and 3.3. The inverse compact MGS representation was recently introduced, requiring a triangular solve instead of the recursive construction of the correction matrix from a blocktriangular inverse (presented in Section 3.2). The required number of dot products per iteration is effectively reduced to one for GMRES and s–step Krylov solvers [23, 25]. The inverse compact algorithm maintains loss of orthogonality.
3 QR Update Methods in AA
In this section, we discuss the QRAdd kernel, creftype 7 of Algorithm 2, employed to update the AA iteration space with the vector. We present the baseline QRAdd_MGS (modified Gram Schmidt) algorithm alongside three low synchronization variants of the kernel implemented within AA in SUNDIALS. For each of the QRAdd kernels, we discuss the form of the projectors applied to orthogonalize the given set of vectors, , as well as, their predicted parallel performance.
Throughout this section and the remainder of the paper, matrix entries are referenced via subscripts with 0based indexing. For example, refers to the first row, first column of a matrix , and slices of a matrix are inclusive, i.e. refers to all rows of the matrix and columns .
3.1 Modified Gram Schmidt
The standard approach for updating the factorization within AA is to apply MGS, outlined in Algorithm 3.
1.15 Input: , , and Output: , 1:for do 2: Sync 3: 4:end for 5: Sync 6:In each AA iteration, applying MGS requires dot products; for the orthogonalization against all previous vectors in the AA iteration space, and one for the normalization at the end of the algorithm. This algorithm results in synchronizations across all processes, as the reductions form a chain of dependencies and can not be performed in tandem. The high costs of these synchronizations are exacerbated by the lack of computational workload between each reduction, namely the entire algorithm only requires flops.
3.2 ICWY Modified Gram Schmidt
Lowsynch onereduce GramSchmidt algorithms are based upon two key ideas. First, the compact representation relies on a triangular correction matrix , which contains a strictly lower triangular matrix . One row or block of rows of is computed at a time in a single global reduction. Each row
is obtained within the current step, then the normalization step is lagged and merged into a single reduction. The associated orthogonal projector is based on Rühe [21] and presented in Świryzdowicz et al. [23].
The implied triangular solve requires an additional flops at iteration and thus leads to a slightly higher operation count compared to the original MGS algorithm. The operation increases ICWYMGS complexity by for an overall complexity of , and reduces synchronizations from at iteration to 1. Only one global reduction is required per iteration, and the amount of interprocess communication does not depend upon the number of rank–1 projections applied at each iteration.
When implementing ICWYMGS in the context of AA, lagging the normalization until the subsequent iteration is not an option, as the factorization is applied immediately after updating in the LSP solve on creftype 6 of Algorithm 2. The resulting update algorithm within AA is detailed in Algorithm 4.
Here, we present a two reduction variant of the ICWYMGS algorithm. The formation of the correction operator and the matrix can be merged on creftype 2. With the inclusion of the normalization at the end of the algorithm, this results in two synchronizations per iteration until the AA iteration space is filled. Once AA reaches creftype 5 in Algorithm 2, the oldest vector in the factorization is deleted. While this can be performed with Givens rotations in cases where only and are stored, the explicit storage and application of requires that this correction matrix be updated by introducing a single reduction. Overall, this process results in two global reduction steps per iteration until is reached, after which, there are three global reductions per iteration.
3.3 Cgs2
The CGS2 algorithm is simply CGS with reorthogonalization, which, unlike CGS, is known to mitigate large cancellation errors and maintain numerical stability [1, 6, 15, 21]. Reorthogonalizing the vectors in updates the associated orthogonal projector to include a correction matrix , although this matrix may not be explicitly formed in practice.
In Appendix 1 of [23], the form of the projection and correction matrices was derived within the context of DCGS2 (discussed in Section 3.4), however, the matrices remain the same and are given by
Here, contains the same information as in the ICWYMGS case, namely each row of is generated by the reorthogonalization of the vectors within before updating the factorization with an additional vector.
Algorithm 5 details the process of updating the AA factorization with an additional vector without explicitly forming the correction matrix. Reorthogonalization happens explicitly on creftypeplural 4 and 3 before being added back into the matrix on creftype 5.
1.15 Input: , , and Output: , 1: Sync 2: 3: Sync 4: 5: 6: Sync 7:Overall, this process requires three reductions to be performed per iteration—two for the orthogonalization and reorthogonalization of against all the vectors in and one for the final normalization of . The amount of computation per iteration has increased to approximately or . While the amount of computation is still on the same order as MGS, the workload has approximately doubled, and, fortunately, there are no additional storage requirements.
3.4 Dcgs2
DCGS2, or CGS with delayed reorthogonalization, is based on the delayed CGS2 algorithm introduced by Hernandez et al. in [13]. As the name suggests, the reorthogonalization of the vectors in CGS2 is delayed to the subsequent iteration. In his original derivation, Hernandez notes that this process is tantamount to updating the factorization with a CGS vector, and a deteriorated loss of orthogonality (between –, that of CGS2 and CGS) is often observed.
A stable variant of DCGS2 was derived by Świryzdowicz et al. [23] and Bielich et al. [4] by exploiting the form of the correction matrix and introducing a normalization lag and delayed reorthogonalization. The symmetric correction matrix is the same as in the context of CGS2, namely
When the matrix is split into the pieces and , then applied across two iterations of the DCGS2 algorithm coupled with lagging the normalization of the added vector, the resulting loss of orthogonality is in practice. Performing DCGS2 in this way decreases the number of reductions to one per iteration because the reorthogonalization is performed “onthefly” and essentially operates a single iteration behind. The authors note in their paper that the final iteration of GMRES requires an additional synchronization due to the required final normalization.
Algorithm 6 details our QRAdd_DCGS2 implementation in which there are two synchronizations per iteration. Because AA requires using the factorization to solve the LSP at the end of every iteration, it is not possible to lag the normalization of the vector .
1.15 Input: , , and Output: , 1: Delayed Sync 2:if then 3: Sync 4: 5: 6:end if 7: 8: Sync 9:The global reduction for the reorthogonalization of the vectors can be still be performed at the same time as the orthogonalization of against . The synchronization on creftype 1 is lagged to coincide with creftype 3. The amount of computation for DCGS2 increases over that of CGS2 except on the first iteration where it performs the same number of flops as MGS. Lagging the reorthogonalization and introducing the operation , makes this method , the same as ICWYMGS.
4 Performance Study
In this section we detail the GPU and CPU performance of the QRAdd variants from Section 3. We differentiate between two performance costs: the time required to fill the AA iteration space, defined as “StartUp Iterations” and the time required to add an additional vector to the space once it is filled, termed “Recycle Iteration”. In the case of CPU performance, a parallel strongscaling study is performed. With GPUs, we perform a weakscaling study, because a common goal is to saturate the devices, and thus avoid underutilization.
Matrices and vectors containing rows are partitioned rowwise across processes when performing strongscaling studies on CPUs. That is, each process (or CPU core) contains contiguous rows of a given matrix or vector. For GPU performance studies, we provide the local vector size per GPU participating in the computation. All tests are performed on the LLNL Lassen supercomputer [12], which has two IBM Power9 CPUs per node each with 20 available cores, and each CPU is connected to two Nvidia V100 GPUs. For GPU performance tests each GPU is paired with a single MPI rank as the host process. Each test is performed 20 times and for 10 iterations (for a total of 200 timings); for each individual test the maximum time required by any single process or GPU is recorded, and the minimum time across all tests is presented.
The results presented are based on the AA implementation within SUNDIALS. Parallel CPU tests are implemented with a socalled nodelocal vector abstraction, called the Parallel N_Vector. Similarly, the GPU tests are implemented with the MPIPlusX N_Vector where the CUDA N_Vector is used as the local vector for each MPI rank [3]. Notably, while we use the CUDA N_Vector as the local vector portion, this can be switched with any N_Vector implementation in SUNDIALS. Due to this abstraction, communication between GPUs is still performed by staging data through the host process and not via CUDAaware technologies.
The low synchronization methods added to SUNDIALS leverage the fused dot product operation to perform matrixvector products with . This operation enables computing the dot product of a single vector with multiple vectors (the columns of ) as a single operation requiring only one MPI call. For ICWY and DCGS2 a new N_Vector operation was introduced in SUNDIALS enabling delayed synchronization by separating the local reduction computation and final global reduction into separate operations. Both of these fused operations perform a onetime copy of the independent vector data pointers into a device buffer accessed by the fused operations. In addition to combining multiple reductions into a single call, these operations reduce the number of kernel launches in the GPU case. We discuss further in Section 4.1.
4.1 StartUp Iterations
We first consider the startup iteration times, displayed in Figures 2 and 1. On the CPU we use a global vector of size , and for the GPU a local vector size of . The number of iterations performed is equivalent to the number of vectors, , in the AA iteration space.
The top rows of Figure 1 (CPU) and Figure 2 (GPU) display the time required to fill the AA iteration space. As part of this operation, QRAdd_MGS requires total synchronizations, while QRAdd_ICWY, QRAdd_CGS2, and QRAdd_DCGS require , , and , respectively.
Predicting performance based solely on the number of dot products performed, we expect that QRAdd_ICWY and QRAdd_DCGS2 will become faster than QRAdd_MGS after and QRAdd_CGS2 will outperform QRAdd_MGS after , for processor counts where the synchronization imposed by dot products is the dominate cost. This is consistent with the CPU performance seen in Figure 1. The benefits of applying low synchronization QRAdd algorithms is not observed until the processor count reaches 1600 or higher when the global reduction costs lead to larger bottlenecks. As increases, the disparity between the time required for QRAdd_MGS and the low synchronization variants greatly increases.
Our observed GPU performance depends on the number of kernel calls for each QRAdd subroutine in addition to the number of global reductions. Compared to QRAdd_MGS, QRAdd_ICWY performs fewer CUDA kernels each iteration for , while QRAdd_CGS2 and QRAdd_DCGS2 perform four additional CUDA kernel calls for . Also, all three low synchronization variants perform sequential onCPU updates to , unlike QRAdd_MGS. Finally, in addition to the kernel launches and sequential updates, the low synchronization routines increase the amount of data being moved between the host and the GPU. Specifically, the fused dot products employed by these routines require that at most values be copied to and from the GPU multiple times per iteration while QRAdd_MGS copies only a single value per individual dot product.
All three low synchronization variants incur the cost of two kernel launches that require data transfers to and from the host process. However, in the case of QRAdd_ICWY, the lower computational requirements result in fewer overall CUDA kernel calls than for QRAdd_CGS2 and QRAdd_DCGS2. Specifically, creftype 2 in Algorithm 5 and creftype 4 in Algorithm 6 correspond to additional vector updates that are not required by QRAdd_ICWY; hence their performance times remain higher than QRAdd_ICWY independent of GPU count or . There are minimal performance gains when using QRAdd_CGS2 or QRAdd_DCGS2 for filling the AA iteration space, as QRAdd_DCGS2 is only slightly faster than QRAdd_MGS when more than 8 GPUs are used and only for and larger.
4.2 Recycle Iterations
We next consider the cost of the QRAdd kernel once the AA iteration space is filled, namely the time required to orthogonalize a single vector against vectors. Assuming a large iteration count is required for convergence, this cost reflects the expected runtime of the QRAdd kernel for the majority of a given solve. We label these times as the “Recycle Iteration” time in Figures 2 and 1. The CPU and GPU timings presented are for the same global and local vector size as presented in Section 4.1, and the results are for a single iteration. Note that this time does not include the additional synchronization introduced by ICWY into QRDelete, first mentioned in Section 3.2 and discussed further in Section 5.
Once the AA iteration space is filled, the per iteration cost is greatly decreased by using one of the low synchronization orthogonalization algorithms when operating on CPUs (see the bottom row of Figure 1). While the “Startup Iterations” did not result in performance gains on CPUs until after 1600 processes, for , we observe performance gains with QRAdd_ICWY with process counts as low as 784. QRAdd_MGS is still faster at smaller scales for all values of due to the reduced synchronization costs of performing an MPIAll_Reduce on a small, closedset number of processes. The performance gains of the low synchronization algorithms at larger scales ranges from 2–8 speedup for , , and at processes. With this drastic speedup for each iteration, we expect much larger performance gains for test problems that run to convergence.
Unlike the CPU performance for AA, QRAdd_ICWY demonstrates performance gains independent of the number of GPUs participating in the computation or the size of the AA iteration space, . This is seen in the bottom row of Figure 2, which displays the performance for a single “Recycle Iteration” on multiple GPUs. While QRAdd_CGS2 and QRAdd_DCGS2 do not see the same performance improvements as QRAdd_ICWY, they do begin exhibiting faster performance than QRAdd_MGS beginning at 8 GPUs and for and larger, though gains are modest.
5 Numerical Experiments
In this section, we highlight the strongscaling parallel efficiency of standard AA compared with AA with low synchronization orthogonalization for test problems run to convergence. The example problems provided are not exhaustive, but the selected tests do stress the performance of low synchronization AA. In addition, application specific tests are used to demonstrate the benefits of low synchronization AA for both distributed CPU and GPU computing environments. All experiments are performed on the LLNL Lassen supercomputer [12], with the same setup as described in Section 4. To account for machine variability, each run is executed 10 times and we report the minimum.
5.1 Anisotropic 2D Heat Equation + Nonlinear Term
This test problem highlights the performance of all AA variants for various and iterations to convergence. We consider a steadystate 2D heat equation with an additional nonlinear term ,
The chosen analytical solution is
hence, the static term is defined as follows
The spatial derivatives are computed using secondorder centered differences, with the data distributed over points on a uniform spatial grid, resulting in a system of equations of size . The Laplacian term is implemented as a matrixvector product giving the the algebraic system as
where denotes the discrete vector of unknowns. Solving for results in the following FP formulation
We use the SUNDIALS PCG solver to solve the linear system with the hypre PFMG preconditioner performing two relaxation sweeps per iteration. Both the FP nonlinear solver and the PCG linear solver set a stopping criteria with a tolerance of . A zero vector is used as the starting guess in all cases.
5.1.1 Nonlinear Term 1
As a first example, consider the nonlinear reaction term
In this case, AA exhibits rapid convergence when , requiring 11 iterations to converge for all variants. Figure 3 displays the overall time to convergence, split into evaluation time and time spent in AA. In general, the performance is volatile, most likely due to the sparse matrix operations required for the linear solve; thus we focus on the AA performance in the bottom of Figure 3.
The time spent in AA meets expectations, based on the results in Section 4. For such a small , we observe minimal performance improvements over AA with MGS, particularly at small process counts. However, the observed improvements begin around 1600 processes.
5.1.2 Nonlinear Term 2
As a second example, consider
With , AA requires more time to converge than for the in Section 5.1.1, better highlighting potential performance benefits of the low synchronization orthogonalization algorithms.
Figure 4 displays the overall timing results, with the number of iterations required to converge listed beneath each bar. In most cases, the number of iterations is between 27 and 42, depending on the orthogonalization method and processor count. The three exceptions are for AA with DCGS2, which requires more iterations to converge, particularly for 1600–6400 processes. This is expected since our QRAdd_DCGS2 cannot lag the normalization and exhibits the same instability as Hernandez’s original algorithm.
In most cases, AA with DCGS2 results in degraded performance in comparison to the other variants due to the high number of iterations required to converge. AA with CGS2 performs best, starting at 1600 processes and continues to be the fastest through processes. Although QRAdd_ICWY only requires two global synchronizations per iteration, there is an additional synchronization in QRDelete to update the matrix , resulting in an implementation that is more comparable to AA with CGS2 at higher process counts.
5.2 2D Bratu Problem
We next consider an example from [24] (and originating from [8]). The Bratu problem is a nonlinear PDE boundary value problem defined as:
We again solve the problem with centered differencing, resulting in a system of the form
The associated FP function is then
We use a uniform grid of points, which results in a system of equations of size . We select for testing, as it is close to the theoretical critical point, as discussed in [18], and is a difficult problem to solve. The SUNDIALS PCG solver is applied to perform the linear system solve with the hypre PFMG preconditioner performing two relaxation sweeps. Both the nonlinear and linear solver employed a tolerance of . A zero vector is set as the starting guess.
For this example problem, , and the number of iterations required to converge for all variants of AA is less than 30; hence the timing results reflect the performance benefits of the QRAdd subroutines for the startup iterations. The strongscaling timing and convergence results are presented in Figure 5. For this test case, we observe improvements with ICWY from the very beginning with 400 processes (although only slightly in this case). AA with ICWY continues to outperform up to 8100 processes. This is consistent with the results in Section 4.1 in which QRAdd_ICWY performed best for as it only requires two dot products per iteration, one fewer than QRAdd_CGS2 and has approximately the same amount of computation as QRAdd_DCGS2.
5.3 ExpectationMaximization Algorithm for Mixture Densities
For this example, we consider a variation of the expectationmaximization test problem presented in
[24]. Consider a mixture density of three univariate normal densities with a mixture density given by , withMixture proportions
are nonnegative and sum to one. The mixture proportions and variances are assumed to be known and the means
are estimated from a set of unlabeled samples
, or samples of unknown origin. Determining the unknown mean distribution parameters is given by the FP functionwith current mean estimations being applied alongside the known mixture proportions and variances to determine the subsequent estimations until convergence. We keep the same mixture proportions and variances as the original test case, and . We generated samples for the mean distribution set , corresponding to a poorly separated mixture, and used the same AA parameter of as Walker and Ni [24].
For our test case, we estimate a single set of mean distribution parameters redundantly for every entry in a global vector, where the vector takes the form
We do this to simulate a function that requires no communication other than that imposed by AA. The resulting FP function to be solved is then given by
Because is small for this test, we expect to see only modest improvements in performance of the low synchronization routines over MGS. For , ICWY and DCGS2 reduce the number of synchronizations per iteration by one over MGS, with ICWY gaining an additional synchronization after the space is filled.
The test is performed as a weakscaling study with each GPU operating on a local vector size of values. Tests were run with a tolerance of , and each AA version requires 21 iterations to converge, independent of the orthogonalization subroutine used.
The results of these tests are presented in Figure 6 and are consistent with the results of the weakscaling study since no communication is required for . The function evaluation performance remains the same, independent of GPU count or AA variant (differing only slightly for runs on a single node with 4 GPUs). In addition, the time spent in is lower for this example since performance depends on the number of samples in the mixture, while the performance of AA scales with the number of values in the global vector.
Results are consistent with observations in the GPU weakscaling study in Section 4 where all of the QRAdd subroutines performed similar for . Additionally, because is small, we do not incur a large overhead for data movement and multiple kernel launches in comparison to those observed with larger values of , as presented in Section 4.2. Because the majority of the computation has moved to the GPU, we observe the expected crossover point for ICWY with its performance being slightly faster than that of MGS for GPU counts of 16 and larger. There are no consistent improvements observed for CGS2 or DCGS2.
6 Conclusions
Anderson Acceleration (AA) is an efficient method for accelerating the convergence of fixed point solvers, but faces performance challenges in parallel distributed computing environments mainly due to the number of global synchronizations per iteration which is dependent upon the size of the AA iteration space. In this paper, we introduced low synchronization orthogonalization subroutines into AA which effectively reduce the number of global synchronizations to a constant number per iteration independent of the size of the AA iteration space. We presented a performance study that demonstrated the improved strongscalability of AA with these low synchronization QRAdd subroutines when performed in a CPUonly parallel environment, as well as demonstrated performance and implementation concerns for these subroutines when operating in a multiGPU computing environment. Furthermore, our numerical results display a realistic picture of the expected performance of AA in practice that matches the predictions of our performance analysis in Section 4 and suggests the use of ICWY for large values of when operating in a CPUonly parallel environment and as the default method for distributed GPU computing. Overall, this paper provides a comprehensive study of low synchronization orthogonalization routines within AA and their parallel performance benefits.
The software used to generate the results in this paper will be available in a future release of SUNDIALS.
Acknowledgments
Support for this work was provided through the Scientific Discovery through Advanced Computing (SciDAC) project “Frameworks, Algorithms and Scalable Technologies for Mathematics (FASTMath),” funded by the U.S. Department of Energy Office of Advanced Scientific Computing Research.
This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DEAC5207NA27344. Lawrence Livermore National Security, LLC. LLNLPROC827159.
This material is based in part upon work supported by the Department of Energy, National Nuclear Security Administration, under Award Number DENA0003963.
References
 [1] N. N. Abdelmalek, Round off error analysis for GramSchmidt method and solution of linear least squares problems, BIT Numerical Mathematics, 11 (1971), pp. 345–367.
 [2] D. G. Anderson, Iterative procedures for nonlinear integral equations, J. Assoc. Comput. Mach., 12 (1965), pp. 547–560, https://doi.org/10.1145/321296.321305.
 [3] C. J. Balos, D. J. Gardner, C. S. Woodward, and D. R. Reynolds, Enabling GPU accelerated computing in the sundials time integration library, Parallel Computing, (2021), p. 102836, https://doi.org/10.1016/j.parco.2021.102836.
 [4] D. Bielich, J. Langou, S. Thomas, K. Świrydowicz, I. Yamazaki, and E. Boman, Lowsynch GramSchmidt with delayed reorthogonalization for Krylov solvers, Parallel Computing, (2021).
 [5] Å. Björck, Solving linear least squares problems by GramSchmidt orthogonalization, BIT Numerical Mathematics, 7 (1967), pp. 1–21.
 [6] J. W. Daniel, W. B. Gragg, L. Kaufman, and G. W. Stewart, Reorthogonalization and stable algorithms for updating the GramSchmidt QR factorization, Mathematics of Computation, 30 (1976), pp. 772–795.
 [7] H. Fang and Y. Saad, Two classes of multisecant methods for nonlinear acceleration, Numer. Linear Algebra Appl., 16 (2009), pp. 197–221.
 [8] D. A. FrankKamenetskii, Diffusion and Heat Exchange in Chemical Kinetics, Princeton University Press, 2016, https://doi.org/10.1515/9781400877195.
 [9] V. Frayssé, L. Giraud, and H. KharrazAroussi, On the influence of the orthogonalization scheme on the parallel performance of GMRES, in European Conference on Parallel Processing, Springer, 1998, pp. 751–762.
 [10] L. Giraud, J. Langou, M. Rozložník, and J. van den Eshof, Rounding error analysis of the classical GramSchmidt orthogonalization process, Numer. Math., 101 (2005), p. 87–100, https://doi.org/10.1007/s0021100506154.
 [11] L. Giraud, J. Langou, and M. Rozloznik, The loss of orthogonality in the GramSchmidt orthogonalization process, Computers & Mathematics with Applications, 50 (2005), pp. 1069–1075, https://doi.org/10.1016/j.camwa.2005.08.009. Numerical Methods and Computational Mechanics.
 [12] W. A. Hanson, The CORAL supercomputer systems, IBM Journal of Research and Development, 64 (2020), pp. 1:1–1:10.
 [13] V. Hernández, J. E. Román, and A. Tomás, A parallel variant of the GramSchmidt process with reorthogonalization, in PARCO, 2005, pp. 221–228.
 [14] A. C. Hindmarsh, R. Serban, C. J. Balos, D. J. Gardner, D. R. Reynolds, and C. S. Woodward, User documentation for KINSOL v5. 7.0 (SUNDIALS v5. 7.0), 2021, https://computing.llnl.gov/projects/sundials/kinsol.
 [15] W. Hoffmann, Iterative algorithms for GramSchmidt orthogonalization, Computing, 41 (1989), pp. 335–348.
 [16] C. T. Kelley, Numerical methods for nonlinear equations, Acta Numerica, 27 (2018), pp. 207–287.
 [17] J. Loffeld and C. S. Woodward, Considerations on the implementation and use of anderson acceleration on distributed memory and GPUbased parallel computers, in Advances in the Mathematical Sciences, Springer, 2016, pp. 417–436.
 [18] A. Mohsen, A simple solution of the Bratu problem, Computers & Mathematics with Applications, 67 (2014), pp. 26–33, https://doi.org/10.1016/j.camwa.2013.10.003.
 [19] C. C. Paige, The effects of loss of orthogonality on large scale numerical computations, in Proceedings of the International Conference on Computational Science and Its Applications, SpringerVerlag, 05 2018, pp. 429–439.

[20]
B. N. Parlett,
The symmetric eigenvalue problem
, SIAM, 1998.  [21] A. Ruhe, Numerical aspects of GramSchmidt orthogonalization of vectors, Linear algebra and its applications, 52 (1983), pp. 591–601.
 [22] Y. Saad and M. H. Schultz, GMRES: A generalized minimal residual algorithm for solving nonsymmetric linear systems, SIAM Journal on scientific and statistical computing, 7 (1986), pp. 856–869.
 [23] K. Świrydowicz, J. Langou, S. Ananthan, U. Yang, and S. Thomas, Low synchronization Gram–Schmidt and generalized minimal residual algorithms, Numerical Linear Algebra with Applications, 28 (2021), p. e2343, https://doi.org/10.1002/nla.2343.
 [24] H. F. Walker and P. Ni, Anderson acceleration for fixedpoint iterations, SIAM Journal on Numerical Analysis, 49 (2011), pp. 1715–1735, https://doi.org/10.1137/10078356X.
 [25] I. Yamazaki, S. Thomas, M. Hoemmen, E. G. Boman, K. Świrydowicz, and J. J. Elliott, Lowsynchronization orthogonalization schemes for sstep and pipelined Krylov solvers in trilinos, in Proceedings of the 2020 SIAM Conference on Parallel Processing for Scientific Computing, SIAM, 2020, pp. 118–128.
Comments
There are no comments yet.