1 Introduction
Consider three Euclidean spaces with . We consider a map from into and a map from to . The direct sum :
(1) 
is a map from to . When the Jacobian of is invertible in a domain of near a root of , and have Jacobians both of full row rank. In that situation we will call a vector Lagrangian and a constraint. The equation
(2) 
covers systems of equations where a number of equations in the system are dependent on a group of variables while the remaining equations involve all variables . The remaining variables are named in honor of Lagrange. The full row rank assumption for ensures that is solvable in (but not always explicitly). For a constrained optimization problem a Lagrangian is a scalar function in the literature and what we call vector Lagrangian here is its differential. Since our focus will be on vector Lagrangian in this article we will drop the qualifier vector for the remain of the article.
Of particular interest is the case when is affine in . Let be a vector function from an open set in to . For each , assume is a linear function from to . Then
(3) 
is a Lagrangian if it has continuous differential and is of full row rank. We will call this special form explicit Lagrangian.
On the constraint side, from our full row rank assumption defines a submanifold of codimension .
This set up covers three classes of equations encountered in the literature:

The eigenvector/invariant subspace problem:
(4) where is a matrix, is a matrix. In this case and are both matrices and is the space of symmetric matrices. The case where is the eigenvector problem.

The constraint optimization problem, one of the most important problems in applied mathematics. Here where is a real value function and . This is the case of the classical Lagrangian multiplier equations. is the domain where is defined and is the target space of the restrictions on . The system (Eq. 1) gives us the set of critical points.

The nonlinear eigenvalue problem:
(5) Here is a matrix with polynomial entries in . While this is not in the form (Eq. 1) we can impose the constraint (or alternatively for a fixed vector ) This equation is not of explicit form. is of dimension one and is a scalar. and are where is the dimension of . There is an extensive literature for this problem ([10]) and NewtonRaphson is an important algorithm.
An iteration method called Rayleigh quotient iteration (RQI) is among the most powerful methods to compute eigenvalues and vectors. For a vector the Rayleigh quotient is
and the iteration computes
which is shown to have cubic converge in for almost all initial points if is normal and quadratic otherwise on suitable initial points. ([1]) gives an extension of Rayleigh quotient iteration for the invariant subspace problem where is a matrix and is a matrix.
The letter
used in the second example in honor of Lagrange is also often used to denote an eigenvalue. The fact that RQI is related to Lagrange multipliers and NewtonRaphson has been known for a long time but it is probably fair to say the relationship is not yet transparent. We will try to clarify the relationship, namely applying NewtonRaphson iteration to (
Eq. 1), the updating equation for involves solving a linear equation similar to . The iterative series of will have form similar to Rayleigh quotient. When the Lagrangian is of explicit form we can use that form to estimate , resulting in a variation of NewtonRaphson extending RQI. We also obtain an iteration when the Lagrangian is implicit.In practice,
could be spaces of matrices or tensors.
is then a tensor, and so are the Jacobians of and . On the theoretical side we will consider the space involves as vector, leaving the tensor related treatment to specific implementations.2 NewtonRaphson applying to eigenvector problem
To explain the ideas involved here we look at the eigenvector problem in details. Here
and constraint is The Jacobian of the total Lagrangian is
where . Applying the Schur complement formula we can evaluate the Newton step:
and get (with and )
So with we simplify the updating equations to:
From here
We see is proportional to , a result known from classical Rayleigh iteration, and the equation for is exactly the Rayleigh equation. However the formula for is iterative. Starting with the general equation for the explicit system:
From the full rank assumption has a left inverse (for example ). We can solve for :
(6) 
In the eigenvector case this is exactly the Rayleigh quotient. This expression appeared very early in the multiplier method literature as discussed below.
We will show if we modify the NewtonRaphson process to use this expression to solve (Eq. 1), we get a generalization of the Rayleigh quotient iteration that also have quadratic convergent rate in general. We also provide a formula generalizing the Chebyshev iteration. We will show a similar process exists for implicit equation, in particular for nonlinear eigenvalue problem and provide both the RQI and RayleighChebyshev results, in particular recover the cubic convergent algorithm of Schwetlick and Schreiber ([10], [17] [18]
Many problems in numerical analysis and optimization problems could be reduced to the form (Eq. 2). We will apply the results of this paper to a number of situations. In most cases the constraints are at most quadratic, and in many cases the function is also at most quadratic. In those cases, second order methods would have simple form which we will try to exploit. In particular, we have a new cubic algorithm for eigenvectors of nonnormal matrices.
In future research we hope to study inequality constraints and applications of this method to optimization problem.
In the next few sections we discuss a few set ups needed to state the main results.
3 Higher derivatives as tensors
The reader can consult [3] for this section. We use slightly different notations in this paper. Recall we can use tensors to denote linear maps between two vector spaces each represented as tensor. If the domain space is of shape and the range space is of shape then a map between them could be represented as a tensor of shape . The map sending a tensor to the tensor formed by contracting to the right is the linear map represented by .
If is a (may be nonlinear) map between two vector spaces and then its Jacobian is a map in of linear maps between and and is represented by a matrix.
A second derivative is a linear map between and , an element of and can be represented as tensor. We will denote this map as well as this tensor as . In general, we will denote the derivatives as and this is an element of (with copies of and one copy of . We can represent it as a tensor of size .
For vectors consider the tuple
we can define
which is the repeated contraction of . We write
times. With this notation we can write the Taylor series expansion up to order around as:
where .
To summarize, there are two maps related to higher derivatives. The map from to is generally nonlinear resulting in a tensor. For a fixed , that tensor gives a multilinear map acting on the tangent space which is embedded in , sending to . In code, we need two functions for these two maps. The second map is in general just tensor contraction however depending on problem a custom implementation may be useful.
4 Retractions
Consider a submanifold of . Recall the definitions of retractions from [2]:

A first order retraction is a map from to around a point if there exists a neighborhood of in such that:

and the restriction is of class .

for all



A second order retraction on is a first order retraction on that satisfies for all ,
(7)
is the normal space at . The exponent map is a second order retraction. It is shown in that paper that projection and orthographic projections are second order retractions. The following is clear:
Proposition 1
If is a retraction on then is a retraction on , if is a retraction of second order then is a retraction of second order.
From this proposition and the result of [2]), we can retract intermediate iteration points to , as a result can be made elements of while is unchanged. Following the common literature, we call a NewtonRaphson iteration on with no retraction to infeasible start iteration. When there is retraction to we have feasible start iteration. We note Lagrange multiplier with feasible improvement has already been studied been studied, e.g. in [8].
5 NewtonRaphson iteration of Lagrangians
We first examine the NewtonRaphson iteration. The Jacobian of is
(8) 
We will drop and from time to time to save space in the expressions. NewtonRaphson iteration in the framework of constraint optimization has been studied by many authors in the literature, including [7], [8], [9], [16], [19]. We will make the connection between with the eigenvector equation more explicit, and show converges to the Rayleigh quotient expression. We transform the updating equations to a format closer to one derived from Riemannian Newton optimization which motivates the RQI in the next section.
To invert this Jacobian without inverting the whole matrix, we describe two approaches. One approach would focus on first parametrize using the equation . If the tangent space has an explicit description this would help, otherwise it is parametrized by the null space of ( where is a basis of the null space and ’s are the new parameters). This will help reduce the number of variables to . Substitute back to the first equation we get a system of equations and variables ( from and from ) which we can solve. This solves the problem in the most general case, including the case where is not invertible and is closely related to the Riemannian Newton approach on the embedded manifold defined by . We present a second approach, already investigated in [[16]], [19]] where we assume is invertible. This approach reveals some interesting relations mentioned above. As we use the Schur complement formula in this second approach, we will call this solution the Schur form. The approach to parametrize first is called the tangent form. In a sense, the classical Rayleigh iteration and its extension in [1] is a Schur form solution for a modified NewtonRaphson equation.
The Schur complement with respect to the top block is evaluated at and the inverse of the Jacobian applied on is
(9) 
evaluated at . With and the Newton step is with
(10) 
(11) 
On a feasible starting point and we thus have:
(12) 
We also note from the second row block of the system in that case. For nonlinear eigenvalue problem the above process is the Nonlinear inverse iteration in the literature ([12], [10]). We will review this in Section 7.4. We note that Schur complement is widely use in equality constraint optimization, for example see chapter 10 of [6] where we see essentially the above calculation. We summarize the steps in algorithm Algorithm 1.
When is explicit, and are two linear maps:
actually has a simpler form. Write it as
we see it is . Here, means transposing with respect to the first two indices.
(13) 
is a generalization of the operator of eigenvalue problem. We do not need the full inverse of in general, but will need to solve for for some matrix in each iteration. We collect all the result thus far in theorem (Theorem 1)
Theorem 1
The NewtonRaphson iteration equations for are
(14) 
(15) 
If then we have
In the explicit case we have
(16) 
(17) 
Proof
We already proved the implicit case. The explicit case is a direct substitution of in to (Eq. 10).
We note the is updated first, then is used in equation for :
(18) 
While we have noted before is a generalization of the eigenvalue operator, the right hand side of this equation is different from that of Rayleigh quotient. To compute and we compute
In the eigenvector case and are related:
and we have seen the calculation get simplified. In the general case we need to compute both and . The expression of is iterative.
Let us now focus on getting a different expression on . From (Eq. 17)
As before let be a left invert to , we solve for :
(19) 
We see as converges to zero converges to
as noted before. may be of a more general form than , for example we can replace with any map such that is invertible. This general set up is relevant when we discuss the twosided Rayleigh quotient.
We note that the papers [7], [8] found the same expression for the general Rayleigh quotient as an estimate of the Lagrange multiplier, together with a related quasiNewton method. The special case of , appeared much earlier in the literature (for example [11]). According to [19] it is difficult to put a name on it without an extensive search of the literature. Lagrange multiplier approximators are widely used in constrained optimization.
Let . This is a projection. If then it is a projection to the tangent space of at , but in general it is not. Substitute the expression (Eq. 19) in (Eq. 18) and moving term to one side
(20) 
This system of equations for is of similar format to the Riemannian Newton equations. However, it is dependent on which needs to be solved recursively. It is interesting that it involves only here, and not . In the implicit case, assuming is a left inverse of we have with
Thus far the analysis is mostly algebraic, using the Schur complement formula and the special form of .
For second order iteration we note the formula for Chebyshev iteration (we choose Chebyshev over Halley to avoid another operator inversion) for high dimension is given by
(21) 
As explain before with and , is an element of and is an element of and the last term is a contraction of the tensor twice on . We note that
We proceed to use Schur complement to evaluate the second order term to arrive at Chebyshev iteration in algorithm Algorithm 3.
While this looks relatively complex, when , and are at most quadratic the algorithm may be useful. In particular, we have a cubic convergent algorithm for eigenvectors, even when the matrix is not normal: the only nonzero terms are and in the Chebyshev expression.
When is a vector and is represented as a matrix, is a matrix and can be represented as a square matrix, so the calculation is simple. When is a matrix, is could be a higher order tensor, so and in general are complex tensors. The main difficulty of this method is in evaluating these tensors and the inverse of the Schur complement .
6 Rayleigh Quotient Iteration
Motivated by [1] (we learned about [7], [8] late in our research) and the above analysis on Lagrange multipliers, with
in the expression for , it is plausible that the system:
(22)  
would provide a generalization of RQI for to vector Lagrangians. Using augmented Lagrangian technique, [8] proposed a quasiNewton method with this expression of as an estimate for the Lagrange multiplier. He showed that it converges superlinearly in general. We prove our particular process has quadratic convergence and extend our result to to the Chebyshev case.
Similar to the NewtonRaphson case, if is invertible we have a solution to this system providing a procedure generalizing the classical Rayleigh quotient. In fact, let and we see
(23) 
satisfies the above equation by direct calculation. As before we call it the Schur form solution. For the eigen problem this provides exactly the classical algorithm. Likewise, we will see later on the invariant subspace RQI of [1] is the Schur form of an equivariant constrained problem on Grassmann manifolds.
We note is now solely dependent on via Rayleigh quotient expression for . We can reparametrize to express as a map from the tangent space of the constraint set to the image of , both of the same dimension . For the case where and ,
( is a just a collection of Hessians for each constraint equation of ) and the equation reduces to the emmbedded Riemannian Newton updating equation. When the last expression is just the expression for the projected Hessian of . We have the following theorem:
Theorem 2
Assuming that having continuous derivatives to degree , and in a neigborhood of a solution to the equation the map from the tangent space to to the image of is invertible satisfying
(24) 
for any in for some constant . If is a first order retraction then for a starting point close enough to , the system (Eq. 22) provides an update to an iteration process which converges with quadratic rate to a solution of (Eq. 3).
Assuming (Eq. 24), having continuous derivatives to degree and have continuous derivatives to degree . If is a second order retraction, then for a starting point close enough to the RayleighChebyshev iteration with update
where and are defined in the following equations
(25) 
converges cubically to .
If the Rayleigh quotient iteration converges cubically up to degree .
If is invertible the Schur form solution exists
(26) 
Proof
Both algorithms require similar estimates. For optimization problem it is just the Riemannian Newton iteration, so this theorem could be considered an extension to vector Lagrangians. We focus on RayleighChebyshev as regular Rayleigh is the easier version and we will prove it again in the implicit case below. Let be a solution of (Eq. 3). As this is a variant of NewtonRaphson and Chebyshev, implicitly we look at Taylor series expansion of with modifications to make sure the steps belong to the tangent space () and note will remove some extra terms. When the Lagrangian is explicit, the Taylor series expansion with term gets cut off early and the second order term would include second derivative with respect to , as well as the cross term of and . This is the famous cross term in Rayleigh quotient literature linking the estimate of eigenvalues to eigenvector and is responsible for the quadratic convergence rate (see [15]). We have:
(27)  
This is purely algebraic (Taylor series of the two variable vector function ). The first term will converge quadratically while the other two are linear and will be made disappear with an appropriate iteration. We also note the first term is cubic if converges quadratically. Also :
(28) 
So if an iteration step is to cancel out , then also cancel out the same term as . As sends terms of form to zero, we get a simpler equation to solve.
With these two observations, using the notation of higher derivatives as tensor discussed previously we have the Taylor series expansion
Comments
There are no comments yet.