## 1 Introduction

Over the last decade, several algorithms have been proposed for distributed estimation over wireless sensor networks. Different algorithms target different goals. However, the most significant feature among these algorithms has been the basic distribution schemes used for developing the algorithms. Nearly all algorithms have been based on either the incremental scheme or the diffusion scheme

[1].The incremental scheme was first introduced in [2]. The network was distributed using a Hamiltonian cycle and each node was connected to two other nodes in the cycle. The resulting algorithm converged quicker than the diffusion scheme and gave lower misadjustment as well. While the diffusion scheme presented some advantages over the incremental scheme, one significant advantage was the theoretical analysis [3]. Since the analysis of the diffusion scheme is easier to perform, most subsequent algorithms have been based on the diffusion scheme [1].

The analysis of the incremental scheme is not easy to perform. The analysis presented in [2] is incomplete. The transient analysis presented in [2] reaches a certain point before the focus shifts to steady-state results. The reason is that the learning behavior for the incremental scheme is not straightforward. As a result, the subsequent algorithms based on the ILMS algorithm of [2] also follow the same process and leave the transient analysis incomplete. This work presents the complete transient analysis for the ILMS algorithm, including the learning behavior. The results are verified through several experimental results. The experiments are conducted to test all aspects of the analysis and simulation results are used to validate the theoretical findings.

## 2 The ILMS Algorithm

A collection of sensor nodes, spread over a geographical area is being considered. The nodes are connected in a Hamiltonian cycle [2]. Fig. 1 shows an illustration of a possible adaptive wireless sensor network. Connecting the nodes in a cyclic way would result in the incremental strategy.

The unknown parameters are modeled as a vector,

, of size (). The input to a node at any given time instant, , is a () regressor vector, , where is the node index. The resulting observed output for the node is a noise corrupted scalar, , given by(1) |

where

is the zero-mean additive noise with variance

.The incremental least mean square (ILMS) algorithm is given by [2]

(2) | |||||

(3) |

where is the estimate of the unknown vector at time instant for node , is the instantaneous error, is the final estimate for iteration and is the transpose operator.

## 3 Proposed analysis

Let the weight-error vector be given by

(4) |

Using (4) in (2) and simplifying gives

(5) | |||||

where

is an identity matrix of size

. The auto correlation matrix of the input regressor vector is given by , where is the expectation operator.can be decomposed into its component matrices of eigenvalues and eigenvectors. Thus,

, where is the matrix of eigenvectors such that and is a diagonal matrix containing the eigenvalues. Using the matrix , the following transformations are madeThe weight-error update equation thus becomes

(6) | |||||

### 3.1 Mean Analysis

Applying the expectation operator to (6) gives

(7) | |||||

where the data independence assumption separates from the rest of the variables. The second term is 0 as additive noise is independent and zero-mean and . The sufficient condition for stability is evaluated from (7) and is given by

(8) |

where is the maximum eigenvalue of .

### 3.2 Mean-Square Analysis

For the mean-square analysis, the approach of [2] is followed. However, the analysis in [2] was not completed to give the learning behavior for the algorithm. Here, the learning behavior is also given in closed form. Taking the squared weighted -norm of (6) and applying the expectation operator yields

(9) | |||

where is the -norm operator and is a weighting matrix. The weighting matrix is given by

(10) | |||||

The last two terms in (9) are zero since the additive noise is independent. Using the data independence assumption, the remaining two terms are simplified as

where is the additive noise variance at node , is the trace operator and .

The expectation operator from the first term in (3.2) applies to the weighting matrix independently as well since the data is being assumed to be independent [4]. Thus, we have, after simplification

(12) | |||||

Using the operator, (3.2) is further simplified to

where , and , where is given by

(14) |

Now, it can be seen from (14) that remains fixed at every iteration, even if it varies for each node, depending on the individual values of the parameters. Thus, using (3.2), the analysis is initialized as

The first iterative update for node is given by

Similarly, the first update for node is given by

Continuing for node gives

where

Thus, for node , the first update is

where

Finally, the first iteration ends with node and the final update is given by

Moving to iteration , the update for node is given by

For node , we have

Continuing for node , we get

The final update for iteration is given by

Before moving on, we need to link the first update for node with the second update. Beginning with node ,

Similarly, for node , we get, after simplification

Generalizing for node gives

The final update for iteration is thus give by

Since the final update for each iteration is given by the update of node , we focus on node only for now. Continuing for node , the update for iteration is given by

(15) | |||||

Similarly, for iteration , we have

(16) | |||||

Subtracting (15) from (16), rearranging and simplifying gives

(17) | |||||

The term can be written in an iterative way as follows:

(18) |

Inserting (18) into (17) gives the final update recursion

(19) | |||||

Taking the weighting matrix as results in the mean-square-deviation (MSD) and gives the excess mean square error (EMSE).

### 3.3 Steady-State Analysis

For steady-state, we take a look at the relation between the first and second updates for node again. We do this in an iterative way as follows:

where the last step is obtained by further successive iterations and simplifying. Generalizing for iteration gives

(20) |

At steady-state, (20) becomes

(21) |

Rearranging and simplifying (21) gives the steady-state equation

(22) |

Using results in the steady-state mean-square-deviation (MSD) and gives the steady-state excess mean square error (EMSE).

## 4 Results and Discussion

This section compares the theoretical findings presented above with simulation results. A network of nodes is used for the simulations, with the length of the unknown vector being . Results are shown for SNR values of dB, dB and dB. The noise power values at each node for the different SNR values are shown in Fig. 2.

In the first experiment, the input data is assumed to be white. The convergence speed varies in order to have a comparison at both fast convergence and slow convergence. Therefore, the values used are and . The results are shown in Figs. 3 and 4. As can be seen, there is an excellent match between theory and simulation curves for all values of SNR. For comparison, the steady-state results obtained using (22) are also shown.

In the second experiment, the two scenarios are exactly the same except that the input data is now correlated, with the correlation factor being . The results are shown in Figs. 5 and 6. It should be noted that there is a slight discrepancy between the curves during the transient phase. However, this discrepancy is very low and the two curves are closely matched.

## 5 Conclusion

This work presents the complete transient analysis for the incremental LMS algorithm, first presented in [2]. The learning behavior for the transient analysis is presented. Results show that the steady-state results from the transient analysis and the steady-state results from the steady-state analysis are an exact match. Furthermore, the simulation results closely match the theoretical results for different scenarios.

## References

- [1] A. H. Sayed, A. H.: Adaptive Networks. Proceedings of the IEEE. 102, 460-497 (2014).
- [2] Lopes, C. G.; Sayed, A. H.: Incremental adaptive strategies over distributed networks. IEEE Trans. Signal Process. 55, 4064-4077 (2007).
- [3] Lopes, C. G.; Sayed, A. H.: Diffusion least-mean squares over adaptive networks: Formulation and performance analysis. IEEE Trans. Signal Process. 56, 3122-3136 (2008).
- [4] Sayed, A. H.: Fundamentals of Adaptive Filtering. Wiley, New York (2003).

Comments

There are no comments yet.