# Exploiting Convexification for Bayesian Optimal Sensor Placement by Maximization of Mutual Information

Bayesian optimal sensor placement, in its full generality, seeks to maximize the mutual information between uncertain model parameters and the predicted data to be collected from the sensors for the purpose of performing Bayesian inference. Equivalently, the expected information entropy of the posterior of the model parameters is minimized over all possible sensor configurations for a given sensor budget. In the context of structural dynamical systems, this minimization is computationally expensive because of the large number of possible sensor configurations. Here, a very efficient convex relaxation scheme is presented to determine informative and possibly-optimal solutions to the problem, thereby bypassing the necessity for an exhaustive, and often infeasible, combinatorial search. The key idea is to relax the binary sensor location vector so that its components corresponding to all possible sensor locations lie in the unit interval. Then, the optimization over this vector is a convex problem that can be efficiently solved. This method always yields a unique solution for the relaxed problem, which is often binary and therefore the optimal solution to the original problem. When not binary, the relaxed solution is often suggestive of what the optimal solution for the original problem is. An illustrative example using a fifty-story shear building model subject to sinusoidal ground motion is presented, including a case where there are over 47 trillion possible sensor configurations. The solutions and computational effort are compared to greedy and heuristic methods.

## Authors

• 1 publication
• 6 publications
03/16/2018

### A New Result on the Complexity of Heuristic Estimates for the A* Algorithm

Relaxed models are abstract problem descriptions generated by ignoring c...
06/03/2021

### MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood Inference from Sampled Trajectories

Simulation-based inference enables learning the parameters of a model ev...
06/17/2018

### Experimental Design of a Prescribed Burn Instrumentation

Observational data collected during experiments, such as the planned Fir...
02/28/2018

### Sensor and Sink Placement, Scheduling and Routing Algorithms for Connected Coverage of Wireless Sensor Networks

A sensor is a small electronic device which has the ability to sense, co...
12/03/2011

### Information-Maximization Clustering based on Squared-Loss Mutual Information

Information-maximization clustering learns a probabilistic classifier in...
01/07/2020

### On the Uniqueness of Binary Quantizers for Maximizing Mutual Information

We consider a channel with a binary input X being corrupted by a continu...
02/25/2019

### Robust Affordable 3D Haptic Sensation via Learning Deformation Patterns

Haptic sensation is an important modality for interacting with the real ...
##### This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

## 1 Introduction

In system identification and structural health monitoring, a common goal is to inform the parameters governing a predictive model of a structure using relevant data collected from sensors on it. For this purpose, it is often desirable to incorporate a Bayesian model updating framework to quantify uncertainty surrounding the model parameters based on sensor data. (Beck and Katafygiotis, 1998; Beck, 2010; Vanik et al., 2000; Zhang and Au, 2016). For example, in a structural dynamics setting, the sensors would typically measure the physical response of the structure in terms of accelerations, strains, etc. under natural, induced or ambient excitation. The data is then used to inform the parameters of a model, which is typically a linear dynamical finite-element model with viscous damping but need not be. Any available prior information about the model parameters that is available can be incorporated in the Bayesian model updating framework, where prior and posteriorprobability distributions over parameters of interest are used to quantify uncertainty in their values before and after utilization of the data, respectively.

The optimal sensor placement problem may be viewed as optimally distributing sensors over a structure to maximize the information gained from the data about the model parameters, given a fixed sensor budget (Beck et al., 1998; Papadimitriou et al., 2000; Yuen et al., 2001; Papadimitriou, 2004; Yuen and Kuok, 2015; Sun and Büyüköztürk, 2015; Argyris et al., 2018)

. For structures that are not simple geometric shapes, which is typically the case, the optimal sensor placement problem over all possible locations becomes difficult to specify. Therefore, only a pre-determined subset of locations is considered for instrumentation, for instance, the degrees of freedom of a finite-element model of the system. The optimization problem therefore becomes an inherently discrete configuration selection problem. In this case, heuristics such as genetic algorithms, greedy selection, swarm optimization etc.,

(Papadimitriou et al., 2000) (Papadimitriou, 2004) (Sun and Büyüköztürk, 2015) are usually needed in order to obtain a satisfactory sub-optimal solution, as the computational cost of a brute-force combinatorial search becomes prohibitive for realistic problems.

The optimality criterion is defined according to the purpose of collecting the sensor data. In this paper, it is assumed that the data would be used in a Bayesian model updating framework. In this setting, optimally placed sensors would collect data to maximize information gained about the model parameters. Then the mutual information (Cover and Thomas, 2005) between the model parameters and predicted data is a natural choice for the objective function for this optimization problem. Equivalently, the expected entropy of the posterior of the model parameters with respect to the predicted data can be chosen. This entropy-based objective function for optimal sensor placement was introduced in (Beck et al., 1998) (Papadimitriou et al., 2000).

Under the assumptions that imply applicability of the Laplace approximation to the posterior distribution, and assuming sufficiently small prediction errors, the entropy-based optimal sensor placement problem can be simplified to the optimization of the log-determinant of the Fisher information matrix (FIM) of the model parameters (Beck et al., 1998). This gives us a rigorous information-theoretic foundation for one of the proposed choices of the objective function in (Udwadia, 1994), namely, the aforementioned log-determinant, rather than the trace, of the FIM. The main focus of this paper is to show how this log-determinant formulation can lend itself to a fast sub-optimal, and possibly optimal, solution using convex optimization (Boyd and Vandenberghe, 2004) (Joshi and Boyd, 2009), bypassing the need for a discrete optimization over all possible sensor configurations. The method is applied to an illustrative example of a 50-story uniform shear building, where there are over 47 trillion possible sensor configurations and the optimal sensor placement is found using fewer than 100 objective function evaluations.

## 2 Bayesian model updating

This section lays out the framework for Bayesian model updating for structural dynamics which forms the basis for the optimal sensor placement problem. The measured data on which inference is to be performed is assumed to be available in the form of dynamic test data in the time domain. It is also assumed here that the complete system input can be determined given the parameters. In a typical structural dynamics setting, one observes the acceleration, velocity or displacement time history at a certain number of locations. The locations that are selected, could, for instance, correspond to a subset of the degrees-of-freedom of a finite-element model of the structure.

Denote the uncertain system model parameters by . These could correspond, for instance, to sub-structure stiffnesses, Rayleigh damping parameters, etc. in the finite element model used to predict the response of the structure.

Denote by the stochastic prediction of the quantity to be observed at the time instant . Note that , where is the actual number of observed degrees-of-freedom out of a total of potential degrees-of-freedom for observation. The set of stochastic predictions for the observations over the whole duration of the measurements, , is denoted by .

The equation relating stochastic predictions of the observations, , to the deterministic predictions from the model is:

 yn=So(δ)(xn(θs)+ϵn(θe)) (1)

where the vector of deterministic predictions at each time-step of interest is denoted by ; the uncertain prediction-error is denoted by ; is a binary vector that indicates the presence or absence of a sensor at a degree-of-freedom, so that the sum of entries of equals the number of observed DOFs, ; and the sensor selection matrix, , selects the components that are to be observed from the full stochastic prediction. The uncertain prediction-errors are modeled by a probability distribution that depends on uncertain prediction-error parameters denoted by

; for example, using a zero-mean, discrete-time Gaussian white-noise process with a stationary and isotropic co-variance matrix having diagonal entries

, would just be the scalar . For an anisotropic model for the co-variance matrix without prediction error correlation between different locations, would be a vector of variances corresponding to the diagonal entries of the co-variance matrix.

Denote the complete parameter vector by , which is the collection of uncertain system parameters and prediction-error parameters, i.e., . Prior information on the parameter values is expressed in terms of a prior distribution, . The stochastic forward model, typically denoted by , prescribes the probability of observing a particular data set given the parameters. The probability model class, , is used to indicate a specific pairing of stochastic forward model and prior. The stochastic forward model with actual data substituted for , when viewed simply as a function of the uncertain parameters, , is called the likelihood function. Our prediction-error modeling assumptions imply:

 p(y1:N|θ,M)=N∏n=1p(yn|θ,M) (2)

Bayes’ rule implies that the posterior probability density,

, is proportional to the product of the likelihood function and prior probability density, thereby allowing an update of the prior:

 p(θ|y1:N,M)=p(y1:N|θ,M)p(θ|M)p(y1:N|M) (3)

Here, the normalizing factor is called the evidence for the model class provided by the data . It can be determined using the Total Probability Theorem as:

 p(y1:N|M)=∫p(y1:N|θ,M)p(θ|M)dθ (4)

For a given data set, the parameter vector that maximizes the likelihood function is known as the maximum likelihood estimate (MLE), denoted by

. Similarly, the parameter vector that maximizes the posterior probability density is known as the maximum a posteriori (MAP) estimate of the parameters, denoted by .

### 2.1 The prediction-error model

Based on the Principle of Maximum Information Entropy (Beck, 2010) (Cover and Thomas, 2005), we model the prediction errors as a discrete-time Gaussian process with equal variance at every degree of freedom, and uncorrelated across time and location, so that:

 E[ϵnϵTm]=σ2INdδmn where ϵn∼N(⋅|0,σ2INd) i.i.d. (5)

This implies that knowing the prediction-error at one time and location provides no information a priori

about the errors at other times and locations. This Gaussian distribution on the prediction errors gives the maximum entropy (most uncertainty) for a specified zero mean and common variance

for all times and locations. Combining Equations (1), (2) and (5), the stochastic forward model then becomes:

 p(y1:N|θ,M)=N∏n=1N(yn|So(δ)xn(θ),σ2) (6)

The logarithm of this is needed later:

 logp(y1:N|θ,M)=−NNo2log2πσ2−12σ2N∑n=1∥yn−So(δ)xn(θs)∥2 (7)

## 3 Mutual information-based optimal sensor placement

The mutual information between two stochastic variables is a measure of about how much they imply about each other’s possible values. It is a fundamental information-theoretic quantification of their correlation (Cover and Thomas, 2005). In our case, the mutual information between the stochastic predictions for the observed data, , and the parameter vector, , is defined by:

 I(y1:N(δ),θ|M) =Ey1:N(δ),θ|M[logp(y1:N(δ),θ|M)p(y1:N(δ)|M)p(θ|M)] (8) =∫∫p(y1:N(δ),θ|M)[logp(θ|y1:N(δ),M)p(θ|M)]dy1:Ndθ (9) =∫p(y1:N(δ)|M)∫p(θ|y1:N(δ),M)[logp(θ|y1:N(δ),M)p(θ|M)]dθdy1:N (10) =Ey1:N(δ)|MDKL[p(θ|y1:N(δ),M)||p(θ|M)] (11)

where is the Kullback-Leibler (KL) divergence (the inner integral over in Equaton (10)).

In the definition of mutual information in Equation (8), the dependence of the stochastic prediction of the data on the sensor configuration is made explicit but will sometimes be omitted for convenience. Equations (9) and (10) come from the product rule in probability. These equations show that the mutual information can be expressed as the expected Kullback-Leibler (KL) divergence, or expected relative entropy, of the posterior with respect to the prior.

In the optimal sensor placement framework, we wish to maximize the mutual information between the predicted sensor data and the model parameters over all possible configurations of placing sensors at locations. This is equivalent to maximizing the expected KL divergence in Equation (11), which implies that the optimal sensor configuration has the largest expected reduction in entropy of the posterior distribution from the prior. This corresponds to the greatest expected information gain about the parameters from the data during Bayesian updating. In other words, considering all sensor configurations, the optimal one would result in the maximum reduction in parameter uncertainty, on average, upon obtaining the data.

The mutual information in Equation (9) can also be expressed in terms of two specified distributions specified by the model class , i.e., the prior and the stochastic forward model, by using the product rule again:

 I(y1:N(δ);θ|M)=∫∫p(y1:N(δ)|θ,M)p(θ|M)[logp(y1:N(δ)|θ,M)p(y1:N(δ)|M)]dy1:N(δ)dθ (12)

The denominator in the logarithm term is what one would associate with the evidence for the model class , based on given data, as defined in Equation (4). Here, the data is not available and this term is simply a function obtained upon marginalizing out the parameters from the joint on and .

Choosing the mutual information between the uncertain parameters and uncertain data as the objective function to be maximized over the sensor configurations, the optimal sensor placement problem becomes:

 maximizeδ I(y1:N(δ);θ) (13) subject to δi∈{0,1}, i=1,…,Nd and 1Tδ=No.

Equation (12) can be used for the objective function for the problem in (13). Recall that is a binary -dimensional vector specifying the sensor configuration over the degrees of freedom in the system and so it has unit components and zero components. The explicit dependence on has been suppressed in the notation in (13).

In principle, to perform the optimization in Equation (13), we need to search over all possible sensor configurations. Instead, we employ a convex optimization technique that greatly reduces the computational complexity of this search.

## 4 Equivalent entropy-based optimal sensor placement

In this section, an equivalent entropy-based optimal sensor placement problem is formulated and an efficient solution using convex optimization techniques is presented. The Laplace approximation to the posterior in a Bayesian system identification problem is described and used to express the entropy of the posterior.

Using the expression in Equation (9) for mutual information:

 I(y1:N(δ);θ) =Ey1:N(δ),θ[logp(θ|y1:N(δ))p(θ)] (14) =H(p(θ))−Ey1:N(δ)[H(p(θ|y1:N(δ)))]

where is the entropy corresponding to . Here, the prior entropy term is irrelevant in the sensor placement problem and may be discarded to yield an objective function that depends only on the posterior entropy term:

 U(δ)=Ey1:N(δ)[H(p(θ|y1:N(δ)))] (15)

Thus, the objective function maximization in Equation (13) can be replaced by an equivalent minimization of the expected posterior entropy of the model parameters . The idea of choosing the locations of the sensors so that the posterior uncertainty of the parameters is minimized was first introduced in (Papadimitriou et al., 2000) (Beck et al., 1998). In these papers, it is shown that this entropy-based criterion can be simplified to a more manageable log-determinant criterion when the following assumptions are made about the probability model class :

• The model class, upon collection of the data, is globally identifiable and the posterior distribution has a single peak (Beck and Katafygiotis, 1998).

• The Laplace approximation holds; that is, a sufficient number of data points are measured so that the posterior is approximately Gaussian (Beck and Katafygiotis, 1998).

Applying the Gaussian approximation to the posterior, and using to denote the MAP value of based on data :

 p(θ|y1:N(δ))≈N(θ|^θ(y1:N),A−1N(^θ(y1:N),δ)) (16)

where the precision matrix here is given by the negative of the Hessian of the logarithm of the stochastic forward model (Beck and Katafygiotis, 1998):

 [AN(θ,δ)]ij=−∂2logp(y1:N(δ)|θ)∂θiθj (17)

The objective function in Equation (15) is now simply related to the entropy of this approximately Gaussian posterior which may be expressed in terms of the log-determinant of its precision matrix as:

 U(δ)=−12Ey1:N(δ)[logdetAN(^θ(y1:N),δ)] (18)

Note that the optimal sensor location problem is to be solved typically before any response data is available from the real structure to be instrumented, so predictions of the data must be made based on a probability model. This can be constrained based on available information about the structure, either from its design or from preliminary tests, in the form of a structural model and a prior distribution over the system parameters, as shown in Section 2. A nice feature of the log-determinant formulation above, is that the dependence on the data of the approximate posterior distribution over the uncertain parameters is only through the optimal parameters, , that is,

 p(θs|y1:N,M)=p(θs|^θ(y1:N),M) (19)

The precision matrix in Equation (17) of the multivariate Gaussian in Equation (16) can be partitioned into two parts corresponding to the MAP values of the system parameters and the prediction-error parameter, :

 AN(θs)=[BN(θs)00CN(^σ)] (20)

where

 [BN(θs)]ij=−[∂2logp(y1:N|[θs,^σ],M)∂θs,i∂θs,j] (21)

and

 CN(^σ)=−∂2logp(y1:N|[θs,σ],M)∂σ∂σ∣∣∣σ=^σ (22)

The off-diagonal terms in the Hessian of Equation (20) are zero because of our choice of stochastic forward model and prior. Also, in Equations (20) - (22), for fixed system parameters is the MAP prediction-error variance, assuming a uniform prior on over a large positive interval :

 ^σ2(θs)=1NNoN∑n=1∥yn−So(δ)xn(θs)∥2Δ=J(θs) (23)

Equation (23) can be used to re-express the stochastic forward model at the MAP variance parameter:

 p(y1:N|θ=[θs,^σ],M)=[2πeJ(θs)]−NNo/2 (24)

Also, the posterior distribution of the system parameters, with the prediction-error parameter marginalized out, can be expressed as:

 p(θs|y1:N,M)=√detBN(^θs)(2π)NNo/2exp[−12(θs−^θs)TBN(^θs)(θs−^θs)] (25)

where from Equations (7) and (21):

 BN(^θs)=NNo2J(^θs)[∂2J(θs)∂θs∂θs]θs=^θs (26)

Furthermore, using Equation (20), the log-determinant in Equation (18) can be expressed as:

 logdetA(^θ(y1:N),δ)=logCN(^σ)+logdetBN(^θs) (27)

where from Equations (7) and (22):

 CN(^σ)=2NNo^σ2(^θs) (28)

Expanding the second-order derivative in Equation (26) using the expression for from Equation (23):

 [BN(^θs)]pq=1^σ2N∑n=1[∂xn∂θs,pSoSTo∂xTn∂θs,q+ϵn(So∂2xn∂θs,p∂θs,q)T]∣∣∣θs=^θs (29)

It is now assumed that the term involving the second derivative in Equation (29) can be neglected. This can be justified if at least one of the following is true: the prediction errors, , are small in magnitude; or the deterministic predictions, , vary slowly with respect to the system parameters and therefore have small second derivatives. The curvature assumption can be assessed before instrumenting the structure, while the prediction-error assumption can only be validated after checking the agreement between the deterministic predictions and the data.

With this assumption, the relevant portion of the Hessian matrix can be expressed in terms of the prediction sensitivity coefficients at the observed degrees of freedom. The resulting approximation is then:

 [BN(^θs)]pq≈1^σ2N∑n=1[∂xn∂θs,pSoSTo∂xTn∂θs,q] (30)

In addition, notice the dependence on the sensor configuration becomes , so Equation (30) can be transformed to a double-sum:

 [BN(^θs)]pq=1^σ2Nd∑i=1δiN∑n=1∂xn,i∂θs,p∂xTn,i∂θs,q (31)

Equation (31) is a linear sum of contributions of terms from each sensor location. For convenience of notation in the formulation of the optimal sensor placement problem, we introduce the matrix , so:

 [BN(^θs)]pq=1^σ2Qpq(δ;^θs)=1^σ2Nd∑i=1δiQ(i)pq(^θs) (32)

where

 Q(i)pq=N∑n=1∂xn,i∂θs,p∂xTn,i∂θs,q∣∣∣θs=^θs (33)

Having developed an approximate expression for the posterior distribution over the system parameters, we are now in a position to tackle the optimal sensor location problem.

To greatly simplify calculations, we use nominal values and for the system and variance parameters respectively, in place of the optimal parameters and . Then, from Equation (25):

 p(θs|θ0,M)=√detBN(θs,0)(2π)NNo/2exp[−12(θs−θs,0)TBN(θ0)(θs−θs,0)] (34)

Since the designer is uncertain about the nominal values, a prior distribution, , where , is specified for the nominal parameters. Therefore, Equation (34) replaces the dependence of the posterior on unavailable data with prior information that is already available in terms of the nominal parameters. These parameters are either available in the form of structural design information, or in the form of information obtained from preliminary tests of the structure.

Thus, the objective function in Equation (18) becomes, after using Equations (27), (32) and dropping the constant terms and scale factor of :

 h(δ) =−Eθ0[logdetQ(δ,θ0)] (35) =−∫logdetQ(δ,θs,0)p(θs,0)dθs,0

where the expression for the matrix is given in Equations (32) and (33).

Thus, as in (Papadimitriou et al., 2000), the entropy-based optimal sensor location problem requires a minimization of the expected posterior entropy of the system parameters over the binary vector, , to give the optimal sensor configuration, :

 minimizeδ h(δ) (36) subject to δi∈{0,1}, i=1,…,Nd and 1Tδ=No.

This is a combinatorial optimization problem, that is, we wish to minimize

by choosing sensor locations among possible locations. Depending upon the problem, this could become prohibitively expensive. Each sensor configuration requires one evaluation of the expectation of the log-determinant of a computationally expensive prediction. For example, selecting 20 locations from among 50 locations to instrument gives a number of possible combinations of about 47 trillion.

Clearly, a brute-force search for the optimal solution over all possible sensor configurations will often not be a feasible approach. However, it is usually the case that it is not possible to guarantee an optimal solution without an exhaustive search. Heuristic methods such as genetic algorithms can produce an acceptable sub-optimal value if run for long enough. Incremental greedy algorithms are also guaranteed to produce a sub-optimal value within of the optimal (Guestrin et al., 2005), relative to the range from the least to the most optimal configuration. Here, we present an alternative approach along the lines of (Joshi and Boyd, 2009) that applies a convex relaxation technique to provide a sub-optimal solution to the problem.

## 5 Convex relaxation of the combinatorial optimization

In this section, a relaxed version of the optimization problem is set up and argued to be convex. The relevant partial derivatives of the objective function, that is, its gradient and its Hessian matrix, are derived for the purpose of computing the solution. The final step involves replacing the expectation integral over the prior distribution on the uncertain parameters by its Monte Carlo approximation so that the problem may be solved by a generic convex solver.

The original optimization problem in Equation (36) is relaxed to allow the components of the binary vector to take on any real values between zero and one. The problem thus specified turns out to be a convex optimization for which there are efficient algorithms to find the unique solution, as shown in the next subsection. However, we note that the solution to this new optimization problem need not be a binary vector, because it may contain entries between 0 and 1. The solution is still very meaningful, however. If the solution is indeed a binary vector, then it is the optimal solution to the original combinatorial problem. If not, then it is an upper bound to the value of the objective function that also gives guidance to what the optimal solution is likely to be.

We replace the original binary sensor placement vector, , by a vector, , in the hypercube , resulting in the relaxed optimization problem:

 minimizez h(z) (37) subject to − zi≤0 zi≤1, i=1,…,Nd and 1Tz=No.

where the objective function is a straightforward extension of the previous one in Equation (35):

 h(z)=−Eθs[logdetQ(z,θs)] (38)

by replacing with in the definition of is Equation (32). A major advantage of the relaxed problem is that it allows one to use continuous, rather than discrete, optimization packages to solve the problem.

### 5.1 Convex nature of the relaxed optimization

Equation (37) describes a convex optimization problem where the objective function is convex in , and the equality and inequality constraints are affine. The log-determinant function in Equation (38) is convex in (Joshi and Boyd, 2009) and the expectation operator preserves the convexity of its argument. Thus, the problem has a unique global optimum. This can be determined computationally, as long as the objective function can be computed at every .

For larger problems, it is useful to efficiently supply the gradient and Hessian of the objective function with respect to . This avoids expensive computations of their finite-difference counterparts. Fortunately, their analytical expressions are tractable and, along with their Monte Carlo approximations, are described in Appendix A.

### 5.2 Solver for the relaxed optimization

Since the problem in Equation (37) is convex, it may be solved using Newton’s method. In order to apply this in practice, numerical approximations to the objective and its derivatives may be used, as in Equations (68) through (70). The MATLAB routine fmincon was used for the examples in this paper. Since the optimization problem has equality and inequality constraints, a constrained convex optimization solver needs to be employed. The expression for the Newton step is not as straightforward as in the unconstrained case and is given in (Boyd and Vandenberghe, 2004).

## 6 Application to multistory structures

### 6.1 Problem description

The problem of optimally placing a fixed number of sensors over a structure for system identification is considered. The structure is assumed to be well modeled by linear dynamics with classical viscous damping. For optimal design of the sensor layout, the structure is imagined to be subject to a sinusoidal ground acceleration at the base specified by an amplitude and frequency . Therefore, the governing differential equation for its displacement vector as a function of time is given by,

 M¨x(t)+C˙x(t)+Kx(t)=−M1a0sin(ωt), with x(0)=0,˙x(0)=0 (39)

where , and are the mass, damping and stiffness matrices respectively, and is a vector of 1’s.

In this problem, a nominal natural frequency parameter, the Rayleigh damping parameters, input frequency and input amplitude are all taken to be uncertain, i.e. . The prediction-error variance for a stationary, isotropic prediction error co-variance matrix, is the uncertain prediction-error parameter. The designer specifies a prior distribution on these uncertain parameters.

Before converting this system into its modal co-ordinates, , with corresponding natural frequency and damping ratio and respectively, the problem is simplified to take into account uncertainties in modeling the system. This is done largely through a substructuring scheme. The assumptions result in a simple expression, relative to the uncertain parameters, for each mode.

The following simplifications are assumed:

1. and , where the mass and stiffness matrices are factored into an uncertain scalar and a known matrix part. Scalars and will be called the uncertain nominal stiffness and mass of the system respectively.

2. is the uncertain nominal natural frequency parameter. Although and are uncertain, only their square-root ratio is relevant and its uncertainty may be treated explicitly.

3. , is the Rayleigh damping matrix with uncertain parameters and .

As a consequence of these simplifying assumptions, the following conditions hold:

1. The modal frequencies , where

are the known eigenvalues of the

system.

2. The eigenvector matrix

that diagonalizes the system also diagonalizes the system and can be used to solve the original dynamical system in Equation (39).

3. and , such that .

Given these simplifications, a modal solution to the problem may be calculated analytically.

### 6.2 Modal solution

We let , where vector contains the modal displacement co-ordinates, so the original equation may be simplified to:

 ΦTMΦ¨q(t)+ΦTCΦ˙q(t)+ΦTKΦq(t)=ΦTM1a0sin(ωt) (40)

Under the assumptions in Section 6.1

, this simplifies to a decoupled set of modal ordinary differential equations:

 ¨qj(t)+2ζjωj˙qj(t)+ω2jqj(t)=ajsin(ωt) (41)

where

 ω2j =c2jω20 (42) 2ζjωj =α+βω2j (43) aj =−a0μjΦT1 (44)

Each modal equation corresponds to a driven, damped oscillator whose solution is:

 qj(t)=aj [(ω3+ω2jω(2ζ2j−1)ωdj)exp(−ζjωjt)sin(ωdjt) (45) +(2ζjωjω)exp(−ζjωjt)cos(ωdjt)+(ω2j−ω2)sin(ωt) (46) −(2ζjωjω)cos(ωt)]/[(ω2j−ω2)2+(2ζjωjω)2] (47)

The prediction equation for the displacements then becomes,

 xi(t)=Nd∑j=1Φijqj(t) (48)

### 6.3 Sensor placement algorithm

In order to solve the relaxed optimal sensor placement problem in this case, the following operations need to be performed:

1. Determine the eigenvalues and eigenvector matrix of the system

2. Choose an initial guess for the sensor positions that satisfies the constraints

3. Generate samples of from the designer-specified prior distribution

4. For each sample , calculate the gradient of with respect to for

6. Calculate the elementary matrices for every and and store them.

7. Compute the objective function, its gradient and Hessian at the current vector

8. Update to and repeat steps 7 and 8 until satisfaction of the convergence criterion

### 6.4 Illustrative example of uniform shear buildings

Consider the problem of placing a fixed number of displacement sensors on a -DOF uniform shear building, as shown in Figure 1.

In this case, the mass and stiffness matrices are given by:

 M =mM∗=mINd (49) K =kK∗=k⎡⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢⎣2−10…00−12−10…00−12−1…0⋱⋱⋱⋱0…0−12−100…0−11⎤⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥⎦ (50)

The prior parameters are taken to be independent and are distributed as:

 ω0 ∼lnN(⋅|μ=2π,σ=0.25) (51) α ∼lnN(⋅|μ=0.1,σ=0.01) (52) β ∼lnN(⋅|μ=10−4,σ=10−5) (53) a0 ∼N(⋅|μ=0,σ=40% gravitational % acceleration) (54) ω ∼lnN(⋅|μ=2π,σ=0.25) (55)

These choices are all reasonable and what is expected to be encountered in typical situations, although they have not been chosen to model any specific physical system. While the stochastic forward model is different for structures with a different number of stories, the same prior is used for each structure considered in this example.

Table 1 shows the results of solving the relaxed, convex optimization problem for structures with up to 8 degrees-of-freedom on which up to 2 sensors are to be installed. Once the matrices have been computed and stored for samples of , the optimization algorithm always converges to the same solution irrespective of the initial starting point, . This is expected, since the problem is convex and is guaranteed to have a global minimum. These are exact optimal solutions since they give in each case. For each structure (value of ), the sensor configuration remains stable to changes in the number of time-steps or prior samples used (each chosen as either 1000 or 2000). There is a sensor on the roof in every case.

For a more challenging problem we consider the sensor location scheme for a 50-story shear building () to be instrumented with sensors. Some typical results are presented in Figures 2 and 3. Since this problem involves a much larger number of degrees-of-freedom, there is a higher variance associated with the computation of the expectation of the log-determinant of the sensitivity matrix using prior samples. For this example, this shows up as a difference in the value of the non-binary sensor configuration at stories 26 through 28. It is to be noted, however, that there does not appear to be any ambiguity upon rounding the numbers to their binary values. For example, no sensors should be placed at stories 26 and 27 in Figure 2, or at story 27 in Figure 3. The results show that all floors above the 27th, except floors 40, 43, 46 and 49, and no floor below the 28th except floor 2 should be instrumented.

As a reminder, an exhaustive search for the optimal configuration would have required over 47 trillion objective function evaluations. The gradient-based method requires fewer than 100 evaluations with a tolerance of in MATLAB’s fmincon function. Of course, heuristic methods might have provided a good sub-optimal solution but in this example, optimal convergence is achieved by our method because the implied binary solution given by rounding the results in Figures 2 and 3 give essentially the same objective value as the optimum for the relaxed optimization problem.

### 6.5 Comparison of differently chosen sensor configurations

The number of bits of information gain of the optimal configuration, , obtained using the convex technique over some other configurations is of interest here. Table 2 describes the sensor configurations that are considered for the uniform 50-story structure:

The configuration is the sensor configuration obtained using a greedy sequential placement algorithm that picks the best sensor location for the sensor after sensors are in place, starting with just one sensor. This algorithm is called the forward sequential sensor placement algorithm in (Papadimitriou, 2004). The number of objective function evaluations is in this example, compared with fewer than 100 evaluations for the convex algorithm.

Table 3 shows that it is inefficient to place sensors in the lower stories relative to the higher ones. It turns out in this case that . With the convex relaxation scheme, however, it is immediately understood that the solution is optimal because it came out as a binary result (while the greedy solution is always binary but not necessarily optimal).

The commonly-used sensor configuration given by in Table 2 appears to be significantly sub-optimal for the given model class. This configuration is sometimes popular, however, for considerations beyond the model class being considered; for instance, when it is expected that the behavior of the structure is not well understood, and so not well modeled.

### 6.6 Variation of optimal configuration with parameter sample set

For a Monte Carlo sample size, , of 1000 or 2000, the optimal configuration is consistent for smaller structures (). For larger structures (), however, the optimal configuration may differ from sample set to sample set, presumably because the two sample sizes are not large enough for accurate estimation of the expectation. The severity of the problem increases with the size of the structure and the number of parameters used. It turns out that for this example, the majority of the sensor locations agree. Those in agreement are typically found in the upper half of the structure. This may change with the specific model choices made in formulating the problem, however.

Possible solutions to this problem:

• Select a sufficiently large sample size to avoid variation between sample sets

• Avoid naive Monte Carlo integration and use an integral that provides an estimate with smaller variance. The method would depend on the nature of the integrand. An algorithm similar to simulated annealing could be used to compute the expectation value, such as in (Beck and Zuev, 2013)

• Obtain better design information about the uncertain parameters to obtain a possibly more peaked prior with less variation between samples within a sample set, but this is not easy to accomplish

### 6.7 Variation of optimal configuration with number of observation time-steps

For a fixed parameter sample set, there is a variation in the optimal configuration when the number of observation time steps is changed from 1000 to 2000 for structures with . Sensor locations near the roof typically remain the same. However, below some story, the sensor locations can be quite different. This may be because of the significantly different amount of information provided about the dynamic response when the duration of measurement is doubled.

## 7 Concluding remarks

It was noted that the optimal sensor placement problem for system identification using any of the objective functions of (i) mutual information between the predicted sensor data and the system model parameters; (ii) the expected information gain about the parameters from the data; or (iii) the expected posterior uncertainty (entropy) of the parameters, give equivalent results.

Then the original combinatorial entropy-based minimization problem for optimal sensor placement was cast into a relaxed convex optimization problem and solved efficiently to yield sub-optimal, and sometimes optimal, solutions. Locations for which a sensor was unambiguously present or absent were typically obtained, across differing Monte Carlo samples for performing a required expectation and differing number of sample time steps for the data. Some partly ambiguous locations occurred where sample variability led to slightly different results. Ambiguous locations could be resolved using a brute-force combinatorial approach since the problem size for optimizing over just the ambiguous locations would be reduced and manageable.

In future work, it would be beneficial to explore the effects on the optimal sensor locations of spatial correlation between the predicted data at the possible sensor locations. Such a study could build on the preliminary investigation in Chapter IV of (Bhattacharyya, 2016).

no multiple

## Appendix A Derivatives of the objective function

### a.1 Gradient of the objective

 ∂h∂zi(z) =−∂∂zi∫logdetQ(z,θs)p(θs)dθs (56) =−∫∂logdetQ∂zip(θs)dθs (interchange) (57) =−∫tr(Q−1∂Q∂zi)p(θs)dθs (derivative of log-det) (58) =−∫tr(Q−1(z,θs)Q(i)(θs))p(θs)dθs (evaluation of partial) (59)
 =−∫(Na∑j=1Na∑k=1Q−1jk(z,θs)Q(i)kj(θs))p(θs)dθs (trace of product) (60) =−∫(Na∑j=1Na∑k=1Q−1jk(z,θs)Q(i)jk(θs))p(θs)dθs (symmetry) (61) =−Eθs[Q−1(z,θs):Q(i)(θs)] (notation) (62)

### a.2 Hessian of the objective

 ∂2h∂zp∂zq(z) =−∂∂zqEθs[Na∑j=1Na∑k=1Q−1jk(z,θs)Q(p)kj(θs)] (63) =−Eθs⎡⎣Na∑j=1Na∑k=1∂Q−1jk(z,θs)∂zqQ(p)kj(θs)⎤⎦ (interchanges) (64) =Eθs[Na∑j=1Na∑k=1(Q−1∂Q∂zqQ−1)jk(z,θs)Q(p)jk(θs)] (derivative of inverse) (65) =Eθs[tr(Q−1(z)Q(q)Q−1(z)Q(p))(θs)] (derivative, trace) (66) =Eθs[((Q−1(z)Q(q)):((Q−1(z)Q(p))T)(θs)] (associativity, trace) (67)

### a.3 Numerical approximations

Given samples distributed according to the prior specified by the designer of the sensor installation, the expectation integrals in Equations (38), (62) and (67) may be approximated by their corresponding Monte Carlo estimates.

For the objective function,

 h(z)≈−1NkNk∑k=1logdetQ(z,θ(k)s) (68)

For the gradient of the objective function,

 ∂h∂zi(z)≈−1NkNk∑k=1Na∑j=1Na∑k=1Q−1jk(z,θ(k)s)Q(i)jk(θ(k)s) (69)

Finally, for the Hessian of the objective function,

 ∂2h∂zp∂zq(z)≈1NkNk∑k=1Na∑j=1Na∑k=1(Q−1Q(q)Q−1)jk(z,θ(k)s)Q(p)(θ(k)s)kj (70)

Some computational effort may be spared by noting that,

 Q(z,θs) =Nd∑l=1zlN∑n=1∂xl∂θs∂xl∂θsT(tn,θs) (71) =Nd∑l=1zlQ(l)(θs) (72)

Hence, stored values of may be used to determine .

yes

## References

• Beck and Katafygiotis (1998) Beck, J.L.; Katafygiotis, L.S.

Updating Models and Their Uncertainties. I: Bayesian Statistical Framework.

Journal of Engineering Mechanics 1998, 124, 455–461. doi:black10.1061/(ASCE)0733-9399(1998)124:4(455).
• Beck (2010) Beck, J.L. Bayesian System Identification Based on Probability Logic. Structural Control and Health Monitoring 2010, 17, 825–847. doi:black10.1002/stc.424.
• Vanik et al. (2000) Vanik, M.W.; Beck, J.L.; Au, S.K. Bayesian Probabilistic Approach to Structural Health Monitoring. Journal of Engineering Mechanics 2000, 126, 738–745. doi:black10.1061/(ASCE)0733-9399(2000)126:7(738).
• Zhang and Au (2016) Zhang, F.L.; Au, S.K. Fundamental Two-Stage Formulation for Bayesian System Identification, Part II: Application to Ambient Vibration Data. Mechanical Systems and Signal Processing 2016, 66-67, 43–61. doi:black10.1016/j.ymssp.2015.04.024.
• Beck et al. (1998) Beck, J.L.; Papadimitriou, C.; Au, S.K.; Vanik, M.W. Entropy-Based Optimal Sensor Location for Structural Damage Detection. Proceedings Volume 3325, Smart Structures and Materials 1998: Smart Systems for Bridges, Structures, and Highways; Liu, S.C., Ed., 1998, pp. 161–172. doi:black10.1117/12.310604.
• Papadimitriou et al. (2000) Papadimitriou, C.; Beck, J.L.; Au, S.K. Entropy-Based Optimal Sensor Location for Structural Model Updating. Journal of Vibration and Control 2000, 6, 781–800. doi:black10.1177/107754630000600508.
• Yuen et al. (2001) Yuen, K.V.; Katafygiotis, L.S.; Papadimitriou, C.; Mickleborough, N.C. Optimal Sensor Placement Methodology for Identification With Unmeasured Excitation. Journal of Dynamic Systems, Measurement, and Control 2001, 123, 677. doi:black10.1115/1.1410929.
• Papadimitriou (2004) Papadimitriou, C. Optimal Sensor Placement Methodology for Parametric Identification of Structural Systems. Journal of Sound and Vibration 2004, 278, 923–947. doi:black10.1016/j.jsv.2003.10.063.
• Yuen and Kuok (2015) Yuen, K.V.; Kuok, S.C. Efficient Bayesian Sensor Placement Algorithm for Structural Identification: A General Approach for Multi-Type Sensory Systems. Earthquake Engineering & Structural Dynamics 2015, 44, 757–774. doi:black10.1002/eqe.2486.
• Sun and Büyüköztürk (2015) Sun, H.; Büyüköztürk, O. Optimal Sensor Placement in Structural Health Monitoring Using Discrete Optimization. Smart Materials and Structures 2015, 24, 125034. doi:black10.1088/0964-1726/24/12/125034.
• Argyris et al. (2018) Argyris, C.; Chowdhury, S.; Zabel, V.; Papadimitriou, C. Bayesian Optimal Sensor Placement for Crack Identification in Structures Using Strain Measurements. Structural Control and Health Monitoring 2018, 25, e2137. doi:black10.1002/stc.2137.
• Cover and Thomas (2005) Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2005. doi:black10.1002/047174882X.
• Udwadia (1994) Udwadia, F.E. Methodology for Optimum Sensor Locations for Parameter Identification in Dynamic Systems. Journal of Engineering Mechanics 1994, 120, 368–390. doi:black10.1061/(ASCE)0733-9399(1994)120:2(368).
• Boyd and Vandenberghe (2004) Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: New York, NY, USA, 2004.
• Joshi and Boyd (2009) Joshi, S.; Boyd, S. Sensor Selection via Convex Optimization. IEEE Transactions on Signal Processing 2009, 57, 451–462. doi:black10.1109/TSP.2008.2007095.
• Guestrin et al. (2005) Guestrin, C.; Krause, A.; Singh, A.P. Near-Optimal Sensor Placements in Gaussian Processes.

Proceedings of the 22nd international conference on Machine learning - ICML ’05; ACM Press: New York, New York, USA, 2005; pp. 265–272.

doi:black10.1145/1102351.1102385.
• Beck and Zuev (2013) Beck, J.L.; Zuev, K.M.

Asymptotically Independent Markov Sampling: A New Markov Chain Monte Carlo Scheme for Bayesian Inference.

International Journal for Uncertainty Quantification 2013, 3, 445–474.
• Bhattacharyya (2016) Bhattacharyya, P. Optimal Sensor Placement For Bayesian Parametric Identification Of Structures. PhD thesis, California Institute of Technology, 2016.