Joint Positioning and Radio Map Generation Based on Stochastic Variational Bayesian Inference for FWIPS

05/17/2017 ∙ by Caifa Zhou, et al. ∙ ETH Zurich 0

Fingerprinting based WLAN indoor positioning system (FWIPS) provides a promising indoor positioning solution to meet the growing interests for indoor location-based services (e.g., indoor way finding or geo-fencing). FWIPS is preferred because it requires no additional infrastructure for deploying an FWIPS and achieving the position estimation by reusing the available WLAN and mobile devices, and capable of providing absolute position estimation. For fingerprinting based positioning (FbP), a model is created to provide reference values of observable features (e.g., signal strength from access point (AP)) as a function of location during offline stage. One widely applied method to build a complete and an accurate reference database (i.e. radio map (RM)) for FWIPS is carrying out a site survey throughout the region of interest (RoI). Along the site survey, the readings of received signal strength (RSS) from all visible APs at each reference point (RP) are collected. This site survey, however, is time-consuming and labor-intensive, especially in the case that the RoI is large (e.g., an airport or a big mall). This bottleneck hinders the wide commercial applications of FWIPS (e.g., proximity promotions in a shopping center). To diminish the cost of site survey, we propose a probabilistic model, which combines fingerprinting based positioning (FbP) and RM generation based on stochastic variational Bayesian inference (SVBI). This SVBI based position and RSS estimation has three properties: i) being able to predict the distribution of the estimated position and RSS, ii) treating each observation of RSS at each RP as an example to learn for FbP and RM generation instead of using the whole RM as an example, and iii) requiring only one time training of the SVBI model for both localization and RSS estimation. These benefits make it outperforms the previous proposed approaches.



There are no comments yet.


page 1

page 3

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Fingerprinting based WLAN indoor positioning systems (FWIPSs) have been attracting attention [1] from both academia and industry in the last decades for their advantages in the following two aspects: they do not necessarily require special or additional infrastructure and they have limited error bounds because they yield absolute positions. While other indoor positioning systems (IPSs), such as IPS based on radio frequency identification (RFID), ultra wide band (UWB), supersonic (e.g., Crickets) or infrared signals (e.g., Active Badge), require deploying and operating dedicated infrastructure. FWIPSs use existing WLAN access points (APs) and WiFi enabled devices (e.g., mobile phones, and tablets)[2, 3, 4, 5]. Other infrastructure-free IPSs (e.g., pedestrian dead-reckoning based on the built-in inertial measurement units (IMUs) of the mobile devices), they provide position changes while need to be integrated and the errors of positioning grow with time. [6, 7].

Generally, an FWIPS is realized using two stages: offline and online stage. During the offline stage a reference database representing the reference values of observable features as a function of location is created. This database is often named radio map (RM)). For creating this database, a site survey throughout the region of interest (RoI) usually has to be conducted. In this paper we assume that the RM consists of a set of reference points (RPs) and the corresponding readings of received signal strengths (RSSs) from all visible APs. Because of changes of the indoor environment and configuration of the WLAN, the offline mapping process needs to be carried out frequently to keep the RM up-to-date. This time-consuming and labor-intensive site survey significantly impairs the widespread application of FWIPS apart from academic research.

To reduce the time and labor of RM collection and update, a mathematical model for both position estimation and RM generation based on stochastic variational Bayesian inference (SVBI) is proposed herein. SVBI is employed to project the RSSs from the high-dimensional RSS-space to a much lower-dimensional space of so-called latent variables (Fig. (c)c and Fig. (d)d), from which the position and corresponding RSS can then be simultaneously estimated. Via training SVBI based joint position and RSS estimation model using a collected RM, it can achieve to RM generation (Fig. (e)e) by implementing SVBI based position and RSS estimation model using a neural network (NN

) which we train using Keras

111Keras is provided via

There is no paper about Keras. The advantages of Keras are i) built-in parallel programming on CPU and GPU; ii) stochastic gradient descent based backpropagation algorithms included; and iii) callbacks (e.g., early stopper and dropout) available for training the deep learning model and avoid the easy-over-fitting problem.

, a machine learning library for deep learning.

(a) Original RM
(b) Standardized original RM
(c) Value of latent variable ( dim.)
(d) Value of latent variable ( dim.)
(e) Generated RM
(f) Schematic of () positioning using the generated RM
Fig. 1: An example of stochastic variational Bayesian inference (SVBI) based joint FbP and radio map (RM) generation. Except for Fig. (b)b, the range of the RoI in other five figures are the same as Fig. (a)a. The filled circle denotes the reference point (RP) and the filled color indicates the value of the respective variable depicted in each figure. In Fig. (f)f, the meaning of the symbols are: i) the filled diamond depicts the ground truth of the test point (TP); ii) the found three nearest neighbors of each TP are circled by a unfilled circle; iii) the unfilled triangle represents the estimated position of the TP using (); and iv) the three enlarged parts show the positioning errors. Each of the enlarged part has the different enlarging scale.

In Section II we give an overview on selected previous work related to FWIPSs, especially on publications focusing on fingerprinting based positioning (FbP) using NNs, and on joint FbP and RM generation. Fundamentals and formulations of FbP and NNs are presented in Section III. The proposed SVBI model is analyzed in Section IV. An experimental analysis and a discussion of the proposed approach is given in Section V.

Ii Related Work

We give a brief review on previously published FbP approaches, especially the ones using NNs. Then methods for RM generation and update are presented focusing on those which combine FbP with RM generation.

Ii-a Fingerprinting based positioning

There are two types of FbP approaches: deterministic ones and probabilistic ones. A typical representative of the first type is nearest neighbors () where the estimated position of the user is obtained as the average of the RPs associated with the nearest neighbors of the measured RSS in the RSS space of the RM [8, 9]. Probabilistic approaches [4, 10] use the likelihood of the observed RSS and a proper prior distribution of the position to compute the posterior and estimate the users location, e.g. by applying the maximum a posteriori (MAP) principle.

Here we focus on reviewing FbP methods using NNs. These are the typical FbP approaches utilizing machine learning algorithms, see e.g., [11]. In [12, 13], position estimation methods based on NNs

with 1 hidden layer and nonlinear activation functions, i.e. introducing nonlinear transformation to the input of the

NN, are proposed. The positioning methods presented in above publications are deterministic, and they have been shown these to perform than . Apart from position estimation based on NN with a single hidden layer, Zhang et al. [14] employed a stacked denoising auto-encoder (SDAE), which is a deep neural network (DNN) consisting of multiple hidden layers, to FbP . They train SDAE model using 100 RMs which are built from manual site surveys throughout the RoI at different times and each RM is treated as one training example.

The SVBI based position estimation which we propose herein is also realized using NNs. However, it differs from the previous work in two aspects. On the one hand, it is a probabilistic model for position estimation. On the other hand, it can be trained using individual observation, i.e. each RP with the corresponding RSS readings is used to train the NN, instead of treating the entire RM as only one training example.

Ii-B Joint FbP and Rm generation

In this part, we categorize the radio map construction approaches according to whether they combine FbP and RM generation. Herein generation of RM means that an RM is newly constructed from the already existing RM and additionally measured RSS without associated position. Most of the previous publications are focused on generating the RM without jointly estimating the position [12, 11, 15, 16]. In [15]

, Talvitie et al. apply interpolation and extrapolation of

RSS based on the distances in the coordinate space to generate a RM which includes more RPs than the original RM. This helps to include RPs into the RM which could not be occupied during site survey, e.g. because of access restrictions. The disadvantages of this method are twofold: i) the accuracy of the interpolated/extrapolated RM depends on the number and spatial distribution of the RPs included in the collected RM, and ii) it needs a separate FbP approach for position estimation.

There are few publications addressing simultaneous FbP and RM generation. Feng et al. [17] and Majeed et al. [18] employ compressive sensing with regularization and manifold alignment with geometry perturbation to achieve both FbP and RM generation. In [12], Zhou et al. propose a method using NNs with backpropagation to realize joint position estimation and RM generation. However, the intrinsic discrepancy of the dimensionality between RSS readings and the coordinates requires to train the position estimation and RM generation models separately, i.e. there are two different training processes: one for mapping the RSS to the coordinate space to achieve FbP, and another one for transforming the coordinates to RSS space for RM generation.

The proposed approach makes use of a latent representation of the RSS readings via SVBI in order to achieve both position and RSS estimation. These two estimation processes are implemented jointly. In this way, it differs significantly from previous work.

Iii Fundamentals

The goal of this paper is to propose a probabilistic model for achieving joint FbP and RM generation implemented using a DNN unifiedly. Having referred to related work above, we now introduce the notation and mathematical concepts used lateron.

Iii-a Fingerprinting based positioning

During the offline stage, a site survey is conducted to construct the fingerprinting database, namely the radio map (RM). It consists of collecting the RSS readings of signals from all APs at each RP using a suitable device (e.g., the mobile phone) with the WiFi module. The coordinates of the RPs must be either known beforehand or determined independent while collecting the RSS. Assuming that an RM consisting of RPs and the RSS readings from APs are available at each of these RPs, we can express the RM as , where is the matrix of coordinates of the RPs in dimensions (e.g., ), and is the matrix of the readings of RSS. The row in , i.e. , denotes the coordinates of the RP

and the vector

of RSS values associated with that RP. A fingerprinting based positioning approach (e.g., ) can be interpreted as a mapping from RSS space to coordinate space. At the online stage, is applied to compute the current location of the user from currently measured RSS readings , i.e. .

To evaluate the positioning accuracy, a test dataset, which is not used to train the FbP model, i.e. to set up the RM in the case of , is introduced. Similar to the notation used for the , let test dataset consist of test points (TPs) and collect the corresponding coordinates and RSS values in the matrix . The coordinates in are only used for performance evaluation via computing the error (e.g., mean squared error (MSE), root mean squared error (RMSE), Euclidean distance) between and the coordinates estimated from .

Iii-B Neural networks and auto-encoders

We provide a short introduction to NNs and auto-encoders (AEs) here to support the further analysis. More details about NNs and AEs can be found e.g. [19, 12], and [20] respectively.

Iii-B1 Neural networks

A single layer neural network (SNN) consists of

nodes (neurons) which transform the input vector

into an output vector . The elements of are linearly combined using weights collected in a matrix , shifted using biases and then processed by an activation function (Fig. (a)a). The free parameters of an SNN are the weights, biases and activation function. For using the SNN these free parameters need to be determined by optimization such that the error of the output


is minimized for a given training data. Generally, is chosen, and gradient descent (e.g., BFGS, Levenberg Marquardt) is applied to back propagate the error (e.g., squared error between and the output of the SNN, i.e. ) for training and optimizing the SNN w.r.t. and [21].

Generalizing from the SNN, we can progress to a deep neural network (DNN) that consists of layers, whose corresponding configuration is given the weights, biases, and activation function for each layer, i.e. (Fig. (b)b). Herein the number of nodes in each layer is equal to the number of columns of the weight matrix of the corresponding layer. Similarly, the output of the layers DNN can be written as:


For the subsequent analysis, we build several baseline models (BMs) for FbP using an SNN or a DNN, i.e. the feed-in and feed-out of the NNs are the readings of RSS and the corresponding coordinates of the RP. We analysis the performance of the BMs and compare to other models (See Section V).

Iii-B2 Auto-encoders

A deep network with a denoising capacity is required instead of SNN and DNN if the feed-in is strongly affected by noise[20]. An auto-encoder (AE

) is a self-supervised learning model whose feed-out is the same as the feed-in and it is implementable via two pipelined

DNNs, i.e. an encoder, and a decoder, as shown in Fig. (c)c. AEs and their improved versions (e.g., SDAE) are designed to mitigate the noisy inputs. The encoder transforms the input into a latent representation via the first DNN, denoted as , i.e.:


The decoder reconstructs from the latent representation using another DNN , i.e.:


Techniques for training an AE have been proposed by Lee et al. [22], and Hinton et al. [23]. They have been proved successful at training a DNN to learn their parameters such that minimize the difference between and . Above brief introduction to AE is helpful to understand the reason why SVBI based joint position and RSS estimation can be implemented using NNs (See Section IV).

(a) SNN
(b) DNN
(c) AE
Fig. 2: Schematic depiction of SNN, DNN, and AE

Iv stochastic variational Bayesian inference and its unified implementation

Stochastic variational Bayesian inference (SVBI) is applied to formulate a probabilistic model, which not only can model the posterior, i.e. ( is a vector representing the RSS and is the associated coordinates of the RP), used for FbP, but also is capable of modeling the feed-in of the positioning model, i.e. the density of the readings of RSS 222Actually the density of each observation of RSS is dependent on the location where it is measured, i.e. . Herein we use a simplified notation to denote the density function of the RSS..

is the parameters to determine the assumed distribution (e.g., mean and covariance of a Gaussian distribution). Firstly, the analysis on estimating

using SVBI, including some fundamentals, the optimization scheme and the relationship with AEs, is presented. Then we generalize it to estimate

as well as the joint distribution

. Finally, a unified frameworks illustrating the implementation of SVBI and its application to FWIPS, especially for joint FbP and RM generation, is analyzed

Iv-a Svbi based density estimation

Iv-A1 Fundamentals of Svbi

Assume that each of the collected in have been drawn i.i.d. from by the process of collecting the readings of RSS. Based on the manifold assumption that the hyper-dimensional RSS lies on an intrinsic manifold, whose dimension is much lower than that of the RSS[24], we introduce a latent variable , where is the dimension of the intrinsic manifold, to represent the information content of . The density of can be computed via marginalization over the latent space:


where is the joint distribution. By introducing a mapping distribution, i.e. , where is the vector of parameters of the mapping distribution, (5) can be rewritten as:


where denotes th expectation operator and is the distribution of . can be interpreted as an encoder similar to the one within an AE and represents the distribution of the mapping from to . The goal of SVBI

is to maximize the estimated probability of

, which is equivalent to maximizing the logarithm of the density function of it w.r.t. to . Taking the logarithm of (6) and applying Jensen’s Inequality 333Jensen’s Inequality gives the lower bound of convex functions. Recall that is a convex function if, for , for all and . Conversely, is concave if is convex. to it, the lower bound of (6) is:


The second term on the right hand side (r.h.s) of (7

) is Kullback-Leibler divergence between

and , denoted as and introducing for the variational lower bound of the marginalized likelihood, we obtain:


The maximization of w.r.t. and is equivalent to minimizing the second term of r.h.s of (8), because Kullback-Leibler divergence is non-negative. Often, mean-field approach is applied to approximate and [25]. However, this requires the analytical solutions of the expectations[25]. In fact, the analytical solutions of the expectation is unknown for the most cases.

Instead of maximizing using mean-field approach by computing the analytical solution of the expectation, we use gradient based backpropagation to optimize w.r.t. , i.e. step-wisely searching the optimal values of . The step size is :


where is a diagonal matrix containing the adaptive learning rate. Herein RMSprop is used to compute it[26]. However, the indirect dependency on over which the expectation of the first term on r.h.s of (8) is taken, makes it difficult to compute the gradient of the expectation w.r.t. . To estimate the gradient of the expected reconstruction loss, i.e. the difference between the reconstructed from and the measured , w.r.t. , Schulman et al. proposed stochastic computation graphs444This approach is also named pairwise derivative, infinitesimal perturbation analysis, and stochastic backpropagation in other publications. to calculate the gradient[27, 28]. This approach solves the indirect dependency problem by applying re-parameterization trick. More details can be found in e.g. [25, 29, 28, 30].

The re-parameterization trick is formulated in the following way. Assume that the latent variable can be expressed by another deterministic variable , where is an auxiliary variable distributed corresponding to an independent distribution, i.e. , and is an arbitrary distribution (e.g., Gaussian). Under this assumption, the expected reconstruction loss, i.e. the r.h.s of (7), is independent of the dependency on while taking the expectation over and can be estimated by:


In (10), is the number of Monte Carlo sampling (MCS). In the case of the Kullback-Leibler divergence can be computed analytically, another equation for evaluating (7) is:


Both (10) and (11

) are treated as the surrogate losses (i.e. substitute losses) of the variational lower bound. But the latter surrogation typically has lower variance than the former

[25]. With the surrogate loss, it is easier to compute the gradient of the expected loss. Therefore, stochastic gradient descent approaches (e.g., RMSprop or Adam) can be applied to compute the optimal values of and .

Iv-A2 Ae interpretation

We can refer to as the encoder, which estimates the distribution over all given an observation . From the latent variable, can be reconstructed via sampling from , which is equivalent to the decoding process of the AE. Different from the AEs described in Section III-B, both the encoding and decoding procedures of SVBI are probabilistic.

Iv-A3 Svbi under Gaussian assumptions

Herein we assume that there exists a transformation that transforms to a latent variable

, which is standard Normally distributed, i.e.

. A DNN is employed to approximate this transformation. The variational posterior , i.e. the encoder, is assumed to be a multivariate Gaussian:


where can be written as , is the Cholesky decomposition of . This decomposition is to reduce the number of parameters needed be estimated. In this case,

can be re-parametrized by a linear transformation such that

, where . Therefore, the Kullback-Leibler divergence in (11) can be analytically written as (13) under above assumptions [31].


Instead of directly estimating the distribution of , we use maximum likelihood estimation (MLE) to estimate the values of with the assumed intrinsic distribution of , i.e. . The MLE can be implementable using NNs[32]. Herein we assume that is Gaussian distributed, MLE can be realized by a 2 layers NN, whose activation functions are tangent hyperbole and linear respectively[25, 29]. Based on the MLE, we introduce another surrogate loss of (11) given the feed-in dataset :


Iv-B Generalization of Svbi

In the case of fingerprinting based positioning (FbP), we try to use SVBI to maximize the conditional likelihood . The variational lower bound of it can be directly extended from (7):


In [33], Sohn et al. relax the dependency of on in (15) to be statistically independent from it, i.e. . This relaxation makes it easier to be implemented. SVBI can also be generalized to estimate the joint distribution and its variational lower bound is[34]:


Iv-C Nn based implementation of Svbi and its application to Fwips

Iv-C1 General implementation of Svbi

First, we illustrate a general implementation framework of SVBI based on NNs. As shown in Fig. 3555To transform to , it is reshaped to a lower triangular matrix., the sum of and is the estimated value of the variational lower bound. The scenario consists of two modules: recognition module and generative module. The former module takes as the input and yields the manifold representation of the input. The generative module is fed by the manifold representation and trained to approximate the corresponding feed-out. Both of the modules can be implemented using NNs with different activation functions [25].

Fig. 3: General implementation framework of SVBI

This general implementation framework can be applied to estimate 3 types variational lower bound via changing the feed-in and feed-out: i) if and are equal to , the loss is the estimation of (14); ii) if and are a concatenation of and , and respectively, the loss is an approximation of conditional variational lower bound (15); and iii) if and are both concatenation of and

, it is the approximation of the joint variational lower bound (


Fig. 4: Hybrid SVBI based RSS and position estimation

Iv-C2 Svbi based FbP and Rss estimation

We now focus on applying SVBI based density estimation model jointly with the NN based positioning approach to realize both position and RSS estimation as illustrated in Fig. 4. This joint model consists of two paths: i) the position estimation path, and ii) the RSS estimation path. Both of them has the shared recognition module but each has its own generative module for their respective purpose. In the case that there is a pre-collected RM consisting of RPs with associated RSS values, we can train the joint model in two ways: i) separate training, and ii) joint training. In the former case, we train the SVBI base RSS and position estimation paths separately and use it for the purpose of RSS and position estimation, respectively. As for the latter case, both paths are trained simultaneously. This joint SVBI model can be used for both position and RSS estimation, such that it can also be employed to generate the RM by feeding the trained generative model with the arbitrary variable , where is drawn from .

In this paper, we use Keras, a deep learning library based on Tensorflow

[35], to implement all the models and the simulations.

V Experimental Analysis and Discussion

V-a Testbed

To evaluate the performance of the proposed approaches, an FWIPS is deployed in the building where the Institute of Geodesy and Photogrammetry, ETH Zürich is located [36]. This FWIPS requires no additional installation of new APs because enough signals are available throughout the RoI. These APs are deployed for the purpose to provide the Internet access services. Though we cannot control the configurations of the WLAN and APs, we assume that the settings of them are stable. A site survey was conducted throughout the RoI using a mobile phone, Nexus 6P, with a custom made Android application in order to collect a RM. The RP position are determined during the site survey using a total station (TS), Leica MS60.

Fifteen prisms permanently mounted on the ceiling represent the coordinates frame and allow the TS to determine its position any where within the RoI by resection. A custom made frame holding the smart phone and a 360 mini prism allows the TS to automatically track the smart phone and thus synchronously measuring RSS and position. The position data collected by TS have an accuracy on the mm-level and are thus considered as ground truth during the later analysis.

V-B Simulation results of the models without variational inference

Three baseline models (BMs), illustrated in Fig. 5, for comparison are presented and analyzed, including the training and parameter setups of the NNs, and the evaluation criterion. Instead of training the NNs with the original RSS readings and coordinates, we start with pre-processing of the data. i) The RSS value are normalized to the range of .

is therefore a vector of normalized values; and ii) the coordinates are standardized to have zero mean and unit standard deviation. We denote this process

and the result [21]. An example of standardized RM is shown in Fig. (b)b. Regarding the BMs, the first two BMs, shown in Fig. (a)a and Fig. (b)b respectively, are based on a SNN, whose activation function are chosen linear. The difference between the first two BMs is the way of inverse transformation (). The former transforms the output of the SNN to the original coordinate space after output from the SNN, and the latter realizes the inverse transformation directly within the NN. The third BM is a deep learning based positioning models (DLPM) consisting of a layers DNN pipelined with the SNN. Thus DLPM is a layers NN. The activation functions of the layers DNN and SNN are set as rectified linear unit (ReLu)666ReLu is defined as if otherwise , and linear, respectively. Because we implement all the models using Keras and Tensorflow, there are two more parameters: i) batch size to fit the model, and ii) maximum number of failures of the early stopper, denoted as patience of the early stopper (PoE). They are introduced to improve the training efficiency and avoid the over-fitting777We use Xavier approach [37] to initialize the weights and the biases for the NN, and

adaptive moment estimation

(Adam) and RMS propagation (RMSprop)[26] to optimize the weights and the biases via backpropagation.
. To evaluate and compare the performance of the models, we calculate the RMSE from the Euclidean distances between the estimated coordinates and their ground truth

together with the 95% confidence interval

[38]888If the number of trials is larger than 30, the empirical standard deviation is practically useful [38]..

(a) BM with post-processed scaled output
(b) BM with built-in scaled output
(c) DLPM
Fig. 5: Three baseline models
BM (post) BM (built-in) DLPM
RMSE () 2.922 3.043 1.6340.024
SVBI (Sep.) SVBI (Joint) 1-
RMSE () 1.725 1.622 2.622
TABLE I: Basic results of the BMs

The results are shown in Table I. We evaluated each model 36 times with the same configuration of the NNs and the same training data. Since the NNs are randomly initialized each of the 36 evaluation yielded different parameter values of the NNs and thus slightly different results. The batch size and PoE of the BMs were chosen as 50 and 25, respectively. The DNN consists of 3 layers, whose number of nodes is 128, 64, 32 for the 3 layers, respectively. The performance of the BMs with built-in and separate inverse standardization is comparable. The performance of the latter is slightly better than that of the former. DLPM performs approximately 50% better than that of the BM with post inverse standardization with the data used herein. The reduction of RMSE of DLPM is almost 0.9 comparing to that of 999The algorithm used herein is from Scikit-learn, an open source Python package for machine learning [39]. We use the weighted version of it. With our training RM, the optimal number nearest neighbors is 1..

V-C Results with variational inference

The DNN of the shared recognition module is set up like the DLPM used in the previous subsection. The RSS are mapped to a 4 dimensional latent space 101010We choose the dimension of latent space as 4 by trials. In the future work, we would find the optimal dimension of the latent space by cross validation. (i.e. ). To simplify the model, the number of nodes of the Gaussian coder (depicted in Fig. 3) is the same as the dimension of the latent space. Herein we present the results under the assumption that the covariance matrix of is diagonal to reduce the computational burden.

Fig. 6: Cumulative positioning accuracy

V-C1 Svbi based position estimation

To achieve SVBI based position estimation, we pipeline the shared recognition module with the BM with the post inverse standardization. RMSE of the estimated positions then is 1.725 (second part of Table I (SVBI (Sep.))). It is 40% lower than that of BM(post). Compared to DLPM, RMSE of SVBI (Sep.) is slightly higher, however, from Fig. 6, the cumulative positioning accuracy (CPA), defined as the cumulative density function of the positioning errors, for both models are almost the same until 2 .

V-C2 Svbi based Rm generation

To train NNs for the purpose of RM generation, joint training of both paths (Fig. 4) is required. The configuration of the joint training is as follows: i) the set up of the shared recognition module and the position generative module are identical; ii) the DNN for RSS generative module consists also of 3 layers, but with 32, 64 and 128 nodes, respectively; iii) the number of nodes of the Gaussian coder in the RSS generative module is equal to the number of APs (it is 399 herein); and iv) to train both paths simultaneously, we use the weighted sum of the loss of both paths.

We present the results of joint training in Table I (SVBI(Joint)) and Fig. 6. The RMSE of the positioning error of SVBI(Joint) is slightly lower than that of DLPM. Also the CPA of it within 2.5 is 2% higher than that of DLPM.

The result of positioning error analysis of the generated RM is illustrated in Fig. 7. To compare the generated RM to the original RM, we evaluate independently using the same test dataset with (, Fig. (f)f). The CPA of them within 2 are comparable, and the gap between them is less than 2%. Also, with the generated RM, the CPA within 6 and 10 are over 70% and 90% respectively. This level of positioning accuracy is adequate for many applications requiring room level positioning accuracy (e.g., telling apart different shops in a big mall). The benefits of RM generation method are in two aspects: i) it smooths the changes of RSS values caused by the temporal and spatial changes of the RoI (Fig. (e)e); and ii) it can generate the RM for the region where is not covered by the site survey, it can thus reduce the labor of collecting the RM.

Fig. 7: Comparison of cumulative positioning accuracy using
Results from Positioning error () RSS estimation error (dB)111111The definition of RSS estimation error is as follows. Supposed the generated RSS , the error comparing to its ground truth is:
mean RMSE mean RMSE
[15]121212Because the proposed RMgeneration algorithm in [15] cannot achieve FbP, they use () to evaluate the performance. Appr. 5.5 10.2
[18] 4.7 13
[17]131313In [17], they employ affinity propagation based clustering and AP selection methods before applying compressive sensing to FbP and RM generation. The result shown in Table II is the case that no clustering and AP selection. Appr. 2.2
This paper 1.82 2.70 10.03 10.16
TABLE II: Results of SVBI based RM generation

Comparing our results to the ones reported in [15, 17] and [18], the proposed joint FbP and RM generation seems to provide better results in terms of both positioning and RSS estimation error (Table II). Regarding the positioning error, the mean error of the proposed approach is three times lower than that of distance based inter/extrapolation method in [15] and more than two times lower than that of manifold alignment based method in [18]. The mean positioning error is 20% lower than that of the compressive sensing based approach in [17]. For the RSS estimation error, SVBI based and distance based inter/extrapolation methods obtain comparable performance regarding the mean error. Comparing to the manifold alignment based method, the proposed method herein achieves approximately 25% reduction of RMSE. However, The comparison summarized in Table II is just an indication, since each of the results is based on different data and a different indoor environment.

Vi Conclusion

Stochastic variational Bayesian inference (SVBI) is employed herein to accomplish joint fingerprinting based positioning (FbP) and radio map (RM) generation. The performance in terms of both positioning and RSS estimation error of the proposed method is better than that of previous work. The proposed probabilistic model can be implemented based on deep neural networks (DNNs) and trained jointly for both FbP and RM generation. Compared to the FbP approaches based on the single layer neural network (SNN), DNN and nearest neighbors (), the proposed SVBI based position estimation outperforms them. The reduction of root mean squared error (RMSE) of the localization is up to 40% comparing to that of SNN based FbP. And the cumulative positioning accuracy within 4 of the proposed FbP method is up to 92% w.r.t. the positioning error, which is approximately 8% higher than of within 4 . Regarding the performance of SVBI based RM generation, it is comparable to that of the manually collect RM and adequate for the applications, which require room level positioning accuracy.


The China Scholarship Council (CSC) supports the first author during this doctoral studies and the second author as a visiting student at ETH Zürich.


  • [1] S. Adler, S. Schmitt, K. Wolter, and M. Kyas, “A survey of experimental evaluation in indoor localization research,” Indoor Positioning and Indoor Navigation (IPIN), 2015 International Conference on, no. October, pp. 1–10, 2015.
  • [2] C. Lee, Y. Chang, G. Park, J. Ryu, S.-G. Jeong, S. Park, J. W. Park, H. C. Lee, K. shik Hong, and M. H. Lee, “Indoor positioning system based on incident angles of infrared emitters,” in 30th Annual Conference of IEEE Industrial Electronics Society, 2004. IECON 2004, vol. 3, Nov 2004, pp. 2218–2222 Vol. 3.
  • [3] M. Hazas and A. Hopper, “Broadband ultrasonic location systems for improved indoor positioning,” Mobile Computing, IEEE Transactions on, vol. 5, no. 5, pp. 536–547, 2006.
  • [4] M. Youssef and A. Agrawala, “The Horus location determination system,” Wireless Networks, vol. 14, no. 3, pp. 357–374, 2008.
  • [5] H. Liu, H. Darabi, P. Banerjee, and J. Liu, “Survey of wireless indoor positioning techniques and systems,” Systems, Man, and Cybernetics, Part C: Applications and Reviews, IEEE Transactions on, vol. 37, no. 6, pp. 1067–1080, 2007.
  • [6] E. Foxlin, “Pedestrian tracking with shoe-mounted inertial sensors,” IEEE Computer Graphics and Applications, vol. 25, no. 6, pp. 38–46, nov 2005.
  • [7] H. Wang, H. Lenz, A. Szabo, J. Bamberger, and U. D. Hanebeck, “WLAN-Based Pedestrian Tracking Using Particle Filters and Low-Cost MEMS Sensors,” in 2007 4th Workshop on Positioning, Navigation and Communication, vol. 2007, 2007, pp. 1–7.
  • [8] P. B. Padmanabhan, V. N., and V. N., “RADAR: An in-building RF based user location and tracking system,” Proceedings IEEE INFOCOM 2000. Conference on Computer Communications. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies (Cat. No.00CH37064), vol. 2, no. c, pp. 775–784, 2000. [Online]. Available:
  • [9] V. Honkavirta, T. Perala, S. Ali-Loytty, and R. Piche, “A comparative survey of wlan location fingerprinting methods,” in 2009 6th Workshop on Positioning, Navigation and Communication, March 2009, pp. 243–251.
  • [10] S. Niedermayr, A. Wieser, and H. Neuner, “Expressing location uncertainty in combined feature-based and geometric positioning,” in Proceedings European Navigation Conference 2014.   s.l.: EUGIN, 2014, pp. 154–.
  • [11] S. He and S. H. G. Chan, “Wi-Fi fingerprint-based indoor positioning: Recent advances and comparisons,” IEEE Communications Surveys and Tutorials, vol. 18, no. 1, pp. 466–490, 2016.
  • [12] C. Zhou and A. Wieser, “Application of backpropagation neural networks to both stages of fingerprinting based WIPS,” in 2016 Fourth International Conference on Ubiquitous Positioning, Indoor Navigation and Location Based Services (UPINLBS).   Picataway, NJ: IEEE, 2017, pp. 207–217.
  • [13] J. Xu, H. Dai, and W.-h. Ying, “Multi-layer neural network for received signal strength-based indoor localisation,” IET Communications, vol. 10, no. 6, pp. 717–723, 2016. [Online]. Available:
  • [14] W. Zhang, K. Liu, W. Zhang, Y. Zhang, and J. Gu, “Deep Neural Networks for wireless localization in indoor and outdoor environments,” Neurocomputing, vol. 194, pp. 279–287, 2016. [Online]. Available:
  • [15] J. Talvitie, M. Renfors, and E. S. Lohan, “Distance-based interpolation and extrapolation methods for RSS-based localization with indoor wireless signals,” IEEE Transactions on Vehicular Technology, vol. 64, no. 4, pp. 1340–1353, 2015.
  • [16] M. M. Atia, A. Noureldin, and M. J. Korenberg, “Dynamic online-calibrated radio maps for indoor positioning in wireless local area networks,” IEEE Transactions on Mobile Computing, vol. 12, no. 9, pp. 1774–1787, 2013.
  • [17] C. Feng, W. S. A. Au, S. Valaee, and Z. Tan, “Received-signal-strength-based indoor positioning using compressive sensing,” IEEE Transactions on Mobile Computing, vol. 11, no. 12, pp. 1983–1993, 2012.
  • [18] K. Majeed, S. Sorour, T. Y. Al-Naffouri, and S. Valaee, “Indoor localization and radio map estimation using unsupervised manifold alignment with geometry perturbation,” IEEE Transactions on Mobile Computing, vol. 15, no. 11, pp. 2794–2808, 2016.
  • [19] H. B. Demuth, M. H. Beale, O. De Jess, and M. T. Hagan, Neural network design.   Martin Hagan, 2014.
  • [20]

    P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P.-A. Manzagol, “Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion,”

    Journal of Machine Learning Research, vol. 11, no. Dec, pp. 3371–3408, 2010.
  • [21] Y. A. LeCun, L. Bottou, G. B. Orr, and K. R. Müller, “Efficient backprop,”

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

    , vol. 7700 LECTU, pp. 9–48, 2012.
  • [22] H. Lee, C. Ekanadham, and A. Y. Ng, “Sparse deep belief net model for visual area v2,” in Advances in neural information processing systems, 2008, pp. 873–880.
  • [23] G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” science, vol. 313, no. 5786, pp. 504–507, 2006.
  • [24]

    T. Pulkkinen, T. Roos, and P. Myllymäki, “Semi-supervised learning for wlan positioning,” in

    International Conference on Artificial Neural Networks.   Springer, 2011, pp. 355–362.
  • [25] D. P. Kingma and M. Welling, “Auto-Encoding Variational Bayes,” arXiv preprint arXiv:1312.6114, no. Ml, pp. 1–14, 2013. [Online]. Available:
  • [26] D. P. Kingma and J. L. Ba, “Adam: a Method for Stochastic Optimization,” International Conference on Learning Representations 2015, pp. 1–15, 2015.
  • [27] J. Schulman, N. Heess, T. Weber, and P. Abbeel, “Gradient Estimation Using Stochastic Computation Graphs,” Nips, pp. 1–13, 2015. [Online]. Available:
  • [28] J. Paisley, D. Blei, and M. Jordan, “Variational Bayesian Inference with Stochastic Search,” Icml, no. 2000, pp. 1367–1374, 2012. [Online]. Available:
  • [29] D. J. Rezende, S. Mohamed, and D. Wierstra, “Stochastic backpropagation and approximate inference in deep generative models,” JMLR: W&CP, vol. 32, pp. 1278–1286, 2014.
  • [30] L. Devroye, “Sample-based Non-uniform Random Variate Generation,” in Proceedings of the 18th Conference on Winter Simulation, ser. WSC ’86.   New York, NY, USA: ACM, 1986, pp. 260–265. [Online]. Available:
  • [31] K. B. Petersen and M. S. Pedersen, “The matrix cookbook,” nov 2012, version 20121115. [Online]. Available:
  • [32] C. M. Bishop, Pattern Recognition and Machine Learning (Information Science and Statistics).   Secaucus, NJ, USA: Springer-Verlag New York, Inc., 2006.
  • [33] K. Sohn, H. Lee, and X. Yan, “Learning Structured Output Representation using Deep Conditional Generative Models,” Advances in Neural Information Processing Systems, pp. 3483–3491, 2015.
  • [34] D. Kingma, D. Rezende, and M. Welling, “Semi-supervised Learning with Deep Generative Models,” arXiv preprint, pp. 1–9, 2014. [Online]. Available:
  • [35] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean, M. Devin, S. Ghemawat, I. Goodfellow, A. Harp, G. Irving, M. Isard, Y. Jia, R. Jozefowicz, L. Kaiser, M. Kudlur, J. Levenberg, D. Mané, R. Monga, S. Moore, D. Murray, C. Olah, M. Schuster, J. Shlens, B. Steiner, I. Sutskever, K. Talwar, P. Tucker, V. Vanhoucke, V. Vasudevan, F. Viégas, O. Vinyals, P. Warden, M. Wattenberg, M. Wicke, Y. Yu, and X. Zheng, “TensorFlow: Large-scale machine learning on heterogeneous systems,” 2015, software available from [Online]. Available:
  • [36] Y. Gu, C. Zhou, A. Wieser, and Z. Zhou, “Pedestrian Positioning Using WiFi Fingerprints and A Foot-mounted Inertial Sensor,” arXiv preprint arXiv:1704.03346, no. 1, pp. 1–9, 2017.
  • [37] X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” Proceedings of the 13th International Conference on Artificial Intelligence and Statistics (AISTATS), vol. 9, pp. 249–256, 2010.
  • [38] C. A. Peters, “Statistics for analysis of experimental data,” Environmental Engineering Processes Laboratory Manual, pp. 1–25, 2001.
  • [39] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay, “Scikit-learn: Machine learning in Python,” Journal of Machine Learning Research, vol. 12, pp. 2825–2830, 2011.