Skip to main content

Infinite Mixtures of Gaussian Process Experts with Latent Variables and its Application to Terminal Location Estimation from Multiple-Sensor Values

  • Conference paper
  • First Online:
Book cover Intelligent Systems and Applications (IntelliSys 2020)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1252))

Included in the following conference series:

  • 813 Accesses

Abstract

This study proposes a probabilistic method that estimates the locations of sensor nodes (terminals) in a wireless sensor network using the multi-sensor data from sensors located on terminals and the hop counts between terminals. The proposed method involves the use of a novel probabilistic generative model, called the infinite mixture model of Gaussian process experts with latent input variables (imGPE-LVM) that enables us to inversely infer the value of an explanatory variable from that of a response variable for a piece-wise continuous function. Based on an imGPE-LVM, where the sensor data measured by sensors on a terminal are represented by observed variables and the location of the terminal is represented by a latent variable, the proposed method maximizes the posterior probability of the latent variable given sensor values with assuming the terminal location estimated by the DV-Hop algorithm as a prior. This method yields more precise estimates of terminal locations compared with the localization techniques utilizing a Gaussian process latent variable model and those using the DV-Hop algorithm, which is solely based on the hop counts between two terminals.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, Heidelberg (2006)

    MATH  Google Scholar 

  2. Duane, S., Kennedy, A.D., Pendleton, B.J., Roweth, D.: Hybrid Monte Carlo. Phys. Lett. B 195(2), 216–222 (1987)

    Article  MathSciNet  Google Scholar 

  3. Ferris, B., Hähnel, D., Fox, D.: Gaussian processes for signal strength-based location estimation. In: Proceedings of Robotics: Science and Systems (2006)

    Google Scholar 

  4. Ferris, B., Fox, D., Lawrence, N.: WiFi-SLAM using Gaussian process latent variable models. In: Proceedings of the 16th International Joint Conference on Artificial Intelligence, pp. 2480–2485 (2007)

    Google Scholar 

  5. Geman, S., Geman, D.: Stochastic relaxation. Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. Pattern Anal. Mach. Intell. (6), 721–741 (1984)

    Google Scholar 

  6. Mistry, H.P., Mistry, N.H.: RSSI based localization scheme in wireless sensor networks: a survey. In: Proceedings of the 5th International Conference on Advanced Computing and Communication Technologies, pp. 647–652 (2015)

    Google Scholar 

  7. Iwata, T., Yamada, T., Sakurai, Y., Ueda, N.: Sequential modeling of topic dynamics with multiple timescales. ACM Trans. Knowl. Discov. Data 5(4), 19 (2012)

    Article  Google Scholar 

  8. Iwata, T., Duvenaud, D., Ghahramani, Z.: Warped mixtures for nonparametric cluster shapes. In: Proceedings of the 29th Conference on Uncertainty in Artificial Intelligence, pp. 311–319 (2013)

    Google Scholar 

  9. Lawrence, N.D.: Gaussian process latent variable models for visualization of high dimensional data. In: Advances in Neural Information Processing Systems, pp. 329–336 (2004)

    Google Scholar 

  10. Lawrence, N.: Probabilistic non-linear principal component analysis with Gaussian process latent variable models. J. Mach. Learn. Res. 6, 1783–1816 (2005)

    MathSciNet  MATH  Google Scholar 

  11. Murphy, K.P.: Machine Learning: A Probabilistic Perspective. The MIT Press, Cambridge (2012)

    MATH  Google Scholar 

  12. Niculescu, D., Nath, B.: DV based positioning in ad hoc networks. Telecommun. Syst. 22(1–4), 267–280 (2003)

    Article  Google Scholar 

  13. Rasmussen, C.E., Ghahramani, Z.: Infinite mixtures of Gaussian process experts. In: Advances in Neural Information Processing Systems, pp. 881–888 (2002)

    Google Scholar 

  14. Schwaighofer, A., Grigoras, M., Tresp, V., Hoffmann, C.: GPPS: a Gaussian process positioning system for cellular networks. In: Advances in Neural Information Processing Systems, pp. 579–586 (2003)

    Google Scholar 

  15. Takizawa, Y., Takashima, Y., Adachi, N.: Self-Organizing Localization for wireless sensor networks based on neighbor topology. In: Proceedings of the Seventh International Conference on Mobile Ubiquitous Computing, Systems, Services and Technologies, pp. 102–108 (2013)

    Google Scholar 

  16. Wang, J., Hertzmann, A., Blei, D.M.: Gaussian process dynamical models. In: Advances in Neural Information Processing Systems, pp. 1441–1448 (2006)

    Google Scholar 

  17. Zhou, Z., Xiao, M., Liu, L., Chen, Y., Lv, J.: An improved DV-HOP localization algorithm. In: Proceedings of the Second International Symposium on Information Science and Engineering, pp. 598–602 (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ryo Hanafusa .

Editor information

Editors and Affiliations

Appendices

Appendix

A Infinite Mixture of Gaussian Process Experts

In the infinite mixture of Gaussian process experts [13], the input space is (probabilistically) divided by a gating function into regions within which specific separate experts make predictions. Using Gaussian process (GP) models as experts, the model has an additional advantage that the computation for each expert is cubic only in the number of data points in its region, rather than in the entire dataset. Each GP-expert learns different characteristics of the function (e.g., lengths scales, noise variances, etc.).

Let \(y \in \mathbb {R}\) be an output variable, let \(x \in \mathbb {R}\) be an input variable, and let c be the discrete indicator variable assigning a data point to an expert. The joint probability of the output data \(\mathbf{y} = \{ y^{(i)} \}_{i=1}^{N}\) and the corresponding indicators \(\mathbf{c} = \{ c^{(i)} \}_{i=1}^{N}\) given the input data \(\mathbf{x} = \{ x^{(i)} \}_{i=1}^{N}\) is expressed by

$$\begin{aligned} \begin{aligned}&p ( \mathbf { y }, \mathbf{c} | \mathbf { x } , {\varvec{\theta }} ) \\&= \left[ \prod _ { j } p \left( \left\{ y ^ { (i) }: c ^ { (i) } = j \right\} | \left\{ x ^ { (i) } : c ^ { (i) } = j \right\} , \theta ^ { (j) } \right) \right] p ( \mathbf { c } | \mathbf { x } , \phi ), \end{aligned} \end{aligned}$$
(13)

where \({\varvec{\theta }} = \{ \theta ^{(j)} \}\) denotes the set of the parameters in the Gaussian process regressor for each of the experts.

In inference, with the posterior of \(\mathbf{c}\) given, the input data \(\mathbf{x}\) and the output data \(\mathbf{y}\) is calculated by Gibbs sampling. Once the expert each data point is assigned to using the maximum posterior estimation for the posterior of \(\mathbf{c}\) is determined, we obtain the Gaussian process regression by class. For a new input, we output the expectation of the predictions produced by the experts with respect to the posterior distributions of the classes.

B Distance Vector Hop

The distance vector hop (DV-Hop) algorithm [12, 17] estimates terminal locations using the hop count between a terminal and an anchor node with known location in a wireless sensor network. It consists of the following stages:

  1. 1.

    Initially, all anchors transmit their locations to the other terminals in the network. The messages are propagated hop by hop where a hop counter is included in the message. Each node maintains an information table on the anchor nodes and counts the minimum number of hops that separates it from each anchor.

  2. 2.

    When an anchor node receives a message from another, it estimates the average distance in terms of hops using the locations of two anchors and the hop counter, and returns it to the network as a correction factor. An anchor estimates the average distance of a hop accompanied by reception of hop counts from anchor nodes of all unknown terminals, this computes the distance to the anchor node based on hop counts and the minimum hop count.

  3. 3.

    The DV-Hop uses a multilateration method to calculate the unknown terminal’s location according to the distance to each anchor node obtained in the stage 2.

C Details on Experiment

1.1 C.1 Experimental Conditions

  1. 1.

    The scale of the environment is assumed to be \(10.00\ \)m. That is, the environment is an area of 10.00 m \(\times 10.00\) m.

  2. 2.

    The wireless communicable range is assumed to be \(10.00 \times \sqrt{2}/3\ \)m. This setting enables us to prevent a terminal from communicating with all other terminals directly and to preventing generation of a terminal incapable of communicating with others for a small number of terminals.

  3. 3.

    The three terminals used as anchor nodes are randomly selected from the terminals in the environment that are at least \(10.00 \times \sqrt{2}/2 \ \)m away from each other. This constraint on the selection of the three anchors causes the DV-Hop to use the locations of the three and the sensor values on them.

  4. 4.

    We used three kinds of sensors, a microphone, a thermometer, and an illuminometer. Referring to the specification of ADMP504 (microphone), that of LM35D (thermometer), that of PICMD01-1 (illuminometer), we assumed their dynamic ranges are: 20–120 (dB) for the microphone, 0–100 (\(^\circ \)C) for the thermometer, and 0–1500 (lux) for the illuminometer, with errors of \(\pm 1\) (dB) for the microphone, \(\pm 1\) (\(^\circ \)C) for the thermometer, and \(\pm 5\) (lux) for the illuminometer.

  5. 5.

    As sensor values, we used the values sampled from the distributions of physical quantities at the true terminal locations added with the Gaussian noises with means of 0 and standard deviations of one third of the sensor errors.

1.2 C.2 Parameter Settings in Inference in imGPE-LVM

In Sect. 3.3, we described the inference in imGPE-LVM. This appendix describes the parameter settings in the experiment.

In sampling, the sample size is 20, 000 and the burn-in size is 2, 000. The initial locations of the terminals \(\mathbf{x}^{(i)}\), \({i=1}\), \(\ldots \), N, are the locations estimated by DV-Hop, that are denoted by \(\mathbf{x}_\mathrm{dvh}^{(i)}\), \({i=1}\), \(\ldots \), N. The initial classes of the experts \(c^{(i)}\), \({i=1}\), \(\ldots \), N, are set to 0. Those of the parameters of the Gaussian process regression for class 0 are \(\theta _0^{(0)} = 1.00\), \(\theta _1^{(0)} = 1.00\), and \(\theta _2^{(0)} = 1.00 \times 10^{2}\). Those of the parameters, \(\alpha \) and \(\phi \), of the hyper-parameters for the Chinese restaurant process are \(\alpha = 1.00\) and \(\phi = 2.50 \times 10^{-3}\). We set the kernel parameters so that the class 0 is removed during sampling.

In sampling by the Hamiltonian Monte Carlo, (1) for the terminal locations \(\mathbf{x}^{(i)}\), \({i=1}\), \(\ldots \), N, the step size is set to \(1.00 \times 10^{-2}\) and the number of the leap frog steps is set to 10, (2) for the kernel parameters \({\varvec{\theta }}_0\), \({\varvec{\theta }}_1\), and \({\varvec{\theta }}_2\) in the Gaussian process regression, the step size is set to \(1.00 \times 10^{-2}\) and the number of the leap frog steps is set to 2, (3) for the hyper-parameters \(\alpha \) in the Chinese restaurant process, the step size is set to \(1.00 \times 10^{-3}\) and the number of the leap frog steps is set to 2, and (4) for the hyper-parameters \(\phi \) in the Chinese restaurant process, the step size is set to \(1.00 \times 10^{-2}\) and the number of the leap frog steps is set to 2. The parameters of the hyper-parameters \(a_{\alpha }\), \(b_{\alpha }\), \(a_{\phi }\), and \(b_{\phi }\) are fixed to \(a_{\alpha } = 1.00 \times 10^{-2}\), \(b_{\alpha } = 1.00 \times 10^{2}\), \(a_{\phi } = 1.00 \times 10^{-2}\), and \(b_{\phi } = 1.00 \times 10^{2}\), because the hyper-parameters \(\alpha \) and \(\phi \) obey the noninformative Gamma distribution.

The parameters, \(b_{\theta _0}\), \(b_{\theta _1}\), and \(b_{\theta _2}\), of the hyper-parameters are fixed to 1.00. The initial values of the parameters, \(a_{\theta _0}\), \(a_{\theta _1}\), and \(a_{\theta _2}\), the hyper-parameters are set to \(a_{\theta _0} = 1.00 \times 10^{1}\), \(a_{\theta _1} = 1.00 \times 10^{1}\), and \(a_{\theta _2} = 1.00\).

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Hanafusa, R., Ebara, J., Okadome, T. (2021). Infinite Mixtures of Gaussian Process Experts with Latent Variables and its Application to Terminal Location Estimation from Multiple-Sensor Values. In: Arai, K., Kapoor, S., Bhatia, R. (eds) Intelligent Systems and Applications. IntelliSys 2020. Advances in Intelligent Systems and Computing, vol 1252. Springer, Cham. https://doi.org/10.1007/978-3-030-55190-2_24

Download citation

Publish with us

Policies and ethics