Skip to main content

Machine Learning of User Attentions in Sensor Data Visualization

  • Conference paper
  • First Online:

Abstract

In this paper, we propose a method for automatically estimating important points of large sensor data by collecting attention points of the user when visualized, and applying a supervised machine-learning algorithm. For large-scale sensor data, it is difficult to find important points simply through visualization, because such points are buried in a large scope of visualization. We also provide the results of an estimation, the accuracy of which was over 80% for multiple visualizations. In addition, the method has the advantage that the trained model can be reused to any other visualization from the same type of the sensors. We show the results of such reusability for the new type of visualization, which achieved an accuracy rate of 70–80%.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    http://www.ti.com/tool/cc2650stk.

  2. 2.

    https://plot.ly/javascript/.

References

  1. Chahuara, P., Fleury, A., Portet, F., Vacher, M.: On-line human activity recognition from audio and home automation sensors: comparison of sequential and non-sequential models in realistic smart homes. J. Ambient Intell. Smart Environ. 8(4), 399–422 (2016)

    Article  Google Scholar 

  2. Ueda, K., Suwa, H., Arakawa, Y., Yasumoto, K.: Exploring accuracy-cost tradeoff in in-home living activity recognition based on power consumptions and user positions. In: 14th IEEE International Conference on Ubiquitous Computing and Communications (IUCC 2015), pp. 1131–1137 (2015)

    Google Scholar 

  3. Bikakis, N., Sellis, T.: Exploration and visualization in the web of big linked data: a survey of the state of the art. In: LWDM (2016)

    Google Scholar 

  4. Pang, X., Cao, Y., Lau, R.W.H., Chan, A.B.: Directing user attention via visual flow on web designs. ACM Trans. Graph. (TOG) 36, 240 (2016)

    Google Scholar 

  5. Bylinskii, Z., Kim, N.W., O’Donovan, P., Alsheikh, S., Madan, S., Pfister, H., Durand, F., Russell, B., Hertzmann, A.: Learning visual importance for graphic designs and data visualizations. In: Proceedings of 30th Annual ACM Symposium on User Interface Software & Technology (2017)

    Google Scholar 

  6. Kim, N.W., Bylinskii, Z., Borkin, M.A., Gajos, K.Z., Oliva, A., Durand, F., Pfister, H.: BubbleView: an interface for crowdsourcing image importance maps and tracking visual attention. ACM Trans. Comput.-Hum. Interact. 24, 36 (2017). (A Special Issue)

    Article  Google Scholar 

  7. Wongsuphasawat, K., Moritz, D., Anand, A., Mackinlay, J., Howe, B., Heer, J.: Voyager: exploratory analysis via faceted browsing of visualization recommendations. IEEE Trans. Visual. Comput. Graph. 22, 649–658 (2015)

    Article  Google Scholar 

  8. Wongsuphasawat, K., Qu, Z., Moritz, D., Chang, R., Ouk, F., Anand, A., Mackinlay, J., Howe, B., Heer, J.: Voyager 2: augmenting visual analysis with partial view specifications. In: Proceedings of 2017 CHI Conference on Human Factors in Computing Systems (2017)

    Google Scholar 

  9. Walker, J., Borgo, R., Jones, M.W.: TimeNotes: a study on effective chart visualization techniques for time-series data. IEEE Trans. Visual. Comput. Graph. 22(1), 549–558 (2016)

    Article  Google Scholar 

  10. Imoto, M., Itoh, T.: A 3D visualization technique for large scale time-varying data. In: 14th International Conference on Information Visualisation (IV10), pp. 17–22 (2010)

    Google Scholar 

  11. Borkin, M.A., Vo, A.A., Bylinskii, Z., Isola, P., Sunkavalli, S., Oliva, A., Pfister, H.: What makes a visualization memorable? IEEE Trans. Visual. Comput. Graph. 19, 2306–2315 (2013). (Proceedings of InfoVis 2013)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Keita Fujino .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Fujino, K., Inoue, S., Shibata, T. (2018). Machine Learning of User Attentions in Sensor Data Visualization. In: Murao, K., Ohmura, R., Inoue, S., Gotoh, Y. (eds) Mobile Computing, Applications, and Services. MobiCASE 2018. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 240. Springer, Cham. https://doi.org/10.1007/978-3-319-90740-6_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-90740-6_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-90739-0

  • Online ISBN: 978-3-319-90740-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics