skip to main content
10.1145/3282894.3289728acmotherconferencesArticle/Chapter ViewAbstractPublication PagesmumConference Proceedingsconference-collections
research-article

Preliminary Investigation of Object-based Activity Recognition Using Egocentric Video Based on Web Knowledge

Published: 25 November 2018 Publication History

Abstract

This study shows a preliminary investigation of daily activity recognition based on a wearable camera without using training data prepared by a user in her environment. Recently, deep learning frameworks have been publicly available, and we can now easily use deep convolutional neural networks (CNNs) pre-trained on a large image data set. In our method, we first detect objects used in the user's activity from her first-person images using a pre-trained CNN for object recognition. We then estimate an activity of the user using the object detection result because objects used in an activity strongly relate to the activity. To estimate the activity without using training data, we utilize knowledge on the Web because the Web is a repository of knowledge that reflects real-world events and common sense. Specifically, we compute semantic similarity between a list of the detected object names and a name of each activity class based on the Web knowledge. The activity class with the largest similarity value is the estimated activity of the user.

References

[1]
Ling Bao and Stephen S Intille. 2004. Activity recognition from user-annotated acceleration data. In Pervasive 2004. 1--17.
[2]
Pedro F Felzenszwalb, Ross B Girshick, David McAllester, and Deva Ramanan. 2010. Object detection with discriminatively trained part-based models. IEEE Transactions on Pattern Analysis and Machine Intelligence 32, 9 (2010), 1627--1645.
[3]
Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Girshick, Sergio Guadarrama, and Trevor Darrell. 2014. Caffe: Convolutional architecture for fast feature embedding. In ACM Multimedia 2014. 675--678.
[4]
Tilke Judd, Frédo Durand, and Antonio Torralba. 2012. A benchmark of computational models of saliency to predict human fixations. Technical Report.
[5]
Tilke Judd, Krista Ehinger, Frédo Durand, and Antonio Torralba. 2009. Learning to predict where humans look. In ICCV 2009. 2106--2113.
[6]
Joseph Korpela, Kazuyuki Takase, Takahiro Hirashima, Takuya Maekawa, Julien Eberle, Dipanjan Chakraborty, and Karl Aberer. 2015. An energy-aware method for the joint recognition of activities and gestures using wearable sensors. In International Symposium on Wearable Computers (ISWC 2015). 101--108.
[7]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. In NIPS 2012. 1097--1105.
[8]
Jonathan Lester, Tanzeem Choudhury, and Gaetano Borriello. 2006. A practical approach to recognizing physical activities. In Pervasive 2006. 1--16.
[9]
Minghuang Ma, Haoqi Fan, and Kris M. Kitani. 2016. Going Deeper into First-Person Activity Recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10]
Takuya Maekawa, Yasue Kishino, Yasushi Sakurai, and Takayuki Suyama. 2011. Recognizing the Use of Portable Electrical Devices with Hand-worn Magnetic Sensors. In Pervasive 2011. 276--293.
[11]
Takuya Maekawa, Yasue Kishino, Yutaka Yanagisawa, and Yasushi Sakurai. 2012a. Recognizing Handheld Electrical Device Usage with Hand-worn Coil of Wire. In Pervasive 2012. 234--252.
[12]
Takuya Maekawa, Yasue Kishino, Yutaka Yanagisawa, and Yasushi Sakurai. 2012b. WristSense: wrist-worn sensor device with camera for daily activity recognition. In 2012 IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops). 510--512.
[13]
Takuya Maekawa and Shinji Watanabe. 2011. Unsupervised Activity Recognition with User's Physical Characteristics Data. In International Symposium on Wearable Computers (ISWC 2011). 89--96.
[14]
Takuya Maekawa, Yutaka Yanagisawa, Yasue Kishino, Katsuhiko Ishiguro, Koji Kamei, Yasushi Sakurai, and Takeshi Okadome. 2010. Object-based activity recognition with heterogeneous sensors on wrist. In Pervasive 2010. 246--264.
[15]
Takuya Maekawa, Yutaka Yanagisawa, Yasushi Sakurai, Yasue Kishino, Koji Kamei, and Takeshi Okadome. 2009. Web Searching for Daily Living. In SIGIR 2009. 27--34.
[16]
Takuya Maekawa, Yutaka Yanagisawa, Yasushi Sakurai, Yasue Kishino, Koji Kamei, and Takeshi Okadome. 2012. Context-Aware Web Search in Ubiquitous Sensor Environment. ACM Transactions on Internet Technology (ACM TOIT) 11, 3 (2012), 12:1--12:23.
[17]
Matthai Philipose, Kenneth P Fishkin, Mike Perkowitz, Donald J Patterson, Dieter Fox, Henry Kautz, and Dirk Hähnel. 2004. Inferring activities from interactions with objects. IEEE Pervasive Computing 3, 4 (2004), 50--57.
[18]
Hamed Pirsiavash and Deva Ramanan. 2012. Detecting activities of daily living in first-person camera views. In CVPR 2012. 2847--2854.
[19]
John Ross Quinlan. 1993. C4.5: Programs for Machine Learning. Morgan Kaufmann.
[20]
Emmanuel Munguia Tapia, Stephen S Intille, and Kent Larson. 2004. Activity recognition in the home using simple and ubiquitous sensors. In Pervasive 2004. 158--175.
[21]
Po-He Tseng, Ran Carmi, Ian GM Cameron, Douglas P Munoz, and Laurent Itti. 2009. Quantifying center bias of observers in free viewing of dynamic natural scenes. Journal of Vision 9, 7 (2009), 4--4.
[22]
Tim Van Kasteren, Athanasios Noulas, Gwenn Englebienne, and Ben Kröse. 2008. Accurate activity recognition in a home setting. In Ubicomp 2008. 1--9.

Cited By

View all
  • (2021)Exploiting Egocentric Cues for Action Recognition for Ambient Assisted Living ApplicationsEmerging Technologies in Biomedical Engineering and Sustainable TeleMedicine10.1007/978-3-030-14647-4_10(131-158)Online publication date: 18-Aug-2021

Index Terms

  1. Preliminary Investigation of Object-based Activity Recognition Using Egocentric Video Based on Web Knowledge

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    MUM '18: Proceedings of the 17th International Conference on Mobile and Ubiquitous Multimedia
    November 2018
    548 pages
    ISBN:9781450365949
    DOI:10.1145/3282894
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 November 2018

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Activity recognition
    2. egocentric video
    3. object detection

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    MUM 2018

    Acceptance Rates

    MUM '18 Paper Acceptance Rate 37 of 82 submissions, 45%;
    Overall Acceptance Rate 190 of 465 submissions, 41%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)6
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 07 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2021)Exploiting Egocentric Cues for Action Recognition for Ambient Assisted Living ApplicationsEmerging Technologies in Biomedical Engineering and Sustainable TeleMedicine10.1007/978-3-030-14647-4_10(131-158)Online publication date: 18-Aug-2021

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media