skip to main content
10.1145/3461615.3485433acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
abstract

Detecting Face Touching with Dynamic Time Warping on Smartwatches: A Preliminary Study

Authors Info & Claims
Published:17 December 2021Publication History

ABSTRACT

Respiratory diseases such as the novel coronavirus (COVID-19) can be transmitted through people's face-touching behaviors. One of the official recommendations for protecting ourselves from such viruses is to avoid touching our eyes, nose, or mouth with unwashed hands. However, prior work has found that people touch their face 23 times per hour on average without realizing it. Therefore, in this Late-Breaking Work, we explore a possible approach to help users avoid touching their face in daily life by alerting them through a smartwatch application every time a face-touching behavior occurs. We selected 10 everyday activities including several that should be easy to distinguish from face touching and several that should be more challenging. We recruited 10 participants and asked them to perform each activity repeatedly for 3 minutes at their own pace while wearing a Samsung smartwatch. Based on the collected accelerometer data, we used dynamic time warping (DTW) to distinguish between the two groups of activities (i.e., face-touching and non-face-touching), which is a method well-suited for small datasets. Our findings show that the DTW-based classifier is capable of classifying the activities into two groups with high accuracy (i.e., 99.07% for the user-dependent scenario). We demonstrated that smartwatches have the potential to detect face-touching behaviors with the proposed methodology. Future work can explore other classification approaches, collect larger datasets, and consider other sensors to increase the robustness of our results.

Skip Supplemental Material Section

Supplemental Material

References

  1. Ahmad Akl and Shahrokh Valaee. 2010. Accelerometer-based gesture recognition via dynamic-time warping, affinity propagation, compressive sensing. In 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, 2270–2273. https://doi.org/10.1109/ICASSP.2010.5495895Google ScholarGoogle ScholarCross RefCross Ref
  2. Davide Anguita, Alessandro Ghio, Luca Oneto, Xavier Parra, and Jorge Luis Reyes-Ortiz. 2013. A public domain dataset for human activity recognition using smartphones. In Esann, 3.Google ScholarGoogle Scholar
  3. Donald J Berndt and James Clifford. 1994. Using dynamic time warping to find patterns in time series. In KDD workshop, 359–370.Google ScholarGoogle Scholar
  4. Centers for Disease Control and Prevention. 2020. Prevent Getting Sick. Retrieved March 8, 2021 from https://www.cdc.gov/coronavirus/2019-ncov/prevent-getting-sick/index.htmlGoogle ScholarGoogle Scholar
  5. Xiang “Anthony” Chen. 2020. FaceOff: Detecting Face Touching with a Wrist-Worn Accelerometer. CoRR abs/2008.0. Retrieved from https://arxiv.org/abs/2008.01769Google ScholarGoogle Scholar
  6. Anis Davoudi, Amal Asiri Wanigatunga, Matin Kheirkhahan, Duane Benjamin Corbett, Tonatiuh Mendoza, Manoj Battula, Sanjay Ranka, Roger Benton Fillingim, Todd Matthew Manini, and Parisa Rashidi. 2019. Accuracy of Samsung Gear S Smartwatch for Activity Recognition: Validation Study. JMIR Mhealth Uhealth 7, 2: e11270. https://doi.org/10.2196/11270Google ScholarGoogle ScholarCross RefCross Ref
  7. Avrelle Harrington. 2020. Ready to Finally Break Your Bad Habit? Introducing the New Immutouch Habit Band. Retrieved March 26, 2021 from https://immutouch.posthaven.com/ready-to-finally-break-your-bad-habit-introducing-the-new-immutouch-habit-bandGoogle ScholarGoogle Scholar
  8. Joseph A Himle, Deborah Bybee, Lisa A O'Donnell, Addie Weaver, Sarah Vlnka, Daniel T DeSena, and Jessica M Rimer. 2018. Awareness enhancing and monitoring device plus habit reversal in the treatment of trichotillomania: An open feasibility trial. Journal of obsessive-compulsive and related disorders 16: 14–20.Google ScholarGoogle ScholarCross RefCross Ref
  9. International Data Corporation. 2021. Consumer Enthusiasm for Wearable Devices Drives the Market to 28.4% Growth in 2020, According to IDC. Retrieved March 22, 2021 from https://www.idc.com/getdoc.jsp?containerId=prUS47534521Google ScholarGoogle Scholar
  10. Michael I Jordan and Tom M Mitchell. 2015. Machine learning: Trends, perspectives, and prospects. Science 349, 6245: 255–260.Google ScholarGoogle Scholar
  11. Jennifer R Kwapisz, Gary M Weiss, and Samuel A Moore. 2011. Activity Recognition Using Cell Phone Accelerometers. SIGKDD Explor. Newsl. 12, 2: 74–82. https://doi.org/10.1145/1964897.1964918Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Yen Lee Angela Kwok, Jan Gralton, and Mary-Louise McLaws. 2015. Face touching: a frequent habit that has implications for hand hygiene. American journal of infection control 43, 2: 112–114.Google ScholarGoogle Scholar
  13. Jiayang Liu, Lin Zhong, Jehan Wickramasuriya, and Venu Vasudevan. 2009. UWave: Accelerometer-Based Personalized Gesture Recognition and Its Applications. Pervasive Mob. Comput. 5, 6: 657–675. https://doi.org/10.1016/j.pmcj.2009.07.007Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. A E Macias, A la Torre, S Moreno-Espinosa, P E Leal, M T Bourlon, and G M Ruiz-Palacios. 2009. Controlling the novel A (H1N1) influenza virus: don't touch your face! Journal of Hospital Infection 73, 3: 280–281.Google ScholarGoogle Scholar
  15. F R McInnes, M A Jack, and J Laver. 1989. Template adaptation in an isolated word-recognition system. IEE Proceedings I (Communications, Speech and Vision) 136, 2: 119–126.Google ScholarGoogle ScholarCross RefCross Ref
  16. Wannes Meert, Kilian Hendrickx, and Toon Van Craenendonck. 2020. wannesm/dtaidistance v2.0.0. https://doi.org/10.5281/zenodo.3981067Google ScholarGoogle Scholar
  17. Steffen Moritz, Andras Treszl, and Michael Rufer. 2011. A Randomized Controlled Trial of a Novel Self-Help Technique for Impulse Control Disorders: A Study on Nail-Biting. Behavior Modification 35, 5: 468–485. https://doi.org/10.1177/0145445511409395Google ScholarGoogle ScholarCross RefCross Ref
  18. NASA Jet Propulsion Laboratory. 2020. PULSE: A Pendant to Warn You When You Touch Your Face. Retrieved March 26, 2021 from https://medeng.jpl.nasa.gov/covid-19/pulse/Google ScholarGoogle Scholar
  19. Maria Nicola, Zaid Alsafi, Catrin Sohrabi, Ahmed Kerwan, Ahmed Al-Jabir, Christos Iosifidis, Maliha Agha, and Riaz Agha. 2020. The socio-economic implications of the coronavirus pandemic (COVID-19): A review. International journal of surgery (London, England) 78: 185–193. https://doi.org/10.1016/j.ijsu.2020.04.018Google ScholarGoogle Scholar
  20. Jaime Ruiz and Yang Li. 2011. DoubleFlip: A Motion Gesture Delimiter for Mobile Interaction. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’11), 2717–2720. https://doi.org/10.1145/1978942.1979341Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Tizen Developers. An open source, standards-based software platform for multiple device categories. Retrieved April 12, 2021 from https://developer.tizen.org/Google ScholarGoogle Scholar
  22. Richard P Troiano, James J McClain, Robert J Brychta, and Kong Y Chen. 2014. Evolution of accelerometer methods for physical activity research. British journal of sports medicine 48, 13: 1019–1023. https://doi.org/10.1136/bjsports-2014-093546Google ScholarGoogle Scholar
  23. Radu-Daniel Vatavu, Lisa Anthony, and Jacob O Wobbrock. 2012. Gestures as Point Clouds: A $P Recognizer for User Interface Prototypes. In Proceedings of the 14th ACM International Conference on Multimodal Interaction (ICMI ’12), 273–280. https://doi.org/10.1145/2388676.2388732Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Gary M. Weiss, Jessica L. Timko, Catherine M. Gallagher, Kenichi Yoneda, and Andrew J. Schreiber. 2016. Smartwatch-based activity recognition: A machine learning approach. In 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), 426–429. https://doi.org/10.1109/BHI.2016.7455925Google ScholarGoogle ScholarCross RefCross Ref
  25. Jacob O Wobbrock, Andrew D Wilson, and Yang Li. 2007. Gestures without Libraries, Toolkits or Training: A $1 Recognizer for User Interface Prototypes. In Proceedings of the 20th Annual ACM Symposium on User Interface Software and Technology (UIST ’07), 159–168. https://doi.org/10.1145/1294211.1294238Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    ICMI '21 Companion: Companion Publication of the 2021 International Conference on Multimodal Interaction
    October 2021
    418 pages
    ISBN:9781450384711
    DOI:10.1145/3461615

    Copyright © 2021 Owner/Author

    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 17 December 2021

    Check for updates

    Qualifiers

    • abstract
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate453of1,080submissions,42%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format