Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3544793.3560337acmconferencesArticle/Chapter ViewAbstractPublication PagesubicompConference Proceedingsconference-collections
poster

On the Effectiveness of Virtual IMU Data for Eating Detection with Wrist Sensors

Published: 24 April 2023 Publication History

Abstract

The successful training of human activity recognition (HAR) systems typically heavily depends on the availability of sufficient amounts of labeled sensor data. Unfortunately, obtaining large-scale labeled datasets is usually expensive and often limited by practical and / or privacy reasons. Recently, IMUTube was introduced to tackle this data scarcity problem through generating weakly-labeled, virtual IMU data from unconstrained video repositories, such as YouTube. IMUTube was demonstrated to be very effective at classifying locomotion or gym exercises that involve large movements of body parts. Yet, many important daily activities, such as eating, do not exhibit such substantial body (part) movements but are rather based on more subtle, fine-grained motions. This work explores the utility of IMUTube for such subtle motion activities with specific, exemplary application to eating detection. We found that–surprisingly–IMUTube is also very effective for this challenging HAR domain. Our experiment demonstrates that eating recognition systems benefit from virtual IMU data extracted from video datasets with significant improvements of recognition accuracy (increases of 8.4% and 5.9% F1-score absolute for both curated and in-the-wild video datasets, respectively relative to 71.5% F1-score of the baseline), which is encouraging for the broader use of systems like IMUTube.

References

[1]
Abdelkareem Bedri, Diana Li, Rushil Khurana, Kunal Bhuwalka, and Mayank Goel. 2020. Fitbyte: Automatic diet monitoring in unconstrained situations using multimodal sensing on eyeglasses. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1–12.
[2]
Abdelkareem Bedri, Richard Li, Malcolm Haynes, Raj Prateek Kosaraju, Ishaan Grover, Temiloluwa Prioleau, Min Yan Beh, Mayank Goel, Thad Starner, and Gregory Abowd. 2017. EarBit: using wearable sensors to detect eating episodes in unconstrained environments. Proceedings of the ACM on interactive, mobile, wearable and ubiquitous technologies 1, 3 (2017), 1–20.
[3]
Abdelkareem Bedri, Apoorva Verlekar, Edison Thomaz, Valerie Avva, and Thad Starner. 2015. A wearable system for detecting eating activities with proximity sensors in the outer ear. In Proceedings of the 2015 ACM International Symposium on Wearable Computers. 91–92.
[4]
Shengjie Bi, Tao Wang, Ellen Davenport, Ronald Peterson, Ryan Halter, Jacob Sorber, and David Kotz. 2017. Toward a wearable sensor for eating detection. In Proceedings of the 2017 Workshop on Wearable Systems and Applications. 17–22.
[5]
Wenqiang Chen, Shupei Lin, Elizabeth Thompson, and John Stankovic. 2021. SenseCollect: We Need Efficient Ways to Collect On-body Sensor-based Human Activity Data!Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, 3 (2021), 1–27.
[6]
M. Ester, H. Kriegel, J. Sander, and X. Xu. 1996. A density-based algorithm for discovering clusters in large spatial databases with noise. In kdd, Vol. 96. 226–231.
[7]
Will Kay, Joao Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, 2017. The kinetics human action video dataset. arXiv preprint arXiv:1705.06950(2017).
[8]
Hyeokhyen Kwon, Gregory D Abowd, and Thomas Plötz. 2019. Handling annotation uncertainty in human activity recognition. In Proceedings of the 23rd International Symposium on Wearable Computers. 109–117.
[9]
Hyeokhyen Kwon, Catherine Tong, Harish Haresamudram, Yan Gao, Gregory D Abowd, Nicholas D Lane, and Thomas Ploetz. 2020. Imutube: Automatic extraction of virtual on-body accelerometry from video for human activity recognition. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 4, 3 (2020), 1–29.
[10]
Edison Thomaz, Irfan Essa, and Gregory D Abowd. 2015. A practical approach for recognizing eating moments with wrist-mounted inertial sensing. In Proceedings of the 2015 ACM international joint conference on pervasive and ubiquitous computing. 1029–1040.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
UbiComp/ISWC '22 Adjunct: Adjunct Proceedings of the 2022 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the 2022 ACM International Symposium on Wearable Computers
September 2022
538 pages
ISBN:9781450394239
DOI:10.1145/3544793
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 24 April 2023

Check for updates

Author Tags

  1. Daily Activity Recognition
  2. Eating recognition
  3. Virtual IMU data

Qualifiers

  • Poster
  • Research
  • Refereed limited

Conference

UbiComp/ISWC '22

Acceptance Rates

Overall Acceptance Rate 764 of 2,912 submissions, 26%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 111
    Total Downloads
  • Downloads (Last 12 months)62
  • Downloads (Last 6 weeks)5
Reflects downloads up to 17 Oct 2024

Other Metrics

Citations

Cited By

View all

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media