Abstract
This paper reports Bento Packaging Activity Recognition Challenge by team “RitsBen” held in the International Conference on Activity and Behavior Computing (ABC 2021). Our approach used an autocorrelation function in the preprocessing to isolate the data since the dataset was given with repetitive activity. We then use a model that implements convolutional layers and LSTM. The final decision is made by majority vote using sigmoid predictions output from all body parts. The loss is calculated using BCEWithLogitsLoss for each body part. The evaluation results showed that average accuracy of 0.123 was achieved among subjects 1, 2, and 3 in leave-one-subject-out manner. However, we did not achieve high accuracy as the possibility that the extraction of repetitive actions was not correct.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Anguita, D., Ghio, A., Oneto, L., Parra, X., Reyes-Ortiz, J.L.: Human activity recognition on smartphones using a multiclass hardware-friendly support vector machine. In: International Workshop on Ambient Assisted Living, pp. 216–223. Springer, Berlin (2012)
Attal, F., Mohammed, S., Dedabrishvili, M., Chamroukhi, F., Oukhellou, L., Amirat, Y.: Physical human activity recognition using wearable sensors. Sensors 15(12), 31314–31338 (2015)
Bayat, A., Pomplun, M., Tran, D.A.: A study on human activity recognition using accelerometer data from smartphones. Procedia Computer Science 34, 450–457 (2014)
Chen, Y., Xue, Y.: A deep learning approach to human activity recognition based on single accelerometer. In: 2015 IEEE International Conference on Systems, Man, and Cybernetics, pp. 1488–1492. IEEE (2015)
Dang, L.M., Min, K., Wang, H., Piran, M.J., Lee, C.H., Moon, H.: Sensor-based and vision-based human activity recognition: a comprehensive survey. Pattern Recogn. 108, 107561 (2020)
Khan, A.M., Lee, Y.K., Lee, S.Y., Kim, T.S.: A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer. IEEE transactions on information technology in biomedicine 14(5), 1166–1172 (2010)
Kohei, A., Sayeda, S.A., Nazmun, N., Haru, K., Paula, L., Sozo, I.: Summary of the bento packaging activity recognition challenge. In: The 3rd International Conference on Activity and Behavior Computing (2021)
Lara, O.D., Labrador, M.A.: A survey on human activity recognition using wearable sensors. IEEE communications surveys & tutorials 15(3), 1192–1209 (2012)
Sayeda, S.A., Kohei, A., Nazmun, N., Haru, K., Paula, L., Sozo, I.: Bento Packaging Activity Recognition Challenge (2021). https://doi.org/10.21227/cwhs-t440
Tsokov, S., Lazarova, M., Aleksieva-Petrova, A.: An evolutionary approach to the design of convolutional neural networks for human activity recognition. Indian Journal of Computer Science and Engineering 12(2), 499–517 (2021)
Yang, J., Nguyen, M.N., San, P.P., Li, X.L., Krishnaswamy, S.: Deep convolutional neural networks on multichannel time series for human activity recognition. In: Twenty-Fourth International Joint Conference on Artificial Intelligence (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Fujii, A., Yoshida, K., Shirai, K., Murao, K. (2022). Bento Packaging Activity Recognition with Convolutional LSTM Using Autocorrelation Function and Majority Vote. In: Ahad, M.A.R., Inoue, S., Roggen, D., Fujinami, K. (eds) Sensor- and Video-Based Activity and Behavior Computing. Smart Innovation, Systems and Technologies, vol 291. Springer, Singapore. https://doi.org/10.1007/978-981-19-0361-8_16
Download citation
DOI: https://doi.org/10.1007/978-981-19-0361-8_16
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-0360-1
Online ISBN: 978-981-19-0361-8
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)