Abstract
In recent years, with advances in hardware devices (e.g., sensors and microprocessors), increases in the maturity of software technology, increases in popularity of the Internet, and decreases in the costs of technologies, embedded systems have been widely used in various applications, including in crop growth monitoring, commodity defect detection, transportation system management, and vital signs monitoring. However, to obtain sufficient information for analysis, various sensors must usually be installed in environments. This requirement can cause many problems, such as changes in the original environment because of the installation of hardware devices, a time-consuming process for setting up the system, high costs given the use of many hardware devices, and difficulty in system maintenance. Therefore, attempts should be made to collect sufficient information by using limited hardware devices. In addition to the hardware equipment used to collect environmental information, software models must be developed for each application scene for analyzing the collected information. Such models are usually designed according to certain environmental conditions and cannot be updated automatically with changes in the environment, which decreases the flexibility and life cycle of the system on which these models are installed. Therefore, in this study, we developed a Quadro-W (QW) learning method to predict human behavior. QW encompasses humans (who), objects (what), locations (where), and time (when). This system obtained QW information only from the data collected by cameras and did not use additional sensors. This study constructed a behavior prediction model on the basis of the obtained QW information. The developed model can not only make predictions based on the initial environment but also update itself with changes in the environment to increase the system flexibility and life cycle.









































Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The data presented in this study are available upon request from the corresponding author.
References
Wang H, Schmid C (2013) Action recognition with improved trajectories. In: 2013 IEEE international conference on computer vision, pp 3551–3558. https://doi.org/10.1109/ICCV.2013.441
Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. In: NIPS'14: proceedings of the 27th international conference on neural information processing systems, vol 1, pp 568–576
Yue-Hei Ng J, Hausknecht M, Vijayanarasimhan S, Vinyals O, Monga R, Toderici G (2015) Beyond short snippets: deep networks for video classification. In: 2015 IEEE conference on computer vision and pattern recognition (CVPR), pp 4694–4702
Qiu Z, Yao T, Mei T (2017) Learning spatio-temporal representation with pseudo-3d residual networks. In: Proceedings of the IEEE international conference on computer vision, pp 5533–5541
Shou Z, Wang D, Chang S-F (2016) Temporal action localization in untrimmed videos via multi-stage CNNs. In: IEEE conference on computer vision and pattern recognition, pp 1049–1058
Lin T, Zhao X, Su H, Wang C, Yang M (2018) BSN: boundary sensitive network for temporal action proposal generation. In: Proceedings of the European conference on computer vision (ECCV), pp 3–19
LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324
Krizhevsky A, Sutskever I, Hinton GE (2017) Imagenet classification with deep convolutional neural networks. Commun ACM 60(6):84–90
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: The 3rd international conference on learning representations (ICLR2015). https://arxiv.org/abs/1409.1556
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Watkins CJ, Dayan P (1992) Q-learning. Mach Learn 8(3–4):279–292
Reinhard E, Adhikhmin M, Gooch B, Shirley P (2001) Color transfer between images. IEEE Comput Gr Appl 21(5):34–41
Li Z, Jing Z, Yang X, Sun S (2005) Color transfer based remote sensing image fusion using non-separable wavelet frame transform. Pattern Recogn Lett 26(13):2006–2014. https://doi.org/10.1016/j.patrec.2005.02.010
Vincent L, Soille P (1991) Watersheds in digital spaces: an efficient algorithm based on immersion simulations. IEEE Trans Pattern Anal Mach Intell 13(6):583–598
Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: IEEE conference on computer vision and pattern recognition, pp 2117–2125
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580–587
Ren S, He K, Girshick R, Sun J (2015) Faster R-CNN: Towards real-time object detection with region proposal networks. In: NIPS’15: proceedings of the 28th international conference on neural information processing systems, vol 1, pp 91–99
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779–788
Liu W et al (2016) SSD: Single shot multibox detector. In: European conference on computer vision, vol 9905. Springer, pp 21–37
Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention, Springer, pp 234–241
Google, Speech-to-Text: Automatic Speech Recognition, Cloud Speech-to-Text. https://cloud.google.com/speech-to-text. erests include mobile computing, intelligent systems, and multimedia
Mnih V, Kavukcuoglu K, Silver D et al (2015) Human-level control through deep reinforcement learning. Nature 518:529–533. https://doi.org/10.1038/nature14236
Xiao Y, Li J, Zhu Y, Li Q (2020) User behavior prediction of social hotspots based on multimessage interaction and neural network. IEEE Trans Comput Soc Syst 7(2):536–545. https://doi.org/10.1109/TCSS.2020.2969484
Li D, Shen D, Kou Y, Nie T (2019) Integrating sign prediction with behavior prediction for signed heterogeneous information networks. IEEE Access 7:171357–171371. https://doi.org/10.1109/ACCESS.2019.2937508
Jiang W, Lv S, Wang Y, Chen J, Liu X, Sun Y (2021) Computational experimental study on social organization behavior prediction problems. IEEE Trans Comput Soc Syst 8(1):148–160. https://doi.org/10.1109/TCSS.2020.3017818
Woźniak M, Połap D (2020) Intelligent home systems for ubiquitous user support by using neural networks and rule-based approach. IEEE Trans Ind Inf 16(4):2651–2658. https://doi.org/10.1109/TII.2019.2951089
Liu Y, Xie DY, Gao Q, Han J, Wang S, Gao X (2019) Graph and autoencoder based feature extraction for zero-shot learning. In: IJCAI, vol 1, no 2, p 6
Library of Congress (2020) Who is credited with inventing the telephone? https://www.loc.gov/everyday-mysteries/item/who-is-credited-with-inventing-the-telephone/. Accessed 07 Jan 2020
He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Analysis Mach Intell 37(9):1904–1916
Acknowledgements
This work was supported in part by the Ministry of Science and Technology (MOST) of Taiwan under Grant MOST 111-2221-E-218-011 and in part by the Allied Advanced Intelligent Biomedical Research Center, STUST from Higher Education Sprout Project, Ministry of Education, Taiwan.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Cheng, ST., Hsu, CW., Horng, GJ. et al. Quadro-W learning for human behavior prediction in an evolving environment: a case study of the intelligent butler technology. J Supercomput 79, 6309–6346 (2023). https://doi.org/10.1007/s11227-022-04899-1
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-022-04899-1