Abstract
Affective social multimedia computing provides us the opportunity to improve our daily lives. Various things, such as devices in ubiquitous computing environments and autonomous vehicles in real environments considering human beings, can be controlled by analyzing and learning affective social big data. Deep learning is a core learning algorithm for autonomous control; however, it requires huge amounts of learning data, and the process of collecting various types of learning data is expensive. The collection limit of affective social videos for deep learning is resolved by analyzing affective social videos, such as YouTube and Closed Circuit Television (CCTV) videos collected in advance, and generating new affective social videos more as learning data without human beings autonomously controlling other cameras. The control signals of the cameras are generated by Convolutional Neural Network (CNN)-based end-to-end controls. However, images captured consecutively need to be analyzed to improve the quality of the generated control signals. This paper proposes a system that generates affective social videos for deep learning by Convolutional Recurrent Neural Network (CRNN)-based end-to-end controls. The extracted images in affective social videos are utilized for calculating the control signals based on the CRNN. Additional affective social videos are then generated by the extracted consecutive images and camera control signals. The effectiveness of the proposed method was verified in the experiments by comparing the results obtained using the proposed method with those obtained using the traditional CNN. The results showed that the accuracy of the control signals obtained using the proposed method was 56.30% higher than that of the control signals obtained using the traditional CNN.
Similar content being viewed by others
References
Bojarski M, Testa DD, Dworakowski D, Firner B, Flepp B, Goyal P, Jackel LD, Monfort M, Muller U, Zhang J, Zhang X, Zhao J, Zieba K (2016) End to end learning for self-driving cars. https://arxiv.org/abs/1604.07316. Accessed 15 January 2019
Chen Y, Aygün RS (2015) SpriteCam: virtual camera control using sprite. Mutimedia Tools and Applications 74(3):1067–1089. https://doi.org/10.1007/s11042-013-1711-6
Chen C, Seff A, Kornhauser A, Xiao J (2015) DeepDriving: learning affordance for direct perception in autonomous driving. Proceeding the IEEE International Conference on Computer Vision (ICCV). Santiago, Chile:1–9
Codevilla F, Muller M, Lopez A, Koltun V, Dosovitskiy A (2018) End-to-End driving via conditional imitation learning. Proceeding 2018 IEEE International Conference on Robotics and Automation (ICRA2018). Brisbane:1–8
DJI Ground Station https://www.dji.com. Accessed 18 January 2019
Ebeid E, Skriver M, Terkildsen KH, Jensen K, Schultz UP (2018) A survey of open-source UAV flight controllers and flight simulators. Microprocess Microsyst 61:11–20. https://doi.org/10.1016/j.micpro.2018.05.002
GCS Paparazzi https://wiki.paparazziuav.org. Accessed 18 January 2019
Giusti A, Guzzi J, Ciresan DC, He F, Rodríguez JP, Fontana F, Faessler M, Forster C, Schmidhuber J, Caro GD, Scaramuzza D, Gambardella LM (2016) A machine learning approach to visual perception of forest trails for mobile robots. IEEE Robotics and Automation Letters 1(2):661–667. https://doi.org/10.1109/LRA.2015.2509024
Hentati AI, Krichen L, Fourati M, Fourati LC (2018) Simulation tools, environments and frameworks for UAV systems performance analysis. Proceeding 14th International Wireless Communications & Mobile Computing Conference (IWCMC). Limassol, Cyprus:1495–1500
Herzig J, Feigenblat G, Shmueli-Scheuer M, Konopnicki D, Rafaeli A (2016) Predicting customer satisfaction in customer support conversations in social media using affective features. Proceeding 2016 Conference on User Modeling Adaptation and Personalization. Nova Scotia, Canada: 115–119
Hussein A, Elyan E, Gaber MM, Jayne C (2018) Deep imitation learning for 3D navigation tasks. Neural Comput & Applic 29(7):389–404
Kersandt K, Munoz G, Barrado C (2018) Self-training by reinforcement learning for full-autonomous drones of the future. Proceeding IEEE/AIAA 37th Digital Avionics Systems Conference (DASC). London, UK:1–10
Kim J, Chung D, Ko I (2017) A climbing motion recognition method using anatomical information for screen climbing games. Human-centric Computing and Information Sciences 7(25):1–14. https://doi.org/10.1186/s13673-017-0106-5
Lee S, Sung Y, Kim Y, Cha E (2018) Variations of AlexNet and GoogLeNet to improve Korean character recognition performance. Journal of Information Processing Systems 14(1):205–217. https://doi.org/10.3745/JIPS.04.0061
Li Z, Tang J (2016) Weakly supervised deep matrix factorization for social image understanding. IEEE Transaction on Image Processing 26(1):276–288. https://doi.org/10.1109/TIP.2016.2624140
Merino L, Caballero F, Martínez-de-Dios JR, Maza I, Ollero A (2012) An unmanned aircraft system for automatic forest fire monitoring and measurement. Journal of Intelligent & Robotic Systems 65:533–548. https://doi.org/10.1007/s10846-011-9560-x
Mission Planner http://ardupilot.org/planner. Accessed 18 January 2019
Polvara R, Patacchiola M, Sharma S, Wan J, Manning A, Sutton R, Cangelosi A (2018) Toward end-to-end control for UAV autonomous landing via deep reinforcement learning. Proceeding 2018 International Conference on Unmanned Aircraft Systems (ICUAS). Dallas, TX, USA:115–123
Sanchez-Escobedo D, Lin X, Casas JR, Pardas M (2018) Hybridnet for depth estimation and semantic segmentation. Proceeding 2018 IEEE International Conference on Acoustics, Speech and Signal Processing. Alberta, Canada:1563–1567
Shah U, Khawad R, Krishna KM (2016) DeepFly-towards complete autonomous navigation of MAVs with monocular camera. Proceeding the Tenth Indian Conference on Computer Vision, Graphics, and Image Processing (ICVGIP 2016). Indian:1–8
Shah S, Dey D, Lovett C, Kapoor A (2017) AirSim: high-fidelity visual and physical simulation for autonomous vehicles. Proceeding 11th Conference on Field and Service Robotics (FSR 2017). Zurich:1–14
Shah U, Khawd R, Krishna KM (2017) Detecting, localizing, and recognizing trees with a monocular MAV: Towards preventing deforestation. Proceeding 2017 IEEE International Conference on Robotics and Automation (ICRA). Singapore, Singapore:1982–1987
Shi B, Bai X, Yao C (2017) An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition. IEEE Transactions on Pattern Analysis & Machine Intelligence 39(11):2298–2304. https://doi.org/10.1109/TPAMI.2016.2646371
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. https://arxiv.org/pdf/1409.1556.pdf. Accessed 15 January 2019
Smolyanskiy N, Kamenev A, Smith J, Birchfield ST (2017) Toward low-flying autonomous MAV trail navigation using deep neural networks for environmental awareness. Proceeding 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2017). Vancouver, Canada:1–7
Song Y, Kim I (2018) DeepAct: a deep neural network model for activity detection in untrimmed videos. Journal of Information Processing Systems 14(1):150–161. https://doi.org/10.3745/JIPS.04.0059
Su Y, Grauman K (2017) Making 360° video watchable in 2D: learning videography for click free viewing. Proceeding 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI, USA:1368–1376
Sung Y, Kwak J, Yang D, Park Y (2015) Ground station design for the control of multi heterogeneous UAVs. Proceeding 2015 Spring Conference of Korea Multimedia Society. Andong, Repulic of Korea: 828–829
Sung Y, Jin Y, Kwak J, Lee S, Cho K (2018) Advanced camera image cropping approach for CNN-based end-to-end controls on sustainable computing. Sustainability 10(3):1–13. https://doi.org/10.3390/su10030816
Truong MTN, Kim S (2017) Parallel implementation of color-based particle filter for object tracking in embedded systems. Human-centric Computing and Information Sciences 7(2):1–13. https://doi.org/10.1186/s13673-016-0082-1
Wang Y, Wang S, Tang J, Liu H, Li B (2015) Unsupervised sentiment analysis for social media images. Proceeding Twenty-Fourth International Joint Conference on Artificial Intelligence. Buenos Aires, Argentina:2378–2379
Xu A, Liu Z, Guo Y, Sinha V, Akkiraju R (2017) A new chatbot for customer service on social media. Proceeding 2017 CHI Conference on Human Factors in Computing Systems. Colorado, USA: 3506–3510
Yang X, Zhang T, Xu C, Hossain MS (2015) Automatic visual concept learning for social event understanding. IEEE Transactions on Multimedia 17(3):346–358. https://doi.org/10.1109/TMM.2015.2393635
YouTube https://www.youtube.com. Accessed 15 January 2019
Zhou P, Zhou Y, Wu D, Jin H (2016) Differentially private online learning for cloud-based video recommendation with multimedia big data in social networks. IEEE Transactions on Multimedia 18(6):1217–1229. https://doi.org/10.1109/TMM.2016.2537216
Acknowledgements
This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education (2018R1D1A1B07049990).
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Kwak, J., Park, J.H. & Sung, Y. Affective social big data generation algorithm for autonomous controls by CRNN-based end-to-end controls. Multimed Tools Appl 78, 27175–27192 (2019). https://doi.org/10.1007/s11042-019-7703-4
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-019-7703-4