skip to main content
10.1145/3581783.3612223acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Aesthetics-Driven Virtual Time-Lapse Photography Generation

Authors Info & Claims
Published:27 October 2023Publication History

ABSTRACT

Time-lapse videos can visualize the temporal change of dynamic scenes and present wonderful sights with drastic variance in color appearance and rapid movement that interests people. We propose an aesthetics-driven virtual time-lapse photography framework to explore the automatic generation of time-lapse videos in the virtual world, which has potential applications like artistic creation and entertainment in the virtual space. We first define shooting parameters to parameterize the time-lapse photography process and accordingly propose image, video, and time-lapse aesthetic assessments to optimize these parameters, enabling the process to be autonomous and adaptive. We also build an interactive interface to visualize the shooting process and help users conduct virtual time-lapse photography by personalizing shooting parameters according to their aesthetic preferences. Finally, we present a two-stream time-lapse aesthetic model and a time-lapse aesthetic dataset, which can evaluate the aesthetic quality of time-lapse videos. Experimental results demonstrate our method can automatically generate time-lapse videos comparable to those of professional photographers and is more efficient.

Skip Supplemental Material Section

Supplemental Material

ID#mmfp2292-video.mp4

mp4

56.1 MB

References

  1. Eric P Bennett and Leonard McMillan. 2007. Computational time-lapse video. In ACM SIGGRAPH. 102-es.Google ScholarGoogle Scholar
  2. Chia-Chi Cheng, Hung-Yu Chen, and Wei-Chen Chiu. 2020. Time flies: Animating a still image with time-lapse video as reference. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5641--5650.Google ScholarGoogle ScholarCross RefCross Ref
  3. Daniel Cohen-Or, Olga Sorkine, Ran Gal, Tommer Leyvand, and Ying-Qing Xu. 2006. Color harmonization. In ACM SIGGRAPH. 624--630.Google ScholarGoogle Scholar
  4. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Ieee, 248--255.Google ScholarGoogle ScholarCross RefCross Ref
  5. Alex Graves and Alex Graves. 2012. Long short-term memory. Supervised sequence labelling with recurrent neural networks (2012), 37--45.Google ScholarGoogle ScholarCross RefCross Ref
  6. Erik Härkönen, Miika Aittala, Tuomas Kynkäänniemi, Samuli Laine, Timo Aila, and Jaakko Lehtinen. 2022. Disentangling random and cyclic effects in time-lapse sequences. ACM Transactions on Graphics, Vol. 41, 4 (2022), 1--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Identity mappings in deep residual networks. In European Conference on Computer Vision. Springer, 630--645.Google ScholarGoogle ScholarCross RefCross Ref
  8. Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation, Vol. 9, 8 (1997), 1735--1780.Google ScholarGoogle Scholar
  9. Azlan Iqbal, Harold Van Der Heijden, Matej Guid, and Ali Makhmali. 2012. Evaluating the aesthetics of endgame studies: a computational model of human aesthetic perception. IEEE Transactions on Computational Intelligence and AI in Games, Vol. 4, 3 (2012), 178--191.Google ScholarGoogle ScholarCross RefCross Ref
  10. Xin Jin, Hao Lou, Heng Huang, Xinning Li, Xiaodong Li, Shuai Cui, Xiaokun Zhang, and Xiqiao Li. 2022. Pseudo-Labeling and Meta Reweighting Learning for Image Aesthetic Quality Assessment. IEEE Transactions on Intelligent Transportation Systems, Vol. 23, 12 (2022), 25226--25235.Google ScholarGoogle ScholarCross RefCross Ref
  11. Neel Joshi, Wolf Kienzle, Mike Toelle, Matt Uyttendaele, and Michael F Cohen. 2015. Real-time hyperlapse creation via optimal frame selection. ACM Transactions on Graphics, Vol. 34, 4 (2015), 1--9.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Yueying Kao, Chong Wang, and Kaiqi Huang. 2015. Visual aesthetic quality assessment with a regression model. In IEEE International Conference on Image Processing. 1583--1587.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Yan Ke, Xiaoou Tang, and Feng Jing. 2006. The design of high-level features for photo quality assessment. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 1. 419--426.Google ScholarGoogle Scholar
  14. Qi Kuang, Xin Jin, Qinping Zhao, and Bin Zhou. 2019. Deep multimodality learning for UAV video aesthetic quality assessment. IEEE Transactions on Multimedia, Vol. 22, 10 (2019), 2623--2634.Google ScholarGoogle ScholarCross RefCross Ref
  15. Hui-Jin Lee, Ki-Sang Hong, Henry Kang, and Seungyong Lee. 2017. Photo aesthetics analysis via DCNN feature encoding. IEEE Transactions on Multimedia, Vol. 19, 8 (2017), 1921--1932.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Hao Lou, Heng Huang, Chaoen Xiao, and Xin Jin. 2021. Aesthetic Evaluation and Guidance for Mobile Photography. In Proceedings of the 29th ACM International Conference on Multimedia. 2780--2782.Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Xin Lu, Zhe Lin, Hailin Jin, Jianchao Yang, and James Z Wang. 2014. Rapid: Rating pictorial aesthetics using deep learning. In Proceedings of the 22nd ACM international conference on Multimedia. 457--466.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Yiwen Luo and Xiaoou Tang. 2008. Photo and video quality evaluation: Focusing on the subject. In European Conference on Computer Vision. 386--399.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Luca Marchesotti, Florent Perronnin, Diane Larlus, and Gabriela Csurka. 2011. Assessing the aesthetic quality of photographs using generic image descriptors. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 1784--1791.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Ricardo Martin-Brualla, David Gallup, and Steven M Seitz. 2015a. 3d time-lapse reconstruction from internet photos. In Proceedings of the IEEE International Conference on Computer Vision. 1332--1340.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Ricardo Martin-Brualla, David Gallup, and Steven M Seitz. 2015b. Time-lapse mining from internet photos. ACM Transactions on Graphics, Vol. 34, 4 (2015), 1--8.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Seonghyeon Nam, Chongyang Ma, Menglei Chai, William Brendel, Ning Xu, and Seon Joo Kim. 2019. End-to-end time-lapse video synthesis from a single outdoor image. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1409--1418.Google ScholarGoogle ScholarCross RefCross Ref
  23. L Neumann, M Sbert, B Gooch, W Purgathofer, et al. 2005. Defining computational aesthetics. Computational aesthetics in graphics, visualization and imaging (2005), 13--18.Google ScholarGoogle Scholar
  24. Masashi Nishiyama, Takahiro Okabe, Imari Sato, and Yoichi Sato. 2011. Aesthetic quality classification of photographs based on color harmony. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 33--40.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Yuzhen Niu and Feng Liu. 2012. What makes a professional video? A computational aesthetics approach. IEEE Transactions on Circuits and Systems for Video Technology, Vol. 22, 7 (2012), 1037--1049.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Stephen M Pizer, E Philip Amburn, John D Austin, Robert Cromartie, Ari Geselowitz, Trey Greer, Bart ter Haar Romeny, John B Zimmerman, and Karel Zuiderveld. 1987. Adaptive histogram equalization and its variations. Computer vision, graphics, and image processing, Vol. 39, 3 (1987), 355--368.Google ScholarGoogle Scholar
  27. Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. 2017. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition. 652--660.Google ScholarGoogle Scholar
  28. Esa Rahtu, Juho Kannala, Mikko Salo, and Janne Heikkilä. 2010. Segmenting salient objects from images and videos. In European conference on computer vision. Springer, 366--379.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Yogesh Singh Rawat and Mohan S Kankanhalli. 2016. ClickSmart: A context-aware viewpoint recommendation system for mobile photography. IEEE Transactions on Circuits and Systems for Video Technology, Vol. 27, 1 (2016), 149--158.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Yogesh Singh Rawat, Mubarak Shah, and Mohan S Kankanhalli. 2019. Photography and Exploration of Tourist Locations Based on Optimal Foraging Theory. IEEE Transactions on Circuits and Systems for Video Technology, Vol. 30, 7 (2019), 2276--2287.Google ScholarGoogle Scholar
  31. Michael Rubinstein, Ce Liu, Peter Sand, Fredo Durand, and William T Freeman. 2011. Motion denoising with application to time-lapse photography. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 313--320.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Hanghang Tong, Mingjing Li, Hong-Jiang Zhang, Jingrui He, and Changshui Zhang. 2004. Classification of digital photos taken by photographers or home users. In Pacific-Rim Conference on Multimedia. 198--205.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Christos Tzelepis, Eftichia Mavridaki, Vasileios Mezaris, and Ioannis Patras. 2016. Video aesthetic quality assessment using kernel Support Vector Machine with isotropic Gaussian sample uncertainty (KSVM-IGSU). In IEEE International Conference on Image Processing. IEEE, 2410--2414.Google ScholarGoogle ScholarCross RefCross Ref
  34. Miao Wang, Jun-Bang Liang, Song-Hai Zhang, Shao-Ping Lu, Ariel Shamir, and Shi-Min Hu. 2017. Hyper-lapse from multiple spatially-overlapping videos. IEEE Transactions on Image Processing, Vol. 27, 4 (2017), 1735--1747.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. Scott Wehrwein, Kavita Bala, and Noah Snavely. 2020. Scene summarization via motion normalization. IEEE Transactions on Visualization and Computer Graphics, Vol. 27, 4 (2020), 2495--2501.Google ScholarGoogle ScholarCross RefCross Ref
  36. Wei Xiong, Wenhan Luo, Lin Ma, Wei Liu, and Jiebo Luo. 2018. Learning to generate time-lapse videos using multi-stage dynamic generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2364--2373.Google ScholarGoogle ScholarCross RefCross Ref
  37. Hsin-Ho Yeh, Chun-Yu Yang, Ming-Sui Lee, and Chu-Song Chen. 2013. Video aesthetic quality assessment by temporal integration of photo-and motion-based features. IEEE transactions on multimedia, Vol. 15, 8 (2013), 1944--1957.Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Hui Zeng, Zisheng Cao, Lei Zhang, and Alan C Bovik. 2019. A unified probabilistic formulation of image aesthetic assessment. IEEE Transactions on Image Processing, Vol. 29 (2019), 1548--1561.Google ScholarGoogle ScholarCross RefCross Ref
  39. Jiangning Zhang, Chao Xu, Liang Liu, Mengmeng Wang, Xia Wu, Yong Liu, and Yunliang Jiang. 2020. Dtvnet: Dynamic time-lapse video generation via single still image. In European Conference on Computer Vision. Springer, 300--315.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Luming Zhang, Yue Gao, Roger Zimmermann, Qi Tian, and Xuelong Li. 2014. Fusion of multichannel local and global structural cues for photo aesthetics evaluation. IEEE Transactions on Image Processing, Vol. 23, 3 (2014), 1419--1429.Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. Feng Zhou, Sing Bing Kang, and Michael F Cohen. 2014. Time-mapping using space-time saliency. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3358--3365.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Aesthetics-Driven Virtual Time-Lapse Photography Generation

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        MM '23: Proceedings of the 31st ACM International Conference on Multimedia
        October 2023
        9913 pages
        ISBN:9798400701085
        DOI:10.1145/3581783

        Copyright © 2023 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 27 October 2023

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate995of4,171submissions,24%

        Upcoming Conference

        MM '24
        MM '24: The 32nd ACM International Conference on Multimedia
        October 28 - November 1, 2024
        Melbourne , VIC , Australia
      • Article Metrics

        • Downloads (Last 12 months)98
        • Downloads (Last 6 weeks)10

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader