skip to main content
10.1145/3372278.3390678acmconferencesArticle/Chapter ViewAbstractPublication PagesicmrConference Proceedingsconference-collections
research-article

Learning to Select Elements for Graphic Design

Published:08 June 2020Publication History

ABSTRACT

Selecting elements for graphic design is essential for ensuring a correct understanding of clients' requirements as well as improving the efficiency of designers before a fine-designed process. Some semi-automatic design tools proposed layout templates where designers always select elements according to the rectangular boxes that specify how elements are placed. In practice, layout and element selection are complementary. Compared to the layout which can be readily obtained from pre-designed templates, it is generally time-consuming to mindfully pick out suitable elements, which calls for an automation of elements selection. To address this, we formulate element selection as a sequential decision-making process and develop a deep element selection network (DESN). Given a layout file with annotated elements, new graphical elements are selected to form graphic designs based on aesthetics and consistency criteria. To train our DESN, we propose an end-to-end, reinforcement learning based framework, where we design a novel reward function that jointly accounts for visual aesthetics and consistency. Based on this, visually readable and aesthetic drafts can be efficiently generated. We further contribute a layout-poster dataset with exhaustively labeled attributes of poster key elements. Qualitative and quantitative results indicate the efficacy of our approach.

References

  1. Jaime Arguello and Bogeum Choi. 2019. The Effects of Working Memory, Perceptual Speed, and Inhibition in Aggregated Search. TOIS, Vol. 37, 3 (2019), 1--34.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Rudolf Arnheim. 1956. Art and visual perception: a psychology of the creative eye. Philosophy & Phenomenological Research, Vol. 16, 3 (1956), 425.Google ScholarGoogle ScholarCross RefCross Ref
  3. Tao Chen, Ming-Ming Cheng, Ping Tan, Ariel Shamir, and Shi-Min Hu. 2009. Sketch2photo: Internet image montage. TOG, Vol. 28, 5 (2009), 124.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Wengling Chen and James Hays. 2018. Sketchygan: Towards diverse and realistic sketch to image synthesis. In CVPR. 9416--9425.Google ScholarGoogle Scholar
  5. Niranjan Dameravenkata, Josep Bento, and Eamonn Obrienstrain. 2011. Probabilistic document model for automated document composition. (2011), 3--12.Google ScholarGoogle Scholar
  6. Jarrod Gingras. [n.d.]. https://www.realstorygroup.com/Blog/3220-Do-You-Need-Digital-Asset-Management? https://www.realstorygroup.com/Blog/3220-Do-You-Need-Digital-Asset-Management?.Google ScholarGoogle Scholar
  7. Shixiang Gu, Timothy P Lillicrap, Ilya Sutskever, and Sergey Levine. 2016. Continuous deep Q-learning with model-based acceleration. In ICML. 2829--2838.Google ScholarGoogle Scholar
  8. Robert M Haralick. 2005. Statistical and structural approaches to texture. Proc. IEEE, Vol. 67, 5 (2005), 786--804.Google ScholarGoogle ScholarCross RefCross Ref
  9. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In CVPR. 770--778.Google ScholarGoogle Scholar
  10. Elad Hoffer, Itay Hubara, and Daniel Soudry. 2017. Train longer, generalize better: closing the generalization gap in large batch training of neural networks. NeurIPS, 1731--1741.Google ScholarGoogle Scholar
  11. Xiansheng Hua. 2018. Challenges and Practices of Large Scale Visual Intelligence in the Real-World. In ACM MM. 364--364.Google ScholarGoogle Scholar
  12. Nathan Hurst, Wilmot Li, and Kim Marriott. 2009. Review of automatic document formatting. In Proceedings of the 9th ACM symposium on Document engineering. 99--108.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Phillip Isola, Jun Yan Zhu, Tinghui Zhou, and Alexei A. Efros. 2017. Image-to-Image Translation with Conditional Adversarial Networks. In CVPR. 5967--5976.Google ScholarGoogle Scholar
  14. Yu Liu, Tao Mei, and Chang Wen Chen. 2016. Automatic suggestion of presentation image for storytelling. In ICME. 1--6.Google ScholarGoogle Scholar
  15. Xin Lu, Zhe Lin, Hailin Jin, Jianchao Yang, and James Z Wang. 2014. Rapid: Rating pictorial aesthetics using deep learning. In ACMMM. 457--466.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Yiwen Luo and Xiaoou Tang. 2008. Photo and Video Quality Evaluation: Focusing on the Subject. In ECCV.Google ScholarGoogle Scholar
  17. Jana Machajdik and Allan Hanbury. 2010. Affective image classification using features inspired by psychology and art theory. In ACM MM. 83--92.Google ScholarGoogle Scholar
  18. Luana Micallef, Gregorio Palmas, Antti Oulasvirta, and Tino Weinkauf. 2017. Towards Perceptual Optimization of the Visual Design of Scatterplots. TVCG, Vol. 23, 6 (2017), 1588--1599.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Naila Murray, Luca Marchesotti, and Florent Perronnin. 2012. AVA: A large-scale database for aesthetic visual analysis. In CVPR. 2408--2415.Google ScholarGoogle Scholar
  20. Alejandro Newell, Kaiyu Yang, and Jia Deng. 2016. Stacked Hourglass Networks for Human Pose Estimation. In ECCV. 483--499.Google ScholarGoogle Scholar
  21. Peter Odonovan, Aseem Agarwala, and Aaron Hertzmann. 2014. Learning Layouts for Single-PageGraphic Designs. TVCG, Vol. 20, 8 (2014), 1200--1213.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Peter Odonovan, Aseem Agarwala, and Aaron Hertzmann. 2015. DesignScape: Design with Interactive Layout Suggestions. In ACM CHI. 1221--1224.Google ScholarGoogle Scholar
  23. Xufang Pang, Ying Cao, Rynson W H Lau, and Antoni B Chan. 2016. Directing user attention via visual flow on web designs. ACM Transactions on Graphics (TOG), Vol. 35, 6 (2016), 240.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Yuting Qiang, Yanwei Fu, Yanwen Guo, Zhihua Zhou, and Leonid Sigal. 2016. Learning to generate posters of scientific papers. In AAAI. 51--57.Google ScholarGoogle Scholar
  25. Filip Radenović, Giorgos Tolias, and Ondvr ej Chum. 2016. CNN image retrieval learns from BoW: Unsupervised fine-tuning with hard examples. In European conference on computer vision. Springer, 3--20.Google ScholarGoogle ScholarCross RefCross Ref
  26. Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael S Bernstein, et al. 2015. ImageNet Large Scale Visual Recognition Challenge. IJCV, Vol. 115, 3 (2015), 211--252.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In ICLR.Google ScholarGoogle Scholar
  28. Hsiao-Hang Su, Tse-Wei Chen, Chieh-Chi Kao, Winston H Hsu, and Shao-Yi Chien. 2011. Scenic photo quality assessment with bag of aesthetics-preserving features. In ACM MM. 1213--1216.Google ScholarGoogle Scholar
  29. Hossein Talebi and Peyman Milanfar. 2017. NIMA: Neural Image Assessment. TIP, Vol. PP, 99 (2017), 1--1.Google ScholarGoogle Scholar
  30. Xin-Yi Tong, Gui-Song Xia, Fan Hu, Yanfei Zhong, Mihai Datcu, and Liangpei Zhang. 2019. Exploiting deep features for remote sensing image retrieval: A systematic investigation. IEEE Transactions on Big Data (2019).Google ScholarGoogle Scholar
  31. George Tucker, Surya Bhupatiraju, Shixiang Gu, Richard E Turner, Zoubin Ghahramani, and Sergey Levine. 2018. The Mirage of Action-Dependent Baselines in Reinforcement Learning. In ICML. 5015--5024.Google ScholarGoogle Scholar
  32. Nam Vo, Lu Jiang, Chen Sun, Kevin Murphy, Li-Jia Li, Li Fei-Fei, and James Hays. 2019. Composing text and image for image retrieval-an empirical odyssey. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 6439--6448.Google ScholarGoogle ScholarCross RefCross Ref
  33. Liwei Wang, Yin Li, and Svetlana Lazebnik. 2016. Learning Deep Structure-Preserving Image-Text Embeddings. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5005--5013.Google ScholarGoogle ScholarCross RefCross Ref
  34. Yan Wang, Tao Mei, and Xian Sheng Hua. 2011. Community Discovery from Movie and Its Application to Poster Generation. Springer Berlin Heidelberg. 107--117 pages.Google ScholarGoogle Scholar
  35. wikihow. [n.d.]. Design-Posters. https://www.wikihow.com/Design-Posters.Google ScholarGoogle Scholar
  36. Shuai Yang, Jiaying Liu, Wenhan Yang, and Zongming Guo. 2018. Context-Aware Text-Based Binary Image Stylization and Synthesis. TIP, Vol. PP, 99 (2018), 1--1.Google ScholarGoogle Scholar
  37. Sasi Kiran Yelamarthi, M Shiva Krishna Reddy, Ashish Kumar Mishra, and Anurag Mittal. 2018. A Zero-Shot Framework for Sketch Based Image Retrieval. In ECCV. 316--333.Google ScholarGoogle Scholar
  38. Quanzeng You, Jiebo Luo, Hailin Jin, and Jianchao Yang. 2016. Building a large scale dataset for image emotion recognition: the fine print and the benchmark. In AAAI. 308--314.Google ScholarGoogle Scholar
  39. Nanxuan Zhao, Ying Cao, and Rynson WH Lau. 2018. What characterizes personalities of graphic designs? TOG, Vol. 37, 4 (2018), 116.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Xinru Zheng, Xiaotian Qiao, Ying Cao, and Rynson WH Lau. 2019. Content-aware generative modeling of graphic design layouts. ACM Transactions on Graphics (TOG), Vol. 38, 4 (2019), 1--15.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Learning to Select Elements for Graphic Design

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        ICMR '20: Proceedings of the 2020 International Conference on Multimedia Retrieval
        June 2020
        605 pages
        ISBN:9781450370875
        DOI:10.1145/3372278

        Copyright © 2020 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 8 June 2020

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate254of830submissions,31%

        Upcoming Conference

        ICMR '24
        International Conference on Multimedia Retrieval
        June 10 - 14, 2024
        Phuket , Thailand

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader