Abstract
In many poster designs, an image usually will be used as a back-ground image, and text and picture will be carried out on the background image later. For intelligent layout design, cropping a suitable background image should be the first problem to be solved. In this paper, through eye movement experiments, ground truth saliency maps of the posters are obtained. Then, the characteristics of the saliency maps of background images are summarized. The characteristics are mainly the rules of the location and size of the salient areas in the background image. The research found that the salient areas of the poster background images are more concentrated in the upper and middle of the poster image, and they are distributed in an inverted triangle. These rules can cut a more suitable background image for typesetting.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Borji, A., Cheng, M.-M., Jiang, H., et al.: Salient object detection: a benchmark. IEEE Trans. Image Process. 24, 5706–5722 (2015). https://doi.org/10.1109/TIP.2015.2487833
Zhao, R., Ouyang, W., Li, H., et al.: Saliency detection by multi-context deep learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1265–1274 (2015). https://doi.org/10.1109/cvpr.2015.7298731
Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 20, 1254–1259 (1998). https://doi.org/10.1109/34.730558
Liu, T., Yuan, Z., Sun, J., et al.: Learning to detect a salient object. IEEE Trans. Pattern Anal. Mach. Intell. 33, 353–367 (2010). https://doi.org/10.1109/CVPR.2007.383047
Achanta, R., Hemami, S., Estrada, F., et al.: Frequency-tuned salient region detection. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1597–1604. IEEE (2009). https://doi.org/10.1109/cvpr.2009.5206596
Borji, A.: What is a salient object? A dataset and a baseline model for salient object detection. IEEE Trans. Image Process. 24, 742–756 (2014). https://doi.org/10.1109/TIP.2014.2383320
Bylinskii, Z., Kim, N.W., O’Donovan, P., et al.: Learning visual importance for graphic designs and data visualizations. In: Proceedings of the 30th Annual ACM Symposium on User Interface Software and Technology, pp. 57–69 (2017). https://doi.org/10.1145/3126594.3126653
Jahanian, A., Liu, J., Tretter, D.R., et al.: Automatic design of magazine covers (2012). https://doi.org/10.1117/12.914596
Jahanian, A., Vishwanathan, S.V.N., Allebach, J.P.: Learning visual balance from large-scale datasets of aesthetically highly rated images. In: Proceedings of SPIE - The International Society for Optical Engineering, vol. 9394 (2015). https://doi.org/10.1117/12.2084548
Chen, L.-Q., Xie, X., Fan, X., et al.: A visual attention model for adapting images on small displays. Multimed. Syst. 9, 353–364 (2003). https://doi.org/10.1007/s00530-003-0105-4
Suh, B., Ling, H., Bederson, B.B., et al.: Automatic thumbnail cropping and its effectiveness. In: Proceedings of the 16th Annual ACM Symposium on User Interface Software and Technology, pp. 95–104 (2003). https://doi.org/10.1145/964696.964707
Santella, A., Agrawala, M., DeCarlo, D., et al.: Gaze-based interaction for semi-automatic photo cropping. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 771–780 (2006). https://doi.org/10.1145/1124772.1124886
Marchesotti, L., Cifarelli, C., Csurka, G.: A framework for visual saliency detection with applications to image thumbnailing. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 2232–2239. IEEE (2009). https://doi.org/10.1109/iccv.2009.5459467
Chen, J., Bai, G., Liang, S., et al.: Automatic image cropping: a computational complexity study. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 507–515 (2016). https://doi.org/10.1109/cvpr.2016.61
Borji, A., Sihite, D.N., Itti, L.: What stands out in a scene? A study of human explicit saliency judgment. Vis. Res. 91, 62–77 (2013). https://doi.org/10.1016/j.visres.2013.07.016
Xu, J., Jiang, M., Wang, S., et al.: Predicting human gaze beyond pixels. J. Vis. 14, 28 (2014). https://doi.org/10.1167/14.1.28
Koehler, K., Guo, F., Zhang, S., et al.: What do saliency models predict? J. Vis. 14, 14 (2014). https://doi.org/10.1167/14.3.14
Cornia, M., Baraldi, L., Serra, G., et al.: Predicting human eye fixations via an LSTM-based saliency attentive model. IEEE Trans. Image Process. 27, 5142–5154 (2018). https://doi.org/10.1109/TIP.2018.2851672
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Zhu, L., Cao, X., Fang, Y., Zhang, L., Li, X. (2020). Application of Visual Saliency in the Background Image Cutting for Layout Design. In: Meiselwitz, G. (eds) Social Computing and Social Media. Design, Ethics, User Behavior, and Social Network Analysis. HCII 2020. Lecture Notes in Computer Science(), vol 12194. Springer, Cham. https://doi.org/10.1007/978-3-030-49570-1_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-49570-1_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-49569-5
Online ISBN: 978-3-030-49570-1
eBook Packages: Computer ScienceComputer Science (R0)