ABSTRACT
The topic of visual saliency is well-known and spreads across numerous disciplines. In this paper, we examine how saliency models perform on images with specific characteristics. We explore a saliency of 14 different digitized paintings, all representing a biblical scene of The Last Supper. We evaluate the performance of different saliency models. The models are using the traditional approach as well as the deep learning approach. For the evaluation, we use three different metrics, AUC, NSS and CC. As ground truth, we use eye-tracking data from 35 participants. Our analysis shows that deep-learning methods predict the most salient parts of the paintings very similar to real eye fixations. We also checked the consistency of gaze patterns among the participants.
- 2018. ARTstor. http://www.artstor.orgGoogle Scholar
- 2018. prometheus. https://prometheus-bildarchiv.deGoogle Scholar
- A. Borji and L. Itti. 2012. State-of-the-art in visual attention modeling. IEEE transactions on pattern analysis and machine intelligence 35, 1(2012), 185–207.Google Scholar
- A. Borji, H. R Tavakoli, D. N. Sihite, and L. Itti. 2013. Analysis of scores, datasets, and models in visual saliency prediction. In Proceedings of the IEEE international conference on computer vision. 921–928.Google Scholar
- G. T. Buswell. 1935. How people look at pictures: a study of the psychology and perception in art. Univ. Chicago Press.Google Scholar
- Z. Bylinskii, T. Judd, A. Oliva, A. Torralba, and F. Durand. 2018. What do different evaluation metrics tell us about saliency models?IEEE transactions on pattern analysis and machine intelligence 41, 3(2018), 740–757.Google Scholar
- M. Cacciola, G. Occhiuto, and F. C. Morabito. 2013. Artistic complexity and saliency: Two faces of the same coin?International Journal of Information Acquisition 9, 02(2013), 1350010.Google Scholar
- M. Cornia, L. Baraldi, G. Serra, and R. Cucchiara. 2018. Predicting human eye fixations via an lstm-based saliency attentive model. IEEE Transactions on Image Processing 27, 10 (2018), 5142–5154.Google ScholarCross Ref
- I. Fuchs, U. Ansorge, Ch. Redies, and H. Leder. 2011. Salience in paintings: bottom-up influences on eye fixations. Cognitive Computation 3, 1 (2011), 25–36.Google ScholarCross Ref
- L. Itti, C. Koch, and E. Niebur. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence 20, 11 (Nov 1998), 1254–1259.Google ScholarDigital Library
- T. N. Vikram, M. Tscherepanow, and B. Wrede. 2011. A random center surround bottom up visual attention model useful for salient region detection. In 2011 IEEE Workshop on Applications of Computer Vision (WACV). IEEE, 166–173.Google Scholar
Recommendations
Mesh saliency
Research over the last decade has built a solid mathematical foundation for representation and analysis of 3D meshes in graphics and geometric modeling. Much of this work however does not explicitly incorporate models of low-level human visual ...
Mesh saliency
SIGGRAPH '05: ACM SIGGRAPH 2005 PapersResearch over the last decade has built a solid mathematical foundation for representation and analysis of 3D meshes in graphics and geometric modeling. Much of this work however does not explicitly incorporate models of low-level human visual ...
Human Visual Scanpath Prediction Based on RGB-D Saliency
ICIGP '18: Proceedings of the 2018 International Conference on Image and Graphics ProcessingHuman visual perception is considered as a dynamic process of information acquisition, while the visual scanpath can clearly reflect the shift of our eye fixations. In the previous study of visual attention, researchers generally do the saliency ...
Comments