Abstract
Sophisticated image editing techniques like colour and tone adjustments are used to enhance the perceived visual quality of images and are used in a broad variety of applications from professional grade image post-processing to sharing in social media platforms. Given a visually appealing reference image that has some photographic filter or effects applied, it is often desired to apply the same effects on a different target image to provide it the same look and feel. Interpreting the effects applied on such images is not a trivial task and requires knowledge and expertise on advanced image editing techniques, which is not easy. Existing deep learning based techniques fail to directly address this problem and offer partial solutions in the form of Neural Style Transfer, which can be used for texture transfer between images. In this paper, a novel method using a convolutional neural network (CNN) is introduced that can transfer the photographic filter and effects from a given reference image to a desired target image via adaptively predicting the parameters of the transformations that were applied on the reference image. These predicted parameters are then applied to the target image to get the same transformations as that of the reference image. In contrast to the existing stylization methods, the predicted parameters are independent of the semantics of the reference image and is well generalized to transfer complex filters from the reference image to any target image.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Heeger, D.J., et al.: Pyramid-based texture analysis/synthesis. In: International Conference on Image Processing, Washington, DC, USA, vol. 3, pp. 648–651 (1995)
Reinhard, E., et al.: Color transfer between images. In: IEEE Computer Graphics and Applications, vol. 21, no. 5, pp. 34–41, July-August 2001
Lee, J.-Y., et al.: Automatic content-aware color and tone stylization. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)
Deng, Y., et al.: Aesthetic-driven image enhancement by adversarial learning. In: ACM Multimedia (2018)
Hu, Y., et al.: Exposure: a white-box photo post-processing framework. ACM Trans. Graph. 37, 1–17 (2018)
Yan, Z., et al.: Automatic photo adjustment using deep neural networks. ACM Trans. Graph. 35, 1–15 (2016)
Zhu, F., et al.: Exemplar-based image and video stylization using fully convolutional semantic features. IEEE Trans. Image Process. 26, 3542–3555 (2017)
Li, Y., et al.: Universal style transfer via feature transforms. In: International Conference on Neural Information Processing Systems (NIPS) (2017)
Li, Y., et al.: A closed-form solution to photorealistic image stylization. In: ECCV (2018)
Ilie, A., et al.: Ensuring color consistency across multiple cameras. In: Tenth IEEE International Conference on Computer Vision (2005)
Bianco, S., et al.: Artistic photo filter removal using convolutional neural networks. J. Electron. Imaging 27(1), 011004 (2017). https://doi.org/10.1117/1.jei.27.1.011004
He, K., et al.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Sen, M., Chakraborty, P. (2020). A Deep Convolutional Neural Network Based Approach to Extract and Apply Photographic Transformations. In: Nain, N., Vipparthi, S., Raman, B. (eds) Computer Vision and Image Processing. CVIP 2019. Communications in Computer and Information Science, vol 1148. Springer, Singapore. https://doi.org/10.1007/978-981-15-4018-9_14
Download citation
DOI: https://doi.org/10.1007/978-981-15-4018-9_14
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-4017-2
Online ISBN: 978-981-15-4018-9
eBook Packages: Computer ScienceComputer Science (R0)