Skip to main content

Hiding Video in Images: Harnessing Adversarial Learning on Deep 3D-Spatio-Temporal Convolutional Neural Networks

  • Conference paper
  • First Online:
Book cover Computer Vision and Image Processing (CVIP 2022)

Abstract

This work proposes end-to-end trainable models of Generative Adversarial Networks (GAN) for hiding video data inside images. Hiding video inside images is a relatively new topic and has never been attempted earlier to our best knowledge. We propose two adversarial models that hide video data inside images: a base model with Recurrent Neural Networks and a novel model with 3D-spatiotemporal Convolutional Neural Networks. Both the models have two distinct networks: (1) An embedder to extract features from the time variate video data and inject them into the deep latent representations of the image. (2) An extractor that reverse-engineers the embedder function to extract the hidden data inside the encoded image. A multi-discriminator GAN framework with multi-objective training for multimedia hiding is one of the novel contributions of this work.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Agustsson, E., Timofte, R.: NTIRE 2017 challenge on single image super-resolution: dataset and study. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1122–1131 (2017)

    Google Scholar 

  2. Chen B.J., et al.: FULL 4-D quaternion discrete Fourier transform based watermarking for color images. Digit. Signal Process. 28(1), 106–119 (2014), https://doi.org/10.1016/j.dsp.2014.02.010

  3. Das C, Panigrahi S, Sharma,V.K., Mahapatra, K.K.: A novel blind robust image watermarking in dct domain using inter-block coefficient correlation. Int. J. Electron. Commun. 68(3), 244–253 (2014)

    Google Scholar 

  4. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: Conference on Computer Vision and Pattern Recognition (2009)

    Google Scholar 

  5. Fan, L., bing Huang, W., Gan, C., Huang, J., Gong, B.: Controllable image-to-video translation: A case study on facial expression generation. CoRR abs/1808.02992 (2019)

    Google Scholar 

  6. Feng, L.P., Zheng, L.B., C.P.: A DWT-DCT based blind watermarking algorithm for copyright protection. Proc. IEEE ICCIST 7, 455–458 (2010). https://doi.org/10.1109/ICCSIT.2010.5565101

  7. Gandikota, R., Mishra, D.: Hiding audio in images: a deep learning approach. In: Deka, B., Maji, P., Mitra, S., Bhattacharyya, D.K., Bora, P.K., Pal, S.K. (eds.) PReMI 2019. LNCS, vol. 11942, pp. 389–399. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-34872-4_43

    Chapter  Google Scholar 

  8. GitRepo: Video-hiding-GAN (2019). https://github.com/RohitGandikota/Hiding-Video-in-Images-using-Deep-Generative-Adversarial-Networks

  9. Hu, Q., Waelchli, A., Portenier, T., Zwicker, M., Favaro, P.: Video synthesis from a single image and motion stroke. CoRR abs/1812.01874 (2018)

    Google Scholar 

  10. Ian, J., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27, 2672–2680 (2013)

    Google Scholar 

  11. J. Ouyang, G. Coatrieux, B.C.H.S.: Color image watermarking based on quaternion fourier transform and improved uniform log-polar mapping. Comput. Electr. Eng. 46, 419–432 (2015)

    Google Scholar 

  12. Zhu, J., Kaplan, R., Johnson, J., Fei-Fei, L.: HiDDeN: Hiding Data With Deep Networks. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11219, pp. 682–697. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01267-0_40

    Chapter  Google Scholar 

  13. Kandi, H., Mishra, D., Sai Gorthi, S.R.K.: Exploring the learning capabilities of convolutional neural networks for robust image watermarking. Comput. Secur. 65, 2506–2510 (2017). https://doi.org/10.1016/j.cose.2016.11.016

  14. Zhang, K.A., Cuesta-Infante, A., Xu, L., Veeramachaneni, K.: SteganoGAN: high capacity image steganography with GANs. arXiv preprint arXiv:1901.03892 (January 2019)

  15. Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.-H.: Flow-Grounded spatial-temporal video prediction from still images. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11213, pp. 609–625. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01240-3_37

    Chapter  Google Scholar 

  16. Mun, S.M., et al.: A robust blind watermarking using convolutional neural network. arXiv preprint arXiv:1704.03248 (2017)

  17. Pan, Y., Qiu, Z., Yao, T., Li, H., Mei, T.: To create what you tell: generating videos from captions. In: Proceedings of the 25th ACM International Conference on Multimedia, pp. 1789–1798. MM 2017 (2017), http://doi.acm.org/10.1145/3123266.3127905

  18. Sun, Q.T., Niu, Y.G., Wang, Q., Sheng, G.: A blind color image watermarking based on dc component in the spatial domain. Optik 124(23), 6255–6260 (2013). https://doi.org/10.1016/j.ijleo.2013.05.013,https://doi.org/10.1016/j.ijleo.2013.05.013

  19. Vondrick, C., Pirsiavash, H., Torralba, A.: Generating videos with scene dynamics. In: Proceedings of the 30th International Conference on Neural Information Processing Systems, pp. 613–621. NIPS2016, Curran Associates Inc., USA (2016), http://dl.acm.org/citation.cfm?id=3157096.3157165

  20. Wu, Y., Lim, J., Yang, M.: object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 37(9), 1834–1848 (2015). https://doi.org/10.1109/TPAMI.2014.2388226

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rohit Gandikota .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Gandikota, R., Mishra, D., Brown, N.B. (2023). Hiding Video in Images: Harnessing Adversarial Learning on Deep 3D-Spatio-Temporal Convolutional Neural Networks. In: Gupta, D., Bhurchandi, K., Murala, S., Raman, B., Kumar, S. (eds) Computer Vision and Image Processing. CVIP 2022. Communications in Computer and Information Science, vol 1776. Springer, Cham. https://doi.org/10.1007/978-3-031-31407-0_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-31407-0_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-31406-3

  • Online ISBN: 978-3-031-31407-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics