skip to main content
10.1145/3645259.3645277acmotherconferencesArticle/Chapter ViewAbstractPublication PagesipmvConference Proceedingsconference-collections
research-article

Generative Adversarial Model Equipped with Contrastive Learning in Map Synthesis

Published:03 May 2024Publication History

ABSTRACT

In the dynamic field of urban planning and the context of unprecedented natural events, such as hurricanes, the fast generation of accurate maps from satellite imagery is paramount. While several studies have utilized Generative Adversarial Networks (GANs) for map generation from satellite images, the present work introduces a new approach by integrating contrastive learning into the GAN framework for enhanced map synthesis. Our methodology distinctively employs positive sampling by aligning similar features (e.g., roads) in both satellite images and their corresponding map outputs, and contrasts this with negative samples for disparate elements. This approach effectively replaces the conventional cyclic process in GANs with a more streamlined, unidirectional procedure, leading to improvements in both the quality of the synthesized maps and computational efficiency. We show the effectiveness of our proposed model, offering an advancement in map generation for remote sensing applications.

References

  1. Iman Abbasnejad, Fabio Zambetta, Flora Salim, Timothy Wiley, Jeffrey Chan, Russell Gallagher, and Ehsan Abbasnejad. 2023. SCONE-GAN: Semantic Contrastive Learning-Based Generative Adversarial Network for an End-to-End Image Translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1111–1120.Google ScholarGoogle ScholarCross RefCross Ref
  2. Hamed Alqahtani, Manolya Kavakli-Thorne, and Gulshan Kumar. 2021. Applications of generative adversarial networks (gans): An updated review. Archives of Computational Methods in Engineering 28 (2021), 525–552.Google ScholarGoogle ScholarCross RefCross Ref
  3. Asha Anoosheh, Eirikur Agustsson, Radu Timofte, and Luc Van Gool. 2018. Combogan: Unrestrained scalability for image domain translation. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops. 783–790.Google ScholarGoogle ScholarCross RefCross Ref
  4. Lubin Bai, Shihong Du, Xiuyuan Zhang, Haoyu Wang, Bo Liu, and Song Ouyang. 2022. Domain Adaptation for Remote Sensing Image Semantic Segmentation: An Integrated Approach of Contrastive Learning and Adversarial Learning. IEEE Transactions on Geoscience and Remote Sensing 60 (2022), 1–13. https://doi.org/10.1109/TGRS.2022.3198972Google ScholarGoogle ScholarCross RefCross Ref
  5. Tianfeng Chai and Roland R Draxler. 2014. Root mean square error (RMSE) or mean absolute error (MAE). Geoscientific model development discussions 7, 1 (2014), 1525–1534.Google ScholarGoogle Scholar
  6. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. 2020. A simple framework for contrastive learning of visual representations. In International conference on machine learning. PMLR, 1597–1607.Google ScholarGoogle Scholar
  7. Xu Chen, Songqiang Chen, Tian Xu, Bangguo Yin, Jian Peng, Xiaoming Mei, and Haifeng Li. 2020. SMAPGAN: Generative adversarial network-based semisupervised styled map tile generation method. IEEE Transactions on Geoscience and Remote Sensing 59, 5 (2020), 4388–4406.Google ScholarGoogle ScholarCross RefCross Ref
  8. Yuan Chen, Yang Zhao, Wei Jia, Li Cao, and Xiaoping Liu. 2020. Adversarial-learning-based image-to-image transformation: A survey. Neurocomputing 411 (2020), 468–486.Google ScholarGoogle ScholarCross RefCross Ref
  9. Ugur Demir and Gozde Unal. 2018. Patch-based image inpainting with generative adversarial networks. arXiv preprint arXiv:1803.07422 (2018).Google ScholarGoogle Scholar
  10. Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, Kun Zhang, and Dacheng Tao. 2019. Geometry-consistent generative adversarial networks for one-sided unsupervised domain mapping. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2427–2436.Google ScholarGoogle ScholarCross RefCross Ref
  11. Swetava Ganguli, Pedro Garzon, and Noa Glaser. 2019. GeoGAN: A conditional GAN with reconstruction and style loss to generate standard layer of maps from satellite images. arXiv preprint arXiv:1902.05611 (2019).Google ScholarGoogle Scholar
  12. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. Advances in neural information processing systems 27 (2014).Google ScholarGoogle Scholar
  13. Yao Gou, Min Li, Yu Song, Yujie He, and Litao Wang. 2023. Multi-feature contrastive learning for unpaired image-to-image translation. Complex & Intelligent Systems 9, 4 (2023), 4111–4122.Google ScholarGoogle ScholarCross RefCross Ref
  14. Michael Gutmann and Aapo Hyvärinen. 2010. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics(Proceedings of Machine Learning Research, Vol. 9), Yee Whye Teh and Mike Titterington (Eds.). PMLR, Chia Laguna Resort, Sardinia, Italy, 297–304. https://proceedings.mlr.press/v9/gutmann10a.htmlGoogle ScholarGoogle Scholar
  15. Junlin Han, Mehrdad Shoeiby, Lars Petersson, and Mohammad Ali Armin. 2021. Dual contrastive learning for unsupervised image-to-image translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 746–755.Google ScholarGoogle ScholarCross RefCross Ref
  16. Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. 2020. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 9729–9738.Google ScholarGoogle ScholarCross RefCross Ref
  17. Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. 2017. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems 30 (2017).Google ScholarGoogle Scholar
  18. Yi-Yen Hsieh, Yu-Chi Lee, and Chia-Hsiang Yang. 2020. A CycleGAN Accelerator for Unsupervised Learning on Mobile Devices. In 2020 IEEE International Symposium on Circuits and Systems (ISCAS). 1–5. https://doi.org/10.1109/ISCAS45731.2020.9180845Google ScholarGoogle ScholarCross RefCross Ref
  19. Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. 2017. Image-To-Image Translation With Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).Google ScholarGoogle ScholarCross RefCross Ref
  20. Hadi Kazemi, Sobhan Soleymani, Fariborz Taherkhani, Seyed Iranmanesh, and Nasser Nasrabadi. 2018. Unsupervised image-to-image translation using domain-specific variational information bound. Advances in neural information processing systems 31 (2018).Google ScholarGoogle Scholar
  21. Ashish Kumar, Abeer Alsadoon, PWC Prasad, Salma Abdullah, Tarik A Rashid, Duong Thu Hang Pham, and Tran Quoc Vinh Nguyen. 2022. Generative adversarial network (GAN) and enhanced root mean square error (ERMSE): deep learning for stock price movement prediction. Multimedia Tools and Applications (2022), 1–19.Google ScholarGoogle Scholar
  22. Andrew L Maas, Awni Y Hannun, Andrew Y Ng, 2013. Rectifier nonlinearities improve neural network acoustic models. In Proc. icml, Vol. 30. Atlanta, GA, 3.Google ScholarGoogle Scholar
  23. Arpan Mahara and Naphtali Rishe. [n. d.]. Integrating Location Information as Geohash Codes in Convolutional Neural Network-Based Satellite Image Classification. ([n. d.]).Google ScholarGoogle Scholar
  24. Aaron van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018).Google ScholarGoogle Scholar
  25. Taesung Park, Alexei A Efros, Richard Zhang, and Jun-Yan Zhu. 2020. Contrastive learning for unpaired image-to-image translation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX 16. Springer, 319–345.Google ScholarGoogle Scholar
  26. Junghoon Seo, Taewon Yoon, Jinwoo Kim, and Kin Choong Yow. 2017. One-to-one example-based automatic image coloring using deep convolutional generative adversarial network. Journal of Advances in Information Technology Vol 8, 2 (2017).Google ScholarGoogle Scholar
  27. Jieqiong Song, Hao Chen, Chun Du, and Jun Li. 2023. Semi-MapGen: Translation of Remote Sensing Image Into Map via Semisupervised Adversarial Learning. IEEE Transactions on Geoscience and Remote Sensing 61 (2023), 1–19. https://doi.org/10.1109/TGRS.2023.3263897Google ScholarGoogle ScholarCross RefCross Ref
  28. Jieqiong Song, Jun Li, Hao Chen, and Jiangjiang Wu. 2021. MapGen-GAN: A fast translator for remote sensing image to map via unsupervised adversarial learning. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 14 (2021), 2341–2357.Google ScholarGoogle ScholarCross RefCross Ref
  29. Zhou Wang, Alan C Bovik, Hamid R Sheikh, and Eero P Simoncelli. 2004. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing 13, 4 (2004), 600–612.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Wang Yuanji, Li Jianhua, Lu Yi, Fu Yao, and Jiang Qinzhong. 2003. Image quality evaluation based on image weighted separating block peak signal to noise ratio. In International Conference on Neural Networks and Signal Processing, 2003. Proceedings of the 2003, Vol. 2. IEEE, 994–997.Google ScholarGoogle ScholarCross RefCross Ref
  31. Ying Zhang, Yifang Yin, Roger Zimmermann, Guanfeng Wang, Jagannadan Varadarajan, and See-Kiong Ng. 2020. An Enhanced GAN Model for Automatic Satellite-to-Map Image Conversion. IEEE Access 8 (2020), 176704–176716. https://doi.org/10.1109/ACCESS.2020.3025008Google ScholarGoogle ScholarCross RefCross Ref
  32. Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. 2017. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision. 2223–2232.Google ScholarGoogle ScholarCross RefCross Ref

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    IPMV '24: Proceedings of the 2024 6th International Conference on Image Processing and Machine Vision
    January 2024
    129 pages
    ISBN:9798400708473
    DOI:10.1145/3645259

    Copyright © 2024 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 3 May 2024

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited
  • Article Metrics

    • Downloads (Last 12 months)27
    • Downloads (Last 6 weeks)27

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format