Abstract
Unmanned Aerial Vehicle (UAV) visual geo-localization aims to match images of the same geographic target captured from different views, i.e., the UAV view and the satellite view. It is very challenging due to the large appearance differences in UAV-satellite image pairs. Previous works map images captured by UAVs and satellites to a shared feature space and employ a classification framework to learn location-dependent features while neglecting the overall distribution shift between the UAV view and the satellite view. In this paper, we address these limitations by introducing distribution alignment of the two views to shorten their distance in a common space. Specifically, we propose an end-to-end network, called PVDA (Progressive View Distribution Alignment). During training, feature encoder, location classifier, and view discriminator are jointly optimized by a novel progressive adversarial learning strategy. Competition between feature encoder and view discriminator prompts both of them to be stronger. It turns out that the adversarial learning is progressively emphasized until UAV-view images are indistinguishable from satellite-view images. As a result, the proposed PVDA becomes powerful in learning location-dependent yet view-invariant features with good scalability towards unseen images of new locations. Compared to the state-of-the-art methods, the proposed PVDA requires less inference time but has achieved superior performance on the University-1652 dataset.
This work was supported in part by the National Natural Science Foundation of China under Grant No. 62171295, and in part by the Liaoning Provincial Natural Science Foundation under Grant No. 2021-MS-266, and in part by the Applied Basic Research Project of Liaoning Province under Grant 2023JH2/101300204, and in part by the Shenyang Science and Technology Innovation Program for Young and Middle-aged Scientists under Grant No. RC210427.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Wilson, D., Alshaabi, T., Van Oort, C., Zhang, X., Nelson, J., Wshah, S.: Object tracking and geo-localization from street images. Remote Sens. 14(11), 1–26 (2022)
Erra, U., Capece, N.: Engineering an advanced geo-location augmented reality framework for smart mobile devices. J. Ambient. Intell. Humaniz. Comput. 10(1), 255–265 (2019)
Senlet, T., Elgammal, A.: Satellite image based precise robot localization on sidewalks. In: IEEE International Conference on Robotics and Automation, pp. 2647–2653 (2012)
Lin, T.Y., Cui, Y., Belongie, S., Hays, J.: Learning deep representations for ground-to-aerial geolocalization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5007–5015 (2015)
Zhai, M., Bessinger, Z., Workman, S., Jacobs, N.: Predicting ground-level scene layout from aerial imagery. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 867–875 (2017)
Tian, Y., Chen, C., Shah, M.: Cross-view image matching for geo-localization in urban environments. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3608–3616 (2017)
Liu, L., Li, H.: Lending orientation to neural networks for cross-view geo-localization. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5624–5633 (2019)
Zheng, Z., Wei, Y., Yang, Y.: University-1652: a multi-view multi-source benchmark for drone-based geo-localization. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1395–1403 (2020)
Ding, L., Zhou, J., Meng, L., Long, Z.: A practical cross-view image matching method between UAV and satellite for UAV-based geo-localization. Remote Sens. 13(47), 1–20 (2021)
Wang, T., et al.: Each part matters: local patterns facilitate cross-view geo-localization. IEEE Trans. Circ. Syst. Video Technol. 32(2), 867–879 (2021)
Lin, J., et al.: Joint representation learning and keypoint detection for cross-view geo-localization. IEEE Trans. Image Process. 31, 3780–3792 (2022)
Zhuang, J., Dai, M., Chen, X., Zheng, E.: A faster and more effective cross-view matching method of UAV and satellite images for UAV geolocalization. Remote Sens. 13(3979), 1–16 (2021)
Dai, M., Hu, J., Zhuang, J., Zheng, E.: A transformer-based feature segmentation and region alignment method for UAV-view geo-localization. IEEE Trans. Circ. Syst. Video Technol. 32(7), 4376–4389 (2022)
Regmi, K., Shah, M.: Bridging the domain gap for ground-to-aerial image matching. In: Proceedings of the IEEE/CVF Conference on Computer Vision (2019)
Shi, Y., Liu, L., Yu, X., Li, H.: Spatial-aware feature aggregation for cross-view image based geo-localization. In: 33rd Conference on Neural Information Processing Systems (2019)
Guo, Y., Choi, M., Li, K., Boussaid, F.: Soft exemplar highlighting for cross-view image-based geo-localization. IEEE Trans. Image Process. 31, 2094–2105 (2022)
Zhang, X., et al.: SSA-Net: spatial scale attention network for image-based geo-localization. IEEE Geosci. Remote Sens. Lett. 19, 1–5 (2022)
Tian, X., Shao, J., Ouyang, D., Shen, H.T.: UAV-satellite view synthesis for cross-view geo-localization. IEEE Trans. Circ. Syst. Video Technol. 32(7), 4804–4815 (2022)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv e-prints (2016)
Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 248–255. IEEE (2009)
Dosovitskiy, A., et al.: An image is worth 16 \(\times \) 16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, C., Liu, J., Qiu, H., Li, Z., Shi, X. (2023). View Distribution Alignment with Progressive Adversarial Learning for UAV Visual Geo-Localization. In: Jin, Z., Jiang, Y., Buchmann, R.A., Bi, Y., Ghiran, AM., Ma, W. (eds) Knowledge Science, Engineering and Management. KSEM 2023. Lecture Notes in Computer Science(), vol 14118. Springer, Cham. https://doi.org/10.1007/978-3-031-40286-9_20
Download citation
DOI: https://doi.org/10.1007/978-3-031-40286-9_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-40285-2
Online ISBN: 978-3-031-40286-9
eBook Packages: Computer ScienceComputer Science (R0)