Abstract
As an essential part of geocoding, address matching has gained increasing research attention. Due to the long-distance dependency and unstructured property, existing address-matching methods hardly capture the contextual and implicit semantic information of unstructured text addresses. This paper presents a Contrastive Learning model based on Global and Local representation Interaction for Address Matching (referred to as CLGLIAM), which introduces a novel global and local interaction network to enhance the discrimination ability of the model on the hard negative address by associating the relationship between the global and local address representation explicitly. Simultaneously, to improve the generalization and transferability of the model, we utilize contrastive learning to enrich the data sample and extricate the model from task-specific knowledge. Furthermore, extensive experiments are conducted on Shenzhen and national address datasets to verify the effectiveness of our approach. Our model achieves state-of-the-art F1 scores of 99.26 and 98.50 on the two datasets, respectively. And the extended hard negative experiments further demonstrate the better performance of CLGLIAM in terms of semantic discrimination.
Similar content being viewed by others
Availability of data and materials
The 498,294 records of the corpus derived from the Shenzhen Address Database are available in Zenodo with the identifiers https://doi.org/10.5281/zenodo.3477007. Complete corpus from the National Address Dataset cannot be made publicly available to protect personal information and to follow the national policy on data security.
References
Monteiro BR, Davis CA Jr, Fonseca F (2016) A survey on the geographic scope of textual documents. Comput Geosci 96:23–34
Drummond WJ (1995) Address matching: Gis technology for mapping human activity patterns. J Am Plann Assoc 61(2):240–251
Edwards SE, Strauss B, Miranda ML (2014) Geocoding large population-level administrative datasets at highly resolved spatial scales. Trans GIS 18(4):586–603
Li F, Lu Y, Mao X, Duan J, Liu X (2022) Multi-task deep learning model based on hierarchical relations of address elements for semantic address matching. Neural Comput Appl 34(11):8919–8931
Recchia G, Louwerse M (2013) A comparison of string similarity measures for toponym matching. In Proceedings of The First ACM SIGSPATIAL international workshop on computational models of place, COMP ’13, New York NY, USA 2013. Association for Computing Machinery pp 54–61
Kılınç D (2016) An accurate toponym-matching measure based on approximate string matching. J Inf Sci 42(2):138–149
Tian Q, Ren F, Hu T, Liu J, Li R, Du Q (2016) Using an optimized chinese address matching method to develop a geocoding service: a case study of shenzhen, china. ISPRS Int J Geo Inf 5(5):65
Comber S, Arribas-Bel D (2019) Machine learning innovations in address matching: A practical comparison of word2vec and crfs. Trans GIS 23(2):334–348
Mengjun K, Qingyun D, Mingjun W (2015) A new method of chinese address extraction based on address tree model. Acta Geodaetica et Cartographica Sinica 44(1):99
Koumarelas I, Kroschk A, Mosley C, Naumann F (2018) Experience: Enhancing address matching with geocoding and similarity measure selection. J Data Inform Quality (JDIQ) 10(2):1–16
Santos R, Murrieta-Flores P, Martins B (2018) Learning to combine multiple string similarity metrics for effective toponym matching. Int J Digital Earth 11(9):913–938
Acheson E, Volpi M, Purves RS (2020) Machine learning for cross-gazetteer matching of natural features. Int J Geogr Inf Sci 34(4):708–734
Santos R, Murrieta-Flores P, Calado P, Martins B (2018) Toponym matching through deep neural networks. Int J Geogr Inf Sci 32(2):324–348
Lin Y, Kang M, Wu Y, Du Q, Liu T (2020) A deep learning architecture for semantic address matching. Int J Geogr Inf Sci 34(3):559–576
Malaviya C, Bhagavatula C, Bosselut A, Choi Y (2020) Commonsense knowledge base completion with structural and semantic context. Proceed AAAI Confer Artif Int 34(3):2925–2933
Wang Z, Li J (2016) Text-enhanced representation learning for knowledge graph. In Proceedings of the twenty-fifth international joint conference on artificial intelligence, IJCAI’16, AAAI Press pp 1293–1299
Gao T, Yao X, Chen D (2021) SimCSE: Simple contrastive learning of sentence embeddings. In Proceedings of the 2021 conference on empirical methods in natural language processing, online and Punta Cana, Dominican Republic, November 2021. Association for computational linguistics pp 6894–6910
Mohiuddin T, Joty S (2019) Revisiting adversarial autoencoder for unsupervised word translation with cycle consistency and improved training. In Proceedings of the 2019 Conference of the North American Chapter of the association for computational linguistics: human language technologies, vol 1 (Long and Short Papers), Minneapolis, Minnesota, June 2019. Association for computational linguistics pp 3857–3867
Janson S, Gogoulou E, Ylipää E, Gyllensten AC, Sahlgren M (2021) Semantic re-tuning with contrastive tension
Xu B, Luo Z, Huang L, Liang B, Xiao Y, Yang D, Wang W (2018) Metic: Multi-instance entity typing from corpus. In Proceedings of the 27th ACM International conference on information and knowledge management, CIKM ’18, New York, NY USA. Association for Computing Machinery pp 903-912
Nizzoli L, Avvenuti M, Tesconi M, Cresci S (2020) Geo-semantic-parsing: Ai-powered geoparsing by traversing semantic knowledge graphs. Decis Support Syst 136:113346
Wu T, Qi G, Luo B, Zhang L, Wang H (2019) Language-independent type inference of the instances from multilingual wikipedia. Int J Semant Web Inf Syst 15(22–46):04
Lafferty JD, McCallum A, Pereira FCN (2001) Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of the eighteenth international conference on machine learning, ICML ’01, San Francisco, CA, USA. Morgan Kaufmann Publishers Inc pp 282–289
Lev Q, Mikolov T (2014) Distributed representations of sentences and documents. In International conference on machine learning, PMLR pp 1188–1196
Cho K, van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Doha, Qatar, October 2014. Association for computational linguistics pp 1724–1734
Chen Q, Zhu X, Ling Z-H, Wei S, Jiang H,Inkpen D (2017) Enhanced LSTM for natural language inference.In Proceedings of the 55th annual meeting of the association for computational linguistics (vol 1: Long Papers), Vancouver, Canada, July 2017. Association for computational linguistics pp 1657–1668
Li J, Shang J, McAuley J (2022) UCTopic: Unsupervised contrastive learning for phrase representations and topic mining.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics, pp 6159– 6169
Li M, Lin XX, Chen X, Chang J, Zhang Q, Wang F, Wang T, Liu Z, Chu W, Zhao D, Yan R (2022) Keywords and instances: A hierarchical contrastive learning framework unifying hybrid granularities for text generation.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 4432–4441
Wang Z, Wang P, Huang L, Sun X, Wang H (2022) Incorporating hierarchy into text encoder: a contrastive learning approach for hierarchical text classification.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin,Ireland, May 2022. Association for computational linguistics pp 7109–7119
Li Y, Liu F, Collier N, Korhonen A, Vulić I (2022) Improving word translation via two-stage contrastive learning. In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 4353–4374
Wu B, Zhang Z, Wang J, Zhao H (2022) Sentence-aware contrastive learning for open-domain passage retrieval. In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 1062–1074
Zhang Y, Zhu H, Wang Y, Xu N, Li X, Zhao B (2022) A contrastive framework for learning sentence representations from pairwise and triple-wise perspective in angular space.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin Ireland, May 2022. Association for computational linguistics pp 4892–4903
Ge S, Mishra S, Li C-L, Wang H, Jacobs D (2021) Robust contrastive learning using negative samples with diminished semantics. Adv Neural Inf Process Syst 34:27356–27368
van den Oord A, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748
Khosla P, Teterwak P, Wang C, Sarna A, Tian Y, Isola P, Maschinot A, Liu C, Krishnan D (2020) Supervised contrastive learning. Adv Neural Inf Process Syst 33:18661–18673
Lin Y, Liu Z, Sun M, Liu Y, Zhu X (2015) Learning entity and relation embeddings for knowledge graph completion. Proceedings of the AAAI Conference on Artificial Intelligence 29(1)
Karpukhin V, Oguz B, Min S, Lewis P, Wu L, Edunov S, Chen D, Yih W-t (2020) Dense passage retrieval for open-domain question answering.In Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), Online, November 2020. Association for computational linguistics pp 6769–6781
Lee J, Sung M, Kang J, Chen D (2021) Learning dense representations of phrases at scale. In Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol 1: Long Papers), Online, August 2021. Association for computational linguistics pp 6634–6647
Meng Y, Xiong C, Bajaj P, Bennett P, Han J, Song X et al (2021) Coco-lm: Correcting and contrasting text sequences for language model pretraining. Adv Neural Inf Process Syst 34:23102–23114
Peters ME, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. In Proceedings of the 2018 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long Papers), New Orleans, Louisiana, June 2018. Association for computational linguistics pp 2227–2237
Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877–1901
Devlin J, Chang M-W, Lee K, Toutanova K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long and Short Papers), Minneapolis, Minnesota, June 2019. Association for computational linguistics pp 4171–4186
Qin L, Chen Q, Xie T, Li Q, Lou J-G, Che W, Kan M-Y (2022) GL-CLeF: A global–local contrastive learning framework for cross-lingual spoken language understanding.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 2677–2686
Zhang S, Cheng H, Gao J, Poon H (2022) Optimizing bi-encoder for named entity recognition via contrastive learning. arXiv preprint arXiv:2208.14565
Lu X, Deng Y, Sun T, Gao Y, Feng J, Sun X, Sutcliffe R (2022) Mkpm: Multi keyword-pair matching for natural language sentences. Appl Intell 52(2):1878–1892
Pascual D, Brunner G, Wattenhofer R Telling (2021) BERT’s full story: from local attention to global aggregation. In Proceedings of the 16th conference of the european chapter of the association for computational linguistics: Main Volume, Online, April 2021. Association for Computational Linguistics pp 105–124
Vázquez R, Celikkanat H, Ravishankar V, Creutz M, Tiedemann J (2022) A closer look at parameter contributions when training neural language and translation models. In Proceedings of the 29th international conference on computational linguistics, Gyeongju, Republic of Korea, October 2022. International committee on computational linguistics pp 4788–4800
Mengjun K, Yue L (2019) Shenzhen address corpus (part)(version v1.0).zenodo .https://doi.org/10.5281/zenodo.3477633
Wang T, Guo J, Wu Z, Xu T (2021) Ifta: Iterative filtering by using tf-aicl algorithm for chinese encyclopedia knowledge refinement. Appl Intell 51:6265–6293
Levenshtein VI, et al.(1966) Binary codes capable of correcting deletions, insertions, and reversals. In Soviet physics doklady, vol 10 Soviet Union pp 707–710
Jaccard P (1908) Nouvelles recherches sur la distribution florale. Bull Soc Vaud Sci Nat 44(223–70):01
Breiman L (2001) Random forests. Machine Learning 45(1):5–32
Hearst MA, Dumais ST, Osuna E, Platt J, Scholkopf B (1998) Support vector machines. IEEE Int Syst Appl 13(4):18–28
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L Gomez AN, Kaiser Ł,Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems 30
Ri R,Yamada I, Tsuruoka Y (2022) m LUKE: The power of entity representations in multilingual pretrained language models. In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin,Ireland, May 2022. Association for computational linguistics, pp 7316–7330
Fedus W, Zoph B, Shazeer N (2022) Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. J Mach Learn Res 23(1):5232–5270
Loshchilov I, Hutter F (2019) Decoupled weight decay regularization. In international conference on learning representations
Funding
This work is supported by the Key Cooperation Project of the Chongqing Municipal Education Commission(Grant No. HZ2021008) and Research Project of Graduate Education and Teaching Reform of Chongqing Municipal Education Commission (Grant No. yjg223087).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare no conflicts of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix 1 National dataset example
Appendix 1 National dataset example
Due to privacy issues, the National address dataset cannot be open-sourced. Therefore, we provide some national address data to help readers better understand the content and composition of the National address dataset. For more information, please refer to Table 8.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Lei, J., Wu, C. & Wang, Y. CLGLIAM: contrastive learning model based on global and local semantic interaction for address matching. Appl Intell 53, 29267–29281 (2023). https://doi.org/10.1007/s10489-023-05089-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-05089-z