Skip to main content
Log in

CLGLIAM: contrastive learning model based on global and local semantic interaction for address matching

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

As an essential part of geocoding, address matching has gained increasing research attention. Due to the long-distance dependency and unstructured property, existing address-matching methods hardly capture the contextual and implicit semantic information of unstructured text addresses. This paper presents a Contrastive Learning model based on Global and Local representation Interaction for Address Matching (referred to as CLGLIAM), which introduces a novel global and local interaction network to enhance the discrimination ability of the model on the hard negative address by associating the relationship between the global and local address representation explicitly. Simultaneously, to improve the generalization and transferability of the model, we utilize contrastive learning to enrich the data sample and extricate the model from task-specific knowledge. Furthermore, extensive experiments are conducted on Shenzhen and national address datasets to verify the effectiveness of our approach. Our model achieves state-of-the-art F1 scores of 99.26 and 98.50 on the two datasets, respectively. And the extended hard negative experiments further demonstrate the better performance of CLGLIAM in terms of semantic discrimination.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Availability of data and materials

The 498,294 records of the corpus derived from the Shenzhen Address Database are available in Zenodo with the identifiers https://doi.org/10.5281/zenodo.3477007. Complete corpus from the National Address Dataset cannot be made publicly available to protect personal information and to follow the national policy on data security.

References

  1. Monteiro BR, Davis CA Jr, Fonseca F (2016) A survey on the geographic scope of textual documents. Comput Geosci 96:23–34

    Article  Google Scholar 

  2. Drummond WJ (1995) Address matching: Gis technology for mapping human activity patterns. J Am Plann Assoc 61(2):240–251

    Article  Google Scholar 

  3. Edwards SE, Strauss B, Miranda ML (2014) Geocoding large population-level administrative datasets at highly resolved spatial scales. Trans GIS 18(4):586–603

    Article  Google Scholar 

  4. Li F, Lu Y, Mao X, Duan J, Liu X (2022) Multi-task deep learning model based on hierarchical relations of address elements for semantic address matching. Neural Comput Appl 34(11):8919–8931

    Article  Google Scholar 

  5. Recchia G, Louwerse M (2013) A comparison of string similarity measures for toponym matching. In Proceedings of The First ACM SIGSPATIAL international workshop on computational models of place, COMP ’13, New York NY, USA 2013. Association for Computing Machinery pp 54–61

  6. Kılınç D (2016) An accurate toponym-matching measure based on approximate string matching. J Inf Sci 42(2):138–149

    Article  Google Scholar 

  7. Tian Q, Ren F, Hu T, Liu J, Li R, Du Q (2016) Using an optimized chinese address matching method to develop a geocoding service: a case study of shenzhen, china. ISPRS Int J Geo Inf 5(5):65

    Article  Google Scholar 

  8. Comber S, Arribas-Bel D (2019) Machine learning innovations in address matching: A practical comparison of word2vec and crfs. Trans GIS 23(2):334–348

    Article  Google Scholar 

  9. Mengjun K, Qingyun D, Mingjun W (2015) A new method of chinese address extraction based on address tree model. Acta Geodaetica et Cartographica Sinica 44(1):99

    Google Scholar 

  10. Koumarelas I, Kroschk A, Mosley C, Naumann F (2018) Experience: Enhancing address matching with geocoding and similarity measure selection. J Data Inform Quality (JDIQ) 10(2):1–16

    Article  Google Scholar 

  11. Santos R, Murrieta-Flores P, Martins B (2018) Learning to combine multiple string similarity metrics for effective toponym matching. Int J Digital Earth 11(9):913–938

    Article  Google Scholar 

  12. Acheson E, Volpi M, Purves RS (2020) Machine learning for cross-gazetteer matching of natural features. Int J Geogr Inf Sci 34(4):708–734

    Article  Google Scholar 

  13. Santos R, Murrieta-Flores P, Calado P, Martins B (2018) Toponym matching through deep neural networks. Int J Geogr Inf Sci 32(2):324–348

    Article  Google Scholar 

  14. Lin Y, Kang M, Wu Y, Du Q, Liu T (2020) A deep learning architecture for semantic address matching. Int J Geogr Inf Sci 34(3):559–576

    Article  Google Scholar 

  15. Malaviya C, Bhagavatula C, Bosselut A, Choi Y (2020) Commonsense knowledge base completion with structural and semantic context. Proceed AAAI Confer Artif Int 34(3):2925–2933

    Google Scholar 

  16. Wang Z, Li J (2016) Text-enhanced representation learning for knowledge graph. In Proceedings of the twenty-fifth international joint conference on artificial intelligence, IJCAI’16, AAAI Press pp 1293–1299

  17. Gao T, Yao X, Chen D (2021) SimCSE: Simple contrastive learning of sentence embeddings. In Proceedings of the 2021 conference on empirical methods in natural language processing, online and Punta Cana, Dominican Republic, November 2021. Association for computational linguistics pp 6894–6910

  18. Mohiuddin T, Joty S (2019) Revisiting adversarial autoencoder for unsupervised word translation with cycle consistency and improved training. In Proceedings of the 2019 Conference of the North American Chapter of the association for computational linguistics: human language technologies, vol 1 (Long and Short Papers), Minneapolis, Minnesota, June 2019. Association for computational linguistics pp 3857–3867

  19. Janson S, Gogoulou E, Ylipää E, Gyllensten AC, Sahlgren M (2021) Semantic re-tuning with contrastive tension

  20. Xu B, Luo Z, Huang L, Liang B, Xiao Y, Yang D, Wang W (2018) Metic: Multi-instance entity typing from corpus. In Proceedings of the 27th ACM International conference on information and knowledge management, CIKM ’18, New York, NY USA. Association for Computing Machinery pp 903-912

  21. Nizzoli L, Avvenuti M, Tesconi M, Cresci S (2020) Geo-semantic-parsing: Ai-powered geoparsing by traversing semantic knowledge graphs. Decis Support Syst 136:113346

    Article  Google Scholar 

  22. Wu T, Qi G, Luo B, Zhang L, Wang H (2019) Language-independent type inference of the instances from multilingual wikipedia. Int J Semant Web Inf Syst 15(22–46):04

    Google Scholar 

  23. Lafferty JD, McCallum A, Pereira FCN (2001) Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of the eighteenth international conference on machine learning, ICML ’01, San Francisco, CA, USA. Morgan Kaufmann Publishers Inc pp 282–289

  24. Lev Q, Mikolov T (2014) Distributed representations of sentences and documents. In International conference on machine learning, PMLR pp 1188–1196

  25. Cho K, van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Doha, Qatar, October 2014. Association for computational linguistics pp 1724–1734

  26. Chen Q, Zhu X, Ling Z-H, Wei S, Jiang H,Inkpen D (2017) Enhanced LSTM for natural language inference.In Proceedings of the 55th annual meeting of the association for computational linguistics (vol 1: Long Papers), Vancouver, Canada, July 2017. Association for computational linguistics pp 1657–1668

  27. Li J, Shang J, McAuley J (2022) UCTopic: Unsupervised contrastive learning for phrase representations and topic mining.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics, pp 6159– 6169

  28. Li M, Lin XX, Chen X, Chang J, Zhang Q, Wang F, Wang T, Liu Z, Chu W, Zhao D, Yan R (2022) Keywords and instances: A hierarchical contrastive learning framework unifying hybrid granularities for text generation.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 4432–4441

  29. Wang Z, Wang P, Huang L, Sun X, Wang H (2022) Incorporating hierarchy into text encoder: a contrastive learning approach for hierarchical text classification.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin,Ireland, May 2022. Association for computational linguistics pp 7109–7119

  30. Li Y, Liu F, Collier N, Korhonen A, Vulić I (2022) Improving word translation via two-stage contrastive learning. In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 4353–4374

  31. Wu B, Zhang Z, Wang J, Zhao H (2022) Sentence-aware contrastive learning for open-domain passage retrieval. In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 1062–1074

  32. Zhang Y, Zhu H, Wang Y, Xu N, Li X, Zhao B (2022) A contrastive framework for learning sentence representations from pairwise and triple-wise perspective in angular space.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin Ireland, May 2022. Association for computational linguistics pp 4892–4903

  33. Ge S, Mishra S, Li C-L, Wang H, Jacobs D (2021) Robust contrastive learning using negative samples with diminished semantics. Adv Neural Inf Process Syst 34:27356–27368

  34. van den Oord A, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748

  35. Khosla P, Teterwak P, Wang C, Sarna A, Tian Y, Isola P, Maschinot A, Liu C, Krishnan D (2020) Supervised contrastive learning. Adv Neural Inf Process Syst 33:18661–18673

    Google Scholar 

  36. Lin Y, Liu Z, Sun M, Liu Y, Zhu X (2015) Learning entity and relation embeddings for knowledge graph completion. Proceedings of the AAAI Conference on Artificial Intelligence 29(1)

  37. Karpukhin V, Oguz B, Min S, Lewis P, Wu L, Edunov S, Chen D, Yih W-t (2020) Dense passage retrieval for open-domain question answering.In Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), Online, November 2020. Association for computational linguistics pp 6769–6781

  38. Lee J, Sung M, Kang J, Chen D (2021) Learning dense representations of phrases at scale. In Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol 1: Long Papers), Online, August 2021. Association for computational linguistics pp 6634–6647

  39. Meng Y, Xiong C, Bajaj P, Bennett P, Han J, Song X et al (2021) Coco-lm: Correcting and contrasting text sequences for language model pretraining. Adv Neural Inf Process Syst 34:23102–23114

  40. Peters ME, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. In Proceedings of the 2018 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long Papers), New Orleans, Louisiana, June 2018. Association for computational linguistics pp 2227–2237

  41. Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877–1901

    Google Scholar 

  42. Devlin J, Chang M-W, Lee K, Toutanova K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long and Short Papers), Minneapolis, Minnesota, June 2019. Association for computational linguistics pp 4171–4186

  43. Qin L, Chen Q, Xie T, Li Q, Lou J-G, Che W, Kan M-Y (2022) GL-CLeF: A global–local contrastive learning framework for cross-lingual spoken language understanding.In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin, Ireland, May 2022. Association for computational linguistics pp 2677–2686

  44. Zhang S, Cheng H, Gao J, Poon H (2022) Optimizing bi-encoder for named entity recognition via contrastive learning. arXiv preprint arXiv:2208.14565

  45. Lu X, Deng Y, Sun T, Gao Y, Feng J, Sun X, Sutcliffe R (2022) Mkpm: Multi keyword-pair matching for natural language sentences. Appl Intell 52(2):1878–1892

    Article  Google Scholar 

  46. Pascual D, Brunner G, Wattenhofer R Telling (2021) BERT’s full story: from local attention to global aggregation. In Proceedings of the 16th conference of the european chapter of the association for computational linguistics: Main Volume, Online, April 2021. Association for Computational Linguistics pp 105–124

  47. Vázquez R, Celikkanat H, Ravishankar V, Creutz M, Tiedemann J (2022) A closer look at parameter contributions when training neural language and translation models. In Proceedings of the 29th international conference on computational linguistics, Gyeongju, Republic of Korea, October 2022. International committee on computational linguistics pp 4788–4800

  48. Mengjun K, Yue L (2019) Shenzhen address corpus (part)(version v1.0).zenodo .https://doi.org/10.5281/zenodo.3477633

  49. Wang T, Guo J, Wu Z, Xu T (2021) Ifta: Iterative filtering by using tf-aicl algorithm for chinese encyclopedia knowledge refinement. Appl Intell 51:6265–6293

    Article  Google Scholar 

  50. Levenshtein VI, et al.(1966) Binary codes capable of correcting deletions, insertions, and reversals. In Soviet physics doklady, vol 10 Soviet Union pp 707–710

  51. Jaccard P (1908) Nouvelles recherches sur la distribution florale. Bull Soc Vaud Sci Nat 44(223–70):01

    Google Scholar 

  52. Breiman L (2001) Random forests. Machine Learning 45(1):5–32

    Article  MATH  Google Scholar 

  53. Hearst MA, Dumais ST, Osuna E, Platt J, Scholkopf B (1998) Support vector machines. IEEE Int Syst Appl 13(4):18–28

    Article  Google Scholar 

  54. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L Gomez AN, Kaiser Ł,Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems 30

  55. Ri R,Yamada I, Tsuruoka Y (2022) m LUKE: The power of entity representations in multilingual pretrained language models. In Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), Dublin,Ireland, May 2022. Association for computational linguistics, pp 7316–7330

  56. Fedus W, Zoph B, Shazeer N (2022) Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. J Mach Learn Res 23(1):5232–5270

    MathSciNet  Google Scholar 

  57. Loshchilov I, Hutter F (2019) Decoupled weight decay regularization. In international conference on learning representations

Download references

Funding

This work is supported by the Key Cooperation Project of the Chongqing Municipal Education Commission(Grant No. HZ2021008) and Research Project of Graduate Education and Teaching Reform of Chongqing Municipal Education Commission (Grant No. yjg223087).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jianjun Lei.

Ethics declarations

Conflicts of interest

The authors declare no conflicts of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix 1 National dataset example

Appendix 1 National dataset example

Due to privacy issues, the National address dataset cannot be open-sourced. Therefore, we provide some national address data to help readers better understand the content and composition of the National address dataset. For more information, please refer to Table 8.

Table 8 Samples of the national address data

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lei, J., Wu, C. & Wang, Y. CLGLIAM: contrastive learning model based on global and local semantic interaction for address matching. Appl Intell 53, 29267–29281 (2023). https://doi.org/10.1007/s10489-023-05089-z

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-05089-z

Keywords

Navigation