skip to main content
research-article

Multi-aspect Understanding with Cooperative Graph Attention Networks for Medical Dialogue Information Extraction

Published: 14 November 2023 Publication History

Abstract

Medical dialogue information extraction is an important but challenging task for Electronic Medical Records. Existing medical information extraction methods ignore the crucial information of sentence and multi-level dependency in dialogue, which limits their effectiveness for capturing essential medical information. To address these issues, we present a novel Multi-aspect Understanding with Cooperative Graph Attention Networks for Medical Dialogue Information Extraction to capture multi-aspect sentence information and multi-level dependency information from the dialogue. First, we propose the multi-aspect sentence encoder to capture various features from different perspectives. Second, we propose double graph attention networks to model the dependency features from intra-window and inter-window, respectively. Extensive experiments on a benchmark dataset have well-validated the effectiveness of the proposed method.

References

[1]
Daniel Adiwardana, Minh-Thang Luong, David R. So, Jamie Hall, Noah Fiedel, Romal Thoppilan, Zi Yang, Apoorv Kulshreshtha, Gaurav Nemade, Yifeng Lu, and Quoc V. Le. 2020. Towards a human-like open-domain chatbot. Retrieved from https://arxiv.org/abs/2001.09977
[2]
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. In Proceedings of NeurIPS, Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (Eds.). Retrieved from https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html
[3]
Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. 2014. Spectral networks and locally connected networks on graphs. In Proceedings of ICLR.
[4]
Razvan C. Bunescu and Raymond J. Mooney. 2005. A shortest path dependency kernel for relation extraction. In Proceedings of EMNLP. 724–731.
[5]
Kyunghyun Cho, Bart Van Merrienboer, Çaglar Gülçehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. Proceedings of EMNLP, Alessandro Moschitti, Bo Pang, and Walter Daelemans (Eds.). 1724–1734. https://dblp.org/rec/conf/emnlp/ChoMGBBSB14.bib
[6]
Thomas G. Dietterich, Richard H. Lathrop, and Tomás Lozano-Pérez. 1997. Solving the multiple instance problem with axis-parallel rectangles. Artif. Intell. 89, 1-2 (1997), 31–71.
[7]
Nan Du, Kai Chen, Anjuli Kannan, Linh Tran, Yuhui Chen, and Izhak Shafran. 2019. Extracting symptoms and their status from clinical conversations. In Proceedings of ACL. 915–925.
[8]
Zhijiang Guo, Yan Zhang, and Wei Lu. 2019. Attention guided graph convolutional networks for relation extraction. Retrieved from http://arxiv.org/abs/1906.07510
[9]
Raphael Hoffmann, Congle Zhang, Xiao Ling, Luke Zettlemoyer, and Daniel S. Weld. 2011. Knowledge-based weak supervision for information extraction of overlapping relations. In Proceedings of ACL. 541–550.
[10]
Guoliang Ji, Kang Liu, Shizhu He, and Jun Zhao. 2017. Distant supervision for relation extraction with sentence-level attention and entity descriptions. In Proceedings of AAAI. 3060–3066.
[11]
Kenton Lee, Luheng He, Mike Lewis, and Luke Zettlemoyer. 2017. End-to-end neural coreference resolution. In Proceedings of EMNLP. 188–197.
[12]
Jiao Li, Yueping Sun, Robin J. Johnson, Daniela Sciaky, Chih-Hsuan Wei, Robert Leaman, Allan Peter Davis, Carolyn J. Mattingly, Thomas C. Wiegers, and Zhiyong Lu. 2016. BioCreative V CDR task corpus: A resource for chemical disease relation extraction. Database J. Biol. Databases Cur. 2016 (2016).
[13]
Yaliang Li, Chaochun Liu, Nan Du, Wei Fan, Qi Li, Jing Gao, Chenwei Zhang, and Hao Wu. 2020. Extracting medical knowledge from crowdsourced question answering website. IEEE Trans. Big Data 6, 2 (2020), 309–321.
[14]
Xinzhu Lin, Xiahui He, Qin Chen, Huaixiao Tou, Zhongyu Wei, and Ting Chen. 2019. Enhancing dialogue symptom diagnosis with global attention and symptom graph. In Proceedings of EMNLP-IJCNLP. 5032–5041.
[15]
Yankai Lin, Shiqi Shen, Zhiyuan Liu, Huanbo Luan, and Maosong Sun. 2016. Neural relation extraction with selective attention over instances. In Proceedings of ACL. 2124–2133.
[16]
Tianyi Liu, Xinsong Zhang, Wanhao Zhou, and Weijia Jia. 2018. Neural relation extraction via inner-sentence noise reduction and transfer learning. In Proceedings of EMNLP. 2195–2204.
[17]
Wenge Liu, Jianheng Tang, Jinghui Qin, Lin Xu, Zhen Li, and Xiaodan Liang. 2020. MedDG: A large-scale medical consultation dataset for building medical dialogue system. Retrieved from https://arxiv.org/abs/2010.07497
[18]
Yang Liu, Furu Wei, Sujian Li, Heng Ji, Ming Zhou, and Houfeng Wang. 2015. A dependency-based neural network for relation classification. In Proceedings of ACL. 285–290. Retrieved from http://aclweb.org/anthology/P/P15/P15-2047.pdf
[19]
Tomas Mikolov, Ilya Sutskever, Kai Chen, Gregory S. Corrado, and Jeffrey Dean. 2013. Distributed representations of words and phrases and their compositionality. In Proceedings of NIPS. 3111–3119.
[20]
Mike Mintz, Steven Bills, Rion Snow, and Dan Jurafsky. 2009. Distant supervision for relation extraction without labeled data. In Proceedings of ACL-IJCNLP. 1003–1011.
[21]
Makoto Miwa and Mohit Bansal. 2016. End-to-end relation extraction using LSTMs on sequences and tree structures. In Proceedings of ACL. 1105–1116.
[22]
Sebastian Riedel, Limin Yao, and Andrew McCallum. 2010. Modeling relations and their mentions without labeled text. In Proceedings of ECML/PKDD. 148–163.
[23]
Adam Santoro, David Raposo, David G. T. Barrett, Mateusz Malinowski, Razvan Pascanu, Peter W. Battaglia, and Tim Lillicrap. 2017. A simple neural network module for relational reasoning. In Proceedings of NIPS. 4967–4976.
[24]
Cicero Nogueira Dos Santos, Bing Xiang, and Bowen Zhou. 2015. Classifying relations by ranking with convolutional neural networks. Comput. Sci. 86, 86 (2015), 132–137.
[25]
Ge Shi, Chong Feng, Lifu Huang, Boliang Zhang, Heng Ji, Lejian Liao, and He-Yan Huang. 2018. Genre separation network with adversarial training for cross-genre relation extraction. In Proceedings of EMNLP. 1018–1023.
[26]
Ge Shi, Chong Feng, Wenfu Xu, Lejian Liao, and Heyan Huang. 2020. Penalized multiple distribution selection method for imbalanced data classification. Knowledge-Based Syst. 196 (2020), 105833.
[27]
Chaitanya P. Shivade, Preethi Raghavan, Eric Fosler-Lussier, Peter J. Embi, Noémie Elhadad, Stephen B. Johnson, and Albert M. Lai. 2014. A review of approaches to identifying patient phenotype cohorts using electronic health records. J. Am. Medical Info. Assoc. 21, 2 (2014), 221–230.
[28]
Nitish Srivastava, Geoffrey E. Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1 (2014), 1929–1958.
[29]
Martin Sundermeyer, Ralf Schluter, and Hermann Ney. 2012. LSTM neural networks for language modeling. In Proceedings of INTERSPEECH. 601–608.
[30]
Nhi-Ha T. Trinh, Soo Jeong Youn, Jessica Sousa, Susan Regan, C. Andres Bedoya, Trina E. Chang, Maurizio Fava, and Albert Yeung. 2011. Using electronic medical records to determine the diagnosis of clinical depression. Int. J. Med. Inform. 80, 7 (2011), 533–540.
[31]
Özlem Uzuner, Brett R. South, Shuying Shen, and Scott L. DuVall. 2011. 2010 i2b2/VA challenge on concepts, assertions, and relations in clinical text. J. Am. Med. Inform. Assoc. 18, 5 (2011), 552–556.
[32]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Proceedings of NIPS. 6000–6010.
[33]
Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. 2017. Graph attention networks. Retrieved from https://arxiv.org/abs/1710.10903
[34]
Danqing Wang, Pengfei Liu, Yining Zheng, Xipeng Qiu, and Xuanjing Huang. 2020. Heterogeneous graph neural networks for extractive document summarization. In Proceedings of ACL. 6209–6219.
[35]
Yan Xu, Etsuko Ishii, Samuel Cahyawijaya, Zihan Liu, Genta Indra Winata, Andrea Madotto, Dan Su, and Pascale Fung. 2022. Retrieval-free knowledge-grounded dialogue response generation with adapters. In Proceedings of DialDoc@ACL, Song Feng, Hui Wan, Caixia Yuan, and Han Yu (Eds.). Association for Computational Linguistics, 93–107.
[36]
Xu Yan, Lili Mou, Ge Li, Yunchuan Chen, Hao Peng, and Zhi Jin. 2015. Classifying relations via long short term memory networks along shortest dependency path. Comput. Sci. 42, 1 (2015), 56–61.
[37]
Zhi-Xiu Ye and Zhen-Hua Ling. 2019. Distant supervision relation extraction with intra-bag and inter-bag attentions. In Proceedings of NAACL. 2810–2819.
[38]
Dmitry Zelenko, Chinatsu Aone, and Anthony Richardella. 2002. Kernel methods for relation extraction. In Proceedings of EMNLP.
[39]
Daojian Zeng, Kang Liu, Yubo Chen, and Jun Zhao. 2015. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of EMNLP. 1753–1762.
[40]
D. Zeng, K. Liu, S. Lai, G. Zhou, and J. Zhao. 2014. Relation classification via convolutional deep neural network. In Proceedings of COLING. 2335–2344.
[41]
Daojian Zeng, Kang Liu, Siwei Lai, Guangyou Zhou, and Jun Zhao. 2014. Relation classification via convolutional deep neural network. In Proceedings of COLING. 2335–2344.
[42]
Yuanzhe Zhang, Zhongtao Jiang, Tao Zhang, Shiwan Liu, Jiarun Cao, Kang Liu, Shengping Liu, and Jun Zhao. 2020. MIE: A medical information extractor towards medical dialogues. In Proceedings of ACL. 6460–6469.
[43]
Yuhao Zhang, Peng Qi, and Christopher D. Manning. 2018. Graph convolution over pruned dependency trees improves relation extraction. In Proceedings of EMNLP. 2205–2215.
[44]
Yufeng Zhang, Xueli Yu, Zeyu Cui, Shu Wu, Zhongzhen Wen, and Liang Wang. 2020. Every document owns its structure: Inductive text classification via graph neural networks. In Proceedings of ACL. 334–339.
[45]
Bo Zheng, Haoyang Wen, Yaobo Liang, Nan Duan, Wanxiang Che, Daxin Jiang, Ming Zhou, and Ting Liu. 2020. Document modeling with graph attention networks for multi-grained machine reading comprehension. In Proceedings of ACL. 6708–6718.
[46]
Guodong Zhou, Jian Su, Jie Zhang, and Min Zhang. 2005. Exploring various knowledge in relation extraction. In Proceedings of ACL. 427–434.
[47]
Peng Zhou, Wei Shi, Jun Tian, Zhenyu Qi, Bingchen Li, Hongwei Hao, and Bo Xu. 2016. Attention-based bidirectional long short-term memory networks for relation classification. In Proceedings of ACL. 207–212.

Cited By

View all
  • (2024)Counterfactual Graph Convolutional Learning for Personalized RecommendationACM Transactions on Intelligent Systems and Technology10.1145/365563215:4(1-20)Online publication date: 18-Jun-2024
  • (2024)DIPE: a diagnosis-assisted inquiry point extractor towards medical dialoguesApplied Intelligence10.1007/s10489-024-06138-x55:3Online publication date: 27-Dec-2024
  • (undefined)Cascaded Alternating Refinement Transformer for Few-shot Medical Image SegmentationACM Transactions on Intelligent Systems and Technology10.1145/3709145

Index Terms

  1. Multi-aspect Understanding with Cooperative Graph Attention Networks for Medical Dialogue Information Extraction

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Intelligent Systems and Technology
    ACM Transactions on Intelligent Systems and Technology  Volume 14, Issue 6
    December 2023
    493 pages
    ISSN:2157-6904
    EISSN:2157-6912
    DOI:10.1145/3632517
    • Editor:
    • Huan Liu
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 14 November 2023
    Online AM: 05 September 2023
    Accepted: 11 August 2023
    Revised: 30 March 2023
    Received: 17 August 2022
    Published in TIST Volume 14, Issue 6

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Natural language processing
    2. relation extraction
    3. medical dialogue information extraction
    4. graph attention network

    Qualifiers

    • Research-article

    Funding Sources

    • National Natural Science Foundation of China
    • MOE (Ministry of Education in China) Project of Humanities and Social Sciences
    • Yunnan Natural Science Foundation of China

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)67
    • Downloads (Last 6 weeks)2
    Reflects downloads up to 02 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Counterfactual Graph Convolutional Learning for Personalized RecommendationACM Transactions on Intelligent Systems and Technology10.1145/365563215:4(1-20)Online publication date: 18-Jun-2024
    • (2024)DIPE: a diagnosis-assisted inquiry point extractor towards medical dialoguesApplied Intelligence10.1007/s10489-024-06138-x55:3Online publication date: 27-Dec-2024
    • (undefined)Cascaded Alternating Refinement Transformer for Few-shot Medical Image SegmentationACM Transactions on Intelligent Systems and Technology10.1145/3709145

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media