skip to main content
10.1145/3289600.3290984acmconferencesArticle/Chapter ViewAbstractPublication PageswsdmConference Proceedingsconference-collections
research-article

Domain Adaptation for Commitment Detection in Email

Published:30 January 2019Publication History

ABSTRACT

People often make commitments to perform future actions. Detecting commitments made in email (e.g., "I'll send the report by end of day'') enables digital assistants to help their users recall promises they have made and assist them in meeting those promises in a timely manner. In this paper, we show that commitments can be reliably extracted from emails when models are trained and evaluated on the same domain (corpus). However, their performance degrades when the evaluation domain differs. This illustrates the domain bias associated with email datasets and a need for more robust and generalizable models for commitment detection. To learn a domain-independent commitment model, we first characterize the differences between domains (email corpora) and then use this characterization to transfer knowledge between them. We investigate the performance of domain adaptation, namely transfer learning, at different granularities: feature-level adaptation and sample-level adaptation. We extend this further using a neural autoencoder trained to learn a domain-independent representation for training samples. We show that transfer learning can help remove domain bias to obtain models with less domain dependence. Overall, our results show that domain differences can have a significant negative impact on the quality of commitment detection models and that transfer learning has enormous potential to address this issue.

References

  1. Ron Artstein and Massimo Poesio. 2008. Inter-coder agreement for computational linguistics. Computational Linguistics, Vol. 34, 4 (2008), 555--596. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Hosein Azarbonyad, Mostafa Dehghani, Kaspar Beelen, Alexandra Arkut, Maarten Marx, and Jaap Kamps. 2017. Words Are Malleable: Computing Semantic Shifts in Political and Media Discourse. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management . 1509--1518. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Victoria Bellotti, Nicolas Ducheneaut, Mark Howard, and Ian Smith. 2003. Taking email to task: the design and evaluation of a task management centered email tool. In Proceedings of the SIGCHI conference on Human factors in computing systems. 345--352. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. 2010. A Theory of Learning from Different Domains. Machine Learning, Vol. 79, 1--2 (2010), 151--175. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Yoshua Bengio et almbox. 2009. Learning deep architectures for AI. Foundations and trends® in Machine Learning, Vol. 2, 1 (2009), 1--127. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Paul N Bennett and Jaime Carbonell. 2005. Detecting action-items in e-mail. In Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval . 585--586. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. John Blitzer, Mark Dredze, and Fernando Pereira. 2007. Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In Proceedings of the 45th annual meeting of the association of computational linguistics. 440--447.Google ScholarGoogle Scholar
  8. Konstantinos Bousmalis, George Trigeorgis, Nathan Silberman, Dilip Krishnan, and Dumitru Erhan. 2016. Domain separation networks. In Proceedings of Advances in Neural Information Processing Systems. 343--351. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Minmin Chen, Zhixiang Xu, Kilian Weinberger, and Fei Sha. 2012. Marginalized denoising autoencoders for domain adaptation. arXiv preprint arXiv:1206.4683 (2012). Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Xilun Chen and Claire Cardie. 2018. Multinomial Adversarial Networks for Multi-Domain Text Classification. arXiv preprint arXiv:1802.05694 (2018).Google ScholarGoogle Scholar
  11. Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014).Google ScholarGoogle Scholar
  12. William W Cohen, Vitor R Carvalho, and Tom M Mitchell. 2004. Learning to Classify Email into“Speech Acts”. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing .Google ScholarGoogle Scholar
  13. Simon Corston-Oliver, Eric Ringger, Michael Gamon, and Richard Campbell. 2004. Task-focused summarization of email. Text Summarization Branches Out (2004).Google ScholarGoogle Scholar
  14. Wenyuan Dai, Qiang Yang, Gui-Rong Xue, and Yong Yu. 2007. Boosting for transfer learning. In Proceedings of the 24th International Conference on Machine Learning. 193--200. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Rachele De Felice. 2013. A corpus-based classification of commitments in Business English. In Yearbook of Corpus Linguistics and Pragmatics 2013. 153--171.Google ScholarGoogle Scholar
  16. Michael Gamon, Saliha Azzam, Yizheng Cai, Nicholas Caldwell, and Ye-Yi Wang. 2013. Automatic Task Extraction and Calendar Entry. (2013). US Patent App. 13/170,660.Google ScholarGoogle Scholar
  17. Yaroslav Ganin and Victor Lempitsky. 2014. Unsupervised domain adaptation by backpropagation. arXiv preprint arXiv:1409.7495 (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, Francc ois Laviolette, Mario Marchand, and Victor Lempitsky. 2016. Domain-adversarial training of neural networks. Journal of Machine Learning Research, Vol. 17, 1 (2016), 2096--2030. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Xavier Glorot, Antoine Bordes, and Yoshua Bengio. 2011. Domain adaptation for large-scale sentiment classification: A deep learning approach. In Proceedings of the 28th International Conference on Machine Learning (ICML-11). 513--520. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Jacek Gwizdka. 2002. Reinventing the inbox: supporting the management of pending tasks in email. In CHI'02 Extended Abstracts on Human Factors in Computing Systems. 550--551. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation, Vol. 9, 8 (1997), 1735--1780. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Jing Jiang and ChengXiang Zhai. 2007. Instance weighting for domain adaptation in NLP. In Proceedings of the 45th annual meeting of the association of computational linguistics. 264--271.Google ScholarGoogle Scholar
  23. Anup K Kalia, Hamid R Motahari-Nezhad, Claudio Bartolini, and Munindar P Singh. 2013. Monitoring commitments in people-driven service engagements. In Services Computing (SCC), 2013 IEEE International Conference on. 160--167. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Bryan Klimt and Yiming Yang. 2004. The enron corpus: A new dataset for email classification research. In Proceedings of European Conference on Machine Learning. 217--226. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Wouter M. Kouw, Laurens J.P. van der Maaten, Jesse H. Krijthe, and Marco Loog. 2016. Feature-Level Domain Adaptation. Journal of Machine Learning Research, Vol. 17, 171 (2016), 1--32. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Andrew Lampert, Robert Dale, and Cecile Paris. 2008. Requests and Commitments in Email are More Complex Than You Think: Eight Reasons to be Cautious. In Proceedings of the Australasian Language Technology Association Workshop 2008. 64--72.Google ScholarGoogle Scholar
  27. Andrew Lampert, Robert Dale, and Cecile Paris. 2010. Detecting emails containing requests for action. In Proceedings of Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics. 984--992. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Andrew Lampert, Cécile Paris, Robert Dale, et almbox. 2007. Can requests-for-action and commitments-to-act be reliably identified in email messages. In Proceedings of the 12th Australasian Document Computing Symposium. 48--55.Google ScholarGoogle Scholar
  29. Guillaume Lample, Alexis Conneau, Marc'Aurelio Ranzato, Ludovic Denoyer, and Herve Jegou. 2018. Word translation without parallel data. In Proceedings of International Conference on Learning Representations .Google ScholarGoogle Scholar
  30. Bin Li, Qiang Yang, and Xiangyang Xue. 2009. Transfer learning for collaborative filtering via a rating-matrix generative model. In Proceedings of the 26th annual International Conference on Machine Learning . 617--624. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Chu-Cheng Lin, Dongyeop Kang, Michael Gamon, Madian Khabsa, Ahmed Hassan Awadallah, and Patrick Pantel. 2017. Actionable Email Intent Modeling with Reparametrized RNNs. arXiv preprint arXiv:1712.09185 (2017).Google ScholarGoogle Scholar
  32. Pengfei Liu, Xipeng Qiu, and Xuanjing Huang. 2017. Adversarial Multi-task Learning for Text Classification. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics . 1--10.Google ScholarGoogle ScholarCross RefCross Ref
  33. Tomas Mikolov, Quoc V Le, and Ilya Sutskever. 2013a. Exploiting similarities among languages for machine translation. arXiv preprint arXiv:1309.4168 (2013).Google ScholarGoogle Scholar
  34. Tomas Mikolov, Quoc V. Le, and Ilya Sutskever. 2013b. Exploiting Similarities among Languages for Machine Translation. In Proceedings of International Conference on Learning Representations .Google ScholarGoogle Scholar
  35. Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013c. Distributed representations of words and phrases and their compositionality. In Proceedings of Advances in Neural Information Processing Systems . 3111--3119. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. Hamid R Motahari Nezhad, Kalpa Gunaratna, and Juan Cappi. 2017. eassistant: Cognitive assistance for identification and auto-triage of actionable conversations. In Proceedings of the 26th International Conference on World Wide Web Companion. 89--98. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Douglas Oard, William Webber, David Kirsch, and Sergey Golitsynskiy. 2015. Avocado research email collection. Linguistic Data Consortium (2015).Google ScholarGoogle Scholar
  38. Sinno Jialin Pan and Qiang Yang. 2010. A survey on transfer learning. IEEE Transactions on knowledge and data engineering, Vol. 22, 10 (2010), 1345--1359. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Chris Quirk, Pallavi Choudhury, Jianfeng Gao, Hisami Suzuki, Kristina Toutanova, Michael Gamon, Wen-tau Yih, Lucy Vanderwende, and Colin Cherry. 2012. MSR SPLAT, a language analysis toolkit. In Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Demonstration Session. 21--24. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Rajat Raina, Andrew Y Ng, and Daphne Koller. 2006. Constructing informative priors using transfer learning. In Proceedings of the 23rd International Conference on Machine learning. 713--720. Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. Herbert Robbins and Sutton Monro. 1951. A stochastic approximation method. The annals of mathematical statistics (1951), 400--407.Google ScholarGoogle Scholar
  42. Hidetoshi Shimodaira. 2000. Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of statistical planning and inference, Vol. 90, 2 (2000), 227--244.Google ScholarGoogle ScholarCross RefCross Ref
  43. Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to sequence learning with neural networks. In Proceedings of Advances in neural information processing systems. 3104--3112. Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Gokhan Tur, Andreas Stolcke, Lynn Voss, Stanley Peters, Dilek Hakkani-Tur, John Dowding, Benoit Favre, Raquel Fernández, Matthew Frampton, Mike Frandsen, et almbox. 2010. The CALO meeting assistant system. IEEE Transactions on Audio, Speech, and Language Processing, Vol. 18, 6 (2010), 1601--1611.Google ScholarGoogle Scholar
  45. Steve Whittaker and Candace Sidner. 1996. Email overload: exploring personal information management of email. In Proceedings of the SIGCHI conference on Human factors in computing systems . 276--283. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Pengcheng Wu and Thomas G Dietterich. 2004. Improving SVM accuracy by training on auxiliary data sources. In Proceedings of the twenty-first International Conference on Machine learning. 110. Google ScholarGoogle ScholarDigital LibraryDigital Library
  47. Fuzhen Zhuang, Xiaohu Cheng, Ping Luo, Sinno Jialin Pan, and Qing He. 2015. Supervised Representation Learning: Transfer Learning with Deep Autoencoders. In Proceedings of the 24th International Conference on Artificial Intelligence . 4119--4125. Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. Fuzhen Zhuang, Ping Luo, Hui Xiong, Yuhong Xiong, Qing He, and Zhongzhi Shi. 2010. Cross-domain learning from multiple sources: A consensus regularization perspective. IEEE Transactions on Knowledge and Data Engineering, Vol. 22, 12 (2010), 1664--1678. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Domain Adaptation for Commitment Detection in Email

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            WSDM '19: Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining
            January 2019
            874 pages
            ISBN:9781450359405
            DOI:10.1145/3289600

            Copyright © 2019 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 30 January 2019

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            WSDM '19 Paper Acceptance Rate84of511submissions,16%Overall Acceptance Rate498of2,863submissions,17%

            Upcoming Conference

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader