Skip to main content
Log in

Learning to Predict U.S. Policy Change Using New York Times Corpus with Pre-Trained Language Model

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

With the process of economic globalization and political multi-polarization accelerating, it is especially important to predict policy change in the United States. While current research has not taken advantage of the rapid advancement in the natural language processing and the relationship between news media and policy change, we propose a BERT-based model to predict policy change in the United States, using news published by the New York Times. Specifically, we propose a large-scale news corpus from the New York Times covers the period from 2006 to 2018. Then we use the corpus to fine-tune the pre-trained BERT language model to determine whether the news is on the front page, which corresponds to the policy priority. We propose a BERT-based Policy Change Index (BPCI) for the United States to predict the policy change in the future short period of time. Experimental results in the New York Times Corpus demonstrate the validity of the proposed method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Ansolabehere S, Iyengar S (1994) Riding the wave and claiming ownership over issues: the joint effects of advertising and news coverage in campaigns[J]. Public Opinion Quarterly 58(3):335–357

    Article  Google Scholar 

  2. Bengio Y, Ducharme R, Vincent P, Jauvin C (2003) A neural probabilistic language model. J Mach Learn Res, 1137–1155

  3. Carlstrom CT, Zaman S (2014) Using an improved Taylor rule to predict when policy changes will occur[J]. Economic Commentary, (2014–02).

  4. Chan JTK, Zhong W (2018) Reading China: predicting policy change with machine learning[J]. AEI Economics Working Paper

  5. DellaVigna S, Kaplan E (2007) The fox news effect: media bias and voting[J]. Q J Econ 122(3):1187–1234

    Article  Google Scholar 

  6. Devlin J, Chang MW, Lee K et al. (2018) Bert: Pre-training of deep bidirectional transformers for language understanding[J]. arXiv preprint arXiv:1810.04805

  7. Enikolopov R, Petrova M, Zhuravskaya E (2011) Media and political persuasion: evidence from Russia[J]. Am Econ Rev 101(7):3253–3285

    Article  Google Scholar 

  8. Gerber AS, Karlan D, Bergan D (2009) Does the media matter? A field experiment measuring the effect of newspapers on voting behavior and political opinions[J]. Am Econ J Appl Econ 1(2):35–52

    Article  Google Scholar 

  9. Girshick R (2015) Fast r-cnn[C]//Proceedings of the IEEE international conference on computer vision. 1440–1448

  10. Graff D (1995) North American News Text Corpus[C]// Linguistic Data Consortium. LDC95T21

  11. Graves A, Mohamed A, Hinton G (2013) Speech recognition with deep recurrent neural networks[C]//2013 IEEE international conference on acoustics, speech and signal processing. IEEE, 6645–6649

  12. Grossman M (2005) Role theory and foreign policy change: the transformation of Russian foreign policy in the 1990s[J]. International Politics 42(3):334–351

    Article  Google Scholar 

  13. Hearings. The Policy Agendas Project at the University of Texas at Austin, 2017. www.comparativeagendas.net. Accessed September 26, 2017.

  14. Hochreiter S, Schmidhuber J (1997) Long short-term memory.[J]. Neural Comput 9(8):1735–1780

    Article  Google Scholar 

  15. Joulin A, Grave E, Bojanowski P, Mikolov T (2017) Bag of Tricks for Efficient Text Classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics. 427–431.

  16. Katz DM, Bommarito II, Michael J et al. (2014) Predicting the behavior of the supreme court of the united states: A general approach[J]. arXiv preprint arXiv:1407.6333

  17. Katz DM, Bommarito MJ II, Blackman J (2017) A general approach for predicting the behavior of the supreme court of the United States. PLoS One 12(4):e0174698

    Article  Google Scholar 

  18. Kim Y (2014) Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). 1746-1751

  19. Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks[C]//Advances in neural information processing systems. 1097–1105

  20. Lai S, Xu L, Liu K, Zhao J (2015) Recurrent convolutional neural networks for text classification. In AAAI conference on artificial intelligence

  21. Le Q, Mikolov T (2014) Distributed representations of sentences and documents[C]//International conference on machine learning. 1188–1196

  22. LeCun Y, Bengio Y, Hinton G (2015) Deep learning[J]. Nature 521(7553):436–444

    Article  Google Scholar 

  23. Liu P, Qiu X, Huang X (2016) Recurrent neural network for text classification with multi-task learning. In Proceedings of International Joint Conference on Artificial Intelligence, 2873–2879. AAAI Press

  24. Mikolov T, Sutskever I, Chen K et al (2013) Distributed representations of words and phrases and their compositionality[J]. Adv Neural Inf Proces Syst 26:3111–3119

    Google Scholar 

  25. Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training. URL https://s3-us-west-2.amazonaws. com/openai-assets/researchcovers/languageunsupervised/language understanding paper. pdf

  26. Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks[J]. IEEE Trans Signal Process 45(11):2673–2681

    Article  Google Scholar 

  27. Sim Y, Routledge B, Smith NA (2016) Friends with motives: using text to infer influence on SCOTUS. In proceedings of the 2016 conference on empirical methods in natural language processing (pp. 1724-1733)

  28. Sun C, Qiu X, Xu Y, et al. (2019) How to Fine-Tune BERT for Text Classification?[J]. arXiv preprint arXiv: 1905.05583

  29. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Polosukhin I (2017) Attention is all you need. In Advances in neural information processing systems, 5998–6008

  30. Woon J (2009) Change we can believe in? Using political science to predict policy change in the Obama presidency[J]. PS: Political Science & Politics 42(2):329–333

    Google Scholar 

  31. Xiao S, Yan J, Farajtabar M, et al. (2019) Learning time series associated event sequences with recurrent point process networks[J]. IEEE Transactions on Neural Networks and Learning Systems, 1–13

  32. Yan J, Zhang C, Zha H, et al. (2015) On machine learning towards predictive sales pipeline analytics[C]//twenty-ninth AAAI conference on artificial intelligence

  33. Yan J, Xiao S, Li C, et al. (2016) Modeling Contagious Merger and Acquisition via Point Processes with a Profile Regression Prior[C]//IJCAI. 2690–2696.

  34. Yano T, Smith NA, Wilkerson JD (2012) Textual predictors of bill survival in congressional committees. In proceedings of the 2012 conference of the north American chapter of the Association for Computational Linguistics: human language technologies (pp. 793-802). Association for Computational Linguistics

  35. Zhou P, Shi W, Tian J, Qi Z, Li B, Hao H, Xu B (2016) Attention-based bidirectional long short-term memory networks for relation classification. In Proceedings of Annual Meeting of the Association for Computational Linguistics 207–212

  36. Zirn C, Meusel R, Stuckenschmidt H (2015) Lost in discussion? Tracking opinion groups in complex political discussions by the example of the fomc meeting transcriptions. In Proceedings of the International Conference Recent Advances in Natural Language Processing, 747–753.

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China under 61775175, 61771378 and 61601355, and by the Key Research and Development Program of Shaanxi Province - Key Industry Innovation Chain (Group) - Industrial Field under No.2019ZDLGY10-06.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jiaji Wu.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, G., Wu, J., Tan, M. et al. Learning to Predict U.S. Policy Change Using New York Times Corpus with Pre-Trained Language Model. Multimed Tools Appl 79, 34227–34240 (2020). https://doi.org/10.1007/s11042-020-08946-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-020-08946-y

Keywords

Navigation