Abstract
With the process of economic globalization and political multi-polarization accelerating, it is especially important to predict policy change in the United States. While current research has not taken advantage of the rapid advancement in the natural language processing and the relationship between news media and policy change, we propose a BERT-based model to predict policy change in the United States, using news published by the New York Times. Specifically, we propose a large-scale news corpus from the New York Times covers the period from 2006 to 2018. Then we use the corpus to fine-tune the pre-trained BERT language model to determine whether the news is on the front page, which corresponds to the policy priority. We propose a BERT-based Policy Change Index (BPCI) for the United States to predict the policy change in the future short period of time. Experimental results in the New York Times Corpus demonstrate the validity of the proposed method.
Similar content being viewed by others
References
Ansolabehere S, Iyengar S (1994) Riding the wave and claiming ownership over issues: the joint effects of advertising and news coverage in campaigns[J]. Public Opinion Quarterly 58(3):335–357
Bengio Y, Ducharme R, Vincent P, Jauvin C (2003) A neural probabilistic language model. J Mach Learn Res, 1137–1155
Carlstrom CT, Zaman S (2014) Using an improved Taylor rule to predict when policy changes will occur[J]. Economic Commentary, (2014–02).
Chan JTK, Zhong W (2018) Reading China: predicting policy change with machine learning[J]. AEI Economics Working Paper
DellaVigna S, Kaplan E (2007) The fox news effect: media bias and voting[J]. Q J Econ 122(3):1187–1234
Devlin J, Chang MW, Lee K et al. (2018) Bert: Pre-training of deep bidirectional transformers for language understanding[J]. arXiv preprint arXiv:1810.04805
Enikolopov R, Petrova M, Zhuravskaya E (2011) Media and political persuasion: evidence from Russia[J]. Am Econ Rev 101(7):3253–3285
Gerber AS, Karlan D, Bergan D (2009) Does the media matter? A field experiment measuring the effect of newspapers on voting behavior and political opinions[J]. Am Econ J Appl Econ 1(2):35–52
Girshick R (2015) Fast r-cnn[C]//Proceedings of the IEEE international conference on computer vision. 1440–1448
Graff D (1995) North American News Text Corpus[C]// Linguistic Data Consortium. LDC95T21
Graves A, Mohamed A, Hinton G (2013) Speech recognition with deep recurrent neural networks[C]//2013 IEEE international conference on acoustics, speech and signal processing. IEEE, 6645–6649
Grossman M (2005) Role theory and foreign policy change: the transformation of Russian foreign policy in the 1990s[J]. International Politics 42(3):334–351
Hearings. The Policy Agendas Project at the University of Texas at Austin, 2017. www.comparativeagendas.net. Accessed September 26, 2017.
Hochreiter S, Schmidhuber J (1997) Long short-term memory.[J]. Neural Comput 9(8):1735–1780
Joulin A, Grave E, Bojanowski P, Mikolov T (2017) Bag of Tricks for Efficient Text Classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics. 427–431.
Katz DM, Bommarito II, Michael J et al. (2014) Predicting the behavior of the supreme court of the united states: A general approach[J]. arXiv preprint arXiv:1407.6333
Katz DM, Bommarito MJ II, Blackman J (2017) A general approach for predicting the behavior of the supreme court of the United States. PLoS One 12(4):e0174698
Kim Y (2014) Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). 1746-1751
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks[C]//Advances in neural information processing systems. 1097–1105
Lai S, Xu L, Liu K, Zhao J (2015) Recurrent convolutional neural networks for text classification. In AAAI conference on artificial intelligence
Le Q, Mikolov T (2014) Distributed representations of sentences and documents[C]//International conference on machine learning. 1188–1196
LeCun Y, Bengio Y, Hinton G (2015) Deep learning[J]. Nature 521(7553):436–444
Liu P, Qiu X, Huang X (2016) Recurrent neural network for text classification with multi-task learning. In Proceedings of International Joint Conference on Artificial Intelligence, 2873–2879. AAAI Press
Mikolov T, Sutskever I, Chen K et al (2013) Distributed representations of words and phrases and their compositionality[J]. Adv Neural Inf Proces Syst 26:3111–3119
Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training. URL https://s3-us-west-2.amazonaws. com/openai-assets/researchcovers/languageunsupervised/language understanding paper. pdf
Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks[J]. IEEE Trans Signal Process 45(11):2673–2681
Sim Y, Routledge B, Smith NA (2016) Friends with motives: using text to infer influence on SCOTUS. In proceedings of the 2016 conference on empirical methods in natural language processing (pp. 1724-1733)
Sun C, Qiu X, Xu Y, et al. (2019) How to Fine-Tune BERT for Text Classification?[J]. arXiv preprint arXiv: 1905.05583
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Polosukhin I (2017) Attention is all you need. In Advances in neural information processing systems, 5998–6008
Woon J (2009) Change we can believe in? Using political science to predict policy change in the Obama presidency[J]. PS: Political Science & Politics 42(2):329–333
Xiao S, Yan J, Farajtabar M, et al. (2019) Learning time series associated event sequences with recurrent point process networks[J]. IEEE Transactions on Neural Networks and Learning Systems, 1–13
Yan J, Zhang C, Zha H, et al. (2015) On machine learning towards predictive sales pipeline analytics[C]//twenty-ninth AAAI conference on artificial intelligence
Yan J, Xiao S, Li C, et al. (2016) Modeling Contagious Merger and Acquisition via Point Processes with a Profile Regression Prior[C]//IJCAI. 2690–2696.
Yano T, Smith NA, Wilkerson JD (2012) Textual predictors of bill survival in congressional committees. In proceedings of the 2012 conference of the north American chapter of the Association for Computational Linguistics: human language technologies (pp. 793-802). Association for Computational Linguistics
Zhou P, Shi W, Tian J, Qi Z, Li B, Hao H, Xu B (2016) Attention-based bidirectional long short-term memory networks for relation classification. In Proceedings of Annual Meeting of the Association for Computational Linguistics 207–212
Zirn C, Meusel R, Stuckenschmidt H (2015) Lost in discussion? Tracking opinion groups in complex political discussions by the example of the fomc meeting transcriptions. In Proceedings of the International Conference Recent Advances in Natural Language Processing, 747–753.
Acknowledgements
This work was supported by the National Natural Science Foundation of China under 61775175, 61771378 and 61601355, and by the Key Research and Development Program of Shaanxi Province - Key Industry Innovation Chain (Group) - Industrial Field under No.2019ZDLGY10-06.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhang, G., Wu, J., Tan, M. et al. Learning to Predict U.S. Policy Change Using New York Times Corpus with Pre-Trained Language Model. Multimed Tools Appl 79, 34227–34240 (2020). https://doi.org/10.1007/s11042-020-08946-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-020-08946-y