Abstract
This paper concerns an application of recurrent neural networks to text synthesis in the word level, with the help of keywords. First, a Parts Of Speech tagging library is employed to extract verbs and nouns from the texts used in our work, a part of which are then considered, after automatic eliminations, as the aforementioned keywords. Our ultimate aim is to train a recurrent neural network to map the keyword sequence of a text to the entire text. Successive reformulations of the keyword and full-text word sequences are performed, so that they can serve as the input and target of the network as efficiently as possible. The predicted texts are understandable enough, and the model performance depends on the problem difficulty, determined by the percentage of full-text words that are considered as keywords, that ranges from 1/3 to 1/2 approximately, the training memory cost, mainly affected by the network architecture, as well as the similarity between different texts, which determines the best architecture.
Similar content being viewed by others
Notes
It cannot be characterized as test set just because early stopping is one of the training termination criteria, and is applied after 1 min of no update in the minimum error.
It is not offered with lower size.
References
Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT Press, Cambridge
Goldberg Y (2016) A primer on neural network models for natural language processing. J Artif Intell Res 57:345–420
Sutskever I, Martens J, Hinton J (2011) Generating text with recurrent neural networks. In: Proceedings of the 28th international conference on machine learning—Bellevue, WA, USA, pp 1017-1024
Kolokas N, Drosou A, Tzovaras D (2018) Keywords-to-text synthesis using recurrent neural network. In: Iliadis L, Maglogiannis I, Plagianakos V (eds) Artificial intelligence applications and innovations—AIAI 2018—IFIP advances in information and communication technology. Springer, Cham. vol 519, pp 85–96
Shah P, Perez-Iratxeta C, Andrade M (2003) Information extraction from full text scientific articles: where are the keywords? BMC Bioinf BioMed Central 4:20
Andrade MA, Valencia A (1998) Automatic extraction of keywords from scientific text: application to the knowledge domain of protein families. Bioinformatics 14(7):600–607
Serra PMUX (2010) A look into the past: analysis of trends and topics in proceedings of sound and music computing conference. In: Sound and music computing conference
HaCohen-Kerner Y (2003) Automatic extraction of keywords from abstracts. In: International conference on knowledge-based and intelligent information and engineering systems, Springer, Berlin. vol 2773, pp 843–849
Litvak M, Last M (2008) Graph-based keyword extraction for single-document summarization. In: Proceedings of the workshop on multi-source multilingual information extraction and summarization
Al-Hashemi R (2010) Text summarization extraction system (TSES) using extracted keywords. Int Arab J Technol 1(4):164–168
Sneiders E (1999) Automated FAQ answering: continued experience with shallow language understanding. In: Association for the advancement of artificial intelligence, pp 97–107
Smit B (2002) Atlas. ti for quality in qualitative research: a CAQDAS project. Perspect Educ 20(3):65–76
Al Kadri HMF, Al Moamary MS, van der Vleuten C (2009) Students’ and teachers’ perceptions of clinical assessment program: a qualitative study in a PBL curriculum. BMC Res Notes 2:263
Kasper R (1989) A flexible interface for linking applications to Penman’s sentence generator. In: Proceeding HLT ’89—Proceedings of the workshop on speech and natural language, pp 153–158
Bateman JA (1997) Sentence generation and systemic grammar: an introduction. Iwanami lecture series: language sciences, p 8
Feiner S, McKeown K (1991) Automating the generation of coordinated multimedia explanations. Computer 24(10):33–41
Bernauer J, Gumrich K, Kutz S, Lindner P, Pretschner DP (1991) An interactive report generator for bone scan studies. In: Proceedings of the annual symposium on computer application in medical care, pp 858–860
Carbonell J (1970) AI in CAI: an artificial-intelligence approach to computer-assisted instruction. IEEE Trans Man Mach Syst 11(4):190–202
Levy S (2016) The brain is here—and it’s already inside your phone, https://www.wired.com/2016/08/an-exclusive-look-at-how-ai-and-machine-learning-work-at-apple/
Kahn J (2017) Apple engineers share behind-the-scenes evolution of Siri & more on Apple Machine Learning Journal, https://9to5mac.com/2017/08/23/evolution-siri-machine-learning-journal/
Vogels W (2016) Bringing the Magic of Amazon AI and Alexa to Apps on AWS, https://www.allthingsdistributed.com/2016/11/amazon-ai-and-alexa-for-all-aws-apps.html
Brownlee J (2017) How to develop an encoder-decoder model for sequence-to-sequence prediction in Keras, https://machinelearningmastery.com/develop-encoder-decoder-model-sequence-sequence-prediction-keras/
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: ICLR
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. In: NIPS
Jean S, Cho K, Memisevic R, Bengio Y (2014) On using very large target vocabulary for neural machine translation, https://arxiv.org/pdf/1412.2007v2.pdf
Kalchbrenner N, Blunsom P (2013) Recurrent continuous translation models. In: EMNLP
Zhang B, Xiong D, Su J, Duan H (2017) A context-aware recurrent encoder for neural machine translation . IEEE/ACM Trans Audio Speech Lang Process 25(12):2424–2432
Su J, Zeng J, Xiong D, Liu Y, Wang M, Xie J (2018) A hierarchy-to-sequence attentional neural machine translation model. IEEE/ACM Trans Audio Speech Lang Process 26(3):623–632
Li Q, Wong DF, Chao LS, Zhu M, Xiao T (2018) Linguistic knowledge-aware neural machine translation. IEEE/ACM Trans Audio Speech Lang Process 26(12):2341–2354
Zhang H, Li J, Ji Y, Yue H (2017) Understanding subtitles by character-level sequence-to-sequence learning. IEEE Trans Ind Inf 13(2):616–624
Dethlefs N (2017) Domain transfer for deep natural language generation from abstract meaning representations. IEEE Comput Intell Mag 12(3):18–28
Weston J, Chopra S, Bordes A (2014) Memory networks. In: ICLR
Kumar A, Irsoy O, Su J, Bradbury J, English R, Pierce B, Ondruska P, Iyyer M, Gulrajani I, Socher R (2015) Ask me anything: dynamic memory networks for natural language processing. In: International conference on machine learning
Zaman MMA, Mishu SZ (2017) Convolutional recurrent neural network for question answering. In: 3rd International conference on electrical information and communication technology (EICT)
Tan C, Wei F, Zhou Q, Yang N, Du B, Lv W, Zhou M (2018) Context-aware answer sentence selection with hierarchical gated recurrent neural networks. IEEE/ACM Trans Audio Speech Lang Process 26(3):540–549
Wu F, Duan X, Xiao J, Zhao Z, Tang S, Zhang Y, Zhuang Y (2017) Temporal interaction and causal influence in community-based question answering. IEEE Trans Knowl Data Eng 29(10):2304–2317
Graves A, Mohamed A, Hinton G (2013) Speech recognition with deep recurrent neural networks. In: ICASSP, pp 6645–6649
Pascanu R, Gülçehre Ç, Cho K, Bengio Y (2014) How to construct deep recurrent neural networks. In: ICLR
Chung J, Gulcehre C, Cho, K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. NIPS deep learning workshop
Del-Agua MÁ, Giménez A, Sanchis A, Civera J, Juan A (2018) Speaker-adapted confidence measures for ASR using deep bidirectional recurrent neural networks. IEEE/ACM Trans Audio Speech Lang Process 26(7):1198–1206
Enarvi S, Smit P, Virpioja S, Kurimo M (2017) Automatic speech recognition with very large conversational finnish and estonian vocabularies. IEEE/ACM Trans Audio Speech Lang Process 25(11):2085–2097
Kim M, Cao B, Mau T, Wang J (2017) Speaker-independent silent speech recognition from flesh-point articulatory movements using an LSTM neural network. IEEE/ACM Trans Audio Speech Lang Process 25(12):2323–2336
Ruan Y-P, Chen Q, Ling Z-H (2017) A sequential neural encoder with latent structured description for modeling sentences. IEEE/ACM Trans Audio Speech Lang Process 26(2):231–242
Baktha K, Tripathy BK (2017) Investigation of recurrent neural networks in the field of sentiment analysis. In: International conference on communication and signal processing (ICCSP)
Hassan A, Mahmood A (2018) Convolutional recurrent deep learning model for sentence classification. IEEE Access 6:13949–13957
Wu D, Chi M (2017) Long short-term memory with quadratic connections in recursive neural networks for representing compositional semantics. IEEE Access 5:16077–16083
Güngör O, Üsküdarli S, Güngör T (2018) Recurrent neural networks for Turkish named entity recognition. In: 26th signal processing and communications applications conference (SIU)
Zhang X, Li X, An J, Gao L, Hou B, Li C (2017) Natural language description of remote sensing images based on deep learning. In: IEEE international geoscience and remote sensing symposium (IGARSS)
Park CC, Kim Y, Kim G (2018) Retrieval of sentence sequences for an image stream via coherence recurrent convolutional networks. IEEE Trans Pattern Anal Mach Intell 40(4):945–957
Karpathy A, Fei-Fei L (2017) Deep visual-semantic alignments for generating image descriptions. IEEE Trans Pattern Anal Mach Intell 39(4):664–676
Li L, Tang S, Zhang Y, Deng L, Tian Q (2018) GLA: global-local attention for image description. IEEE Trans Multimedia 20(3):726–737
Yang L, Hu H (2017) TVPRNN for image caption generation. Electron Lett 53(22):1471–1473
Vinyals O, Toshev A, Bengio S, Erhan D (2017) Show and tell: lessons learned from the 2015 MSCOCO image captioning challenge. IEEE Trans Pattern Anal Mach Intell 39(4):652–663
Cascianelli S, Costante G, Ciarfuglia TA, Valigi P, Fravolini ML (2018) Full-GRU natural language video description for service robotics applications. IEEE Robot Autom Lett 3(2):841–848
Peng Y, Qi J, Yuan Y (2018) Modality-specific cross-modal similarity measurement with recurrent attention network. IEEE Trans Image Process 27(11):5585–5599
Zheng H-T, Wang W, Chen W, Sangaiah AK (2017) Automatic generation of news comments based on gated attention neural networks. IEEE Access 6:702–710
Zhang R, Meng F, Zhou Y, Liu B (2018) Relation classification via recurrent neural network with attention and tensor layers. Big Data Min Anal 1(3):234–244
Tang Z, Wang D, Chen Y, Li L, Abel A (2018) Phonetic temporal neural model for language identification. IEEE/ACM Trans Audio Speech Lang Process 26(1):134–144
Zheng K, Yan WQ, Nand P (2018) Video dynamics detection using deep neural networks. IEEE Trans Emerg Top Comput Intell 2(3):224–234
Donahue J, Hendricks LA, Rohrbach M, Venugopalan S, Guadarrama S, Saenko K, Darrell T (2017) Long-term recurrent convolutional networks for visual recognition and description. IEEE Trans Pattern Anal Mach Intell 39(4):677–691
Yang Y, Zhou J, Ai J, Bin Y, Hanjalic A, Shen HT, Ji Y (2018) Video captioning by adversarial LSTM. IEEE Trans Image Process 27(11):5600–5611
Liu Z-C, Ling Z-H, Dai L-R (2018) Statistical parametric speech synthesis using generalized distillation framework. IEEE Signal Process Lett 25(5):695–699
Gonzalez JA, Cheah LA, Gomez AM, Green PD, Gilbert JM, Ell SR, Moore RK, Holdsworth E (2017) Direct speech reconstruction from articulatory sensor data by machine learning. IEEE/ACM Trans Audio Speech Lang Process 25(12):2362–2374
Feng X, Huang L, Qin B, Lin Y, Ji H, Liu T (2017) Multi-level cross-lingual attentive neural architecture for low resource name tagging. Tsinghua Sci Technol 22(6):633–645
Abroyan N (2017) Convolutional and recurrent neural networks for real-time data classification. In: Seventh international conference on innovative computing technology (INTECH)
Ling Z-H, Ai Y, Gu Y, Dai L-R (2018) Waveform modeling and generation using hierarchical recurrent neural networks for speech bandwidth extension. IEEE/ACM Trans Audio Speech Lang Process 26(5):883–894
Wang W, Sheng Y, Wang J, Zeng X, Ye X, Huang Y, Zhu M (2017) HAST-IDS: learning hierarchical spatial-temporal features using deep neural networks to improve intrusion detection. IEEE Access 6:1792–1806
Naseer S, Saleem Y, Khalid S, Bashir MK, Han J, Iqbal MM, Han K (2018) Enhanced network anomaly detection based on deep neural networks. IEEE Access 6:48231–48246
Bengio Y, Ducharme R, Vincent P (2001) A neural probabilistic language model. In: Leen TK, Dietterich TG, Tresp V (eds) NIPS’2000, MIT Press, Cambridge, pp 932–938
Wang Zhish, Lin J, Wang Zhongf (2018) Hardware-oriented compression of long short-term memory for efficient inference. IEEE Signal Process Lett 25(7):984–988
Chen C, Ding H, Peng H, Zhu H, Wang Y, Shi C-JR (2018) OCEAN: an on-chip incremental-learning enhanced artificial neural network processor with multiple gated-recurrent-unit accelerators. IEEE J Emerg Sel Top Circuits Syst 8(3):519–530
Goldberg Y, Hirst G (2017) Neural network methods for natural language processing. Morgan & Claypool, San Rafael
allrecipes - Beef Steak Recipes, http://allrecipes.com/recipes/475/meat-and-poultry/beef/steaks/
Natural Language Toolkit (NLTK), http://www.nltk.org
Brownlee J (2017) How to develop word embeddings in python with Gensim, https://machinelearningmastery.com/develop-word-embeddings-python-gensim/
Atabay D., pyrenn: A recurrent neural network toolbox for Python and Matlab, Institute for Energy Economy and Application Technology - Technische Universität, München, http://pyrenn.readthedocs.io/en/latest/
Keras Documentation - Models - Sequential, https://keras.io/models/sequential/
scipy.optimize.basinhopping, https://docs.scipy.org/doc/scipy-0.18.1/reference/generated/scipy.optimize.basinhopping.html
Acknowledgements
This work has been partially supported by the European Commission through project Scan4Reco funded by the European Union H2020 programme under Grant Agreement No. 665091. The opinions expressed in this paper are those of the authors and do not necessarily reflect the views of the European Commission.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Kolokas, N., Drosou, A. & Tzovaras, D. Text synthesis from keywords: a comparison of recurrent-neural-network-based architectures and hybrid approaches. Neural Comput & Applic 32, 4259–4274 (2020). https://doi.org/10.1007/s00521-019-04435-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-019-04435-y