ISCA Archive Interspeech 2019
ISCA Archive Interspeech 2019

Improving Keyword Spotting and Language Identification via Neural Architecture Search at Scale

Hanna Mazzawi, Xavi Gonzalvo, Aleks Kracun, Prashant Sridhar, Niranjan Subrahmanya, Ignacio Lopez Moreno, Hyun Jin Park, Patrick Violette

In this paper we present a novel Neural Architecture Search (NAS) framework to improve keyword spotting and spoken language identification models. Even with the huge success of deep neural networks (DNNs) in many different domains, finding the best network architecture is still a laborious task and very computationally expensive at best with existing searching approaches. Our search approach efficiently and robustly finds better model sequences with respect to hand-designed systems. We do this by constructing architectures incrementally, using a custom mutation algorithm and leveraging the power of parameter transfer between layers. We demonstrate that our approach can automatically design DNNs with an order of magnitude fewer parameters that achieves better performance than the current best models. It leads to significant performance improvements: up to 4.09% accuracy increase for language identification (6.1% if we allow an increase in the number of parameters) and 0.3% for phoneme classification in keyword spotting with half the size of the model.


doi: 10.21437/Interspeech.2019-1916

Cite as: Mazzawi, H., Gonzalvo, X., Kracun, A., Sridhar, P., Subrahmanya, N., Moreno, I.L., Park, H.J., Violette, P. (2019) Improving Keyword Spotting and Language Identification via Neural Architecture Search at Scale. Proc. Interspeech 2019, 1278-1282, doi: 10.21437/Interspeech.2019-1916

@inproceedings{mazzawi19_interspeech,
  author={Hanna Mazzawi and Xavi Gonzalvo and Aleks Kracun and Prashant Sridhar and Niranjan Subrahmanya and Ignacio Lopez Moreno and Hyun Jin Park and Patrick Violette},
  title={{Improving Keyword Spotting and Language Identification via Neural Architecture Search at Scale}},
  year=2019,
  booktitle={Proc. Interspeech 2019},
  pages={1278--1282},
  doi={10.21437/Interspeech.2019-1916}
}