Loading [a11y]/accessibility-menu.js
Sampling-Based Pruned Knowledge Distillation for Training Lightweight RNN-T | IEEE Journals & Magazine | IEEE Xplore

Sampling-Based Pruned Knowledge Distillation for Training Lightweight RNN-T


Abstract:

We present a novel training method for small-scale RNN-T models, widely used in real-world speech recognition applications. Despite efforts to scale down models for edge ...Show More

Abstract:

We present a novel training method for small-scale RNN-T models, widely used in real-world speech recognition applications. Despite efforts to scale down models for edge devices, the demand for even smaller and more compact speech recognition models persists to accommodate a broader range of devices. In this letter, we propose Sampling-based Pruned Knowledge Distillation (SP-KD) for training lightweight RNN-T models. In contrast to the conventional knowledge distillation techniques, the proposed method enables student models to distill knowledge from the distribution of teacher models, which is estimated by considering not only the best paths but also less likely paths. Additionally, we leverage pruning the output lattice of RNN-T to comprehensively transfer knowledge from teacher models to student models. Experimental results demonstrate that our proposed method outperforms the baseline in training tiny RNN-T models.
Published in: IEEE Signal Processing Letters ( Volume: 32)
Page(s): 631 - 635
Date of Publication: 13 January 2025

ISSN Information:


Contact IEEE to Subscribe

References

References is not available for this document.