Loading [MathJax]/extensions/MathMenu.js
Generative Data Augmentation via Wasserstein Autoencoder for Text Classification | IEEE Conference Publication | IEEE Xplore

Generative Data Augmentation via Wasserstein Autoencoder for Text Classification


Abstract:

Generative latent variable models are commonly used in text generation and augmentation. However generative latent variable models such as the variational autoencoder(VAE...Show More

Abstract:

Generative latent variable models are commonly used in text generation and augmentation. However generative latent variable models such as the variational autoencoder(VAE) experience a posterior collapse problem ignoring learning for a subset of latent variables during training. In particular, this phenomenon frequently occurs when the VAE is applied to natural language processing, which may degrade the reconstruction performance. In this paper, we propose a data augmentation method based on the pre-trained language model (PLM) using the Wasserstein autoencoder (WAE) structure. The WAE was used to prevent a posterior collapse in the generative model, and the PLM was placed in the encoder and decoder to improve the augmentation performance. We evaluated the proposed method on seven benchmark datasets and proved the augmentation effect.
Date of Conference: 19-21 October 2022
Date Added to IEEE Xplore: 25 November 2022
ISBN Information:

ISSN Information:

Conference Location: Jeju Island, Korea, Republic of

Contact IEEE to Subscribe

References

References is not available for this document.