Abstract
Zero-shot learning aims to transfer knowledge from the seen classes to unseen ones through some high-level semantics (e.g., per-class attributes), enabling the learning model to recognize novel classes without retraining. Among them, the generative methods adopt the scheme that synthesizes samples for the unseen classes, thereby converting the task into a standard classification problem. However, most existing work inevitably suffers from the domain shift problem when only the seen classes are used for supervision. Furthermore, they can not fully leverage the semantic information in data synthesis due to the limited expressiveness of the generator. In this paper, we develop a novel network, named stack-VAE, to alleviate the above problems. The proposal mainly consists of a generative module and a feature core agent. Specifically, we design the generator based on hierarchical VAE, which exploits multi-layer Gaussian distribution to improve the expressiveness, thereby better mimicking the real data distribution of the unseen classes. Besides, we propose a feature core agent based objective, which is beneficial to mitigate seen class bias by enforcing the inter-class separability and reducing the intra-class scatter. Experimental results conducted on three widely used datasets, i.e., AWA2, SUN, CUB, show that the proposed network outperforms the baselines and achieves a new state-of-the-art.
This work was supported in part by the National Natural Science Foundation of China under Grant 62172109 and Grant 62072118, in part by the Natural Science Foundation of Guangdong Province under Grant 2020A1515011361, and in part by the High-level Talents Programme of Guangdong Province under Grant 2017GC010556.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Chao, W.-L., Changpinyo, S., Gong, B., Sha, F.: An empirical study and analysis of generalized zero-shot learning for object recognition in the wild. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 52–68. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_4
Felix, R., Kumar, B.G.V., Reid, I.D., Carneiro, G.: Multi-modal cycle-consistent generalized zero-shot learning. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 21–37 (2018)
Frome, A., et al.: DeViSE: a deep visual-semantic embedding model. In: Advances in Neural Information Processing Systems 26, vol. 26, pp. 2121–2129 (2013)
Fu, Y., Hospedales, T.M., Xiang, T., Gong, S.: Transductive multi-view zero-shot learning. IEEE Trans. Pattern Anal. Mach. Intell. 37(11), 2332–2345 (2015)
Gao, R., Hou, X., Qin, J., Liu, L., Zhu, F., Zhang, Z.: A joint generative model for zero-shot learning. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 631–646 (2018)
Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems 27, vol. 27, pp. 2672–2680 (2014)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778 (2016)
Huang, Y., Deng, Z., Wu, T.: Learning discriminative latent features for generalized zero-and few-shot learning. In: 2020 IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6 (2020)
Keshari, R., Singh, R., Vatsa, M.: Generalized zero-shot learning via over-complete distribution. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13300–13308 (2020)
Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. In: ICLR 2014: International Conference on Learning Representations (ICLR) 2014 (2014)
Lampert, C.H., Nickisch, H., Harmeling, S.: Attribute-based classification for zero-shot visual object categorization. IEEE Trans. Pattern Anal. Mach. Intell. 36(3), 453–465 (2014)
Larochelle, H., Erhan, D., Bengio, Y.: Zero-data learning of new tasks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 2, pp. 646–651 (2008)
Li, J., Jing, M., Lu, K., Ding, Z., Zhu, L., Huang, Z.: Leveraging the invariant side of generative zero-shot learning. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7402–7411 (2019)
Li, J., Jing, M., Lu, K., Zhu, L., Yang, Y., Huang, Z.: Alleviating feature confusion for generative zero-shot learning. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 1587–1595 (2019)
Louizos, C., Swersky, K., Li, Y., Welling, M., Zemel, R.: The variational fair autoencoder. In: ICLR 2016: International Conference on Learning Representations 2016 (2016)
Ma, P., Hu, X.: A variational autoencoder with deep embedding model for generalized zero-shot learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 11733–11740 (2020)
van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(86), 2579–2605 (2008)
Patterson, G., Hays, J.: Sun attribute database: discovering, annotating, and recognizing scene attributes. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2751–2758 (2012)
Qin, P., Wang, X., Chen, W., Zhang, C., Xu, W., Wang, W.Y.: Generative adversarial zero-shot relational learning for knowledge graphs. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 8673–8680 (2020)
Rezende, D.J., Mohamed, S., Wierstra, D.: Stochastic backpropagation and approximate inference in deep generative models. In: Proceedings of the 31st International Conference on Machine Learning, pp. 1278–1286 (2014)
Schonfeld, E., Ebrahimi, S., Sinha, S., Darrell, T., Akata, Z.: Generalized zero- and few-shot learning via aligned variational autoencoders. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8247–8255 (2019)
Shigeto, Y., Suzuki, I., Hara, K., Shimbo, M., Matsumoto, Y.: Ridge regression, hubness, and zero-shot learning. In: Proceedings of the 2015th European Conference on Machine Learning and Knowledge Discovery in Databases - Volume Part I, pp. 135–151 (2015)
Sønderby, C.K., Raiko, T., Maaløe, L., Sønderby, S.K., Winther, O.: Ladder variational autoencoders. In: 29th Annual Conference on Neural Information Processing Systems (NIPS 2016), vol. 29, pp. 3738–3746 (2016)
Sun, Y., et al.: Circle loss: a unified perspective of pair similarity optimization. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6398–6407 (2020)
Vahdat, A., Kautz, J.: NVAE: a deep hierarchical variational autoencoder. In: Advances in Neural Information Processing Systems, vol. 33 (2020)
Verma, V.K., Arora, G., Mishra, A., Rai, P.: Generalized zero-shot learning via synthesized examples. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4281–4289 (2018)
Vyas, M.R., Venkateswara, H., Panchanathan, S.: Leveraging seen and unseen semantic relationships for generative zero-shot learning. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12375, pp. 70–86. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58577-8_5
Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The Caltech-UCSD birds-200-2011 dataset. California Institute of Technology (2011)
Wang, F., Xiang, X., Cheng, J., Yuille, A.L.: NormFace: L 2 hypersphere embedding for face verification. In: Proceedings of the 25th ACM International Conference on Multimedia, pp. 1041–1049 (2017)
Wei, K., Yang, M., Wang, H., Deng, C., Liu, X.: Adversarial fine-grained composition learning for unseen attribute-object recognition. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 3741–3749 (2019)
Xian, Y., Lampert, C.H., Schiele, B., Akata, Z.: Zero-shot learning-a comprehensive evaluation of the good, the bad and the ugly. IEEE Trans. Pattern Anal. Mach. Intell. 41(9), 2251–2265 (2019)
Xian, Y., Lorenz, T., Schiele, B., Akata, Z.: Feature generating networks for zero-shot learning. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5542–5551 (2018)
Zhang, Z., Saligrama, V.: Zero-shot learning via semantic similarity embedding. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 4166–4174 (2015)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Xie, J., Wu, J., Liang, T., Meng, M. (2021). Stack-VAE Network for Zero-Shot Learning. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds) Neural Information Processing. ICONIP 2021. Lecture Notes in Computer Science(), vol 13111. Springer, Cham. https://doi.org/10.1007/978-3-030-92273-3_21
Download citation
DOI: https://doi.org/10.1007/978-3-030-92273-3_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-92272-6
Online ISBN: 978-3-030-92273-3
eBook Packages: Computer ScienceComputer Science (R0)