skip to main content
10.1145/3474085.3478873acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
abstract

Few-shot Learning for Multi-Modality Tasks

Authors Info & Claims
Published:17 October 2021Publication History

ABSTRACT

Recent deep learning methods rely on a large amount of labeled data to achieve high performance. These methods may be impractical in some scenarios, where manual data annotation is costly or the samples of certain categories are scarce (e.g., tumor lesions, endangered animals and rare individual activities). When only limited annotated samples are available, these methods usually suffer from the overfitting problem severely, which degrades the performance significantly. In contrast, humans can recognize the objects in the images rapidly and correctly with their prior knowledge after exposed to only a few annotated samples. To simulate the learning schema of humans and relieve the reliance on the large-scale annotation benchmarks, researchers start shifting towards the few-shot learning problem: they try to learn a model to correctly recognize novel categories with only a few annotated samples.

References

  1. Flood Sung, Yongxin Yang, Li Zhang, Tao Xiang, Philip H.S. Torr, Timothy M. Hospedales, Learning to Compare: Relation Network for Few-Shot Learning, CVPR 2018Google ScholarGoogle Scholar
  2. Yikai Wang, Li Zhang, Yuan Yao, Yanwei Fu, How to trust unlabeled data? Instance Credibility Inference for Few-Shot Learning, IEEE TPAMI 2021Google ScholarGoogle Scholar
  3. Chen Liu, Yanwei Fu, Chengming Xu, Siqian Yang, Jilin Li, Chengjie Wang, Li Zhang, Learning a Few-shot Embedding Model with Contrastive Learning, AAAI, 2021Google ScholarGoogle Scholar
  4. Yikai Wang, Chengming Xu, Chen Liu, Li Zhang, Yanwei Fu, Instance Credibility Inference for Few-Shot Learning, CVPR, 2020Google ScholarGoogle ScholarCross RefCross Ref
  5. Hongguang Zhang, Li Zhang, Xiaojuan Qi, Hongdong Li, Philip H.S. Torr, Piotr Koniusz, Few-shot Action Recognition with Permutation-invariant Attention, ECCV, 2020Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Yuqian Fu, Li Zhang, Junke Wang, Yanwei Fu, Yu-Gang Jiang, Depth Guided Adaptive Meta-Fusion Network for Few-shot Video Recognition, ACM MM 2020 Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. S. Kevin Zhou, H. Greenspan, C. Davatzikos, J.S. Duncan, B. van Ginneken, A. Madabhushi, J.L. Prince, D. Rueckert, and R.M. Summers, ?A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises," Proceedings of the IEEE, 2021Google ScholarGoogle Scholar
  8. Q. Yao, L. Xiao, P. Liu, and S. Kevin Zhou, ?Label-free segmentation of COVID-19 lesions in lung CT," IEEE Trans. on Medical Imaging, 2021.Google ScholarGoogle Scholar
  9. G. Shi, L. Xiao, Y. Chen, and S. Kevin Zhou, "Marginal loss and exclusion loss for partially supervised multi-organ segmentation," Medical Image Analysis, 2021.Google ScholarGoogle Scholar
  10. H. Li, H. Han, Z. Li, L. Wang, Z. Wu, J. Lu, and S. Kevin Zhou, "High-resolution chest X-ray bone suppression using unpaired CT structural priors," IEEE Trans. on Medical Imaging, 2020.Google ScholarGoogle Scholar
  11. C. Huang, H. Han, Q. Yao, S. Zhu, and S. Kevin Zhou, "3D U2Net: A 3D universal u-net for multi-domain medical image segmentation," MICCAI, 2019.Google ScholarGoogle Scholar
  12. B. Li, B. Yang, C. Liu, F. Liu, R. Ji, Q. Ye, "Beyond Max-Margin: Class Margin Equilibrium for Few-shot Object Detection," CVPR 2021.Google ScholarGoogle Scholar
  13. B. Liu, Y. Ding, J. Jiao, X. Ji, Q. Ye, Anti-aliasing Semantic Reconstruction for Few-Shot Semantic Segmentation, CVPR 2021.Google ScholarGoogle Scholar
  14. B. Yang, C Liu, B. Li, J. Jiao, Q. Ye, "Prototype Mixture Models for Few-shot Semantic Segmentation," ECCV 2020.Google ScholarGoogle Scholar
  15. B. Liu, J. Jiao, Q. Ye, "Harmonic Feature Activation for Few-Shot Semantic Segmentation," IEEE Trans. Image Process, 30(2):3142 - 3153, 2021Google ScholarGoogle ScholarCross RefCross Ref
  16. Xiaoyu Tao, Xiaopeng Hong, Xinyuan Chang, Songlin Dong, Xing Wei, Yihong Gong. Few-Shot Class-Incremental Learning. CVPR, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  17. Songlin Dong, Xiaopeng Hong, Xiaoyu Tao, Xinyuan Chang, Xing Wei, Yinghong Gong. Few-Shot Class-Incremental Learning via Relation Knowledge Distillation. AAAI, 2021.Google ScholarGoogle Scholar
  18. Xiaoyu Tao, Xinyuan Chang, Xiaopeng Hong, Songlin Dong, Xing Wei, Yihong Gong. Topology-Preserving Class-Incremental Learning. ECCV, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  19. Xiaoyu Tao, Xiaopeng Hong, Xinyuan Chang, Yihong Gong. Bi-objective Continual Learning: Learning "New' while Consolidating "Known,' AAAI, 2020.Google ScholarGoogle Scholar
  20. Chaofan Chen, Xiaoshan Yang, Changsheng Xu, Xuhui Huang, Zhe Ma: ECKPN: Explicit Class Knowledge Propagation Network for Transductive Few-shot Learning, CVPR 2021Google ScholarGoogle Scholar
  21. Yi Huang, Xiaoshan Yang, Junyu Gao, Jitao Sang, Changsheng Xu, Knowledge-driven Egocentric Multimodal Activity Recognition, TOMM, 2020 Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Fan Qi, Xiaoshan Yang, Changsheng Xu, Emotion Knowledge Driven Video Highlight Detection, IEEE TMM, 2020Google ScholarGoogle Scholar
  23. Li Liu, Wanli Ouyang, Xiaogang Wang, Paul Fieguth, Jie Chen, Xinwang Liu, Matti Pietikäinen; Deep Learning for Generic Object Detection: A Survey, IJCV, 2019.Google ScholarGoogle Scholar
  24. Xiawu Zheng, Rongrong Ji_, Qiang Wang, Yuhang Chen, Baochang Zhang, Jie Chen, Qixiang Ye, Feiyue Huang, Yonghong Tian, MIGO-NAS: Towards Fast and Generalizable Neural Architecture Search, IEEE TPAMI, 2021.Google ScholarGoogle Scholar
  25. Hongliang He, Yao Ding, Guoli Song, Pengxu Wei, Jie Chen, DGN: Direction Guided Network for Nuclear Instance Segmentation, ICCV, 2021Google ScholarGoogle Scholar
  26. Mengjun Cheng, Zishang Kong, Jie Chen, Learnable Oriented-Derivative Network for Polyp Segmentation, MICCAI 2021Google ScholarGoogle Scholar

Index Terms

  1. Few-shot Learning for Multi-Modality Tasks

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      MM '21: Proceedings of the 29th ACM International Conference on Multimedia
      October 2021
      5796 pages
      ISBN:9781450386517
      DOI:10.1145/3474085

      Copyright © 2021 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 17 October 2021

      Check for updates

      Qualifiers

      • abstract

      Acceptance Rates

      Overall Acceptance Rate995of4,171submissions,24%

      Upcoming Conference

      MM '24
      MM '24: The 32nd ACM International Conference on Multimedia
      October 28 - November 1, 2024
      Melbourne , VIC , Australia
    • Article Metrics

      • Downloads (Last 12 months)60
      • Downloads (Last 6 weeks)5

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader