CutFreq: Cut-and-Swap Frequency Components for Low-Level Vision Augmentation

Authors

  • Hongyang Chen Xi’an Jiaotong University
  • Kaisheng Ma Tsinghua University

DOI:

https://doi.org/10.1609/aaai.v38i2.27868

Keywords:

CV: Low Level & Physics-based Vision, CV: Computational Photography, Image & Video Synthesis, CV: Representation Learning for Vision

Abstract

Low-level vision plays a crucial role in a wide range of imaging quality and image recognition applications. However, the limited size, quality, and diversity of datasets often pose significant challenges for low-level tasks. Data augmentation is the most effective and practical way of sample expansion, but the commonly used augmentation methods in high-level tasks have limited improvement in the low-level due to the boundary effects or the non-realistic context information. In this paper, we propose the Cut-and-Swap Frequency Components (CutFreq) method for low-level vision, which aims to preserve high-level representations with directionality and improve image synthesis quality. Observing the significant frequency domain differences between reconstructed images and real ones, in CutFreq, we propose to transform the input and real images separately in the frequency domain, then define two stages for the model training process, and finally swap the specified frequency bands respectively and inversely transform to generate augmented samples. The experimental results show the superior performance of CutFreq on five low-level vision tasks. Moreover, we demonstrate the effectiveness of CutFreq in the low-data regime. Code is available at https://github.com/DreamerCCC/CutFreq.

Published

2024-03-24

How to Cite

Chen, H., & Ma, K. (2024). CutFreq: Cut-and-Swap Frequency Components for Low-Level Vision Augmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(2), 1072-1080. https://doi.org/10.1609/aaai.v38i2.27868

Issue

Section

AAAI Technical Track on Computer Vision I