Training a linear classifier or lightweight model on top of pretrained vision model outputs, so-called 'frozen features', leads to impressive performance on a number of downstream few-shot tasks. Currently, frozen features are not modified during training. On the other hand, when networks are trained directly on images, data augmentation is a standard recipe that improves performance with no substantial overhead. In this paper, we conduct an extensive pilot study on few-shot image classification that explores applying data augmentations in the frozen feature space, dubbed 'frozen feature augmentation (FroFA)', covering twenty augmentations in total. Our study demonstrates that adopting a deceptively simple point-wise FroFAs, such as brightness, can improve few-shot performance consistently across three network architectures, three large pretraining datasets, and eight transfer datasets.
@InProceedings{Baer2024,
author = {Andreas B\"ar and Neil Houlsby and Mostafa Dehghani and Manoj Kumar},
booktitle = {Proc.\ of CVPR},
title = {{Frozen Feature Augmentation for Few-Shot Image Classification}},
month = jun,
year = {2024},
address = {Seattle, WA, USA},
pages = {16046--16057},
}