DFM-X: Augmentation by Leveraging Prior Knowledge of Shortcut Learning logo

DFM-X: Augmentation by Leveraging Prior Knowledge of Shortcut Learning

We propose a data augmentation strategy, named DFM-X, that leverages knowledge about frequency shortcuts, encoded in Dominant Frequencies Maps computed for image classification models.

GitHub Link

The GitHub link is https://github.com/nis-research/dfmx-augmentation

Introduce

The 'DFM-X Augmentation by Leveraging Prior Knowledge of Shortcut Learning' is a data augmentation strategy designed to improve the generalization and robustness of neural network models. It introduces DFM-X, a technique that uses Dominant Frequencies Maps (DFMs) to select training images for augmentation. By including frequencies from other classes' DFMs, the models are encouraged to learn deeper and task-related semantics, reducing shortcut learning. This method enhances model robustness against corruptions and adversarial attacks, and can be combined with other augmentation techniques. The code repository provides installation instructions and scripts for computing DFMs, evaluating model performance, and training with DFM-X. The strategy is presented in the context of the International Conference on Computer Vision Workshops (ICCVW) by Shunxin Wang, Christoph Brune, Raymond Veldhuis, and Nicola Strisciuglio. We propose a data augmentation strategy, named DFM-X, that leverages knowledge about frequency shortcuts, encoded in Dominant Frequencies Maps computed for image classification models.

Content

Neural networks are prone to learn easy solutions from superficial statistics in the data, namely shortcut learning, which impairs generalization and robustness of models. We propose a data augmentation strategy, named DFM-X, that leverages knowledge about frequency shortcuts, encoded in Dominant Frequencies Maps computed for image classification models. We randomly select X% training images of certain classes for augmentation, and process them by retaining the frequencies included in the DFMs of other classes. This strategy compels the models to leverage a broader range of frequencies for classification, rather than relying on specific frequency sets. Thus, the models learn more deep and task-related semantics compared to their counterpart trained with standard setups. Unlike other commonly used augmentation techniques which focus on increasing the visual variations of training data, our method targets exploiting the original data efficiently, by distilling prior knowledge about destructive learning behavior of models from data. Our experimental results demonstrate that DFM-X improves robustness against common corruptions and adversarial attacks. It can be seamlessly integrated with other augmentation techniques to further enhance the robustness of models.

Alternatives & Similar Tools

LongLLaMA-handle very long text contexts, up to 256,000 tokens logo

LongLLaMA is a large language model designed to handle very long text contexts, up to 256,000 tokens. It's based on OpenLLaMA and uses a technique called Focused Transformer (FoT) for training. The repository provides a smaller 3B version of LongLLaMA for free use. It can also be used as a replacement for LLaMA models with shorter contexts.