CUDA: Curriculum of Data Augmentation for Long-tailed Recognition
Presenter: Zheda Mai
ICLR 2023 Oral
Class Imbalanced Problem
Goal: Train the model so as it’s not being influenced by the class distribution
Previous Approaches
A key problem: Overfitting to minorities
Previous Approaches
The key problem
Detailed analysis about DA in class imbalanced problem is missing
Findings
Intuition: STRONG augmentation to minority , WEAK augmentation to majority
____
____
balanced test acc
Findings
Findings
What about in a balanced dataset?
Augmentation strength for class 50-99
Why? How does it happen?
[1]Decoupling representation and classifier for long-tailed recognition, ICLR 2020
[2]:{BOIL}: Towards representation change for few-shot learning, ICLR 2021
Tools we use to explain why
Why?
Balanced CIFAR-100
Imbalanced CIFAR-100
partial aug: aug 0-50 classes
class index sorted by sample #
Why?
How does partial aug reduce weight norm for aug classes?
Now, we can explain why
partial aug -> poor acc for aug classes
How does partial aug reduce weight norm for aug classes?
FS high
FS low
the classifier for this class can easily find a pattern
the classifier for this class can NOT easily find a pattern
variation of gradients for classifier is low
variation of gradients for classifier is high
Remaining Question
CUDA: Curriculum of Data Augmentation
Core idea:
A model should only learn harder samples when it has mastered most easy samples
2
# of correct predictions > threshold
3
3
# of correct predictions < threshold
1
1
CUDA: Curriculum of Data Augmentation
Advantages of CUDA
CUDA improves existing method
Takeaways
Findings
CUDA
Questions?