Deep neural networks trained via empirical risk minimisation often exhibit significant performance disparities across groups, particularly when group and task labels are spuriously correlated (e.g., "grassy background" and "cows"). In this talk, I will first argue that previously proposed bias mitigation methods that aim to address this issue often either rely on group labels for training or validation, or require an extensive hyperparameter search, even when this assumption is not explicitly made. Such data and computational requirements hinder the practical deployment of these methods, especially when datasets are too large to be group-annotated, computational resources are limited, and models are trained through already complex pipelines. With this in mind, I will outline some of the challenges that may need to be addressed to design practical bias mitigation methods. Then, I will describe Targeted Augmentations for Bias mitigation (TAB), a new approach that considers these design principles. I will conclude by showing how TAB, a simple hyperparameter-free framework that leverages the entire training history of a helper model to identify spurious samples, improves worst-group performance without any group information or model selection.
"You can also join us on Zoom":https://cam-ac-uk.zoom.us/j/83400335522?pwd=LkjYvMOvVpMbabOV1MVTm8QU6DrGN7.1