BREEDS: Benchmarks for Subpopulation Shift

Shibani Santurkar · Dimitris Tsipras · Aleksander Madry

Keywords: [ benchmarks ] [ distribution shift ] [ hierarchy ] [ robustness ]

[ Abstract ]
[ Paper ]
Thu 6 May 9 a.m. PDT — 11 a.m. PDT


We develop a methodology for assessing the robustness of models to subpopulation shift---specifically, their ability to generalize to novel data subpopulations that were not observed during training. Our approach leverages the class structure underlying existing datasets to control the data subpopulations that comprise the training and test distributions. This enables us to synthesize realistic distribution shifts whose sources can be precisely controlled and characterized, within existing large-scale datasets. Applying this methodology to the ImageNet dataset, we create a suite of subpopulation shift benchmarks of varying granularity. We then validate that the corresponding shifts are tractable by obtaining human baselines. Finally, we utilize these benchmarks to measure the sensitivity of standard model architectures as well as the effectiveness of existing train-time robustness interventions.

Chat is not available.