Poster
Specialized Foundation Models Struggle to Beat Supervised Baselines
Zongzhe Xu · Ritvik Gupta · Wenduo Cheng · Alexander Shen · Junhong Shen · Ameet Talwalkar · Mikhail Khodak
Hall 3 + Hall 2B #40
Following its success for vision and text, the "foundation model" (FM) paradigmpretraining large models on massive data, then fine-tuning on target taskshas rapidly expanded to domains in the sciences, engineering, healthcare, and beyond. Has this achieved what the original FMs accomplished, i.e. the supplanting of traditional supervised learning in their domains? To answer we look at three modalitiesgenomics, satellite imaging, and time serieswith multiple recent FMs and compare them to a standard supervised learning workflow: model development, hyperparameter tuning, and training, all using only data from the target task. Across these three specialized domains, we find that it is consistently possible to train simple supervised modelsno more complicated than a lightly modified wide ResNet or UNetthat match or even outperform the latest foundation models. Our work demonstrates that the benefits of large-scale pretraining have yet to be realized in many specialized areas, reinforces the need to compare new FMs to strong, well-tuned baselines, and introduces two new, easy-to-use, open-source, and automated workflows for doing so.
Live content is unavailable. Log in and register to view live content