Skip to yearly menu bar Skip to main content


Poster

RouteLLM: Learning to Route LLMs from Preference Data

Isaac Ong · Amjad Almahairi · Vincent Wu · Wei-Lin Chiang · Tianhao Wu · Joseph E Gonzalez · M Kadous · Ion Stoica

Hall 3 + Hall 2B #153
[ ] [ Project Page ]
Wed 23 Apr 7 p.m. PDT — 9:30 p.m. PDT

Abstract:

Large language models (LLMs) excel at a wide range of tasks, but choosing the right model often involves balancing performance and cost. Powerful models offer better results but are expensive, while smaller models are more cost-effective but less capable. To address this trade-off, we introduce a training framework for learning efficient router models that dynamically select between a stronger and weaker LLM during inference. Our framework leverages human preference data and employs data augmentation techniques to enhance performance. Evaluations on public benchmarks show that our approach can reduce costs by over 2 times without sacrificing response quality. Moreover, our routers exhibit strong generalization capabilities, maintaining performance even when routing between LLMs not included in training. This highlights the potential of our framework to deliver cost-effective, high-performance LLM solutions.

Live content is unavailable. Log in and register to view live content