LLMRANK: UNDERSTANDING LLM STRENGTHS FOR MODEL ROUTING

Authors

  • Shubham Agrawal Zeno AI Author
  • Prasang Gupta Independent Researcher Author

Keywords:

LLMRank, RouterBench, lightweight proxy solver

Abstract

The rapid growth of large language models (LLMs) with diverse capabilities, latency and computational costs presents a critical deployment challenge: selecting the most suitable model for each prompt to optimize the trade-off between performance and efficiency. We introduce LLMRank, a prompt-aware routing framework that leverages rich, human-readable features extracted from prompts, including task type, reasoning patterns, complexity indicators, syntactic cues, and signals from a lightweight proxy solver. Unlike prior one-shot routers that rely solely on latent embeddings, LLMRank predicts per-model utility using a neural ranking model trained on RouterBench, comprising 36,497 prompts spanning 11 benchmarks and 11 state-of-the-art LLMs, from small efficient models to large frontier systems. Our approach achieves up to 89.2% of oracle utility, while providing interpretable feature attributions that explain routing decisions. Extensive studies demonstrate the importance of multifaceted feature extraction and the hybrid ranking objective, highlighting the potential of feature-driven routing for efficient and transparent LLM deployment.

Downloads

Published

2025-12-31

How to Cite

Agrawal, S. ., & Gupta, P. . (2025). LLMRANK: UNDERSTANDING LLM STRENGTHS FOR MODEL ROUTING. Journal of Smart Computing and Quantum Technologies, 1(1), 54-65. https://technology.tresearch.ee/index.php/JSCQT/article/view/56