Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
12 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ComplexityNet: Increasing LLM Inference Efficiency by Learning Task Complexity (2312.11511v3)

Published 12 Dec 2023 in cs.CL, cs.AI, and cs.LG

Abstract: We present ComplexityNet, a streamlined LLM designed for assessing task complexity. This model predicts the likelihood of accurate output by various LLMs, each with different capabilities. Our initial application of ComplexityNet involves the Mostly Basic Python Problems (MBPP) dataset. We pioneered the creation of the first set of labels to define task complexity. ComplexityNet achieved a notable 79% accuracy in determining task complexity, a significant improvement over the 34% accuracy of the original, non fine-tuned model. Furthermore, ComplexityNet effectively reduces computational resource usage by 90% compared to using the highest complexity model, while maintaining a high code generation accuracy of 86.7%. This study demonstrates that fine-tuning smaller models to categorize tasks based on their complexity can lead to a more balanced trade-off between accuracy and efficiency in the use of LLMs. Our findings suggest a promising direction for optimizing LLM applications, especially in resource-constrained environments.

Citations (1)

Summary

We haven't generated a summary for this paper yet.