MLE-STAR: Agentic AutoML System
This presentation introduces MLE-STAR, a novel agentic system for automated machine learning engineering that combines large language models with external knowledge retrieval and component-level refinement. The system achieves a 64% medal rate on Kaggle competitions by retrieving state-of-the-art solutions from the web and iteratively refining code at the block level through ablation studies and ensemble strategies, substantially outperforming previous AutoML agents.Script
Previous machine learning agents generate solutions from their internal knowledge alone, limiting their ability to adapt to new techniques. MLE-STAR breaks this pattern by retrieving state-of-the-art model solutions from the web and refining them through targeted code block exploration, achieving a 64% medal rate on Kaggle competitions.
MLE-STAR operates through three breakthrough mechanisms. First, it sources task-specific model solutions directly from the web, capturing techniques published after the Large Language Model's training cutoff. Second, it performs ablation studies to pinpoint which code blocks matter most, then refines those components through nested iteration. Third, it systematically constructs ensembles from diverse candidates, reducing variance and boosting competitive performance.
Let's examine how these specialized agents coordinate their work.
The system orchestrates specialized agents in two major phases. On the left, retrieval and merging agents gather external solutions and combine them intelligently, retaining only modifications that improve validation scores. On the right, ablation analysis pinpoints the most impactful code blocks, triggering deep exploration cycles that converge on high-performing implementations through iterative feedback.
On 22 Kaggle competition tasks, MLE-STAR achieves a 64% medal rate, more than doubling the 25.8% baseline of the AIDE agent. The ensemble module proves especially powerful, systematically pushing solutions into gold medal territory by combining complementary strengths of diverse candidates.
MLE-STAR's architecture solves a fundamental tension in automated machine learning. By retrieving external knowledge, it remains current as the field evolves. By targeting critical code components through ablation, it avoids the inefficiency of undirected mutations. And its modular agent design extends naturally to vision, language, and multimodal pipelines, positioning the framework as a long-term platform for competitive ML automation.
MLE-STAR demonstrates that automated machine learning can move beyond static rule sets and internal knowledge, leveraging the live ecosystem of published techniques to achieve competitive-level performance. Visit EmergentMind.com to explore this research and create your own AI-narrated presentations.