Can lightweight LLMs gain strong agentic capabilities via pre-training?
Determine whether lightweight large language models can acquire strong agentic capabilities solely through pre-training, without relying on post-augmentation approaches such as post-training procedures or external agentic frameworks, in order to establish the feasibility of native agentic competence in small-scale models.
Sponsor
References
Nevertheless, existing studies leave a critical open question unanswered: Can lightweight LLMs acquire strong agentic capabilities through pre-training, rather than post-augmentation, such as post-training or agentic frameworks?
— Youtu-LLM: Unlocking the Native Agentic Potential for Lightweight Large Language Models
(2512.24618 - Lu et al., 31 Dec 2025) in Section 1 Introduction