Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TextPruner: A Model Pruning Toolkit for Pre-Trained Language Models (2203.15996v1)

Published 30 Mar 2022 in cs.CL

Abstract: Pre-trained LLMs have been prevailed in natural language processing and become the backbones of many NLP tasks, but the demands for computational resources have limited their applications. In this paper, we introduce TextPruner, an open-source model pruning toolkit designed for pre-trained LLMs, targeting fast and easy model compression. TextPruner offers structured post-training pruning methods, including vocabulary pruning and transformer pruning, and can be applied to various models and tasks. We also propose a self-supervised pruning method that can be applied without the labeled data. Our experiments with several NLP tasks demonstrate the ability of TextPruner to reduce the model size without re-training the model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Ziqing Yang (29 papers)
  2. Yiming Cui (80 papers)
  3. Zhigang Chen (102 papers)
Citations (11)