PIEClass: Weakly-Supervised Text Classification with Prompting and Noise-Robust Iterative Ensemble Training (2305.13723v2)
Abstract: Weakly-supervised text classification trains a classifier using the label name of each target class as the only supervision, which largely reduces human annotation efforts. Most existing methods first use the label names as static keyword-based features to generate pseudo labels, which are then used for final classifier training. While reasonable, such a commonly adopted framework suffers from two limitations: (1) keywords can have different meanings in different contexts and some text may not have any keyword, so keyword matching can induce noisy and inadequate pseudo labels; (2) the errors made in the pseudo label generation stage will directly propagate to the classifier training stage without a chance of being corrected. In this paper, we propose a new method, PIEClass, consisting of two modules: (1) a pseudo label acquisition module that uses zero-shot prompting of pre-trained LLMs (PLM) to get pseudo labels based on contextualized text understanding beyond static keyword matching, and (2) a noise-robust iterative ensemble training module that iteratively trains classifiers and updates pseudo labels by utilizing two PLM fine-tuning methods that regularize each other. Extensive experiments show that PIEClass achieves overall better performance than existing strong baselines on seven benchmark datasets and even achieves similar performance to fully-supervised classifiers on sentiment classification tasks.
- Eugene Agichtein and Luis Gravano. 2000. Snowball: extracting relations from large plain-text collections. In Digital library.
- Learning from rules generalizing labeled exemplars. In ICLR.
- Data programming for learning discourse structure. In ACL.
- Avrim Blum and Tom M. Mitchell. 1998. Combining labeled and unlabeled data with co-training. In COLT.
- Importance of semantic representation: Dataless classification. In AAAI.
- Data programming using continuous and quality-guided labeling functions. In AAAI.
- Knowprompt: Knowledge-aware prompt-tuning with synergistic optimization for relation extraction. In WWW.
- ELECTRA: Pre-training text encoders as discriminators rather than generators. In ICLR.
- Commonsense knowledge mining from pretrained models. In EMNLP-IJCNLP.
- Rlprompt: Optimizing discrete text prompts with reinforcement learning. In EMNLP.
- Bert: Pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT.
- Prompt-learning for fine-grained entity typing. arXiv preprint arXiv:2108.10604.
- Evgeniy Gabrilovich and Shaul Markovitch. 2007. Computing semantic relatedness using wikipedia-based explicit semantic analysis. In IJCAI.
- Making pre-trained language models better few-shot learners. In ACL-IJCNLP.
- Efficient (soft) q-learning for text generation with limited good data. In EMNLP Findings.
- Ptr: Prompt tuning with rules for text classification. arXiv preprint arXiv:2105.11259.
- Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification. In ACL.
- Few-shot fine-grained entity typing with automatic label interpretation and instance generation. In KDD.
- MEGClass: Text classification with extremely weak supervision via mutually-enhancing text granularities. In EMNLP.
- Samuli Laine and Timo Aila. 2017. Temporal ensembling for semi-supervised learning. In ICLR.
- Co-training improves prompt-based learning for large language models. In ICML.
- Ken Lang. 1995. Newsweeder: Learning to filter netnews. In ICML.
- The power of scale for parameter-efficient prompt tuning. In EMNLP.
- Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. In ACL-IJCNLP.
- Pre-trained token-replaced detection model as few-shot learner. In COLING.
- Roberta: A robustly optimized bert pretraining approach. ArXiv, abs/1907.11692.
- Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. In ICLR.
- Learning word vectors for sentiment analysis. In ACL.
- Semi-supervised data programming with subset selection. In ACL-IJCNLP Findings.
- Learning to robustly aggregate labeling functions for semi-supervised data programming. In ACL Findings.
- Julian McAuley and Jure Leskovec. 2013. Hidden factors and hidden topics: understanding rating dimensions with review text. In RecSys’13.
- LOPS: Learning order inspired pseudo-label selection for weakly supervised text classification. In EMNLP Findings.
- Dheeraj Mekala and Jingbo Shang. 2020. Contextualized weak supervision for text classification. In ACL.
- Weakly-supervised neural text classification. In CIKM.
- Distantly-supervised named entity recognition with noise-robust learning and language model augmented self-training. In EMNLP.
- Text classification using label names only: A language model self-training approach. In EMNLP.
- OpenAI. 2023. Gpt-4 technical report. ArXiv, abs/2303.08774.
- Seongmin Park and Jihwa Lee. 2022. LIME: Weakly-supervised text classification without seeds. In COLING.
- Graph alignment with noisy supervision. In WWW.
- Language models as knowledge bases? In EMNLP-IJCNLP.
- Automatic rule induction for efficient semi-supervised learning. In EMNLP Findings.
- Language models are unsupervised multitask learners.
- SQuAD: 100,000+ questions for machine comprehension of text. In EMNLP.
- Data programming: Creating large training sets, quickly. In NIPS.
- Denoising multi-source weak supervision for neural text classification. In EMNLP Findings.
- Evan Sandhaus. 2008. The New York Times Annotated Corpus.
- Automatically identifying words that can serve as labels for few-shot text classification. In COLING.
- AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts. In EMNLP.
- Learning with weak supervision for email intent detection. In SIGIR.
- Language models in the loop: Incorporating prompting into weak supervision. ArXiv, abs/2205.02318.
- Yangqiu Song and Dan Roth. 2014. On dataless hierarchical text classification. In AAAI.
- Document modeling with gated recurrent neural network for sentiment classification. In EMNLP.
- Doc2cube: Allocating documents to text cube without labeled data. In ICDM.
- Paroma Varma and Christopher Ré. 2018. Snuba: Automating weak supervision to label training data. Proc. VLDB Endow., 12:223–236.
- X-class: Text classification with extremely weak supervision. In NAACL.
- Prompting electra: Few-shot learning with discriminative pre-trained models. In EMNLP.
- Hierarchical attention networks for document classification. In NAACL-HLT.
- Prompt tuning for discriminative pre-trained language models. In ACL Findings.
- Weakly supervised text classification using supervision signals from a language model. In NAACL Findings.
- Weakly-supervised text classification based on keyword graph. In EMNLP.
- Prompt-based rule discovery and boosting for interactive weakly-supervised learning. In ACL.
- Character-level convolutional networks for text classification. In NIPS.
- Motifclass: Weakly supervised text classification with higher-order metadata information. In WSDM.
- Unsupervised key event detection from massive text corpora. In KDD.
- Empower entity set expansion via language model probing. In ACL.
- Zhilu Zhang and Mert R. Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. In NIPS.
- Factual probing is [MASK]: Learning vs. learning to recall. In NAACL.
- Weaker than you think: A critical look at weakly supervised learning. In ACL.
- Yunyi Zhang (39 papers)
- Minhao Jiang (10 papers)
- Yu Meng (92 papers)
- Yu Zhang (1399 papers)
- Jiawei Han (263 papers)