Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Training Classifiers with Natural Language Explanations (1805.03818v4)

Published 10 May 2018 in cs.CL

Abstract: Training accurate classifiers requires many labels, but each label provides only limited information (one bit for binary classification). In this work, we propose BabbleLabble, a framework for training classifiers in which an annotator provides a natural language explanation for each labeling decision. A semantic parser converts these explanations into programmatic labeling functions that generate noisy labels for an arbitrary amount of unlabeled data, which is used to train a classifier. On three relation extraction tasks, we find that users are able to train classifiers with comparable F1 scores from 5-100$\times$ faster by providing explanations instead of just labels. Furthermore, given the inherent imperfection of labeling functions, we find that a simple rule-based semantic parser suffices.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Braden Hancock (12 papers)
  2. Paroma Varma (6 papers)
  3. Stephanie Wang (18 papers)
  4. Martin Bringmann (3 papers)
  5. Percy Liang (239 papers)
  6. Christopher RĂ© (194 papers)
Citations (149)

Summary

We haven't generated a summary for this paper yet.