Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 71 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 23 tok/s Pro
GPT-5 High 17 tok/s Pro
GPT-4o 111 tok/s Pro
Kimi K2 161 tok/s Pro
GPT OSS 120B 412 tok/s Pro
Claude Sonnet 4 35 tok/s Pro
2000 character limit reached

Linear Classifier: An Often-Forgotten Baseline for Text Classification (2306.07111v1)

Published 12 Jun 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Large-scale pre-trained LLMs such as BERT are popular solutions for text classification. Due to the superior performance of these advanced methods, nowadays, people often directly train them for a few epochs and deploy the obtained model. In this opinion paper, we point out that this way may only sometimes get satisfactory results. We argue the importance of running a simple baseline like linear classifiers on bag-of-words features along with advanced methods. First, for many text data, linear methods show competitive performance, high efficiency, and robustness. Second, advanced models such as BERT may only achieve the best results if properly applied. Simple baselines help to confirm whether the results of advanced models are acceptable. Our experimental results fully support these points.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. A training algorithm for optimal margin classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, pages 144–152. ACM Press.
  2. LexGLUE: A benchmark dataset for legal language understanding in English. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, pages 4310–4330.
  3. BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, pages 4171–4186.
  4. LIBLINEAR: a library for large linear classification. Journal of Machine Learning Research, 9:1871–1874.
  5. Rong-En Fan and Chih-Jen Lin. 2007. A study on threshold selection for multi-label classification. Technical report, Department of Computer Science, National Taiwan University.
  6. On the cost-effectiveness of stacking of neural and non-neural methods for text classification: Scenarios and performance prediction. In Findings of the Association for Computational Linguistics: ACL/IJCNLP, pages 4003–4014.
  7. Deep Learning. The MIT Press.
  8. Karen S. Jones. 1972. A statistical interpretation of term specificity and its application in retrieval. Journal of Documentation, 28(1):11–20.
  9. Ken Lang. 1995. Newsweeder: Learning to filter netnews. In Proceedings of the Twelfth International Conference on Machine Learning, pages 331–339.
  10. RCV1: A new benchmark collection for text categorization research. Journal of Machine Learning Research, 5:361–397.
  11. On the use of unrealistic predictions in hundreds of papers evaluating graph representations. In Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence (AAAI).
  12. Hans Peter Luhn. 1958. The automatic creation of literature abstracts. IBM Journal of Research and Development, 2(2):159–165.
  13. Optimizing f-measures by cost-sensitive classification. In Advances in Neural Information Processing Systems, volume 27.
  14. A comparison of svm against pre-trained language models (plms) for text classification tasks. In Machine Learning, Optimization, and Data Science, pages 304–313. Springer Nature Switzerland.
  15. Yiming Yang. 2001. A study on thresholding strategies for text categorization. In Proceedings of the 24th ACM International Conference on Research and Development in Information Retrieval, pages 137–145, New Orleans, US. ACM Press, New York, US.
  16. PECOS: Prediction for enormous and correlated output spaces. Journal of Machine Learning Research, 23(98):1–32.
Citations (8)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube