Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CPL-NoViD: Context-Aware Prompt-based Learning for Norm Violation Detection in Online Communities (2305.09846v3)

Published 16 May 2023 in cs.CL and cs.SI

Abstract: Detecting norm violations in online communities is critical to maintaining healthy and safe spaces for online discussions. Existing machine learning approaches often struggle to adapt to the diverse rules and interpretations across different communities due to the inherent challenges of fine-tuning models for such context-specific tasks. In this paper, we introduce Context-aware Prompt-based Learning for Norm Violation Detection (CPL-NoViD), a novel method that employs prompt-based learning to detect norm violations across various types of rules. CPL-NoViD outperforms the baseline by incorporating context through natural language prompts and demonstrates improved performance across different rule types. Significantly, it not only excels in cross-rule-type and cross-community norm violation detection but also exhibits adaptability in few-shot learning scenarios. Most notably, it establishes a new state-of-the-art in norm violation detection, surpassing existing benchmarks. Our work highlights the potential of prompt-based learning for context-sensitive norm violation detection and paves the way for future research on more adaptable, context-aware models to better support online community moderators.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Investigating toxicity across multiple Reddit communities, users, and moderators. In Companion proceedings of the web conference 2020, 294–298.
  2. Deep learning for hate speech detection in tweets. In Proceedings of the 26th international conference on World Wide Web companion, 759–760.
  3. Few-shot Text Classification With Distributional Signatures. In International Conference on Learning Representations.
  4. Random search for hyper-parameter optimization. Journal of machine learning research, 13(2).
  5. Social network sites: Definition, history, and scholarship. Journal of computer-mediated Communication, 13(1): 210–230.
  6. Language models are few-shot learners. Advances in neural information processing systems, 33: 1877–1901.
  7. You can’t stay here: The efficacy of reddit’s 2015 ban examined through hate speech. Proceedings of the ACM on Human-Computer Interaction, 1(CSCW): 1–22.
  8. Cropf, R. A. 2008. Benkler, Y.(2006). The Wealth of Networks: How Social Production Transforms Markets and Freedom. New Haven and London: Yale University Press. 528 pp. $40.00 (papercloth). Social Science Computer Review, 26(2): 259–261.
  9. No country for old members: User lifecycle and linguistic change in online communities. In Proceedings of the 22nd international conference on World Wide Web, 307–318.
  10. Automated hate speech detection and the problem of offensive language. In Proceedings of the international AAAI conference on web and social media, volume 11, 512–515.
  11. The spreading of misinformation online. Proceedings of the national academy of Sciences, 113(3): 554–559.
  12. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186.
  13. OpenPrompt: An Open-source Framework for Prompt-learning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 105–113.
  14. The benefits of Facebook “friends:” Social capital and college students’ use of online social network sites. Journal of computer-mediated communication, 12(4): 1143–1168.
  15. Reddit rules! characterizing an ecosystem of governance. In Proceedings of the International AAAI Conference on Web and Social Media, volume 12.
  16. Large scale crowdsourcing and characterization of twitter abusive behavior. In Proceedings of the international AAAI conference on web and social media, volume 12.
  17. Using convolutional neural networks to classify hate-speech. In Proceedings of the first workshop on abusive language online, 85–90.
  18. Making Pre-trained Language Models Better Few-shot Learners. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 3816–3830.
  19. Hanu, L.; and Unitary team. 2020. Detoxify. https://github.com/unitaryai/detoxify.
  20. Reading between the tweets: Deciphering ideological stances of interconnected mixed-ideology communities. arXiv preprint arXiv:2402.01091.
  21. Long short-term memory. Neural computation, 9(8): 1735–1780.
  22. The more (and the more compatible) the merrier: Multiple group memberships and identity compatibility as predictors of adjustment after life transitions. British Journal of Social Psychology, 48(4): 707–733.
  23. Human-machine collaboration for content regulation: The case of reddit automoderator. ACM Transactions on Computer-Human Interaction (TOCHI), 26(5): 1–35.
  24. Does transparency in moderation really matter? User behavior after content removal explanations on reddit. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW): 1–27.
  25. Online harassment and content moderation: The case of blocklists. ACM Transactions on Computer-Human Interaction (TOCHI), 25(2): 1–33.
  26. A Just and Comprehensive Strategy for Using NLP to Address Online Abuse. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 3658–3666.
  27. P-Tuning: Prompt Tuning Can Be Comparable to Fine-tuning Across Scales and Tasks. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 61–68. Dublin, Ireland: Association for Computational Linguistics.
  28. The coddling of the American mind: How good intentions and bad ideas are setting up a generation for failure. Penguin.
  29. Media manipulation and disinformation online.
  30. Abusive language detection in online user content. In Proceedings of the 25th international conference on world wide web, 145–153.
  31. Detecting Community Sensitive Norm Violations in Online Conversations. In Findings of the Association for Computational Linguistics: EMNLP 2021, 3386–3397.
  32. Reducing Gender Bias in Abusive Language Detection. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2799–2804.
  33. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32.
  34. Improving language understanding by generative pre-training.
  35. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1): 5485–5551.
  36. Quick, community-specific learning: How distinctive toxicity norms are maintained in political subreddits. In Proceedings of the International AAAI Conference on Web and Social Media, volume 14, 557–568.
  37. It’s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2339–2352.
  38. Shirky, C. 2008. Here comes everybody: The power of organizing without organizations. Penguin.
  39. Tufekci, Z. 2017. Twitter and tear gas: The power and fragility of networked protest. Yale University Press.
  40. Wang, W. Y. 2017. ” liar, liar pants on fire”: A new benchmark dataset for fake news detection. arXiv preprint arXiv:1705.00648.
  41. The network basis of social support: A network is more than the sum of its ties. In Networks in the global village, 83–118. Routledge.
  42. Huggingface’s transformers: State-of-the-art natural language processing. arXiv preprint arXiv:1910.03771.
  43. Ex machina: Personal attacks seen at scale. In Proceedings of the 26th international conference on world wide web, 1391–1399.
  44. Learning from bullying traces in social media. In Proceedings of the 2012 conference of the North American chapter of the association for computational linguistics: Human language technologies, 656–666.
  45. Identifying semantic edit intentions from revisions in wikipedia. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2000–2010.
  46. Prompt-Based Meta-Learning For Few-shot Text Classification. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 1342–1357.
  47. Conversations Gone Awry: Detecting Early Signs of Conversational Failure. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1350–1361.
  48. Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners. In International Conference on Learning Representations.
Citations (3)

Summary

We haven't generated a summary for this paper yet.