Papers
Topics
Authors
Recent
2000 character limit reached

Generative Monoculture in Large Language Models (2407.02209v1)

Published 2 Jul 2024 in cs.CL and cs.AI

Abstract: We introduce {\em generative monoculture}, a behavior observed in LLMs characterized by a significant narrowing of model output diversity relative to available training data for a given task: for example, generating only positive book reviews for books with a mixed reception. While in some cases, generative monoculture enhances performance (e.g., LLMs more often produce efficient code), the dangers are exacerbated in others (e.g., LLMs refuse to share diverse opinions). As LLMs are increasingly used in high-impact settings such as education and web search, careful maintenance of LLM output diversity is essential to ensure a variety of facts and perspectives are preserved over time. We experimentally demonstrate the prevalence of generative monoculture through analysis of book review and code generation tasks, and find that simple countermeasures such as altering sampling or prompting strategies are insufficient to mitigate the behavior. Moreover, our results suggest that the root causes of generative monoculture are likely embedded within the LLM's alignment processes, suggesting a need for developing fine-tuning paradigms that preserve or promote diversity.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (55)
  1. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682, 2022.
  2. Holistic evaluation of language models. Annals of the New York Academy of Sciences, 1525(1):140–146, 2023.
  3. Foundational challenges in assuring alignment and safety of large language models. arXiv preprint arXiv:2404.09932, 2024.
  4. Decodingtrust: A comprehensive assessment of trustworthiness in gpt models. arXiv preprint arXiv:2306.11698, 2023.
  5. Hallucinating law: Legal mistakes with large language models are pervasive, Jan 2024.
  6. HaluEval: A large-scale hallucination evaluation benchmark for large language models. In Houda Bouamor, Juan Pino, and Kalika Bali, editors, Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 6449–6464, Singapore, December 2023. Association for Computational Linguistics.
  7. Siren’s song in the ai ocean: a survey on hallucination in large language models. arXiv preprint arXiv:2309.01219, 2023.
  8. Persistent anti-muslim bias in large language models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, pages 298–306, 2021.
  9. Unveiling the implicit toxicity in large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 1322–1338, 2023.
  10. The dark side of language models: Exploring the potential of llms in multimedia disinformation generation and dissemination. Machine Learning with Applications, page 100545, 2024.
  11. Exploring the deceptive power of llm-generated fake news: A study of real-world detection challenges. arXiv preprint arXiv:2403.18249, 2024.
  12. Fine-grained spoiler detection from large-scale review corpora. In Anna Korhonen, David R. Traum, and Lluís Màrquez, editors, Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers, pages 2605–2610. Association for Computational Linguistics, 2019.
  13. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  14. Competition-level code generation with alphacode. Science, 378(6624):1092–1097, 2022.
  15. Do users write more insecure code with ai assistants? In Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, pages 2785–2799, 2023.
  16. Asleep at the keyboard? assessing the security of github copilot’s code contributions. In 2022 IEEE Symposium on Security and Privacy (SP), pages 754–768. IEEE, 2022.
  17. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730–27744, 2022.
  18. Whose opinions do language models reflect? arXiv preprint arXiv:2303.17548, 2023.
  19. Vishakh Padmakumar and He He. Does writing with language models reduce content diversity? arXiv preprint arXiv:2309.05196, 2023.
  20. A roadmap to pluralistic alignment. arXiv preprint arXiv:2402.05070, 2024.
  21. Algorithmic monoculture and social welfare. Proceedings of the National Academy of Sciences, 118(22):e2018340118, 2021.
  22. Picking on the same person: Does algorithmic monoculture lead to outcome homogenization? Advances in Neural Information Processing Systems, 35:3663–3678, 2022.
  23. Proving test set contamination in black box language models. arXiv preprint arXiv:2310.17623, 2023.
  24. Perils of self-feedback: Self-bias amplifies in large language models. arXiv preprint arXiv:2402.11436, 2024.
  25. Sentence-bert: Sentence embeddings using siamese bert-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 11 2019.
  26. Winnowing: local algorithms for document fingerprinting. In Proceedings of the 2003 ACM SIGMOD international conference on Management of data, pages 76–85, 2003.
  27. Bryson Lingenfelter. blingenf/copydetect: Code plagiarism detection tool. Accessed: 2024-03-20 at https://github.com/blingenf/copydetect.
  28. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633–2650, 2021.
  29. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  30. Quantifying language models’ sensitivity to spurious features in prompt design or: How i learned to start worrying about prompt formatting. arXiv preprint arXiv:2310.11324, 2023.
  31. In-context impersonation reveals large language models’ strengths and biases. Advances in Neural Information Processing Systems, 36, 2024.
  32. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  33. Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality, March 2023.
  34. Microsoft. Openai models - azure openai service. https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models, 2024. accessed: 2024-04-08.
  35. The curious case of neural text degeneration. arXiv preprint arXiv:1904.09751, 2019.
  36. HuggingFace. distilbert/distilbert-base-uncased-finetuned-sst-2-english. Accessed: 2024-03-20 at https://huggingface.co/distilbert/distilbert-base-uncased-finetuned-sst-2-english.
  37. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pages 1631–1642, 2013.
  38. Hanna M Wallach. Topic modeling: beyond bag-of-words. In Proceedings of the 23rd international conference on Machine learning, pages 977–984, 2006.
  39. A survey of topic modeling in text mining. Int. J. Adv. Comput. Sci. Appl.(IJACSA), 6(1), 2015.
  40. Maarten Grootendorst. Bertopic: Neural topic modeling with a class-based tf-idf procedure. arXiv preprint arXiv:2203.05794, 2022.
  41. HuggingFace. Maartengr/bertopic_wikipedia. Accessed: 2024-03-20 at {https://huggingface.co/MaartenGr/BERTopic_Wikipedia}.
  42. CodeForces. Codeforces. Accessed: 2024-03-20 at https://codeforces.com/.
  43. Anthropic. Claude-3 language model. https://www.anthropic.com/, 2024. Accessed: 2024-04-08.
  44. Effibench: Benchmarking the efficiency of automatically generated code. arXiv preprint arXiv:2402.02037, 2024.
  45. Generating diverse code explanations using the gpt-3 large language model. In Proceedings of the 2022 ACM Conference on International Computing Education Research-Volume 2, pages 37–39, 2022.
  46. Using an llm to help with code understanding. In 2024 IEEE/ACM 46th International Conference on Software Engineering (ICSE), pages 881–881. IEEE Computer Society, 2024.
  47. HuggingFace. sentence-transformers/all-minilm-l6-v2. Accessed: 2024-03-20 at https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2.
  48. Understanding the effects of RLHF on LLM generalisation and diversity. In The Twelfth International Conference on Learning Representations, 2024.
  49. OpenAI Community. Web chat default temperature for gpt-3.5 and 4. https://community.openai.com/t/web-chat-default-temperature-for-gpt-3-5-and-4/167356/5, 2023. accessed: 2024-04-08.
  50. OpenAI. Chat completions - openai api. https://platform.openai.com/docs/api-reference/chat/create, 2024. accessed: 2024-04-08.
  51. NLTK: The natural language toolkit. In Proceedings of the ACL Interactive Poster and Demonstration Sessions, pages 214–217, Barcelona, Spain, July 2004. Association for Computational Linguistics.
  52. Michael Kerrisk. time(1) — linux manual page, 2023.
  53. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems, 35:24824–24837, 2022.
  54. Code llama: Open foundation models for code. arXiv preprint arXiv:2308.12950, 2023.
  55. Starcoder: may the source be with you! arXiv preprint arXiv:2305.06161, 2023.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 24 likes about this paper.