Llama meets EU: Investigating the European Political Spectrum through the Lens of LLMs (2403.13592v2)
Abstract: Instruction-finetuned LLMs inherit clear political leanings that have been shown to influence downstream task performance. We expand this line of research beyond the two-party system in the US and audit Llama Chat in the context of EU politics in various settings to analyze the model's political knowledge and its ability to reason in context. We adapt, i.e., further fine-tune, Llama Chat on speeches of individual euro-parties from debates in the European Parliament to reevaluate its political leaning based on the EUandI questionnaire. Llama Chat shows considerable knowledge of national parties' positions and is capable of reasoning in context. The adapted, party-specific, models are substantially re-aligned towards respective positions which we see as a starting point for using chat-based LLMs as data-driven conversational engines to assist research in political science.
- Palm 2 technical report.
- Adapting large language models via reading comprehension.
- Deep reinforcement learning from human preferences.
- Scaling instruction-finetuned language models.
- Bold: Dataset and metrics for measuring biases in open-ended language generation. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’21, page 862–872, New York, NY, USA. Association for Computing Machinery.
- From pretraining data to language models to downstream tasks: Tracking the trails of political biases leading to unfair NLP models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 11737–11762, Toronto, Canada. Association for Computational Linguistics.
- Opiniongpt: Modelling explicit biases in instruction-tuned llms.
- The political ideology of conversational ai: Converging evidence on chatgpt’s pro-environmental, left-libertarian orientation. ArXiv, abs/2301.01768.
- LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations.
- Mixtral of experts.
- Scalable agent alignment via reward modeling: a research direction. CoRR, abs/1811.07871.
- euandi2019: Project description and datasets documentation. SSRN Electronic Journal.
- OpenAI. 2023. Gpt-4 technical report.
- Nils Reimers. 2021. Easy NMT - Easy to use, state-of-the-art Neural Machine Translation.
- Whose opinions do language models reflect? In Proceedings of the 40th International Conference on Machine Learning, ICML’23. JMLR.org.
- Learning to summarize with human feedback. In Advances in Neural Information Processing Systems, volume 33, pages 3008–3021. Curran Associates, Inc.
- Llama 2: Open foundation and fine-tuned chat models.
- Chain of thought prompting elicits reasoning in large language models. ArXiv, abs/2201.11903.
- Ilias Chalkidis (40 papers)
- Stephanie Brandl (14 papers)