LLMs for Targeted Sentiment in News Headlines: Exploring the Descriptive-Prescriptive Dilemma
Abstract: News headlines often evoke sentiment by intentionally portraying entities in particular ways, making targeted sentiment analysis (TSA) of headlines a worthwhile but difficult task. Due to its subjectivity, creating TSA datasets can involve various annotation paradigms, from descriptive to prescriptive, either encouraging or limiting subjectivity. LLMs are a good fit for TSA due to their broad linguistic and world knowledge and in-context learning abilities, yet their performance depends on prompt design. In this paper, we compare the accuracy of state-of-the-art LLMs and fine-tuned encoder models for TSA of news headlines using descriptive and prescriptive datasets across several languages. Exploring the descriptive--prescriptive continuum, we analyze how performance is affected by prompt prescriptiveness, ranging from plain zero-shot to elaborate few-shot prompts. Finally, we evaluate the ability of LLMs to quantify uncertainty via calibration error and comparison to human label variation. We find that LLMs outperform fine-tuned encoders on descriptive datasets, while calibration and F1-score generally improve with increased prescriptiveness, yet the optimal level varies.
- Opinion mining of news headlines using SentiWordNet. In 2016 Symposium on Colossal Data Analysis and Networking (CDAN), pages 1–5.
- Open-source large language models outperform crowd workers and approach chatgpt in text-annotation tasks.
- Sentiments and emotions evoked by news headlines of coronavirus disease (covid-19) outbreak. Humanities and Social Sciences Communications, 7(1).
- Sentiment Analysis in the News. ArXiv:1309.6202 [cs].
- Katarzyna Baraniak and Marcin Sydow. 2021. A dataset for Sentiment analysis of Entities in News headlines (SEN). Procedia Computer Science, 192:3627–3636.
- Target Two Birds With One SToNe: Entity-Level Sentiment and Tone Analysis in Croatian News Headlines. In Proceedings of the 9th Workshop on Slavic Natural Language Processing 2023 (SlavicNLP 2023), pages 78–85, Dubrovnik, Croatia. Association for Computational Linguistics.
- Language models are few-shot learners.
- Ting-Yun Chang and Robin Jia. 2023. Data curation alone can stabilize in-context learning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8123–8144, Toronto, Canada. Association for Computational Linguistics.
- Generative approach to Aspect Based Sentiment Analysis with GPT Language Models. Procedia Computer Science, 229:284–293.
- SemEval-2017 Task 5: Fine-Grained Sentiment Analysis on Financial Microblogs and News. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), pages 519–535, Vancouver, Canada. Association for Computational Linguistics.
- Pre-training polish transformer-based language models at scale. In Artificial Intelligence and Soft Computing, pages 301–314. Springer International Publishing.
- Bert: Pre-training of deep bidirectional transformers for language understanding.
- MAD-TSC: A Multilingual Aligned News Dataset for Target-dependent Sentiment Classification. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8286–8305, Toronto, Canada. Association for Computational Linguistics.
- Felix Hamborg and Karsten Donnay. 2021. NewsMTSC: A Dataset for (Multi-)Target-dependent Sentiment Classification in Political News Articles. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 1663–1675, Online. Association for Computational Linguistics.
- Towards Target-Dependent Sentiment Classification in News Articles. In Diversity, Divergence, Dialogue, Lecture Notes in Computer Science, pages 156–166, Cham. Springer International Publishing.
- Reducing Sentiment Bias in Language Models via Counterfactual Evaluation. ArXiv:1911.03064 [cs].
- Stock trend prediction using news sentiment analysis. International Journal of Computer Science and Information Technology, 8:67–76.
- Roberta: A robustly optimized bert pretraining approach.
- Nikola Ljubešić and Davor Lauc. 2021. BERTi\’c – The Transformer Language Model for Bosnian, Croatian, Montenegrin and Serbian. ArXiv:2104.09243 [cs].
- Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity.
- Dealing with Disagreements: Looking Beyond the Majority Vote in Subjective Annotations. Transactions of the Association for Computational Linguistics, 10:92–110.
- M. Melih Mutlu and Arzucan Özgür. 2022. A Dataset and BERT-based Models for Targeted Sentiment Analysis on Turkish Texts. ArXiv:2205.04185 [cs].
- László Nemes and Attila Kiss. 2021. Prediction of stock values changes using sentiment analysis of stock news headlines. Journal of Information and Telecommunication, 5(3):375–394.
- Gpt-4 technical report.
- Automated annotation with generative AI requires validation. CoRR, abs/2306.00176.
- Targeted sentiment analysis: A data-driven categorization.
- Combining Confidence Elicitation and Sample-based Methods for Uncertainty Quantification in Misinformation Mitigation. ArXiv:2401.08694 [cs].
- Two contrasting data annotation paradigms for subjective NLP tasks. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 175–190, Seattle, United States. Association for Computational Linguistics.
- Longitudinal analysis of sentiment and emotion in news media headlines using automated labelling with Transformer language models. PLOS ONE, 17(10):e0276367.
- A spanish dataset for targeted sentiment analysis of political headlines.
- Holli Semetko and Patti Valkenburg. 2000. Framing european politics: A content analysis of press and television news. Journal of Communication, 50:93 – 109.
- Multilingual Entity-Centered Sentiment Analysis Evaluated by Parallel Corpora. In Proceedings of the International Conference Recent Advances in Natural Language Processing 2011, pages 770–775, Hissar, Bulgaria. Association for Computational Linguistics.
- Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback. ArXiv:2305.14975 [cs].
- Llama 2: Open foundation and fine-tuned chat models.
- Want to reduce labeling cost? GPT-3 can help. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 4195–4205, Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Zhengxuan Wu and Desmond C. Ong. 2021. Context-Guided BERT for Targeted Aspect-Based Sentiment Analysis. Proceedings of the AAAI Conference on Artificial Intelligence, 35(16):14094–14102.
- Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs. ArXiv:2306.13063 [cs].
- A BERT Fine-tuning Model for Targeted Sentiment Analysis of Chinese Online Course Reviews. International Journal on Artificial Intelligence Tools, 29(07n08):2040018.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.