Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TEII: Think, Explain, Interact and Iterate with Large Language Models to Solve Cross-lingual Emotion Detection (2405.17129v2)

Published 27 May 2024 in cs.CL and cs.AI

Abstract: Cross-lingual emotion detection allows us to analyze global trends, public opinion, and social phenomena at scale. We participated in the Explainability of Cross-lingual Emotion Detection (EXALT) shared task, achieving an F1-score of 0.6046 on the evaluation set for the emotion detection sub-task. Our system outperformed the baseline by more than 0.16 F1-score absolute, and ranked second amongst competing systems. We conducted experiments using fine-tuning, zero-shot learning, and few-shot learning for LLM-based models as well as embedding-based BiLSTM and KNN for non-LLM-based techniques. Additionally, we introduced two novel methods: the Multi-Iteration Agentic Workflow and the Multi-Binary-Classifier Agentic Workflow. We found that LLM-based approaches provided good performance on multilingual emotion detection. Furthermore, ensembles combining all our experimented models yielded higher F1-scores than any single approach alone.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (12)
  1. Emotion detection using a bidirectional long-short term memory (bilstm) neural network. International Journal of Current Pharmaceutical Review and Research, Vol 4, no 11:1718–1732.
  2. Cross-lingual emotion detection. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, pages 6948–6958, Marseille, France. European Language Resources Association.
  3. Findings of the wassa 2024 exalt shared task on explainability for cross-lingual emotion in tweets. In Proceedings of the 14th Workshop of on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis@ACL 2024, Bangkok, Thailand.
  4. Gorilla: Large language model connected with massive apis. Preprint, arXiv:2305.15334.
  5. Communicative agents for software development. Preprint, arXiv:2307.07924.
  6. Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face. Preprint, arXiv:2303.17580.
  7. Reflexion: Language agents with verbal reinforcement learning. Preprint, arXiv:2303.11366.
  8. M2SA: Multimodal and multilingual model for sentiment analysis of tweets. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 10833–10845, Torino, Italia. ELRA and ICCL.
  9. Chain-of-thought prompting elicits reasoning in large language models. Preprint, arXiv:2201.11903.
  10. Autogen: Enabling next-gen llm applications via multi-agent conversation. Preprint, arXiv:2308.08155.
  11. Wenbiao Yin and Lin Shang. 2022. Efficient nearest neighbor emotion classification with BERT-whitening. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 4738–4745, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  12. Qimin Zhou and Hao Wu. 2018. NLP at IEST 2018: BiLSTM-attention and LSTM-attention via soft voting in emotion classification. In Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 189–194, Brussels, Belgium. Association for Computational Linguistics.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Long Cheng (77 papers)
  2. Qihao Shao (1 paper)
  3. Christine Zhao (1 paper)
  4. Sheng Bi (28 papers)
  5. Gina-Anne Levow (5 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets