Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Emoji Prediction in Tweets using BERT (2307.02054v3)

Published 5 Jul 2023 in cs.CL and cs.AI

Abstract: In recent years, the use of emojis in social media has increased dramatically, making them an important element in understanding online communication. However, predicting the meaning of emojis in a given text is a challenging task due to their ambiguous nature. In this study, we propose a transformer-based approach for emoji prediction using BERT, a widely-used pre-trained LLM. We fine-tuned BERT on a large corpus of text (tweets) containing both text and emojis to predict the most appropriate emoji for a given text. Our experimental results demonstrate that our approach outperforms several state-of-the-art models in predicting emojis with an accuracy of over 75 percent. This work has potential applications in natural language processing, sentiment analysis, and social media marketing.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Muhammad Osama Nusrat (2 papers)
  2. Zeeshan Habib (1 paper)
  3. Mehreen Alam (3 papers)
  4. Saad Ahmed Jamal (5 papers)
Citations (3)