Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LT@Helsinki at SemEval-2020 Task 12: Multilingual or language-specific BERT? (2008.00805v1)

Published 3 Aug 2020 in cs.CL

Abstract: This paper presents the different models submitted by the LT@Helsinki team for the SemEval 2020 Shared Task 12. Our team participated in sub-tasks A and C; titled offensive language identification and offense target identification, respectively. In both cases we used the so-called Bidirectional Encoder Representation from Transformer (BERT), a model pre-trained by Google and fine-tuned by us on the OLID and SOLID datasets. The results show that offensive tweet classification is one of several language-based tasks where BERT can achieve state-of-the-art results.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Marc Pàmies (7 papers)
  2. Emily Öhman (6 papers)
  3. Kaisla Kajava (2 papers)
  4. Jörg Tiedemann (41 papers)
Citations (10)