Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KoreALBERT: Pretraining a Lite BERT Model for Korean Language Understanding (2101.11363v1)

Published 27 Jan 2021 in cs.CL and cs.LG

Abstract: A Lite BERT (ALBERT) has been introduced to scale up deep bidirectional representation learning for natural languages. Due to the lack of pretrained ALBERT models for Korean language, the best available practice is the multilingual model or resorting back to the any other BERT-based model. In this paper, we develop and pretrain KoreALBERT, a monolingual ALBERT model specifically for Korean language understanding. We introduce a new training objective, namely Word Order Prediction (WOP), and use alongside the existing MLM and SOP criteria to the same architecture and model parameters. Despite having significantly fewer model parameters (thus, quicker to train), our pretrained KoreALBERT outperforms its BERT counterpart on 6 different NLU tasks. Consistent with the empirical results in English by Lan et al., KoreALBERT seems to improve downstream task performance involving multi-sentence encoding for Korean language. The pretrained KoreALBERT is publicly available to encourage research and application development for Korean NLP.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Jaewoong Yoon (2 papers)
  2. Bonggyu Hwang (1 paper)
  3. Seongho Joe (11 papers)
  4. Seungjai Min (7 papers)
  5. Youngjune Gwon (20 papers)
  6. HyunJae Lee (10 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.