2000 character limit reached
CMV-BERT: Contrastive multi-vocab pretraining of BERT (2012.14763v2)
Published 29 Dec 2020 in cs.CL
Abstract: In this work, we represent CMV-BERT, which improves the pretraining of a LLM via two ingredients: (a) contrastive learning, which is well studied in the area of computer vision; (b) multiple vocabularies, one of which is fine-grained and the other is coarse-grained. The two methods both provide different views of an original sentence, and both are shown to be beneficial. Downstream tasks demonstrate our proposed CMV-BERT are effective in improving the pretrained LLMs.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.