Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Effective Domain Adaptive Post-Training Method for BERT in Response Selection (1908.04812v2)

Published 13 Aug 2019 in cs.CL and cs.LG

Abstract: We focus on multi-turn response selection in a retrieval-based dialog system. In this paper, we utilize the powerful pre-trained LLM Bi-directional Encoder Representations from Transformer (BERT) for a multi-turn dialog system and propose a highly effective post-training method on domain-specific corpus. Although BERT is easily adopted to various NLP tasks and outperforms previous baselines of each task, it still has limitations if a task corpus is too focused on a certain domain. Post-training on domain-specific corpus (e.g., Ubuntu Corpus) helps the model to train contextualized representations and words that do not appear in general corpus (e.g., English Wikipedia). Experimental results show that our approach achieves new state-of-the-art on two response selection benchmarks (i.e., Ubuntu Corpus V1, Advising Corpus) performance improvement by 5.9% and 6% on R@1.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Taesun Whang (9 papers)
  2. Dongyub Lee (9 papers)
  3. Chanhee Lee (14 papers)
  4. Kisu Yang (7 papers)
  5. Dongsuk Oh (7 papers)
  6. Heuiseok Lim (49 papers)
Citations (26)