Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adversarial Training for Community Question Answer Selection Based on Multi-scale Matching (1804.08058v2)

Published 22 Apr 2018 in cs.CL

Abstract: Community-based question answering (CQA) websites represent an important source of information. As a result, the problem of matching the most valuable answers to their corresponding questions has become an increasingly popular research topic. We frame this task as a binary (relevant/irrelevant) classification problem, and present an adversarial training framework to alleviate label imbalance issue. We employ a generative model to iteratively sample a subset of challenging negative samples to fool our classification model. Both models are alternatively optimized using REINFORCE algorithm. The proposed method is completely different from previous ones, where negative samples in training set are directly used or uniformly down-sampled. Further, we propose using Multi-scale Matching which explicitly inspects the correlation between words and ngrams of different levels of granularity. We evaluate the proposed method on SemEval 2016 and SemEval 2017 datasets and achieves state-of-the-art or similar performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Xiao Yang (158 papers)
  2. Madian Khabsa (38 papers)
  3. Miaosen Wang (11 papers)
  4. Wei Wang (1793 papers)
  5. Ahmed Awadallah (27 papers)
  6. Daniel Kifer (65 papers)
  7. C. Lee Giles (69 papers)
Citations (22)