Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HELP: A Dataset for Identifying Shortcomings of Neural Models in Monotonicity Reasoning (1904.12166v1)

Published 27 Apr 2019 in cs.CL

Abstract: Large crowdsourced datasets are widely used for training and evaluating neural models on natural language inference (NLI). Despite these efforts, neural models have a hard time capturing logical inferences, including those licensed by phrase replacements, so-called monotonicity reasoning. Since no large dataset has been developed for monotonicity reasoning, it is still unclear whether the main obstacle is the size of datasets or the model architectures themselves. To investigate this issue, we introduce a new dataset, called HELP, for handling entailments with lexical and logical phenomena. We add it to training data for the state-of-the-art neural models and evaluate them on test sets for monotonicity phenomena. The results showed that our data augmentation improved the overall accuracy. We also find that the improvement is better on monotonicity inferences with lexical replacements than on downward inferences with disjunction and modification. This suggests that some types of inferences can be improved by our data augmentation while others are immune to it.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Hitomi Yanaka (30 papers)
  2. Koji Mineshima (20 papers)
  3. Daisuke Bekki (15 papers)
  4. Kentaro Inui (119 papers)
  5. Satoshi Sekine (11 papers)
  6. Lasha Abzianidze (16 papers)
  7. Johan Bos (27 papers)
Citations (59)