Constructing meaningful negative examples for source–grounding consistency
Identify a principled procedure to construct meaningful negative examples for a binary classification formulation of translation that labels source–grounding pairs (S, z) as consistent or inconsistent, specifically defining what grounding data should constitute inconsistency with a given source communication so that the reduction is well-posed for learning and evaluation.
References
One might hope to apply these active learning insights to translation by reducing it to binary classification: classify examples x=(S,z) as positive when grounding data z is consistent with source S and negative otherwise. However, while source-grounding pairs from translation training data provide natural positive examples, it's unclear how to construct meaningful negative examples---what would constitute grounding data that's 'inconsistent' with a source?