Improving Bias Mitigation through Bias Experts in Natural Language Understanding (2312.03577v1)
Abstract: Biases in the dataset often enable the model to achieve high performance on in-distribution data, while poorly performing on out-of-distribution data. To mitigate the detrimental effect of the bias on the networks, previous works have proposed debiasing methods that down-weight the biased examples identified by an auxiliary model, which is trained with explicit bias labels. However, finding a type of bias in datasets is a costly process. Therefore, recent studies have attempted to make the auxiliary model biased without the guidance (or annotation) of bias labels, by constraining the model's training environment or the capability of the model itself. Despite the promising debiasing results of recent works, the multi-class learning objective, which has been naively used to train the auxiliary model, may harm the bias mitigation effect due to its regularization effect and competitive nature across classes. As an alternative, we propose a new debiasing framework that introduces binary classifiers between the auxiliary model and the main model, coined bias experts. Specifically, each bias expert is trained on a binary classification task derived from the multi-class classification task via the One-vs-Rest approach. Experimental results demonstrate that our proposed strategy improves the bias identification ability of the auxiliary model. Consequently, our debiased model consistently outperforms the state-of-the-art on various challenge datasets.
- Mitigating dataset bias by using per-sample gradient. In The Eleventh International Conference on Learning Representations, ICLR 2023. OpenReview.net.
- Learning de-biased representations with biased representations. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, volume 119 of Proceedings of Machine Learning Research, pages 528–539. PMLR.
- Adversarial filters of dataset biases. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, volume 119 of Proceedings of Machine Learning Research, pages 1078–1088. PMLR.
- Rubi: Reducing unimodal biases for visual question answering. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, pages 839–850. Curran Associates, Inc.
- Don’t take the easy way out: Ensemble based methods for avoiding known dataset biases. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, pages 4067–4080. Association for Computational Linguistics.
- Learning to model and ignore dataset bias with mixed capacity ensembles. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 3031–3045. Association for Computational Linguistics.
- BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, pages 4171–4186. Association for Computational Linguistics.
- The advantages of multiple classes for reducing overfitting from test set reuse. In Proceedings of the 36th International Conference on Machine Learning, ICML 2020, volume 97 of Proceedings of Machine Learning Research, pages 1892–1900. PMLR.
- End-to-end self-debiasing framework for robust NLU training. In Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, pages 1923–1929. Association for Computational Linguistics.
- Annotation artifacts in natural language inference data. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, pages 107–112. Association for Computational Linguistics.
- Unlearn dataset bias in natural language inference by fitting the residual. In Proceedings of the 2nd Workshop on Deep Learning Approaches for Low-Resource NLP, DeepLo@EMNLP-IJCNLP 2019, pages 132–142. Association for Computational Linguistics.
- Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, pages 770–778. IEEE Computer Society.
- Learning not to learn: Training deep neural networks with biased data. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2019, pages 9012–9020. Computer Vision Foundation / IEEE.
- Learning debiased classifier with biased committee. In Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, pages 18403–18415. Curran Associates, Inc.
- Learning debiased representation via disentangled feature augmentation. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, pages 25123–25133. Curran Associates, Inc.
- Revisiting the importance of amplifying bias for debiasing. In Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, pages 14974–14981. AAAI Press.
- Just train twice: Improving group robustness without training group information. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, volume 139 of Proceedings of Machine Learning Research, pages 6781–6792. PMLR.
- Feature-level debiased natural language understanding. In Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, pages 13353–13361. AAAI Press.
- End-to-end bias mitigation by modelling biases in corpora. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, pages 8706–8716. Association for Computational Linguistics.
- Right for the wrong reasons: Diagnosing syntactic heuristics in natural language inference. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, pages 3428–3448. Association for Computational Linguistics.
- Learning from failure: De-biasing classifier from biased classifier. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, pages 20673–20684. Curran Associates, Inc.
- Hypothesis only baselines in natural language inference. In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics, *SEM@NAACL-HLT 2018, pages 180–191. Association for Computational Linguistics.
- "why should I trust you?": Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 2016, pages 1135–1144. ACM.
- Ryan M. Rifkin and Aldebaro Klautau. 2004. In defense of one-vs-all classification. J. Mach. Learn. Res., 5:101–141.
- Learning from others’ mistakes: Avoiding dataset biases without modeling them. In The Ninth International Conference on Learning Representations, ICLR 2021. OpenReview.net.
- Towards debiasing fact verification models. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, pages 3417–3423. Association for Computational Linguistics.
- FEVER: a large-scale dataset for fact extraction and verification. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, pages 809–819. Association for Computational Linguistics.
- Well-read students learn better: On the importance of pre-training compact models. arXiv preprint arXiv:1908.08962.
- Mitigating spurious correlation in natural language understanding with counterfactual inference. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, pages 11308–11321. Association for Computational Linguistics.
- Mind the trade-off: Debiasing NLU models without degrading the in- distribution performance. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, pages 8717–8729. Association for Computational Linguistics.
- Towards debiasing NLU models from unknown biases. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, pages 7597–7610. Association for Computational Linguistics.
- Sphereface2: Binary classification is all you need for deep face recognition. In The Tenth International Conference on Learning Representations, ICLR 2022. OpenReview.net.
- A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, pages 1112–1122. Association for Computational Linguistics.
- Increasing robustness to spurious correlations using forgettable examples. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL 2021, pages 3319–3332. Association for Computational Linguistics.
- PAWS: paraphrase adversaries from word scrambling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, pages 1298–1308. Association for Computational Linguistics.
- Zhilu Zhang and Mert R. Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, pages 8792–8802. Curran Associates, Inc.
- Object recognition with and without objects. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI 2017, pages 3609–3615. ijcai.org.
- Eojin Jeon (2 papers)
- Mingyu Lee (10 papers)
- Juhyeong Park (3 papers)
- Yeachan Kim (12 papers)
- Wing-Lam Mok (2 papers)
- SangKeun Lee (18 papers)