Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multilingual Multi-Aspect Explainability Analyses on Machine Reading Comprehension Models (2108.11574v3)

Published 26 Aug 2021 in cs.CL and cs.AI

Abstract: Achieving human-level performance on some of the Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained LLMs (PLMs). However, the internal mechanism of these artifacts remains unclear, placing an obstacle for further understanding these models. This paper focuses on conducting a series of analytical experiments to examine the relations between the multi-head self-attention and the final MRC system performance, revealing the potential explainability in PLM-based MRC models. To ensure the robustness of the analyses, we perform our experiments in a multilingual way on top of various PLMs. We discover that passage-to-question and passage understanding attentions are the most important ones in the question answering process, showing strong correlations to the final performance than other parts. Through comprehensive visualizations and case studies, we also observe several general findings on the attention maps, which can be helpful to understand how these models solve the questions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yiming Cui (80 papers)
  2. Wei-Nan Zhang (19 papers)
  3. Wanxiang Che (152 papers)
  4. Ting Liu (329 papers)
  5. Zhigang Chen (102 papers)
  6. Shijin Wang (69 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.