Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

JaQuAD: Japanese Question Answering Dataset for Machine Reading Comprehension (2202.01764v1)

Published 3 Feb 2022 in cs.CL, cs.AI, and cs.LG

Abstract: Question Answering (QA) is a task in which a machine understands a given document and a question to find an answer. Despite impressive progress in the NLP area, QA is still a challenging problem, especially for non-English languages due to the lack of annotated datasets. In this paper, we present the Japanese Question Answering Dataset, JaQuAD, which is annotated by humans. JaQuAD consists of 39,696 extractive question-answer pairs on Japanese Wikipedia articles. We finetuned a baseline model which achieves 78.92% for F1 score and 63.38% for EM on test set. The dataset and our experiments are available at https://github.com/SkelterLabsInc/JaQuAD.

Overview of JaQuAD: A Japanese Question Answering Dataset for Machine Reading Comprehension

The paper entitled "JaQuAD: Japanese Question Answering Dataset for Machine Reading Comprehension" by ByungHoon So, Kyuhong Byun, Kyungwon Kang, and Seongjin Cho addresses the construction and analysis of a human-annotated question-answering dataset in the Japanese language. The challenges being tackled include the scarcity of annotated datasets that hinder advancements in non-English NLP applications, particularly those focused on machine reading comprehension (MRC) tasks.

While significant strides have been made in question answering (QA) systems within English, performance for other languages like Japanese lag behind due to dataset limitations. In response, the authors introduce JaQuAD, a novel resource that comprises 39,696 extractive question-answer pairs derived from Japanese Wikipedia articles. By curating such a collection, the authors aim to fill the gap and provide the requisite data foundation for training and evaluating Japanese-language MRC models.

The rigor in the annotation process is highlighted, as the dataset was curated by human annotators, ensuring high semantic fidelity and practical relevance of the question-answer pairs extracted. This human-inclusive methodology is critical for the complexity inherent in natural language, where semantics often transcend mere syntactic representation.

Experimental Results

To demonstrate the applicability of the JaQuAD dataset, a baseline model was finetuned, which achieved an F1 score of 78.92% and an Exact Match (EM) score of 63.38% on the test set. These metrics are pivotal in quantifying the model's precision in both recognizing and accurately extracting the answer span. The provision of these scores as benchmarks enables comparative evaluations of subsequent advancements in Japanese MRC models.

Implications and Future Directions

The development and release of JaQuAD underpin promising implications for both computational linguistics research and real-world applications. Practically, the dataset facilitates enhancements in non-English language processing tools, enabling the broader deployment of QA systems in various socio-linguistic contexts. Theoretically, it lays a foundation for cross-linguistic studies in QA and offers insights into the generalizability of current architectures across different linguistic structures.

Future developments may include leveraging this dataset to explore transfer learning scenarios where MRC models trained on English datasets are adapted to Japanese, potentially reducing the extensive labeled data requirements. Additionally, the dataset could help catalyze the development of multilingual models that inherently support various languages through integrated architectures, which are increasingly pertinent due to globalization and multicultural communication demands.

In summary, JaQuAD represents a pivotal step in democratizing QA technology for Japanese-language users and contributes significantly to the ecosystem of multilingual natural language processing research. The availability of the dataset on an accessible platform further encourages collaborative improvements and potentially fosters innovation extending well beyond the current baseline results documented in this paper.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. ByungHoon So (1 paper)
  2. Kyuhong Byun (1 paper)
  3. Kyungwon Kang (3 papers)
  4. Seongjin Cho (1 paper)
Citations (9)