Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MAVEN-ERE: A Unified Large-scale Dataset for Event Coreference, Temporal, Causal, and Subevent Relation Extraction (2211.07342v1)

Published 14 Nov 2022 in cs.CL

Abstract: The diverse relationships among real-world events, including coreference, temporal, causal, and subevent relations, are fundamental to understanding natural languages. However, two drawbacks of existing datasets limit event relation extraction (ERE) tasks: (1) Small scale. Due to the annotation complexity, the data scale of existing datasets is limited, which cannot well train and evaluate data-hungry models. (2) Absence of unified annotation. Different types of event relations naturally interact with each other, but existing datasets only cover limited relation types at once, which prevents models from taking full advantage of relation interactions. To address these issues, we construct a unified large-scale human-annotated ERE dataset MAVEN-ERE with improved annotation schemes. It contains 103,193 event coreference chains, 1,216,217 temporal relations, 57,992 causal relations, and 15,841 subevent relations, which is larger than existing datasets of all the ERE tasks by at least an order of magnitude. Experiments show that ERE on MAVEN-ERE is quite challenging, and considering relation interactions with joint learning can improve performances. The dataset and source codes can be obtained from https://github.com/THU-KEG/MAVEN-ERE.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Xiaozhi Wang (51 papers)
  2. Yulin Chen (134 papers)
  3. Ning Ding (122 papers)
  4. Hao Peng (291 papers)
  5. Zimu Wang (15 papers)
  6. Yankai Lin (125 papers)
  7. Xu Han (270 papers)
  8. Lei Hou (127 papers)
  9. Juanzi Li (144 papers)
  10. Zhiyuan Liu (433 papers)
  11. Peng Li (390 papers)
  12. Jie Zhou (687 papers)
Citations (50)

Summary

Analysis of the Maven-Ere Dataset for Event Relation Extraction

The paper "Maven-Ere: A Unified Large-scale Dataset for Event Coreference, Temporal, Causal, and Subevent Relation Extraction" presents a comprehensive dataset designed to address limitations in the existing datasets used for Event Relation Extraction (ERE) tasks within the domain of NLP. The authors identify two primary limitations in existing datasets: their small scale and the lack of unified annotations for various event relations, which prevents comprehensive model training and evaluation.

Significance of the Dataset

The Maven-Ere dataset is positioned as a substantial advancement in ERE research, boasting a scale that eclipses existing counterparts by at least an order of magnitude. Specifically, it comprises 103,193 event coreference chains, 1,216,217 temporal relations, 57,992 causal relations, and 15,841 subevent relations. This magnitude facilitates the training of data-hungry AI models while supporting diverse applications that require an understanding of the complex relationships between events.

Methodology and Dataset Characteristics

The extensive annotation process for Maven-Ere implements refined methodologies to ensure the dataset's comprehensiveness and reliability, particularly in distinguishing between four types of relations: coreference, temporal, causal, and subevent. To handle the intricate task of temporal relation extraction, the authors innovate with a timeline annotation scheme that allows annotators to sequentially arrange event starting and ending points, bypassing the laborious nature of pairwise comparison within large temporal datasets.

By annotating coreference relations initially, further annotation tasks for temporal, causal, and subevent relations are optimized, leveraging transitivity and other intrinsic relational constraints to improve annotation efficiency and quality. Transitivity is employed in causal relations to reduce redundancy in annotations, allowing some relations to be inferred automatically based on others.

Experimental Results and Implications

Experiments conducted using state-of-the-art LLMs, such as RoBERTa, underline the challenging nature of ERE tasks on the Maven-Ere dataset. While the performance on tasks like event coreference benefitted from the dataset's scale, other tasks like causal and subevent relation extraction remain exigent, indicating room for model improvements. Notably, joint training on multiple relation types shows performance enhancement, emphasizing the potential of integrated approaches that capitalize on inter-relational dependencies.

The larger scale of Maven-Ere provides clearer insight into the model's capacity to generalize across documents, reflecting the challenges of long-range dependency modeling frequently encountered in event relation tasks. Moreover, analysis reveals substantial improvements in model performance correlating with increased training data, suggesting the dataset's sufficiency for model training and potential for advancing various NLP applications.

Future Directions and Challenges

The Maven-Ere dataset opens several avenues for future research and exploration in NLP. The substantial data volume and unified relational annotations create opportunities for developing more sophisticated neural models capable of integrating multiple types of event relations. Moreover, the explicit linkage provided by subevent and causal relations offers a frame for hierarchical event understanding, which could be significant for downstream tasks like narrative comprehension and automated reasoning.

Implementation of the dataset across multilingual applications and diverse linguistic corpora remains an area for expansion, poised to drive further advancements in cross-linguistic AI models and their capabilities. Additionally, while Maven-Ere's timeline annotation reduces some burdens, exploring alternative approaches to automate or semi-automate complex annotations presents an ongoing challenge, the solution to which will likely advance the broader field of machine learning and artificial intelligence.

Overall, the Maven-Ere dataset represents a critical resource for extending the boundaries of ERE research and underscores the importance of scale and integration in augmenting the robustness and efficacy of NLP models.