- The paper introduces a novel sequence-to-structure model that unifies trigger detection and argument labeling for event extraction.
- It employs a constrained decoding algorithm and curriculum learning to align with predefined schemas and improve accuracy.
- Experimental results on ACE05-EN and ERE-EN datasets demonstrate competitive F1-scores and effective transfer learning capabilities.
Text2Event: Controllable Sequence-to-Structure Generation for End-to-End Event Extraction
The paper presents a novel approach to the complex task of event extraction using a sequence-to-structure generation paradigm, called Text2Event, to transform text into event records efficiently. Traditional methods typically decompose the task into subtasks such as detecting triggers and identifying arguments. Text2Event aims to alleviate these complexities by utilizing a unified model that handles event extraction end-to-end.
Methodology
Text2Event employs a sequence-to-structure network, a constrained decoding algorithm for real-time event knowledge injection, and a curriculum learning strategy for model training:
- Sequence-to-Structure Network: The authors leverage a neural network to concurrently process triggers, arguments, and their associated labels. This network does not require fine-grained token-level annotations, which enhances its efficiency.
- Constrained Decoding Algorithm: This component ensures that event extraction aligns with pre-defined schemas, injecting domain-specific knowledge into the model's inference phase, thus improving accuracy.
- Curriculum Learning: To facilitate model training, a staged approach is employed. The model is initially trained on simplified substructure tasks and progressively exposed to more complex full-structure tasks.
Experimental Results
The authors conducted extensive experiments on multiple datasets including ACE05-EN and ERE-EN. Text2Event demonstrated competitive results:
- On ACE05-EN, Text2Event, without fine-grained annotations, achieved results rivaling those of state-of-the-art models that employ hierarchical and entity-based annotations.
- The transfer learning abilities of Text2Event were tested by pre-training on a subset of events and fine-tuning on new events, resulting in F1-score improvements.
Implications and Future Work
Text2Event introduces a new paradigm in the field of event extraction by modeling tasks in a simplified and unified framework. This approach not only enhances data efficiency but also encourages knowledge transfer across different event types. The implications for AI research are significant, as the sequence-to-structure approach could be adapted to other information extraction tasks such as N-ary relation extraction. Future developments might focus on refining the sequence-to-structure model and applying it to more diverse datasets and tasks.
In summary, Text2Event represents a sophisticated integration of neural generation models and event extraction requirements, offering a promising direction for research focused on natural language understanding.