Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Hierarchical Graph Network for Multi-hop Question Answering (1911.03631v4)

Published 9 Nov 2019 in cs.CL

Abstract: In this paper, we present Hierarchical Graph Network (HGN) for multi-hop question answering. To aggregate clues from scattered texts across multiple paragraphs, a hierarchical graph is created by constructing nodes on different levels of granularity (questions, paragraphs, sentences, entities), the representations of which are initialized with pre-trained contextual encoders. Given this hierarchical graph, the initial node representations are updated through graph propagation, and multi-hop reasoning is performed via traversing through the graph edges for each subsequent sub-task (e.g., paragraph selection, supporting facts extraction, answer prediction). By weaving heterogeneous nodes into an integral unified graph, this hierarchical differentiation of node granularity enables HGN to support different question answering sub-tasks simultaneously. Experiments on the HotpotQA benchmark demonstrate that the proposed model achieves new state of the art, outperforming existing multi-hop QA approaches.

Hierarchical Graph Network for Multi-Hop Question Answering

The paper "Hierarchical Graph Network for Multi-hop Question Answering" introduces a novel method, Hierarchical Graph Network (HGN), to address the complexities inherent in multi-hop question answering (QA) tasks. Multi-hop QA is a nuanced development in machine comprehension where the goal is to derive answers that require integrating information from various documents or paragraphs—a departure from traditional one-hop QA, where answers are contained within a single paragraph. The authors leverage a hierarchical graph structure to improve inference processes by integrating multiple levels of textual granularity, ultimately enhancing performance on multi-hop reasoning tasks.

Core Contributions and Methodology

The primary contribution of this paper lies in the construction of a hierarchical graph that incorporates nodes representing questions, paragraphs, sentences, and entities. This multi-level structure enables the effective propagation of information, structured to facilitate the answering of complex multi-hop questions. The representations of these nodes are initialized using pre-trained contextual encoders, specifically BERT and RoBERTa, and refined through graph propagation techniques.

Key steps in the methodology include:

  • Graph Construction: Identifying multi-hop relevant paragraphs and establishing connections among sentences/entities via paragraphs using hyperlinks.
  • Context Encoding: Employing Transformer-based models like RoBERTa to encode contexts, leading to rich initial node representations.
  • Graph Reasoning: Utilizing Graph Attention Networks (GATs) to perform message-passing across the hierarchically structured graph nodes. This step updates the node representations essential for further reasoning.
  • Multi-task Prediction: Simultaneous performance of sub-tasks such as paragraph selection, supporting facts prediction, and final answer prediction through multi-task learning.

This extensive framework results in improved predictions by allowing joint answer/evidence deduction, offering effective supervision signals during reasoning processes. The paper delivers detailed insights into the nuances of constructing and utilizing hierarchical graphs for QA.

Experimental Results and Analysis

Experiments conducted on the HotpotQA benchmark reveal that the HGN model establishes new state-of-the-art performances, significantly surpassing existing approaches in multi-hop QA tasks. Notably, results on both the Distractor and Fullwiki settings underscore the efficacy of HGN, recording improvements of 1.48% in the joint F1 score metric on the Distractor setting while showing effective performance in the Fullwiki context, a more challenging dataset with fewer explicit clues.

The paper further supports its findings through a robust ablation paper that verifies the added value of hierarchical graphs, entity nodes, and other components. By demonstrating that even incremental graph complexity leads to measurable improvements, the paper provides empirical reinforcement to the design decisions inherent in HGN.

Implications and Future Directions

The implications of this research are manifold, particularly in enhancing methods for machine-reading tasks necessitating sophisticated reasoning. The hierarchical graph approach suggests broader applicability beyond HotpotQA, pointing towards potential integration within other multi-paragraph comprehension datasets, possibly expanding capabilities in open-domain question answering systems.

Looking forward, future research can focus on optimizing graph construction techniques for longer documents, enhancing retrieval algorithms, or integrating joint training paradigms between graph networks and retrievers to further bolster QA performance across challenging benchmarks.

In summary, "Hierarchical Graph Network for Multi-hop Question Answering" stands out due to its well-founded techniques in improving multi-hop reasoning, setting a precedence for graph-based model enhancements in complex QA scenarios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yuwei Fang (31 papers)
  2. Siqi Sun (46 papers)
  3. Zhe Gan (135 papers)
  4. Rohit Pillai (4 papers)
  5. Shuohang Wang (69 papers)
  6. Jingjing Liu (139 papers)
Citations (164)