Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Hierarchy-Aware T5 with Path-Adaptive Mask Mechanism for Hierarchical Text Classification (2109.08585v1)

Published 17 Sep 2021 in cs.CL and cs.AI

Abstract: Hierarchical Text Classification (HTC), which aims to predict text labels organized in hierarchical space, is a significant task lacking in investigation in natural language processing. Existing methods usually encode the entire hierarchical structure and fail to construct a robust label-dependent model, making it hard to make accurate predictions on sparse lower-level labels and achieving low Macro-F1. In this paper, we propose a novel PAMM-HiA-T5 model for HTC: a hierarchy-aware T5 model with path-adaptive mask mechanism that not only builds the knowledge of upper-level labels into low-level ones but also introduces path dependency information in label prediction. Specifically, we generate a multi-level sequential label structure to exploit hierarchical dependency across different levels with Breadth-First Search (BFS) and T5 model. To further improve label dependency prediction within each path, we then propose an original path-adaptive mask mechanism (PAMM) to identify the label's path information, eliminating sources of noises from other paths. Comprehensive experiments on three benchmark datasets show that our novel PAMM-HiA-T5 model greatly outperforms all state-of-the-art HTC approaches especially in Macro-F1. The ablation studies show that the improvements mainly come from our innovative approach instead of T5.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Wei Huang (318 papers)
  2. Chen Liu (206 papers)
  3. Yihua Zhao (4 papers)
  4. Xinyun Yang (1 paper)
  5. Zhaoming Pan (5 papers)
  6. Zhimin Zhang (97 papers)
  7. Guiquan Liu (8 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.