Utilizing Local Hierarchy with Adversarial Training for Hierarchical Text Classification (2402.18825v2)
Abstract: Hierarchical text classification (HTC) is a challenging subtask of multi-label classification due to its complex taxonomic structure. Nearly all recent HTC works focus on how the labels are structured but ignore the sub-structure of ground-truth labels according to each input text which contains fruitful label co-occurrence information. In this work, we introduce this local hierarchy with an adversarial framework. We propose a HiAdv framework that can fit in nearly all HTC models and optimize them with the local hierarchy as auxiliary information. We test on two typical HTC models and find that HiAdv is effective in all scenarios and is adept at dealing with complex taxonomic hierarchies. Further experiments demonstrate that the promotion of our framework indeed comes from the local hierarchy and the local hierarchy is beneficial for rare classes which have insufficient training data.
- Improved training of mixture-of-experts language gans. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE.
- Hyperbolic interaction model for hierarchical multi-label classification. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 7496–7503.
- Hierarchy-aware label semantics matching network for hierarchical text classification. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4370–4379, Online. Association for Computational Linguistics.
- HTCInfoMax: A global model for hierarchical text classification via information maximization. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3259–3265, Online. Association for Computational Linguistics.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- Generative adversarial networks. Communications of the ACM, 63(11):139–144.
- Neural topic modeling with cycle-consistent adversarial training. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 9018–9030, Online. Association for Computational Linguistics.
- Hierarchical verbalizer for few-shot hierarchical text classification. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2918–2933, Toronto, Canada. Association for Computational Linguistics.
- SMART: Robust and efficient fine-tuning for pre-trained natural language models through principled regularized optimization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2177–2190, Online. Association for Computational Linguistics.
- Exploiting global and local hierarchies for hierarchical text classification. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 4030–4039, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Rie Johnson and Tong Zhang. 2015. Effective use of word order for text categorization with convolutional neural networks. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 103–112, Denver, Colorado. Association for Computational Linguistics.
- Catgan: Category-aware generative adversarial networks with hierarchical evolutionary learning for category text generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 8425–8432.
- Hierarchical text classification with reinforced label assignment. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 445–455, Hong Kong, China. Association for Computational Linguistics.
- Mobashir Sadat and Cornelia Caragea. 2022. Hierarchical multi-label classification of scientific documents. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 8923–8937, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Coldgans: Taming language gans with cautious sampling strategies. Advances in Neural Information Processing Systems, 33:18978–18989.
- Carlos N Silla and Alex A Freitas. 2011. A survey of hierarchical classification across different application domains. Data Mining and Knowledge Discovery, 22(1):31–72.
- Zlpr: A novel loss for multi-label classification. arXiv preprint arXiv:2208.02955.
- Infobert: Improving robustness of language models from an information theoretic perspective. In International Conference on Learning Representations.
- Neural topic modeling with bidirectional adversarial training. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 340–350, Online. Association for Computational Linguistics.
- Atm: Adversarial-neural topic model. Information Processing & Management, 56(6):102098.
- Incorporating hierarchy into text encoder: a contrastive learning approach for hierarchical text classification. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7109–7119, Dublin, Ireland. Association for Computational Linguistics.
- HPT: Hierarchy-aware prompt tuning for hierarchical text classification. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 3740–3751, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Toward adversarial training on contextualized language representation. In International Conference on Learning Representations.
- Learning to learn and predict: A meta-learning approach for multi-label classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4354–4364, Hong Kong, China. Association for Computational Linguistics.
- Conditional adversarial networks for multi-domain text classification. In Proceedings of the Second Workshop on Domain Adaptation for NLP, pages 16–27, Kyiv, Ukraine. Association for Computational Linguistics.
- Co-regularized adversarial learning for multi-domain text classification. In International Conference on Artificial Intelligence and Statistics, pages 6690–6701. PMLR.
- Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems, 34:28877–28888.
- La-hcn: Label-based attention for hierarchical multi-label text classification neural network. Expert Systems with Applications, page 115922.
- Hierarchical multi-label text classification: Self-adaption semantic awareness network integrating text topic and label level information. In International Conference on Knowledge Science, Engineering and Management, pages 406–418. Springer.
- Hierarchy-aware global model for hierarchical text classification. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1106–1117, Online. Association for Computational Linguistics.
- Freelb: Enhanced adversarial training for natural language understanding. In International Conference on Learning Representations.
- HiTIN: Hierarchy-aware tree isomorphism network for hierarchical text classification. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7809–7821, Toronto, Canada. Association for Computational Linguistics.
- Web of Science Dataset. Mendeley Data. PID https://doi.org/10.17632/9rw3vkcfy4.6.
- Rcv1: A new benchmark collection for text categorization research. Goldsmiths, University of London. PID https://trec.nist.gov/data/reuters/reuters.html.
- Evan Sandhaus. 2008. The new york times annotated corpus. Linguistic Data Consortium, Philadelphia, ISLRN 429-488-225-160-9.
- Zihan Wang (181 papers)
- Peiyi Wang (48 papers)
- Houfeng Wang (43 papers)