Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 97 tok/s
Gemini 2.5 Pro 50 tok/s Pro
GPT-5 Medium 37 tok/s
GPT-5 High 28 tok/s Pro
GPT-4o 110 tok/s
GPT OSS 120B 468 tok/s Pro
Kimi K2 236 tok/s Pro
2000 character limit reached

Human Parity on CommonsenseQA: Augmenting Self-Attention with External Attention (2112.03254v3)

Published 6 Dec 2021 in cs.CL, cs.AI, and cs.LG

Abstract: Most of today's AI systems focus on using self-attention mechanisms and transformer architectures on large amounts of diverse data to achieve impressive performance gains. In this paper, we propose to augment the transformer architecture with an external attention mechanism to bring external knowledge and context to bear. By integrating external information into the prediction process, we hope to reduce the need for ever-larger models and increase the democratization of AI systems. We find that the proposed external attention mechanism can significantly improve the performance of existing AI systems, allowing practitioners to easily customize foundation AI models to many diverse downstream applications. In particular, we focus on the task of Commonsense Reasoning, demonstrating that the proposed external attention mechanism can augment existing transformer models and significantly improve the model's reasoning capabilities. The proposed system, Knowledgeable External Attention for commonsense Reasoning (KEAR), reaches human parity on the open CommonsenseQA research benchmark with an accuracy of 89.4\% in comparison to the human accuracy of 88.9\%.

Citations (51)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces KEAR, a method that integrates external attention with self-attention to enhance commonsense reasoning and achieve human parity on CommonsenseQA.
  • KEAR combines knowledge from ConceptNet, Wiktionary, and training data by concatenating external knowledge with input text to improve model performance.
  • The approach achieves 89.4% accuracy on CommonsenseQA and demonstrates improvements across various Transformer-based encoders, indicating its broad applicability.

Achieving Human Parity on CommonsenseQA with External Knowledge Integration

This paper (2112.03254) introduces Knowledgeable External Attention for commonsense Reasoning (KEAR), a novel approach to enhance the performance of Transformer models in commonsense reasoning tasks. By augmenting the self-attention mechanism with external attention, KEAR leverages external knowledge sources to improve reasoning capabilities. The paper demonstrates that KEAR achieves human parity on the CommonsenseQA benchmark, surpassing previous state-of-the-art models.

Methodology: External Attention and Knowledge Integration

The KEAR framework enhances the self-attention mechanism of Transformer models by incorporating external knowledge. Instead of relying solely on the input data, KEAR retrieves related context and knowledge from external sources and integrates it into the model. This approach mimics human intelligence by utilizing external resources like search engines and dictionaries to navigate the world. The process involves concatenating relevant knowledge from external sources with the input text and feeding it into the Transformer model. Figure 1

Figure 1: KEAR integrates retrieved knowledge from external sources with the input, leveraging text-level concatenation without altering the model architecture.

Specifically, the input to the first Transformer layer is modified as follows:

H0=[X;K]=[x1,...,xN,x1K,...,xNkK]H_0 = [X; K] = [x_1, ..., x_N, x^K_1, ..., x^K_{N_k}]

where XX represents the input text, and KK represents the external knowledge in text format. This concatenation allows the model to perform self-attention across both the input and the external knowledge, enhancing its reasoning capacity.

External Knowledge Sources: ConceptNet, Wiktionary, and Training Data

KEAR leverages three external knowledge sources to complement the input questions and choices.

  1. Knowledge Graph (ConceptNet): Relevant relation triples are retrieved from ConceptNet to provide curated facts for commonsense reasoning.
  2. Dictionary (Wiktionary): Definitions of key concepts from Wiktionary are used to provide accurate semantic explanations, especially for less frequent words.
  3. Training Data: Relevant questions and answers are retrieved from the training data to provide additional context and information.

The knowledge from these sources is concatenated to form a final knowledge input:

K=[KKG;Kdict;Ktrain]K = [K_{KG}; K_{dict}; K_{train}]

where KKGK_{KG}, KdictK_{dict}, and KtrainK_{train} represent the knowledge retrieved from the knowledge graph, dictionary, and training data, respectively.

Experimental Results: Achieving Human Parity

The paper presents empirical results on the CommonsenseQA benchmark, demonstrating that KEAR achieves human parity with an accuracy of 89.4\% on the test set, compared to human accuracy of 88.9\%. The results also show that each of the three external knowledge sources contributes to the performance improvement.

The authors experiment with various encoders, including ELECTRA, DeBERTa, and DeBERTaV3, and find that KEAR consistently improves the performance of these models. The best single model (DeBERTaV3-large + KEAR) achieves 91.2\% accuracy on the dev set. Furthermore, an ensemble of 39 models with KEAR reaches 93.4\% accuracy on the dev set.

Implications and Future Directions

The KEAR framework offers several practical benefits, including reduced dependence on large-scale models, computationally efficient knowledge retrieval methods, and easy adoption by existing systems. The success of KEAR suggests that integrating external knowledge is a promising direction for improving the reasoning capabilities of AI systems. Future research could explore other NLP tasks to improve LLM performance with external knowledge.

Github Logo Streamline Icon: https://streamlinehq.com