Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Structured Self-Attention Weights Encode Semantics in Sentiment Analysis (2010.04922v1)

Published 10 Oct 2020 in cs.CL and cs.AI

Abstract: Neural attention, especially the self-attention made popular by the Transformer, has become the workhorse of state-of-the-art NLP models. Very recent work suggests that the self-attention in the Transformer encodes syntactic information; Here, we show that self-attention scores encode semantics by considering sentiment analysis tasks. In contrast to gradient-based feature attribution methods, we propose a simple and effective Layer-wise Attention Tracing (LAT) method to analyze structured attention weights. We apply our method to Transformer models trained on two tasks that have surface dissimilarities, but share common semantics---sentiment analysis of movie reviews and time-series valence prediction in life story narratives. Across both tasks, words with high aggregated attention weights were rich in emotional semantics, as quantitatively validated by an emotion lexicon labeled by human annotators. Our results show that structured attention weights encode rich semantics in sentiment analysis, and match human interpretations of semantics.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zhengxuan Wu (37 papers)
  2. Thanh-Son Nguyen (8 papers)
  3. Desmond C. Ong (26 papers)
Citations (15)

Summary

We haven't generated a summary for this paper yet.