Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Representing Affect Information in Word Embeddings (2209.10583v1)

Published 21 Sep 2022 in cs.CL

Abstract: A growing body of research in NLP and natural language understanding (NLU) is investigating human-like knowledge learned or encoded in the word embeddings from LLMs. This is a step towards understanding what knowledge LLMs capture that resembles human understanding of language and communication. Here, we investigated whether and how the affect meaning of a word (i.e., valence, arousal, dominance) is encoded in word embeddings pre-trained in large neural networks. We used the human-labeled dataset as the ground truth and performed various correlational and classification tests on four types of word embeddings. The embeddings varied in being static or contextualized, and how much affect specific information was prioritized during the pre-training and fine-tuning phase. Our analyses show that word embedding from the vanilla BERT model did not saliently encode the affect information of English words. Only when the BERT model was fine-tuned on emotion-related tasks or contained extra contextualized information from emotion-rich contexts could the corresponding embedding encode more relevant affect information.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yuhan Zhang (38 papers)
  2. Wenqi Chen (7 papers)
  3. Ruihan Zhang (26 papers)
  4. Xiajie Zhang (4 papers)
Citations (3)