Papers
Topics
Authors
Recent
2000 character limit reached

Emotion-Enhanced Multi-Task ACSA Framework

Updated 2 December 2025
  • The paper presents an emotion-enhanced multi-task ACSA framework that integrates discrete emotional dimensions into sentiment polarity prediction using a unified encoder–decoder approach.
  • It employs parallel sentiment and emotion classification heads built on a Flan-T5 architecture, leveraging joint training and VAD-based refinement for consistent labeling.
  • Experimental results on benchmark datasets demonstrate improved F1 scores, validating the effectiveness of combining discrete Ekman emotions with continuous VAD information.

Emotion-enhanced multi-task Aspect Category Sentiment Analysis (ACSA) frameworks advance sentiment analysis by integrating discrete emotional dimensions into the sentiment polarity modeling process. The explicit joint modeling of sentiment polarity and category-specific emotions—anchored in Ekman’s six basic emotions and neutral—enables LLMs to capture fine-grained affective cues for each aspect category within text. This paradigm addresses the limitation of traditional ACSA, which typically restricts annotation and prediction to coarse sentiment polarity, thereby neglecting the underlying affective dimensions that shape sentiment expressions. Coupled with a Valence-Arousal-Dominance (VAD)–guided refinement mechanism, the framework ensures emotional label fidelity, elevating both the reliability and the interpretability of aspect-sensitive sentiment predictions (Chai et al., 24 Nov 2025).

1. Model Design and Multi-Task Framework

The core of the emotion-enhanced multi-task ACSA approach is a unified, generative encoder–decoder architecture based on Flan-T5 (3B parameters). The model processes an input sentence SS and a predefined set of aspect categories C={c1,...,cm}C = \{c_1, ..., c_m\}, and employs two parallel heads formulated as different target output sequences:

  • Sentiment Polarity Head: Generates ordered category–polarity pairs (ci,yi)(c_i, y_i), where yi{positive,neutral,negative}y_i \in \{\text{positive}, \text{neutral}, \text{negative}\}.
  • Aspect Emotion Classification Head: Generates ordered category–emotion pairs (ci,ei)(c_i, e_i), with eiE={anger,disgust,fear,joy,sadness,surprise,neutral}e_i \in E = \{\text{anger}, \text{disgust}, \text{fear}, \text{joy}, \text{sadness}, \text{surprise}, \text{neutral}\}.

Both heads share all encoder–decoder parameters; differentiation arises solely through target serialization. The multi-task design allows simultaneous learning of coarse sentiment and fine-grained emotion, capturing more nuanced affective features associated with each aspect (Chai et al., 24 Nov 2025).

2. Task Specification and Training Objective

The formalization comprises two generative tasks:

  • Aspect-Category Sentiment Analysis (ACSA): Predicts Ysen={(ci,yi)}Y^{\mathrm{sen}} = \{ (c_i, y_i) \} for all ciCc_i \in C.
  • Aspect Emotion Classification (AEC): Predicts Yemo={(ci,ei)}Y^{\mathrm{emo}} = \{ (c_i, e_i) \} for all ciCc_i \in C.

Negative log-likelihood loss is used for both tasks: Lsen(θ)=E[t=1Tlogpθ(ytseny<tsen,tsen(x))]\mathcal{L}_{\mathrm{sen}}(\theta) = -\mathbb{E} \left[ \sum_{t=1}^T \log p_\theta(y_t^{\mathrm{sen}} \mid y_{<t}^{\mathrm{sen}}, t^{\mathrm{sen}}(x)) \right]

Lemo(θ)=E[t=1Tlogpθ(ytemoy<temo,temo(x))]\mathcal{L}_{\mathrm{emo}}(\theta) = -\mathbb{E} \left[ \sum_{t=1}^T \log p_\theta(y_t^{\mathrm{emo}} \mid y_{<t}^{\mathrm{emo}}, t^{\mathrm{emo}}(x)) \right]

The joint training objective is

L(θ)=αLsen(θ)+(1α)Lemo(θ)\mathcal{L}(\theta) = \alpha\,\mathcal{L}_{\mathrm{sen}}(\theta) + (1-\alpha)\,\mathcal{L}_{\mathrm{emo}}(\theta)

where α[0,1]\alpha \in [0,1] balances the two tasks. Empirically, α=0.6\alpha = 0.6 yields optimal F1 on validation data. No task-specific regularization beyond AdamW weight decay is used (Chai et al., 24 Nov 2025).

3. Generative Emotion Description and Labeling

Emotion assignment leverages LLM-driven generative capabilities. Each sentence is decomposed (via GPT-4o-mini prompts) into aspect-focused sub-sentences {si}\{s_i\}, each aligned with a specific (ci,yi)(c_i, y_i). The emotion generation prompt for each sis_i is:

“Given sub-sentence: ‘…’, the aspect category: <<c_i>>, the sentiment polarity: <<y_i>>. From {anger,disgust,fear,joy,sadness,surprise,neutral}\{\text{anger}, \text{disgust}, \text{fear}, \text{joy}, \text{sadness}, \text{surprise}, \text{neutral}\}, select the most appropriate emotion label.”

This process operationalizes category-level emotion annotation, providing explicit emotional context per aspect that complements coarse polarity judgments (Chai et al., 24 Nov 2025).

4. VAD-Based Emotion Refinement and Consistency

To enhance emotion label fidelity, a VAD-based refinement mechanism is introduced:

  • Continuous VAD Prediction: For each sis_i, a DeBERTa model fine-tuned on EmoBank predicts (v^i,a^i,d^i)[1,5]3(\hat v_i, \hat a_i, \hat d_i) \in [1,5]^3, which are normalized to [1,1][-1,1]:

v=(v^i3)/2,a=(a^i3)/2,d=(d^i3)/2v = (\hat v_i - 3)/2,\quad a = (\hat a_i - 3)/2,\quad d = (\hat d_i - 3)/2

  • Centroid Mapping: Each Ekman emotion ee has a fixed VAD centroid (ve,ae,de)(v^*_e, a^*_e, d^*_e).
  • Discrete VAD-Mapped Emotion:

eivad=argmineE(v,a,d)(ve,ae,de)22e^{\mathrm{vad}}_i = \arg\min_{e\in E} \| (v,a,d) - (v^*_e, a^*_e, d^*_e) \|_2^2

  • Consistency and Refinement: If the initial LLM-generated eie_i matches eivade^{\mathrm{vad}}_i, it is retained. Otherwise, the LLM is prompted for revision given sis_i, cic_i, yiy_i, the original eie_i, and eivade^{\mathrm{vad}}_i, directing alignment with both context and VAD space. This process ensures emotion labels are semantically and affectively consistent (Chai et al., 24 Nov 2025).

5. Training Setup and Benchmark Evaluation

The framework is evaluated on SemEval-2015/2016 Restaurant (Rest15/Rest16) and Laptop (Lap15/Lap16) datasets. Each is split into 90% train, 10% validation (held out), with official test splits retained. Experimental results are averaged over three random seeds.

Hyperparameter details:

  • Backbone: Flan-T5 (3B)
  • Emotion generator: GPT-4o-mini
  • VAD predictor: DeBERTa fine-tuned on EmoBank
  • Optimization: AdamW, learning rate 3×1053 \times 10^{-5}, batch size 4, epochs 10
  • Task balance parameter: α\alpha selected via grid search, optimal at 0.6
  • Early stopping: based on validation ACSA F1

Each batch computes Lsen\mathcal{L}_{\mathrm{sen}} and Lemo\mathcal{L}_{\mathrm{emo}} from the same inputs with separate serialization. Parameters θ\theta are updated to minimize the joint loss L(θ)\mathcal{L}(\theta) (Chai et al., 24 Nov 2025).

6. Quantitative Analysis and Ablation Studies

Empirical evaluation demonstrates that the emotion-enhanced multi-task ACSA framework yields superior performance over baseline models, including backbone Flan-T5 and PBJM. F1 score improvements are observed on all benchmarks:

Dataset Full Model Flan-T5 PBJM Δ\Delta (Flan-T5) Δ\Delta (PBJM)
Rest15 81.65 80.21 67.58 +1.44 +14.07
Rest16 85.01 84.58 75.03 +0.43 +9.98
Lap15 78.17 75.89 62.41 +2.28 +15.76
Lap16 67.95 66.36 55.62 +1.59 +12.33

Ablation results indicate:

  • Removal of emotion task supervision reduces F1 (e.g., Lap15: –2.28).
  • Removal of VAD-based emotion refinement induces further declines (e.g., Rest15: –2.86). This demonstrates that both explicit emotion label learning and VAD-guided consistency contribute to the observed improvements in sentiment polarity prediction (Chai et al., 24 Nov 2025).

7. Integration of Discrete and Dimensional Affect

The framework synthesizes discrete (Ekman’s emotion taxonomy) and continuous (VAD) affective modeling. The dual approach fortifies the model's capacity to represent not only high-level sentiment polarity, but also the nuanced emotional landscape underpinning aspect-category judgments. The explicit VAD-based consistency check, followed by LLM-mediated label revision, reduces subjective misalignment that may arise from generative emotion prediction. A plausible implication is that such hybrid modeling may facilitate more interpretable and robust sentiment analysis for applications requiring psychologically grounded affective interpretations (Chai et al., 24 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Emotion-Enhanced Multi-Task ACSA Framework.