Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Document-Level Supervision for Multi-Aspect Sentiment Analysis Without Fine-grained Labels (2310.06940v1)

Published 10 Oct 2023 in cs.CL

Abstract: Aspect-based sentiment analysis (ABSA) is a widely studied topic, most often trained through supervision from human annotations of opinionated texts. These fine-grained annotations include identifying aspects towards which a user expresses their sentiment, and their associated polarities (aspect-based sentiments). Such fine-grained annotations can be expensive and often infeasible to obtain in real-world settings. There is, however, an abundance of scenarios where user-generated text contains an overall sentiment, such as a rating of 1-5 in user reviews or user-generated feedback, which may be leveraged for this task. In this paper, we propose a VAE-based topic modeling approach that performs ABSA using document-level supervision and without requiring fine-grained labels for either aspects or sentiments. Our approach allows for the detection of multiple aspects in a document, thereby allowing for the possibility of reasoning about how sentiment expressed through multiple aspects comes together to form an observable overall document-level sentiment. We demonstrate results on two benchmark datasets from two different domains, significantly outperforming a state-of-the-art baseline.

Introduction

The importance of aspect-based sentiment analysis (ABSA) is evident given the explosion of user-generated textual content. However, the conventional approach to ABSA requires extensive fine-grained annotations which render it impractical in many applications. This research proposes a novel Variational Auto-Encoder (VAE) based topic modeling technique that leverages document-level sentiment ratings for ABSA without the need for fine-grained labels on aspects or sentiments. The significance of this work lies in its ability to deduce multiple aspects and sentiments within a document using only the overarching sentiment score, thereby presenting a compelling solution for analyzing user feedback efficiently.

Methodology

The proposed model departs from traditional topic models by using document-level sentiment scores instead of aspect-level annotations. It infers topic distributions within documents through a VAE, where the input to the encoder is the token embeddings from a pretrained transformer. Significantly, the researchers freeze the transformer weights during training, which is crucial for model performance. Topics are then associated with aspects or sentiments, and the model employs a pooled sentiment representation for each aspect to forecast the overall sentiment. This framework permits extracting multiple aspects and their sentiments from a single document, forging a link between latent aspect-sentiment configurations and observable document-level sentiments.

Evaluation

For evaluation, the paper harnesses datasets from the restaurant and laptop domains, comparing results to JASen, which represents the state-of-the-art in unsupervised ABSA. Quantitative results showcase superior performance for both aspect detection and aspect-sentiment pairing, outperforming the baseline by significant margins across both domains. Additionally, they qualitatively demonstrate the model's ability to discern topically relevant terms for various aspects and sentiments, further underscoring the efficacy of the proposed approach.

Conclusion & Outlook

This work addresses a crucial gap in ABSA research by providing a robust model that operates without granular annotations, circumventing the cost-intensive and laborious process of data labeling. The effectiveness of the model in real-world datasets indicates its promise for practical applications. The paper concludes by suggesting enhancements to the aspect-sentiment detection accuracy, positing the inclusion of a minimal set of labeled examples to guide the model. Such continuations of the research are likely to further optimize the trade-off between the need for labeled data and the desire for comprehensive, nuanced sentiment analysis.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Latent target-opinion as prior for document-level sentiment classification: A variational approach from fine-grained perspective. In Proceedings of the Web Conference 2021, WWW ’21, page 553–564, New York, NY, USA. Association for Computing Machinery.
  2. W2vlda: Almost unsupervised system for aspect based sentiment analysis. In Expert Systems with Applications, volume 91, pages 127–137.
  3. An unsupervised neural attention model for aspect extraction. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 388–397, Vancouver, Canada. Association for Computational Linguistics.
  4. Open-domain targeted sentiment analysis via span-based extraction and classification. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 537–546.
  5. Weakly-supervised aspect-based sentiment analysis via joint aspect-sentiment topic embedding. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6989–6999, Online. Association for Computational Linguistics.
  6. Diederik P Kingma and Max Welling. 2013. Auto-encoding variational bayes.
  7. Exploiting coherence for the simultaneous discovery of latent facets and associated sentiments. In Proceedings of the 2011 SIAM International Conference on Data Mining (SDM), pages 498–509.
  8. Dual graph convolutional networks for aspect-based sentiment analysis. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 6319–6329.
  9. Chenghua Lin and Yulan He. 2009. Joint sentiment/topic model for sentiment analysis. In Proceedings of the 18th ACM Conference on Information and Knowledge Management, CIKM ’09, page 375–384, New York, NY, USA. Association for Computing Machinery.
  10. Bing Liu. 2012. Sentiment analysis and opinion mining. volume 5, pages 1–167. Morgan & Claypool Publishers.
  11. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
  12. Efficient estimation of word representations in vector space. In 1st International Conference on Learning Representations, ICLR 2013, Scottsdale, Arizona, USA, May 2-4, 2013, Workshop Track Proceedings.
  13. Justifying recommendations using distantly-labeled reviews and fine-grained aspects. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 188–197, Hong Kong, China. Association for Computational Linguistics.
  14. Opinion mining and sentiment analysis. volume 2, pages 1–135. Now Publishers, Inc.
  15. Knowing what, how and why: A near complete solution for aspect-based sentiment analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 8600–8607.
  16. A disentangled adversarial neural topic model for separating opinions from plots in user reviews. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2870–2883, Online. Association for Computational Linguistics.
  17. SemEval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), pages 19–30, San Diego, California. Association for Computational Linguistics.
  18. SemEval-2015 task 12: Aspect based sentiment analysis. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), pages 486–495, Denver, Colorado. Association for Computational Linguistics.
  19. Radim Rehurek and Petr Sojka. 2011. Gensim–python framework for vector space modelling. NLP Centre, Faculty of Informatics, Masaryk University, Brno, Czech Republic, 3(2).
  20. Adapt or get left behind: Domain adaptation through bert language model finetuning for aspect-target sentiment classification. In Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020),.
  21. Akash Srivastava and Charles Sutton. 2017. Autoencoding variational inference for topic models. In International Conference on Learning Representations (ICLR).
  22. Dependency graph enhanced dual-transformer structure for aspect-based sentiment classification. In Proceedings of the 58th annual meeting of the association for computational linguistics, pages 6578–6588.
  23. Stéphan Tulkens and Andreas van Cranenburgh. 2020. Embarrassingly simple unsupervised aspect extraction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 3182–3187, Online. Association for Computational Linguistics.
  24. Coupled multi-layer attentions for co-extraction of aspect and opinion terms. In Proceedings of the AAAI conference on artificial intelligence, volume 31.
  25. Wei Xue and Tao Li. 2018. Aspect based sentiment analysis with gated convolutional networks. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics.
  26. Aspect-based sentiment classification with aspect-specific graph convolutional networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing.
  27. Joint aspect-sentiment analysis with minimal user guidance. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, page 1241–1250, New York, NY, USA. Association for Computing Machinery.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)