Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Investigating Cross-Domain Behaviors of BERT in Review Understanding (2306.15123v2)

Published 27 Jun 2023 in cs.CL and cs.CE

Abstract: Review score prediction requires review text understanding, a critical real-world application of natural language processing. Due to dissimilar text domains in product reviews, a common practice is fine-tuning BERT models upon reviews of differing domains. However, there has not yet been an empirical study of cross-domain behaviors of BERT models in the various tasks of product review understanding. In this project, we investigate text classification BERT models fine-tuned on single-domain and multi-domain Amazon review data. In our findings, though single-domain models achieved marginally improved performance on their corresponding domain compared to multi-domain models, multi-domain models outperformed single-domain models when evaluated on multi-domain data, single-domain data the single-domain model was not fine-tuned on, and on average when considering all tests. Though slight increases in accuracy can be achieved through single-domain model fine-tuning, computational resources and costs can be reduced by utilizing multi-domain models that perform well across domains.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Approaches to cross-domain sentiment analysis: A systematic literature review. In IEEE Access, volume 5, pages 16173–16192.
  2. A brief analysis of amazon online reviews. In 2019 Sixth International Conference on Social Networks Analysis, Management and Security (SNAMS), pages 555–560.
  3. Adaptation algorithms for neural network-based speech recognition: An overview. In IEEE Open Journal of Signal Processing, pages 33–66.
  4. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186.
  5. Improving children’s speech recognition through out-of-domain data augmentation. In INTERSPEECH 2016, pages 1598–1602.
  6. Ruining He and Julian McAuley. 2016. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In proceedings of the 25th international conference on world wide web, pages 507–517.
  7. Domain-specific knowledge distillation yields smaller and better models for conversational commerce. In Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5), pages 151–160.
  8. Elshrif Ibrahim Elmurngi and Abdelouahed Gherbi. 2018. Unfair reviews detection on amazon reviews using sentiment analysis with supervised learning techniques. In Journal of Computer Science, pages 714–726.
  9. Raviraj Joshi and Anupam Singh. 2022. A simple baseline for domain adaptation in end to end asr systems using synthetic data. In Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5), pages 217–223.
  10. Does bert need domain adaptation for clinical negation detection? In Journal of the American Medical Informatics Association, volume 27, page 584–591.
  11. Cross-domain sentiment aware word embeddings for review sentiment analysis. In International Journal of Machine Learning and Cybernetics, volume 12, pages 343–354.
  12. Image-based recommendations on styles and substitutes. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval, pages 43–52.
  13. Domain adaptation via teacher-student learning for end-to-end speech recognition. In IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pages 268–275.
  14. Transferable high-level representations of bert for cross-domain sentiment classification. In Proceedings on the International Conference on Artificial Intelligence, pages 135–141.
  15. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. In NeurIPS EMC2 Workshop.
  16. An unsupervised deep domain adaptation approach for robust speech recognition. In Neurocomputing, volume 257, pages 79–87.
  17. Are we there yet? exploring clinical domain knowledge of bert models. In Proceedings of the 20th Workshop on Biomedical Language Processing, pages 41–53.
  18. End-to-end asr: from supervised to semi-supervised learning with modern architectures. In ICML Self-supervision in Audio and Speech.
  19. Sentiment analysis on large scale amazon product reviews. In 2018 IEEE international conference on innovative research and development (ICIRD), pages 1–6.
  20. Calendar graph neural networks for modeling time structures in spatiotemporal user behaviors. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pages 2581–2589.
  21. Modeling co-evolution of attributed and structural information in graph sequence. IEEE Transactions on Knowledge and Data Engineering.

Summary

We haven't generated a summary for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com