Investigating Cross-Domain Behaviors of BERT in Review Understanding (2306.15123v2)
Abstract: Review score prediction requires review text understanding, a critical real-world application of natural language processing. Due to dissimilar text domains in product reviews, a common practice is fine-tuning BERT models upon reviews of differing domains. However, there has not yet been an empirical study of cross-domain behaviors of BERT models in the various tasks of product review understanding. In this project, we investigate text classification BERT models fine-tuned on single-domain and multi-domain Amazon review data. In our findings, though single-domain models achieved marginally improved performance on their corresponding domain compared to multi-domain models, multi-domain models outperformed single-domain models when evaluated on multi-domain data, single-domain data the single-domain model was not fine-tuned on, and on average when considering all tests. Though slight increases in accuracy can be achieved through single-domain model fine-tuning, computational resources and costs can be reduced by utilizing multi-domain models that perform well across domains.
- Approaches to cross-domain sentiment analysis: A systematic literature review. In IEEE Access, volume 5, pages 16173–16192.
- A brief analysis of amazon online reviews. In 2019 Sixth International Conference on Social Networks Analysis, Management and Security (SNAMS), pages 555–560.
- Adaptation algorithms for neural network-based speech recognition: An overview. In IEEE Open Journal of Signal Processing, pages 33–66.
- Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186.
- Improving children’s speech recognition through out-of-domain data augmentation. In INTERSPEECH 2016, pages 1598–1602.
- Ruining He and Julian McAuley. 2016. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In proceedings of the 25th international conference on world wide web, pages 507–517.
- Domain-specific knowledge distillation yields smaller and better models for conversational commerce. In Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5), pages 151–160.
- Elshrif Ibrahim Elmurngi and Abdelouahed Gherbi. 2018. Unfair reviews detection on amazon reviews using sentiment analysis with supervised learning techniques. In Journal of Computer Science, pages 714–726.
- Raviraj Joshi and Anupam Singh. 2022. A simple baseline for domain adaptation in end to end asr systems using synthetic data. In Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5), pages 217–223.
- Does bert need domain adaptation for clinical negation detection? In Journal of the American Medical Informatics Association, volume 27, page 584–591.
- Cross-domain sentiment aware word embeddings for review sentiment analysis. In International Journal of Machine Learning and Cybernetics, volume 12, pages 343–354.
- Image-based recommendations on styles and substitutes. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval, pages 43–52.
- Domain adaptation via teacher-student learning for end-to-end speech recognition. In IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pages 268–275.
- Transferable high-level representations of bert for cross-domain sentiment classification. In Proceedings on the International Conference on Artificial Intelligence, pages 135–141.
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. In NeurIPS EMC2 Workshop.
- An unsupervised deep domain adaptation approach for robust speech recognition. In Neurocomputing, volume 257, pages 79–87.
- Are we there yet? exploring clinical domain knowledge of bert models. In Proceedings of the 20th Workshop on Biomedical Language Processing, pages 41–53.
- End-to-end asr: from supervised to semi-supervised learning with modern architectures. In ICML Self-supervision in Audio and Speech.
- Sentiment analysis on large scale amazon product reviews. In 2018 IEEE international conference on innovative research and development (ICIRD), pages 1–6.
- Calendar graph neural networks for modeling time structures in spatiotemporal user behaviors. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pages 2581–2589.
- Modeling co-evolution of attributed and structural information in graph sequence. IEEE Transactions on Knowledge and Data Engineering.