Papers
Topics
Authors
Recent
Search
2000 character limit reached

LLMvsSmall Model? Large Language Model Based Text Augmentation Enhanced Personality Detection Model

Published 12 Mar 2024 in cs.CL | (2403.07581v1)

Abstract: Personality detection aims to detect one's personality traits underlying in social media posts. One challenge of this task is the scarcity of ground-truth personality traits which are collected from self-report questionnaires. Most existing methods learn post features directly by fine-tuning the pre-trained LLMs under the supervision of limited personality labels. This leads to inferior quality of post features and consequently affects the performance. In addition, they treat personality traits as one-hot classification labels, overlooking the semantic information within them. In this paper, we propose a LLM based text augmentation enhanced personality detection model, which distills the LLM's knowledge to enhance the small model for personality detection, even when the LLM fails in this task. Specifically, we enable LLM to generate post analyses (augmentations) from the aspects of semantic, sentiment, and linguistic, which are critical for personality detection. By using contrastive learning to pull them together in the embedding space, the post encoder can better capture the psycho-linguistic information within the post representations, thus improving personality detection. Furthermore, we utilize the LLM to enrich the information of personality labels for enhancing the detection performance. Experimental results on the benchmark datasets demonstrate that our model outperforms the state-of-the-art methods on personality detection.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. The Nature and Structure of Correlations Among Big Five Ratings: The Halo-Alpha-Beta Model. Journal of personality and social psychology, 97: 1142–56.
  2. The Role of the Five-Factor Model in Personality Assessment and Treatment Planning. Clinical Psychology: Science and Practice, 23(4): 365–381.
  3. A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 632–642.
  4. Language Models are Few-Shot Learners. In Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M.; and Lin, H., eds., Advances in Neural Information Processing Systems, volume 33, 1877–1901.
  5. Improving Contrastive Learning of Sentence Embeddings from AI Feedback. In Findings of the Association for Computational Linguistics: ACL 2023, 11122–11138.
  6. Survey Analysis of Machine Learning Methods for Natural Language Processing for MBTI Personality Type Prediction. Technical report, Stanford University.
  7. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In Korhonen, A.; Traum, D.; and Màrquez, L., eds., Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2978–2988.
  8. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186.
  9. On Text-based Personality Computing: Challenges and Future Directions. In Findings of the Association for Computational Linguistics: ACL 2023, 10861–10879.
  10. SimCSE: Simple Contrastive Learning of Sentence Embeddings. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 6894–6910.
  11. PANDORA Talks: Personality and Demographics on Reddit. In Proceedings of the Ninth International Workshop on Natural Language Processing for Social Media, 138–152.
  12. Dimensionality Reduction by Learning an Invariant Mapping. In 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), volume 2, 1735–1742.
  13. Distilling the Knowledge in a Neural Network. arXiv:1503.02531.
  14. Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes. In Findings of the Association for Computational Linguistics: ACL 2023, 8003–8017.
  15. Weighted Distillation with Unlabeled Examples. In Koyejo, S.; Mohamed, S.; Agarwal, A.; Belgrave, D.; Cho, K.; and Oh, A., eds., Advances in Neural Information Processing Systems, volume 35, 7024–7037.
  16. Is ChatGPT a Good Personality Recognizer? A Preliminary Study. arXiv:2307.03952.
  17. Automatic Text-based Personality Recognition on Monologues and Multiparty Dialogues Using Attentive Networks and Contextual Embeddings. arXiv:1911.09304.
  18. Automatic Text-Based Personality Recognition on Monologues and Multiparty Dialogues Using Attentive Networks and Contextual Embeddings (Student Abstract). volume 34, 13821–13822.
  19. The Relationship Between Personality Traits and the Processing of Emotion Words: Evidence from Eye-Movements in Sentence Reading. Journal of Psycholinguistic Research, 1–27.
  20. Myers-Briggs Personality Classification and Personality-Specific Language Generation Using Pre-trained Language Models. arXiv:1907.06333.
  21. Construction of MBTI Personality Estimation Model Considering Emotional Information. In Proceedings of the 35th Pacific Asia Conference on Language, Information and Computation, 262–269. Shanghai, China.
  22. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv:1907.11692.
  23. Hierarchical Modeling for User Personality Prediction: The Role of Message-Level Attention. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 5306–5316.
  24. Training language models to follow instructions with human feedback. In Koyejo, S.; Mohamed, S.; Agarwal, A.; Belgrave, D.; Cho, K.; and Oh, A., eds., Advances in Neural Information Processing Systems, volume 35, 27730–27744.
  25. Personality, Gender, and Age in the Language of Social Media: The Open-Vocabulary Approach. PLOS ONE, 8(9): 1–16.
  26. A Survey of Automatic Personality Detection from Texts. In Proceedings of the 28th International Conference on Computational Linguistics, 6284–6295.
  27. ERNIE 3.0: Large-scale Knowledge Enhanced Pre-training for Language Understanding and Generation. arXiv:2107.02137.
  28. Personality Predictions Based on User Behavior on the Facebook Social Media Platform. IEEE Access, 6: 61959–61969.
  29. Personality Prediction System from Facebook Users. Procedia Computer Science, 116: 604–611. Discovery and innovation of computer science technology in artificial intelligence era: The 2nd International Conference on Computer Science and Computational Intelligence (ICCSCI 2017).
  30. The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods. Journal of Language and Social Psychology, 29(1): 24–54.
  31. Want To Reduce Labeling Cost? GPT-3 Can Help. In Findings of the Association for Computational Linguistics: EMNLP 2021, 4195–4205.
  32. Self-Instruct: Aligning Language Models with Self-Generated Instructions. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 13484–13508.
  33. Is ChatGPT a Good Sentiment Analyzer? A Preliminary Study. arXiv:2304.04339.
  34. Automatically Select Emotion for Response via Personality-affected Emotion Transition. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 5010–5020.
  35. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), 1112–1122.
  36. Deep learning-based personality recognition from text posts of online social networks. Applied Intelligence, 48.
  37. ConSERT: A Contrastive Framework for Self-Supervised Sentence Representation Transfer. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 5065–5075.
  38. Multi-Document Transformer for Personality Detection. volume 35, 14221–14229.
  39. Improving Dialog Systems for Negotiation with Personality Modeling. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 681–693.
  40. Orders Are Unwanted: Dynamic Deep Graph Convolutional Network for Personality Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, 13896–13904.
  41. Psycholinguistic Tripartite Graph Network for Personality Detection. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 4229–4239.
  42. Yarkoni, T. 2010. Personality in 100,000 Words: A large-scale analysis of personality and word use among bloggers. Journal of Research in Personality, 44(3): 363–373.
  43. OPT: Open Pre-trained Transformer Language Models. arXiv:2205.01068.
  44. Benchmarking Large Language Models for News Summarization. arXiv:2301.13848.
  45. Understanding bag-of-words model: a statistical framework. International Journal of Machine Learning and Cybernetics, 1: 43–52.
  46. Contrastive Graph Transformer Network for Personality Detection. In Raedt, L. D., ed., Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, 4559–4565.
Citations (7)

Summary

  • The paper presents an innovative method that uses LLMs to generate supplementary text analyses, overcoming data scarcity in personality detection.
  • The approach enriches personality labels with graded, detailed explanations, moving beyond traditional one-hot encoding for improved generalization.
  • Performance evaluations demonstrate significant gains in macro-F1 scores on benchmark datasets, validating the effectiveness of LLM-driven augmentation.

LLM Based Text Augmentation Enhanced Personality Detection Model

Personality detection from social media posts is the task of inferring personality traits based on written content. This task faces significant challenges, primarily due to the scarcity of labeled data and the nuanced nature of personality traits. The paper presents an innovative approach by leveraging LLMs to enhance the personality detection capabilities of smaller models through text augmentation.

Introduction to Personality Detection Challenges

Personality detection seeks to identify individual personality traits, such as those defined by the Myers-Briggs Type Indicator (MBTI), from social media texts. Traditional methods have relied heavily on manually crafted features derived from psychological models like LIWC. However, these methods are constrained by the limited availability of self-reported personality data and the oversimplification introduced by treating personality traits as one-hot vectors, ignoring their semantic richness.

Current approaches attempt to utilize pre-trained LLMs fine-tuned on the available data, but these tend to generate suboptimal post representations. The proposed model addresses these limitations by using LLMs not as direct predictors, but as knowledge bases to generate insightful text augmentations, enriching both data and labels in the personality detection task.

LLM-Driven Text Augmentation

The model takes advantage of LLMs to generate analyses that supplement the original social media posts. These analyses focus on key aspects: semantic content, sentiment dynamics, and linguistic style, all of which are known to correlate with personality traits.

The LLM is prompted to produce additional commentary on each post, creating a richer, multi-dimensional dataset that provides more nuanced signals for personality detection. This augmented dataset is used during training to enhance the embedding space, enabling more robust psycho-linguistic profiling. Notably, the LLM-derived augmentations do not incur additional computational cost at inference time. Figure 1

Figure 1: An example of personality detection using the LLM: The result provided by the LLM is ISFJ, while the actual ground truth is ENFP.

Enriching Personality Label Information

The paper also leverages LLMs to go beyond one-hot personality labels, which typically fail to capture the full complexity of personality dimensions. By generating detailed explanations for each of the MBTI traits from various psychological dimensions, the LLM provides enriched label information.

This approach results in a set of soft labels that offer a graded understanding of personality traits, addressing issues such as measurement error and improving model generalization. These enriched labels facilitate better alignment of user representations with personality traits during training.

Performance Evaluation

Experiments demonstrate that the proposed model, which the paper refers to as Text Augmentation Enhanced (TAE) personality detection, significantly outperforms baseline models on two benchmark datasets. The model's superior performance is attributed to the inclusion of LLM-generated semantic, sentiment, and linguistic text analyses, as well as the enriched personality labels. Figure 2

Figure 2: An overview of our TAE.

Ablation studies confirm the importance of each component, showing that each type of augmentation contributes to the final performance. Utilizing LLMs for both data augmentation and label information results in substantial improvements in macro-F1 scores across dimensions of the personality spectrum.

Conclusion

The paper introduces a robust framework that capitalizes on LLM capabilities for enhancing personality trait detection from text. By generating knowledgeable augmentations and enriching label semantics, the approach bridges the gap between limited datasets and the complex nature of personality traits. The results underscore the value of integrating LLMs in data augmentation for personality profiling tasks, paving the way for future research to explore other psychological traits and improve model explainability.

The findings suggest numerous paths for further development, such as using larger LLMs or different LLMs, including domain-specific knowledge to improve text augmentations, and extending this framework to cover additional personality dimensions and psychological attributes.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.