Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
91 tokens/sec
GPT-4o
12 tokens/sec
Gemini 2.5 Pro Pro
o3 Pro
5 tokens/sec
GPT-4.1 Pro
15 tokens/sec
DeepSeek R1 via Azure Pro
33 tokens/sec
Gemini 2.5 Flash Deprecated
12 tokens/sec
2000 character limit reached

User Embedding Model for Personalized Language Prompting (2401.04858v1)

Published 10 Jan 2024 in cs.CL, cs.AI, cs.IR, and cs.LG

Abstract: Modeling long histories plays a pivotal role in enhancing recommendation systems, allowing to capture user's evolving preferences, resulting in more precise and personalized recommendations. In this study we tackle the challenges of modeling long user histories for preference understanding in natural language. Specifically, we introduce a new User Embedding Module (UEM) that efficiently processes user history in free-form text by compressing and representing them as embeddings, to use them as soft prompts to a LM. Our experiments demonstrate the superior capability of this approach in handling significantly longer histories compared to conventional text based prompting methods, yielding substantial improvements in predictive performance. The main contribution of this research is to demonstrate the ability to bias LLMs with user signals represented as embeddings.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. Palm 2 technical report. CoRR, abs/2305.10403.
  2. Sparks of artificial general intelligence: Early experiments with GPT-4. CoRR, abs/2303.12712.
  3. Scaling instruction-finetuned language models. CoRR, abs/2210.11416.
  4. LongT5: Efficient text-to-text transformer for long sequences. In Findings of the Association for Computational Linguistics: NAACL 2022, pages 724–736, Seattle, United States. Association for Computational Linguistics.
  5. F. Maxwell Harper and Joseph A. Konstan. 2016. The movielens datasets: History and context. ACM Trans. Interact. Intell. Syst., 5(4):19:1–19:19.
  6. Lora: Low-rank adaptation of large language models. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net.
  7. Pratyush Kumar. 2023. Large language models humanize technology. CoRR, abs/2305.05576.
  8. The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  9. Automatic prompt rewriting for personalized text generation.
  10. Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, pages 4582–4597. Association for Computational Linguistics.
  11. Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
  12. Useridentifier: Implicit user representations for simple and effective personalized sentiment analysis. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, Seattle, WA, United States, July 10-15, 2022, pages 3449–3456. Association for Computational Linguistics.
  13. Cross-task generalization via natural language crowdsourcing instructions. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, pages 3470–3487. Association for Computational Linguistics.
  14. Learning to compress prompts with gist tokens.
  15. Sentence-t5: Scalable sentence encoders from pre-trained text-to-text models. In Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, May 22-27, 2022, pages 1864–1874. Association for Computational Linguistics.
  16. OpenAI. 2023. GPT-4 technical report. CoRR, abs/2303.08774.
  17. Training language models to follow instructions with human feedback. In NeurIPS.
  18. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, July 6-12, 2002, Philadelphia, PA, USA, pages 311–318. ACL.
  19. Train short, test long: Attention with linear biases enables input length extrapolation. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net.
  20. Reasoning with language model prompting: A survey. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5368–5393, Toronto, Canada. Association for Computational Linguistics.
  21. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 21:140:1–140:67.
  22. COMET: A neural framework for MT evaluation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, pages 2685–2702. Association for Computational Linguistics.
  23. Lamp: When large language models meet personalization. CoRR, abs/2304.11406.
  24. Multitask prompted training enables zero-shot task generalization. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net.
  25. Human language modeling. In Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, May 22-27, 2022, pages 622–636. Association for Computational Linguistics.
  26. Roformer: Enhanced transformer with rotary position embedding. CoRR, abs/2104.09864.
  27. Llama 2: Open foundation and fine-tuned chat models. CoRR, abs/2307.09288.
  28. Finetuned language models are zero-shot learners. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net.
  29. Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS.
  30. Retrieval-augmented multimodal language modeling. In International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pages 39755–39769. PMLR.
  31. Useradapter: Few-shot user learning in sentiment analysis. In Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, Online Event, August 1-6, 2021, volume ACL/IJCNLP 2021 of Findings of ACL, pages 1484–1488. Association for Computational Linguistics.
Citations (6)

Summary

  • The paper presents a novel User Embedding Module (UEM) that captures extended user histories as compact embeddings.
  • It leverages personalized soft prompts to tailor language model outputs without high computational costs.
  • Empirical results demonstrate up to a 0.25 improvement in F1 score over baseline methods in text-to-text tasks.

Introduction to User Embedding for LLMs

LLMs (LMs) have remarkably transformed the way artificial intelligence systems understand and generate human language. Their applications span from chatbots and recommendation systems to more complex tasks, such as reasoning and creativity assistance. These models have largely benefited from the mass of data available on the internet, which has taught them to adapt rapidly and generalize across a wide range of contexts and questions. However, personalization has remained a nuanced challenge, particularly when it comes to understanding and integrating long user histories into the responses and suggestions these models provide.

Advancements in Personalized Language Prompting

The primary innovation presented in the discussed research is the User Embedding Module (UEM), a technique designed to capture and represent extended user interaction histories as compact embeddings. These embeddings can be used as personalized soft prompts for LMs, thereby influencing the model's outputs to align better with an individual's preferences and behavior. This approach is markedly more efficient than previous methods, which often involved sifting through shorter segments of user history or various prompt rewriting techniques. In contrast, the UEM is capable of capturing an entire user's history without the prohibitive computational costs typically associated with larger data sequences.

Empirical Findings and Methodology

The research deployed the UEM within a text-to-text generation framework, where tasks are framed as language generation conditioned on input. Their empirical results highlighted the efficiency of UEM, showing a considerable improvement in predictive performance, with up to a 0.25 improvement in F1 scores over baseline methods for handling user history. The user history was represented as various embeddings covering aspects like movie titles, genres, ratings, and descriptions, subsequently processed through the UEM. This enables the LM, coupled with UEM, to handle user histories with more depth — up to 50 items in this experiment — significantly surpassing the limitations of traditional text-based LMs.

Looking Forward

The findings of this paper have immediate implications for enhancing the personalization of LMs in various applications, as seen in performance improvements when the UEM is implemented. In the future, exploring different ways to fine-tune such LMs using parameter-efficient techniques or extending these methods to multimodal signals holds potential for further advancements.

In conclusion, this research successfully takes a stride towards integrating longer user histories into LMs by representing these histories with an embedding-based approach. The approach not only enhances the performance of LMs in predicting user preferences but does so with computational efficiency, raising the possibility for more personalized, context-aware AI applications in the near future.