Neeko: Leveraging Dynamic LoRA for Efficient Multi-Character Role-Playing Agent (2402.13717v3)
Abstract: LLMs have revolutionized open-domain dialogue agents but encounter challenges in multi-character role-playing (MCRP) scenarios. To address the issue, we present Neeko, an innovative framework designed for efficient multiple characters imitation. Unlike existing methods, Neeko employs a dynamic low-rank adapter (LoRA) strategy, enabling it to adapt seamlessly to diverse characters. Our framework breaks down the role-playing process into agent pre-training, multiple characters playing, and character incremental learning, effectively handling both seen and unseen roles. This dynamic approach, coupled with distinct LoRA blocks for each character, enhances Neeko's adaptability to unique attributes, personalities, and speaking patterns. As a result, Neeko demonstrates superior performance in MCRP over most existing methods, offering more engaging and versatile user interaction experiences. Code and data are available at https://github.com/weiyifan1023/Neeko.
- Intrinsic dimensionality explains the effectiveness of language model fine-tuning. arXiv preprint arXiv:2012.13255.
- Large language models meet harry potter: A dataset for aligning dialogue agents with characters. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 8506–8520.
- Lifelong language pretraining with distribution-specialized experts. In International Conference on Machine Learning, pages 5383–5395. PMLR.
- Adanet: Adaptive structural learning of artificial neural networks. In International conference on machine learning, pages 874–883. PMLR.
- Learning factored representations in a deep mixture of experts. arXiv preprint arXiv:1312.4314.
- Transformer feed-forward layers are key-value memories. arXiv preprint arXiv:2012.14913.
- Meet your favorite character: Open-domain chatbot mimicking fictional characters with only a few utterances. arXiv preprint arXiv:2204.10825.
- Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
- Lorahub: Efficient cross-task generalization via dynamic lora composition. arXiv preprint arXiv:2307.13269.
- Chatharuhi: Reviving anime character in reality via large language model. arXiv preprint arXiv:2308.09597.
- Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81.
- Moelora: An moe-based parameter efficient fine-tuning method for multi-task medical applications. arXiv preprint arXiv:2310.18339.
- Moelora: Contrastive learning guided mixture of experts on parameter-efficient fine-tuning for large language models.
- Unsupervised enrichment of persona-grounded dialog with background stories. arXiv preprint arXiv:2106.08364.
- OpenAI. 2023. Chatgpt: Optimizing language models for dialogue.
- Character-llm: A trainable agent for role-playing. arXiv preprint arXiv:2310.10158.
- Roleeval: A bilingual role evaluation benchmark for large language models. arXiv preprint arXiv:2312.16132.
- Telling stories through multi-user dialogue by modeling character relations. arXiv preprint arXiv:2105.15054.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
- Charactereval: A chinese benchmark for role-playing conversational agent evaluation. arXiv preprint arXiv:2401.01275.
- Dylora: Parameter-efficient tuning of pre-trained models using dynamic search-free low-rank adaptation. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 3266–3279.
- Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models. arXiv preprint arXiv:2310.00746.
- Assessing knowledge editing in language models via relation perspective. arXiv preprint arXiv:2311.09053.
- Fei Ye and Adrian G Bors. 2023. Lifelong compression mixture model via knowledge relationship graph. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 10900–10908.
- Melo: Enhancing model editing with neuron-indexed dynamic lora. arXiv preprint arXiv:2312.11795.
- Spiking generative networks empowered by multiple dynamic experts for lifelong learning. Expert Systems with Applications, 238:121845.
- Narrativeplay: Interactive narrative understanding. arXiv preprint arXiv:2310.01459.
- Characterglm: Customizing chinese conversational ai characters with large language models. arXiv preprint arXiv:2311.16832.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.