CHOPS: CHat with custOmer Profile Systems for Customer Service with LLMs (2404.01343v4)
Abstract: Businesses and software platforms are increasingly turning to LLMs such as GPT-3.5, GPT-4, GLM-3, and LLaMa-2 for chat assistance with file access or as reasoning agents for customer service. However, current LLM-based customer service models have limited integration with customer profiles and lack the operational capabilities necessary for effective service. Moreover, existing API integrations emphasize diversity over the precision and error avoidance essential in real-world customer service scenarios. To address these issues, we propose an LLM agent named CHOPS (CHat with custOmer Profile in existing System), designed to: (1) efficiently utilize existing databases or systems for accessing user information or interacting with these systems following existing guidelines; (2) provide accurate and reasonable responses or carry out required operations in the system while avoiding harmful operations; and (3) leverage a combination of small and large LLMs to achieve satisfying performance at a reasonable inference cost. We introduce a practical dataset, the CPHOS-dataset, which includes a database, guiding files, and QA pairs collected from CPHOS, an online platform that facilitates the organization of simulated Physics Olympiads for high school teachers and students. We have conducted extensive experiments to validate the performance of our proposed CHOPS architecture using the CPHOS-dataset, with the aim of demonstrating how LLMs can enhance or serve as alternatives to human customer service. Code for our proposed architecture and dataset can be found at {https://github.com/JingzheShi/CHOPS}.
- Customer interactions: Rocking them with llms. https://www.cmswire.com/customer-experience/customer-interactions-rocking-them-with-llms/, 2022.
- Pioneering a new era of automated customer service with large language models (llms). https://www.databricks.com/solutions/accelerators/llms-customer-service-and-support, 2022.
- Universal sentence encoder, 2018.
- Harrison Chase. LangChain, October 2022. URL https://github.com/langchain-ai/langchain.
- Benchmarking large language models in retrieval-augmented generation, 2023.
- Databricks. Databricks: Llms for customer service and support. https://www.databricks.com/solutions/accelerators/llms-customer-service-and-support.
- Exploring agent-based architectures for enhancing large language model performance on mathematical puzzles. In Proceedings of the International Conference on Artificial Intelligence and Natural Language Processing, pp. 123–130, 2023.
- Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 320–335, 2022.
- Chatdb: Augmenting llms with databases as their symbolic memory, 2023a.
- Chatdb: Augmenting llms with databases as their symbolic memory. arXiv preprint arXiv:2306.03901, 2023b.
- A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions, 2023.
- Triviaqa: A large scale distantly supervised challenge dataset for reading comprehension, 2017.
- Study and analysis of chat gpt and its impact on different fields of study. International Journal of Innovative Science and Research Technology, 8(3), 2023.
- Retrieval-augmented generation for knowledge-intensive nlp tasks, 2021.
- Julian McAuley. Recommender systems and personalization datasets. https://cseweb.ucsd.edu/~jmcauley/datasets.html#amazon_reviews.
- OpenAI. Chatgpt, 2022. URL https://openai.com/blog/chatgpt.
- OpenAI. Gpt-4 technical report, 2023.
- Gorilla: Large language model connected with massive apis. arXiv preprint arXiv:2305.15334, 2023.
- Toolllm: Facilitating large language models to master 16000+ real-world apis, 2023.
- SQuAD: 100,000+ questions for machine comprehension of text. In Jian Su, Kevin Duh, and Xavier Carreras (eds.), Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp. 2383–2392, Austin, Texas, November 2016. Association for Computational Linguistics. doi: 10.18653/v1/D16-1264. URL https://aclanthology.org/D16-1264.
- Toolalpaca: Generalized tool learning for language models with 3000 simulated cases, 2023.
- Llama 2: Open foundation and fine-tuned chat models, 2023.
- Bhaskar Tripathi. Pdf-gpt. https://github.com/bhaskatripathi/pdfGPT, 2023.
- Though Vector. Customer support on twitter. https://www.kaggle.com/datasets/thoughtvector/customer-support-on-twitter/data, 2018.
- Chain-of-thought prompting elicits reasoning in large language models, 2023.
- Large language models are better reasoners with self-verification, 2023.
- Jocelyn Wulf. Pioneering a new era of automated customer service with large language models (llms). https://medium.com/@jocwulf/, 2022.
- Tree of thoughts: Deliberate problem solving with large language models, 2023.
- Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414, 2022.
- Cumulative reasoning with large language models, 2023.