Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CHOPS: CHat with custOmer Profile Systems for Customer Service with LLMs (2404.01343v4)

Published 31 Mar 2024 in cs.CL and cs.AI

Abstract: Businesses and software platforms are increasingly turning to LLMs such as GPT-3.5, GPT-4, GLM-3, and LLaMa-2 for chat assistance with file access or as reasoning agents for customer service. However, current LLM-based customer service models have limited integration with customer profiles and lack the operational capabilities necessary for effective service. Moreover, existing API integrations emphasize diversity over the precision and error avoidance essential in real-world customer service scenarios. To address these issues, we propose an LLM agent named CHOPS (CHat with custOmer Profile in existing System), designed to: (1) efficiently utilize existing databases or systems for accessing user information or interacting with these systems following existing guidelines; (2) provide accurate and reasonable responses or carry out required operations in the system while avoiding harmful operations; and (3) leverage a combination of small and large LLMs to achieve satisfying performance at a reasonable inference cost. We introduce a practical dataset, the CPHOS-dataset, which includes a database, guiding files, and QA pairs collected from CPHOS, an online platform that facilitates the organization of simulated Physics Olympiads for high school teachers and students. We have conducted extensive experiments to validate the performance of our proposed CHOPS architecture using the CPHOS-dataset, with the aim of demonstrating how LLMs can enhance or serve as alternatives to human customer service. Code for our proposed architecture and dataset can be found at {https://github.com/JingzheShi/CHOPS}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Customer interactions: Rocking them with llms. https://www.cmswire.com/customer-experience/customer-interactions-rocking-them-with-llms/, 2022.
  2. Pioneering a new era of automated customer service with large language models (llms). https://www.databricks.com/solutions/accelerators/llms-customer-service-and-support, 2022.
  3. Universal sentence encoder, 2018.
  4. Harrison Chase. LangChain, October 2022. URL https://github.com/langchain-ai/langchain.
  5. Benchmarking large language models in retrieval-augmented generation, 2023.
  6. Databricks. Databricks: Llms for customer service and support. https://www.databricks.com/solutions/accelerators/llms-customer-service-and-support.
  7. Exploring agent-based architectures for enhancing large language model performance on mathematical puzzles. In Proceedings of the International Conference on Artificial Intelligence and Natural Language Processing, pp.  123–130, 2023.
  8. Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp.  320–335, 2022.
  9. Chatdb: Augmenting llms with databases as their symbolic memory, 2023a.
  10. Chatdb: Augmenting llms with databases as their symbolic memory. arXiv preprint arXiv:2306.03901, 2023b.
  11. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions, 2023.
  12. Triviaqa: A large scale distantly supervised challenge dataset for reading comprehension, 2017.
  13. Study and analysis of chat gpt and its impact on different fields of study. International Journal of Innovative Science and Research Technology, 8(3), 2023.
  14. Retrieval-augmented generation for knowledge-intensive nlp tasks, 2021.
  15. Julian McAuley. Recommender systems and personalization datasets. https://cseweb.ucsd.edu/~jmcauley/datasets.html#amazon_reviews.
  16. OpenAI. Chatgpt, 2022. URL https://openai.com/blog/chatgpt.
  17. OpenAI. Gpt-4 technical report, 2023.
  18. Gorilla: Large language model connected with massive apis. arXiv preprint arXiv:2305.15334, 2023.
  19. Toolllm: Facilitating large language models to master 16000+ real-world apis, 2023.
  20. SQuAD: 100,000+ questions for machine comprehension of text. In Jian Su, Kevin Duh, and Xavier Carreras (eds.), Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp.  2383–2392, Austin, Texas, November 2016. Association for Computational Linguistics. doi: 10.18653/v1/D16-1264. URL https://aclanthology.org/D16-1264.
  21. Toolalpaca: Generalized tool learning for language models with 3000 simulated cases, 2023.
  22. Llama 2: Open foundation and fine-tuned chat models, 2023.
  23. Bhaskar Tripathi. Pdf-gpt. https://github.com/bhaskatripathi/pdfGPT, 2023.
  24. Though Vector. Customer support on twitter. https://www.kaggle.com/datasets/thoughtvector/customer-support-on-twitter/data, 2018.
  25. Chain-of-thought prompting elicits reasoning in large language models, 2023.
  26. Large language models are better reasoners with self-verification, 2023.
  27. Jocelyn Wulf. Pioneering a new era of automated customer service with large language models (llms). https://medium.com/@jocwulf/, 2022.
  28. Tree of thoughts: Deliberate problem solving with large language models, 2023.
  29. Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414, 2022.
  30. Cumulative reasoning with large language models, 2023.
Citations (2)

Summary

  • The paper presents the CHOPS framework that integrates LLMs with customer profiles to achieve context-aware customer service.
  • It introduces a Classifier-Executor-Verifier architecture that optimizes query processing and ensures response accuracy.
  • Experimental results on the CPHOS-dataset demonstrate improved cost-effectiveness and reliability in sensitive customer service settings.

The paper "CHOPS: CHat with custOmer Profile Systems for Customer Service with LLMs" introduces a novel framework leveraging LLMs to enhance customer service automation capabilities by effectively integrating these models with customer profile data and existing operational systems. The authors propose a system named CHOPS, which stands for Chat with custOmer Profile in existing System. This system is designed to address limitations in current LLM applications that do not sufficiently incorporate customer profiles or operational integration capabilities necessary for precise and error-free interactions in customer service contexts.

Key Contributions:

  1. Integration with Customer Profiles: CHOPS was developed to efficiently utilize existing databases and systems, ensuring it can access user information or interact with these systems based on predefined guidelines. By doing so, it aims to provide accurate and contextually relevant responses while minimizing the risk of executing harmful or incorrect operations.
  2. Architecture Design:
    • The CHOPS framework employs a Classifier-Executor-Verifier architecture to enhance interaction accuracy and control inference costs.
    • Classifier: Determines whether a user query relates to guide files, API usage, or requires basic user information.
    • Executor: Processes the user query using the extracted information to generate responses or execute operations.
    • Verifier: Checks the executor's responses to ensure they are correct before finalizing the interaction with users, a structure inspired by previous advancements in self-verification methods to improve model reliability.
  3. Focus on API Integration: Unlike previous approaches focused on widely diverse API sets, CHOPS narrows down the utilization to critical APIs, thereby prioritizing high accuracy in user status modifications, particularly in sensitive sectors like banking. This approach enlightens the need for specific customer service datasets that reflect real-world scenarios more accurately.
  4. CPHOS-dataset: To facilitate thorough evaluation, the researchers introduce a novel dataset named CPHOS, derived from an online platform used to simulate Physics Olympiads. The dataset includes databases, guiding files, and collected question-answer pairs. This practical dataset aims to bridge existing gaps by simulating real interactions requiring database queries or the utilization of guiding files.

Experimental Validation:

The authors report extensive experimentation demonstrating the efficacy of the CHOPS architecture. Testing with the CPHOS-dataset, they claim that their architecture can achieve substantial improvements in customer service accuracy by outperforming a naive use of more powerful LLMs in terms of both cost-effectiveness and reliability. Specifically, performance improvements were consistent even when leveraging smaller model variants due to the architectural specialization of CHOPS.

Technical Details:

  • Use of LLMs: LLMs like GPT-3.5, GPT-4, GLM-3, and LLaMa-2 are exploited for their ability to generate natural language responses. The integration emphasizes blending model strengths across tasks, balancing performance and inference costs.
  • RAG Framework: Retrieval-Augmented Generation (RAG) techniques are employed to incorporate external knowledge sources into LLM responses efficiently, aligning with vector-based data retrieval systems.

Conclusion:

The framework proposes a refined method to integrate LLMs into practical customer service systems effectively. The results suggest that the CHOPS architecture not only enhances interaction accuracy and cost efficiency but also establishes a flexible design paradigm proficient at accommodating LLMs within operational settings that require interaction with both data-driven APIs and directive guide documents. The development and usage of the CPHOS-dataset further illustrate the framework’s applicability in practical business scenarios, laying a prospective pathway for future customer service automation tools that integrate advanced LLMs.

X Twitter Logo Streamline Icon: https://streamlinehq.com