Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Handling Long-Tail Queries with Slice-Aware Conversational Systems (2104.13216v1)

Published 26 Apr 2021 in cs.LG and cs.AI

Abstract: We have been witnessing the usefulness of conversational AI systems such as Siri and Alexa, directly impacting our daily lives. These systems normally rely on machine learning models evolving over time to provide quality user experience. However, the development and improvement of the models are challenging because they need to support both high (head) and low (tail) usage scenarios, requiring fine-grained modeling strategies for specific data subsets or slices. In this paper, we explore the recent concept of slice-based learning (SBL) (Chen et al., 2019) to improve our baseline conversational skill routing system on the tail yet critical query traffic. We first define a set of labeling functions to generate weak supervision data for the tail intents. We then extend the baseline model towards a slice-aware architecture, which monitors and improves the model performance on the selected tail intents. Applied to de-identified live traffic from a commercial conversational AI system, our experiments show that the slice-aware model is beneficial in improving model performance for the tail intents while maintaining the overall performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Cheng Wang (386 papers)
  2. Sun Kim (26 papers)
  3. Taiwoo Park (4 papers)
  4. Sajal Choudhary (4 papers)
  5. Sunghyun Park (38 papers)
  6. Young-Bum Kim (22 papers)
  7. Ruhi Sarikaya (16 papers)
  8. Sungjin Lee (46 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.