Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DORIC : Domain Robust Fine-Tuning for Open Intent Clustering through Dependency Parsing (2303.09827v1)

Published 17 Mar 2023 in cs.CL and cs.AI

Abstract: We present our work on Track 2 in the Dialog System Technology Challenges 11 (DSTC11). DSTC11-Track2 aims to provide a benchmark for zero-shot, cross-domain, intent-set induction. In the absence of in-domain training dataset, robust utterance representation that can be used across domains is necessary to induce users' intentions. To achieve this, we leveraged a multi-domain dialogue dataset to fine-tune the LLM and proposed extracting Verb-Object pairs to remove the artifacts of unnecessary information. Furthermore, we devised the method that generates each cluster's name for the explainability of clustered results. Our approach achieved 3rd place in the precision score and showed superior accuracy and normalized mutual information (NMI) score than the baseline model on various domain datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jihyun Lee (25 papers)
  2. Seungyeon Seo (4 papers)
  3. Yunsu Kim (40 papers)
  4. Gary Geunbae Lee (53 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.