Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Label-aware Multi-level Contrastive Learning for Cross-lingual Spoken Language Understanding (2205.03656v2)

Published 7 May 2022 in cs.CL and cs.AI

Abstract: Despite the great success of spoken language understanding (SLU) in high-resource languages, it remains challenging in low-resource languages mainly due to the lack of labeled training data. The recent multilingual code-switching approach achieves better alignments of model representations across languages by constructing a mixed-language context in zero-shot cross-lingual SLU. However, current code-switching methods are limited to implicit alignment and disregard the inherent semantic structure in SLU, i.e., the hierarchical inclusion of utterances, slots, and words. In this paper, we propose to model the utterance-slot-word structure by a multi-level contrastive learning framework at the utterance, slot, and word levels to facilitate explicit alignment. Novel code-switching schemes are introduced to generate hard negative examples for our contrastive learning framework. Furthermore, we develop a label-aware joint model leveraging label semantics to enhance the implicit alignment and feed to contrastive learning. Our experimental results show that our proposed methods significantly improve the performance compared with the strong baselines on two zero-shot cross-lingual SLU benchmark datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shining Liang (9 papers)
  2. Linjun Shou (53 papers)
  3. Jian Pei (104 papers)
  4. Ming Gong (246 papers)
  5. Wanli Zuo (7 papers)
  6. Xianglin Zuo (3 papers)
  7. Daxin Jiang (138 papers)
Citations (7)