Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Cross-lingual Spoken Language Understanding with Regularized Representation Alignment (2009.14510v1)

Published 30 Sep 2020 in cs.CL and cs.LG

Abstract: Despite the promising results of current cross-lingual models for spoken language understanding systems, they still suffer from imperfect cross-lingual representation alignments between the source and target languages, which makes the performance sub-optimal. To cope with this issue, we propose a regularization approach to further align word-level and sentence-level representations across languages without any external resource. First, we regularize the representation of user utterances based on their corresponding labels. Second, we regularize the latent variable model (Liu et al., 2019) by leveraging adversarial training to disentangle the latent variables. Experiments on the cross-lingual spoken language understanding task show that our model outperforms current state-of-the-art methods in both few-shot and zero-shot scenarios, and our model, trained on a few-shot setting with only 3\% of the target language training data, achieves comparable performance to the supervised training with all the training data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zihan Liu (102 papers)
  2. Genta Indra Winata (94 papers)
  3. Peng Xu (357 papers)
  4. Zhaojiang Lin (45 papers)
  5. Pascale Fung (151 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.