Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Importance of Word Order Information in Cross-lingual Sequence Labeling (2001.11164v4)

Published 30 Jan 2020 in cs.CL

Abstract: Word order variances generally exist in different languages. In this paper, we hypothesize that cross-lingual models that fit into the word order of the source language might fail to handle target languages. To verify this hypothesis, we investigate whether making models insensitive to the word order of the source language can improve the adaptation performance in target languages. To do so, we reduce the source language word order information fitted to sequence encoders and observe the performance changes. In addition, based on this hypothesis, we propose a new method for fine-tuning multilingual BERT in downstream cross-lingual sequence labeling tasks. Experimental results on dialogue natural language understanding, part-of-speech tagging, and named entity recognition tasks show that reducing word order information fitted to the model can achieve better zero-shot cross-lingual performance. Furthermore, our proposed methods can also be applied to strong cross-lingual baselines, and improve their performances.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zihan Liu (102 papers)
  2. Genta Indra Winata (94 papers)
  3. Samuel Cahyawijaya (75 papers)
  4. Andrea Madotto (64 papers)
  5. Zhaojiang Lin (45 papers)
  6. Pascale Fung (150 papers)
Citations (3)