Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Meta-Learning for Effective Multi-task and Multilingual Modelling (2101.10368v3)

Published 25 Jan 2021 in cs.CL

Abstract: Natural language processing (NLP) tasks (e.g. question-answering in English) benefit from knowledge of other tasks (e.g. named entity recognition in English) and knowledge of other languages (e.g. question-answering in Spanish). Such shared representations are typically learned in isolation, either across tasks or across languages. In this work, we propose a meta-learning approach to learn the interactions between both tasks and languages. We also investigate the role of different sampling strategies used during meta-learning. We present experiments on five different tasks and six different languages from the XTREME multilingual benchmark dataset. Our meta-learned model clearly improves in performance compared to competitive baseline models that also include multi-task baselines. We also present zero-shot evaluations on unseen target languages to demonstrate the utility of our proposed model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ishan Tarunesh (5 papers)
  2. Sushil Khyalia (8 papers)
  3. Vishwajeet Kumar (23 papers)
  4. Ganesh Ramakrishnan (88 papers)
  5. Preethi Jyothi (51 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.