Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SkillNet-X: A Multilingual Multitask Model with Sparsely Activated Skills (2306.16176v1)

Published 28 Jun 2023 in cs.CL

Abstract: Traditional multitask learning methods basically can only exploit common knowledge in task- or language-wise, which lose either cross-language or cross-task knowledge. This paper proposes a general multilingual multitask model, named SkillNet-X, which enables a single model to tackle many different tasks from different languages. To this end, we define several language-specific skills and task-specific skills, each of which corresponds to a skill module. SkillNet-X sparsely activates parts of the skill modules which are relevant either to the target task or the target language. Acting as knowledge transit hubs, skill modules are capable of absorbing task-related knowledge and language-related knowledge consecutively. Based on Transformer, we modify the multi-head attention layer and the feed forward network layer to accommodate skill modules. We evaluate SkillNet-X on eleven natural language understanding datasets in four languages. Results show that SkillNet-X performs better than task-specific baselines and two multitask learning baselines (i.e., dense joint model and Mixture-of-Experts model). Furthermore, skill pre-training further improves the performance of SkillNet-X on almost all datasets. To investigate the generalization of our model, we conduct experiments on two new tasks and find that SkillNet-X significantly outperforms baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Zhangyin Feng (14 papers)
  2. Yong Dai (33 papers)
  3. Fan Zhang (686 papers)
  4. Duyu Tang (65 papers)
  5. Xiaocheng Feng (54 papers)
  6. Shuangzhi Wu (29 papers)
  7. Bing Qin (186 papers)
  8. Yunbo Cao (43 papers)
  9. Shuming Shi (126 papers)

Summary

We haven't generated a summary for this paper yet.