Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dual-Balancing for Multi-Task Learning (2308.12029v2)

Published 23 Aug 2023 in cs.LG and cs.AI

Abstract: Multi-task learning (MTL), a learning paradigm to learn multiple related tasks simultaneously, has achieved great success in various fields. However, task balancing problem remains a significant challenge in MTL, with the disparity in loss/gradient scales often leading to performance compromises. In this paper, we propose a Dual-Balancing Multi-Task Learning (DB-MTL) method to alleviate the task balancing problem from both loss and gradient perspectives. Specifically, DB-MTL ensures loss-scale balancing by performing a logarithm transformation on each task loss, and guarantees gradient-magnitude balancing via normalizing all task gradients to the same magnitude as the maximum gradient norm. Extensive experiments conducted on several benchmark datasets consistently demonstrate the state-of-the-art performance of DB-MTL.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Baijiong Lin (15 papers)
  2. Weisen Jiang (15 papers)
  3. Feiyang Ye (17 papers)
  4. Yu Zhang (1400 papers)
  5. Pengguang Chen (20 papers)
  6. Ying-Cong Chen (47 papers)
  7. Shu Liu (146 papers)
  8. James T. Kwok (65 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.