Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Same Representation, Different Attentions: Shareable Sentence Representation Learning from Multiple Tasks (1804.08139v1)

Published 22 Apr 2018 in cs.CL and cs.AI

Abstract: Distributed representation plays an important role in deep learning based natural language processing. However, the representation of a sentence often varies in different tasks, which is usually learned from scratch and suffers from the limited amounts of training data. In this paper, we claim that a good sentence representation should be invariant and can benefit the various subsequent tasks. To achieve this purpose, we propose a new scheme of information sharing for multi-task learning. More specifically, all tasks share the same sentence representation and each task can select the task-specific information from the shared sentence representation with attention mechanism. The query vector of each task's attention could be either static parameters or generated dynamically. We conduct extensive experiments on 16 different text classification tasks, which demonstrate the benefits of our architecture.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Renjie Zheng (29 papers)
  2. Junkun Chen (27 papers)
  3. Xipeng Qiu (257 papers)
Citations (30)

Summary

We haven't generated a summary for this paper yet.