TensorCoder: Dimension-Wise Attention via Tensor Representation for Natural Language Modeling (2008.01547v2)
Abstract: Transformer has been widely-used in many NLP tasks and the scaled dot-product attention between tokens is a core module of Transformer. This attention is a token-wise design and its complexity is quadratic to the length of sequence, limiting its application potential for long sequence tasks. In this paper, we propose a dimension-wise attention mechanism based on which a novel LLMing approach (namely TensorCoder) can be developed. The dimension-wise attention can reduce the attention complexity from the original $O(N2d)$ to $O(Nd2)$, where $N$ is the length of the sequence and $d$ is the dimensionality of head. We verify TensorCoder on two tasks including masked LLMing and neural machine translation. Compared with the original Transformer, TensorCoder not only greatly reduces the calculation of the original model but also obtains improved performance on masked LLMing task (in PTB dataset) and comparable performance on machine translation tasks.
- Shuai Zhang (319 papers)
- Peng Zhang (641 papers)
- Xindian Ma (6 papers)
- Junqiu Wei (4 papers)
- Ningning Wang (28 papers)
- Qun Liu (230 papers)