Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Post-Processing of Word Representations via Variance Normalization and Dynamic Embedding (1808.06305v3)

Published 20 Aug 2018 in cs.CL

Abstract: Although embedded vector representations of words offer impressive performance on many NLP applications, the information of ordered input sequences is lost to some extent if only context-based samples are used in the training. For further performance improvement, two new post-processing techniques, called post-processing via variance normalization (PVN) and post-processing via dynamic embedding (PDE), are proposed in this work. The PVN method normalizes the variance of principal components of word vectors while the PDE method learns orthogonal latent variables from ordered input sequences. The PVN and the PDE methods can be integrated to achieve better performance. We apply these post-processing techniques to two popular word embedding methods (i.e., word2vec and GloVe) to yield their post-processed representations. Extensive experiments are conducted to demonstrate the effectiveness of the proposed post-processing techniques.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Bin Wang (751 papers)
  2. Fenxiao Chen (5 papers)
  3. Angela Wang (7 papers)
  4. C. -C. Jay Kuo (177 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.