Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KeyVec: Key-semantics Preserving Document Representations (1709.09749v1)

Published 27 Sep 2017 in cs.CL, cs.LG, and cs.NE

Abstract: Previous studies have demonstrated the empirical success of word embeddings in various applications. In this paper, we investigate the problem of learning distributed representations for text documents which many machine learning algorithms take as input for a number of NLP tasks. We propose a neural network model, KeyVec, which learns document representations with the goal of preserving key semantics of the input text. It enables the learned low-dimensional vectors to retain the topics and important information from the documents that will flow to downstream tasks. Our empirical evaluations show the superior quality of KeyVec representations in two different document understanding tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Bin Bi (24 papers)
  2. Hao Ma (116 papers)

Summary

We haven't generated a summary for this paper yet.