Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Simplify and Robustify Negative Sampling for Implicit Collaborative Filtering (2009.03376v1)

Published 7 Sep 2020 in cs.LG and cs.IR

Abstract: Negative sampling approaches are prevalent in implicit collaborative filtering for obtaining negative labels from massive unlabeled data. As two major concerns in negative sampling, efficiency and effectiveness are still not fully achieved by recent works that use complicate structures and overlook risk of false negative instances. In this paper, we first provide a novel understanding of negative instances by empirically observing that only a few instances are potentially important for model learning, and false negatives tend to have stable predictions over many training iterations. Above findings motivate us to simplify the model by sampling from designed memory that only stores a few important candidates and, more importantly, tackle the untouched false negative problem by favouring high-variance samples stored in memory, which achieves efficient sampling of true negatives with high-quality. Empirical results on two synthetic datasets and three real-world datasets demonstrate both robustness and superiorities of our negative sampling method.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jingtao Ding (50 papers)
  2. Yuhan Quan (6 papers)
  3. Quanming Yao (102 papers)
  4. Yong Li (628 papers)
  5. Depeng Jin (72 papers)
Citations (88)

Summary

We haven't generated a summary for this paper yet.