Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
153 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Double Correction Framework for Denoising Recommendation (2405.11272v3)

Published 18 May 2024 in cs.IR and cs.AI

Abstract: As its availability and generality in online services, implicit feedback is more commonly used in recommender systems. However, implicit feedback usually presents noisy samples in real-world recommendation scenarios (such as misclicks or non-preferential behaviors), which will affect precise user preference learning. To overcome the noisy samples problem, a popular solution is based on dropping noisy samples in the model training phase, which follows the observation that noisy samples have higher training losses than clean samples. Despite the effectiveness, we argue that this solution still has limits. (1) High training losses can result from model optimization instability or hard samples, not just noisy samples. (2) Completely dropping of noisy samples will aggravate the data sparsity, which lacks full data exploitation. To tackle the above limitations, we propose a Double Correction Framework for Denoising Recommendation (DCF), which contains two correction components from views of more precise sample dropping and avoiding more sparse data. In the sample dropping correction component, we use the loss value of the samples over time to determine whether it is noise or not, increasing dropping stability. Instead of averaging directly, we use the damping function to reduce the bias effect of outliers. Furthermore, due to the higher variance exhibited by hard samples, we derive a lower bound for the loss through concentration inequality to identify and reuse hard samples. In progressive label correction, we iteratively re-label highly deterministic noisy samples and retrain them to further improve performance. Finally, extensive experimental results on three datasets and four backbones demonstrate the effectiveness and generalization of our proposed framework.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. GoRec: A Generative Cold-start Recommendation Framework. (2023).
  2. Unified Representation Learning for Discrete Attribute Enhanced Completely Cold-Start Recommendation. IEEE Transactions on Big Data (2024), 1–12.
  3. Mitigating Recommendation Biases via Group-Alignment and Global-Uniformity in Representation Learning. ACM Transactions on Intelligent Systems and Technology (2024).
  4. Olivier Catoni. 2012. Challenging the empirical mean and empirical variance: a deviation study. In Annales de l’IHP Probabilités et statistiques, Vol. 48. 1148–1185.
  5. Denoising self-attentive sequential recommendation. In Proceedings of the 16th ACM Conference on Recommender Systems. 92–101.
  6. A generalized catoni’s m-estimator under finite α𝛼\alphaitalic_α-th moment assumption with α∈𝛼absent\alpha\initalic_α ∈(1, 2). Electronic Journal of Statistics 15, 2 (2021), 5523–5544.
  7. LCD: Adaptive Label Correction for Denoising Music Recommendation. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management. 3903–3907.
  8. Simplify and robustify negative sampling for implicit collaborative filtering. Advances in Neural Information Processing Systems 33 (2020), 1094–1105.
  9. Sampler design for bayesian personalized ranking by leveraging view data. IEEE transactions on knowledge and data engineering 33, 2 (2019), 667–681.
  10. Personalized ranking for non-uniformly sampled items. In Proceedings of KDD Cup 2011. PMLR, 231–247.
  11. Self-guided learning to denoise for robust recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 1412–1422.
  12. Automated Data Denoising for Recommendation. arXiv preprint arXiv:2305.07070 (2023).
  13. A loss curvature perspective on training instability in deep learning. arXiv preprint arXiv:2110.04369 (2021).
  14. Exponential and moment inequalities for U-statistics. In High Dimensional Probability II. Springer, 13–38.
  15. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval. 639–648.
  16. Neural collaborative filtering. In Proceedings of the 26th international conference on world wide web. 173–182.
  17. PU learning for matrix completion. In International conference on machine learning. PMLR, 2445–2453.
  18. Collaborative filtering for implicit feedback datasets. In 2008 Eighth IEEE international conference on data mining. Ieee, 263–272.
  19. Deep learning. nature 521, 7553 (2015), 436–444.
  20. Modeling user exposure in recommendation. In Proceedings of the 25th international conference on World Wide Web. 951–961.
  21. AutoDenoise: Automatic Data Instance Denoising for Recommendations. In Proceedings of the ACM Web Conference 2023. 1003–1011.
  22. Foundations of Machine Learning. (Aug 2012).
  23. One-class collaborative filtering. In 2008 Eighth IEEE international conference on data mining. IEEE, 502–511.
  24. On the Theories Behind Hard Negative Sampling for Recommendation. In Proceedings of the ACM Web Conference 2023. 812–822.
  25. Training region-based object detectors with online hard example mining. In Proceedings of the IEEE conference on computer vision and pattern recognition. 761–769.
  26. Collaborative-Enhanced Prediction of Spending on Newly Downloaded Mobile Games under Consumption Uncertainty. WWW2024, Industry Track (2024).
  27. Neighborhood-Enhanced Supervised Contrastive Learning for Collaborative Filtering. IEEE Transactions on Knowledge and Data Engineering (2023).
  28. Does Every Data Instance Matter? Enhancing Sequential Recommendation by Eliminating Unreliable Data.. In IJCAI. 1579–1585.
  29. Learning to denoise unreliable interactions for graph collaborative filtering. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 122–132.
  30. M2GRL: A multi-task multi-view graph representation learning framework for web-scale recommender systems. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining. 2349–2358.
  31. Denoised Self-Augmented Learning for Social Recommendation. arXiv preprint arXiv:2305.12685 (2023).
  32. Denoising implicit feedback for recommendation. In Proceedings of the 14th ACM international conference on web search and data mining. 373–381.
  33. Neural graph collaborative filtering. In Proceedings of the 42nd international ACM SIGIR conference on Research and development in Information Retrieval. 165–174.
  34. Intersectional Two-sided Fairness in Recommendation. In Proceedings of the ACM on Web Conference 2024. New York, NY, USA, 3609–3620.
  35. Unbiased Delayed Feedback Label Correction for Conversion Rate Prediction. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2456–2466.
  36. Learning robust recommenders through cross-model agreement. In Proceedings of the ACM Web Conference 2022. 2015–2025.
  37. Efficient Bi-Level Optimization for Recommendation Denoising. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2502–2511.
  38. Implicit feedbacks are not always favorable: Iterative relabeled one-class collaborative filtering against noisy interactions. In Proceedings of the 29th ACM International Conference on Multimedia. 3070–3078.
  39. Leveraging post-click feedback for content recommendations. In Proceedings of the 13th ACM Conference on Recommender Systems. 278–286.
  40. A neural influence diffusion model for social recommendation. In Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval. 235–244.
  41. Sample selection with uncertainty of losses for learning with noisy labels. arXiv preprint arXiv:2106.00445 (2021).
  42. Unsupervised deep embedding for clustering analysis. In International conference on machine learning. PMLR, 478–487.
  43. Improving Implicit Feedback-Based Recommendation through Multi-Behavior Alignment. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval. 932–941.
  44. IDVT: Interest-aware Denoising and View-guided Tuning for Social Recommendation. arXiv preprint arXiv:2308.15926 (2023).
  45. Hyperbolic Graph Learning for Social Recommendation. IEEE Transactions on Knowledge and Data Engineering (2023), 1–14. https://doi.org/10.1109/TKDE.2023.3343402
  46. Towards robust neural graph collaborative filtering via structure denoising and embedding perturbation. ACM Transactions on Information Systems 41, 3 (2023), 1–28.
  47. Hierarchical item inconsistency signal learning for sequence denoising in sequential recommendation. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management. 2508–2518.
  48. LightFR: Lightweight Federated Recommendation with Privacy-preserving Matrix Factorization. ACM Trans. Inf. Syst. 41, 4, Article 90 (mar 2023), 28 pages.
  49. Robust Recommender System: A Survey and Future Directions. arXiv preprint arXiv:2309.02057 (2023).
  50. Uncovering User Interest from Biased and Noised Watch Time in Video Recommendation. arXiv preprint arXiv:2308.08120 (2023).
  51. Improving top-N recommendation performance using missing data. Mathematical Problems in Engineering 2015 (2015).
  52. Knowledge-refined Denoising Network for Robust Recommendation. arXiv preprint arXiv:2304.14987 (2023).
Citations (6)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com