ICDM: Debiasing ID-based CF Models
- ICDM is a module that disentangles true collaborative signals from popularity and conformity biases in ID-based CF models.
- It employs similarity and orthogonality losses to separate debiased representations from biased factors for improved alignment.
- Integrated in the DAS framework, ICDM leverages multi-view contrastive learning to robustly align collaborative and semantic signals.
The ID-based Collaborative Filtering Debias Module (ICDM) is a key component in modern large-scale recommender systems that addresses entrenched biases in ID-based collaborative filtering (CF) representations. Implemented within the single-stage Dual-Aligned Semantic IDs (DAS) framework, ICDM is designed to disentangle signal and bias in user and item representations, producing debiased embeddings for contrastive alignment with semantic code-based identifiers, thereby improving downstream recommendation performance and preserving semantic integrity (Ye et al., 14 Aug 2025).
1. Purpose and Placement of ICDM
ICDM targets two pervasive forms of bias in CF models: item popularity bias (a small subset of items dominating clicks) and user conformity bias (users disproportionately clicking on popular items). These biases distort learned embeddings and, without correction, can pollute the alignment with semantic identifiers derived from multi-modal LLM (MLLM) features. ICDM operates between the User & Item Semantic Model (which outputs quantized, codebook-based semantic IDs ) and the downstream Multi-view Dual-Aligned Mechanism (MDAM), which aligns collaborative and semantic signals.
The module splits each user or item ID-based representation into a “true” (signal) factor and a “bias” factor, enforces orthogonality and similarity constraints for disentanglement, and utilizes only the debiased factors in the downstream alignment. This ensures that the semantic code space benefits from collaborative cues without inheriting distortion from CF biases.
2. Formal Objectives and Loss Functions
Let denote raw CF embeddings for users and items. Four multi-layer perceptron (MLP) encoders decompose these:
- User-Interest (debiased):
- User-Conformity (bias):
- Item-Content (debiased):
- Item-Popularity (bias):
Biased representations are constructed as and .
Disentangling Losses:
Two constraints are enforced:
- Similarity Penalty
- Orthogonality Penalty
The combined disentangling loss:
0
Supervised CF Objectives:
- Biased CF Loss:
1
- Unbiased CF Loss:
2
The total ICDM loss:
3
with 4 controlling the tradeoff.
3. Multi-view Contrastive Alignment and Mutual Information Framework
ICDM supplies debiased CF signals that are contrastively aligned with Semantic IDs across six cross-views using InfoNCE objectives, which lower-bound mutual information between respective pairs:
- Dual User-to-Item:
- 5: 6
- 7: 8
- Dual User-to-User / Item-to-Item:
- 9: 0
- 1: 2
- Dual Co-occurrence:
- User co-click 3: 4
- Item co-click 5: 6
Total alignment loss:
7
Each term encourages mutual information maximization between semantic and collaborative representations, facilitating more robust multimodal recommendation.
4. Joint Training Objective and Unified Optimization
ICDM operates within a single-stage end-to-end system. The complete loss integrates semantic quantization, ICDM, and multi-view alignment:
8
with hyperparameters 9 tuning the relative importance of each sub-module.
All modules—including RQ-VAE codebooks for semantic vector quantization, four MLPs for factor disentanglement, and the contrastive aligner—are optimized jointly via AdamW. The one-stage nature of DAS avoids information loss and inflexibility associated with prior two-stage approaches.
5. Implementation Workflow
A high-level pseudocode outline for a training iteration is as follows:
4
6. Hyperparameter Scheme, Ablations, and Empirical Findings
Key Hyperparameters
| Module | Value(s) | Notes |
|---|---|---|
| Semantic quantization | 0, 1 | 3 codebooks, 512 codes/level, dim=32 |
| ICDM: 2, 3 | 4, 5 | Disentangle term, UISM-ICDM loss balance |
| Alignment: 6, 7 | 8, 9 | InfoNCE temperature, co-occ. bank 0 |
| Optimization | AdamW, lr 1 | batch 2, 163A100 GPUs |
Offline Ablations
Incremental ablations highlight that inclusion of ICDM improves AUC from 0.8048 to 0.8050 and GAUC from 0.7433 to 0.7434. Removal of any dual-u2i, dual-i2i/u2u, or co-occurrence terms each reduces AUC by 0.4–0.6 ‰, indicating the necessity of multi-view contrastive mechanisms for optimal information integration.
Online Performance
Deployment results show a +3.48% overall increase in eCPM, with +8.98% in cold-start user segments. This suggests effective transfer of collaborative signals to semantic ID space and improved generalizability in data-sparse regimes.
7. Significance and Role within Industrial Recommender Systems
ICDM functions as a debiasing intermediary that ensures collaborative signals can be contrastively integrated with semantic codes free from the confounding effects of popularity and conformity. In one-stage frameworks such as DAS, this enables simultaneous optimization of semantic quantization, collaborative filtering, and contrastive alignment. The design permits maximization of mutual information, minimizes information loss, and allows fine-grained control over the transfer of collaborative cues (Ye et al., 14 Aug 2025). The result is more robust, accurate, and generalizable recommender systems, as validated in extensive industrial deployments.