Session-Based CTR Prediction via BiLSTM
- The paper introduces a session-based CTR prediction model that segments user behaviors into temporally coherent sessions and uses self-attention enhanced bidirectional LSTM to capture intra- and inter-session dynamics.
- Its methodology integrates bias-encoded multi-head self-attention and a local activation unit within the BiLSTM framework, achieving notable AUC improvements (e.g., from 0.6343 to 0.6375) on industrial datasets.
- The significance lies in unifying detailed session segmentation with advanced deep learning techniques, making the approach adaptable for applications in both advertising and recommendation systems.
Session-based prediction via Bidirectional LSTM is a framework for modeling user behavior sequences in click-through rate (CTR) prediction. This approach decomposes long-term behavior histories into temporally coherent sessions, extracts session-level interest representations using self-attention with bias encoding, and captures inter-session interest dynamics through bidirectional LSTM. The Deep Session Interest Network (DSIN) exemplifies this methodology by further weighting session contributions to prediction using a local activation unit, resulting in improved performance over previous models in industrial advertising and recommender system applications (Feng et al., 2019).
1. Session Segmentation and Representation
User behavior histories, denoted , are partitioned into a sequence of sessions based on temporal proximity. A new session is initiated when the time gap between consecutive behaviors exceeds 30 minutes, i.e., if . The -th session is , where up to user actions are retained per session. This segmentation is grounded on the empirical observation that behaviors are highly homogeneous within sessions but heterogeneous across sessions.
2. Intra-Session Interest Extraction via Biased Multi-Head Self-Attention
Each session embedding is enhanced by a three-way bias encoding to inject: (a) session index , (b) time-step within the session, and (c) embedding dimension . The bias tensor is , with , , . This tensor is added to session inputs as for all .
Each session then undergoes multi-head self-attention. is divided into heads of size , and for head ,
The result across heads is concatenated, projected, passed through a feed-forward network (FFN) with residual connections and layer normalization, yielding . Session-level interest embedding is then obtained by average pooling over the session’s steps:
3. Inter-Session Modeling Using Bidirectional LSTM
Session interest embeddings serve as the input sequence to a standard bidirectional LSTM. For each timestep (corresponding to session ), cell computations proceed as follows:
Forward and backward passes generate and , concatenated as . This encodes both preceding and succeeding session interest dynamics for each session position, capturing the temporal evolution and context of user interests.
4. Session Relevance Weighting via Local Activation Unit
For prediction, session contributions are adaptively weighted via a local activation unit (an attention mechanism at the session level). Given the target item embedding , attention scores over both raw session interests and context-enriched Bi-LSTM outputs are computed:
where and are learnable parameter matrices. and are final summary vectors aggregating session information, modulated by their relevance to the target item.
5. Prediction, Training, and Implementation Details
The final prediction vector is constructed by concatenating user profile embedding , item profile embedding , and the two session summary vectors:
where the multilayer perceptron (MLP) outputs the click-through probability via a sigmoid activation. End-to-end training minimizes the binary cross-entropy loss:
Key implementation details include: embedding dimension shared across all inputs; actions per session (typical ), most recent sessions (); single-layer Bi-LSTM per direction (hidden size ); residual connections and layer normalization in attention/FFN blocks; and standard dropout in MLPs.
6. Applications and Empirical Performance
Evaluation was conducted on two datasets: the Alimama advertising dataset (26 million ad logs over 1 million users and 800,000 ads) and the Alibaba production recommender dataset (6 billion logs, 100 million users, 70 million items). Models were trained on seven days of data and evaluated on the eighth.
The Area Under the Receiver Operating Characteristic Curve (AUC) was employed as the performance metric. DSIN achieved AUC = 0.6375 on the advertising dataset, surpassing DIEN (0.6343), and AUC = 0.6515 on the recommender dataset. This establishes DSIN as outperforming prior state-of-the-art methods under these benchmarks (Feng et al., 2019).
7. Significance and Extensions
Session-based modeling via Bi-LSTM, as implemented by DSIN, unifies fine-grained session segmentation, bias-enhanced intra-session interest extraction, bidirectional inter-session interest modeling, and adaptive session relevance weighting. The empirical findings underscore the value of respecting session boundaries and tracking both homogeneous within-session behavior and heterogeneous cross-session dynamics for CTR prediction. A plausible implication is that such layered modeling architectures may generalize to other temporal user modeling tasks beyond CTR, provided similar behavioral session structures and interest dynamics are present (Feng et al., 2019).