Distance-Dependent Indian Buffet Process
- The dd-IBP is a non-exchangeable Bayesian nonparametric model that extends the Indian buffet process by incorporating explicit distance metrics to capture spatio-temporal and covariate-based feature dependencies.
- It employs a generative process where each data point draws new features based on a normalized proximity matrix and inherits features through distance-dependent connections.
- The model has demonstrated practical benefits in applications such as Alzheimer’s MRI classification and EEG reconstruction, despite inference challenges due to its complex sampling-based approach.
The distance-dependent Indian buffet process (dd-IBP) is a non-exchangeable Bayesian nonparametric prior for infinite latent feature models. It extends the Indian buffet process (IBP) by encouraging data points proximate in an explicit distance metric to share more latent features, thereby capturing spatio-temporal and covariate-based dependencies that violate the exchangeability assumption inherent in the classical IBP. The dd-IBP offers a flexible framework in which the pairwise structure of data determines both feature allocation and sharing, with recoverability of the standard IBP in the limiting case of uniform proximity.
1. Generative Framework and Model Specification
The dd-IBP defines a prior over an binary matrix representing feature inclusion for objects (or "customers") and countably infinite latent features ("dishes"). The generative process proceeds as follows (Gershman et al., 2011):
- New-dish ownership: Each customer draws and owns new features, where is a normalization of proximity weights and is the mass parameter.
- Distance-dependent connections: For every existing feature , customer chooses a customer to "link" with probability , where is a non-increasing decay function mapping distances to affinities.
- Dish inheritance: For feature with owner , customer sets if and only if a directed path exists from to in the corresponding connectivity graph defined by the link assignments.
This yields a binary matrix as a deterministic function of ownership and connectivity assignments. The joint prior is: with Poisson and multivariate multinomial factors for new-dish ownership and connection assignments, respectively.
The dd-IBP reduces to the original IBP under a sequential distance structure ( for ) and constant decay .
2. Distance Metrics, Decay Functions, and Affinity Structure
The proximity structure is encoded by a user-supplied symmetric (or optionally asymmetric) distance matrix , with typical choices:
- Temporal ( or for )
- Spatial (Euclidean/geodesic in )
- Covariate-based (arbitrary non-negative dissimilarities)
A decay kernel then translates these distances to unnormalized affinities, with normalization yielding the normalized proximity matrix . Standard forms include:
- Exponential:
- Logistic:
- Window:
- Constant:
These choices enable flexible modeling of smooth or blockwise structure in latent feature sharing.
3. Theoretical Feature-Sharing Properties
Let denote the total number of features for object and the number shared by and . Define the reachability indicator as 1 if can reach the owner of a feature, 0 otherwise.
Under finite : As , the normalized ratios and converge in distribution to the expected reachabilities under the proximity structure.
A salient feature is that the fraction of features shared between and varies smoothly with their pairwise distance, in contrast to the IBP, where and for all , enforcing universal exchangeability in the infinite feature limit.
4. Inference Methodology
Inference in the dd-IBP is typically performed via Markov chain Monte Carlo. In a generic latent-feature model with , the posterior over is proportional to the product of the likelihood, dd-IBP prior factors, and hyperpriors.
Key Gibbs/MH steps include:
- Hyperparameter : Conjugate gamma update if is gamma.
- Feature assignments for owned dishes: For each customer and active feature , sample new connection proportional to .
- Owner allocations (dish birth and death): Metropolis proposals for new Poisson draws and feature reallocations, with acceptance given by likelihood ratio.
- Model parameters (): Updated by Gibbs or Metropolis–Hastings as appropriate; for the linear-Gaussian likelihood (), the marginal likelihood can be computed in closed form after integrating out the weight matrix .
The joint prior and latent variable dependencies render marginalization of intractable, requiring explicit sampling of the connectivity graph .
5. Relation to the IBP and Other Non-Exchangeable Extensions
The dd-IBP strictly generalizes the IBP. It recovers the IBP in the sequential, uniform-affinity case. The standard IBP imposes exchangeability, ensuring that all pairs share of their features as . The dd-IBP allows arbitrary sharing patterns dictated by the distance metric and decay function.
Comparison with the attraction Indian buffet distribution (AIBD) (Warr et al., 2021) highlights several distinctions:
| Property | dd-IBP | AIBD |
|---|---|---|
| Marginal on | Intractable, requires sampling | Explicit, tractable PMF |
| Feature counts | Dependent on , typically IBP | Exactly as in the IBP |
| Exchangeability | Only under special proximity and kernel choices | Recovers IBP for or uniform |
| Asymmetric D | Supported | Requires symmetric distance matrix |
A key distinction is that the dd-IBP’s feature introduction process is itself distance-dependent (via ), while AIBD preserves the IBP’s new-feature marginal and imposes dependence only in the sharing of existing features.
6. Practical Impact and Applications
The dd-IBP is most appropriate in settings where latent structure is expected to follow non-exchangeable, locally correlated patterns. Notable case studies include:
- Alzheimer’s MRI classification: Age-sequential distance, exponential decay , resulting in improved AUC for feature-driven classifiers compared to IBP, the dependent IBP, and hierarchical beta processes.
- EEG missing-data reconstruction: Temporal distance from timestamps, with the dd-IBP linear-Gaussian model yielding lower squared imputation error than baseline priors (Gershman et al., 2011).
The model augments classical infinite latent feature models by introducing flexible, distance-aware structure, with applications in longitudinal studies, spatial genomics, and similar domains.
7. Limitations and Computational Considerations
Inference with the dd-IBP can be computationally intensive, as the marginal over is intractable, necessitating sampling of the full connectivity graph. Mixing can be slow due to dependencies among assignments. AIBD offers a tractable alternative in cases where symmetry is present and maintenance of IBP-like feature marginals is desired (Warr et al., 2021). Nonetheless, the dd-IBP offers unique modeling flexibility for hierarchical and asymmetric dependency structures not accessible to exchangeable or permutation-invariant priors.
A plausible implication is that, in large-scale applications where interpretability of distance effects and computational efficiency are equally important, tradeoffs between dd-IBP and alternatives such as AIBD or other dependent feature allocation models must be carefully considered.