Hierarchical ICL Maximization
- Hierarchical ICL Maximization is a clustering framework that integrates global ICL maximization via a hybrid genetic algorithm with hierarchical construction through Dirichlet regularization.
- The method optimizes the integrated classification likelihood to simultaneously determine the optimal partition and number of clusters using robust genetic operations.
- Empirical validations on diverse datasets demonstrate its capability to recover both fine- and coarse-grained structures, offering interpretable multi-scale representations.
Hierarchical ICL maximization is a model-based clustering framework that unifies the simultaneous discovery of cluster structure and model selection within discrete latent variable models (DLVMs) via the integrated classification likelihood (ICL) criterion. The approach consists of a two-phase procedure: (1) global maximization of the ICL using a hybrid genetic algorithm to yield a partition and cluster number, and (2) construction of a multi-scale hierarchy by regularizing with the Dirichlet prior parameter and employing a log-linear merge rule. The method rigorously connects regularization, model-based clustering, and hierarchy construction, with direct optimization performed on the exact integrated likelihood and probabilistic assignments.
1. The Integrated Classification Likelihood (ICL) in DLVMs
The ICL objective quantifies the plausibility of a clustering in a generative latent variable model. For data with latent hard assignments , cluster-proportions , and within-cluster parameters , the ICL is defined as:
$\ICLex(Z; \alpha, \beta) = \log \int p(X \mid Z, \theta)\, p(\theta \mid \beta) [p(Z \mid \Pi)\, p(\Pi \mid \alpha)] \, d\theta\, d\Pi = \log p(X \mid Z, \beta) + \log p(Z \mid \alpha)$
where the first term is the integrated conditional log-likelihood, and the second term results from Dirichlet–multinomial conjugacy:
This decomposition enables principled model selection (number of clusters ) and accounts for the Occam’s razor effect (penalizing over-complex partitions).
2. Global ICL Maximization via Hybrid Genetic Algorithm
Standard greedy optimization of $\ICLex$ is prone to sub-optimal local maxima. The hybrid genetic algorithm (GA) overcomes this by employing a population of set partitions (not label vectors), eliminating label-switching redundancy and enabling effective structural crossover:
- Population: Each individual is a partition .
- Fitness: $\mathrm{fit}(P) = \ICLex(P)$.
- Genetic operations:
- Selection: Rank-based.
- Crossover: Offspring blocks are all non-empty intersections of two parents.
- Mutation: Randomly splits one block.
- Local search: Greedy merges and swaps within the offspring, always increasing ICL.
- Complexity: Each crossover and mutation is ; swap/merge is ; operations focus on offspring only.
Pseudocode excerpt:
1 2 3 4 5 6 7 8 |
Initialize population P_1,…,P_V (seeded + local swap).
For g = 1…G_max:
Carry over best solution.
For (V–1) pairs:
Crossover → child partition.
Locally merge/split/swap (ICL-increasing).
Add to next generation.
Return best solution found. |
This procedure jointly selects both the optimal partition and the cluster number .
3. Hierarchical Construction Through Dirichlet Prior Regularization
Given an ICL-optimal partition, the second phase builds a hierarchy by leveraging the Dirichlet prior parameter as a regularization knob controlling granularity:
- Linearized criterion: For small ,
$\ICLlin(Z, \alpha) = I(Z) + (K-1) \log \alpha$
where
- Greedy merge rule: At each step, merge the pair for which
is maximized. This is the largest interval where the merge dominates the parent partition.
- Hierarchy construction: For decreasing , update partitions and using the above rule until a single cluster remains.
Hierarchy pseudocode:
1 2 3 4 5 6 |
Input: Partition Z^(K), intercept I(·), α^(K) = 1.
For k = K…2:
For pairs g<h: compute ΔI_{g,h}, α_{g,h}
Select (g*, h*) with max α_{g,h}
Merge, update Z^(k-1), α^(k-1)
Return sequence of partitions and α-path. |
Non-dominant merges (where the new α increases) are pruned, producing an approximate Pareto frontier of optimal granularities.
4. Theoretical Interpretation and Regularization Path
The hierarchy induced by varying corresponds to a multi-scale view of the data: high yields many clusters; low merges aggressively, yielding a coarser clustering. The linearized criterion is affine in , and the explicit α-path offers interpretable model selection and visualization possibilities (dendrograms with axes ; optimal leaf ordering via block-sum minimization).
A plausible implication is that the α-regularization perspective connects hierarchical clustering with Bayesian model selection explicitly, such that the entire sequence of clusterings from fine to coarse can be understood as traversing a regularization path in model space (Côme et al., 2020).
5. Empirical Validation and Performance
Extensive empirical evaluation demonstrates the superiority and robustness of hierarchical ICL maximization:
- SBM (stochastic block model) simulations: Hybrid GA recovers all underlying clusters in up to 93% of runs (NMI ≈ 1.0), outperforming greedy and EM/BIC/AIC approaches.
- Mixtures of multinomials: Hybrid GA attains the correct number of clusters more frequently and with higher ICL/NMI than standard methods.
- Real networks: On diverse graphs (Blogs, Books, Football, Jazz), the procedure consistently yields the highest $\ICLex$ and lowest variance, even with small population sizes.
- French parliamentary co-voting: The method initially discovers fine block co-clusters, then hierarchical pruning reveals interpretable coarser structures aligned with real political groupings (NMI ≈ 0.8 with party labels).
- All cases: The hierarchy enables interpretation at multiple scales by extracting a nested set of partitions and their merge α-paths (Côme et al., 2020).
6. Practical Considerations and Generalization
Key hyperparameters include population size (), mutation rate (), and maximum generations () for the GA, as well as the initial value and updating schedule for in Phase II. The GA-based maximization is parallelizable and scalable due to efficient representations and operations. Merges are computationally trivial at each step (), and the linearization in is robust for sufficiently small .
A plausible extension is that this framework applies to any DLVM with tractable integrated classification likelihood, including degree-corrected SBMs, mixtures of multinomials, and relational models with conjugate priors. The entire approach is model-agnostic within the family of tractable DLVMs, yielding both cluster structures and interpretable hierarchical relationships under the unified principle of ICL maximization (Côme et al., 2020).