Oxytrees: Efficient Proxy-Based Biclustering Trees
- Oxytrees are proxy-based biclustering model trees that facilitate efficient bipartite learning on sparse interaction matrices.
- They employ proxy-based acceleration to compress interaction matrices into row and column proxies, dramatically reducing computational complexity.
- Their integration of kernelized leaf models using Kronecker product kernels yields expressive local predictions with faster training and inference.
Oxytrees are proxy-based biclustering model trees designed for efficient and interpretable bipartite learning—predicting binary interactions for pairs from two feature domains. They address scale, inductive generalization, and interpretability in settings where interactions are organized as sparse matrices, such as drug–target, RNA–disease, or regulatory networks. Oxytrees improve upon previous biclustering forests by compressing the interaction matrix into row and column proxies, enabling fast construction and inference with shallower trees whose leaves host expressive linear models parameterized by Kronecker product kernels (Ilídio et al., 16 Nov 2025).
1. Bipartite Learning Setting and Motivation
Bipartite learning seeks to infer a function given samples and , forming a sparse bipartite interaction matrix . Challenges arise from:
- Scalability: and may both reach thousands.
- Generalization: Models must predict for unseen instances (inductive setting).
- Interpretability: The goal is a two-dimensional partitioning—biclustering—that remains comprehensible and locally predictive.
Previous state-of-the-art biclustering forests such as BICTR scale as and utilize constant-value leaf predictions, limiting both efficiency and model expressivity. Oxytrees integrate proxy-based acceleration, efficient batch inference, and model tree structure with linear (kernelized) leaves to address these limitations (Ilídio et al., 16 Nov 2025).
2. Proxy-based Biclustering and Split Scoring
Oxytrees exploit the property that impurity measures on interaction submatrices can be reformulated in a proxy-friendly form:
For variance impurity, and 0. This enables computation of two proxy matrices per node:
- Row proxy: 1
- Column proxy: 2
Horizontal splits use prefix sums of 3, while vertical splits use 4, ensuring that all split candidates per feature are evaluated in 5 per threshold. The computational complexity is reduced to 6, a theoretical and observed 7–8 fold speedup relative to traditional biclustering trees (Ilídio et al., 16 Nov 2025).
3. Model Tree Construction and Leaf Modeling
Oxytrees construct a binary tree top-down. At each node, 9 features are randomly sampled from 0 or 1, akin to extremely randomized trees. Each candidate split is evaluated for impurity reduction 2, computed expediently via the proxy mechanism. Splits with the maximum 3 are chosen, and growth halts if 4 falls below a minimum or 5.
Distinctly, Oxytrees assign each leaf a local response matrix 6 and fit a linear model parameterized by a Kronecker product kernel. This approach improves expressivity over mean-value leaves and yields shallower, more efficient trees (Ilídio et al., 16 Nov 2025).
4. Efficient Batch Leaf Assignment and Inference
For prediction over Cartesian products 7, Oxytrees avoid redundant traversals. The inference procedure is:
- Horizontal splits partition 8, passing all of 9 downstream.
- Vertical splits partition 0, passing all of 1 downstream.
- At each leaf, all predictions for 2 are computed in batch via matrix multiplication.
This routine requires each test instance be visited only once per tree depth, with empirical complexity 3, compared to 4 for previous state-of-the-art methods (Ilídio et al., 16 Nov 2025).
5. Kronecker Product Kernel Models in Leaves
Each leaf utilizes Regularized Least Squares regression with a Kronecker product kernel 5. Given training Gram matrices 6 and 7 for leaf instances, eigendecompositions 8 and 9 produce the optimal weight matrix:
0
Predictions for test batch similarity matrices 1, 2 are:
3
This structure allows efficient, batch-wise computation for any leaf-level domain sizes (Ilídio et al., 16 Nov 2025).
6. Empirical Results and Performance Benchmarks
Oxytrees were evaluated on 15 bipartite datasets, including drug–nuclear receptor, kinase inhibitors, and lncRNA–disease, with sizes ranging from 4 to 5 and densities 1–20%. Baselines included MLPs on concatenated features, local models (RLS-avg, BLMNII, WkNNIR), RLS-Kron, NRLMF, and BICTR.
- Validation encompassed instance-wise + dyad-wise splits yielding transductive (TD), semi-inductive (LT/TL), and fully inductive (TT) test sets; positive-unlabeled (PU) masking at 0–75%.
- Metrics included AUROC, AUPRC, and Friedman + Nemenyi tests.
Key outcomes:
- Predictive performance: Oxytrees match or outperform BICTR and RLS-Kron, notably in inductive scenarios.
- Efficiency: Oxytrees train %%%%46547%%%% faster and predict %%%%48049%%%% faster than BICTR (p < 0).
- Model parsimony: Oxytrees reach 98% accuracy with 142% fewer trees due to expressive leaves.
- Proxy mechanism and leaf model ablation confirm that these components are crucial; removing leaf models degrades performance.
- Robustness: Maintained accuracy under PU-masking and with increased minimum leaf size (Ilídio et al., 16 Nov 2025).
7. Software Implementation and Reproducibility
A Python package "bipartite_learn" provides the Oxytrees framework, including:
OxytreeClassifierclass withfit(X1, X2, Y),predict(X1, X2),predict_proba(), andscore()- hyperparameters for number of trees, minimum leaf size, feature subsample size, impurity, regularization (2), and kernel functions 3, 4
- efficient batch inference
- access to all 15 preprocessed datasets and evaluation/PU-masking utilities
All results are reproducible via open-source code, with tutorials and Jupyter notebooks available at https://github.com/pedroilidio/oxytrees2025.
| Component | Characteristic | Complexity/Efficiency |
|---|---|---|
| Split Scoring | Proxy row/col matrices (5) | 6 |
| Leaf Model | Kronecker-kernel RLS | Efficient per-leaf SVD |
| Inference | Batch, single-pass per test set | 7 |
Oxytrees thereby constitute an interpretable, fast, and accurate methodology for inductive bipartite learning, combining proxy-based biclustering, expressive kernelized leaf modeling, and scalable batch inference (Ilídio et al., 16 Nov 2025).