DiffRed: Dimensionality Reduction guided by stable rank
Abstract: In this work, we propose a novel dimensionality reduction technique, DiffRed, which first projects the data matrix, A, along first $k_1$ principal components and the residual matrix $A{*}$ (left after subtracting its $k_1$-rank approximation) along $k_2$ Gaussian random vectors. We evaluate M1, the distortion of mean-squared pair-wise distance, and Stress, the normalized value of RMS of distortion of the pairwise distances. We rigorously prove that DiffRed achieves a general upper bound of $O\left(\sqrt{\frac{1-p}{k_2}}\right)$ on Stress and $O\left(\frac{(1-p)}{\sqrt{k_2*\rho(A{*})}}\right)$ on M1 where $p$ is the fraction of variance explained by the first $k_1$ principal components and $\rho(A{*})$ is the stable rank of $A{*}$. These bounds are tighter than the currently known results for Random maps. Our extensive experiments on a variety of real-world datasets demonstrate that DiffRed achieves near zero M1 and much lower values of Stress as compared to the well-known dimensionality reduction techniques. In particular, DiffRed can map a 6 million dimensional dataset to 10 dimensions with 54% lower Stress than PCA.
- (2005). MDS as a Psychological Model, pages 359–388. Springer New York, New York, NY.
- Ntire 2017 challenge on single image super-resolution: Dataset and study. In 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 1122–1131.
- Ntire 2017 challenge on single image super-resolution: Dataset and study. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.
- The fast johnson–lindenstrauss transform and approximate nearest neighbors. SIAM Journal on Computing, 39(1):302–322.
- Dimensionality reduction: theoretical perspective on practical measures. In NIPS, volume 32.
- Random projection in dimensionality reduction: Applications to image and text data. KDD ’01, page 245–250.
- Generalized multidimensional scaling: A framework for isometry-invariant partial surface matching. Proceedings of the National Academy of Sciences of the United States of America, 103:1168 – 1172.
- Robust euclidean embedding. ICML ’06, page 169–176, New York, NY, USA.
- Calibration by correlation using metric embedding from nonmetric similarities. IEEE Trans. on PAMI, 35(10):2357–2370.
- Gpu accelerated t-distributed stochastic neighbor embedding. Journal of Parallel and Distributed Computing, 131:1–13.
- Optimal Approximate Matrix Product in Terms of Stable Rank. In Chatzigiannakis, I., Mitzenmacher, M., Rabani, Y., and Sangiorgi, D., editors, 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016), volume 55 of Leibniz International Proceedings in Informatics (LIPIcs), pages 11:1–11:14, Dagstuhl, Germany. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik.
- Automated hate speech detection and the problem of offensive language. In Proceedings of the 11th International AAAI Conference on Web and Social Media, ICWSM ’17, pages 512–515.
- Reducing high-dimensional data by principal component analysis vs. random projection for nearest neighbor classification. In ICMLA’06, pages 245–250.
- Toward a quantitative survey of dimension reduction techniques. IEEE Transactions on Visualization and Computer Graphics, 27:2153–2173.
- The fast johnson-lindenstrauss transform is even faster.
- Fast randomized pca for sparse data.
- Fiorini, S. (2016). gene expression cancer RNA-Seq. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5R88H.
- F.R.S., K. P. (1901). Liii. on lines and planes of closest fit to systems of points in space. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science, 2(11):559–572.
- Dimensionality reduction for supervised learning with reproducing kernel hilbert spaces. J. Mach. Learn. Res., 5:73–99.
- Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions.
- Learning-based low-rank approximations. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
- Aptos 2019 blindness detection.
- Restricted eigenvalue from stable rank with applications to sparse linear regression. In Bubeck, S., Perchet, V., and Rigollet, P., editors, Conference On Learning Theory, COLT 2018, Stockholm, Sweden, 6-9 July 2018, volume 75 of Proceedings of Machine Learning Research, pages 1011–1041. PMLR.
- Krizhevsky, A. (2009). Learning multiple layers of features from tiny images.
- Kruskal, J. B. (1964). Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis. Psychometrika, 29(1):1–27.
- Lewis, D. (1997). Reuters-21578 Text Categorization Collection. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C52G6M.
- Visualizing high-dimensional data: Advances in the past decade. IEEE Transactions on Visualization and Computer Graphics, 23(3):1249–1268.
- Umap: Uniform manifold approximation and projection for dimension reduction.
- Bank Marketing. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5K306.
- Predicting internet network distance with coordinates-based approaches. In Proceedings.Twenty-First Annual Joint Conference of the IEEE Computer and Communications Societies, volume 1, pages 170–179 vol.1.
- Finding the homology of submanifolds with high confidence from random samples. Discrete and Computational Geometry, 39(1-3):419–441.
- Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830.
- Distributional scaling: An algorithm for structure-preserving embedding of metric and nonmetric spaces. J. Mach. Learn. Res., 5:399–420.
- Sampling from large matrices: An approach through geometric functional analysis. J. ACM, 54(4):21–es.
- Hanson-wright inequality and sub-gaussian concentration.
- Schmidt, B. (2018). Stable random projection: Lightweight, general-purpose dimensionality reduction for digitized libraries. Journal of Cultural Analytics, 3(1).
- Spearman, C. (1904). “general intelligence”, objectively determined and measured. The American Journal of Psychology, 15(2):201–292.
- Torgerson, W. (1952). Multidimensional scaling i. theory and method. Psychometrika, 17(4):401–419.
- Visualizing data using t-SNE. Journal of Machine Learning Research, 9:2579–2605.
- Dimensionality reduction: a comparative review. J Mach Learn Res, 10:66–71.
- Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. https://github.com/zalandoresearch/fashion-mnist.
- Robust dimensionality reduction for high-dimension data. In 2008 46th Annual Allerton Conference on Communication, Control, and Computing, pages 1291–1298.
- Yin, H. (2007). Nonlinear dimensionality reduction and data visualization: A review. International Journal of Automation and Computing, 4:294–303.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.