Papers
Topics
Authors
Recent
Search
2000 character limit reached

Multidimensional Scaling (MDS)

Updated 5 April 2026
  • Multidimensional Scaling (MDS) is a set of techniques that embed high-dimensional dissimilarity data into low-dimensional Euclidean spaces while preserving pairwise distances.
  • It encompasses classical, metric, non-metric, Sammon mapping, and Isomap variants, each using distinct optimization methods to faithfully reconstruct data geometry.
  • MDS techniques scale to large datasets via landmark and interpolation methods and extend to multi-view, non-Euclidean, and infinite-dimensional frameworks.

Multidimensional Scaling (MDS) is a foundational family of techniques in geometry-aware data analysis and manifold learning. It seeks a configuration of points in a geometric space such that the pairwise distances among these points reflect a given matrix of input dissimilarities or distances. Applications encompass dimensionality reduction, visualization, feature learning, computational linguistics, shape matching, and network analysis. MDS encompasses classical (eigendecomposition-based) algorithms, iterative stress-minimization procedures, nonlinear and manifold variants such as Isomap, as well as scalability adaptations for large data and extensions to multiple, heterogeneous dissimilarity sources.

1. Mathematical Foundations and Classical MDS

Let {xi}i=1n\{x_i\}_{i=1}^{n} denote a finite collection of objects, each typically represented as a point in Rd\mathbb{R}^d (but not necessarily with explicit coordinates), and let D=(dij)D = (d_{ij}) be an n×nn \times n matrix of pairwise dissimilarities dijd_{ij}. The central goal is to find {yi}i=1n⊂Rp\{y_i\}_{i=1}^n \subset \mathbb{R}^p for p≪np \ll n such that their Euclidean distances ∥yi−yj∥\|y_i - y_j\| closely approximate the given dissimilarities.

Classical MDS is based on the following sequence (Ghojogh et al., 2020, Adams et al., 2019):

  • Compute squared distances D(2)=(dij2)D^{(2)}=(d_{ij}^2).
  • Center with H=In−(1/n)1n1nTH = I_n - (1/n) \mathbf{1}_n \mathbf{1}_n^T.
  • Form the double-centered Gram matrix: Rd\mathbb{R}^d0.
  • Compute the eigendecomposition Rd\mathbb{R}^d1.
  • Select the top Rd\mathbb{R}^d2 eigenpairs Rd\mathbb{R}^d3 to form the embedding: The Rd\mathbb{R}^d4th point Rd\mathbb{R}^d5 has coordinates Rd\mathbb{R}^d6.

This configuration uniquely minimizes the so-called strain (Frobenius norm between empirical and reconstructed Gram matrices), achieving exact recovery when the input matrix Rd\mathbb{R}^d7 is Euclidean (Adams et al., 2019).

2. Variants: Metric, Non-metric, Sammon, and Isomap

  • Metric MDS: Seeks Rd\mathbb{R}^d8 minimizing

Rd\mathbb{R}^d9

for fixed positive weights D=(dij)D = (d_{ij})0 (often D=(dij)D = (d_{ij})1). The objective is nonconvex and typically tackled by gradient-based or majorization algorithms such as SMACOF (Ghojogh et al., 2020, Boyarski et al., 2017).

  • Non-metric MDS: Preserves only the rank order of dissimilarities. Introduces a monotonic, typically isotonic-transform D=(dij)D = (d_{ij})2 and minimizes

D=(dij)D = (d_{ij})3

alternating between updating D=(dij)D = (d_{ij})4 and fitting D=(dij)D = (d_{ij})5 (Ghojogh et al., 2020).

  • Sammon Mapping: A weighted form where

D=(dij)D = (d_{ij})6

with normalization D=(dij)D = (d_{ij})7, emphasizing fidelity for small distances (Ghojogh et al., 2020).

  • Isomap: Replaces D=(dij)D = (d_{ij})8 by an estimated geodesic distance, using shortest paths on a neighborhood graph. After transforming these distances to a centered Gram matrix, Isomap applies the classical spectral embedding, capturing nonlinear manifold structures (Ghojogh et al., 2020).

3. Computation and Scalability for Large Data

For D=(dij)D = (d_{ij})9, classical algorithms become impractical due to cubic complexity. Partition-based approaches decompose the global problem:

  • Landmark MDS: Selects n×nn \times n0 "landmarks", runs classical MDS on them, and triangulates new points via analytic formulas (Gower’s interpolation) (Delicado et al., 2020).
  • Interpolation MDS, Reduced MDS, Pivot MDS, Divide-and-Conquer MDS, Fast MDS: Variants that partition, align, and stitch together local MDS results; see Table 1.
Algorithm Complexity (Time) Memory Comment
Landmark MDS n×nn \times n1 n×nn \times n2 Efficient, low-memory
Interp MDS n×nn \times n3 n×nn \times n4 Fastest in benchmark
Pivot MDS n×nn \times n5 n×nn \times n6 Direct inner-product
Div&Conq MDS n×nn \times n7 n×nn \times n8 More robust, slightly biased
Fast MDS n×nn \times n9 dijd_{ij}0 Recursive, higher const

Empirical evaluations on million-point EMNIST images demonstrated that all these methods closely match the quality of classical MDS but differ in time/memory footprint (Delicado et al., 2020).

4. Extensions: Multiple Views, Non-Euclidean Embeddings, and Feature Learning

  • Multi-view MDS (MVMDS): For dijd_{ij}1 heterogeneous distance matrices (views) on the same objects, minimizes a weighted sum of stresses with automatic weight learning:

dijd_{ij}2

with controller parameter dijd_{ij}3. Alternating minimization yields both the embedding dijd_{ij}4 and optimal view-weights dijd_{ij}5 (Bai et al., 2016).

  • Feature Learning via MDS: MDS can serve as a feature learning framework when applied to high-level, semantics-sensitive distances (e.g., spatial pyramid matching) between objects rather than raw pixel or Euclidean distances (Wang et al., 2013). The process learns vector codes so that Euclidean distances between codes reflect semantic similarity.
  • Non-Euclidean Targets: MDS can also be adapted to hyperbolic or spherical target spaces, with corresponding definitions of distance and gradient, enabling more faithful embeddings for inherently hierarchical or non-flat data (Cvetkovski et al., 2011, Liu et al., 2022).

5. Infinite, Continuous, and Operator-Theoretic Generalizations

MDS extends to infinite metric measure spaces dijd_{ij}6 by defining the "double-centered" kernel:

dijd_{ij}7

and embedding via the spectrum of the associated Hilbert-Schmidt operator dijd_{ij}8 on dijd_{ij}9. The canonical embedding {yi}i=1n⊂Rp\{y_i\}_{i=1}^n \subset \mathbb{R}^p0 (with {yi}i=1n⊂Rp\{y_i\}_{i=1}^n \subset \mathbb{R}^p1) optimally minimizes infinite-sample "strain" (Adams et al., 2019, Kassab, 2019).

Continuous MDS frameworks permit the analysis of limit behavior as {yi}i=1n⊂Rp\{y_i\}_{i=1}^n \subset \mathbb{R}^p2, establishing {yi}i=1n⊂Rp\{y_i\}_{i=1}^n \subset \mathbb{R}^p3-consistency of embeddings and proposing "Approximate Lipschitz Embedding" (ALE) to guarantee equicontinuity and uniform convergence (Trosset et al., 2024).

6. Optimization, Computational Complexity, and Algorithmic Guarantees

  • Majorization and Iterative Schemes: The SMACOF algorithm and its variants leverage quadratic majorization of the stress surface, ensuring monotonicity (Boyarski et al., 2017, Rajawat et al., 2016).
  • Complexity and Hardness: Global minimization of metric MDS stress is NP-hard, even in one dimension and for bounded-diameter graphs, ruling out efficient exact algorithms (Demaine et al., 2021). On the other hand, polynomial-time approximation schemes (PTAS) are available in special cases (e.g., small-diameter graphs), and quasi-polynomial approximations have been developed for general dissimilarity matrices with controlled aspect ratios (Bakshi et al., 2023).
  • Stochastic and Streaming Algorithms: Stochastic SMACOF and related incremental algorithms provide scalable and provably convergent solutions for large, dynamically evolving networks (Rajawat et al., 2016).

7. Applications and Theoretical Impact

Applications span manifold learning, visualization, feature extraction for object recognition, cross-linguistic semantic mapping, shape analysis, and cooperative network localization (Ghojogh et al., 2020, Wang et al., 2013, Klis et al., 2020, Rajawat et al., 2016). MDS forms the foundation for related methods such as Isomap and Spectral Generalized MDS, enables kernel-based, out-of-sample extensions, and admits both supervised and statistically informed extensions (e.g., {yi}i=1n⊂Rp\{y_i\}_{i=1}^n \subset \mathbb{R}^p4-informed MDS for microbiome studies) (Kim et al., 2023).

Theoretical advances include:

These developments collectively underscore MDS as a central, extensible tool for distance-preserving geometric embedding, equipped with rigorous foundations, practical adaptations, and across-the-board impact in data science and computational geometry.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Multidimensional Scaling (MDS).