Papers
Topics
Authors
Recent
Search
2000 character limit reached

Faster Support Vector Machines

Published 20 Aug 2018 in cs.LG, cs.DM, cs.DS, and stat.ML | (1808.06394v3)

Abstract: The time complexity of support vector machines (SVMs) prohibits training on huge data sets with millions of data points. Recently, multilevel approaches to train SVMs have been developed to allow for time-efficient training on huge data sets. While regular SVMs perform the entire training in one -- time consuming -- optimization step, multilevel SVMs first build a hierarchy of problems decreasing in size that resemble the original problem and then train an SVM model for each hierarchy level, benefiting from the solved models of previous levels. We present a faster multilevel support vector machine that uses a label propagation algorithm to construct the problem hierarchy. Extensive experiments indicate that our approach is up to orders of magnitude faster than the previous fastest algorithm while having comparable classification quality. For example, already one of our sequential solvers is on average a factor 15 faster than the parallel ThunderSVM algorithm, while having similar classification quality.

Citations (27)

Summary

  • The paper’s main contribution is a multilevel strategy that leverages label propagation to accelerate SVM training.
  • It utilizes a hierarchy of coarsened problems for efficient initial training followed by uncoarsening to refine support vectors.
  • Experimental results demonstrate orders of magnitude speed-ups while maintaining robust classification accuracy on diverse datasets.

Faster Support Vector Machines

The paper "Faster Support Vector Machines" (1808.06394) proposes a multilevel approach to enhance the efficiency of Support Vector Machines (SVMs), allowing them to handle massive datasets effectively. The authors introduce the use of label propagation algorithms for constructing problem hierarchies, achieving significant speed improvements over existing methods without compromising classification accuracy.

Introduction to Multilevel SVMs

The paper begins by addressing the limitations of traditional SVMs due to their computational complexity, which becomes prohibitive for large-scale datasets. Typical SVM training involves solving a complex optimization problem with time complexity ranging between O(n2)O(n^2) and O(n3)O(n^3). The multilevel SVM approach simplifies this by creating a hierarchy of smaller, related problems.

Multilevel Paradigm and Hierarchy Construction

The multilevel process involves creating a hierarchy starting from the largest, original problem and progressively coarsening it. Each level in this hierarchy represents a reduced version of the problem:

  • Coarsening: It uses a label propagation clustering technique to quickly reduce the size of the problem.
  • Initial Training: A smaller version of the problem is solved to find initial SVM parameters.
  • Uncoarsening: Gradually moving back to the original problem size, refining the model with insights learned at previous levels. Figure 1

    Figure 1: A non-separable binary classification problem solved with SVM and slack variables. Non-support vectors are gray shaded. The mapping phi is assumed to be the identity in this example.

Implementation Details

Coarsening Using Label Propagation

Label propagation is applied to cluster the graph representation of data points efficiently. This clustering determines proximity and helps define how the problem can be reduced through contraction.

Efficient Training

In multilevel SVMs, training initially occurs at the smallest level. Due to its reduced size, optimization is rapid, speeding up model selection. The algorithm uses pre-determined kernel parameters, significantly minimizing computational overhead compared to exhaustive grid searches or Uniform Design processes.

Uncoarsening Strategy

During uncoarsening, the support vectors are refined as they re-expand to larger problem sizes. Structural insights from the coarse levels help optimize parameters progressively as the hierarchy is traversed. Figure 2

Figure 2

Figure 2

Figure 2: Distribution of speed-ups of our algorithm over mlsvm-AMG when running algorithms on the same k-folds. From left to right: speed-ups of LPSVM over mlsvm-AMG. speed-ups of LPSVMfast_\text{fast}

Experimental Results

Extensive experiments show the proposed approach reduces computation time significantly, sometimes by orders of magnitude, compared to leading algorithms such as ThunderSVM. The performance is tested across various datasets, from standard UCI benchmarks to massive data challenges, showing comparable or improved classification quality.

  • Performance Metrics: G-mean is used to handle imbalanced datasets, ensuring sensitivity and specificity aren't compromised.
  • Speed Comparisons: Compared to mlsvm-AMG, the algorithm achieves geometric mean speed-ups across all tested instances.

Conclusion

This multilevel approach presents a substantial advancement in efficiently scaling SVMs to large data sizes while maintaining robust classification quality. The hierarchical model selection strategy enables quick optimization, making the approach versatile across domains requiring rapid data-driven insights.

Future work could explore automatic parameter tuning and extend the multilevel strategy to other kernel-based learning models, potentially broadening its application scope further. The KaSVM open-source framework supports ongoing development and application studies.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.