Papers
Topics
Authors
Recent
2000 character limit reached

One-Class SVM: Smooth CVaR Optimization

Updated 4 December 2025
  • The paper introduces a novel OC-SVM framework that integrates smooth CVaR surrogates with signature-based embeddings to achieve tractable risk calibration.
  • The methodology uses shuffle-product identities to derive closed-form polynomial surrogates, leading to explicit error bounds and enhanced hypothesis testing.
  • Empirical evaluations in anomalous diffusion and RNA modification detection demonstrate improved type I/II error control and increased detection power over traditional methods.

One-class Support Vector Machine (OC-SVM) algorithms optimising smooth Conditional Value-at-Risk (CVaR) objectives constitute a significant advance in novelty detection within path spaces, connecting sequential data analysis, statistical learning, and probability in function spaces. This class of algorithms exploits signature-based feature embeddings and the shuffle-product structure, enabling closed-form polynomial surrogates for risk-sensitive test statistics and new theoretical guarantees for error control and statistical power in hypothesis testing settings (Gasteratos et al., 2 Dec 2025).

1. Signature-based Features and Smooth CVaR Surrogates

Let XμX \sim \mu denote a path whose signature SN(X)S_N(X) is taken up to truncation level NN. To approximate the positive part [u]+[u]^+ on [K,K][-K,K], a polynomial Qn(u)=i=0naiuiQ_n(u) = \sum_{i=0}^n a_i u^i is introduced. The smooth CVaR surrogate is then defined by

fαn(ρ)=ρ+11αEμ[Qn(w,SN(X)ρ)],ρ[K,K]f^n_\alpha(\rho) = \rho + \frac{1}{1-\alpha} E_\mu\left[ Q_n(\langle w, S_N(X) \rangle - \rho) \right], \quad \rho \in [-K, K]

Employing the shuffle-product identity, $(\langle w, S \rangle)^i = \langle w^{\shuffle i}, S \rangle$, Theorem 3.1 shows the surrogate may be rewritten as

$E_\mu\left[ Q_n(\langle w, S(X) \rangle - \rho) \right] = \langle Q_n^{\shuffle}(w - \rho 1), E_\mu[S(X)] \rangle$

where $Q_n^{\shuffle}(\ell) = \sum_{i=0}^n a_i \ell^{\shuffle i} \in (T^{nN}(\mathbb{R}^d))^*$. The result is an explicit polynomial fαn(ρ)=m=0nbmρmf^n_\alpha(\rho) = \sum_{m=0}^n b_m \rho^m whose coefficients bmb_m depend only on Eμ[S(X)]E_\mu[S(X)] and shuffle-powers of ww. This surrogate admits closed-form computation, substantially improving tractability for high-dimensional path data.

2. OC-SVM Formulations and Optimisation Problems

The OC-SVM framework is captured as minimising a regularised CVaR of negative scoring functionals. Given a feature map φ\varphi, typically the truncated signature SNS_N or its infinite-level version, the population-level objective reads

w=argminwH{CVaRα(w,φ(X)H)+12wH2}w^* = \arg\min_{w \in H} \left\{ \mathrm{CVaR}_\alpha\left(-\langle w, \varphi(X) \rangle_H \right) +\frac{1}{2} \|w\|_H^2 \right\}

Replacing CVaRα\mathrm{CVaR}_\alpha by the smooth surrogate yields the smooth-CVaR OC-SVM problem,

(w,ρ)=argminwH,ρ[K,K]{12wH2+fαn(ρ)}(w^*, \rho^*) = \arg\min_{w \in H, \rho \in [-K,K]} \left\{ \frac{1}{2} \|w\|_H^2 + f^n_\alpha(\rho) \right\}

For empirical OC-SVM with finite samples {xi}\{x_i\}, the unconstrained primal is

(v^,ρ^)=argminvH,ρR{12vH2+[1γni=1n[ρv,φ(xi)H]+ρ]}(\hat v, \hat \rho) = \arg\min_{v \in H, \rho \in \mathbb{R}} \left\{ \frac{1}{2} \|v\|_H^2 + \left[ \frac{1}{\gamma n} \sum_{i=1}^n [\rho - \langle v, \varphi(x_i) \rangle_H]^+ - \rho \right] \right\}

A constrained quadratic program variant introduces slack variables ξi0\xi_i \ge 0: minv,ρ,ξ  12vH2+1γni=1nξiρs.t.v,φ(xi)Hρξi,  ξi0\min_{v, \rho, \xi} \; \frac{1}{2} \|v\|_H^2 + \frac{1}{\gamma n} \sum_{i=1}^n \xi_i - \rho \quad \text{s.t.} \quad \langle v, \varphi(x_i) \rangle_H \ge \rho - \xi_i,\; \xi_i \ge 0

3. Dual Formulation and Signature Kernels

The dual form of the constrained OC-SVM is

minαRn  12αKαs.t.0αi1γn,i=1nαi=1\min_{\alpha \in \mathbb{R}^n} \; \frac{1}{2} \alpha^\top K \alpha \quad \text{s.t.} \quad 0 \le \alpha_i \le \frac{1}{\gamma n}, \quad \sum_{i=1}^n \alpha_i = 1

for kernel matrix Kij=φ(xi),φ(xj)HK_{ij} = \langle \varphi(x_i), \varphi(x_j) \rangle_H. With φ=SN\varphi = S_N, the signature kernel is

κN(x,y)=SN(x),SN(y)HN\kappa_N(x,y) = \langle S_N(x), S_N(y) \rangle_{H_N}

Given solution α\alpha, the primal vector is w=iαiSN(xi)w = \sum_i \alpha_i S_N(x_i) and the test-score for a new path xx is f(x)=w,SN(x)HN=i=1nαiκN(xi,x)f(x) = \langle w, S_N(x) \rangle_{H_N} = \sum_{i=1}^n \alpha_i \kappa_N(x_i, x). In the smooth-CVaR population version, Eμ[S(X)]E_\mu[S(X)] enters via the closed-form surrogate, replacing empirical averages.

4. Theoretical Error Bounds: Type I and Power

Denoting Ωr={x:w,SN(x)>r}\Omega_r = \{x : \langle w, S_N(x) \rangle > r\}, the following bounds are established:

  • Type I Error (Theorem 3.4): If μ\mu obeys an (a,c)(a,c) transportation-cost inequality, including Gaussian and RDE laws, then there exist constants C1,C2C_1, C_2 such that

μ(w,SN(X)>r)C2exp{C122((r/A)2p(r/A)2p/N)}\mu\left(\langle w, S_N(X) \rangle > r\right) \le C_2 \exp \left\{ - \frac{C_1^2}{2} \left( (r/A)^{2p} \vee (r/A)^{2p/N} \right) \right\}

where A=wHNdNCN/2A = \|w\|_H \sqrt{N} d^N C^{N/2} and p(0,1]p \in (0, 1] depends on the deviation a(t)t2pa(t) \gtrsim t^{2p}. Solving μ()α\mu(\cdot) \le \alpha provides a quantile bound r(α)r^*(\alpha) and super-uniform p-values

p(x)=C2exp{C122((w,SN(x)/A)2p(w,SN(x)/A)2p/N)}μ(w,SN(X)w,SN(x))p(x) = C_2 \exp \left\{ - \frac{C_1^2}{2} \left( (\langle w, S_N(x) \rangle / A)^{2p} \vee (\langle w, S_N(x) \rangle / A)^{2p/N} \right) \right\} \le \mu(\langle w, S_N(X) \rangle \ge \langle w, S_N(x) \rangle)

  • Type II Error (Power, Theorem 3.3): For alternatives ν\nu with finite first moment,

ν(w,SN(X)r)1C(r/wH)pN(11N+1NE)dpNp/2N\nu( \langle w, S_N(X) \rangle \le r ) \ge 1 - C \left( r / \|w\|_H \right)^{- \frac{p}{N}(1 - \frac{1}{N} + \frac{1}{N} E_*) } d^p N^{p / 2N}

with E=a1((νμ))+Eμ[Xγp]E_* = a^{-1}((\nu \mid \mu)) + E_\mu [ \|X\|^p_\gamma ] for E1E_* \le 1, and (νμ)(\nu \mid \mu) the relative entropy. Thus, finite relative entropy ensures nontrivial lower bounds on power.

5. Algorithmic Procedure and Practical Considerations

Population-level smooth-CVaR OC-SVM is implemented via the following high-level steps:

  1. Empirical expected signature: E^=(1/m)i=1mSN(X(i))\hat E = (1/m) \sum_{i=1}^m S_N(X^{(i)})
  2. Surrogate objective: $\hat f_n(\rho) = \rho + \frac{1}{1-\alpha} \langle Q_n^{\shuffle}(w - \rho 1), \hat E \rangle$
  3. Joint optimisation: minw,ρ{12wH2+f^n(ρ)}\min_{w, \rho} \{ \frac{1}{2} \|w\|_H^2 + \hat f_n(\rho) \}
    • Solved by alternation or explicit polynomial root-finding when dim\dim \ll \infty.
  4. Test statistic: score(x)=w,SN(x)\text{score}(x) = \langle w^*, S_N(x) \rangle
  5. Hypothesis rejection: Reject H0H_0 if score(x)>ρ\text{score}(x) > \rho^*.

For sample-based OC-SVM, standard primal/dual QP with signature kernels is used (e.g., LIBSVM, ThunderSVM). At test time, iαiκN(xi,x)\sum_i \alpha_i \kappa_N(x_i, x) is compared to the learned bias ρ^\hat \rho.

6. Empirical Evaluation: Diffusion and Molecular Biology

Anomalous Diffusion

  • Setup: Binary discrimination between standard Brownian motion (μ\mu) and “spiked-BM” (νϵ\nu_\epsilon) defined by Xt=Bt+ϵ[tθ]+1X_t = B_t + \epsilon \sqrt{ [t-\theta]^+ } \wedge 1, θUnif[0,1]\theta \sim \mathrm{Unif}[0,1].
  • Statistic: Signature-based distance f(x)=SN(x)E[SN(X)]f(x) = \|S_N(x) - E[S_N(X)]\|; also, linear form in S(x)S(x).
  • Results:
    • AUROC vs ϵ2\epsilon^2: Monotonic increase, no sharp phase at ϵ=8\epsilon = \sqrt{8}.
    • Type I and II control: Empirical p-values (n=1000n=1000 calibration) give marginal FDR 0.1\le 0.1 but high conditional variability. Weibull tail-bound (Theorem 3.4) with 10510^5 samples yields super-uniform p-values, tighter FDR and FPR control.
    • Comparison: Signature-based distance outperforms TAMSD and is competitive with kernelised OC-SVM.

RNA Modification Detection

  • Data: Synthetic 100-nt oligos, three modifications (inosine, m5C, Ψ\Psi) at fixed positions; Nanopore direct RNA reads (Leger et al. 2021).
  • Preprocessing: Dorado basecalling, Uncalled4 event alignment, per-base segmentation.
  • Methods:
    • OC-SVM on signature features (N=6N=6, time-augment and invisibility-reset), $3000$ unmodified reads per site, p-values from 10510^5 held-out reads.
    • OC-SVM on standard 2D features (mean current and dwell time).
  • Results: At BH–FDR level $0.20$, signature OC-SVM yields substantially higher recall (power) for all modification types, with type I error controlled at nominal level.

7. Connections, Scope, and Implications

These developments bridge hypothesis testing, path signatures (Lyons et al.), transportation-cost inequalities (Gasteratos and Jacquier 2023), and robust machine learning. The use of smooth CVaR surrogates via shuffle-product identities establishes new analytic techniques for risk calibration and empirical p-value calculation. Non-asymptotic bounds on error rates generalise beyond Gaussian settings to laws of rough differential equation solutions, supporting broader applications in anomalous diffusion analysis and molecular biology. A plausible implication is further cross-fertilisation with time-series anomaly detection and functional data analysis, leveraging closed-form population objectives and signature kernel methods.

The principal contribution is the integration of population-level risk surrogates, shuffle-product algebra, and theoretical guarantees for novelty detection (Gasteratos et al., 2 Dec 2025). This framework enables more refined control of type I and type II errors and supports robust calibration for high-dimensional non-Euclidean data spaces.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to One-Class SVM Algorithms Optimising Smooth CVaR Objectives.