Papers
Topics
Authors
Recent
Search
2000 character limit reached

Discriminant Gain in ISAC and Edge AI Systems

Updated 8 January 2026
  • Discriminant Gain is a metric that quantifies class separability using Gaussian formulations, offering a clear measure of inference accuracy and error bounds.
  • It replaces traditional MSE criteria by optimally allocating power via closed-form water-filling solutions, focusing on the most discriminative features.
  • DG governs the trade-off between sensing and communication, enabling efficient system design and benchmarking in ISAC and ISEA applications.

Discriminant Gain (DG) quantifies class separability in feature space and has recently become a core metric for characterizing inference performance in task-oriented integrated sensing and communication (ISAC) and integrated sensing and edge AI (ISEA) systems. Unlike classical mean squared error (MSE)-based criteria, DG directly connects with detection-theoretic limits, admitting tractable expressions and closed-form optimization in Gaussian mixture settings. It governs the tradeoff between sensing and communication resources, optimally allocates power to maximize inference accuracy, and enables the principled design and benchmarking of ISAC/ISEA pipelines.

1. Formal Definition and Mathematical Foundations

DG is defined under the assumption that feature vectors xCM\bm{x} \in \mathbb{C}^M for each class \ell follow a complex Gaussian distribution. The pairwise discriminant gain between classes ii and jj is

DG(i,j)=(μiμj)Σ1(μiμj)=m=0M1(μi,mμj,m)2σm2\mathrm{DG}(i,j) = (\bm{\mu}_i - \bm{\mu}_j)^\top \bm{\Sigma}^{-1}(\bm{\mu}_i - \bm{\mu}_j) = \sum_{m=0}^{M-1} \frac{(\mu_{i,m} - \mu_{j,m})^2}{\sigma_m^2}

where μi\bm{\mu}_i is the mean of class ii and Σ=diag(σ02,,σM12)\bm{\Sigma} = \operatorname{diag}(\sigma_0^2, \ldots, \sigma_{M-1}^2) the feature covariance.

For multiclass problems, DGminminijDG(i,j)\mathrm{DG}_{\min} \triangleq \min_{i \neq j} \mathrm{DG}(i, j) governs the worst-case separability, serving as a lower-complexity surrogate for inference-oriented system design (Dong et al., 23 Oct 2025).

Alternative formulations use the symmetric Kullback-Leibler divergence in multi-view edge AI settings. For views indexed by kk and PkP_k as the subspace projection at sensor kk,

Gk(,)=(μμ)PkC1Pk(μμ)G_k(\ell, \ell') = (\mu_\ell - \mu_{\ell'})^\top P_k\, C^{-1}\, P_k (\mu_\ell - \mu_{\ell'})

with CC the shared covariance (Chen et al., 2023).

2. Connection to Inference Error Bounds

DG provides a tight link to Bayesian inference errors. For two classes and one-dimensional features,

Pe(2)=Q(DG2),DG=(μ0μ1)2σ2P_e^{(2)} = Q\left(\sqrt{\frac{\mathrm{DG}}{2}}\right), \qquad \mathrm{DG} = \frac{(\mu_0 - \mu_1)^2}{\sigma^2}

where Q(x)Q(x) is the Gaussian Q-function. The relation extends to vector features and multiclass cases using the minimum pairwise DG, confining the inference error probability as

Pe(L1)Q(DGmin2)P_e \geq (L-1) Q\left(\sqrt{\tfrac{\mathrm{DG}_{\min}}{2}}\right)

Thus, increasing DG—particularly DGmin\mathrm{DG}_{\min}—monotonically decreases a lower bound on the inference error probability, establishing DG as a critical system-level surrogate (Dong et al., 23 Oct 2025).

In multi-view settings, DG predicts the exponential rate at which the entropy (uncertainty) of the predicted class distribution decays with the number of aggregated views KK. The global DG exponent controls how rapidly

Hs(L1)exp(κξK),K1H_s \approx (L - 1) \exp(-\kappa\,\xi\,K), \quad K \gg 1

where HsH_s is an entropy surrogate, κ\kappa a constant, and ξ\xi the asymptotic average DG (Chen et al., 2023).

3. Task-Oriented DG Maximization and System Models

DG-centric system optimization replaces MSE with the direct maximization of DG under power/resource constraints. In the compress-and-estimate ISAC link, each transformed feature x~n\tilde{x}_n is transmitted over a fading channel, and the effective per-subcarrier DG is

DGn=hnbn2(μ0,nμ1,n)2hnbn2σn2+σw2\mathrm{DG}_n = \frac{|h_n b_n|^2(\mu_{0,n} - \mu_{1,n})^2}{|h_n b_n|^2 \sigma_n^2 + \sigma_w^2}

where bnb_n is the transmission gain, hnh_n the channel coefficient, and σw2\sigma_w^2 the communication noise.

The DG-maximization under a total power budget PcP_{\sf c} becomes

max{bn}n=0N1DGns.t. n=0N1bn2νn2Pc\max_{\{b_n\}} \sum_{n=0}^{N-1} \mathrm{DG}_n \qquad \text{s.t. } \sum_{n=0}^{N-1} |b_n|^2 \nu_n^2 \le P_{\sf c}

with νn2=E[x~n2]\nu_n^2 = \mathbb{E}[|\tilde{x}_n|^2] (Dong et al., 23 Oct 2025).

In edge AI, local and global DGs are constructed via symmetric KL divergence in the projected subspace, and DG's subspace geometry determines how well class means are separated across the pooled sensor network (Chen et al., 2023).

4. Closed-Form DG-Optimal Power Allocation and Water-Filling Structure

The DG-maximization problem is convex in the allocated per-feature powers {bn2}\{|b_n|^2\} and admits a closed-form, water-filling-type solution: bnDG=ejhnhn[σwhnμ0,nμ1,nνnλσn2σw2σn2]+b_n^{\rm DG} = \frac{e^{-j \angle h_n}}{|h_n|} \sqrt{\left[ \frac{\sigma_w |h_n| |\mu_{0, n}-\mu_{1, n}|}{\nu_n \sqrt{\lambda} \sigma_n^2} - \frac{\sigma_w^2}{\sigma_n^2} \right]^+} with λ\lambda a Lagrange multiplier for the power constraint.

Letting pn=bn2νn2p_n = |b_n|^2 \nu_n^2,

pnDG=[μ0,nμ1,nσn2σwλσw2νn2]+p_n^{\rm DG} = \left[\frac{|\mu_{0,n}-\mu_{1,n}|}{\sigma_n^2} \frac{\sigma_w}{\sqrt{\lambda}} - \frac{\sigma_w^2}{\nu_n^2} \right]^+

Power is assigned only to features with sufficiently high discrimination-to-noise ratios, turning off weak subcarriers and concentrating resources on the most informative dimensions. This distinguishes DG-water-filling from its MSE-based counterpart, which allocates power more uniformly, even to low-informative features (Dong et al., 23 Oct 2025).

5. Comparison with MSE-Optimal and Traditional Criteria

Under MSE-optimal design, the system solves

min{bn}nMSEns.t. nbn2νn2Pc\min_{\{b_n\}} \sum_n \mathrm{MSE}_n \quad \text{s.t.} \ \sum_n |b_n|^2 \nu_n^2 \le P_{\sf c}

with a similar water-filling solution but lacking the emphasis on discrimination power: bnMSE=ejhnhn[σwhnνnλσw2σn2]+b_n^{\rm MSE} = \frac{e^{-j \angle h_n}}{|h_n|} \sqrt{\left[ \frac{\sigma_w |h_n|}{\nu_n \sqrt{\lambda}} - \frac{\sigma_w^2}{\sigma_n^2} \right]^+} The DG-optimal approach introduces an extra factor μ0,nμ1,nσn2\tfrac{|\mu_{0, n} - \mu_{1, n}|}{\sigma_n^2}, biasing power allocation toward features with higher class separability. In the low-SNR regime, DG-maximization achieves substantially better power efficiency by shutting off weak subcarriers. In the high-SNR regime, the distinction between DG- and MSE-optimal allocations vanishes as all channels are used and the benefit per dB equalizes (Dong et al., 23 Oct 2025).

6. Multi-View Aggregation, Channel Effects, and Discriminant Loss

In ISEA, aggregated DG grows linearly with the number of views/sensors, and the geometry of the pooled subspace (as determined by Pˉ=1KkPk\bar{P} = \tfrac{1}{K} \sum_k P_k) controls overall class separability: G,=KD,,D,=(μμ)PˉC1Pˉ(μμ)G_{\ell, \ell'} = K D_{\ell, \ell'}, \quad D_{\ell, \ell'} = (\mu_\ell - \mu_{\ell'})^\top \bar{P}\,C^{-1}\,\bar{P}\,(\mu_\ell - \mu_{\ell'}) Sensing uncertainty, measured by entropy surrogates, decays exponentially with the product of the global discriminant gain and the number of views: Hs(L1)exp(κξK)H_s \approx (L-1) \exp(-\kappa\,\xi\,K) When transmission occurs over noisy (e.g., AirComp) channels, channel-induced discriminant loss attenuates DG: Aloss=D~(γair)DˉA_{\rm loss} = \frac{\tilde{D}(\gamma_{\rm air})}{\bar{D}} The uncertainty scaling law becomes Hs(γair)(L1)exp(κξAlossK)H_s(\gamma_{\rm air}) \approx (L-1) \exp(-\kappa\,\xi\,A_{\rm loss}\,K), where Aloss1A_{\rm loss}\leq 1 quantifies the effective reduction in discriminability caused by channel noise or distortion (Chen et al., 2023).

7. Operational Insights: Power-Efficient Inference and Resource Tradeoffs

DG-optimal designs yield several system-level advantages:

  • Selective Feature Activation: Simulation and empirical studies confirm that for fixed inference accuracy, DG-based resource allocation requires substantially less power than MSE-optimal design by focusing on the most discriminative features or views (Dong et al., 23 Oct 2025).
  • Sensing-Communication Tradeoff: Power savings achieved by DG-optimal communication can be redirected to improve sensing quality, leading to joint design strategies for radar and communication subsystems (Dong et al., 23 Oct 2025).
  • Adaptive Aggregation and Access Mode: In multi-view settings, the exponential convergence rate remains until attenuated by channel effects. Adaptive switching between over-the-air computing and orthogonal access based on the ratio N/KN/K (number of antennas to sensors) allows the system to maintain high global DG and rapid uncertainty decay (Chen et al., 2023).

DG thus underpins the design of resource-constrained, inference-optimal ISAC and ISEA links, establishing a unified metric that generalizes across model classes, channel effects, and practical hardware constraints.

References: (Dong et al., 23 Oct 2025, Chen et al., 2023)

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Discriminant Gain (DG).