SAE Output Modification with Scores
- SAE output modification using scores is a framework where quantitative metrics guide feature selection, output adjustment, and model interpretability.
- It leverages input and output scores to steer activations, ensuring improved semantic alignment and enhanced fairness through precise interventions.
- The approach integrates multi-objective metrics and statistical calibration, achieving two- to three-fold improvements over traditional unsupervised techniques.
SAE output modification using scores refers to the principled use of scoring mechanisms—quantitative metrics or score functions—to select, adjust, reweight, and evaluate the activation or output of sparse autoencoders (SAEs) or broader model components. In contemporary machine learning, scores serve as both optimization objectives and interpretable diagnostic tools for pushing model outputs toward desired properties or controlling their behavior post-training. SAE output modification via scores is now central in applications including model steering, topic alignment, fairness optimization, multi-objective aggregation, and rigorous statistical evaluation.
1. Conceptual Foundations: Scores in SAE Output Control
Scores are quantitative functions designed to measure aspects central to model function—such as interpretability, semantic alignment, output influence, or calibration. In the context of SAE output modification, scores can act as:
- Selection criteria—identifying which features or components should be activated, suppressed, or modulated.
- Adjustment metrics—determining the magnitude or direction of feature intervention for output steering.
- Evaluation measures—quantifying the degree of success or fidelity after modification.
The underlying mathematical structure of a score varies with application: it can be an activation overlap (e.g., input score), intervention effect (e.g., output score), semantic distance, entropy, score change under variable transformation, or a fairness discrepancy.
2. Score-Driven Feature Selection and Steering in SAEs
Recent models of steering in SAEs distinguish “input features” and “output features,” using scores to target the latter for controlling LLM output (Arad et al., 26 May 2025). Two scores central to this paradigm are:
- Input Score (): Measures the token-activation overlap between SAE feature activations and logit lens projections. High indicates features tied to input patterns.
- Output Score (): Captures the effect of feature intervention on model output probabilities. Computed as the difference in rank-weighted token probabilities with and without feature intervention:
with .
Empirically, the decoupling between input and output scores enables identification of features that are more effective at steering output (high , low ), yielding two- to three-fold improvements over unsupervised interventions when filtered appropriately (Arad et al., 26 May 2025).
3. Topic Alignment: Semantic Scoring and Layer-Level Modification
Semantic alignment methodologies utilize scores to quantify neuron-specific affinity to an alignment topic or text (Joshi et al., 14 Jun 2025). The scoring workflow comprises:
- Activation summary: For each neuron , normalize the aggregated activation across prompt tokens:
where .
- Semantic distance: Embed prompts and reference alignment text, compute minimum distance .
- Score aggregation: For neuron ,
The final score is normalized between $0$ and $1$.
Score-weighted modification utilizes these neuron scores to emphasize topic-aligned neurons. The “Swap” approach applies context-sensitive selection:
This construct enables dynamic, token-aware steering towards the alignment topic. The contamination metric quantifies misalignment as:
Experiments confirm the approach enhances language acceptability and reduces training/inference time compared to classical fine-tuning (Joshi et al., 14 Jun 2025).
4. Dense Latents: Activation Density and Antipodality Scores
Dense SAE latents—units that fire on 10–50% of tokens—are analyzed via their activation density and geometric antipodality (Sun et al., 18 Jun 2025). Key scoring mechanisms include:
- Antipodality Score ():
where is the cosine similarity.
- Nullspace Coefficient ():
Dense latents with high regulate entropy via alignment with the quasi-nullspace of the unembedding matrix.
Dense latents are classified into position (context tracking), context-binding, nullspace, alphabet, POS, and PCA categories. Their functional roles shift from structure in early layers to semantic/contextual and output-oriented signals in later layers, and interventions guided by activation or antipodality scores enable fine-grained SAE output control (Sun et al., 18 Jun 2025).
5. Optimization Objectives: Score-Based Training and Fairness
In DNN-based source enhancement, “score” refers to non-differentiable objective metrics (e.g., PESQ, STOI) for perceptual quality. Rather than minimizing MSE, the policy gradient method optimizes the DNN to maximize expected OSQA scores by:
- Sampling outputs from the DNN-estimated PDF,
- Computing OSQA scores for each sample,
- Subtracting the baseline average score,
- Updating DNN parameters with a stochastic gradient weighted by score improvement:
This framework yields measurable improvements in OSQA metrics, even when MSE is not minimized (Koizumi et al., 2018).
For fairness in probabilistic classification, score transformation modifies outputs to fair scores using a closed-form solution derived via Lagrangian duality. The transformation is parameterized by a vector encoding fairness corrections:
Score-based postprocessing or pre-processing yields outputs with bounded fairness disparity and preserves score-based metrics (Brier score, AUC) (Wei et al., 2019).
6. Multi-Objective Score Spaces and Preference Aggregation
Applying the probability integral transform (PIT), raw objectives are mapped into comparable scores:
This uniform scaling enables aggregation and ordering of multi-objective outputs for SAE systems. Trade-off mappings between preference space and score space are learned via neural networks, accommodating inherent non-linearities. This process improves proximity to desired trade-offs and eliminates auxiliary scaling parameters, facilitating more interpretable and user-aligned output modifications (Hönel et al., 2022).
7. Statistical Calibration and Score Aggregation
The scale-invariant Continuous Ranked Probability Score (CRPS) is reparameterized using the probability integral transform:
with the quantile. The transformed score
lies in and aggregates across variables via convolution:
Closed-form CDFs using Fresnel integrals enable exact p-value tests for statistical accuracy, transferable to SAE output evaluation. The method is robust to scale and sensitive to median forecast proximity, though insensitive to location bias (Nane et al., 2023).
8. Theoretical Advances: Change of Variables and Flexible Score Matching
Score functions transform under smooth invertible mappings as:
This enables decoupling the SAE training and generation spaces, guiding output correction via the transformed score gradient. Generalized Sliced Score Matching further extends loss projection from linear to arbitrary smooth functions, capturing nonlinear data structures for more flexible SAE output refinement (Robbins, 10 Dec 2024).
Conclusion
SAE output modification using scores is a multifaceted framework encompassing selection, adaptation, fairness correction, semantic alignment, aggregation, and real-time steering. It harnesses quantitative, interpretable functions—from activation-based criteria and semantic distances to statistical calibration and fairness constraints—for principled intervention in autoencoder models and broader DNNs. Theoretical advances in score transformation, variable change, and matching losses broaden the scope for fine-grained, user-aligned, and performance-competitive output control. Empirical results consistently demonstrate enhancements over baseline and unsupervised methods, while ongoing research continues to refine the targeting and interpretability of these score-driven interventions.