Knowledge Attribution Score (KAS)
- Knowledge Attribution Score (KAS) is a continuous metric that measures the support or contradiction of textual claims by leveraging structured knowledge graphs.
- It integrates per-claim scoring, triplet similarity, and asymmetric normalization to deliver granular, reliable assessments for fact-checking and source attribution.
- KAS is applied in both claim verification and retrieval-augmented generation frameworks, guiding decision-making in high-stakes, automated information pipelines.
The Knowledge Attribution Score (KAS) is a continuous metric designed to quantify the degree to which textual content or database items are supported or implicated by underlying knowledge structures. It serves as both a reliability indicator in claim-level fact verification against structured knowledge graphs and a responsibility score for source tracing in retrieval-augmented generation (RAG) systems. In recent research, KAS arises in two principal frameworks: ClaimVer for claim-level text verification against trusted knowledge graphs (Dammu et al., 12 Mar 2024), and RAGOrigin for identifying the origins of corrupted outputs in poisoned RAG databases (Zhang et al., 17 Sep 2025). KAS outputs a scalar score, typically normalized or bounded, enabling granular downstream decision-making in high-stakes information pipelines.
1. Motivation and Conceptual Rationale
KAS emerged from a need for fine-grained validation and attribution under conditions of uncertain, incomplete, or adversarially manipulated knowledge. Traditional fact-checking methods provide discrete judgments (“true/false/unsupported”) at the sentence or document level and lack actionable, interpretable feedback. KAS addresses these limitations through:
- Granularity: Each claim or candidate text is scored for degree of support, contradiction, or influence, rather than only binary adjudication.
- Continuity: Enables ranking, thresholding, and learning from a real-valued scale instead of coarse categories.
- Asymmetry: Heavier penalization for misinformation versus commensurate rewarding of robust support, preventing overconfident aggregation when contradictory evidence is present.
- Attribution: Traces outcome responsibility to specific evidence, claims, or database items, supporting explainability and accountability in automated systems.
- Downstream Utility: Applicability to ranking candidates, gating for publication, and training feedback loops.
These features position KAS as a foundational reliability signal in modern knowledge-centric NLP applications.
2. Mathematical Formulation and Computation
ClaimVer Framework (Dammu et al., 12 Mar 2024)
KAS in ClaimVer is defined for an input text partitioned into claims:
- Per-claim score:
- Triplet Match Score (TMS):
- is the cosine similarity between claim and triplet text embeddings.
- is the fraction of claim entities covered by triplet entities.
- Hyperparameters: (default).
- Aggregate and non-symmetric normalization:
RAGOrigin Framework (Zhang et al., 17 Sep 2025)
KAS (called the “responsibility score”) for candidate document in narrowed attribution scope :
- Raw signals:
- Retrieval ranking
- Semantic relevance (proxy LLM log-prob of reconstructing question from )
- Influence-on-generation (proxy LLM log-prob of generating the incorrect answer from )
- Standardization:
- Aggregate score:
- Default weights .
- Clustering: -means on values, flagging the cluster with higher mean KAS as poisoned.
3. Computation Pipeline and Example Procedures
ClaimVer Pipeline
- NER and KG linking of claim spans.
- Coreference resolution for standalone claim formation.
- Triplet retrieval via Woolnet (multi-node BFS up to 3 hops, max 4 paths).
- Claim attribution: LLM outputs + rationale.
- Per-claim KAS calculation: Assign cs, compute TMS, aggregate.
- Final score: Sum weighted contributions, apply asymmetric sigmoid.
Example
For the claim “Apollo 11 landed on the Moon in 1969,” relevant KG triplets are retrieved, leading to:
- Attributable, cs
- , , TMS
- Weighted sum: $1.63$; KAS
RAGOrigin Pipeline
- Scope narrowing: Retrieve and chunk knowledge base by similarity to question; blocks tested for their ability to reproduce the misgenerated answer.
- Score computation: For candidates in the final scope, calculate .
- Z-normalization: Standardize scores.
- Score aggregation: Compute KAS for each candidate.
- Clustering: Apply k-means; flag items as poisoned according to cluster mean.
- Dynamic adaptation: No manual thresholds; block size and weights tested for stability.
4. Properties, Sensitivities, and Hyperparameter Effects
ClaimVer (Dammu et al., 12 Mar 2024)
- Range:
- Interpretability: KAS near zero implies strong contradiction; near one, strong support.
- Score Dynamics: Attributable labels drive KAS up by TMS; contradictions penalize via TMS and asymmetric .
- Sensitivity: Weak triplet matches (low , ) yield low impact even for “Attributable” claims; contradictions with moderate match are aggressively penalized.
- Granularity: “Extrapolatory” claims receive partial credit; omission or neutral triplets do not contribute.
- Extensibility: Parameters can be fine-tuned as needed.
RAGOrigin (Zhang et al., 17 Sep 2025)
- Standardization: Z-scoring of all components is essential for consistent thresholding.
- Component contribution: Ablations indicate all three signals are necessary for high accuracy; absence of z-normalization or adaptive scope sharply degrades detection.
- Block size and model selection: Robust to parameter variations; default , weights = 1, proxy model flexible.
5. Integration and Downstream Applications
ClaimVer
- Pipeline role: Final reliability metric for multi-claim text.
- Uses: Ranks candidate texts, thresholds for review/publishing, fine-tuning models to maximize factuality, and interactive UI feedback showing attribution scores.
- User trust: Provides both fine-grained explanations and a summary scalar, aiding transparency and cognitive load reduction.
RAGOrigin
- Provenance forensics: Identifies and isolates sources of generation corruption (knowledge poisoning).
- Automated operation: Works in black-box manner across multiple datasets, attack types, retriever models, and LLMs.
- No manual thresholding: Clustering adapts automatically to context and event.
- Efficiency: Average $2$ seconds, \$0.0002 per query.
6. Empirical Findings and Observed Performance
ClaimVer (Dammu et al., 12 Mar 2024)
- Examples: High KAS correlates with claims marked mostly “Attributable”; low KAS dominated by “Contradictory.”
- Classification F1: Underlying claim classification reaches F1.
- Quality correlation: Higher KAS aligns with factual user impressions and clear KG evidence.
- Asymmetry effect: empirically selected; ablations suggest symmetric sigmoid over-penalizes contradictions.
RAGOrigin (Zhang et al., 17 Sep 2025)
- Detection Accuracy (DACC): across 5 QA datasets and 9 attacks; FPR , FNR .
- Removal efficacy: Flags enable removal of poisoned texts and drop attack success rates to zero, outperforming prior techniques (RAGForensics).
- Robustness: Stable across paraphrased questions, varying poisoned document numbers, different retrievers and LLMs, multi-attacker scenarios, adaptive attacks, and dynamic database changes.
- Component ablations: Use of all signals and adaptive scope is critical for maximal robustness.
7. Context and Plausible Implications
KAS bridges interpretable reliability assessment and source attribution in high-stakes information systems. In ClaimVer, KAS empowers nuanced trust modeling and granular evidence weighting in claim verification, allowing stakeholders to differentiate content based on degrees of knowledge graph support. In RAGOrigin, KAS enables forensic traceability of corrupted knowledge, mitigating the risk of semantic poisoning and unobtrusively adapting to attack variations and database volatility.
A plausible implication is that KAS will be increasingly central in both development of robust fact-checking pipelines and real-time provenance tracking in retrieval-augmented LLMs. Its formalization supports integration into broader feedback loops, user-facing reliability visualization, and autonomous review or gating of model outputs, providing a principled tool for both explainability and operational defense in automated knowledge systems.