- The paper introduces Zero-NatVer, a system that leverages instruction-tuned LLMs to perform fact verification without relying on annotated training data.
- It employs a structured pipeline including chunking, alignment, NatOp assignment, and DFA-based proof execution to generate explainable verdicts.
- Evaluations show an average accuracy improvement of 8.96 points over baselines and demonstrate strong zero-shot transfer across diverse datasets.
Zero-Shot Fact Verification via Natural Logic and LLMs
The paper under review presents a novel approach to fact verification (FV) by leveraging zero-shot capabilities of instruction-tuned LLMs, specifically implementing a method named Zero-NatVer. This approach is grounded in natural logic (NatLog), thereby enhancing explainability and providing faithful justifications for the verification process.
Major Contributions and Findings
Zero-NatVer addresses the dependency of traditional NatLog-based FV systems on substantial annotated training data by employing zero-shot methods to operate without such data. The system uses a natural logic framework bolstered by the generalization potential of instruction-tuned LLMs, offering an alternative that reduces the need for data-intensive operations.
In evaluations, the proposed method was compared with existing fact verification systems in two setups: zero-shot generalization and zero-shot transfer. The results reveal several significant findings:
- Zero-Shot Generalization: Zero-NatVer demonstrated superior accuracy, with an average improvement of 8.96 points over the best-performing baseline when evaluated on English datasets. Additionally, when compared to traditional direct-QA methods using the same backbone models, Zero-NatVer showed competitive performance, underscoring its effectiveness in generating explainable results while maintaining accuracy.
- Zero-Shot Transfer: Tested across various datasets, Zero-NatVer outperformed other NatLog-based systems that relied on FEVER-like data for training. This suggests that the method can effectively generalize to different domains without requiring retraining on domain-specific data.
Methodology
The authors implement Zero-NatVer in a structured pipeline comprising four steps:
- Chunking: The claims are segmented into smaller chunks that can be independently verified. This segmentation is performed using constrained LLM decoding to prevent hallucination.
- Alignment: The claim chunks are aligned with corresponding evidence. This process includes generating alignment explanations to transfer global information for subsequent inference stages.
- NatOp Assignment: Built on a question-answering ensemble framework, Zero-NatVer assigns natural logic operators (NatOps) to aligned pairs. Calibration issues are mitigated using a weighted ensemble of 10 question prompts per candidate NatOp.
- Proof Execution: Proofs are executed on a deterministic finite automaton (DFA), producing the claim's final verdict.
Evaluation and Results
The methodology is evaluated on a diverse set of datasets containing both real-world and artificial claims. On English datasets, Zero-NatVer outperformed its NatLog baselines and was competitive with non-NatLog direct-answering models. Zero-NatVer displayed strong performance across multilingual datasets without needing language-specific modifications, demonstrating the potential for broad applicability.
Implications and Future Directions
The implementation of Zero-NatVer addresses the scalability concerns associated with data-dependent natural logic systems, enabling broader applications in domains where labeled training data is scarce or costly to produce. The results highlight the feasibility of leveraging LLMs for zero-shot fact-checking with high accuracy, providing valuable insights into the combination of neural and symbolic reasoning for explainability.
Future investigations could focus on further refining NatOp assignments, exploring more sophisticated techniques to resolve multi-NatOp conflicts, and enhancing the expressivity of natural logic to capture more complex reasoning phenomena. Moreover, adaptations to specific domains or languages could be explored to further extend the reach and applicability of zero-shot fact verification systems.