Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TRACE: TRansformer-based Attribution using Contrastive Embeddings in LLMs (2407.04981v1)

Published 6 Jul 2024 in cs.CL and cs.LG

Abstract: The rapid evolution of LLMs represents a substantial leap forward in natural language understanding and generation. However, alongside these advancements come significant challenges related to the accountability and transparency of LLM responses. Reliable source attribution is essential to adhering to stringent legal and regulatory standards, including those set forth by the General Data Protection Regulation. Despite the well-established methods in source attribution within the computer vision domain, the application of robust attribution frameworks to natural language processing remains underexplored. To bridge this gap, we propose a novel and versatile TRansformer-based Attribution framework using Contrastive Embeddings called TRACE that, in particular, exploits contrastive learning for source attribution. We perform an extensive empirical evaluation to demonstrate the performance and efficiency of TRACE in various settings and show that TRACE significantly improves the ability to attribute sources accurately, making it a valuable tool for enhancing the reliability and trustworthiness of LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Cheng Wang (386 papers)
  2. Xinyang Lu (15 papers)
  3. See-Kiong Ng (103 papers)
  4. Bryan Kian Hsiang Low (77 papers)

Summary

We haven't generated a summary for this paper yet.