Papers
Topics
Authors
Recent
2000 character limit reached

Zero-Shot Attribution for Large Language Models: A Distribution Testing Approach (2506.20197v1)

Published 25 Jun 2025 in cs.LG, cs.AI, and cs.SE

Abstract: A growing fraction of all code is sampled from LLMs. We investigate the problem of attributing code generated by LLMs using hypothesis testing to leverage established techniques and guarantees. Given a set of samples $S$ and a suspect model $\mathcal{L}*$, our goal is to assess the likelihood of $S$ originating from $\mathcal{L}*$. Due to the curse of dimensionality, this is intractable when only samples from the LLM are given: to circumvent this, we use both samples and density estimates from the LLM, a form of access commonly available. We introduce $\mathsf{Anubis}$, a zero-shot attribution tool that frames attribution as a distribution testing problem. Our experiments on a benchmark of code samples show that $\mathsf{Anubis}$ achieves high AUROC scores ( $\ge0.9$) when distinguishing between LLMs like DeepSeek-Coder, CodeGemma, and Stable-Code using only $\approx 2000$ samples.

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.