Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TOPFORMER: Topology-Aware Authorship Attribution of Deepfake Texts with Diverse Writing Styles (2309.12934v3)

Published 22 Sep 2023 in cs.CL

Abstract: Recent advances in LLMs have enabled the generation of open-ended high-quality texts, that are non-trivial to distinguish from human-written texts. We refer to such LLM-generated texts as deepfake texts. There are currently over 72K text generation models in the huggingface model repo. As such, users with malicious intent can easily use these open-sourced LLMs to generate harmful texts and dis/misinformation at scale. To mitigate this problem, a computational method to determine if a given text is a deepfake text or not is desired--i.e., Turing Test (TT). In particular, in this work, we investigate the more general version of the problem, known as Authorship Attribution (AA), in a multi-class setting--i.e., not only determining if a given text is a deepfake text or not but also being able to pinpoint which LLM is the author. We propose TopFormer to improve existing AA solutions by capturing more linguistic patterns in deepfake texts by including a Topological Data Analysis (TDA) layer in the Transformer-based model. We show the benefits of having a TDA layer when dealing with imbalanced, and multi-style datasets, by extracting TDA features from the reshaped $pooled_output$ of our backbone as input. This Transformer-based model captures contextual representations (i.e., semantic and syntactic linguistic features), while TDA captures the shape and structure of data (i.e., linguistic structures). Finally, TopFormer, outperforms all baselines in all 3 datasets, achieving up to 7\% increase in Macro F1 score. Our code and datasets are available at: https://github.com/AdaUchendu/topformer

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Whodunit? learning to contrast for authorship attribution. In AACL, pages 1142–1157, 2022.
  2. Extracting training data from large language models. In USENIX Security Symposium, volume 6, 2021.
  3. Token prediction as implicit classification to identify llm-generated text. arXiv preprint arXiv:2311.08723, 2023.
  4. Z. Chen and H. Liu. Stadee: Statistics-based deep detection of machine generated text. In International Conference on Intelligent Computing, pages 732–743. Springer, 2023.
  5. Acceptability judgements via examining the topology of attention maps. In Proceedings of the 2022 EMNLP, volume 2022, pages 88–107, 2022.
  6. Toxicity in chatgpt: Analyzing persona-assigned language models. In Proceedings of the 2023 EMNLP, 2023.
  7. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  8. P. Doshi and W. Zadrozny. Movie genre detection using topological data analysis. In Statistical Language and Speech Processing: 6th International Conference, SLSP 2018, Mons, Belgium, October 15–16, 2018, Proceedings 6, pages 117–128. Springer, 2018.
  9. Llm-as-a-coauthor: The challenges of detecting llm-human mixcase. arXiv preprint arXiv:2401.05952, 2024.
  10. Gltr: Statistical detection and visualization of generated text. ACL 2019, page 111, 2019.
  11. Topological signature of 19th century novelists: Persistent homology in text mining. big data and cognitive computing, 2(4):33, 2018.
  12. Story trees: Representing documents using topological persistence. In Proceedings of the Thirteenth LREC 2022, pages 2413–2429, 2022.
  13. A survey of topological machine learning methods. Frontiers in Artificial Intelligence, 4:681108, 2021.
  14. Automatic detection of generated text is easiest when humans are fooled. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1808–1822, 2020.
  15. A watermark for large language models. In International Conference on Machine Learning, pages 17061–17084. PMLR, 2023.
  16. Outfox: Llm-generated essay detection through in-context learning with adversarially generated examples. arXiv preprint arXiv:2307.11729, 2023.
  17. J-guard: Journalism guided adversarially robust detection of ai-generated news. In IJCNLP-AACL, pages 484–497, 2023a.
  18. Stylometric detection of ai-generated text in twitter timelines. arXiv preprint arXiv:2303.03697, 2023b.
  19. How reliable are ai-generated-text detectors? an assessment framework using evasive soft prompts. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 1337–1349, 2023c.
  20. Artificial text detection via examining the topology of attention maps. In Proceedings of the 2021 EMNLP, pages 635–649, 2021.
  21. Do language models plagiarize? In Proceedings of the ACM Web Conference 2023, pages 3637–3647, 2023.
  22. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
  23. Fighting fire with fire: The dual role of llms in crafting and detecting elusive disinformation. In Proceedings of the 2023 EMNLP, pages 14279–14305, 2023.
  24. Do neural networks trained with topological features learn different internal representations? In NeurIPS Workshop on Symmetry and Geometry in Neural Representations, pages 122–136. PMLR, 2023.
  25. Detectgpt: Zero-shot machine-generated text detection using probability curvature. In International Conference on Machine Learning, pages 24950–24962. PMLR, 2023.
  26. E. Munch. A user’s guide to topological data analysis. Journal of Learning Analytics, 4(2):47–61, 2017.
  27. I. Perez and R. Reinauer. The topological bert: Transforming attention into topology for natural language processing. arXiv preprint arXiv:2206.15195, 2022.
  28. Persistent topology of syntax. Mathematics in Computer Science, 12(1):33–50, 2018.
  29. Topological analysis of syntactic structures. arXiv preprint arXiv:1903.05181, 2019.
  30. A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922, 2023.
  31. Visualizing and measuring the geometry of bert. Advances in Neural Information Processing Systems, 32, 2019.
  32. D. Rosati. Synscipass: detecting appropriate uses of scientific text generation. In Proceedings of the Third Workshop on Scholarly Document Processing, pages 214–222, 2022.
  33. Topological data analysis for discourse semantics? In Proceedings of the 13th International Conference on Computational Semantics-Student Papers, pages 34–43, 2019.
  34. Societal biases in language generation: Progress and challenges. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4275–4293, 2021.
  35. Bert rediscovers the classical nlp pipeline. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4593–4601, 2019.
  36. Topological data analysis for speech processing. arXiv preprint arXiv:2211.17223, 2022.
  37. On the effectiveness of persistent homology. Advances in Neural Information Processing Systems, 35:35432–35448, 2022.
  38. Argumentative topology: Finding loop (holes) in logic. arXiv preprint arXiv:2011.08952, 2020.
  39. Authorship attribution for neural text generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 8384–8395, 2020.
  40. Turingbench: A benchmark environment for turing test in the age of neural text generation. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 2001–2016, 2021.
  41. Attribution and obfuscation of neural text authorship: A data mining perspective. SIGKDD Explorations, page vol. 25, 2023a.
  42. Does Human Collaboration Enhance the Accuracy of Identifying LLM-Generated Deepfake Texts? In 11th AAAI Conf. on Human Computation and Crowdsourcing (HCOMP), Nov. 2023b.
  43. Gpt-who: An information density-based machine-generated text detector. arXiv preprint arXiv:2310.06202, 2023.
  44. Dialogue term extraction using transfer learning and topological data analysis. In Proceedings of the 23rd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 564–581, 2022.
  45. M4: Multi-generator, multi-domain, and multi-lingual black-box machine-generated text detection. 2023.
  46. Topological analysis of contradictions in text. In Proceedings of the 45th International ACM SIGIR, pages 2478–2483, 2022.
  47. X. Zhu. Persistent homology: An introduction and a new text representation for natural language processing. In IJCAI, pages 1953–1959, 2013.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Adaku Uchendu (16 papers)
  2. Thai Le (38 papers)
  3. Dongwon Lee (65 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.