Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Model Extraction and Adversarial Transferability, Your BERT is Vulnerable! (2103.10013v1)

Published 18 Mar 2021 in cs.CL

Abstract: Natural language processing (NLP) tasks, ranging from text classification to text generation, have been revolutionised by the pre-trained LLMs, such as BERT. This allows corporations to easily build powerful APIs by encapsulating fine-tuned BERT models for downstream tasks. However, when a fine-tuned BERT model is deployed as a service, it may suffer from different attacks launched by malicious users. In this work, we first present how an adversary can steal a BERT-based API service (the victim/target model) on multiple benchmark datasets with limited prior knowledge and queries. We further show that the extracted model can lead to highly transferable adversarial attacks against the victim model. Our studies indicate that the potential vulnerabilities of BERT-based API services still hold, even when there is an architectural mismatch between the victim model and the attack model. Finally, we investigate two defence strategies to protect the victim model and find that unless the performance of the victim model is sacrificed, both model ex-traction and adversarial transferability can effectively compromise the target models

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Xuanli He (43 papers)
  2. Lingjuan Lyu (131 papers)
  3. Qiongkai Xu (33 papers)
  4. Lichao Sun (186 papers)
Citations (87)

Summary

We haven't generated a summary for this paper yet.