Scalable Backdoor Detection in Neural Networks (2006.05646v1)
Abstract: Recently, it has been shown that deep learning models are vulnerable to Trojan attacks, where an attacker can install a backdoor during training time to make the resultant model misidentify samples contaminated with a small trigger patch. Current backdoor detection methods fail to achieve good detection performance and are computationally expensive. In this paper, we propose a novel trigger reverse-engineering based approach whose computational complexity does not scale with the number of labels, and is based on a measure that is both interpretable and universal across different network and patch types. In experiments, we observe that our method achieves a perfect score in separating Trojaned models from pure models, which is an improvement over the current state-of-the art method.
- Haripriya Harikumar (8 papers)
- Vuong Le (22 papers)
- Santu Rana (68 papers)
- Sourangshu Bhattacharya (26 papers)
- Sunil Gupta (78 papers)
- Svetha Venkatesh (160 papers)