RESEARCH

CORE MACHINE LEARNING

Adversarial Example Games

December 03, 2020

Abstract

The existence of adversarial examples capable of fooling trained neural network classifiers calls for a much better understanding of possible attacks to guide the development of safeguards against them. This includes attack methods in the challenging non-interactive blackbox setting, where adversarial attacks are generated without any access, including queries, to the target model. Prior attacks in this setting have relied mainly on algorithmic innovations derived from empirical observations (e.g., that momentum helps), lacking principled transferability guarantees. In this work, we provide a theoretical foundation for crafting transferable adversarial examples to entire hypothesis classes. We introduce Adversarial Example Games (AEG), a framework that models the crafting of adversarial examples as a min-max game between a generator of attacks and a classifier. AEG provides a new way to design adversarial examples by adversarially training a generator and a classifier from a given hypothesis class (e.g., architecture). We prove that this game has an equilibrium, and that the optimal generator is able to craft adversarial examples that can attack any classifier from the corresponding hypothesis class. We demonstrate the efficacy of AEG on the MNIST and CIFAR-10 datasets, out- performing prior state-of-the-art approaches with an average relative improvement of 27.5% and 47.2% against undefended and robust models respectively.

Download the Paper

AUTHORS

Written by

Pascal Vincent

Hugo Berard

Andre Cianflone

Gauthier Gidel

Joey Bose

Simon Lacoste-Julien

Will Hamilton

Publisher

NeurIPS

Research Topics

Core Machine Learning

Related Publications

August 12, 2024

CORE MACHINE LEARNING

Contrastive Predict-and-Search for Mixed Integer Linear Programs

Arman Zharmagambetov, Yuandong Tian, Aaron Ferber, Bistra Dilkina, Taoan Huang

August 12, 2024

August 09, 2024

CORE MACHINE LEARNING

Benchmarking Attacks on Learning with Errors

Emily Wenger, Eshika Saxena, Mohamed Malhou, Ellie Thieu, Kristin Lauter

August 09, 2024

August 02, 2024

CORE MACHINE LEARNING

GenCO: Generating Diverse Designs with Combinatorial Constraints

Arman Zharmagambetov, Yuandong Tian

August 02, 2024

July 29, 2024

COMPUTER VISION

CORE MACHINE LEARNING

Factorizing Text-to-Video Generation by Explicit Image Conditioning

Rohit Girdhar, Mannat Singh, Andrew Brown, Quentin Duval, Samaneh Azadi, Saketh Rambhatla, Mian Akbar Shah, Xi Yin, Devi Parikh, Ishan Misra

July 29, 2024

Help Us Pioneer The Future of AI

We share our open source frameworks, tools, libraries, and models for everything from research exploration to large-scale production deployment.