Creating and running an attack¶
After setting up a pretrained model, we can now create an attack.
Importing the attack class¶
Attacks in torchattack are exposed as classes.
To create the classic attack — Fast Gradient Sign Method (FGSM), for instance, we can import the FGSM
class from torchattack.
import torch
from torchattack import AttackModel, FGSM
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = AttackModel.from_pretrained(model_name='resnet50').to(device)
transform, normalize = model.transform, model.normalize
attack = FGSM(model, normalize, device)
Note the three arguments, model
, normalize
, and device
.
model
specifies the pretrained model to attack, often acting as the surrogate model for transferable black-box attacks.normalize
is the normalize function associated with the model's weight, that we automatically resolved earlier.device
is the device to run the attack on.
All attacks in torchattack, if not explicitly stated, require these three arguments. (1)
- This is the thin layer of abstraction that we were talking about.
Most adversarial attacks, especially gradient-based attacks, are projected onto the \(\ell_p\) ball, where \(p\) is the norm, to restrict the magnitude of the perturbation and maintain imperceptibility.
Attacks like FGSM takes an additional argument, eps
, to specify the radius \(\varepsilon\) of the \(\ell_\infty\) ball.
For an 8-bit image with pixel values in the range of \([0, 255]\), the tensor represented image is within \([0, 1]\) (after division by 255). As such, common values for eps
are 8/255
or 16/255
, or simply 0.03
.
Different attacks hold different arguments. For instance, the Momentum-Iterative FGSM (MI-FGSM) attack accepts the additional steps
and decay
as arguments.
from torchattack import MIFGSM
attack = MIFGSM(model, normalize, device, eps=8 / 255, steps=10, decay=1.0)
Finally, please take a look at the actual implementation of FGSM here torchattack.FGSM
(expand collapsed fgsm.py
source code).
The create_attack
method¶
torchattack provides an additional helper, create_attack
, to create an attack by its name.
To initialize the same FGSM attack.
To specify the common eps
argument.
For additional attack specific arguments, pass them as keyword arguments.
For generative attacks (BIA, CDA, and LTP, etc.) that do not require passing the model
and normalize
arguments.
attack = create_attack('BIA', device, eps=10 / 255, weights='DEFAULT')
attack = create_attack('CDA', device, eps=10 / 255, checkpoint_path='path/to/checkpoint.pth')
Generative attacks are models specifically trained to generate adversarial perturbation directly.
These models have already been given a pretrained model to attack, which acts as the discriminator in their setups. In inference mode, they load checkpoints like any other pretrained model from torchvision. The checkpoints provided by torchattack are sourced from their original repositories.
Running the attack¶
Similar to PyTorch models, torchattack attacks implement the forward
method, to run the attack on a batch of images.
Calling the attack class itself also directs to the forward
method.
Here is a simple example for a batch of dummy images of shape \((4, 3, 224, 224)\) (typical ImageNet images), and their corresponding labels (1000 classes).