Defenses Functions
Available functions:
adversarial_training
(model, x, y, epsilon=0.01)
: Adversarial Training defense.feature_squeezing
(model, bit_depth=4)
: Feature Squeezing defense.gradient_masking
(model, mask_threshold=0.1)
: Gradient Masking defense.input_transformation
(model, transformation_function=None)
: Input Transformation defense.defensive_distillation
(model, teacher_model, temperature=2)
: Defensive Distillation defense.
Adversarial Training
Adversarial training is a method where the model is trained on both the original and adversarial examples, aiming to make the model more robust to adversarial attacks.
Feature Squeezing
Feature squeezing reduces the number of bits used to represent the input features, which can remove certain adversarial perturbations.
Gradient Masking
Gradient masking modifies the gradients during training to make them less informative for adversarial attackers.
Input Transformation
Input transformation applies a transformation to the input data before feeding it to the model, aiming to remove adversarial perturbations.
Defensive Distillation
Defensive distillation trains a student model to mimic the predictions of a teacher model, which is often a more robust model.
Last updated