ART 1.7.0
This release of ART v1.7.0 introduces many new evasion and inference attacks providing support for the evaluation of malware or tabular data classification, new query-efficient black-box (GeoDA) and strong white-box (Feature Adversaries) evaluation methods. Furthermore, this release introduces an easy to use estimator for Espresso ASR models to facilitate ASR research and connect Espresso and ART. This release also introduces support for binary classification with single outputs in neural networks classifiers and selected attacks. Many more new features and details can be found below:
Added
- Added LowProFool evasion attack for imperceptible attacks on tabular data classification in
art.attacks.evasion.LowProFool
. (#1063) - Added Over-the-Air-Flickering attack in PyTorch for evasion on video classifiers in
art.attacks.evasion.OverTheAirFlickeringPyTorch
. (#1077, #1102) - Added API for speech recognition estimators compatible with Imperceptible ASR attack in PyTorch. (#1052)
- Added Carlini&Wagner evasion attack with perturbations in L0-norm in
art.attacks.evasion.CarliniL0Method
. (#844, #1109) - Added support for Deep Speech v3 in
PyTorchDeepSpeech
estimator. (#1107) - Added support for TensorBoard collecting evolution of norms (L1, L2, and Linf) of loss gradients per batch, adversarial patch, and total loss and its model-specific components where available (e.g. PyTochFasterRCNN) in
AdversarialPatchPyTorch
,AdversarialPatchTensorFlow
,FastGradientMethod
, and allProjectedGradientDescent*
attacks. (#1071) - Added
MalwareGDTensorFlow
attack for evasion on malware classification of portable executables supporting append based, section insertion, slack manipulation, and DOS header attacks. (#1015) - Added Geometric Decision-based Attack (GeoDA) in
art.attacks.evasion.GeoDA
for query-efficient black-box attacks on decision labels using DCT noise. (#1001) - Added Feature Adversaries framework-specific in PyTorch and TensorFlow v2 as efficient white-box attack generating adversarial examples imitating intermediate representations at multiple layers in
art.attacks.evasion.FeatureAdversaries*
. (#1128, #1142, #1156) - Added attribute inference attack based on membership inference in
art.attacks.inference.AttributeInferenceMembership
. (#1132) - Added support for binary classification with neural networks with a single output neuron in
FastGradientMethod
, and allProjectedGradientDescent*
attacks. Neural network binary classifiers with a single output require settingnb_classes=2
and labelsy
in shape (nb_samples, 1) or (nb_samples,) containing 0 or 1. Backward compatibility for binary classifiers with two outputs is guaranteed withnb_classes=2
and labelsy
one-hot-encoded in shape (nb_samples, 2). (#1118) - Added estimator for Espresso ASR models in
art.estimators.speech_recognition.PyTorchEspresso
with support for attacks withFastGradientMethod
,ProjectedGradientDescent
andImperceptibleASRPyTorch
. (#1036) - Added deprecation warnings for
art.classifiers
andart.wrappers
to be replace withart.estimators
. (#1154)
Changed
- Changed
art.utils.load_iris
to use Iris dataset fromsklearn.datasets
instead ofarchive.ics.uci.edu
. (#1097 ) - Changed
HopSkipJump
to check for NaN in the adversarial example candidates and return original (benign) sample if at least one NaN is detected. (#1124) - Changed
SquareAttack
to accept user-defined loss and adversarial criterium definitions to enable black-box attacks on all machine learning tasks on images beyond classification. (#1127) - Changed
PyTorchFasterRCNN.loss_gradients
to process each sample separately to avoid issues with gradient propagation withtorch>=1.7
. (#1138)
Removed
[None]