π‘ Guidelineο
Use python 3.10
git clone git@github.com:IRT-SystemX/adaro-rl.git
pip install -e adaro-rl
π¦ Content of the Libraryο
adaro_rl/
βββ attacks/
βββ pipelines/
βββ wrappers/
βββ zoo/
βββ __init__.py
src/adaro_rl/attacks/contains the implementation of many adversarial attacks in a unified framework.src/adaro_rl/pipelines/contains the implementation of scripts to train, test, attack, and adversarially train agents.src/adaro_rl/wrappers/contains the implementations of wrappers of gymnasium environments to apply attacks from different perspectives (the agent or the adversary) and with different supports for perturbation (observations or environment).src/adaro_rl/zoo/contains the implementation of tools and functionality to run the scripts.
π‘οΈ Adversarial Attacksο
Available Methodsο
Random Attacksο
RandomUniformAttackorRUA: generates perturbation with random uniform sampling.RandomNormalAttackorRNA: generates perturbation with random normal sampling.RandomSignedAttackorRSA: generates perturbation with random sampling between -1 and 1.
Fast Gradients Method Based Attacksο
FGM_DforFastGradientMethod_DiscreteAction: generates perturbation with fast gradient method to attack discrete action policies.FGM_CforFastGradientMethod_ContinuousAction: generates perturbation with fast gradient method to attack continuous action policies.FGM_VforFastGradientMethod_V_Critic: generates perturbation with fast gradient method to attack value critics.FGM_QCforFastGradientMethod_Q_Critic: generates perturbation with fast gradient method to attack the q critic of a q actor-critic agent.FGM_QACforFastGradientMethod_Q_ActorCritic: generates perturbation with fast gradient method to attack both models of a q actor-critic agent.
Fast Gradients Method Based Attacksο
FGSM_Dfor the sign version ofFGM_D.FGSM_Cfor the sign version ofFGM_C.FGSM_Vfor the sign version ofFGM_V.FGSM_QCfor the sign version ofFGM_QC.FGSM_QACfor the sign version ofFGM_QAC.
Parameters of the Attacksο
The common parameters for all attacks are the following:
epsdefines the amount of perturbation applied with an attack.normdefines the norm of the delta. Options: 0, 1, 2, β¦ inf.max_epsdefines the maximum amount allowed on each dimension of the perturbation. It can be a number, the same for each dimension. It can also be an array, defining a specific amount of perturbation for each dimension. By default, max_eps is set as delta.is_proportional_maskis an array of booleans that defines for each dimension, if the perturbation has to be applied proportionally to the value of the support on which the perturbation is applied. By default, is_prop_perturb_mask is None.
Other parameters for gradient attacks are the following:
targetdefines the target followed by the attack. Options: βuntargetedβ, βtargetedβ, βminβ, βmaxβ or βtarget_fctβ. Default is βuntargetedβ.
π οΈ Usageο
Usage of Attacks Onlyο
The first way to use the ADARO-RL library is to directly apply the attack within your infrastructure with your training and testing scripts.
from adaro_rl.attacks import make_attack
attack = make_attack(attack_name, **attack_kwargs)
perturbation = attack.generate_perturbation(observation)
adv_obs = observation + perturbation
Here, use one of the
attack_namelisted inadaro_rl.attacks.attack_names.Add the
attack_kwargsneeded for the attack_name specified.Use an
observationof your environment|dataset.
Usage with the Scriptsο
The second way to use the ADARO-RL library is to use the application available in adaro_rl in command line. This enables to train, test, attack, and adversarial train agents, it is needed to choose a configuration of an existing use case provided in the adaro_rl.zoo, add your configurations in the zoo, or provide a new zoo module to use.
Command Linesο
Here are some basic command line examples :
adaro_rl train --config 'HalfCheetah-v5' --zoo 'adaro_rl.zoo'
adaro_rl test --config 'HalfCheetah-v5' --zoo 'adaro_rl.zoo'
adaro_rl online_attack --config 'HalfCheetah-v5' --zoo 'adaro_rl.zoo' --attack-name 'FGM_D' --eps 0.01 --norm "2"
The different pipelines that can be executed by the application are the following:
traintestonline_attackadversarial_train
Use any given script with the option --help or check the source codes available in adaro-rl/pipelines to see how to use them.
Pythonο
You can also use the scripts directly in Python. Here are some examples equivalent to the ones presented above in command lines:
import adaro_rl.zoo as zoo
config = zoo.configs['HalfCheetah-v5']
adaro_rl.train(config=config)
adaro_rl.test(config=config)
adaro_rl.online_attack(config=config, attack_name='FGM_D', eps=0.1, norm=2)
Examplesο
Example scripts and jupyter notebooks are available in examples folder of the repository. You will find there :
Attacks examples and
examples/attacks.py: a jupyter notebook and a python script showing the usages of the attacks available.Pipelines examples and
examples/pipelines.sh: a jupyter notebook and a bash script showing the usages of the high level pipelines available.