What to do after the poisoning attack? #1636
-
Hi, I try to figure out how to use the If I executed the function do I have to put my poisoned examples back into the original training data or do I need to declare an extra poisoned dataset? And after testing I have to do a predicition with the poisoned sample right? - Because I can only trigger my backdoor if the pattern is inside an image. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 9 replies
-
Yes, these poisoning attacks create and return a poisoned dataset or samples by calling the method
After generating the poisoned dataset you have to used it to train a model. Then evaluate the model on test datasets samples that are benign and on samples containing the backdoor. This notebook contains the complete workflow: https://github.com/Trusted-AI/adversarial-robustness-toolbox/blob/main/notebooks/poisoning_attack_clean_label_backdoor.ipynb |
Beta Was this translation helpful? Give feedback.
Hi @EvilWatermelon
Yes, these poisoning attacks create and return a poisoned dataset or samples by calling the method
poison
.PoisoningAttackBackdoor
applies the poison to all samples and you have to mix them with benign samples yourself if you don't want to have all samples poisoned or use in combination withHiddenTriggerBackdoor
andPoisoningAttackCleanLabelBackdoor
which have arguments defining the fraction of samples that should be poisonous.After generating the poisoned dataset you have to used it to train a model. Then evaluate the model on test datasets samples that are benign and on samples containing the backdoor.
This notebook contains the complete workflow: https://github.com…